General Configurations if Big Data Processing License is Enabled

Figure 7-31 DMT Configurations window


This image displays the DMT Configurations window.

Table 7-10 Fields in the DMT Configuration window and their Description

Property Name Property Value
Generic
T2T Mode Select the mode of T2T to be used for execution of Data Mapping definition, from the list. The options are Default (for Java engine) and CPP (for CPP engine).
H2T Mode

Select the mode of H2T to be used for execution of Data Mapping definition, from the list. The options are Default, Sqoop, and OLH.

OLH (Oracle Loader for Hadoop) must have been installed and configured in your system. For more information on how to use OLH for H2T, see Oracle® Loader for Hadoop (OLH) Configuration section in OFS Analytical Applications Infrastructure Administration Guide.

Sqoop should have been installed and configured in your system. For more information, see the Sqoop Configuration section in OFS Analytical Applications Infrastructure Administration Guide. Additionally, you should register the cluster information of the source Information domain using the Register Cluster tab.

T2H Mode

Select the mode of T2H to be used for execution of Data Mapping definition, from the list. The options are Default and Sqoop.

For the Default option, additional configurations are required, which is explained in the Data Movement from RDBMS Source to HDFS Target (T2H) section in OFS Analytical Applications Infrastructure Administration Guide. Additionally, you should register the cluster information of the target Information domain using the Register Cluster tab.

For the Sqoop option, Sqoop should have been installed and configured in your system. For more information, see the Sqoop Configuration section in OFS Analytical Applications Infrastructure Administration Guide. Additionally, you should register the cluster information of the source Information domain using the Register Cluster tab.

PLC Mode Select the mode of execution to be used for Post Load Changes definition, from the list. The options are Default (for Java engine) and CPP (for CPP engine).
SCD MODE

This field is applicable only if SCD uses a merge approach.

  • DEFAULT_V1- Select this option to perform SCD execution using JAVA engine with a single Merge query for both Update and Insert. This is the default execution mode.
  • DEFAULT _V2- Select this option to perform SCD execution using JAVA engine with a Merge query for updates and Insert query for inserts. Since Insert is a separate query, the sequence used for SKEY will be incremented only for the required records making the SKEY column value continuous.
  • CPP_V1- Select this option to perform SCD execution using CPP engine with a single Merge query for both Update and Insert. This is the default execution mode.
  • CPP_V2- Select this option to perform SCD execution using CPP engine with a Merge query for updates and Insert query for inserts. Since Insert is a separate query, the sequence used for SKEY will be incremented only for the required records making the SKEY column value continuous.
  • BACKDATED_V1-Backdated support for CPP_V1.
  • BACKDATED_V2- Backdated support for CPP_V2.

Note: For the Backdated Executions containing type 2 column mappings, below column mappings are mandatory :

  • Start date
  • End date
Is Hive Local

This is applicable for T2H and F2H.

Select Yes if HiveServer is running locally to OFSAA, else select No, from the drop-down list.

Validate Definition Query on Save Select Yes to validate the SQL Query of the Data Mapping definition on save.
Generic Working Directory Specify the path of the HDFS working directory for generic operations. By default, the path is set as /user/ofsaa/GenericPath.
Allow Pre806 Data File Path

This field is applicable only in case of upgrade from an earlier version to 8.1.0.0.0 version. If yours is a fresh installation of 8.1.0.0.0 version using Full installer, this field is not applicable.

For F2T, the path for Data File in versions before 8.0.6.0.0 is /<ftpshare>/STAGE/<FileBasedSource>/<MISDate>/<dataFile.dat>. In 8.1.0.0.0, it is changed to /ftpshare/<INFODOM>/dmt/source/<Data Source Code>/data/<MISDATE>/<dataFile.dat>.

Select Yes to allow the old Data File path in 8.1.0.0.0 version.

SMG Mode

By default, the Source Model Generation (SMG) mode is set as Dictionary.

When SMG Mode is selected as Dictionary, the time taken for generating Source models of Views from the database is optimized.

Select Default for the earlier mode.

Allow Pre806 T2F File Path

In the versions before 8.0.6.0.0, the T2F extract file path is <ftpshare>/STAGE/<SOURCE_CODE>/<MISDATE>.

Select Yes, if you want to set the preceding extract path.

If you select No, the extract file path is set to <ftpshare>/<INFODOM>/dmt/def/<DEFINITION_CODE>/<BATCH_ID>_<TASK_ID>/<MISDATE>.

Sqoop

(This section is applicable only if you select Sqoop for T2H Mode or H2T Mode.).

Sqoop Mode

Select Client to execute Sqoop in client mode or select Cluster to execute Sqoop in cluster mood, from the drop-down list.

If you select Cluster as Sqoop Mode, you should register the cluster from Register Cluster tab. For more details, see Registering a Cluster.

Note: Copying of any Sqoop jars and Hadoop/Hive configuration XMLs to OFSAAI is not required in cluster mode.

Sqoop Working Directory Specify the path of the HDFS working directory for Sqoop related operations.

WebLog

(This section is applicable only for L2H)

Keep Weblog Processed File

Select Yes or No from the drop-down list.

Yes- The working directory will be retained with the processed WebLog files. If the data loading was successful, the WebLog file name will be appended with Processed. Else, the WebLog file name will be appended with Working.

No- The working directory will be deleted after data loading.

Weblog Temp File Ext Enter the extension of the Weblog temporary file.
Weblog Working Directory Enter the name of the temporary working directory in HDFS.
File Encryption
Encryption At rest Select Yes from the drop-down list, if encryption is required for T2F or H2F and decryption is required for F2T or F2H.
Key File Name Enter the name of the Key File that you used for encrypting the Data File.
Key File Path Enter the absolute path of the Key File that you used for encrypting the Data File.

Note:

You can use the BackendServerProperties.conf in the ficdb/conf layer to support the required Timezone and Time Format in the CPP logs.