- Administration and Configuration Guide
- Data Management Tools (DMT) Module Configurations
- Oracle® Loader for Hadoop (OLH) Configuration
- Steps for Configuring OLH
Steps for Configuring OLH
- Unzip the OLH Package downloaded from the Oracle site in the VM where OFSAAI is
installed.Location: Inside the Home Directory of the user where OFSAAI is installed.
- Set
OLH_HOMEenvironment variable in the.profilefile.OLH_HOME contains the directories such asbin,examples,jlibandlib.Step 2: Configuring the Property- Set the following property in the
jdbcOutput.xmlfile, which is present in the$FIC_DB_HOME/conf/location:<property><name>oracle.hadoop.loader.connection.defaultExecuteBatch</name><value>100</value></property> - From the DMT Configurations window, set H2T Mode as OLH.
- From the Register Cluster tab in the DMT Configurations window, register a cluster with Source Information domain name as the Cluster Name.
For more information, see the DMT Configurations section in the OFS Analytical Applications Infrastructure User Guide.Step 3: Copy Configuration XMLs from Hadoop Cluster- Copy the following files from the Hadoop Cluster to the Configuration
File Path given in the Cluster Configurations window of the registered
cluster.
core-site.xml hdfs-site.xml mapred-site.xml hive-site.xml yarn-site.xml
Note:
Only Client Configuration Properties are required. If Cloudera Manager is used, the same can be downloaded directly, which will contain only the client properties. - Modify the following property in the mapred-site.xml file in the
$FIC_HOME/conf
folder:
<property><name>mapred.child.java.opts</name><value>-Xmx4096m</value></property><property><name>mapreduce.job.outputformat.class</name><value>oracle.hadoop.loader.lib.output.JDBCOutputFormat</value></property><property><name>mapreduce.output.fileoutputformat.outputdir</name><value>(Any temporary directory)</value></property><property><name>oracle.hadoop.loader.defaultDateFormat</name><value>yyyy-MM-dd</value></property>Note:
If proxy user is enabled and the Job is submitted by the same, the user should be created in every node of the Hadoop Cluster. For more information, see CDH4 Security Guide. The version of jars to be copied differs depending upon the version of CDH configured.
Step 4: Copy the required Jars- Copy
commons-httpclient-3.1.jarfrom the CDH installation libraries into the$OLH_HOME/jlibfolder. - If OFSAA is using Apache driver:
- Usually jars such as
hive-exec-*.jar,libfb303-*.jar,hive-service-*.jar, hive-metastore-*.jarare present in theext/libfolder and are added to the Classpath. In case of any ClassNotFound Exception, perform the following steps:- Edit the
oracle.hadoop.loader.libjarsproperty present in theOLH_HOME/doc/oraloader-conf.xmlfile to accommodate the newly added jars. That is,$FIC_HOME/ext/lib/ hive-exec-*.jar(repeat for each of the mentioned jars) - Copy the entire property to the
FIC_DB_HOME/conf/dtextInput.xmlfile.
- Edit the
- If OFSAA is using Cloudera Connectors:
- Note that OLH is not qualified on Cloudera Connectors. Perform
the following workaround:
- Copy the following jars (Apache Drivers) to the
OLH_HOME/jlibfolder:hive-exec-*.jar, libfb303-*.jar, hive-service-*.jar, hive-metastore-*.jarUsually these jars are added to the Classpath. In case of any ClassNotFound Exception, perform the following steps:
- Edit the
oracle.hadoop.loader.libjarsproperty present inOLH_HOME/doc/oraloader-conf.xmlto accommodate the newly added jars. That is,${oracle.hadoop.loader.olh_home}/jlib/ hive-exec-*.jar
- Edit the
- Copy the following jars (Apache Drivers) to the
- Note that OLH is not qualified on Cloudera Connectors. Perform
the following workaround:
- Cop0y the entire property to the
FIC_DB_HOME/conf/dtextInput.xmlfile.Note:
Add the aforementioned jars only if OLH task is to be run. If any other OFSAA task is running, do not keep a copy of the jars in theOLH_HOME/jlibfolder.
- Set the following property in the