- Administration and Configuration Guide
- Data Management Tools (DMT) Module Configurations
- Oracle® Loader for Hadoop (OLH) Configuration
- Steps for Configuring OLH
Steps for Configuring OLH
- Unzip the OLH Package downloaded from the Oracle site in the VM where OFSAAI is
installed.Location: Inside the Home Directory of the user where OFSAAI is installed.
- Set
OLH_HOME
environment variable in the.profile
file.OLH_HOME contains the directories such asbin
,examples
,jlib
andlib
.Step 2: Configuring the Property- Set the following property in the
jdbcOutput.xml
file, which is present in the$FIC_DB_HOME/conf/
location:<property>
<name>oracle.hadoop.loader.connection.defaultExecuteBatch</name>
<value>100</value>
</property>
- From the DMT Configurations window, set H2T Mode as OLH.
- From the Register Cluster tab in the DMT Configurations window, register a cluster with Source Information domain name as the Cluster Name.
For more information, see the DMT Configurations section in the OFS Analytical Applications Infrastructure User Guide.Step 3: Copy Configuration XMLs from Hadoop Cluster- Copy the following files from the Hadoop Cluster to the Configuration
File Path given in the Cluster Configurations window of the registered
cluster.
core-site.xml
hdfs-site.xml
mapred-site.xml
hive-site.xml
yarn-site.xml
Note:
Only Client Configuration Properties are required. If Cloudera Manager is used, the same can be downloaded directly, which will contain only the client properties. - Modify the following property in the mapred-site.xml file in the
$FIC_HOME/conf
folder:
<property>
<name>mapred.child.java.opts</name>
<value>-Xmx4096m</value>
</property>
<property>
<name>mapreduce.job.outputformat.class</name>
<value>oracle.hadoop.loader.lib.output.JDBCOutputFormat</value>
</property>
<property>
<name>mapreduce.output.fileoutputformat.outputdir</name>
<value>(Any temporary directory)</value>
</property>
<property>
<name>oracle.hadoop.loader.defaultDateFormat</name>
<value>yyyy-MM-dd</value>
</property>
Note:
If proxy user is enabled and the Job is submitted by the same, the user should be created in every node of the Hadoop Cluster. For more information, see CDH4 Security Guide. The version of jars to be copied differs depending upon the version of CDH configured.
Step 4: Copy the required Jars- Copy
commons-httpclient-3.1.jar
from the CDH installation libraries into the$OLH_HOME/jlib
folder. - If OFSAA is using Apache driver:
- Usually jars such as
hive-exec-*.jar
,libfb303-*.jar
,hive-service-*
.jar, hive-metastore-*.jar
are present in theext/lib
folder and are added to the Classpath. In case of any ClassNotFound Exception, perform the following steps:- Edit the
oracle.hadoop.loader.libjars
property present in theOLH_HOME/doc/oraloader-conf.xml
file to accommodate the newly added jars. That is,$FIC_HOME/ext/lib/ hive-exec-*.jar
(repeat for each of the mentioned jars) - Copy the entire property to the
FIC_DB_HOME/conf/dtextInput.xml
file.
- Edit the
- If OFSAA is using Cloudera Connectors:
- Note that OLH is not qualified on Cloudera Connectors. Perform
the following workaround:
- Copy the following jars (Apache Drivers) to the
OLH_HOME/jlib
folder:hive-exec-*.jar, libfb303-*.jar, hive-service-*.jar, hive-metastore-*.jar
Usually these jars are added to the Classpath. In case of any ClassNotFound Exception, perform the following steps:
- Edit the
oracle.hadoop.loader.libjars
property present inOLH_HOME/doc/oraloader-conf.xml
to accommodate the newly added jars. That is,${oracle.hadoop.loader.olh_home}/jlib/ hive-exec-*.jar
- Edit the
- Copy the following jars (Apache Drivers) to the
- Note that OLH is not qualified on Cloudera Connectors. Perform
the following workaround:
- Cop0y the entire property to the
FIC_DB_HOME/conf/dtextInput.xml
file.Note:
Add the aforementioned jars only if OLH task is to be run. If any other OFSAA task is running, do not keep a copy of the jars in theOLH_HOME/jlib
folder.
- Set the following property in the