Oozie is an Apache open source project, originally developed at Yahoo. Oozie is a general purpose scheduling system for multistage Hadoop jobs.
Workflow
. The Oozie workflows are DAG (Directed cyclic graph) of actions.Coordinator
.Bundle
and can be scheduled on a Oozie server for execution.Oozie support most of the Hadoop Jobs as Oozie Action Nodes like: MapRedude
, Java
, FileSystem
(HDFS operations), Hive
, Hive2
, Pig
, Spark
, SSH
, Shell
, DistCp
and Sqoop
. It provides a decision capability using a Decision Control Node
action and Parallel execution of the jobs using Fork-Join Control Node
. It allow users to configure email option for Success/Failure notification of the Workflow using Email
action.
Oozie Version | Release Date |
---|---|
4.3.0 | 2016-12-02 |
Pre-requisites
This article demonstrated installing oozie-4.3.0 on Hadoop 2.7.3
Step1: Dist file
Get oozie tar.gz file from http://www-eu.apache.org/dist/oozie/4.3.0/ and extract it
cd $HOME
tar -xvf oozie-4.3.0.tar.gz
Step2: Build Oozie
cd $HOME/oozie-4.3.0/bin
./mkdistro.sh -DskipTests
Step3: Server Installation
Copy the built binaries to the home directory as ‘oozie’
cd $HOME
cp -R $HOME/oozie-4.3.0/distro/target/oozie-4.3.0-distro/oozie-4.3.0 .
Step 3.1: libext Create libext directory inside oozie directory
cd $HOME/oozie
mkdir libext
Note: ExtJS (2.2+) library (optional, to enable Oozie webconsole) But, The ExtJS library is not bundled with Oozie because it uses a different license :( Now you need to put hadoop jars inside libext directory, else it will throw below error in oozie.log file
WARN ActionStartXCommand:523 - SERVER[data01.teg.io] USER[hadoop] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-161215143751620-oozie-hado-W] ACTION[0000000-161215143751620-oozie-hado-W@mr-node] Error starting action [mr-node]. ErrorType [TRANSIENT], ErrorCode [JA009], Message [JA009: Cannot initialize Cluster. Please check your configuration for mapreduce.framework.name and the correspond server addresses.]
So, let's put below jars inside libext directory
cp $HADOOP_HOME/share/hadoop/common/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/common/lib/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/hdfs/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/hdfs/lib/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/mapreduce/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/mapreduce/lib/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/yarn/*.jar oozie/libext/
cp $HADOOP_HOME/share/hadoop/yarn/lib/*.jar oozie/libext/
Step 3.2: Oozie Impersonate To avoid impersonate error on oozie, modify core-site.xml like below
<!-- OOZIE -->
<property>
<name>hadoop.proxyuser.[OOZIE_SERVER_USER].hosts</name>
<value>[OOZIE_SERVER_HOSTNAME]</value>
</property>
<property>
<name>hadoop.proxyuser.[OOZIE_SERVER_USER].groups</name>
<value>[USER_GROUPS_THAT_ALLOW_IMPERSONATION]</value>
</property>
Assuming, my oozie user is huser and host is localhost and group is hadoop
<!-- OOZIE -->
<property>
<name>hadoop.proxyuser.huser.hosts</name>
<value>localhost</value>
</property>
<property>
<name>hadoop.proxyuser.huser.groups</name>
<value>hadoop</value>
</property>
Note : You can use * in all values, in case of confusion
Step 3.3: Prepare the war
cd $HOME/oozie/bin
./oozie-setup.sh prepare-war
This will create oozie.war file inside oozie directory. If this war will be used further, you may face this error :
ERROR ActionStartXCommand:517 - SERVER[data01.teg.io] USER[hadoop] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-161220104605103-oozie-hado-W] ACTION[0000000-161220104605103-oozie-hado-W@mr-node] Error, java.lang.NoSuchFieldError: HADOOP_CLASSPATH
Why? because, The oozie compilation produced Hadoop 2.6.0 jars even when specifying Hadoop 2.7.3 with the option "-Dhadoop.version=2.7.3".
So, to avoid this error, copy the oozie.war file to a different directory
mkdir $HOME/oozie_war_dir
cp $HOME/oozie/oozie.war $HOME/oozie_war_dir
cd $HOME/oozie_war_dir
jar -xvf oozie.war
rm -f oozie.war/WEB-INF/lib/hadoop-*.jar
rm -f oozie.war/WEB-INF/lib/hive-*.jar
rm oozie.war
jar -cvf oozie.war ./*
cp oozie.war $HOME/oozie/
Then, regenerate the oozie.war binaries for oozie with a prepare-war
cd $HOME/oozie/bin
./oozie-setup.sh prepare-war
Step 3.4: Create sharelib on HDFS
cd $HOME/oozie/bin
./oozie-setup.sh sharelib create -fs hdfs://localhost:9000
Now, this sharelib set up may give you below error:
org.apache.oozie.service.ServiceException: E0104: Could not fully initialize service [org.apache.oozie.service.ShareLibService], Not able to cache sharelib. An Admin needs to install the sharelib with oozie-setup.sh and issue the 'oozie admin' CLI command to update the sharelib
To avoid this, modify oozie-site.xml like below
cd $HOME/oozie
vi conf/oozie-site.xml
Add below property
<property>
<name>oozie.service.HadoopAccessorService.hadoop.configurations</name>
<value>*=/usr/local/hadoop/etc/hadoop/</value>
</property>
The value should be your $HADOOP_HOME/etc/hadoop, where all hadoop configuration files are present.
Step 3.5 : Create Oozie DB
cd $HOME/oozie
./bin/ooziedb.sh create -sqlfile oozie.sql -run
Step 3.6 : Start Daemon
To start Oozie as a daemon use the following command:
./bin/oozied.sh start
To stop
./bin/oozied.sh stop
check logs for errors, if any
cd $HOME/oozie/logs
tail -100f oozie.log
Use the following command to check the status of Oozie from command line:
$ ./bin/oozie admin -oozie http://localhost:11000/oozie -status
System mode: NORMAL
Step 4: Client Installation
$ cd
$ cp oozie/oozie-client-4.3.0.tar.gz .
$ tar -xvf oozie-client-4.3.0.tar.gz
$ mv oozie-client-3.3.2 oozie-client
$ cd bin
Add $HOME/oozie-client/bin to PATH variable in .bashrc file and restart your terminal or do
source $HOME/.bashrc
For more details on set up, you can refer this URL https://oozie.apache.org/docs/4.3.0/DG_QuickStart.html
Now you can submit hadoop jobs to oozie in your terminal.
To run an example, you can follow this URL and set up your first example to run https://oozie.apache.org/docs/4.3.0/DG_Examples.html
You may face below error while running the map reduce example in above URL
java.io.IOException: java.net.ConnectException: Call From localhost.localdomain/127.0.0.1 to 0.0.0.0:10020 failed on connection exception: java.net.ConnectException: Connection refused; For more details see: http://wiki.apache.org/hadoop/ConnectionRefused
Solution: Start mr-jobhistory-server.sh
cd $HADOOP_HOME/sbin
./mr-jobhistory-server.sh start historyserver
Another point to note about modifying job.properties file is :
nameNode=hdfs://localhost:9000
jobTracker=localhost:8032
in your case, this can be different, as I am using apache hadoop, you may be using cloudera/hdp/anything
To run spark job, I have tried running in local[*], yarn-client and
yarn-cluster as master, but succeeded in local[*] only