Tags
Today I would like to explain how I managed to compile and install Apache Oozie 4.0.1 against the lastest stable Hadoop version 2.4.1
Prerequisites:
- Hadoop 2.4.1 : installation explained in another post
- Maven
- Java 1.6+
- Unix/Mac machine
Download Oozie
wget http://apache.hippo.nl/oozie/4.0.1/oozie-4.0.1.tar.gz tar xzvf oozie-4.0.1.tar.gz cd oozie-4.0.1
Building against Hadoop 2.4.1
By default Oozie builds against Hadoop 1.1.1, so to build against Hadoop 2.4.1, we will have to configure maven dependencies in pom.xml
Change hadoop-2 maven profile
In the downloaded Oozie source code (pom.xml), the hadoop-2 maven profile specifies hadoop.version & hadoop.auth.version to be 2.3.0. So we change them to use 2.4.1
<profile> <id>hadoop-2</id> <activation> <activeByDefault>false</activeByDefault> </activation> <properties> <hadoop.version>2.4.1</hadoop.version> <hadoop.auth.version>2.4.1</hadoop.auth.version> <pig.classifier>h2</pig.classifier> <sqoop.classifier>hadoop200</sqoop.classifier> </properties> </profile>
Change Hadooplibs maven module
Next step is to configure hadooplibs maven module to build libs for 2.4.1 version. So we change the pom.xml in hadoop-2,hadoop-distcp-2 & hadoop-test-2 maven modules within Hadooplibs maven module
cd hadooplibs File hadoop-2/pom.xml : change hadoop-client & hadoop-auth dependency version to 2.4.1 File hadoop-distcp-2/pom.xml: change hadoop-distcp version to 2.4.1 File hadoop-test-2/pom.xml: change hadoop-minicluster version to 2.4.1
Build Oozie distro
Use Maven profile hadoop-2 to compile Oozie 4.0.1 against Hadoop 2.4.1
cd .. bin/mkdistro.sh -P hadoop-2 -DskipTests or mvn clean package assembly:single -P hadoop-2 -DskipTests
Setup Oozie server
Copy the Oozie distro to new directory
cd .. mkdir Oozie cp -R oozie-4.0.1/distro/target/oozie-4.0.1-distro/oozie-4.0.1/ Oozie cd oozie mkdir libext cp -R ../oozie-4.0.1/hadooplibs/hadoop-2/target/hadooplibs/hadooplib-2.4.1.oozie-4.0.1/* libext wget -P libext http://extjs.com/deploy/ext-2.2.zip
Prepare the Oozie war
./bin/oozie-setup.sh prepare-war
Create Sharelib Directory on HDFS
Following command will internally issue a HDFS create directory command to the Name node running at hdfs://localhost:9000 and then copy the shared library to that directory.
./bin/oozie-setup.sh sharelib create -fs hdfs://localhost:9000
*make sure you select the right port number, otherwise you might get some error like Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message end-group tag did not match expected tag. That is the case when oozie tries to talk to some other web service instead of HFDS FS.
Oozie Database
./bin/ooziedb.sh create -sqlfile oozie.sql -run
Configure Hadoop
Configure the Hadoop cluster with proxyuser for the Oozie process. The following two properties are required in Hadoop etc/hadoop/core-site.xml. If you are using Hadoop higher than version 1.1.0, you can use wildcards to specify the properties in configuration files. Replace “gaurav” with the username you would be running Oozie with.
<property> <name>hadoop.proxyuser.gaurav.hosts</name> <value>*</value> </property> <property> <name>hadoop.proxyuser.gaurav.groups</name> <value>*</value> </property>
Start Oozie
./bin/oozied.sh start
Oozie should now be accessible at http://localhost:11000/oozie
Submit a Test Workflow
Now we will try to submit a Workflow provided in the examples with Oozie: map-reduce. First we need to copy the examples directory in Oozie to your home directory on hdfs and then we submit the oozie job
From Hadoop Directory: bin/hdfs dfs -put path-to-oozie-directory/examples examples From Oozie Directory: bin/oozie job -oozie http://localhost:11000/oozie/ -config examples/apps/map-reduce/job.properties -run
You might need to change job.properties before your submit the workflow to use the correct NameNode and JobTracker ports. If you are running Yarn ( MapReduce 2 ) then JobTracker will be referencing to the ResourceManager port.
nameNode=hdfs://localhost:9000 jobTracker=localhost:8032 queueName=default examplesRoot=examples oozie.wf.application.path=${nameNode}/user/${user.name}/${examplesRoot}/apps/map-reduce outputDir=map-reduce
Status and output of Workflow
Map-Reduce submitted in Oozie : http://localhost:11000/oozie/
Status Of Map-Reduce in Hadoop Cluster
Map-Reduce Finished status in Oozie
That’s It
So we have successfully configured Oozie 4.0.1 with Hadoop 2.4.1 and were also able to submit a Job. In the next post we will talk about other aspects of Oozie, like sub-workflow and how we can link or make workflows depend on each other.
Possible Issues
Java heap space or PermGen space
While running maven command to compile you might faced either of PermGen space or OutOfMemory Java heap space error. So in that case you need to increase the memory allocated to maven process
export 'MAVEN_OPTS=-Xmx1024m -XX:MaxPermSize=128m'
Hadoop History server
The Oozie server needs to talk to Hadoop History server, to know the previous state of the Jobs, so we need to keep history server started while running Oozie. This error occurs when you try to run a workflow.
sbin/mr-jobhistory-daemon.sh start historyserver
Error related to impersonation
RemoteException: User: oozie is not allowed to impersonate oozie. This is caused when you fail to configure proper hadoop.proxyuser.oozie.hosts and hadoop.proxyuser.oozie.groups properties in Hadoop, make sure you use wildcards only if Hadoop is 1.1.0+ version.
InvalidProtocolBufferException
Failed on local exception: com.google.protobuf.InvalidProtocolBufferException: Protocol message end-group tag did not match expected tag.;
This happens when you have compiled your oozie with a Protobuf library which is incompatible with the one used in Hadoop. For my use case I had compiled Oozie 4.0.1 with Protobuf 2.5.0 to work with Hadoop 2.4.1
Pingback: Error on running multiple Workflow in OOZIE-4.1.0
Pingback: 在OOZIE-4.1.0中运行多个工作流时出错 – FIXBBS
Excellent article. This actually resolved my problem in oozie.
LikeLike
Hi Gaurav,
my oozie user name is vidya.pandey… I am getting 501 impersonation user. Please let me know how to give this name “vidya.pandey” in hadoop core-site.xml file
LikeLike
extjs is corrupted… need to download from site
wget http://dev.sencha.com/deploy/ext-2.2.zip
LikeLike
What changes i need to do in pom.xml file of oozie-4.2.0 if i’m using hadoop-2.7.1 with jdk 7
LikeLike
[INFO] Scanning for projects…
Downloading: http://repo1.maven.org/maven2/org/apache/apache/16/apache-16.pom
Downloading: http://repository.codehaus.org/org/apache/apache/16/apache-16.pom
Downloading: https://repository.apache.org/content/groups/snapshots/org/apache/apache/16/apache-16.pom
Downloading: http://www.datanucleus.org/downloads/maven2/org/apache/apache/16/apache-16.pom
[ERROR] [ERROR] Some problems were encountered while processing the POMs:
[FATAL] Non-resolvable parent POM for org.apache.oozie:oozie-main:4.2.0: Could not transfer artifact org.apache:apache:pom:16 from/to central (http://repo1.maven.org/maven2): proxy.host.net: Name or service not known and ‘parent.relativePath’ points at wrong local POM @ line 21, column 11
@
[ERROR] The build could not read 1 project -> [Help 1]
[ERROR]
[ERROR] The project org.apache.oozie:oozie-main:4.2.0 (/usr/local/oozie/pom.xml) has 1 error
[ERROR] Non-resolvable parent POM for org.apache.oozie:oozie-main:4.2.0: Could not transfer artifact org.apache:apache:pom:16 from/to central (http://repo1.maven.org/maven2): proxy.host.net: Name or service not known and ‘parent.relativePath’ points at wrong local POM @ line 21, column 11 -> [Help 2]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/ProjectBuildingException
[ERROR] [Help 2] http://cwiki.apache.org/confluence/display/MAVEN/UnresolvableModelException
ERROR, Oozie distro creation failed
plz help me to resolve this error
LikeLike
Hi Gaurav
Can you please post the step by step instructions for installing Oozie 4.2.0 on Ubuntu 14.04 LTS with hadoop version 2.6.0 installed? Would be very helpful if you post that. I tried several times installing this Oozie, the build is not at all successful. Please help.
LikeLike
hi gaurav,
I am following your blog to install oozie in ubuntu 14.04 and hadoop 2.5dch5.3.2
m getting the following error
oozie disrto failed
i configured pom.xml to chnge java version n hadoop vesion.
n the maven hadooplibs also..
LikeLike
Hi
I’m trying to install Oozie 4.1.0 on Ubuntu 14.04. I got the following error when I run either of the following commands: Can someone help me to install Oozie sucessfully? I got maven 3.0.5 sucessfully installed. I also updated the pom.xml with the java version 1.7 and hadoop version 2.6.0., and pig, hive, sqoop versions under the directory Oozie 4.1.0 and the other pom.xml in directory hadooplibs/hadoop-2, hadoop-test-2, hadoop-distcp-2.
Command(s):
bin/mkdistro.sh -DskipTests -Dhadoopversion=2.6.0
sudo mvn clean package assembly:single -P hadoop-2 -DskipTests
Error:
[INFO] ————————————————————————
[INFO] Reactor Summary:
[INFO]
[INFO] Apache Oozie Main …………………………… SUCCESS [25.375s]
[INFO] Apache Oozie Client …………………………. SUCCESS [17.633s]
[INFO] Apache Oozie Hadoop 1.1.1.oozie-4.1.0 …………. SUCCESS [2.480s]
[INFO] Apache Oozie Hadoop Distcp 1.1.1.oozie-4.1.0 …… SUCCESS [0.243s]
[INFO] Apache Oozie Hadoop 1.1.1.oozie-4.1.0 Test …….. SUCCESS [3.646s]
[INFO] Apache Oozie Hadoop Utils 1.1.1.oozie-4.1.0 ……. SUCCESS [0.939s]
[INFO] Apache Oozie Hadoop 2.3.0.oozie-4.1.0 …………. SUCCESS [9.939s]
[INFO] Apache Oozie Hadoop 2.3.0.oozie-4.1.0 Test …….. SUCCESS [2.736s]
[INFO] Apache Oozie Hadoop Distcp 2.3.0.oozie-4.1.0 …… SUCCESS [0.288s]
[INFO] Apache Oozie Hadoop Utils 2.3.0.oozie-4.1.0 ……. SUCCESS [1.890s]
[INFO] Apache Oozie Hadoop 0.23.5.oozie-4.1.0 ………… SUCCESS [6.159s]
[INFO] Apache Oozie Hadoop 0.23.5.oozie-4.1.0 Test ……. SUCCESS [2.442s]
[INFO] Apache Oozie Hadoop Distcp 0.23.5.oozie-4.1.0 ….. SUCCESS [0.196s]
[INFO] Apache Oozie Hadoop Utils 0.23.5.oozie-4.1.0 …… SUCCESS [1.166s]
[INFO] Apache Oozie Hadoop Libs …………………….. SUCCESS [3.046s]
[INFO] Apache Oozie Hbase 0.94.2.oozie-4.1.0 …………. SUCCESS [0.696s]
[INFO] Apache Oozie Hbase Libs ……………………… SUCCESS [0.415s]
[INFO] Apache Oozie HCatalog 0.13.1.oozie-4.1.0 ………. SUCCESS [8.483s]
[INFO] Apache Oozie HCatalog Libs …………………… SUCCESS [0.949s]
[INFO] Apache Oozie Share Lib Oozie …………………. FAILURE [3.043s]
[INFO] Apache Oozie Share Lib HCatalog ………………. SKIPPED
[INFO] Apache Oozie Core …………………………… SKIPPED
[INFO] Apache Oozie Docs …………………………… SKIPPED
[INFO] Apache Oozie Share Lib Pig …………………… SKIPPED
[INFO] Apache Oozie Share Lib Hive ………………….. SKIPPED
[INFO] Apache Oozie Share Lib Sqoop …………………. SKIPPED
[INFO] Apache Oozie Share Lib Streaming ……………… SKIPPED
[INFO] Apache Oozie Share Lib Distcp ………………… SKIPPED
[INFO] Apache Oozie WebApp …………………………. SKIPPED
[INFO] Apache Oozie Examples ……………………….. SKIPPED
[INFO] Apache Oozie Share Lib ………………………. SKIPPED
[INFO] Apache Oozie Tools ………………………….. SKIPPED
[INFO] Apache Oozie MiniOozie ………………………. SKIPPED
[INFO] Apache Oozie Distro …………………………. SKIPPED
[INFO] Apache Oozie ZooKeeper Security Tests …………. SKIPPED
[INFO] ————————————————————————
[INFO] BUILD FAILURE
[INFO] ————————————————————————
[INFO] Total time: 1:57.349s
[INFO] Finished at: Tue Sep 08 10:15:41 CDT 2015
[INFO] Final Memory: 57M/242M
[INFO] ————————————————————————
[ERROR] Failed to execute goal on project oozie-sharelib-oozie: Could not resolve dependencies for project org.apache.oozie:oozie-sharelib-oozie:jar:4.1.0: Failed to collect dependencies for [com.googlecode.json-simple:json-simple:jar:1.1 (compile), org.apache.oozie:oozie-hadoop:jar:2.6.0.oozie-4.1.0 (provided), org.apache.oozie:oozie-hadoop-test:jar:2.6.0.oozie-4.1.0 (test), junit:junit:jar:4.10 (test), org.apache.oozie:oozie-hadoop-utils:jar:2.6.0.oozie-4.1.0 (compile)]: Failed to read artifact descriptor for org.apache.oozie:oozie-hadoop:jar:2.6.0.oozie-4.1.0: Could not transfer artifact org.apache.oozie:oozie-hadoop:pom:2.6.0.oozie-4.1.0 from/to Codehaus repository (http://repository.codehaus.org/): repository.codehaus.org: Name or service not known: Unknown host repository.codehaus.org: Name or service not known -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :oozie-sharelib-oozie
LikeLike
Hi Murali,
I am also getting same error. If u ve got the solution for above post. then pls post me steps to resolve the above issue.
LikeLike
Pingback: Hue – Hadoop可视化分析平台搭建 » Vanjor
please follow the below link for the ext-js else no use
http://dev.sencha.com/deploy/ext-2.2.zip
LikeLike
I spent more than half a day figuring out shared lib error – finally found this blog (this is not a spam) – https://hadooptutorial.info/oozie-share-lib-does-not-exist-error/
I had to specify the path to hadoop-conf in oozie-site.xml and use hdfs:/// protocol in the workflowapp properties. It did it.
LikeLike
where we can check the history server status and commad to run it.
LikeLike
Hi Gaurav,
Very well documented. Thnaks for the steps. Colud you please help me with below error.
I am installing Oozie 4.0 on hadoop 7.0 with Hive 0.14. Have made all and only changes you specified in pom.xml file.
I am using below command to install.
bin/mkdistro.sh -DskipTests -Dhadoopversion=2.7.0 -DjavaVersion=1.7 -DtargetJavaVersion=1.7
Below is the error. Searched a lot over ner. But no luck.
[ERROR] Failed to execute goal on project oozie-hcatalog: Could not resolve dependencies for project org.apache.oozie:oozie-hcatalog:jar:0.5.0.oozie-4.0.1: Failed to collect dependencies at org.apache.hcatalog:hcatalog-server-extensions:jar:0.5.0-incubating -> org.apache.hcatalog:hcatalog-core:jar:0.5.0-incubating -> org.apache.hive:hive-builtins:jar:0.10.0 -> org.apache.hive:hive-exec:jar:0.10.0 -> org.apache.hive:hive-metastore:jar:0.10.0 -> javax.jdo:jdo2-api:jar:2.3-ec: Failed to read artifact descriptor for javax.jdo:jdo2-api:jar:2.3-ec: Could not transfer artifact javax.jdo:jdo2-api:pom:2.3-ec from/to Codehaus repository (http://repository.codehaus.org/): sun.security.validator.ValidatorException: PKIX path building failed: sun.security.provider.certpath.SunCertPathBuilderException: unable to find valid certification path to requested target -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :oozie-hcatalog
LikeLike
Nice information, I got successfully installed oozie with your commands.
Thank you very much
LikeLiked by 1 person
Hello ,
I may need some help running oozie 4.1.0 with hadoop 2.6.0.
I am running the map-reduce job via oozie:
./oozie job -oozie http://localhost:11000/oozie/ -config ../examples/apps/map-reduce/job.properties -run
I am seeing the job fail:
2015-03-30 09:14:58,982 WARN JobSubmitter:259 – SERVER[localhost.localdomain] No job jar file set. User classes may not be found. See Job or Job#setJar(String).
2015-03-30 09:14:59,691 INFO MapReduceActionExecutor:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] checking action, hadoop job ID [job_1427400522986_0018] status [RUNNING]
2015-03-30 09:14:59,694 INFO ActionStartXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] [0000000-150330091436044-oozie-oozi-W@mr-node]Action status=RUNNING
2015-03-30 09:14:59,694 INFO ActionStartXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] [0000000-150330091436044-oozie-oozi-W@mr-node]Action updated in DB!
2015-03-30 09:15:23,327 INFO CallbackServlet:541 – SERVER[localhost.localdomain] USER[-] GROUP[-] TOKEN[-] APP[-] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] callback for action [0000000-150330091436044-oozie-oozi-W@mr-node]
2015-03-30 09:15:24,329 INFO MapReduceActionExecutor:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] action completed, external ID [job_1427400522986_0018]
2015-03-30 09:15:24,333 WARN MapReduceActionExecutor:544 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] LauncherMapper died, check Hadoop LOG for job [localhost:8032:job_1427400522986_0018]
2015-03-30 09:15:25,099 INFO ActionEndXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@mr-node] ERROR is considered as FAILED for SLA
2015-03-30 09:15:25,227 INFO ActionStartXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@fail] Start action [0000000-150330091436044-oozie-oozi-W@fail] with user-retry state : userRetryCount [0], userRetryMax [0], userRetryInterval [10]
2015-03-30 09:15:25,228 INFO ActionStartXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@fail] [0000000-150330091436044-oozie-oozi-W@fail]Action status=DONE
2015-03-30 09:15:25,228 INFO ActionStartXCommand:541 – SERVER[localhost.localdomain] USER[oozie] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000000-150330091436044-oozie-oozi-W] ACTION[0000000-150330091436044-oozie-oozi-W@fail] [0000000-150330091436044-oozie-oozi-W@fail]Action updated in DB!
2015-03-30 09:15:53,940 INFO StatusTransitService$StatusTransitRunnable:541 – SERVER[localhost.localdomain] USER[-] GROUP[-] Acquired lock for [org.apache.oozie.service.StatusTransitService]
I do know why it would except the WARN:
Could you be so kind and help me on this I am stuck.
New to this technology.
LikeLike
Hello,
I managed to move forward but when I tried:
covage@datanode2:~/oozie$ ./bin/oozie-setup.sh sharelib create -fs hdfs://localhost:9001
I got:
setting CATALINA_OPTS=”$CATALINA_OPTS -Xmx1024m”
the destination path for sharelib is: /user/covage/share/lib/lib_20150124195507
Error: E0902: Exception occured: [Call to localhost/127.0.0.1:9001 failed on connection exception: java.net.ConnectException: Connexion refusée]
Stack trace for the error was (for debug purposes):
org.apache.oozie.service.HadoopAccessorException: E0902: Exception occured: [Call to localhost/127.0.0.1:9001 failed on connection exception: java.net.ConnectException: Connexion refusée]
at org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:431)
at org.apache.oozie.tools.OozieSharelibCLI.run(OozieSharelibCLI.java:158)
at org.apache.oozie.tools.OozieSharelibCLI.main(OozieSharelibCLI.java:56)
Caused by: java.net.ConnectException: Call to localhost/127.0.0.1:9001 failed on connection exception: java.net.ConnectException: Connexion refusée
at org.apache.hadoop.ipc.Client.wrapException(Client.java:1136)
at org.apache.hadoop.ipc.Client.call(Client.java:1112)
at org.apache.hadoop.ipc.RPC$Invoker.invoke(RPC.java:229)
at com.sun.proxy.$Proxy3.getProtocolVersion(Unknown Source)
at org.apache.hadoop.ipc.RPC.getProxy(RPC.java:411)
at org.apache.hadoop.hdfs.DFSClient.createRPCNamenode(DFSClient.java:135)
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:276)
at org.apache.hadoop.hdfs.DFSClient.(DFSClient.java:241)
at org.apache.hadoop.hdfs.DistributedFileSystem.initialize(DistributedFileSystem.java:100)
at org.apache.hadoop.fs.FileSystem.createFileSystem(FileSystem.java:1411)
at org.apache.hadoop.fs.FileSystem.access$200(FileSystem.java:66)
at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:1429)
at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:254)
at org.apache.oozie.service.HadoopAccessorService$2.run(HadoopAccessorService.java:423)
at org.apache.oozie.service.HadoopAccessorService$2.run(HadoopAccessorService.java:421)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1136)
at org.apache.oozie.service.HadoopAccessorService.createFileSystem(HadoopAccessorService.java:421)
… 2 more
Caused by: java.net.ConnectException: Connexion refusée
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:511)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:481)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:453)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:579)
at org.apache.hadoop.ipc.Client$Connection.access$2100(Client.java:202)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1243)
at org.apache.hadoop.ipc.Client.call(Client.java:1087)
… 19 more
Anyone knowing what is going on ?
Why do you think the Call to localhost/127.0.0.1:9001 failed on connection exception: java.net.ConnectException ?
Thanks for your help.
Rgds,
Sebastien
LikeLike
I found out what was the problem, port 9001 was not opened, i had to change to port 9000 and then it worked fine
LikeLike
is your hadoop processes are running including job history server.
LikeLike
Hello Gaurav,
i tried to install Oozie 4.1.0 with Hadoop 2.6.0 following your tutorial and got the following errors after running:
mvn clean package assembly:single -P hadoop-2 -DskipTests
[ERROR] Failed to execute goal on project oozie-sharelib-oozie: Could not resolve dependencies for project org.apache.oozie:oozie-sharelib-oozie:jar:4.1.0: The following artifacts could not be resolved: org.apache.oozie:oozie-hadoop:jar:2.6.0.oozie-4.1.0, org.apache.oozie:oozie-hadoop-test:jar:2.6.0.oozie-4.1.0, org.apache.oozie:oozie-hadoop-utils:jar:2.6.0.oozie-4.1.0: Could not find artifact org.apache.oozie:oozie-hadoop:jar:2.6.0.oozie-4.1.0 in central (http://repo1.maven.org/maven2) -> [Help 1]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/DependencyResolutionException
[ERROR]
[ERROR] After correcting the problems, you can resume the build with the command
[ERROR] mvn -rf :oozie-sharelib-oozie
Not very familiar with oozie and hadoop, I may have missed something but not sure what. Could you please help me a but in order that I can move forward
LikeLike
i am getting similar error :
Could not find artifact org.apache.oozie:oozie-hadoop:jar:2.4.1.oozie-4.1.0 in central (http://repo1.maven.org/maven2)
and it looks like that jar is not there in the repo
http://search.maven.org/#search%7Cga%7C1%7Ca%3A%22oozie-hadoop%22
any help?
LikeLike
Hello,
I reached the same conclusion and then decided to use Oozie 4.1 with by default config, i.e. with hadoop 2.3. It looks like it is not a problem, even if I am actually using hadoop 2.6 version.
I hope this help.
Rgds,
Sebastien
LikeLike
Great. I will try that out.
Thanks
LikeLike
Hello,
I also tried to build oozie 4.1.0 for hadoop 2.6.0 with default config (i.e. hadoop 2.3.0). But build fails at oozie-tools, with following error –
[ERROR] Failed to execute goal org.apache.maven.plugins:maven-compiler-plugin:2.3.2:testCompile (default-cli) on project oozie-tools: Compilation failure: Compilation failure:
[ERROR] /usr/local/oozie-4.1.0/tools/src/test/java/org/apache/oozie/tools/TestOozieDBCLI.java:[97,64] error: cannot find symbol
[ERROR] class TestOozieDBCLI
[ERROR] /usr/local/oozie-4.1.0/tools/src/test/java/org/apache/oozie/tools/TestOozieDBCLI.java:[138,12] error: cannot find symbol
[ERROR] class TestOozieDBCLI
[ERROR] /usr/local/oozie-4.1.0/tools/src/test/java/org/apache/oozie/tools/TestOozieSharelibCLI.java:[193,12] error: cannot find symbol
[ERROR] class TestOozieSharelibCLI
[ERROR] /usr/local/oozie-4.1.0/tools/src/test/java/org/apache/oozie/tools/TestOozieMySqlDBCLI.java:[106,12] error: cannot find symbol
[ERROR] -> [Help 1]
[ERROR]
Did you get anysuch error?
LikeLike
Hi Gaurav,
I am trying to build oozie with hadoop-2.4.0.
I followed the steps, and make changes in the required pom.xml(s).
but when i executed the command
bin/mkdistro.sh -P hadoop-2 -DskipTests
i got following error
[INFO] Scanning for projects…
Downloading: https://repository.cloudera.com/artifactory/ext-release-local/org/apache/felix/maven-bundle-plugin/2.4.0/maven-bundle-plugin-2.4.0.pom
^MDownloading: http://repo.maven.apache.org/maven2/org/apache/felix/maven-bundle-plugin/2.4.0/maven-bundle-plugin-2.4.0.pom
^M[ERROR] The build could not read 1 project -> [Help 1]
[ERROR]
[ERROR] The project org.apache.oozie:oozie-zookeeper-security-tests:4.1.0 (/var/opt/oozie-4.1.0/zookeeper-security-tests/pom.xml) has 1 error
[ERROR] Unresolveable build extension: Plugin org.apache.felix:maven-bundle-plugin:2.4.0 or one of its dependencies could not be resolved: Failed to collect dependencies for org.apache.felix:maven-bundle-plugin:jar:2.4.0 (): Failed to read artifact descriptor for org.apache.felix:maven-bundle-plugin:jar:2.4.0: Could not transfer artifact org.apache.felix:maven-bundle-plugin:pom:2.4.0 from/to repository.cloudera.com (https://repository.cloudera.com/artifactory/ext-release-local): Connection to https://repository.cloudera.com refused: Connection refused -> [Help 2]
[ERROR]
[ERROR] To see the full stack trace of the errors, re-run Maven with the -e switch.
[ERROR] Re-run Maven using the -X switch to enable full debug logging.
[ERROR]
[ERROR] For more information about the errors and possible solutions, please read the following articles:
[ERROR] [Help 1] http://cwiki.apache.org/confluence/display/MAVEN/ProjectBuildingException
[ERROR] [Help 2] http://cwiki.apache.org/confluence/display/MAVEN/PluginResolutionException
ERROR, Oozie distro creation failed
Can u please help me in this.
Thanks
-Aditya Pathak
LikeLike
Did you get any solution ?
LikeLike
Sorry the exception was
2014-09-30 18:16:21,410 WARN Credentials:103 – Null token ignored for oozie mr token
2014-09-30 18:16:21,411 WARN MapReduceActionExecutor:542 – USER[impadmin] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] Exception in check(). Message[JA017: Unknown hadoop job [job_local1782273335_0005] associated with action [0000003-140930181109732-oozie-impa-W@mr-node]. Failing this action!]
org.apache.oozie.action.ActionExecutorException: JA017: Unknown hadoop job [job_local1782273335_0005] associated with action [0000003-140930181109732-oozie-impa-W@mr-node]. Failing this action!
at org.apache.oozie.action.hadoop.JavaActionExecutor.check(JavaActionExecutor.java:1009)
at org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:947)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:215)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:60)
at org.apache.oozie.command.XCommand.call(XCommand.java:280)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
2014-09-30 18:16:21,411 WARN ActionStartXCommand:542 – USER[impadmin] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] Error starting action [mr-node]. ErrorType [FAILED], ErrorCode [JA017], Message [JA017: Unknown hadoop job [job_local1782273335_0005] associated with action [0000003-140930181109732-oozie-impa-W@mr-node]. Failing this action!]
org.apache.oozie.action.ActionExecutorException: JA017: Unknown hadoop job [job_local1782273335_0005] associated with action [0000003-140930181109732-oozie-impa-W@mr-node]. Failing this action!
at org.apache.oozie.action.hadoop.JavaActionExecutor.check(JavaActionExecutor.java:1009)
at org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:947)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:215)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:60)
at org.apache.oozie.command.XCommand.call(XCommand.java:280)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
2014-09-30 18:16:21,411 WARN ActionStartXCommand:542 – USER[impadmin] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] Failing Job due to failed action [mr-node]
2014-09-30 18:16:21,412 WARN LiteWorkflowInstance:542 – USER[impadmin] GROUP[-] TOKEN[] APP[map-reduce-wf] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] Workflow Failed. Failing node [mr-node]
2014-09-30 18:16:21,488 WARN Configuration:2358 – file:/tmp/hadoop-impadmin/mapred/local/localRunner/impadmin/job_local1782273335_0005/job_local1782273335_0005.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring.
2014-09-30 18:16:21,489 WARN Configuration:2358 – file:/tmp/hadoop-impadmin/mapred/local/localRunner/impadmin/job_local1782273335_0005/job_local1782273335_0005.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring.
2014-09-30 18:16:21,491 WARN Configuration:2358 – file:/usr/local/Oozie/oozie-4.0.0/bin/action.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.retry.interval; Ignoring.
2014-09-30 18:16:21,491 WARN Configuration:2358 – file:/usr/local/Oozie/oozie-4.0.0/bin/action.xml:an attempt to override final parameter: mapreduce.job.end-notification.max.attempts; Ignoring.
2014-09-30 18:16:21,510 WARN JobSubmitter:150 – Hadoop command-line option parsing not performed. Implement the Tool interface and execute your application with ToolRunner to remedy this.
2014-09-30 18:16:21,512 WARN JobSubmitter:259 – No job jar file set. User classes may not be found. See Job or Job#setJar(String).
2014-09-30 18:16:21,653 INFO CallbackServlet:539 – USER[-] GROUP[-] TOKEN[-] APP[-] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] callback for action [0000003-140930181109732-oozie-impa-W@mr-node]
2014-09-30 18:16:21,662 ERROR CompletedActionXCommand:536 – USER[-] GROUP[-] TOKEN[] APP[-] JOB[0000003-140930181109732-oozie-impa-W] ACTION[0000003-140930181109732-oozie-impa-W@mr-node] XException,
org.apache.oozie.command.CommandException: E0800: Action it is not running its in [PREP] state, action [0000003-140930181109732-oozie-impa-W@mr-node]
at org.apache.oozie.command.wf.CompletedActionXCommand.eagerVerifyPrecondition(CompletedActionXCommand.java:85)
at org.apache.oozie.command.XCommand.call(XCommand.java:251)
at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
2014-09-30 18:16:21,663 WARN CallableQueueService$CallableWrapper:542 – USER[-] GROUP[-] TOKEN[-] APP[-] JOB[-] ACTION[-] exception callable [callback], E0800: Action it is not running its in [PREP] state, action [0000003-140930181109732-oozie-impa-W@mr-node]
org.apache.oozie.command.CommandException: E0800: Action it is not running its in [PREP] state, action [0000003-140930181109732-oozie-impa-W@mr-node]
at org.apache.oozie.command.wf.CompletedActionXCommand.eagerVerifyPrecondition(CompletedActionXCommand.java:85)
at org.apache.oozie.command.XCommand.call(XCommand.java:251)
at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:744)
2014-09-30 18:16:22,105 INFO CoordMaterializeTriggerService$CoordMaterializeTriggerRunnable:539 – USER[-] GROUP[-] TOKEN[-] APP[-] JOB[-] ACTION[-] CoordMaterializeTriggerService – Curr Date= Tue Sep 30 18:21:22 IST 2014, Num jobs to materialize = 0
LikeLike
Hi Sanjay,
I haven’t seen this error before, but can you check if your HistoryServer is running, because to me it looks like Oozie is not able to find the status of some earlier job you ran on Oozie cluster.
And do you see any exceptions in ResourceManager logs
LikeLike
Hi Gaurav,
I am getting below exception job is getting failed while submiting
I am using Oozie4 + hadoop2.4.0
Any Idea..?
node] Error starting action [mr-node]. ErrorType [TRANSIENT], ErrorCode [JA009], Message [JA009: Cannot delete /user/impadmin/oozie-impa/0000002-140930181109732-oozie-impa-W/mr-node–map-reduce.tmp. Name node is in safe mode.
The reported blocks 119 needs additional 20 blocks to reach the threshold 0.9990 of total blocks 139.
The number of live datanodes 1 has reached the minimum number 0. Safe mode will be turned off automatically once the thresholds have been reached.
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkNameNodeSafeMode(FSNamesystem.java:1195)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInternal(FSNamesystem.java:3336)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.deleteInt(FSNamesystem.java:3296)
at org.apache.hadoop.hdfs.server.namenode.FSNamesystem.delete(FSNamesystem.java:3280)
at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.delete(NameNodeRpcServer.java:727)
at org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.delete(ClientNamenodeProtocolServerSideTranslatorPB.java:547)
at org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:585)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:928)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)
at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1548)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)
LikeLike
hi ,I got the error:
my job.properties :
nameNode=hdfs://10.64.22.249:8020
jobTracker=10.64.22.250:8032 #my active resource manager and
#yarn.resourcemanager.port =8032
queueName=default
examplesRoot=examples
user.name=oozie
oozie.wf.application.path=${nameNode}/user/${user.name}/${examplesRoot}/apps/map-reduce
outputDir=map-reduce
the error:
2014-10-15 11:00:07,171 WARN ActionStartXCommand:542 – USER[oozie] GROUP[-] TOKEN[-] APP[map-reduce-wf] JOB[0000006-141015104311118-oozie-omm-W] ACTION[0000006-141015104311118-oozie-omm-W@mr-node] Error starting action [mr-node]. ErrorType [TRANSIENT], ErrorCode [ JA006], Message [ JA006: Connection refused]
org.apache.oozie.action.ActionExecutorException: JA006: Connection refused
at org.apache.oozie.action.ActionExecutor.convertExceptionHelper(ActionExecutor.java:412)
at org.apache.oozie.action.ActionExecutor.convertException(ActionExecutor.java:392)
at org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:794)
at org.apache.oozie.action.hadoop.JavaActionExecutor.start(JavaActionExecutor.java:945)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:215)
at org.apache.oozie.command.wf.ActionStartXCommand.execute(ActionStartXCommand.java:60)
at org.apache.oozie.command.XCommand.call(XCommand.java:280)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:326)
at org.apache.oozie.service.CallableQueueService$CompositeCallable.call(CallableQueueService.java:255)
at org.apache.oozie.service.CallableQueueService$CallableWrapper.run(CallableQueueService.java:175)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Caused by: java.net.ConnectException: Connection refused
at sun.nio.ch.SocketChannelImpl.checkConnect(Native Method)
at sun.nio.ch.SocketChannelImpl.finishConnect(SocketChannelImpl.java:739)
at org.apache.hadoop.net.SocketIOWithTimeout.connect(SocketIOWithTimeout.java:206)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:529)
at org.apache.hadoop.net.NetUtils.connect(NetUtils.java:493)
at org.apache.hadoop.ipc.Client$Connection.setupConnection(Client.java:604)
at org.apache.hadoop.ipc.Client$Connection.setupIOstreams(Client.java:699)
at org.apache.hadoop.ipc.Client$Connection.access$2800(Client.java:367)
at org.apache.hadoop.ipc.Client.getConnection(Client.java:1462)
at org.apache.hadoop.ipc.Client.call(Client.java:1381)
at org.apache.hadoop.ipc.Client.call(Client.java:1363)
at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)
at com.sun.proxy.$Proxy39.getDelegationToken(Unknown Source)
at org.apache.hadoop.mapreduce.v2.api.impl.pb.client.MRClientProtocolPBClientImpl.getDelegationToken(MRClientProtocolPBClientImpl.java:211)
at org.apache.hadoop.mapred.YARNRunner.getDelegationTokenFromHS(YARNRunner.java:210)
at org.apache.hadoop.mapred.YARNRunner.addHistoryToken(YARNRunner.java:197)
at org.apache.hadoop.mapred.YARNRunner.submitJob(YARNRunner.java:282)
at org.apache.hadoop.mapreduce.JobSubmitter.submitJobInternal(JobSubmitter.java:432)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1285)
at org.apache.hadoop.mapreduce.Job$10.run(Job.java:1282)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1612)
at org.apache.hadoop.mapreduce.Job.submit(Job.java:1282)
at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:562)
at org.apache.hadoop.mapred.JobClient$1.run(JobClient.java:557)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1612)
at org.apache.hadoop.mapred.JobClient.submitJobInternal(JobClient.java:557)
at org.apache.hadoop.mapred.JobClient.submitJob(JobClient.java:548)
at org.apache.oozie.action.hadoop.JavaActionExecutor.submitLauncher(JavaActionExecutor.java:779)
… 10 more
2014-10-15 11:00:07,184 INFO ActionStartXCommand:539 – USER[oozie] GROUP[-] TOKEN[-] APP[map-reduce-wf] JOB[0000006-141015104311118-oozie-omm-W] ACTION[0000006-141015104311118-oozie-omm-W@mr-node] Next Retry, Attempt Number [1] in [60,000] milliseconds
LikeLike
my jobhistory was not setup.now it’s ok.
LikeLiked by 1 person