Member since
04-29-2016
9
Posts
2
Kudos Received
0
Solutions
08-15-2016
05:08 PM
@Ankit A Are you able to run the Spark job from the shell/command line? If so, then you may want to use Shell Action instead. Oozie Spark Action in HDP 2.3.4 is still in tech preview and not supported yet. The below tech note was released with the recommendation to use Shell Actions or Java Actions instead. https://community.hortonworks.com/content/kbentry/51582/how-to-use-oozie-shell-action-to-run-a-spark-job-i-1.html -------------------- Begin Tech Note -------------------- Because spark action in oozie is not supported in HDP 2.3.x and HDP 2.4.0, there is no workaround especially in kerberos environment. We can use either java action or shell action to launch spark job in oozie workflow. In this article, we will discuss how to use oozie shell action to run a spark job in kerberos environment. Prerequisite: 1. Spark client is installed on every host where nodemanager is running. This is because we have no control over which node the 2. Optionally, if the spark job need to interact with hbase cluster, hbase client need to be installed on every host as well. Steps: 1. Create a shell script with the spark-submit command. For example, in the script.sh: /usr/hdp/current/spark-client/bin/spark-submit --keytab keytab --principal ambari-qa-falconJ@FALCONJSECURE.COM --class org.apache.spark.examples.SparkPi --master yarn-client --driver-memory 500m --num-executors 1 --executor-memory 500m --executor-cores 1 spark-examples.jar 3 2. Prepare kerberos keytab which will be used by the spark job. For example, we use ambari smoke test user, the keytab is already generated by Ambari in/etc/security/keytabs/smokeuser.headless.keytab. 3. Create the oozie workflow with a shell action which will execute the script created above, for example, in the workflow.xml: <workflow-app name="WorkFlowForShellAction" xmlns="uri:oozie:workflow:0.4">
<start to="shellAction"/>
<action name="shellAction">
<shell xmlns="uri:oozie:shell-action:0.2">
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node>
<exec>script.sh</exec>
<file>/user/oozie/shell/script.sh#script.sh</file>
<file>/user/oozie/shell/smokeuser.headless.keytab#keytab</file>
<file>/user/oozie/shell/spark-examples.jar#spark-examples.jar</file>
<capture-output/>
</shell>
<ok to="end"/>
<error to="killAction"/>
</action>
<kill name="killAction">
<message>"Killed job due to error"</message>
</kill>
<end name="end"/>
</workflow-app> 4. Create the oozie job properties file. For example, in job.properties: nameNode=falconJ2.sec.support.com:8050 queueName=default oozie.wf.application.path=${nameNode}/user/oozie/shell oozie.use.system.libpath=true 5. Upload the following files created above to the oozie workflow application path in HDFS (In this example: /user/oozie/shell): - workflow.xml - smokeuser.headless.keytab - script.sh - spark uber jar (In this example: /usr/hdp/current/spark-client/lib/spark-examples*.jar) - Any other configuration file mentioned in workflow (optional) 6. Execute the oozie command to run this workflow. For example: oozie job -oozie http://<oozie-server>:11000/oozie -config job.properties -run -------------------- End Tech Note -------------------- See similar/related response here: https://community.hortonworks.com/questions/22772/oozie-spark-action-giving-key-not-found-spark-home.html#answer-45981
... View more
05-29-2016
01:02 PM
Thanks that worked
... View more
05-10-2016
08:26 AM
No problems, it can happen to the best of us. Also glad to hear that it worked for your, because many people have been having troubles with Oozie Sqoop actions recently. By the way, for me it worked without specifying "archive" in the workflow file, I just uploaded mysql-connector jar to Oozie Sqoop share lib. Happy sqooping!
... View more