Member since
04-03-2019
962
Posts
1743
Kudos Received
146
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 15009 | 03-08-2019 06:33 PM | |
| 6179 | 02-15-2019 08:47 PM | |
| 5101 | 09-26-2018 06:02 PM | |
| 12606 | 09-07-2018 10:33 PM | |
| 7447 | 04-25-2018 01:55 AM |
08-29-2016
01:30 AM
@Mike Frampton - Is this resolved? If yes then can you please accept the appropriate answer?
... View more
08-24-2016
06:17 PM
@da li - I believe you have passed impersonation issues after referring below answers. are you still facing same issue? If not then I would suggest you to accept the appropriate answer and start a new question if you have any further issues.
... View more
08-24-2016
06:09 PM
@Sami Ahmad Try below command hadoop distcp -cp hdfs:///user/sami/ hdfs:///user/zhang
... View more
08-22-2016
11:59 PM
3 Kudos
Below are the steps to run Hive(TEZ) query in a shell script using Oozie shell action . 1. Configure job.properties Example: #*************************************************
# job.properties
#*************************************************
nameNode=hdfs://<namenode-fqdn>:8020
jobTracker=<resourcemanager-host-fqdn>:8050
queueName=default
examplesRoot=examples
oozie.use.system.libpath=true
oozie.wf.application.path=${nameNode}/user/${user.name}/ . 2. Configure Workflow.xml Example: <?xml version="1.0" encoding="UTF-8"?>
<workflow-app xmlns="uri:oozie:workflow:0.4"
name="test-shell-with-kerberos-wf">
<global>
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node>
<configuration>
<property>
<name>mapred.job.queue.name</name>
<value>${queueName}</value>
</property>
<property>
<name>tez.queue.name</name>
<value>${queueName}</value>
</property>
</configuration>
</global>
<credentials>
<credential name="hive_credentials" type="hcat">
<property>
<name>hcat.metastore.uri</name>
<value>thrift://<metastore-server>:9083</value>
</property>
<property>
<name>hcat.metastore.principal</name>
<value>hive/_HOST@REALM</value>
</property>
</credential>
</credentials>
<start to="run-shell-script"/>
<action name="run-shell-script" cred="hive_credentials">
<shell xmlns="uri:oozie:shell-action:0.2">
<job-tracker>${jobTracker}</job-tracker>
<name-node>${nameNode}</name-node>
<configuration>
<property>
<name>mapred.job.queue.name</name>
<value>${queueName}</value>
</property>
<property>
<name>tez.lib.uris</name>
<value>/hdp/apps/<VERSION>/tez/tez.tar.gz</value>
</property>
</configuration>
<exec>/user/<username>/hive.sh</exec>
<file>/user/<username>/hive.sh#hive.sh</file>
</shell>
<ok to="end"/>
<error to="killnode"/>
</action>
<kill name="killnode">
<message>Job failed, error message[${wf:errorMessage(wf:lastErrorNode())}]</message>
</kill>
<end name="end"/>
</workflow-app> . 3. Write sample shell script Example: #!/bin/bash
hive --hiveconf mapreduce.job.credentials.binary=$HADOOP_TOKEN_FILE_LOCATION --hiveconf tez.credentials.path=$HADOOP_TOKEN_FILE_LOCATION -e 'select * from test_hive;'
4. Upload workflow.xml and shell script to "oozie.wf.application.path" defined in job.properties . 5. Follow below command to run Oozie workflow oozie job -oozie http://<oozie-server-hostname>:11000/oozie -config /$PATH/job.properties -run . Please comment if you have any question! Happy Hadooping!! 🙂
... View more
Labels:
08-20-2016
05:21 AM
Thanks @Vijay Kumar J Can you please accept the answer.
... View more
08-19-2016
06:18 PM
3 Kudos
@Vijay Kumar J - Do you have execute permissions on your shell script? Have you mentioned shell script correctly(with file and script tags) in your oozie workflow.xml? Please refer this tutorial for more details: https://github.com/crazyadmins/oozie-tutorials/tree/master/workflows/shell
... View more
08-19-2016
02:32 AM
@Mike Frampton - If you restart Oozie first before Yarn then there is a chance that first restart still read old configs. Then you restarted Yarn and Hive which deployed updated hive-site.xml and yarn-site.xml/mapred-site.xml. Next Oozie restart got the updated config files. I could be wrong but this is one of the possible reason.
... View more
08-18-2016
06:29 AM
2 Kudos
@Fasil Ahamed 2) Local to DataNode machine on which the particular map program runs. Please note that Datanode is only storage component, job will be run by Nodemanager on that particular slave machine. Generally slave node has both Datanode and Nodemanagers deployed. Datanode for storage purpose and Nodemanagers for running Mapper/Reducer/Tez etc. containers.
... View more