Member since
04-03-2019
962
Posts
1743
Kudos Received
146
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 14998 | 03-08-2019 06:33 PM | |
| 6178 | 02-15-2019 08:47 PM | |
| 5098 | 09-26-2018 06:02 PM | |
| 12591 | 09-07-2018 10:33 PM | |
| 7446 | 04-25-2018 01:55 AM |
11-30-2016
07:37 PM
4 Kudos
@João Souza Yes! You can use Hive/Pig/Spark actions in appropriate order as per your requirement and control the flow ( like if Hive is successful then more to Pig node and so on or else go to fail node. ) OR You can create a shell script and put calls to your Hive/Pig/Spark scripts in an appropriate order and use Oozie's shell action to execute the script. If your cluster is Kerberized then I would not suggest to use shell action as it will create lot of issues related to authentication. Hope this information helps! Happy Hadooping! 🙂
... View more
11-30-2016
07:31 PM
2 Kudos
@Mohan V Can you please try to start Datanode manually (without Ambari) with DEBUG logs? Here is the command 1. Login to problematic Datanode by 'hdfs' user 2. Run below commands: #Command1 export HADOOP_ROOT_LOGGER=DEBUG,console #Command2 hdfs datanode Note - This will print output on screen and will try to start your Datanode, please do not press 'ctrl+c' until you get ERROR/Exception 🙂 Hope this information helps you to troubleshoot your issue! Happy Hadooping 🙂
... View more
11-30-2016
07:22 PM
4 Kudos
@Brian Ramsel Did you remove ZEPPELIN and AMBARI_METRICS from Ambari in past by any chance? If that is the case, some traces might have left in Ambari DB. Can you please follow below steps: 1. Take backup of Ambari DB to be on safer side. 2. Login to Ambari DB, check below tables and remove any leftover traces of AMBARI_METRICS and ZEPPELIN hostcomponentstate
hostcomponentdesiredstate
servicecomponentdesiredstate
servicedesiredstate
serviceconfighosts
serviceconfigmapping
serviceconfig
requestresourcefilter
requestoperationlevel
clusterservices
clusterconfig Hope this information helps! 🙂
... View more
11-30-2016
07:10 PM
3 Kudos
@shyam gurram This is correct. Falcon uses Oozie for job scheduling. If you delete Oozie, falcon will get broken --> You need to remove Falcon first and then Oozie. Please note - If you have any jobs scheduled from Falcon, please make sure that you have taken care of the backup plan before going ahead with service deletion.
... View more
11-30-2016
05:53 PM
3 Kudos
@suresh krish Unfortunately you cannot do it without restarting Oozie Here are the steps: 1. Add the below set of lines after log4j.appender.oozie and before log4j.appender.oozieops:
log4j.appender.oozieError=org.apache.log4j.rolling.RollingFileAppender
log4j.appender.oozieError.RollingPolicy=org.apache.oozie.util.OozieRollingPolicy
log4j.appender.oozieError.File=${oozie.log.dir}/oozie-error.log
log4j.appender.oozieError.Append=true
log4j.appender.oozieError.layout=org.apache.log4j.PatternLayout
log4j.appender.oozieError.layout.ConversionPattern=%d{ISO8601} %5p %c{1}:%L - SERVER[${oozie.instance.id}] %m%n
log4j.appender.oozieError.RollingPolicy.FileNamePattern=${log4j.appender.oozieError.File}-%d{yyyy-MM-dd-HH}
log4j.appender.oozieError.RollingPolicy.MaxHistory=720
log4j.appender.oozieError.filter.1 = org.apache.log4j.varia.LevelMatchFilter
log4j.appender.oozieError.filter.1.levelToMatch = WARN
log4j.appender.oozieError.filter.2 = org.apache.log4j.varia.LevelMatchFilter
log4j.appender.oozieError.filter.2.levelToMatch = ERROR
log4j.appender.oozieError.filter.3 = org.apache.log4j.varia.LevelMatchFilter
log4j.appender.oozieError.filter.3.levelToMatch = FATAL
log4j.appender.oozieError.filter.4 = org.apache.log4j.varia.DenyAllFilter
2. Modify below props: Earlier value: log4j.logger.org.apache.oozie=WARN, oozie New value: log4j.logger.org.apache.oozie=ALL, oozie, oozieError
3. Restart Oozie service via Ambari. Note - This would help in getting the job error log for the new jobs launched after restart of oozie service.
... View more
11-30-2016
02:55 PM
3 Kudos
Step by step guide to configure Oozie HA in kerberized environment http://crazyadmins.com/tag/oozie-ha-loadbalancer/
... View more
11-28-2016
06:11 PM
3 Kudos
Here is the scenario: 1. I have workflow.xml which contains hive action. . 2. I have added <job-xml> tag inside hive action and provided path to the hive-site.xml (say /tmp/hive-site.xml) . 3. I have added hive-site.xml to the ${wf.application.path}/lib directory as well. . 4. I have hive-site.xml to the oozie sharelib under: /user/oozie/sharelib/lib_<timestamp>/oozie/hive-site.xml /user/oozie/sharelib/lib_<timestamp>/sqoop/hive-site.xml /user/oozie/sharelib/lib_<timestamp>/hive/hive-site.xml . 5. My simple hive workflow is failing with below error: Oozie Hive action configuration
=================================================================
Using action configuration file /hadoop/data01/hadoop/yarn/local/usercache/root/appcache/application_1443111597609_2691/container_1443111597609_2691_01_000002/action.xml
------------------------
Setting env property for mapreduce.job.credentials.binary to: /hadoop/data01/hadoop/yarn/local/usercache/root/appcache/application_1443111597609_2691/container_1443111597609_2691_01_000002/container_tokens
------------------------
------------------------
Setting env property for tez.credentials.path to: /hadoop/data01/hadoop/yarn/local/usercache/root/appcache/application_1443111597609_2691/container_1443111597609_2691_01_000002/container_tokens
------------------------
<<< Invocation of Main class completed <<<
Failing Oozie Launcher, Main class [org.apache.oozie.action.hadoop.HiveMain], main() threw exception, hive-site.xml (Permission denied)
java.io.FileNotFoundException: hive-site.xml (Permission denied)
at java.io.FileOutputStream.open(Native Method)
at java.io.FileOutputStream.<init>(FileOutputStream.java:221)
at java.io.FileOutputStream.<init>(FileOutputStream.java:110)
at org.apache.oozie.action.hadoop.HiveMain.setUpHiveSite(HiveMain.java:166)
at org.apache.oozie.action.hadoop.HiveMain.run(HiveMain.java:196)
at org.apache.oozie.action.hadoop.LauncherMain.run(LauncherMain.java:38)
at org.apache.oozie.action.hadoop.HiveMain.main(HiveMain.java:66)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at org.apache.oozie.action.hadoop.LauncherMapper.map(LauncherMapper.java:225)
at org.apache.hadoop.mapred.MapRunner.run(MapRunner.java:54)
at org.apache.hadoop.mapred.MapTask.runOldMapper(MapTask.java:430)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:342)
at org.apache.hadoop.mapred.YarnChild$2.run(YarnChild.java:168)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:415)
at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1594)
at org.apache.hadoop.mapred.YarnChild.main(YarnChild.java:163)
Oozie Launcher failed, finishing Hadoop job gracefully . How to resolve? . As we know that having multiple jar files with different version in oozie sharelib can cause classnotfound exceptions, in a similar way having multiple copies of configuration file also can cause conflicts. In this case Oozie might be trying to override hive-site.xml in nodemanager's local filesystem (/hadoop/yarn/local/usercache/<username>/appcache/application_id/blah/blah with one of the copy taken from sharelib or ${wf.application.path}/lib or from the <job-xml>. To resolve such conflicts, We have to remove extra copies of hive-site.xml from all the above mentioned location. Oozie uses hive-site.xml from /etc/oozie/conf/action-conf/hive/hive-site.xml 🙂 . To repeat, This issue was resolved by removing hive-site.xml from below locations:
1. oozie sharelib (it was present at multiple locations in oozie sharelib)
. 2. from ${wf.application.path}/lib/ directory.
. 3. From workflow.xml (removed <job-xml> part) . By default Oozie takes this file from /etc/oozie/conf/action-conf/hive/hive-site.xml . With Oozie nothing is Easy 😉 Please comment if you have any feedback/questions/suggestions. Happy Hadooping!! 🙂
... View more
Labels:
11-28-2016
05:40 PM
3 Kudos
@Vincent Jiang I have resolved similar issue Please check - https://community.hortonworks.com/questions/65860/oozie-workflow-calls-sqoop-import-sql-server-to-hi.html#answer-68743
... View more
11-28-2016
05:38 PM
5 Kudos
@PJ If you are doing manual registration then you don't need to setup passwordless ssh. Ambari needs passwordless ssh or root password only to bootstrap(install and configure) Ambari agents. You have already done that so no need to configure! Hope this information helps 🙂
... View more
11-28-2016
05:00 PM
4 Kudos
@Imtiaz Yousaf
/Other people who are reading this. This issue was resolved by removing hive-site.xml from below locations: 1. oozie sharelib (it was present at multiple locations in oozie sharelib) 2. from ${wf.application.path}/lib/ directory. 3. From workflow.xml (removed <job-xml> part) By default Oozie takes this file from /etc/oozie/conf/action-conf/hive/hive-site.xml Issue has been resolved! 🙂
... View more