Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1969 | 07-09-2019 12:53 AM | |
| 11879 | 06-23-2019 08:37 PM | |
| 9143 | 06-18-2019 11:28 PM | |
| 10127 | 05-23-2019 08:46 PM | |
| 4577 | 05-20-2019 01:14 AM |
07-22-2015
10:36 PM
Is the HBase service up and running? Could you check with the 'service' commands, and by visiting the HMaster and HRegionServer web UIs?
... View more
07-22-2015
09:32 AM
A direct property to change this is now available in the most recent CM 5.4.x releases. Once you upgrade, please remember to remove the env-var workaround so things continue to be configured consistently.
... View more
07-21-2015
07:09 AM
How are you bulk loading, specifically? Could you chown the prepared files as 'hbase' before you trigger the bulk load, or use the SecureBulkLoad technique offered by HBase?
... View more
07-20-2015
11:05 PM
In MultipleOutput's write(K, V, String) method, you can pass a relative path for the baseOutputPath (last) argument such as "my-sub-dir/my-filename-prefix". This will allow you to produce subdirectories under the job output directory. Please ensure to use only relative paths, and not absolute ones, as otherwise your output paths will not be respected by the FileOutputCommitter process.
... View more
07-20-2015
08:36 PM
1 Kudo
Thank you for following up! Please consider marking the thread resolved so others with similar questions may be able to locate topics with solutions faster.
... View more
07-20-2015
08:35 PM
Are you ensuring to call multipleOutputs.close() in the cleanup phase? If that is not done, the file won't be closed up, and would be left hanging open for an hour during which you can't see any written data. If you're already doing that, please share your code here so we can take a look at what may be going wrong.
... View more
07-20-2015
12:42 PM
While CDH releases are not vanilla Apache Hadoop ones (they carry additional backports), you are correct in that YARN-2424 isn't available in CDH 5.2.0. My bad for having you try that. It is available from CDH 5.3.0 onwards. Thanks for taking the time to mark a solution!
... View more
07-20-2015
02:37 AM
Thank you - it appears the logs are masking the thread failures at the INFO level. Could you change your logging config into DEBUG for the RM and retry to see if you notice any exceptions pertaining to the handler threads starting up? You can switch it in the log4j.properties file under /etc/hadoop/conf/ on the RM host. I see another email in the cdh-user@ group where something similar was observed for a user on Ubuntu, but that happened to them cause they had a mix-up with the zookeeper package provided by Ubuntu's own repository vs. CDH ones being installed. I am not sure if your Debian repository carries a zookeeper package as well but you should probably check via dpkg/etc. to ensure its a CDH version package and not a Debian-default one.
... View more
07-20-2015
02:35 AM
2 Kudos
Sqoop CLI currently does not have auto-retry capabilities for the situation you've mentioned. Oozie actions, as suggested by Ben, do have a retry specifier that you can leverage: http://archive.cloudera.com/cdh5/cdh/5/oozie/WorkflowFunctionalSpec.html#a18_User-Retry_for_Workflow_Actions_since_Oozie_3.1
... View more
07-19-2015
04:59 AM
Could you attach your full RM log here, via pastebin.com or such service, and also the .out file you will find alongside the .log file? The abort's happening cause we're unable to run some vital threads within the RM. From the short log excerpt though, the reason is not very clear.
... View more