Member since
07-31-2013
1924
Posts
462
Kudos Received
311
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
1543 | 07-09-2019 12:53 AM | |
9292 | 06-23-2019 08:37 PM | |
8050 | 06-18-2019 11:28 PM | |
8676 | 05-23-2019 08:46 PM | |
3473 | 05-20-2019 01:14 AM |
07-20-2014
06:37 AM
1 Kudo
Since Oozie lacks knowledge of where your HBase configs lie, you will need to pass the client hbase-site.xml file (placed somewhere on HDFS, by copying from /etc/hbase/conf/hbase-site.xml on any HBase gateway node) via the <job-xml>…</job-xml> option. Alternatively, try the below command instead (will not be sufficient for secured clusters, which need further properties), replacing zk-host1,zk-host2,zk-host3 with your actual 3 hosts appropriately: sqoop import -Dhbase.zookeeper.quorum=zk-host1,zk-host2,zk-host3 --connect jdbc:oracle:thin:@XXX:port/XXX --username XXX --password XXX --table XXX -m 1 --incremental lastmodified --last-value '2014-06-23' --check-column XXX --append --hbase-table XXX --column-family info --hbase-row-key XXX --hbase-bulkload
... View more
07-20-2014
12:45 AM
Where exactly is the hive.metastore.sasl.enabled property applied? Are you certain it is applied to the running HiveMetaStore server? Does a regular Hive CLI configured with hive.metastore.uris instead of DB properties run properly (i.e. show tables, etc. work fine)?
... View more
07-20-2014
12:42 AM
1 Kudo
Please post your cluster's memory configuration, such as the resource MB offered by the NodeManagers, and individual MapReduce settings of AM, Map and Reduce task memories. It appears that the cluster's unable to schedule more than 1 or 2 containers at a time, causing the job to eternally hang cause Oozie runs 2x AMs grabbing 2x containers already.
... View more
07-19-2014
11:19 PM
1 Kudo
Can you post more details on what you mean by 'multiple applications' (and how many, exactly), as well as your scheduler configuration? What behaviour do you notice exactly when you say they all 'stop'. Do you mean their AppMasters run but the actual application containers (i.e. map or reduce tasks) do not run, or do you mean they all just fail?
... View more
07-19-2014
11:16 PM
Do you perhaps have safety valve overrides in your Hue Configuration page in CM that is setting the default mapred_clusters to a MR1 location? If so, please remove them away since you have switched over to YARN now. Doing that will resolve it. Also ensure that Hue's MapReduce Cluster setting is set to your YARN service and not the MR1 service.
... View more
07-19-2014
10:53 PM
2 Kudos
Yes, the reason of the 200k default is to warn you that you may be facing a small files issue in your cluster, or that you may be close to requiring to expand further horizontally. Having more number of blocks raises the heap requirement at the DataNodes. The threshold warning exists to also notify you about this (that you may need to soon raise the DN heap size to allow it to continue serving blocks at the same performance). With CM5 we have revised the number to 600k, given memory optimisation improvements for DNs in CDH4.6+ and CDH5.0+. You can feel free to raise the threshold via the CM -> HDFS -> Configuration -> Monitoring section fields, but do look into if your users have begun creating too many tiny files as it may hamper their job performance with overheads of too many blocks (and thereby, too many mappers).
... View more
07-19-2014
10:23 PM
It is difficult to say if you are hitting a bug without looking at relevant Checkpointer placed entries in the StandbyNameNode (SBN) logs. There may be issues with transferring the file between the SBN and the NN, probably cause of timeouts or otherwise.
... View more
07-19-2014
10:17 PM
If CM does not present a UI field for an advanced tuning property, you can rely on the Configuration Snippet (Safety Valve) fields to set them in manually. More on this at: http://www.cloudera.com/content/cloudera-content/cloudera-docs/CM4Ent/4.8.1/Cloudera-Manager-Managing-Clusters/cmmc_safety_valve.html
... View more
07-19-2014
10:06 PM
2 Kudos
You receive the error because the 'hbase' user does not have a login shell assigned to it. You can set a shell for the 'hbase' user on the machine, to allow direct 'su' based login to that user, by following http://www.cyberciti.biz/faq/howto-set-bash-as-your-default-shell/ However, if your goal is to simply use the 'hbase' user for running superuser level commands, we instead recommend using 'sudo' style commands. For example: ~> sudo -u hbase hbase hbck ~> sudo -u hbase hbase shell You can also invoke a shell as the 'hbase' user in certain cases, via: ~> sudo -u hbase /bin/bash
... View more
07-14-2014
11:46 AM
Your java program needs to include cluster client configs on its classpath for the Configuration class to be able to read and discover the actual MR cluster automatically. Typically you can achieve this by adding the directory /etc/hadoop/conf to your classpath, if you are not launching your custom application using the 'hadoop jar' command (which auto-sets the desirable classpath).
... View more