Member since
09-24-2015
816
Posts
488
Kudos Received
189
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3171 | 12-25-2018 10:42 PM | |
| 14192 | 10-09-2018 03:52 AM | |
| 4763 | 02-23-2018 11:46 PM | |
| 2481 | 09-02-2017 01:49 AM | |
| 2910 | 06-21-2017 12:06 AM |
04-11-2016
03:20 PM
2 Kudos
As far as I know, up to Knox-0.6 (currently used in HDP) there can be only one Auth. provider per topology. Last time when I faced such a requirement I created two topologies, one for each Auth. provider. Edit: KNOX-236 is going to provide such a feature but "in the future".
... View more
04-11-2016
03:13 PM
1 Kudo
No, it must be numeric because according to the specs: "By default sqoop will use query select min(<split-by>), max(<split-by>) from <table name> to find out boundaries for creating splits." The alternative is to use --boundary-query which also requires numeric columns. Otherwise the Sqoop job will fail. If you don't have such a column in your table the only workaround is to use only 1 mapper: "-m 1".
... View more
04-11-2016
02:23 PM
Can you double check Base URLs of HDP and HDP-UTILS you used to register you new version. If one of them is wrong, or not accessible then the registration will fail. You can test using wget from your Ambari server node: wget <HDP-Base-URL>/repodata/repomd.xml
wget <HDP-UTILS-Base-URL>/repodata/repomd.xml
... View more
04-10-2016
07:29 AM
Hi @Thomas Forsyth, welcome to HCC! You can give Ben points by accepting his answer. There is an "Accept" link below each answer. But in this case only @Maharaj Muthusamy who asked the question can do it.
... View more
04-10-2016
01:36 AM
1 Kudo
Hi @Rajendra Vechalapu, you can omit setting master in your source, see this example: val conf = new SparkConf().setAppName("Spark Pi")
val spark = new SparkContext(conf) You can then launch your application using spark-submit and provide the master there there using "--master" and "--deploy-mode" options. Refer to Spark programming guide for this and other useful hints. Edit: When you run spark-submit on Sandbox, be sure to supply additional arguments for master, num-executors, driver-memory, executor-memory, and executor-cores as given below. Note that larger values for last 4 arguments will not work on the Sandbox! Follow (and you can also try) this example computing Pi in Python (as any user who has access to HDFS/Yarn): cd /usr/hdp/current/spark-client/
spark-submit --master yarnclient --num-executors 1 --driver-memory 512m --executor-memory 512m --executor-cores 1 examples/src/main/python/pi.py 10
"--master yarncluster" works too. You can also set these 4 in spark-env in Ambari. They are already there but commented out, and not all with values like here. See also Spark guide on HDP.
... View more
04-09-2016
04:29 AM
3 Kudos
I saw this, and I think it's fixed in latest Ambari versions. Can you try: cd /usr/lib/python2.6/site-packages/resource_monitoring/
python psutil/build.py and restart metrics monitors.
... View more
04-09-2016
01:24 AM
1 Kudo
Your port number for Kafka broker is wrong, the default on Sandbox is 6667. You are using 2181 which is the Zookeeper's port.
... View more
04-09-2016
01:20 AM
Hi @Wellington De Oliveira, change that symlink, it's still pointing to the old version of the connector: cd /usr/share/java
rm mysql-connector-java.jar
ln -s mysql-connector-java-5.1.38.jar mysql-connector-java.jar Check the name of the new version, not sure is it as I wrote.
... View more
04-09-2016
12:56 AM
1 Kudo
Use fsck, it's a tool of choice to manage HDFS. "Orphans" are corrupted files (with missing blocks) in HDFS lingo. You can use "-move" or "-delete" options to move corrupted files to /lost+found or to delete them. fsck will also tell you about under-replicated blocks (having at least 1 replica but less than configured replication factor) but HDFS will repair them little by little by creating missing replicas.
... View more
04-08-2016
06:59 AM
Hi @Sunile Manjee, have you seen the answers? Please consider to accept/upvote the helpful ones.
... View more