Member since
10-01-2015
3933
Posts
1150
Kudos Received
374
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 3560 | 05-03-2017 05:13 PM | |
| 2935 | 05-02-2017 08:38 AM | |
| 3184 | 05-02-2017 08:13 AM | |
| 3147 | 04-10-2017 10:51 PM | |
| 1624 | 03-28-2017 02:27 AM |
08-16-2016
11:33 AM
HDP 2.3.6 and HDP 2.4.2 are pretty much the same except for Hive 1.2.1 and Spark 1.6. Kafka mirror maker is now supported and perhaps your issue is resolved in 2.3.6 as well. http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.6/bk_HDP_RelNotes/content/new_features.html I think however there's a documentation issue in the release notes that contradicts this statement and I will follow up. Please try upgrading to 2.3.6 or test this on another cluster with 2.3.6 to confirm.
... View more
08-16-2016
11:20 AM
1 Kudo
I remember having this issue a long time ago. It's amazing it still lingers. I don't remember my workaround but perhaps the dirty way to clean up is to comment out the kill and shutdown statements http://stackoverflow.com/questions/16658779/twitter-storm-example-running-in-local-mode-cannot-delete-file
... View more
08-15-2016
04:33 PM
@Nic Hopper yes so now you're good to go as you have access to HDFS via CLI (Command Shell). You can now upload and download files using CLI sudo -u username hdfs dfs -put file /user/username/
sudo -u username hdfs dfs -get file /user/username/file if you have your username available locally on the OS then you can skip the part with sudo -u username just change to your userID and run those commands su username
hdfs dfs -ls
hdfs dfs -put
hdfs dfs -get also as a good practice, please accept the answer to close this thread.
... View more
08-13-2016
09:45 PM
Run the following command, then repeat the earlier commands sudo -u hdfs hdfs dfsadmin -safemode leave
... View more
08-13-2016
05:40 PM
1 Kudo
You might have to distribute the new binary across the cluster. looks like you're hitting Spark-8032 which was committed after Spark 1.3.1, you should consider upgrading your cluster to latest HDP. https://github.com/apache/spark/commit/22703dd79fecc844d68033358f3201fd8a8f95cb
... View more
08-13-2016
05:21 PM
1 Kudo
You can't install Spark 1.5.2 on HDP 2.4.x unless you build it from source and deploy it yourself. you can get the sides from Apache Spark webpage. as long as you keep your Spark init scripts in a non-hdp specific directory you should be good. make sure /usr/bin/spark does not collide with HDP spark. It's best if you wait for SAP Bods 1.3 which I believe will support spark 1.6. in the next major HDP release will be able to run two independent versions of spark simultaneously (1.6.2 and 2.0 technical preview).
... View more
08-13-2016
04:08 PM
2 Kudos
create your user directory sudo -u hdfs hdfs dfs -mkdir /user/username
sudo -u hdfs hdfs dfs -chown -R hdfs:username /user/username
... View more
08-11-2016
07:50 PM
Please consider publishing an article on this, others will find it useful as it's not an obvious find.
... View more
08-11-2016
06:10 PM
@Biswajit Chakraborty I can't believe I didn't think of this earlier, take a look at https://apache.googlesource.com/mrunit/+/e43ef01dd1199a7eb0963edbf05258a8609bf0dc/src/test/java/org/apache/hadoop/mrunit/mapreduce/TestDistributedCache.java This is MRUnit's own DistributedCache test with examples how to set it up.
... View more
08-11-2016
06:05 PM
unfortunately there are none that I know of. We do have one engineer who's providing support for his unit-testing framework and it has mapreduce unit testing capabilities, though I am not sure if distributedcache testing is supported. https://github.com/sakserv/hadoop-mini-clusters
... View more