Member since
02-02-2016
583
Posts
518
Kudos Received
98
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3174 | 09-16-2016 11:56 AM | |
1354 | 09-13-2016 08:47 PM | |
5344 | 09-06-2016 11:00 AM | |
3093 | 08-05-2016 11:51 AM | |
5168 | 08-03-2016 02:58 PM |
04-14-2016
03:41 PM
1 Kudo
@Maeve Ryan You can decrease space for non-HDFS use by setting dfs.datanode.du.reserved to a low value. It's set per disk volume. You can also freed it up by deleting any unwanted files from the datanode machine such as hadoop logs, any non hadoop related files (other information on the disk), etc. It cannot be done by using any hadoop commands.
... View more
04-14-2016
12:18 PM
Did you tried restarting secondary namnode? if not then first we would try to do restart.
... View more
04-13-2016
03:46 PM
2 Kudos
@wayne2chicagoPlease try below command. Instead of "in" it should be "on" and for username we only need to put back quote. hive> show grant user `ambari-qa` on ALL;
... View more
04-13-2016
02:45 PM
2 Kudos
@Adi Jabkowsky Since you added new high capacity nodes into the cluster therefore first it would be recommended to run HDFS balance multiple times specially when your cluster is less load or any idle window(for best output) and we should start balancer threshold value from a high number to lower value, Example start with 10 and end with 5. Though it is not guaranteed that you will get fully balanced cluster as per threshold in single shot but balancer will try it best during multiple runs. I think it tries to balance 10G of data at once. Also namenode has limits on how many blocks can be moved at one time, In addition to that you may also consider increasing below properties values to optimize balancer.
dfs.datanode.max.xcievers / dfs.datanode.max.transfer.threads dfs.datanode.balance.max.concurrent.moves dfs.datanode.balance.bandwidthPerSec
... View more
04-13-2016
01:01 PM
1 Kudo
Can you try to set below parameter on hive shell and run again? set hive.exec.stagingdir=/tmp/
... View more
04-11-2016
11:20 AM
2 Kudos
Can you please provide logs or error messages which you are getting? Also did you tried using below parameters on hive shell? set hive.execution.engine=spark;
set spark.master=yarn-client;
set spark.eventLog.enabled=true;
set spark.executor.memory=512m;
set spark.executor.cores=2;
... View more
03-28-2016
06:08 PM
1 Kudo
@mike pal Thanks Mike for confirming, would be great if you can click on best answer 🙂
... View more