Member since
08-08-2017
1652
Posts
30
Kudos Received
11
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 1970 | 06-15-2020 05:23 AM | |
| 16069 | 01-30-2020 08:04 PM | |
| 2108 | 07-07-2019 09:06 PM | |
| 8247 | 01-27-2018 10:17 PM | |
| 4674 | 12-31-2017 10:12 PM |
11-10-2017
12:50 PM
hi Aditya , on each worker machine we have 5 volumes , and we not want to stay with 4 volume on the problematic workers , so about option 2 we not want to remove the volume , second what is the meaning to set the dfs.datanode.failed.volumes.tolerated to 1 ? after HDFS restart - it will fix the problem ?
... View more
11-10-2017
12:44 PM
hi Jay - grep dfs.datanode.failed.volumes.tolerated /etc/hadoop/conf/hdfs-site.xml <name>dfs.datanode.failed.volumes.tolerated</name> this already set in the xml file
... View more
11-10-2017
12:37 PM
in my Ambari cluster we have some services that not installed yet as graphite service as described here ( in the picture ) what are the API commands that required in order to identify which service is/are available to install and the API that install the service
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
11-10-2017
12:29 PM
here the permissions : ls -ltr /xxxxx/sdc/hadoop/hdfs/data/ drwxr-xr-x. 3 hdfs hadoop 4096 current -rw-r--r--. 1 hdfs hadoop 28 in_use.lock
... View more
11-10-2017
12:27 PM
on the second approach if we removed the folder /xxxx/sdc/hadoop/hdfs/data on the problematic worker and then we restart the HDFS component on the worker then it will create the folder - data again ?
... View more
11-10-2017
12:25 PM
hi Aditya , not clear for me if we change the fs.datanode.failed.volumes.tolerated to then it will affected all workers machine and we have problem only on worker01 , so do you mean that we need to change it to 1 and restart the HDFS service and then return it to 0?
... View more
11-10-2017
10:54 AM
when we start the data node on one of the workers machine we get: ERROR datanode.DataNode (DataNode.java:secureMain(2691)) - Exception in secureMain
org.apache.hadoop.util.DiskChecker$DiskErrorException: Too many failed volumes - current valid volumes: 4, volumes configured: 5, volumes failed: 1, volume failures tolerated: 0 and this WARN checker.StorageLocationChecker (StorageLocationChecker.java:check(208)) - Exception checking StorageLocation [DISK]file:/grid/sdc/hadoop/hdfs/data/
org.apache.hadoop.util.DiskChecker$DiskErrorException: Directory is not writable: /xxxx/sdc/hadoop/hdfs/data what are the steps that needs to fix it?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
11-06-2017
09:51 AM
in ambari GUI we want to add the graphite service so we select actions --> add service --> check the box of graphite --> assign masters --> assign slaves and masters but when click next to the next screen "assign slaves and masters" this screen stay only 2 second and then we get the customize services ( so we cant check the boxes in the screen of assign slaves and masters because its disappeared after 2 seconds , why this happened ?
... View more
Labels:
- Labels:
-
Apache Ambari
-
Apache Hadoop
11-05-2017
10:24 AM
yes the IP's and hostname are now ok , but still cant start the ambari-agent do you think need to restart the proccess that hold the port - 10000 ?
... View more
11-05-2017
09:22 AM
when we run the netstat -tnlpa | grep 10000 , we get tcp 0 0 45.89.12.111:10000 45.89.12.110:44570 ESTABLISHED 15598/java
tcp 0 0 45.89.12.111:10000 45.89.12.110:55109 ESTABLISHED 15598/java regarding the iptables it is stooped , and we get the output - connection refused from nc command , and the full machine name is - machine-master03.pop.com
... View more