Member since
07-30-2019
111
Posts
185
Kudos Received
35
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
3072 | 02-07-2018 07:12 PM | |
2311 | 10-27-2017 06:16 PM | |
2628 | 10-13-2017 10:30 PM | |
4823 | 10-12-2017 10:09 PM | |
1215 | 06-29-2017 10:19 PM |
04-24-2017
08:13 PM
1 Kudo
Hi @Michael Häusler, this may be caused by HDFS-9958. I see that HDFS-9958 is not fixed in HDP 2.4.2 but it was fixed in HDP 2.4.3. If you can see this consistently I'd recommend upgrading to check whether that fixes the problem. If you have a support contract we can provide you with a hotfix release.
... View more
04-23-2017
08:43 PM
Check your rack setting for the DataNode. If you don't see the problem you can post the output of the following command and someone may be able to point out the error. hdfs dfsadmin -report
... View more
04-18-2017
02:46 PM
Hi @Sedat Kestepe, take a look at rack awareness. https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-common/RackAwareness.html Here's how you can configure racks using Ambari https://docs.hortonworks.com/HDPDocuments/Ambari-2.2.0.0/bk_Ambari_Users_Guide/content/ch03s11.html HDFS will avoid placing all block replicas in the same rack to avoid data loss in case of a rack failure. You may be able to use this to achieve what you want.
... View more
04-06-2017
04:33 PM
1 Kudo
Using RAID can reduce the availability and fault tolerance of HDFS. It certainly reduces the overall performance as compared to JBOD. We strongly recommend configuring your disks as JBOD since HDFS already stores data redundantly by replicating across nodes/racks and can automatically recover from disk and node failures.
... View more
04-05-2017
01:46 PM
I'd also post this question on the Ambari track to check why Ambari didn't detect the DataNodes doing down. Also from your logs it is hard to say why the DataNode went down. I again recommend increasing the DataNode heap allocation via Ambari. Also check that your nodes are provisioned with sufficient amount of RAM.
... View more
04-05-2017
01:41 PM
Ok looks like you have automatic failover enabled. I am not sure why you get the EOFException. Look through your NameNode logs to see if there are any errors.
... View more
04-05-2017
01:36 PM
The Mover will move blocks within the same node when possible and thus try to avoid network activity. If that is not possible (e.g. when a node doesn't have SSD or when the local SSDs are full), it will move block replicas across the network to another node that has the target media. I've edited my answer.
... View more
04-04-2017
09:06 PM
@Riccardo Iacomini, are you asking about the HDFS move/rename command? Move is purely a metadata operation on the NameNode and does not result in any data movement until the HDFS Mover utility is run. https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/ArchivalStorage.html#Mover_-_A_New_Data_Migration_Tool Edit: The Mover will move blocks within the same node when possible and thus try to avoid network activity.
If
that is not possible (e.g. when a node doesn't have SSD or when the
local SSDs are full), it will move block replicas across the network to
another node that has the target media.
... View more
04-03-2017
11:14 PM
1 Kudo
You likely have Kerberos enabled. The DataNode process starts as root so it can bind a privileged port (<1024) for data transfer. Then it launches another process as user hdfs. You should not kill either process. The "refused to connect" error looks like some network connectivity issue in your environment, or you are hitting the wrong port number. See if you can find the correct info port from either configuration or from the DataNodes tab of the NameNode web UI.
... View more
04-03-2017
06:23 PM
1 Kudo
Are you using HDP and did you enable NameNode HA using Ambari? If so then you should have automatic failover configured. Automatic Failover requires the ZooKeeper service instances and ZooKeeper FailoverControllers to be up and running. If you setup HA manually, then you may need to transition one of the NNs to active status manually as described here: https://hadoop.apache.org/docs/stable/hadoop-project-dist/hadoop-hdfs/HDFSHighAvailabilityWithQJM.html
... View more