Created 01-17-2019 02:47 PM
We are trying to fetch file from HDFS (hdfs://10.140.176.148:9000/),we are able to list the files on HDFS however while fetching the file we are getting below error : hdfs.DFSClient: Failed to connect to /192.168.20.8:50010 for block, add to deadNodes and continue. org.apache.hadoop.net.ConnectTimeoutException: 60000 millis timeout while waiting for channel to be ready for connect. ch : java.nio.channels.SocketChannel[connection-pending remote=/192.168.20.8:50010] it seems the IP to which the request is internally redirected is not reachable from our client servers : Client server : 10.141.80.117 10.141.80.119 10.141.80.128 10.141.80.129 10.143.168.75 Could you please help us to resolve the issue .
Created 01-18-2019 04:58 AM
@Jay Kumar SenSharma can you plz help me with that?
Created 01-18-2019 05:35 AM
Is there any specific reason you are using IP Address instead of Hostnames ? Using Hostnames instead of IP address is always recommended.
.
Also the error shows
Failed to connect to /192.168.20.8:50010
.
So from the machine where you are running the HDFS fetch commands are you able to access the mentioned DataNode host & port?
# telnet 192.168.20.8 50010 (OR) # nc -v 192.168.20.8 50010
.
It might be either Port blocking issue / like firewall issue.
Or the DataNode on the mentioned IPAddress / port might not be listening.
So please SSH to the DataNode host "192.168.20.8" and check if the port is opened and listening on 50010 ?
# netstat -tnlpa | grep -i DataNode
.
If you see that port is not opened OR if it is not bind to the intended network interface (ip address) then please fix it or share the DataNode logs to see if there are any errors
Created 01-18-2019 06:09 AM
@Jay Kumar SenSharma sure , i will check and update. Thanks!!