Archives of Support Questions (Read Only)

This is an archived board for historical reference. Information and links may no longer be available or relevant
Announcements
This board is archived and read-only for historical reference. To ask a new question, please post a new topic on the appropriate active board.

java.io.FileNotFoundException: File file:/dfs/dn does not exist

avatar
Explorer

Hello, Folks 
I got partial cloudera 5.8.2 server running. 
So far I have ZooKeeper and Name Node, NFS Gateway service is running. 
I can't get HDFS services to start on data node. 

 

Keep getting - the Following error 

 

java.io.FileNotFoundException: File file:/dfs/dn does not exist

 

Google is not very helpful either - 

Any suggestion ...? Please advise. 

 

---------

DFS Info on data node 1. 

 

root@hadoop-dn01:/# ll dfs
total 20
drwx------ 5 hdfs hdfs 4096 Sep 30 22:32 ./
drwxr-xr-x 26 root root 4096 Sep 30 14:09 ../
drwx------ 3 hdfs hdfs 4096 Sep 26 16:19 dn/
drwx------ 2 hdfs hdfs 4096 Sep 26 20:19 nn/
drwxr-xr-x 2 hdfs hdfs 4096 Sep 30 22:32 snn/

 

======

 

root@hadoop-dn01:/# ll -l /dfs/dn
total 24
drwx------ 3 hdfs hdfs 4096 Sep 26 16:19 ./
drwx------ 5 hdfs hdfs 4096 Sep 30 22:32 ../
drwx------ 2 root root 16384 Sep 26 16:19 lost+found/

-------

 

Installation_dataNode_09_30_16.JPG

 

 


1 ACCEPTED SOLUTION

avatar
Explorer

Ok, problem indeed with permission. 

/DFS ==>  root
/dfs/dn01 ==> root 

============== 
Correct permission:  

Solution:: 

 

/dfs ==> cloudera-scm:cloudera-scm
/dfs/dn01 ==> hdfs:hdfs

chown cloudera-scm:cloudera-scm dfs


cd /dfs

chown hdfs:hdfs dn01



View solution in original post

4 REPLIES 4

avatar
Rising Star

Hello,

 

I have a couple of questions:

 

1. Have you checked the value of your datanode directory in Cloudera Manager (CM -> HDFS -> Configuration-> DataNode Data Directory)?  It should state /dfs/dn.

 

2. Run the following command:

 

sudo chown -R hdfs:hdfs /dfs/dn

 

3. If that does not fix the issue, delete /dfs/dn, and restart HDFS. 

avatar
Explorer

I check the configuration and start/restart the HDFS services. No luck ..still seeing the same problem. 

I even try the different path and mount the storage to a different location. Still same problem.  

 

Old /dfs/dn

new /dn01

root@hadoop-dn01:/# ll /dn01
total 24
drwx------ 3 hdfs hdfs 4096 Sep 26 16:19 ./
drwxr-xr-x 27 root root 4096 Oct 3 18:01 ../
drwx------ 2 hdfs hdfs 16384 Sep 26 16:19 lost+found/
root@hadoop-dn01:/#

 

---------------------

 Oct 3, 6:45:31.023 PM DEBUG org.apache.htrace.core.Tracer
span.receiver.classes = ; loaded no span receivers
Oct 3, 6:45:31.161 PM DEBUG org.apache.hadoop.io.nativeio.NativeIO
Initialized cache for IDs to User/Group mapping with a cache timeout of 14400 seconds.
Oct 3, 6:45:31.161 PM WARN org.apache.hadoop.hdfs.server.datanode.DataNode
Invalid dfs.datanode.data.dir /dn01 :
EPERM: Operation not permitted
at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmodImpl(Native Method)
at org.apache.hadoop.io.nativeio.NativeIO$POSIX.chmod(NativeIO.java:230)
at org.apache.hadoop.fs.RawLocalFileSystem.setPermission(RawLocalFileSystem.java:708)
at org.apache.hadoop.fs.ChecksumFileSystem$1.apply(ChecksumFileSystem.java:500)
at org.apache.hadoop.fs.ChecksumFileSystem$FsOperation.run(ChecksumFileSystem.java:481)
at org.apache.hadoop.fs.ChecksumFileSystem.setPermission(ChecksumFileSystem.java:497)
at org.apache.hadoop.util.DiskChecker.mkdirsWithExistsAndPermissionCheck(DiskChecker.java:140)
at org.apache.hadoop.util.DiskChecker.checkDir(DiskChecker.java:156)
at org.apache.hadoop.hdfs.server.datanode.DataNode$DataNodeDiskChecker.checkDir(DataNode.java:2458)
at org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2500)
at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2482)
at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2374)
at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:2421)
at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2598)
at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2622)
Oct 3, 6:45:31.164 PM FATAL org.apache.hadoop.hdfs.server.datanode.DataNode
Exception in secureMain
java.io.IOException: All directories in dfs.datanode.data.dir are invalid: "/dn01/"
at org.apache.hadoop.hdfs.server.datanode.DataNode.checkStorageLocations(DataNode.java:2509)
at org.apache.hadoop.hdfs.server.datanode.DataNode.makeInstance(DataNode.java:2482)
at org.apache.hadoop.hdfs.server.datanode.DataNode.instantiateDataNode(DataNode.java:2374)
at org.apache.hadoop.hdfs.server.datanode.DataNode.createDataNode(DataNode.java:2421)
at org.apache.hadoop.hdfs.server.datanode.DataNode.secureMain(DataNode.java:2598)
at org.apache.hadoop.hdfs.server.datanode.DataNode.main(DataNode.java:2622)
Oct 3, 6:45:31.166 PM INFO org.apache.hadoop.util.ExitUtil
Exiting with status 1
Oct 3, 6:45:31.169 PM INFO org.apache.hadoop.hdfs.server.datanode.DataNode
SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down DataNode at hadoop-dn01.fhmc.local/10.29.1.4
************************************************************/

--------------------

 

Same problem!! -

avatar
Explorer
Has anybody has any suggestion?
still getting same error.

avatar
Explorer

Ok, problem indeed with permission. 

/DFS ==>  root
/dfs/dn01 ==> root 

============== 
Correct permission:  

Solution:: 

 

/dfs ==> cloudera-scm:cloudera-scm
/dfs/dn01 ==> hdfs:hdfs

chown cloudera-scm:cloudera-scm dfs


cd /dfs

chown hdfs:hdfs dn01