<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: cant start DataNode  from ambari cluster in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193396#M155460</link>
    <description>&lt;P&gt;here the permissions :&lt;/P&gt;&lt;P&gt;ls -ltr /xxxxx/sdc/hadoop/hdfs/data/ &lt;/P&gt;&lt;P&gt;drwxr-xr-x. 3 hdfs hadoop 4096  current&lt;/P&gt;&lt;P&gt;-rw-r--r--. 1 hdfs hadoop   28  in_use.lock &lt;/P&gt;</description>
    <pubDate>Fri, 10 Nov 2017 20:29:46 GMT</pubDate>
    <dc:creator>mike_bronson7</dc:creator>
    <dc:date>2017-11-10T20:29:46Z</dc:date>
    <item>
      <title>cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193392#M155456</link>
      <description>&lt;P&gt;when we start the data node on one of the workers machine we get:&lt;/P&gt;&lt;P&gt;ERROR datanode.DataNode (DataNode.java:secureMain(2691)) - Exception in secureMain
org.apache.hadoop.util.DiskChecker$DiskErrorException: Too many failed volumes - current valid volumes: 4, volumes configured: 5, volumes failed: 1, volume failures tolerated: 0&lt;/P&gt;&lt;P&gt;and this &lt;/P&gt;&lt;P&gt;WARN  checker.StorageLocationChecker (StorageLocationChecker.java:check(208)) - Exception checking StorageLocation [DISK]file:/grid/sdc/hadoop/hdfs/data/
org.apache.hadoop.util.DiskChecker$DiskErrorException: Directory is not writable: /xxxx/sdc/hadoop/hdfs/data&lt;/P&gt;&lt;P&gt;what are the steps that needs to fix it?&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 18:54:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193392#M155456</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T18:54:10Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193393#M155457</link>
      <description>&lt;P&gt; &lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Check if you have write permissions for '/xxxx/sdc/hadoop/hdfs/data' . Change the ownership to hdfs:hadoop&lt;/P&gt;&lt;PRE&gt;chown hdfs:hadoop /xxxx/sdc/hadoop/hdfs/data&lt;/PRE&gt;&lt;P&gt; If you are okay with failed volumes ,then you can change  'dfs.datanode.failed.volumes.tolerated' to 1 or another solution is to  remove the above directory(/xxxx/sdc/hadoop/hdfs/data) from 'dfs.datanode.data.dir'&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;Aditya&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:13:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193393#M155457</guid>
      <dc:creator>asirna</dc:creator>
      <dc:date>2017-11-10T20:13:43Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193394#M155458</link>
      <description>&lt;P&gt;hi Aditya , not clear for me if we change the fs.datanode.failed.volumes.tolerated to then it will affected all workers machine and we have problem only on worker01 , so do you mean that we need to change it to 1 and restart the HDFS service and then return it to 0?&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:25:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193394#M155458</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:25:14Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193395#M155459</link>
      <description>&lt;P&gt; on the second approach if we  removed the folder /xxxx/sdc/hadoop/hdfs/data on the problematic worker  and then we restart the HDFS component on the worker then it will create the folder - data again ?&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:27:04 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193395#M155459</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:27:04Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193396#M155460</link>
      <description>&lt;P&gt;here the permissions :&lt;/P&gt;&lt;P&gt;ls -ltr /xxxxx/sdc/hadoop/hdfs/data/ &lt;/P&gt;&lt;P&gt;drwxr-xr-x. 3 hdfs hadoop 4096  current&lt;/P&gt;&lt;P&gt;-rw-r--r--. 1 hdfs hadoop   28  in_use.lock &lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:29:46 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193396#M155460</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:29:46Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193397#M155461</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;/A&gt;&lt;/P&gt;&lt;PRE&gt;WARN checker.StorageLocationChecker (StorageLocationChecker.java:check(208)) - Exception checking 
StorageLocation [DISK]file:/grid/sdc/hadoop/hdfs/data/ org.apache.hadoop.util.DiskChecker$DiskErrorException: Directory is not writable: /xxxx/sdc/hadoop/hdfs/data&lt;/PRE&gt;&lt;P&gt;The above error can occur sometimes whet the Hard Disk/Filesystem has gone bad and the filesystem is in Read-Only mode. Remounting might help. Please check for any hardware errors. Check the harddisk and remount the Volume.&lt;BR /&gt;&lt;BR /&gt;Also it will be good to see "/etc/hadoop/conf/hdfs-site.xml" property "dfs.datanode.failed.volumes.tolerated" this will set the disk failure tolerance.&lt;BR /&gt;&lt;/P&gt;&lt;PRE&gt;&amp;lt;property&amp;gt;
     &amp;lt;name&amp;gt;dfs.datanode.failed.volumes.tolerated&amp;lt;/name&amp;gt;
     &amp;lt;value&amp;gt;1&amp;lt;/value&amp;gt;
&amp;lt;/property&amp;gt; &lt;/PRE&gt;&lt;P&gt;.&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:31:48 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193397#M155461</guid>
      <dc:creator>jsensharma</dc:creator>
      <dc:date>2017-11-10T20:31:48Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193398#M155462</link>
      <description>&lt;P&gt; &lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;1) First solution is to try changing the ownership of the directory and restart. If this works then no need to change anything.&lt;/P&gt;&lt;P&gt;2) If #1 doesn't work and you are ok to remove this volume, then remove the directory from "dfs.datanode.data.dir" and let the value of 'dfs.datanode.failed.volumes.tolerated' remain 0&lt;/P&gt;&lt;P&gt;3) If you do not want to remove this volume and you are okay with this failed volume and continue then set 'dfs.datanode.failed.volumes.tolerated' to 1&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:35:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193398#M155462</guid>
      <dc:creator>asirna</dc:creator>
      <dc:date>2017-11-10T20:35:43Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193399#M155463</link>
      <description>&lt;P&gt;hi Jay -&lt;/P&gt;&lt;P&gt; grep dfs.datanode.failed.volumes.tolerated  /etc/hadoop/conf/hdfs-site.xml &lt;/P&gt;&lt;P&gt;      &amp;lt;name&amp;gt;dfs.datanode.failed.volumes.tolerated&amp;lt;/name&amp;gt;&lt;/P&gt;&lt;P&gt;this already set in the xml file&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:44:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193399#M155463</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:44:31Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193400#M155464</link>
      <description>&lt;P&gt;hi Aditya , on each worker machine we have 5 volumes , and we not want to stay with 4 volume on the problematic workers , so about option 2 we not want to remove the volume , second what is the meaning to set the dfs.datanode.failed.volumes.tolerated to 1 ? after HDFS restart - it will fix the problem ? &lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:50:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193400#M155464</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:50:33Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193401#M155465</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;BR /&gt;&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml" target="_blank"&gt;https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-hdfs/hdfs-default.xml&lt;/A&gt;&lt;/P&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;BR /&gt;dfs.datanode.failed.volumes.tolerated&lt;/TD&gt;&lt;TD&gt;0&lt;/TD&gt;&lt;TD&gt;The number of volumes that are allowed to fail before a datanode stops offering service. By default any volume failure will cause a datanode to shutdown.&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;The default value is 0.   Please set it to 1 and then try again.  Or please fix the Failed Volume.&lt;/P&gt;&lt;P&gt;.&lt;BR /&gt;&lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;&lt;/A&gt; &lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:53:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193401#M155465</guid>
      <dc:creator>jsensharma</dc:creator>
      <dc:date>2017-11-10T20:53:20Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193402#M155466</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;If you set dfs.datanode.failed.volumes.tolerated to 'x', it will allow maximum of 'x' no of volumes to be failed. So HDFS restart should fix it.&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:56:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193402#M155466</guid>
      <dc:creator>asirna</dc:creator>
      <dc:date>2017-11-10T20:56:19Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193403#M155467</link>
      <description>&lt;P&gt;hi Jay , I have idea but not sure about this so I need your advice , on the problematic worker we have extra volume - sdg , and the bad volume is sdf , so maybe we need to umount the sdf and mount the volume sdg in place sdf , and change the DataNode directories from ambari GUI from sdf to sdg - and then restart the component HDFS on the worker , what you think ? &lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 20:56:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193403#M155467</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T20:56:31Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193404#M155468</link>
      <description>&lt;P&gt;another remark if I set this value to 1 it mean that HDFS will start up in spite the volume is bad ? or not in use  , &lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 21:14:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193404#M155468</guid>
      <dc:creator>mike_bronson7</dc:creator>
      <dc:date>2017-11-10T21:14:31Z</dc:date>
    </item>
    <item>
      <title>Re: cant start DataNode  from ambari cluster</title>
      <link>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193405#M155469</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/26229/uribarih.html" nodeid="26229"&gt;@Michael Bronson&lt;/A&gt;&lt;P&gt;Yes. It will startup inspite the volume is bad. If you dont want this to happen you might have to replace your failed volume with a new volume (ie unmount old one and mount new one)&lt;/P&gt;</description>
      <pubDate>Fri, 10 Nov 2017 21:22:47 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/cant-start-DataNode-from-ambari-cluster/m-p/193405#M155469</guid>
      <dc:creator>asirna</dc:creator>
      <dc:date>2017-11-10T21:22:47Z</dc:date>
    </item>
  </channel>
</rss>

