<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Volume failure reported while disks seem fine in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/270252#M4106</link>
    <description>&lt;P&gt;Yes even in my case fstab issue was not there. Rather i could clearly find the issue from datanode logs under /var/log/ and fix the issue. I have a blog on the same as below. Please comment on blog if it helps.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;A href="https://itdoctorharsha.wordpress.com/2019/08/23/hadoop-data-volume-failures-and-solution-cloudera/" target="_blank" rel="noopener"&gt;https://itdoctorharsha.wordpress.com/2019/08/23/hadoop-data-volume-failures-and-solution-cloudera/&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
    <pubDate>Wed, 11 Sep 2019 22:40:00 GMT</pubDate>
    <dc:creator>Lambzee</dc:creator>
    <dc:date>2019-09-11T22:40:00Z</dc:date>
    <item>
      <title>Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22706#M4095</link>
      <description>&lt;P&gt;I just upgraded our cluster from CDH 5.0.1 to 5.2.1, using parcels and following&amp;nbsp;the provided instructions.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;After the upgrade has finished, the health test "Data Directory Status" is critical for one of the data nodes. The reported error message is "&lt;SPAN&gt;The DataNode has 1 volume failure(s)". By running 'hdfs dfsadmin -report' I can also confirm that the available HDFS space on that node is approximately 4 TB less than on the other&amp;nbsp;&lt;/SPAN&gt;nodes, indicating that one of the disks is not being used.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;However, when checking the status of the actual disks and regular file system we can not find anything that seems wrong. All disks are mounted and seem to be working as they should. There is also an in_use.lock file in the dfs/nn directory on all of the disks.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;How can I get more detailed information about which volume the DataNode is complaining about, and what the issue might be?&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Best Regards&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;\Knut Nordin&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 16 Dec 2014 13:00:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22706#M4095</guid>
      <dc:creator>Knut N</dc:creator>
      <dc:date>2014-12-16T13:00:16Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22775#M4096</link>
      <description>Couple of places you can check:&lt;BR /&gt;- Host Inspector in CM&lt;BR /&gt;- CM agent logs under /var/log/cloudera-scm-agent&lt;BR /&gt;&lt;BR /&gt;&lt;BR /&gt;</description>
      <pubDate>Wed, 17 Dec 2014 08:25:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22775#M4096</guid>
      <dc:creator>GautamG</dc:creator>
      <dc:date>2014-12-17T08:25:53Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22777#M4097</link>
      <description>&lt;P&gt;Thank you for the suggestions!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Host Inspector reports that everything is OK. I have checked the agent logs and also the log for the DataNode, but can not really find anything that gives any clues about this particular error.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is there a command for making the DataNode report which physical directories it is using and what it thinks their status is?&lt;/P&gt;</description>
      <pubDate>Wed, 17 Dec 2014 08:49:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22777#M4097</guid>
      <dc:creator>Knut N</dc:creator>
      <dc:date>2014-12-17T08:49:10Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22832#M4098</link>
      <description>&lt;P&gt;Not sure if this is helpful, but here are some charts of volume failures and disk capacity the faultly node, and for a healthy one as reference. The strange thing is that all disks seem to be in use on both nodes, but the total configured capacity is different.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/771i537E98846C1E72F5/image-size/original?v=mpbl-1&amp;amp;px=-1" title="node3.png" alt="node3.png" border="0" /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/774i39556C7FBDB76117/image-size/original?v=mpbl-1&amp;amp;px=-1" title="node2.png" alt="node2.png" border="0" /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Knut&lt;/P&gt;</description>
      <pubDate>Thu, 18 Dec 2014 10:44:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22832#M4098</guid>
      <dc:creator>Knut N</dc:creator>
      <dc:date>2014-12-18T10:44:25Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22897#M4099</link>
      <description>For a start, ensure your Cloudera Manager is also 5.2.1. We informally required that the versions be matched but it is now a requirement from 5.2 onwards. CM version should be equal or higher than CDH version.&lt;BR /&gt;&lt;BR /&gt;Next, go to Cloudera Management services -&amp;gt; even search -&amp;gt; critical. See if the volume related error is listed. If so, please paste details here.</description>
      <pubDate>Fri, 19 Dec 2014 00:49:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22897#M4099</guid>
      <dc:creator>GautamG</dc:creator>
      <dc:date>2014-12-19T00:49:22Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22906#M4100</link>
      <description>&lt;P&gt;I can confirm that CM is running version 5.2.1, so that should not be an issue.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When searching for critical log messages I came across these:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;It appears that another namenode 56559@&lt;EM&gt;datanode_hostname&lt;/EM&gt;&amp;nbsp;has already locked the storage directory&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;EM&gt;datanode_hostname&lt;/EM&gt;:50010:DataXceiver error processing WRITE_BLOCK operation src: /10.8.19.28:37029 dst: /10.8.19.14:50010&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have only ever had one namenode, and none on the datanode host, so I am a bit confused by what the first message really means. The second one I suspect might be a secondary failure of the first one?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards&lt;/P&gt;&lt;P&gt;Knut&lt;/P&gt;</description>
      <pubDate>Fri, 19 Dec 2014 07:53:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/22906#M4100</guid>
      <dc:creator>Knut N</dc:creator>
      <dc:date>2014-12-19T07:53:01Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/23240#M4101</link>
      <description>&lt;P&gt;The source of this error has been found. It turned out that /etc/fstab&amp;nbsp;on this node was badly configured, so that one of the disks was mounted twice as two separate data directories. Interestingly, this has not been causing any visible errors until upgrading to CDH 5.2.1. Nice that it was pointed out to us by this version though.&lt;/P&gt;</description>
      <pubDate>Fri, 02 Jan 2015 08:14:37 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/23240#M4101</guid>
      <dc:creator>Knut N</dc:creator>
      <dc:date>2015-01-02T08:14:37Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/23241#M4102</link>
      <description>Glad the issue is resolved, thanks for letting us know.</description>
      <pubDate>Fri, 02 Jan 2015 08:24:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/23241#M4102</guid>
      <dc:creator>GautamG</dc:creator>
      <dc:date>2015-01-02T08:24:29Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/40919#M4103</link>
      <description>&lt;P&gt;Hi,&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;can you tell us please how it was pointed out ? where to check if this is the case ?&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;regards&lt;/P&gt;</description>
      <pubDate>Tue, 17 May 2016 14:13:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/40919#M4103</guid>
      <dc:creator>zeeshan.khan</dc:creator>
      <dc:date>2016-05-17T14:13:38Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/48428#M4104</link>
      <description>&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;BR /&gt;&lt;P&gt;can you tell us please how it was pointed out ? where to check if this is the case ?&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here's one way of checking for duplicate mounts:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;mount |wc -l &amp;gt; a
mount |sort -u|uniq|wc -l &amp;gt; b
cmp a b&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If there's no output, there are no duplicate mounts.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;HTH&lt;/P&gt;</description>
      <pubDate>Sat, 10 Dec 2016 23:28:00 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/48428#M4104</guid>
      <dc:creator>bramd</dc:creator>
      <dc:date>2016-12-10T23:28:00Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/64532#M4105</link>
      <description>&lt;P&gt;In my case I did moved the existing dn directory as it has very less data and started the services.&lt;/P&gt;&lt;P&gt;There was no disk failure or issue with fstab file or mounting.&lt;/P&gt;</description>
      <pubDate>Mon, 12 Feb 2018 13:48:07 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/64532#M4105</guid>
      <dc:creator>SSH007</dc:creator>
      <dc:date>2018-02-12T13:48:07Z</dc:date>
    </item>
    <item>
      <title>Re: Volume failure reported while disks seem fine</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/270252#M4106</link>
      <description>&lt;P&gt;Yes even in my case fstab issue was not there. Rather i could clearly find the issue from datanode logs under /var/log/ and fix the issue. I have a blog on the same as below. Please comment on blog if it helps.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;A href="https://itdoctorharsha.wordpress.com/2019/08/23/hadoop-data-volume-failures-and-solution-cloudera/" target="_blank" rel="noopener"&gt;https://itdoctorharsha.wordpress.com/2019/08/23/hadoop-data-volume-failures-and-solution-cloudera/&lt;/A&gt;&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 11 Sep 2019 22:40:00 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Volume-failure-reported-while-disks-seem-fine/m-p/270252#M4106</guid>
      <dc:creator>Lambzee</dc:creator>
      <dc:date>2019-09-11T22:40:00Z</dc:date>
    </item>
  </channel>
</rss>

