<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Non-DFS storage occupied in Hadoop mount in Linux server in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220714#M182599</link>
    <description>&lt;P&gt;Thank you very much &lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158"&gt;@Vinicius Higa Murakami&lt;/A&gt;&lt;/P&gt;</description>
    <pubDate>Tue, 26 Jun 2018 12:39:23 GMT</pubDate>
    <dc:creator>karthik_chandra</dc:creator>
    <dc:date>2018-06-26T12:39:23Z</dc:date>
    <item>
      <title>Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220700#M182585</link>
      <description>&lt;P&gt;I am working in a 16 node cluster and recently i received an issue with the Non-DFS storage where the /hadoop/ mount which I am using is being consumed for non-DFS data and when I observed i found lot of blk_12345 etc. named .meta and original files. where each file is of 128 MB size and .meta file of size 1.1 MB (Totally all the files in total are consuming 1.7 TB of total cluster storage). Please let me know if I can remove these files and what is the impact if I remove them.&lt;/P&gt;&lt;P&gt;And what is the reason they are created?&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 13:20:13 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220700#M182585</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2022-09-16T13:20:13Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220701#M182586</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/81338/karthikchandrashekhar.html" nodeid="81338"&gt;@Karthik Chandrashekhar&lt;/A&gt;!&lt;/P&gt;&lt;P&gt;I'm not sure if i get you right, but my advice would be to not delete these files. It belongs to HDFS Datanode, the blk_12345 dir carries some blocks+meta = data stored in HDFS. &lt;BR /&gt;If you want to know which file belongs to which block, you can use the following command:&lt;BR /&gt;&lt;/P&gt;&lt;PRE&gt;[hdfs@node2 ~]$ cd /hadoop/hdfs/data/current/BP-686380642-172.25.33.129-1527546468579/current/finalized/subdir0/subdir0/
[hdfs@node2 subdir0]$ ls | head -2
blk_1073741825
blk_1073741825_1001.meta
[hdfs@node2 ~]$ hdfs fsck / -files -locations -blocks -blockId blk_1073741825
Connecting to namenode via &lt;A href="http://node3:50070/fsck?ugi=hdfs&amp;amp;files=1&amp;amp;locations=1&amp;amp;blocks=1&amp;amp;blockId=blk_1073741825+&amp;amp;path=%2F" target="_blank"&gt;http://node3:50070/fsck?ugi=hdfs&amp;amp;files=1&amp;amp;locations=1&amp;amp;blocks=1&amp;amp;blockId=blk_1073741825+&amp;amp;path=%2F&lt;/A&gt;
FSCK started by hdfs (auth:SIMPLE) from /MYIP at Tue Jun 12 14:54:08 UTC 2018

Block Id: blk_1073741825
Block belongs to: /hdp/apps/2.6.4.0-91/mapreduce/mapreduce.tar.gz
No. of Expected Replica: 3
No. of live Replica: 3
No. of excess Replica: 0
No. of stale Replica: 0
No. of decommissioned Replica: 0
No. of decommissioning Replica: 0
No. of corrupted Replica: 0
Block replica on datanode/rack: node2/default-rack is HEALTHY
Block replica on datanode/rack: node3/default-rack is HEALTHY
Block replica on datanode/rack: node4/default-rack is HEALTHY
&lt;/PRE&gt;Hope this helps! &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;</description>
      <pubDate>Wed, 13 Jun 2018 00:21:30 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220701#M182586</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2018-06-13T00:21:30Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220702#M182587</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158"&gt;@Vinicius Higa Murakami&lt;/A&gt;&lt;P&gt; Is there a way to move the blocks from non-DFS mount /hadoop to DFS mounts like /data and later remove the files if they are yet there on /hadoop mount.&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jun 2018 13:51:11 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220702#M182587</guid>
      <dc:creator>bhanu_pamu</dc:creator>
      <dc:date>2018-06-13T13:51:11Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220703#M182588</link>
      <description>&lt;P&gt;Hi &lt;A href="https://community.hortonworks.com/users/79158/vmurakami.html"&gt;@Vinicius Higa Murakami&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Thank you very much for the reply and the command to see where the block is being used.&lt;/P&gt;&lt;P&gt;But what i don't understand is the file is already present in DFS and occupying specific amount of space.&lt;/P&gt;&lt;P&gt;Why for the same file there is block on non-DFS also?&lt;/P&gt;&lt;P&gt;Is it a kind of backup?&lt;/P&gt;</description>
      <pubDate>Wed, 13 Jun 2018 14:09:02 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220703#M182588</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2018-06-13T14:09:02Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220704#M182589</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/81338/karthikchandrashekhar.html" nodeid="81338"&gt;@Karthik Chandrashekhar&lt;/A&gt;!&lt;BR /&gt;Hm, but these blocks belongs to your  dfs.datanode.data.dir parameter? If so, they should belong to DFS not NON-DFS.&lt;BR /&gt;Cause AFAIK, any data outside of hdfs and written in the same mount disk as dfs.datanode.data.dir path is considered as non-DFS. &lt;BR /&gt;If these blocks doesn't belong to your DFS (NON-DFS) and they're in the same path as your dfs.datanode.data.dir value. Then, we might have an issue there &lt;span class="lia-unicode-emoji" title=":disappointed_face:"&gt;😞&lt;/span&gt; &lt;/P&gt;&lt;P&gt;Btw, could you check your mount points as well?&lt;/P&gt;&lt;P&gt;Hope this helps! &lt;/P&gt;</description>
      <pubDate>Wed, 13 Jun 2018 23:43:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220704#M182589</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2018-06-13T23:43:29Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220705#M182590</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/79163/bhanupamu.html" nodeid="79163"&gt;@Bhanu Pamu&lt;/A&gt;!&lt;BR /&gt;I'm not sure if I get it, but if you have /hadoop (non-dfs + dfs files) and do you wanna move them to /data, guess the best choice would be to add this /data to dfs.datanode.data.dir as well. Then stop the datanodes and move the files from /hadoop to /data. Not sure if this is the best practice or if has another approach to do this, but certainly i'd investigate more on this thought before doing anything under HDFS. &lt;/P&gt;&lt;P&gt;Hope this helps! &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; &lt;/P&gt;</description>
      <pubDate>Thu, 14 Jun 2018 01:06:54 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220705#M182590</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2018-06-14T01:06:54Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220706#M182591</link>
      <description>&lt;P&gt;Hi @&lt;A href="https://community.hortonworks.com/users/79158/vmurakami.html" rel="nofollow noopener noreferrer" target="_blank"&gt;Vinicius Higa Murakami&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;Thanks again.&lt;/P&gt;&lt;P&gt;Please find few of the screenshots below from my system and let me know if the way i am analyzing is right.&lt;/P&gt;&lt;P&gt;DFS Storage:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="78418-dfs-storage.png" style="width: 470px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15844i07AC7DC6E1C107BB/image-size/medium?v=v2&amp;amp;px=400" role="button" title="78418-dfs-storage.png" alt="78418-dfs-storage.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Storage description in Ambari:&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="78419-ambari.png" style="width: 1134px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15845iC92388AFA4341C6F/image-size/medium?v=v2&amp;amp;px=400" role="button" title="78419-ambari.png" alt="78419-ambari.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Non-DFS Storage: Here is where i have the blk_12345 files.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="78420-datanode.png" style="width: 465px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15846i67E60CBDB3798436/image-size/medium?v=v2&amp;amp;px=400" role="button" title="78420-datanode.png" alt="78420-datanode.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="78423-nondfs-size.png" style="width: 1043px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15847iC5988B844FDA64BD/image-size/medium?v=v2&amp;amp;px=400" role="button" title="78423-nondfs-size.png" alt="78423-nondfs-size.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 02:23:58 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220706#M182591</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2019-08-18T02:23:58Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220707#M182592</link>
      <description>&lt;P&gt;Hi @&lt;A href="https://community.hortonworks.com/users/79158/vmurakami.html" rel="nofollow noopener noreferrer" target="_blank"&gt;Vinicius Higa Murakami&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;Please also find the Disk space utilization in my system.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="78426-disk-util-hadoop.png" style="width: 1010px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15843iED7F9B93FE41A5D4/image-size/medium?v=v2&amp;amp;px=400" role="button" title="78426-disk-util-hadoop.png" alt="78426-disk-util-hadoop.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 02:23:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220707#M182592</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2019-08-18T02:23:34Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220708#M182593</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/81338/karthikchandrashekhar.html" nodeid="81338"&gt;@Karthik Chandrashekhar&lt;/A&gt;! &lt;BR /&gt;Sorry for my delay, so basically I couldn't note anything wrong in your configs. &lt;BR /&gt;But one thing, on your disk-util-hadoop,png. Could you check if under the /hadoop mount is there any other subdir (besides the /hadoop/hadoop/hdfs/data)?&lt;/P&gt;&lt;PRE&gt;du --max-depth=1 -h /hadoop/hadoop/&lt;BR /&gt;or&lt;BR /&gt;du --max-depth=1 -h /hadoop/
#And just to check the mountpoints&lt;BR /&gt;lsblk
&lt;/PRE&gt;&lt;P&gt; I think your Non-dfs usage is high, because HDFS is counting other directories under the /hadoop at the same disk.&lt;BR /&gt;And one last thing, how much is set for dfs.datanode.du.reserved? &lt;BR /&gt;Hope this helps!&lt;/P&gt;</description>
      <pubDate>Sat, 16 Jun 2018 01:04:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220708#M182593</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2018-06-16T01:04:49Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220709#M182594</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158"&gt;@Vinicius Higa Murakami&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;Please find below the value of dfs.datanode.du.reserved.&lt;/P&gt;&lt;P&gt;    &amp;lt;property&amp;gt;
      &amp;lt;name&amp;gt;dfs.datanode.du.reserved&amp;lt;/name&amp;gt;
      &amp;lt;value&amp;gt;48204034048&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;&lt;/P&gt;&lt;P&gt;And regarding your question on subdir.&lt;/P&gt;&lt;P&gt;Yes, i have a folder called subdir1 which is filled with this enormous blocks data.&lt;/P&gt;&lt;P&gt;It is available in below path.&lt;/P&gt;&lt;P&gt;/hadoop/hadoop/hdfs/data/current/BP-1468174578-IP_Address-1522297380740/current/finalized&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Mountpoints: lsblk&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;IMG src="https://community.cloudera.com/legacyfs/online/attachments/77734-lsblk.png" /&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 19 Jun 2018 12:43:07 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220709#M182594</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2018-06-19T12:43:07Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220710#M182595</link>
      <description>&lt;P&gt;Hey &lt;A rel="user" href="https://community.cloudera.com/users/81338/karthikchandrashekhar.html" nodeid="81338"&gt;@Karthik Chandrashekhar&lt;/A&gt;! &lt;BR /&gt;Unfortunately, i'm not able to see the content of your attach, could you upload it again, please? &lt;/P&gt;&lt;P&gt;And regarding, your subdir, could you share the output from the following command:&lt;/P&gt;&lt;PRE&gt;du --max-depth=1 -h /hadoop/&lt;BR /&gt;&lt;/PRE&gt;&lt;P&gt;Hope this helps!&lt;/P&gt;</description>
      <pubDate>Wed, 20 Jun 2018 03:01:58 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220710#M182595</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2018-06-20T03:01:58Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220711#M182596</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158" target="_blank"&gt;@Vinicius Higa Murakami&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;Please find the attachment again.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="77768-lsblk.png" style="width: 937px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15842i56D061FA8DA50EB0/image-size/medium?v=v2&amp;amp;px=400" role="button" title="77768-lsblk.png" alt="77768-lsblk.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 02:23:26 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220711#M182596</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2019-08-18T02:23:26Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220712#M182597</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158"&gt;@Vinicius Higa Murakami&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;Can you please let me know if I can delete the subdirNNN files under finalized folder?&lt;/P&gt;&lt;P&gt;And how to permanently stop files being stored in finalized folder?&lt;/P&gt;</description>
      <pubDate>Thu, 21 Jun 2018 11:33:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220712#M182597</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2018-06-21T11:33:21Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220713#M182598</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/81338/karthikchandrashekhar.html" nodeid="81338" target="_blank"&gt;@Karthik Chandrashekhar&lt;/A&gt;! &lt;BR /&gt;Sorry about my delay, so taking a look at your du outputs, it looks like HDFS is doing okay with the DFS total size. &lt;BR /&gt;If you sum the values from 16 hosts under the /hadoop/hadoop/hdfs/data it will be equal to 1.7TB. &lt;BR /&gt;Do you have a specific mount disk for /hadoop/hadoop/hdfs/data? Or it's all under the / directory, and how many disks do you have?&lt;BR /&gt;E.g., in my case, I have a lab and its everything under the / directory in 1 disk.&lt;BR /&gt;&lt;/P&gt;&lt;PRE&gt;[root@c1123-node3 hadoop]# df -h
Filesystem            Size  Used Avail Use% Mounted on
rootfs                1.2T  731G  423G  64% /
overlay               1.2T  731G  423G  64% /
tmpfs                 126G     0  126G   0% /dev
tmpfs                 126G     0  126G   0% /sys/fs/cgroup
/dev/mapper/vg01-vsr_lib_docker
                      1.2T  731G  423G  64% /etc/resolv.conf
/dev/mapper/vg01-vsr_lib_docker
                      1.2T  731G  423G  64% /etc/hostname
/dev/mapper/vg01-vsr_lib_docker
                      1.2T  731G  423G  64% /etc/hosts
shm                    64M   12K   64M   1% /dev/shm
overlay               1.2T  731G  423G  64% /proc/meminfo&lt;/PRE&gt;&lt;P&gt;If I calculate on all hosts the du --max-depth=1-h /hadoop/hdfs/data, I'll get my DFS Usage. &lt;BR /&gt;And if I calculate on all hosts my du --max-depth=1-h / minus the value from HDFS directory, I'll get the total of Non-dfs usage. &lt;BR /&gt;So the math would be:&lt;BR /&gt;DFS Usage = Total DU on the HDFS Path &lt;BR /&gt;NON-DFS Usage = Total DU - (DFS Usage) &lt;BR /&gt;For each disk.&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="77793-screen-shot-2018-06-21-at-145126.png" style="width: 1203px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/15841i264956293F21DB9B/image-size/medium?v=v2&amp;amp;px=400" role="button" title="77793-screen-shot-2018-06-21-at-145126.png" alt="77793-screen-shot-2018-06-21-at-145126.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;And answering your last question, the finalized folder it's used by HDFS to allocated the blocks that have been processed. So deleting these files, it'll probably throw some alerts from HDFS to you (maybe a block missing a replica or some corrupted block).&lt;BR /&gt;I completely understand your concern about your storage getting almost full, but, If you aren't able to delete any data outside of the HDFS, I'd try to delete old and unused files from HDFS (using the HDFS DFS command!), compress any raw data, use more file formats with compression enabled or in last case change the replication-factor to a lower value (kindly remember that changing this, it may cause some problems). Just a friendly reminder, everything under the dfs.datanode.data.dir will be used internally for HDFS storing purposes &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; &lt;BR /&gt;&lt;BR /&gt;Hope this helps! &lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 02:23:18 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220713#M182598</guid>
      <dc:creator>vmurakami</dc:creator>
      <dc:date>2019-08-18T02:23:18Z</dc:date>
    </item>
    <item>
      <title>Re: Non-DFS storage occupied in Hadoop mount in Linux server</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220714#M182599</link>
      <description>&lt;P&gt;Thank you very much &lt;A rel="user" href="https://community.cloudera.com/users/79158/vmurakami.html" nodeid="79158"&gt;@Vinicius Higa Murakami&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 26 Jun 2018 12:39:23 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Non-DFS-storage-occupied-in-Hadoop-mount-in-Linux-server/m-p/220714#M182599</guid>
      <dc:creator>karthik_chandra</dc:creator>
      <dc:date>2018-06-26T12:39:23Z</dc:date>
    </item>
  </channel>
</rss>

