<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: How to increase datanode capacity ? in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121598#M26544</link>
    <description>&lt;P&gt;do the following to increase the dfs size :&lt;/P&gt;&lt;P&gt;Create multiple directories or mount points in the hdfs data path : &lt;/P&gt;&lt;P&gt;by default ambari deployed cluster contain /hadoop/hdfs/data as the data directory , so &lt;/P&gt;&lt;P&gt;with root privileges : &lt;/P&gt;&lt;P&gt;create a directory &lt;/P&gt;&lt;P&gt;1) mkdir /hadoop/hdfs/data1 &lt;/P&gt;&lt;P&gt;2) chown -R hdfs:hadoop /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;3) chmod -R 777 /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;now edit the hdfs configuration :&lt;/P&gt;&lt;P&gt;1) on the cluster click on hdfs , click on configs , in the settings  add the directory separated by comma  under the hdfs.data.dir  property : &lt;/P&gt;&lt;P&gt;ex :  /hadoop/hdfs/data, /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;save the changes and restart the effected &lt;/P&gt;&lt;P&gt;That will increase the disk space , to increase further repeat the same  (or)&lt;/P&gt;&lt;P&gt;lvs resize  /hadoop/hdfs/data directory &lt;/P&gt;,&lt;P&gt;do the following to increase the dfs size : &lt;/P&gt;&lt;P&gt;Create multiple directories or mount points in the hdfs data path : &lt;/P&gt;&lt;P&gt;by default ambari deployed cluster contain /hadoop/hdfs/data as the data directory , so &lt;/P&gt;&lt;P&gt;with root privileges : &lt;/P&gt;&lt;P&gt;create a directory &lt;/P&gt;&lt;P&gt;1) mkdir /hadoop/hdfs/data1 &lt;/P&gt;&lt;P&gt;2) chown -R hdfs:hadoop /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;3) chmod -R 777 /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;now edit the hdfs configuration :&lt;/P&gt;&lt;P&gt;1) on the cluster click on hdfs , click on configs , in the settings  add the directory separated by comma  under the hdfs.data.dir  property : &lt;/P&gt;&lt;P&gt;ex :  /hadoop/hdfs/data, /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;save the changes and restart the effected &lt;/P&gt;&lt;P&gt;That will increase the disk space , to increase further repeat the same  (or)&lt;/P&gt;&lt;P&gt;lvs resize  /hadoop/hdfs/data directory &lt;/P&gt;</description>
    <pubDate>Tue, 04 Oct 2016 01:32:38 GMT</pubDate>
    <dc:creator>nagaiik</dc:creator>
    <dc:date>2016-10-04T01:32:38Z</dc:date>
    <item>
      <title>How to increase datanode capacity ?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121594#M26540</link>
      <description>&lt;P&gt;I am using Ambari and it shows that my data node capacity is only &lt;/P&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;
991.83 MB&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt; and has 283 blocks. (Surprisingly),  Even if it is the default, why is it as low as 991 MB? &lt;/P&gt;&lt;P&gt;I hear that having too many blocks isn't such a good idea. I do not really have space constraints on the machine I am on and we are not planning to have datanode distributed across multiple hosts. &lt;/P&gt;&lt;P&gt;My question is:&lt;/P&gt;&lt;P&gt;1. Is there a maximum limit to size of a datanode? If yes, what is it?&lt;/P&gt;&lt;P&gt;2. What is the easiest and robust way to have multiple datanodes on the same machine without breaking what is up and running in the  existing cluster?&lt;/P&gt;&lt;P&gt;3. I understand that we need to add more directories for new data nodes and specify the path in ambari but what next? &lt;/P&gt;&lt;P&gt;4. what is the optimum block size in ambari? (or if there is some datanode/block size ration for the optimized number?)&lt;/P&gt;&lt;P&gt;5. How to configure the block size through ambari?&lt;/P&gt;&lt;P&gt;6. How to increase size of an existing datanode in ambari?&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 13:11:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121594#M26540</guid>
      <dc:creator>sim6</dc:creator>
      <dc:date>2016-04-29T13:11:56Z</dc:date>
    </item>
    <item>
      <title>Re: How to increase datanode capacity ?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121595#M26541</link>
      <description>&lt;P&gt;HI &lt;A rel="user" href="https://community.cloudera.com/users/6243/simrankauradept.html" nodeid="6243"&gt;@simran kaur&lt;/A&gt;, to answer your questions&lt;/P&gt;&lt;OL&gt;&lt;LI&gt;There is no limit to "size", or capacity of the DN. It's only bound by the number of hard disk slots and capacity of your individual disks. If you have 12 slots and 6T per disk, then it's 72T per node.&lt;/LI&gt;&lt;LI&gt;Datanode is a process managing HDFS files on a machine. You use only 1 DN on the same machine.&lt;/LI&gt;&lt;LI&gt;You specify your DN directories, typically mounting points of your disks in dfs.datanode.data.dir. That's all, HDFS will take care of organizing data there.&lt;/LI&gt;&lt;LI&gt;You configure block size as the dfs.blocksize property in HDFS. The default is 134217728 or 128M.&lt;/LI&gt;&lt;LI&gt;The default of 128M is considered an optimal size for general-purpose clusters. If you keep many large files it can be increased, for example to 256M. &lt;/LI&gt;&lt;LI&gt;And finally your DN capacity of only 991M indicates that something is wrong or you are running a Sandbox on a machine with small capacity. My capacity on my Sandbox is 45G.&lt;/LI&gt;&lt;/OL&gt;</description>
      <pubDate>Fri, 29 Apr 2016 13:35:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121595#M26541</guid>
      <dc:creator>pminovic</dc:creator>
      <dc:date>2016-04-29T13:35:22Z</dc:date>
    </item>
    <item>
      <title>Re: How to increase datanode capacity ?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121596#M26542</link>
      <description>&lt;P&gt;Thank you for your response &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt; That helped. No, I am not running through a sandbox and have installed hdp on a centos machine. Could you please tell what could be the possible reasons for DN capacity to be so low?&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 13:54:04 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121596#M26542</guid>
      <dc:creator>sim6</dc:creator>
      <dc:date>2016-04-29T13:54:04Z</dc:date>
    </item>
    <item>
      <title>Re: How to increase datanode capacity ?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121597#M26543</link>
      <description>&lt;P&gt;Can you check your dfs.datanode.data.dir setting, and confirm that the directories listed there correspond to your disk mounting points. The setting applies to all Data nodes in the cluster, all of them must have the same disk mounting configuration.&lt;/P&gt;</description>
      <pubDate>Fri, 29 Apr 2016 14:46:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121597#M26543</guid>
      <dc:creator>pminovic</dc:creator>
      <dc:date>2016-04-29T14:46:55Z</dc:date>
    </item>
    <item>
      <title>Re: How to increase datanode capacity ?</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121598#M26544</link>
      <description>&lt;P&gt;do the following to increase the dfs size :&lt;/P&gt;&lt;P&gt;Create multiple directories or mount points in the hdfs data path : &lt;/P&gt;&lt;P&gt;by default ambari deployed cluster contain /hadoop/hdfs/data as the data directory , so &lt;/P&gt;&lt;P&gt;with root privileges : &lt;/P&gt;&lt;P&gt;create a directory &lt;/P&gt;&lt;P&gt;1) mkdir /hadoop/hdfs/data1 &lt;/P&gt;&lt;P&gt;2) chown -R hdfs:hadoop /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;3) chmod -R 777 /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;now edit the hdfs configuration :&lt;/P&gt;&lt;P&gt;1) on the cluster click on hdfs , click on configs , in the settings  add the directory separated by comma  under the hdfs.data.dir  property : &lt;/P&gt;&lt;P&gt;ex :  /hadoop/hdfs/data, /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;save the changes and restart the effected &lt;/P&gt;&lt;P&gt;That will increase the disk space , to increase further repeat the same  (or)&lt;/P&gt;&lt;P&gt;lvs resize  /hadoop/hdfs/data directory &lt;/P&gt;,&lt;P&gt;do the following to increase the dfs size : &lt;/P&gt;&lt;P&gt;Create multiple directories or mount points in the hdfs data path : &lt;/P&gt;&lt;P&gt;by default ambari deployed cluster contain /hadoop/hdfs/data as the data directory , so &lt;/P&gt;&lt;P&gt;with root privileges : &lt;/P&gt;&lt;P&gt;create a directory &lt;/P&gt;&lt;P&gt;1) mkdir /hadoop/hdfs/data1 &lt;/P&gt;&lt;P&gt;2) chown -R hdfs:hadoop /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;3) chmod -R 777 /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;now edit the hdfs configuration :&lt;/P&gt;&lt;P&gt;1) on the cluster click on hdfs , click on configs , in the settings  add the directory separated by comma  under the hdfs.data.dir  property : &lt;/P&gt;&lt;P&gt;ex :  /hadoop/hdfs/data, /hadoop/hdfs/data1&lt;/P&gt;&lt;P&gt;save the changes and restart the effected &lt;/P&gt;&lt;P&gt;That will increase the disk space , to increase further repeat the same  (or)&lt;/P&gt;&lt;P&gt;lvs resize  /hadoop/hdfs/data directory &lt;/P&gt;</description>
      <pubDate>Tue, 04 Oct 2016 01:32:38 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/How-to-increase-datanode-capacity/m-p/121598#M26544</guid>
      <dc:creator>nagaiik</dc:creator>
      <dc:date>2016-10-04T01:32:38Z</dc:date>
    </item>
  </channel>
</rss>

