<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Hadoop not utilizing available memory in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30367#M6864</link>
    <description>&lt;P&gt;The yarn.nodemanager.resource.memory-mb parameter was set to 8GB which was causing the problem.&lt;/P&gt;&lt;P&gt;I have reset this to a higher value , which helped me resolve this behavior.&lt;/P&gt;</description>
    <pubDate>Sun, 02 Aug 2015 10:40:32 GMT</pubDate>
    <dc:creator>Bahubali Jain</dc:creator>
    <dc:date>2015-08-02T10:40:32Z</dc:date>
    <item>
      <title>Hadoop not utilizing available memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30366#M6863</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;I have a 2 node cluster (each node having 32 GB RAM and 8 Cores). I have installed CDH 5.4.&lt;/P&gt;&lt;P&gt;But I dont think the available memory is being utilized by hadoop, as the page :8088/cluster/apps shows up only 16 GB in "Memory Total" column.&lt;/P&gt;&lt;P&gt;Only once I could see that the "Memory Total " as 64GB, not sure whats going on.&lt;/P&gt;&lt;P&gt;What could be the reason?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;Baahu&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:36:36 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30366#M6863</guid>
      <dc:creator>Bahubali Jain</dc:creator>
      <dc:date>2022-09-16T09:36:36Z</dc:date>
    </item>
    <item>
      <title>Re: Hadoop not utilizing available memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30367#M6864</link>
      <description>&lt;P&gt;The yarn.nodemanager.resource.memory-mb parameter was set to 8GB which was causing the problem.&lt;/P&gt;&lt;P&gt;I have reset this to a higher value , which helped me resolve this behavior.&lt;/P&gt;</description>
      <pubDate>Sun, 02 Aug 2015 10:40:32 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30367#M6864</guid>
      <dc:creator>Bahubali Jain</dc:creator>
      <dc:date>2015-08-02T10:40:32Z</dc:date>
    </item>
    <item>
      <title>Re: Hadoop not utilizing available memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30369#M6865</link>
      <description>&lt;P&gt;Congratulations on solving your issue. Thank you for sharing the solution as it may also help others. &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 02 Aug 2015 12:57:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Hadoop-not-utilizing-available-memory/m-p/30369#M6865</guid>
      <dc:creator>cjervis</dc:creator>
      <dc:date>2015-08-02T12:57:09Z</dc:date>
    </item>
  </channel>
</rss>

