<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Yarn memory utilization. in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162123#M124502</link>
    <description>&lt;P&gt;Please find my cluster details in my first mail, also I am
using have 2 disk per node.&lt;/P&gt;&lt;P&gt;Document says the recommended configuration as below, that’s
the same I have did in my newer configuration.&lt;/P&gt;&lt;P&gt;yarn.scheduler.minimum-allocation-mb=1024&lt;/P&gt;&lt;P&gt; yarn.scheduler.maximum-allocation-mb=4096&lt;/P&gt;&lt;P&gt; yarn.nodemanager.resource.memory-mb=4096&lt;/P&gt;&lt;P&gt; mapreduce.map.memory.mb=512&lt;/P&gt;&lt;P&gt; mapreduce.map.java.opts=-Xmx409m&lt;/P&gt;&lt;P&gt; mapreduce.reduce.memory.mb=1024&lt;/P&gt;&lt;P&gt; mapreduce.reduce.java.opts=-Xmx819m&lt;/P&gt;&lt;P&gt; yarn.app.mapreduce.am.resource.mb=512&lt;/P&gt;&lt;P&gt; yarn.app.mapreduce.am.command-opts=-Xmx409m&lt;/P&gt;&lt;P&gt; mapreduce.task.io.sort.mb=204&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
    <pubDate>Wed, 03 Aug 2016 13:34:39 GMT</pubDate>
    <dc:creator>arunkumar_d</dc:creator>
    <dc:date>2016-08-03T13:34:39Z</dc:date>
    <item>
      <title>Yarn memory utilization.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162121#M124500</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;I am running a cluster with 15 data node, 15 region server and
16 node manager (of course name node, Secondary name node, Hactive master, Resource
manager). All the machines are m3.large type machine basically so, 2 core processor
and 7.5GB of RAM. &lt;/P&gt;&lt;P&gt;By default it allocates 32GB for the yarn memory and 1vcore.
Here my default configuration and it uses DefaultResourceCalculator. &lt;/P&gt;&lt;P&gt;yarn.scheduler.minimum-allocation-mb: 682&lt;/P&gt;&lt;P&gt;yarn.scheduler.maximum-allocation-mb: 2048&lt;/P&gt;&lt;P&gt;yarn.nodemanager.resource.cpu-vcores :
1&lt;/P&gt;&lt;P&gt;yarn.nodemanager.resource.memory-mb: 2048&lt;/P&gt;&lt;P&gt;when I run a mapreduce job it takes about some 30min to
complete it till the time the yarn memory utilization was high, I thought that
the yarn memory was the issue. So I have doubled the size as below. &lt;/P&gt;&lt;P&gt;yarn.scheduler.minimum-allocation-mb: 1024&lt;/P&gt;&lt;P&gt;yarn.scheduler.maximum-allocation-mb: 4096&lt;/P&gt;&lt;P&gt;yarn.nodemanager.resource.cpu-vcores :
1&lt;/P&gt;&lt;P&gt;yarn.nodemanager.resource.memory-mb: 4096 &lt;/P&gt;&lt;P&gt;Now, yarn  memory increased
from 32Gb to 64GB, but when I run a same mapreduce job with newer configuration
it takes me around 42 min though yarn memory all the 64GB the cluster seems
slower than before. So, I would like to understand the containers resource
allocation and why it’s slow down after I increased the memory also I would
like to see how many containers per cluster and per node (any calculation). Please
suggest me with the recommended configuration in this case. &lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Arun&lt;/P&gt;</description>
      <pubDate>Wed, 03 Aug 2016 13:02:52 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162121#M124500</guid>
      <dc:creator>arunkumar_d</dc:creator>
      <dc:date>2016-08-03T13:02:52Z</dc:date>
    </item>
    <item>
      <title>Re: Yarn memory utilization.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162122#M124501</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/1450/arunkumard.html" nodeid="1450"&gt;@Arunkumar Dhanakumar&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Here is the link for calculating YARN memory:&lt;/P&gt;&lt;P&gt;&lt;A href="https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.0/bk_installing_manually_book/content/determine-hdp-memory-config.html" target="_blank"&gt;https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.3.0/bk_installing_manually_book/content/determine-hdp-memory-config.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;1) How many data nodes do you have?&lt;/P&gt;&lt;P&gt;2) How many disks do you have in each data node?&lt;/P&gt;&lt;P&gt;3) Did you install HBase?&lt;/P&gt;&lt;P&gt;4) How many Cores do you have on each data node?&lt;/P&gt;&lt;P&gt;5) RAM size on each data node?&lt;/P&gt;</description>
      <pubDate>Wed, 03 Aug 2016 13:22:04 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162122#M124501</guid>
      <dc:creator>divakarreddy_a</dc:creator>
      <dc:date>2016-08-03T13:22:04Z</dc:date>
    </item>
    <item>
      <title>Re: Yarn memory utilization.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162123#M124502</link>
      <description>&lt;P&gt;Please find my cluster details in my first mail, also I am
using have 2 disk per node.&lt;/P&gt;&lt;P&gt;Document says the recommended configuration as below, that’s
the same I have did in my newer configuration.&lt;/P&gt;&lt;P&gt;yarn.scheduler.minimum-allocation-mb=1024&lt;/P&gt;&lt;P&gt; yarn.scheduler.maximum-allocation-mb=4096&lt;/P&gt;&lt;P&gt; yarn.nodemanager.resource.memory-mb=4096&lt;/P&gt;&lt;P&gt; mapreduce.map.memory.mb=512&lt;/P&gt;&lt;P&gt; mapreduce.map.java.opts=-Xmx409m&lt;/P&gt;&lt;P&gt; mapreduce.reduce.memory.mb=1024&lt;/P&gt;&lt;P&gt; mapreduce.reduce.java.opts=-Xmx819m&lt;/P&gt;&lt;P&gt; yarn.app.mapreduce.am.resource.mb=512&lt;/P&gt;&lt;P&gt; yarn.app.mapreduce.am.command-opts=-Xmx409m&lt;/P&gt;&lt;P&gt; mapreduce.task.io.sort.mb=204&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Wed, 03 Aug 2016 13:34:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162123#M124502</guid>
      <dc:creator>arunkumar_d</dc:creator>
      <dc:date>2016-08-03T13:34:39Z</dc:date>
    </item>
    <item>
      <title>Re: Yarn memory utilization.</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162124#M124503</link>
      <description>&lt;P&gt;Basically you increased your YARN memory from 32Gb to 64GB, it means you increased all containers memory. Container is a unit for YARN submitting the jobs in-terms of CPU and RAM.&lt;/P&gt;&lt;P&gt;you increased YARN container size then what about Tez container size?&lt;/P&gt;&lt;P&gt;--&amp;gt; ideally tez container size should be multiple of YARN Memory.&lt;/P&gt;&lt;P&gt;--&amp;gt; ideally we can allocate two containers per disk and per CPU.&lt;/P&gt;</description>
      <pubDate>Wed, 03 Aug 2016 13:37:33 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Yarn-memory-utilization/m-p/162124#M124503</guid>
      <dc:creator>divakarreddy_a</dc:creator>
      <dc:date>2016-08-03T13:37:33Z</dc:date>
    </item>
  </channel>
</rss>

