<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: MapReduce application failed with OutOfMemoryError in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/MapReduce-application-failed-with-OutOfMemoryError/m-p/55888#M36632</link>
    <description>&lt;P&gt;is it fail at the Map phase or the reducer? MR jobs running with 6 GB for map and 12 G for reducer and fail, need a review.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;How much data the MR is running on? how many MR jobs do you have? 18K mappers also indicate that you have a lot of small files in your HDFS.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;your cluster consumption of ~300Vcores and 3T memory is meaning that each Vcore is running with an average of 10 G memory, how much nodes you have? from your stat seems your jobs are memory intenstive rather than vcores which this can reflects mainly on Spark jobs and you can manage this to get the optimal Vcores and memory for each Spark job.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;i suspect that you have a specific crazy job and prefer you if you are using Cloudera manager to search it in the application, or try to run the jobs under pools and then identify the problmatic ones.&lt;/P&gt;</description>
    <pubDate>Wed, 14 Jun 2017 22:20:55 GMT</pubDate>
    <dc:creator>Fawze</dc:creator>
    <dc:date>2017-06-14T22:20:55Z</dc:date>
  </channel>
</rss>

