<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark SQL Job stcuk  indefinitely at last task of a stage -- Shows  INFO: BlockManagerInfo : Removed broadcast in memory in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark-SQL-Job-stcuk-indefinitely-at-last-task-of-a-stage/m-p/132544#M95214</link>
    <description>&lt;P&gt;Hi Puneet --as per suggestion I tried with&lt;/P&gt;&lt;P&gt;--driver-memory 4g --num-executors 15 --total-executor-cores 30  --executor-memory 10g  --driver-cores 2&lt;/P&gt;and it failed with &lt;P&gt;Exception in thread "dispatcher-event-loop-3" java.lang.OutOfMemoryError: Java heap space.&lt;/P&gt;&lt;P&gt;What I am suspecting is parttioning pushing huge data on on one or more executors, and it failes....I saw in spark job environment and &lt;TABLE&gt;
&lt;TBODY&gt;&lt;TR&gt;
&lt;TD&gt;spark.yarn.driver.memoryOverhead = 384&lt;/TD&gt;

&lt;/TR&gt;&lt;TR&gt;
&lt;TD&gt;spark.yarn.executor.memoryOverhead = 384&lt;/TD&gt;
&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;/P&gt;whihc is very low.......i refered documentation and its says &lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;CODE&gt;spark.yarn.executor.memoryOverhead&lt;/CODE&gt;&lt;/TD&gt;&lt;TD&gt;executorMemory * 0.10, with minimum of 384 &lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt; How we can set it to 1G or more</description>
    <pubDate>Tue, 19 Jul 2016 16:48:30 GMT</pubDate>
    <dc:creator>pkhare</dc:creator>
    <dc:date>2016-07-19T16:48:30Z</dc:date>
  </channel>
</rss>

