<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark SQL Job stcuk  indefinitely at last task of a stage -- Shows  INFO: BlockManagerInfo : Removed broadcast in memory in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark-SQL-Job-stcuk-indefinitely-at-last-task-of-a-stage/m-p/132540#M95210</link>
    <description>&lt;P&gt;Hi Praveen,&lt;/P&gt;&lt;P&gt;Here are a few points to help:&lt;/P&gt;&lt;P&gt;1. Try running your API without options like "--driver-memory 15g --num-executors 25 --total-executor-cores 60 --executor-memory 15g --driver-cores 2" and check logs for memory allocated to RDDs/DataFrames.&lt;/P&gt;&lt;P&gt;2. Driver doesn't need 15g memory if you are not collecting data on driver. Try setting it to 4g rather. I hope u r not using .collect() or similar operations which collect all data to driver.&lt;/P&gt;&lt;P&gt;3. The error needs fine tuning your configurations between executor memory and driver memory. The total number of executors(25) are pretty much higher considering the memory allocated(15g). Reduce number of executors and consider allocating less memory(4g to start with).&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;Puneet&lt;/P&gt;</description>
    <pubDate>Mon, 18 Jul 2016 15:09:49 GMT</pubDate>
    <dc:creator>psingh15</dc:creator>
    <dc:date>2016-07-18T15:09:49Z</dc:date>
  </channel>
</rss>

