<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark job getting failed with Jupyter notebook in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/93785#M57368</link>
    <description>&lt;P&gt;Hi, the probable root cause is that the spark job submitted by the Jupyter notebook has a different memory config parameters. So I dont think the issue is Jupyter, but rather the executor and driver memory settings. Yarn is not able to provide enough resources (i.e. memory)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;19/08/06 23:10:41 WARN cluster.YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Check your cluster settings:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;- how much memory YARN has allocated in NodeManagers, how big the container could be&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;- what are the submit options of your spark job&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Wed, 07 Aug 2019 07:48:39 GMT</pubDate>
    <dc:creator>Tomas79</dc:creator>
    <dc:date>2019-08-07T07:48:39Z</dc:date>
    <item>
      <title>Spark job getting failed with Jupyter notebook</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/93777#M57364</link>
      <description>&lt;P&gt;I built Spark2 with CDH 5.16 and able to submit scala jobs with no issues. Now I am able to launch pyspark2 and when I am trying to run simple job, its throwing the below error? can you please suggest on this? what is the alternate to submit python jobs on spark jobs apart from Jupyter notebook? Pls advise&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s1"&gt;[I 23:08:33.864 NotebookApp]&lt;/SPAN&gt;&lt;SPAN class="s2"&gt; Adapting to protocol v5.1 for kernel f8d7200b-6718-49f6-86e9-c051fb6d84a6&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;[Stage 0:&amp;gt;&lt;SPAN class="Apple-converted-space"&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;(0 + 0) / 2]Exception in thread "dispatcher-event-loop-0" java.lang.OutOfMemoryError: Java heap space&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at java.util.Arrays.copyOf(Arrays.java:3236)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at java.io.ByteArrayOutputStream.grow(ByteArrayOutputStream.java:118)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at java.io.ByteArrayOutputStream.ensureCapacity(ByteArrayOutputStream.java:93)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at java.io.ByteArrayOutputStream.write(ByteArrayOutputStream.java:153)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at org.apache.spark.util.ByteBufferOutputStream.write(ByteBufferOutputStream.scala:41)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;at java.io.ObjectOutputStream$BlockDataOutputStream.write(ObjectOutputStream.java:1853)&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;Thanks&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s2"&gt;CS&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&amp;nbsp;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s1"&gt;19/08/06 23:10:41 WARN cluster.YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;/SPAN&gt;&lt;/P&gt;
&lt;P class="p1"&gt;&lt;SPAN class="s1"&gt;[Stage 0:&amp;gt;&lt;SPAN class="Apple-converted-space"&gt;&amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &amp;nbsp; &lt;/SPAN&gt;(0 + 0) / 2]19/08/06 23:10:47 WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: Requesting driver to remove executor 2 for reason Container marked as failed: container_1565048178604_0033_01_000003 on host: ukvmlx-rdk-22.rms.com. Exit status: 1. Diagnostics: Exception from container-launch.&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 08 Aug 2019 20:55:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/93777#M57364</guid>
      <dc:creator>Chittu</dc:creator>
      <dc:date>2019-08-08T20:55:09Z</dc:date>
    </item>
    <item>
      <title>Re: Spark job getting failed with Jupyter notebook</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/93785#M57368</link>
      <description>&lt;P&gt;Hi, the probable root cause is that the spark job submitted by the Jupyter notebook has a different memory config parameters. So I dont think the issue is Jupyter, but rather the executor and driver memory settings. Yarn is not able to provide enough resources (i.e. memory)&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;19/08/06 23:10:41 WARN cluster.YarnScheduler: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient resources&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;Check your cluster settings:&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;- how much memory YARN has allocated in NodeManagers, how big the container could be&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&amp;nbsp;- what are the submit options of your spark job&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 07 Aug 2019 07:48:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/93785#M57368</guid>
      <dc:creator>Tomas79</dc:creator>
      <dc:date>2019-08-07T07:48:39Z</dc:date>
    </item>
    <item>
      <title>Re: Spark job getting failed with Jupyter notebook</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/94110#M57557</link>
      <description>&lt;P&gt;Hi Chittu,&lt;BR /&gt;&lt;BR /&gt;Your issue here is that your JVM process is running out of memory, specifically heap space:&lt;BR /&gt;&lt;BR /&gt;&lt;SPAN&gt;java.lang.OutOfMemoryError: Java heap space&lt;/SPAN&gt;&lt;BR /&gt;&lt;BR /&gt;Judging from the output you shared, I believe this is your driver that's running out of memory and so you would need to increase the maximum heap size for the driver. That's done by configuring the spark.driver.memory parameter or by passing the --driver-memory flag to the Spark command being used.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Aug 2019 03:19:45 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/94110#M57557</guid>
      <dc:creator>w@leed</dc:creator>
      <dc:date>2019-08-16T03:19:45Z</dc:date>
    </item>
    <item>
      <title>Re: Spark job getting failed with Jupyter notebook</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/286958#M212769</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As mentioned in the previous posts, did you tried increasing the memory and whether it solved the issue?&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please let us know if you are still facing any issues?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;AKR&lt;/P&gt;</description>
      <pubDate>Mon, 06 Jan 2020 17:26:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spark-job-getting-failed-with-Jupyter-notebook/m-p/286958#M212769</guid>
      <dc:creator>AKR</dc:creator>
      <dc:date>2020-01-06T17:26:06Z</dc:date>
    </item>
  </channel>
</rss>

