<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: spark-submit works on single node only in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20366#M3302</link>
    <description>&lt;P&gt;Thanks Srowen.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;After tunning the Java Heap memory for all nodes through CM and also increased the driver and worker memory to 6GB and 3GB resp the "&lt;SPAN style="color: #666666; font-family: arial, sans-serif; font-size: 11px; font-style: normal; font-variant: normal; font-weight: normal; letter-spacing: normal; line-height: 14px; orphans: auto; text-align: left; text-indent: 0px; text-transform: none; white-space: normal; widows: auto; word-spacing: 0px; -webkit-text-stroke-width: 0px; display: inline !important; float: none; background-color: #ffffff;"&gt;TaskSchedulerImpl: Initial job has not accepted any resources&lt;/SPAN&gt;" issue got resolved.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards,&lt;/P&gt;&lt;P&gt;&amp;nbsp; Shailesh&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Thu, 16 Oct 2014 21:45:39 GMT</pubDate>
    <dc:creator>Shailesh</dc:creator>
    <dc:date>2014-10-16T21:45:39Z</dc:date>
    <item>
      <title>spark-submit works on single node only</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20220#M3298</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have a CDH 5.1.3 cluster running on 4 nodes.&lt;/P&gt;&lt;P&gt;I am executing a TallSkinnySVD program (modified bit to run on big data).&lt;/P&gt;&lt;P&gt;When I execute it on cluster it always shows only one executor. I am specifying the number of executors in command but still it’s not working.&lt;/P&gt;&lt;P&gt;One more strange behaviour, I can see the process on &lt;A href="http://%3cspark-host%3e:4040" target="_blank"&gt;http://&amp;lt;spark-host&amp;gt;:4040&lt;/A&gt; but the same process is not listing on Spark UI (from CM on port 18080).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The command I am using is as below:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;$ sudo -u hdfs spark-submit --executor-memory 3g --driver-memory 6g --num-executors 10 --class org.xyz.spark.examples.mllib.TallSkinnySVD --master spark://myhost:7077 target/sparkExamples-0.0.1-SNAPSHOT.jar hdfs://&amp;lt;hdfshost&amp;gt;:8020/user/shailesh/RData/data7K.csv false&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please let me know if I am missing anything ?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;&amp;nbsp; Shailesh&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:09:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20220#M3298</guid>
      <dc:creator>Shailesh</dc:creator>
      <dc:date>2022-09-16T09:09:53Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit works on single node only</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20222#M3299</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Got resolved with error of&amp;nbsp;submitting job on cluster but now its giving different error. (Its was my fault in code).&lt;/P&gt;&lt;P&gt;Its not able submit job on worker. PFB the log.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;----------------------------------&lt;/P&gt;&lt;P&gt;14/10/15 18:09:08 INFO DAGScheduler: Parents of final stage: List()&lt;BR /&gt;14/10/15 18:09:08 INFO DAGScheduler: Missing parents: List()&lt;BR /&gt;14/10/15 18:09:08 INFO DAGScheduler: Submitting Stage 0 (MappedRDD[2] at map at TallSkinnySVD.scala:84), which has no missing parents&lt;BR /&gt;14/10/15 18:09:08 INFO DAGScheduler: Submitting 7 missing tasks from Stage 0 (MappedRDD[2] at map at TallSkinnySVD.scala:84)&lt;BR /&gt;14/10/15 18:09:08 INFO TaskSchedulerImpl: Adding task set 0.0 with 7 tasks&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;14/10/15 18:09:23 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:09:38 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:09:53 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:10:08 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:10:23 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:10:38 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;BR /&gt;14/10/15 18:10:53 WARN TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;-------------------------------------&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please note I have set 4G Java heap for each node from CM. And there are no other tasks running on the cluster.&lt;/P&gt;&lt;P&gt;Not getting why its giving such error.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks,&lt;/P&gt;&lt;P&gt;&amp;nbsp; Shailesh&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 15 Oct 2014 05:17:05 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20222#M3299</guid>
      <dc:creator>Shailesh</dc:creator>
      <dc:date>2014-10-15T05:17:05Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit works on single node only</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20226#M3300</link>
      <description>&lt;P&gt;TallSkinnySVD calls RowMatrix.computeSVD, and by default it will decide whether to run the computation locally or not. The defaults may be causing the driver to run the computation only, depending on your data.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Oct 2014 07:09:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20226#M3300</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-10-15T07:09:53Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit works on single node only</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20228#M3301</link>
      <description>&lt;P&gt;You are using standalone mode, i.e. the "Spark" service and not YARN? Check to see that the workers are running and healthy. Did executors register at startup? Double-check they have the memory you think. If not, they may not be accepting work because they do not have the memory to allocate that you expect.&lt;/P&gt;</description>
      <pubDate>Wed, 15 Oct 2014 07:12:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20228#M3301</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-10-15T07:12:20Z</dc:date>
    </item>
    <item>
      <title>Re: spark-submit works on single node only</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20366#M3302</link>
      <description>&lt;P&gt;Thanks Srowen.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;After tunning the Java Heap memory for all nodes through CM and also increased the driver and worker memory to 6GB and 3GB resp the "&lt;SPAN style="color: #666666; font-family: arial, sans-serif; font-size: 11px; font-style: normal; font-variant: normal; font-weight: normal; letter-spacing: normal; line-height: 14px; orphans: auto; text-align: left; text-indent: 0px; text-transform: none; white-space: normal; widows: auto; word-spacing: 0px; -webkit-text-stroke-width: 0px; display: inline !important; float: none; background-color: #ffffff;"&gt;TaskSchedulerImpl: Initial job has not accepted any resources&lt;/SPAN&gt;" issue got resolved.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Regards,&lt;/P&gt;&lt;P&gt;&amp;nbsp; Shailesh&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 16 Oct 2014 21:45:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/spark-submit-works-on-single-node-only/m-p/20366#M3302</guid>
      <dc:creator>Shailesh</dc:creator>
      <dc:date>2014-10-16T21:45:39Z</dc:date>
    </item>
  </channel>
</rss>

