<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Spark 1.1.0 on cdh5.1.3 does not work in yarn-cluster mode in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21398#M1388</link>
    <description>&lt;P&gt;Hello!!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have a similar issue,&amp;nbsp;I am having CDH 5 installed on my cluster (version&amp;nbsp;Hadoop 2.3.0-cdh5.1.3)&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;I have installed and configured a prebuilt version of Spark 1.1.0 (Apache Version), built for hadoop 2.3 on my cluster.&lt;/P&gt;&lt;P&gt;when I run the Pi example in the ‘client mode’, it runs succesfully, but it fails in the ‘yarn-cluster’ mode. The spark job is successfully submitted, but fails after sometime saying:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;***********************************&lt;BR /&gt;$ ./bin/spark-submit –class org.apache.spark.examples.SparkPi –master yarn-cluster –num-executors 2 –driver-memory 500m –executor-cores 2 lib/spark-examples*.jar 3&lt;/P&gt;&lt;P&gt;Logs:&lt;BR /&gt;14/11/05 20:47:47 INFO yarn.Client: Application report from ResourceManager:&lt;BR /&gt;application identifier: application_1415193640322_0013&lt;BR /&gt;appId: 13&lt;BR /&gt;clientToAMToken: null&lt;BR /&gt;appDiagnostics: Application application_1415193640322_0013 failed 2 times due to AM Container for appattempt_1415193640322_0013_000002 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException:&lt;BR /&gt;org.apache.hadoop.util.Shell$ExitCodeException:&lt;BR /&gt;***********************************&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can you please suggest any solution. Do you think I should compile the spark code on my cluster.&lt;BR /&gt;Or should I use Spark provided with CDH5.1&lt;/P&gt;&lt;P&gt;Any help will be appreciated!&lt;/P&gt;</description>
    <pubDate>Tue, 11 Nov 2014 13:04:40 GMT</pubDate>
    <dc:creator>Rakesh Gupta</dc:creator>
    <dc:date>2014-11-11T13:04:40Z</dc:date>
    <item>
      <title>Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7736#M1374</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am a newbie to Apache Spark.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have installed CDH-5 using parcels (Beta 2 Version) and installed Spark also&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As per the Spark installation documentation,&amp;nbsp; &lt;A target="_blank" href="http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH5/latest/CDH5-Installation-Guide/cdh5ig_Spark_prerequisites.html#../CDH5-Installation-Guide/../CDH5-Installation-Guide/cdh5ig_Spark_configuring.html"&gt;http://www.cloudera.com/content/cloudera-content/cloudera-docs/CDH5/latest/CDH5-Installation-Guide/cdh5ig_Spark_prerequisites.html#../CDH5-Installation-Guide/../CDH5-Installation-Guide/cdh5ig_Spark_configuring.html&lt;/A&gt;, it is said,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"&amp;nbsp; &lt;STRONG&gt;Note&lt;/STRONG&gt;:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;The current version of CDH 5 does not support running Spark on YARN.&lt;/LI&gt;&lt;LI&gt;The current version of Spark does work in a secure cluster."&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;So, if YARN in CDH-5 does not support Spark, how do we run Spark in CDH-5?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please let me know and also proivde any documentation if available.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 08:55:54 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7736#M1374</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2022-09-16T08:55:54Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7738#M1375</link>
      <description>&lt;P&gt;At the moment, CDH5b2 deploys Spark in "standalone" mode:&amp;nbsp;&lt;A target="_blank" href="https://spark.apache.org/docs/0.9.0/spark-standalone.html"&gt;https://spark.apache.org/docs/0.9.0/spark-standalone.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This simply means Spark tries to manage resources itself, rather than participating in a cluster manager like YARN or Mesos. As an end user, it shouldn't make much difference to you at all. Just fire up the shell and go.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Once a few kinks are worked out, Spark's YARN integration will be used in the future, as I understand.&lt;/P&gt;</description>
      <pubDate>Mon, 24 Mar 2014 13:51:50 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7738#M1375</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-03-24T13:51:50Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7804#M1376</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have installed Spark in Stand alone mode in CDH-5 cluster. But when I start the Spark Master, I am getting the following error -&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/spark/deploy/master/Master
Caused by: java.lang.ClassNotFoundException: org.apache.spark.deploy.master.Master
        at java.net.URLClassLoader$1.run(URLClassLoader.java:202)
        at java.security.AccessController.doPrivileged(Native Method)
        at java.net.URLClassLoader.findClass(URLClassLoader.java:190)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:306)
        at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)
        at java.lang.ClassLoader.loadClass(ClassLoader.java:247)
Could not find the main class: org.apache.spark.deploy.master.Master.  Program will exit.&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have provided the folder '/usr/lib/spark' in class path and also set the variable '&amp;nbsp;SPARK_LIBRARY_PATH=/usr/lib/spark' in the file&amp;nbsp;spark-env.sh.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Still I am facing this error.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I installed SPARK using yum.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Could you please assist? Thanks!&lt;/P&gt;</description>
      <pubDate>Tue, 25 Mar 2014 11:24:44 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7804#M1376</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-03-25T11:24:44Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7806#M1377</link>
      <description>&lt;P&gt;Are you on CDH5 beta 2? It already includes Spark. I wonder if its setup of Spark is interfering with whatever you have installed separately, or vice versa. Can you simply use the built-in deployment? It would be easier.&lt;/P&gt;</description>
      <pubDate>Tue, 25 Mar 2014 11:29:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7806#M1377</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-03-25T11:29:12Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7946#M1378</link>
      <description>I removed the version I installed and used the one available in CDH-5, it worked. Thanks!</description>
      <pubDate>Thu, 27 Mar 2014 12:15:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/7946#M1378</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-03-27T12:15:28Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8304#M1379</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am unable to run even the sample vertification job in Scala. The worker node status is howing as alive with cores&amp;nbsp;&lt;SPAN&gt;4 (0 Used) and memory&amp;nbsp;&lt;SPAN&gt;6.7 GB (0.0 B Used).&amp;nbsp;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;SPAN&gt;&amp;nbsp;&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;&lt;SPAN&gt;But I am repeatedly getting the below error. Could you please assist?&lt;/SPAN&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;PRE&gt;WARN scheduler.TaskSchedulerImpl: Initial job has not accepted any resources; check your cluster UI to ensure that workers are registered and have sufficient memory&lt;/PRE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 02 Apr 2014 14:22:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8304#M1379</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-04-02T14:22:34Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8320#M1380</link>
      <description>&lt;P&gt;I believe that means you've requested more memory for a task than any worker has available, but people more knowledgeable might be able to confirm or deny that.&lt;/P&gt;</description>
      <pubDate>Wed, 02 Apr 2014 17:00:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8320#M1380</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-04-02T17:00:59Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8362#M1381</link>
      <description>&lt;P&gt;But I have not manually requested memory anywhere or set any parameter.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;So, is there a way to control this? Thanks!&lt;/P&gt;</description>
      <pubDate>Thu, 03 Apr 2014 04:38:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8362#M1381</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-04-03T04:38:55Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8472#M1382</link>
      <description>&lt;P&gt;Apache Spark uses the derby database in background and hence only one instance of the 'Spark-Shell' can be connected at any time.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is there any way to configure mysql or any other RDMS and is there any configuration document?&lt;/P&gt;</description>
      <pubDate>Mon, 07 Apr 2014 07:01:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/8472#M1382</guid>
      <dc:creator>ArunShell</dc:creator>
      <dc:date>2014-04-07T07:01:27Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14982#M1383</link>
      <description>&lt;P&gt;My apologies to reawaken this dead thread, but the subject line was still applicable:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am trying to get Spark 1.0 to run on on CDH5 YARN. That means NOT the provided 0.9 RPMs.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Still very new to Spark, I'm trying to get things right between what seems like somewhat contradictory information between Apache's Spark site and the CDH5 installation guide. This begins with precisely *what* is installed on *which* node in the cluster. Apache simply stated "on ALL nodes", without any distinction between node role - namenode? datanode? resource manager? node manager? application master? Then there's the client, say a development machine or similar.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The Cloudera docs state to only start the master on one node - yet, which machine is best chose for this? Coming from Mapreduce this would imply the namenode. But it's not made clear that's the right choice. Should it instead be one of the datanodes (in this case they are also the node managers and application masters for YARN, in other words the worker bees of the cluster)? Or maybe it should be the client machine from which jobs are launched through shell or submit?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The Apache cluster mode overview isn't much help either. I suppose that with Spark still being fresh and new, the doco has to catch up from a typical case of "this was totally obvious to the creators" for us mere mortals.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;What makes this more confusing is that the 1.0 tarball from Apache is one large directory, as opposed to the RPMs which break the packages down. I assume I could simply install the tarball on each and every machine on the cluster, from client to namenode/resource manager to datanode/node manager and just push out configuration settings to all, but that seems somewhat "sloppy".&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;In the end it would be great if it were clear enough to take the 1.0 source and create my own RPMs, aligned with the current 0.9 CDH5 packages and install them only where needed, but for that I need a better understanding of what goes where.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Any suggestions and pointers welcome!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks!&lt;/P&gt;</description>
      <pubDate>Fri, 11 Jul 2014 12:02:54 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14982#M1383</guid>
      <dc:creator>Marakai</dc:creator>
      <dc:date>2014-07-11T12:02:54Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14984#M1384</link>
      <description>&lt;P&gt;You might wait for CDH 5.1.0, which will be released very soon. This deploys Spark 1.0.0+patches on YARN for you.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"Node" means a machine on which you want to run Spark. "Namenode" is for example an HDFS concept. It is not directly related to Spark. You may choose to run a Spark process on a machine that happens to host the namenode, or not. This is why Spark is not describe in terms of, say, HDFS roles.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You do not need to start the Spark master on the HDFS namenode. You didn't have to start the MR jobtracker on the namenode either. On a cluster I ran, I put the master on the namenode just since it's a simple default choice. But any machine that can see HDFS and YARN would be fine; it need not even be running other Hadoop services.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You can easily choose which machines are the Spark workers and which is the master in Cloudera Manager.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;The Spark master is not the same thing as a client. Its role is like that of the jobtracker really. It would not be run outside the cluster. You may be thinking of a driver for your specific app.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The Apache distro is indeed a tarball and it's up to you to deploy it and run it. The role of CDH is to package, deploy and run things for you. The packaging is not at all the same, although the contents (scripts, binaries) are of course the same. You would not try to paste the raw tarball onto CDH nodes.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you want to get adventurous, you can go to all machines and dig into /opt/cloudera/parcels/CDH/lib/spark and replace binaries with a newer compiled version. That's a manual process, and I suppose not 100% guaranteed to work, but you can try it.&lt;/P&gt;</description>
      <pubDate>Fri, 11 Jul 2014 12:28:58 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14984#M1384</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-07-11T12:28:58Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14990#M1385</link>
      <description>&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/133"&gt;@srowen&lt;/a&gt; wrote:&lt;BR /&gt;&lt;P&gt;You might wait for CDH 5.1.0, which will be released very soon. This deploys Spark 1.0.0+patches on YARN for you.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;"Node" means a machine on which you want to run Spark. "Namenode" is for example an HDFS concept. It is not directly related to Spark. You may choose to run a Spark process on a machine that happens to host the namenode, or not. This is why Spark is not describe in terms of, say, HDFS roles.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You do not need to start the Spark master on the HDFS namenode. You didn't have to start the MR jobtracker on the namenode either. On a cluster I ran, I put the master on the namenode just since it's a simple default choice. But any machine that can see HDFS and YARN would be fine; it need not even be running other Hadoop services.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You can easily choose which machines are the Spark workers and which is the master in Cloudera Manager.&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;The Spark master is not the same thing as a client. Its role is like that of the jobtracker really. It would not be run outside the cluster. You may be thinking of a driver for your specific app.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;The Apache distro is indeed a tarball and it's up to you to deploy it and run it. The role of CDH is to package, deploy and run things for you. The packaging is not at all the same, although the contents (scripts, binaries) are of course the same. You would not try to paste the raw tarball onto CDH nodes.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If you want to get adventurous, you can go to all machines and dig into /opt/cloudera/parcels/CDH/lib/spark and replace binaries with a newer compiled version. That's a manual process, and I suppose not 100% guaranteed to work, but you can try it.&lt;/P&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;Ah, thanks so much for that info! This covered a lot of my answers in one fell swoop.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am of course aware that namenode/datanodes and RM/NM are not synonymous, however for simplicity sake in put them together as they frequently are.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;My assumption was that the master was equivalent to a jobtracker, as you said and therefore would frequently be found on a NN/RM node, whereas the workers would go on a DN/NM/AM. Again, lumping those Hadoop components together.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;If not, how would the workers access files on HDFS, unless by streaming? What would that do to performance? It's unfortunate that the Apache docs don't give a really detailed view of the architecture and component interaction both within Spark as well as with the various Hadoop components.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I think it's this that cause my confusion: if we're talking about Spark operating in a YARN environment, then there's a tacit implication of also having a "typical" underlying infrastructure based on your usual Hadoop cluster. Even if we take YARN out of the equation, if your data is on HDFS, then where do the Spark workers need to sit to ensure the maximum access speed? Talk of RDDs is all wel and good, but at some point your data is not all in memory, it's on platters whence it must get INTO memory! &lt;span class="lia-unicode-emoji" title=":slightly_smiling_face:"&gt;🙂&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When's 5.1 coming? &lt;span class="lia-unicode-emoji" title=":winking_face:"&gt;😉&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 11 Jul 2014 13:16:52 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14990#M1385</guid>
      <dc:creator>Marakai</dc:creator>
      <dc:date>2014-07-11T13:16:52Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14992#M1386</link>
      <description>&lt;P&gt;This might help a lot:&amp;nbsp;&lt;A target="_blank" href="http://blog.cloudera.com/blog/2014/05/apache-spark-resource-management-and-yarn-app-models/"&gt;http://blog.cloudera.com/blog/2014/05/apache-spark-resource-management-and-yarn-app-models/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Yes you want Spark executors to end up colocated with datanodes or else data has to be accessed over the network a lot. It works but of course ideally workers all process only local data. You should get that if YARN nodemanagers are colocated with datanodes, since YARN is the thing running Spark's executors in its containers, when using YARN.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Things get confusing because of at least two things. First, there are two different types of YARN deployment, although, I don't think they affect how you think about placing services. But second, there is also "standalone" mode, the default in 0.9.0 and what you are currently using, wherein you actually do separately control where Spark workers run, separately from YARN.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I suppose I'd say the thing that matters is: datanodes and nodemanagers and spark worker services are present on all machines doing work.&lt;/P&gt;</description>
      <pubDate>Fri, 11 Jul 2014 13:49:49 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/14992#M1386</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-07-11T13:49:49Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/15010#M1387</link>
      <description>&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/133"&gt;@srowen&lt;/a&gt; wrote:&lt;BR /&gt;&lt;P&gt;This might help a lot:&amp;nbsp;&lt;A target="_blank" href="http://blog.cloudera.com/blog/2014/05/apache-spark-resource-management-and-yarn-app-models/"&gt;http://blog.cloudera.com/blog/2014/05/apache-spark-resource-management-and-yarn-app-models/&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;[...]&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Things get confusing because of at least two things. First, there are two different types of YARN deployment, although, I don't think they affect how you think about placing services. But second, there is also "standalone" mode, the default in 0.9.0 and what you are currently using, wherein you actually do separately control where Spark workers run, separately from YARN.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I suppose I'd say the thing that matters is: datanodes and nodemanagers and spark worker services are present on all machines doing work.&lt;/P&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&amp;nbsp;Many thanks for the blog link! I need to spend a week just catching up on the Cloudera blog one of these days. You guys produce some fantastic articles in there.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Yes, it was the mild confusion about 0.9 using this "standalone" mode instead of YARN which made me try and for for 1.0. Whether it was something I read on the Apache Spark pages or some other link via Google gave me the impression that the 0.9 RPMs are hard-wired for stand-alone mode and can't be configured for YARN instead. Could have also been a comment/note box in the CDH5 install guide? Anyway, I'm still digging through the config options to see whether 0.9 can be configured for YARN instead of standalone.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;May get around to it this weekend, unless I crash and fall asleep, dang Worldcup! &lt;span class="lia-unicode-emoji" title=":winking_face:"&gt;😉&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 11 Jul 2014 23:13:11 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/15010#M1387</guid>
      <dc:creator>Marakai</dc:creator>
      <dc:date>2014-07-11T23:13:11Z</dc:date>
    </item>
    <item>
      <title>Spark 1.1.0 on cdh5.1.3 does not work in yarn-cluster mode</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21398#M1388</link>
      <description>&lt;P&gt;Hello!!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have a similar issue,&amp;nbsp;I am having CDH 5 installed on my cluster (version&amp;nbsp;Hadoop 2.3.0-cdh5.1.3)&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;I have installed and configured a prebuilt version of Spark 1.1.0 (Apache Version), built for hadoop 2.3 on my cluster.&lt;/P&gt;&lt;P&gt;when I run the Pi example in the ‘client mode’, it runs succesfully, but it fails in the ‘yarn-cluster’ mode. The spark job is successfully submitted, but fails after sometime saying:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;***********************************&lt;BR /&gt;$ ./bin/spark-submit –class org.apache.spark.examples.SparkPi –master yarn-cluster –num-executors 2 –driver-memory 500m –executor-cores 2 lib/spark-examples*.jar 3&lt;/P&gt;&lt;P&gt;Logs:&lt;BR /&gt;14/11/05 20:47:47 INFO yarn.Client: Application report from ResourceManager:&lt;BR /&gt;application identifier: application_1415193640322_0013&lt;BR /&gt;appId: 13&lt;BR /&gt;clientToAMToken: null&lt;BR /&gt;appDiagnostics: Application application_1415193640322_0013 failed 2 times due to AM Container for appattempt_1415193640322_0013_000002 exited with exitCode: 1 due to: Exception from container-launch: org.apache.hadoop.util.Shell$ExitCodeException:&lt;BR /&gt;org.apache.hadoop.util.Shell$ExitCodeException:&lt;BR /&gt;***********************************&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can you please suggest any solution. Do you think I should compile the spark code on my cluster.&lt;BR /&gt;Or should I use Spark provided with CDH5.1&lt;/P&gt;&lt;P&gt;Any help will be appreciated!&lt;/P&gt;</description>
      <pubDate>Tue, 11 Nov 2014 13:04:40 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21398#M1388</guid>
      <dc:creator>Rakesh Gupta</dc:creator>
      <dc:date>2014-11-11T13:04:40Z</dc:date>
    </item>
    <item>
      <title>Re: Spark 1.1.0 on cdh5.1.3 does not work in yarn-cluster mode</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21399#M1389</link>
      <description>&lt;P&gt;Hm, why not just use the Spark that is part of CDH? If you want 1.1, can you update to CDH 5.2? Are there more logs? this isn't the underlying error.&lt;/P&gt;</description>
      <pubDate>Tue, 11 Nov 2014 13:22:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21399#M1389</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-11-11T13:22:12Z</dc:date>
    </item>
    <item>
      <title>Re: Spark 1.1.0 on cdh5.1.3 does not work in yarn-cluster mode</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21400#M1390</link>
      <description>&lt;P&gt;More Logs:&lt;/P&gt;&lt;DIV&gt;Application application_1415193640322_0016 failed 2 times due to Error launching appattempt_1415193640322_0016_000002. Got exception: org.apache.hadoop.yarn.exceptions.YarnException: java.io.EOFException&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.startContainers(ContainerManagerImpl.java:710)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.api.impl.pb.service.ContainerManagementProtocolPBServiceImpl.startContainers(ContainerManagementProtocolPBServiceImpl.java:60)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.proto.ContainerManagementProtocol$ContainerManagementProtocolService$2.callBlockingMethod(ContainerManagementProtocol.java:95)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:587)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1026)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)&lt;/DIV&gt;&lt;DIV&gt;at java.security.AccessController.doPrivileged(Native Method)&lt;/DIV&gt;&lt;DIV&gt;at javax.security.auth.Subject.doAs(Subject.java:415)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)&lt;/DIV&gt;&lt;DIV&gt;Caused by: java.io.EOFException&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readFully(DataInputStream.java:197)&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readUTF(DataInputStream.java:609)&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readUTF(DataInputStream.java:564)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.security.ContainerTokenIdentifier.readFields(ContainerTokenIdentifier.java:151)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.security.token.Token.decodeIdentifier(Token.java:142)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.utils.BuilderUtils.newContainerTokenIdentifier(BuilderUtils.java:262)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.startContainers(ContainerManagerImpl.java:696)&lt;/DIV&gt;&lt;DIV&gt;... 10 more&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native Method)&lt;/DIV&gt;&lt;DIV&gt;at sun.reflect.NativeConstructorAccessorImpl.newInstance(NativeConstructorAccessorImpl.java:57)&lt;/DIV&gt;&lt;DIV&gt;at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)&lt;/DIV&gt;&lt;DIV&gt;at java.lang.reflect.Constructor.newInstance(Constructor.java:526)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.ipc.RPCUtil.instantiateException(RPCUtil.java:53)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.ipc.RPCUtil.unwrapAndThrowException(RPCUtil.java:101)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagementProtocolPBClientImpl.startContainers(ContainerManagementProtocolPBClientImpl.java:99)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.launch(AMLauncher.java:118)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.resourcemanager.amlauncher.AMLauncher.run(AMLauncher.java:249)&lt;/DIV&gt;&lt;DIV&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)&lt;/DIV&gt;&lt;DIV&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)&lt;/DIV&gt;&lt;DIV&gt;at java.lang.Thread.run(Thread.java:744)&lt;/DIV&gt;&lt;DIV&gt;Caused by: org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.yarn.exceptions.YarnException): java.io.EOFException&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.ipc.RPCUtil.getRemoteException(RPCUtil.java:38)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.startContainers(ContainerManagerImpl.java:710)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.api.impl.pb.service.ContainerManagementProtocolPBServiceImpl.startContainers(ContainerManagementProtocolPBServiceImpl.java:60)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.proto.ContainerManagementProtocol$ContainerManagementProtocolService$2.callBlockingMethod(ContainerManagementProtocol.java:95)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:587)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1026)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2013)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:2009)&lt;/DIV&gt;&lt;DIV&gt;at java.security.AccessController.doPrivileged(Native Method)&lt;/DIV&gt;&lt;DIV&gt;at javax.security.auth.Subject.doAs(Subject.java:415)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1614)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2007)&lt;/DIV&gt;&lt;DIV&gt;Caused by: java.io.EOFException&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readFully(DataInputStream.java:197)&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readUTF(DataInputStream.java:609)&lt;/DIV&gt;&lt;DIV&gt;at java.io.DataInputStream.readUTF(DataInputStream.java:564)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.security.ContainerTokenIdentifier.readFields(ContainerTokenIdentifier.java:151)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.security.token.Token.decodeIdentifier(Token.java:142)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.utils.BuilderUtils.newContainerTokenIdentifier(BuilderUtils.java:262)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.server.nodemanager.containermanager.ContainerManagerImpl.startContainers(ContainerManagerImpl.java:696)&lt;/DIV&gt;&lt;DIV&gt;... 10 more&lt;/DIV&gt;&lt;DIV&gt;&amp;nbsp;&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Client.call(Client.java:1409)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.Client.call(Client.java:1362)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:206)&lt;/DIV&gt;&lt;DIV&gt;at com.sun.proxy.$Proxy69.startContainers(Unknown Source)&lt;/DIV&gt;&lt;DIV&gt;at org.apache.hadoop.yarn.api.impl.pb.client.ContainerManagementProtocolPBClientImpl.startContainers(ContainerManagementProtocolPBClientImpl.java:96)&lt;/DIV&gt;&lt;DIV&gt;... 5 more&lt;/DIV&gt;&lt;DIV&gt;. Failing the application.&lt;/DIV&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN style="text-decoration: underline;"&gt;&lt;STRONG&gt;When I go to node Manager logs:&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Log Type: stderr&lt;/P&gt;&lt;P&gt;Log Length: 87&lt;/P&gt;&lt;PRE&gt;Error: Could not find or load main class org.apache.spark.deploy.yarn.ExecutorLauncher&lt;/PRE&gt;</description>
      <pubDate>Tue, 11 Nov 2014 14:09:02 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21400#M1390</guid>
      <dc:creator>Rakesh Gupta</dc:creator>
      <dc:date>2014-11-11T14:09:02Z</dc:date>
    </item>
    <item>
      <title>Re: Spark 1.1.0 on cdh5.1.3 does not work in yarn-cluster mode</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21581#M1391</link>
      <description>&lt;P&gt;Looks like I have to try upgrading cdh to 5.2 and use SPARK that comes with it, but does support all modes of spark. i.e 'yarn-cluster',&amp;nbsp;&lt;SPAN&gt;'yarn-client' etc...&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 14 Nov 2014 11:43:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21581#M1391</guid>
      <dc:creator>Rakesh Gupta</dc:creator>
      <dc:date>2014-11-14T11:43:09Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21746#M1392</link>
      <description>&lt;P&gt;Hi, &lt;SPAN&gt;I'm just a newbie and trying to run an example first to get to know how Spark work&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;SPAN&gt;I follow the&amp;nbsp;&lt;/SPAN&gt;link here: &lt;A target="_blank" href="http://www.cloudera.com/content/cloudera/en/documentation/cdh5/v5-0-0/CDH5-Installation-Guide/cdh5ig_running_spark_apps.html?scroll=concept_w24_rsc_nn_unique_1&amp;nbsp;"&gt;http://www.cloudera.com/content/cloudera/en/documentation/cdh5/v5-0-0/CDH5-Installation-Guide/cdh5ig_running_spark_apps.html?scroll=concept_w24_rsc_nn_unique_1&amp;nbsp;&lt;/A&gt;&lt;/P&gt;&lt;P&gt;I'm trying to run in YARN client mode and&amp;nbsp;got this error:&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Exception in thread "main" java.lang.NoClassDefFoundError: org/apache/spark/examples/SparkPi&lt;BR /&gt;Caused by: java.lang.ClassNotFoundException: org.apache.spark.examples.SparkPi&lt;BR /&gt;at java.net.URLClassLoader$1.run(URLClassLoader.java:202)&lt;BR /&gt;at java.security.AccessController.doPrivileged(Native Method)&lt;BR /&gt;at java.net.URLClassLoader.findClass(URLClassLoader.java:190)&lt;BR /&gt;at java.lang.ClassLoader.loadClass(ClassLoader.java:306)&lt;BR /&gt;at sun.misc.Launcher$AppClassLoader.loadClass(Launcher.java:301)&lt;BR /&gt;at java.lang.ClassLoader.loadClass(ClassLoader.java:247)&lt;BR /&gt;Could not find the main class: org.apache.spark.examples.SparkPi. Program will exit.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm running CDH 5.2p0.36&lt;/P&gt;&lt;P&gt;Please help cause i don't even fully unserstand the guide in the link above.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks&amp;nbsp;you !&lt;/P&gt;</description>
      <pubDate>Tue, 18 Nov 2014 09:47:24 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21746#M1392</guid>
      <dc:creator>MabuXayda</dc:creator>
      <dc:date>2014-11-18T09:47:24Z</dc:date>
    </item>
    <item>
      <title>Re: Spark on YARN in CDH-5</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21748#M1393</link>
      <description>&lt;P&gt;You should use the documentation for CDH 5.2, which you are using and which corresponds to Spark 1.1:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;A target="_blank" href="http://www.cloudera.com/content/cloudera/en/documentation/core/latest/topics/cdh_ig_running_spark_apps.html"&gt;http://www.cloudera.com/content/cloudera/en/documentation/core/latest/topics/cdh_ig_running_spark_apps.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You are looking at docs for CDH 5.0.x, which corresponds to Spark 0.9. A lot has changed since then.&lt;/P&gt;</description>
      <pubDate>Tue, 18 Nov 2014 09:57:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-on-YARN-in-CDH-5/m-p/21748#M1393</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-11-18T09:57:53Z</dc:date>
    </item>
  </channel>
</rss>

