<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Difference between local[*] vs yarn cluster vs yarn client for SparkConf - Java,SparkConf Master URL Configuration in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Difference-between-local-vs-yarn-cluster-vs-yarn-client-for/m-p/161248#M123627</link>
    <description>&lt;P&gt;&lt;STRONG&gt;local[*]&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;new SparkConf() .setMaster("local[2]")&lt;/PRE&gt;&lt;UL&gt;
&lt;LI&gt;This is specific to run the job in local mode&lt;/LI&gt;&lt;LI&gt; This is specifically used to test the code in small amount of data in local environment&lt;/LI&gt;&lt;LI&gt;It Does not provide the advantages of distributed environment&lt;/LI&gt;&lt;LI&gt;* is the number of cpu cores to be allocated to perform the local operation&lt;/LI&gt;&lt;LI&gt;It helps in debugging the code by applying breakpoints while running from Eclipse or IntelliJ&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;STRONG&gt;
&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;yarn-client&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;--master yarn --deploy-mode client&lt;/PRE&gt;&lt;UL&gt;
&lt;LI&gt;Yarn client mode: your driver program is running on the yarn client where you type the command to submit the spark application (may not be a machine in the yarn cluster). In this mode, although the drive program is running on the client machine, the tasks are executed on the executors in the node managers of the YARN cluster&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;STRONG&gt;yarn-cluster&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;--master yarn --deploy-mode cluster&lt;/PRE&gt;&lt;UL&gt;
&lt;LI&gt;This is the most advisable pattern for executing/submitting your spark jobs in production&lt;/LI&gt;&lt;LI&gt;Yarn cluster mode: Your driver program is running on the cluster master machine where you type the command to submit the spark application&lt;/LI&gt;&lt;/UL&gt;</description>
    <pubDate>Wed, 22 Mar 2017 13:31:44 GMT</pubDate>
    <dc:creator>aaditya_a_deshp</dc:creator>
    <dc:date>2017-03-22T13:31:44Z</dc:date>
  </channel>
</rss>

