<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode? in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186374#M148476</link>
    <description>&lt;P&gt;Ok. thanks. Now i understood that by default the amount of memory allotted for an  executor is 1 GB and this value can be controlled through --executor-memory option. Now, I would like to know that By default, how many executors will be created for an application in a node and what is the  total number of executors created in a cluster? How to control number of executors created in a node? Also, by default how many cores will be allotted to an executor in a node(I think that the number of cores allotted for an executor is unlimited in a node.Am i right?)?&lt;/P&gt;</description>
    <pubDate>Tue, 04 Apr 2017 18:56:55 GMT</pubDate>
    <dc:creator>saranpons3</dc:creator>
    <dc:date>2017-04-04T18:56:55Z</dc:date>
    <item>
      <title>What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode?</title>
      <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186372#M148474</link>
      <description>&lt;P&gt;My set up is as follows: 1 laptop where i am running wordcount scala program through spark-submit command. The input for my application is a text file which is placed in HDFS. I'm using Spark's standalone cluster for managing my cluster. I'm running my application on a kind of pseudo distributed mode. while executing spark-submit command i don't use --executor-memory option. My execution command looks like this? I would like to know that how much memory will be allotted to executor by default when --executor-memory option is not given. My interface looks like the one in the image attached.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="14371-screenshot-from-2017-04-04-131643.png" style="width: 1299px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/18806i8D39D217A260E738/image-size/medium?v=v2&amp;amp;px=400" role="button" title="14371-screenshot-from-2017-04-04-131643.png" alt="14371-screenshot-from-2017-04-04-131643.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;spark-submit --class Wordcount --master spark://saravanan:7077 /home/hduser/sparkapp/target/scala-2.11/sparkapp_2.11-0.1.jar hdfs://127.0.0.1:9000//inp_wrd hdfs://127.0.0.1:9000//amazon_wrd_count1&lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 08:12:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186372#M148474</guid>
      <dc:creator>saranpons3</dc:creator>
      <dc:date>2019-08-18T08:12:27Z</dc:date>
    </item>
    <item>
      <title>Re: What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode?</title>
      <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186373#M148475</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/16783/saranpons3.html" nodeid="16783" target="_blank"&gt;@Saravanan Selvam&lt;/A&gt;&lt;P&gt;Default executor memory will shown at spark-&amp;gt;config-&amp;gt;Advance spark-env, for reference check the attached image.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="14373-capture1.png" style="width: 1350px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/18805iA75876940440D3EF/image-size/medium?v=v2&amp;amp;px=400" role="button" title="14373-capture1.png" alt="14373-capture1.png" /&gt;&lt;/span&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 18 Aug 2019 08:12:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186373#M148475</guid>
      <dc:creator>rajsyrus</dc:creator>
      <dc:date>2019-08-18T08:12:19Z</dc:date>
    </item>
    <item>
      <title>Re: What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode?</title>
      <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186374#M148476</link>
      <description>&lt;P&gt;Ok. thanks. Now i understood that by default the amount of memory allotted for an  executor is 1 GB and this value can be controlled through --executor-memory option. Now, I would like to know that By default, how many executors will be created for an application in a node and what is the  total number of executors created in a cluster? How to control number of executors created in a node? Also, by default how many cores will be allotted to an executor in a node(I think that the number of cores allotted for an executor is unlimited in a node.Am i right?)?&lt;/P&gt;</description>
      <pubDate>Tue, 04 Apr 2017 18:56:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186374#M148476</guid>
      <dc:creator>saranpons3</dc:creator>
      <dc:date>2017-04-04T18:56:55Z</dc:date>
    </item>
    <item>
      <title>Re: What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode?</title>
      <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186375#M148477</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/16783/saranpons3.html" nodeid="16783"&gt;@Saravanan Selvam&lt;/A&gt;
&lt;/P&gt;&lt;P&gt;how many executors will be created for an application in a node?what is the total number of executors created in a cluster for an application?&lt;/P&gt;&lt;P&gt;Based on the SPARK_EXECUTOR_INSTANCE by default.&lt;/P&gt;&lt;P&gt;how many cores will be allotted to an executor in a node?&lt;/P&gt;&lt;P&gt;Based on the SPARK_EXECUTOR_CORES by default.&lt;/P&gt;</description>
      <pubDate>Tue, 04 Apr 2017 19:30:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186375#M148477</guid>
      <dc:creator>rajsyrus</dc:creator>
      <dc:date>2017-04-04T19:30:39Z</dc:date>
    </item>
    <item>
      <title>Re: What is the default value of --executor -memory when spark-submit is running in the standalone cluster pseudo distributed mode?</title>
      <link>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186376#M148478</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;Below are the default configuration values which will be considered by the spark job if these are not overriden at the time of submitting job to the required values.&lt;/P&gt;&lt;P&gt;# - SPARK_EXECUTOR_INSTANCES, Number of workers to start (Default: 2)&lt;BR /&gt;# - SPARK_EXECUTOR_CORES, Number of cores for the workers (Default: 1).&lt;BR /&gt;# - SPARK_EXECUTOR_MEMORY, Memory per Worker (e.g. 1000M, 2G) (Default: 1G)&lt;/P&gt;&lt;P&gt;SPARK_EXECUTOR_INSTANCES -&amp;gt; indicates the number of workers to be started, it means for a job maximum this many number of executors it can take from the cluster resource manager.&lt;/P&gt;&lt;P&gt;SPARK_EXECUTOR_CORES -&amp;gt; indicates the number of cores in each executor, it means the spark TaskScheduler will ask this many cores to be allocated/blocked in each of the executor machine.&lt;/P&gt;&lt;P&gt;SPARK_EXECUTOR_MEMORY -&amp;gt; indicates the maximum amount of RAM/MEMORY it requires in each executor.&lt;/P&gt;&lt;P&gt;All these details are asked by the TastScheduler to the cluster manager (it may be a spark standalone, yarn, mesos and can be kubernetes starting from spark 2.0) to provide before actually the job execution starts.&lt;/P&gt;&lt;P&gt;Also, please note that, initial number of executor instances is dependent on "--num-executors" but when the data is more to be processed and "spark.dynamicAllocation.enabled" set true, then it will be dynamically add more executors based on "spark.dynamicAllocation.initialExecutors".&lt;/P&gt;&lt;P&gt;Note: Always "spark.dynamicAllocation.initialExecutors" should be configured greater than "--num-executors".&lt;/P&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;CODE&gt;spark.dynamicAllocation.initialExecutors&lt;/CODE&gt;&lt;/TD&gt;&lt;TD&gt;&lt;CODE&gt;spark.dynamicAllocation.minExecutors&lt;/CODE&gt;&lt;/TD&gt;&lt;TD&gt;Initial number of executors to run if dynamic allocation is enabled. &lt;BR /&gt;&lt;BR /&gt;If `--num-executors` (or `spark.executor.instances`) is set and larger than this value, it will be used as the initial number of executors.&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;CODE&gt;spark.executor.memory&lt;/CODE&gt;&lt;/TD&gt;&lt;TD&gt;1g&lt;/TD&gt;&lt;TD&gt;Amount of memory to use per executor process, in the same format as JVM memory strings with a size unit suffix ("k", "m", "g" or "t") (e.g. &lt;CODE&gt;512m&lt;/CODE&gt;, &lt;CODE&gt;2g&lt;/CODE&gt;).&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;&lt;CODE&gt;spark.executor.cores&lt;/CODE&gt;&lt;/TD&gt;&lt;TD&gt;1 in YARN mode, all the available cores on the worker in standalone and Mesos coarse-grained modes.&lt;/TD&gt;&lt;TD&gt;The number of cores to use on each executor. In standalone and Mesos coarse-grained modes, for more detail, see &lt;A href="http://spark.apache.org/docs/latest/spark-standalone.html#Executors%20Scheduling"&gt;this description&lt;/A&gt;.&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;</description>
      <pubDate>Fri, 26 Jul 2019 11:15:22 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/What-is-the-default-value-of-executor-memory-when-spark/m-p/186376#M148478</guid>
      <dc:creator>Former Member</dc:creator>
      <dc:date>2019-07-26T11:15:22Z</dc:date>
    </item>
  </channel>
</rss>

