<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Extra worker in spark in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/18124#M26698</link>
    <description>&lt;P&gt;Thanks you very much&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;It solved my problem.&lt;/P&gt;</description>
    <pubDate>Wed, 03 Sep 2014 07:04:53 GMT</pubDate>
    <dc:creator>nishi</dc:creator>
    <dc:date>2014-09-03T07:04:53Z</dc:date>
    <item>
      <title>Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17760#M26683</link>
      <description>&lt;P&gt;Hi All,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I have cloudera enterprise data hub edition 5.1.0 installed in a single system. Due to some requirement I need to create one extra worker in spark. Currently, it has 1 master and 1 worker running but I want 1 master and 2 worker. I have tried to create following the guideline of CDH(added SPARK_WORKER_INSTANCES=2) in spark-env.sh. It didn't worked for me.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I followed the same steps in spark out of CDH(just downlaoded from apache website) I am able create extra worker.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Could someone let me know what would be steps for creating extra worker in spark inside CDH 5.1.0?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks in advance.&lt;/P&gt;&lt;P&gt;Nishikant&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:06:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17760#M26683</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2022-09-16T09:06:16Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17762#M26684</link>
      <description>&lt;P&gt;I assume you're working in standalone mode. You can just go to the Spark service in Cloudera Manager, click Instances, click Add Role Instances, and assign other hosts as workers.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;You do not need to install Spark. It is already installed. In fact I would not change its configuration files directly unless you're sure you know what you're doing.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 28 Aug 2014 12:09:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17762#M26684</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-08-28T12:09:41Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17764#M26685</link>
      <description>I had installed spark outside the CDH to verify the steps those are required to create extra worker in standalone mode.&lt;BR /&gt;&lt;BR /&gt;As per your reply go to Spark service in Cloudera Manager, click Instances, click Add Role Instances. here I want to give same host as another worker but it not taking.&lt;BR /&gt;Could you provide some screen shot or something and reply to my mail ID- nkantkumar@gmail.com&lt;BR /&gt;</description>
      <pubDate>Thu, 28 Aug 2014 12:48:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17764#M26685</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2014-08-28T12:48:09Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17820#M26686</link>
      <description>&lt;P&gt;Hi All,&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I am not able to create one extra worker in spark in CDH. I need 2 workers with 1 master in my cdh spark.&lt;/P&gt;&lt;P&gt;CDH spark has 1 master and 1 worker as default , this way I am not able to do group by opearion on streams. because of that I am looking for minimum 2 workers.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks in advance&lt;/P&gt;&lt;P&gt;Nishi&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 29 Aug 2014 07:54:40 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17820#M26686</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2014-08-29T07:54:40Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17834#M26687</link>
      <description>&lt;P&gt;It doesn't make sense to put two workers on one host. One worker can host many executors, and an executor can even run many tasks in parallel. Your default parallelism will be a function of the number of cores, which should much more than 1. As long as your input has more than one partition you'll get parallel execution. If not, use repartition() to make more partitions.&lt;/P&gt;</description>
      <pubDate>Fri, 29 Aug 2014 11:00:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17834#M26687</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-08-29T11:00:43Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17938#M26688</link>
      <description>Thanks a lot&lt;BR /&gt;&lt;BR /&gt;I am going through your suggestion to create more number of partition in RDD to achieve groupByKey on stream data.&lt;BR /&gt;&lt;BR /&gt;Meanwhile, Could you please let me know how to add another worker in different host. I have 2 m/c where cloudera enterprise data hub edition 5.1.0 is installed. I want one master and 2 workers. one worker will be on another machine.</description>
      <pubDate>Mon, 01 Sep 2014 07:45:35 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17938#M26688</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2014-09-01T07:45:35Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17944#M26692</link>
      <description>&lt;P&gt;See my message above about modifying roles. You would just set an additional host to be a worker. I'm assuming you are using standalone mode.&lt;/P&gt;</description>
      <pubDate>Mon, 01 Sep 2014 09:27:17 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17944#M26692</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2014-09-01T09:27:17Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17952#M26697</link>
      <description>Problem is that It doesn't reflect second host information inside, add role instance&amp;gt; select host. It is only showing my current host and all it's information.&lt;BR /&gt;&lt;BR /&gt;Could you please let me know, why it is happening.</description>
      <pubDate>Mon, 01 Sep 2014 11:08:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/17952#M26697</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2014-09-01T11:08:14Z</dc:date>
    </item>
    <item>
      <title>Re: Extra worker in spark</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/18124#M26698</link>
      <description>&lt;P&gt;Thanks you very much&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;It solved my problem.&lt;/P&gt;</description>
      <pubDate>Wed, 03 Sep 2014 07:04:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Extra-worker-in-spark/m-p/18124#M26698</guid>
      <dc:creator>nishi</dc:creator>
      <dc:date>2014-09-03T07:04:53Z</dc:date>
    </item>
  </channel>
</rss>

