<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Idle Spark Shells in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40997#M28460</link>
    <description>&lt;P&gt;FWIW, there is a safety valve setting in CM for spark-defaults.conf&lt;/P&gt;</description>
    <pubDate>Wed, 18 May 2016 20:09:44 GMT</pubDate>
    <dc:creator>scrawford</dc:creator>
    <dc:date>2016-05-18T20:09:44Z</dc:date>
    <item>
      <title>Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40881#M28454</link>
      <description>&lt;P&gt;&lt;SPAN&gt;We have some users who start Spark shells and leave them open indefinitely. Without using dynamic resource allocation to deallocate executors - would it be possible to write&amp;nbsp;something to poll YARN to determine if a Spark shell isn't doing anything, and after X time period of inactivity, kill it?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 10:19:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40881#M28454</guid>
      <dc:creator>Braundmeier</dc:creator>
      <dc:date>2022-09-16T10:19:55Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40883#M28455</link>
      <description>&lt;P&gt;Heh, that is a large part of what dynamic allocation was meant for, so you could have a long running process that could only consume resources when it's active. and a shell sitting open is a prime example of that.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;To some degree you can manage this via resource pools in YARN, and restrict a user, group or perhaps type of usage to a certain set of resources. This would be a pretty crude limit though, just a cap on the problem. Open shells would still keep resources.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Timing out shells is tricky because you lose work and state; that's probably pretty surprising.&lt;/P&gt;&lt;P&gt;Really you want dynamic allocation for this.&lt;/P&gt;</description>
      <pubDate>Mon, 16 May 2016 22:11:59 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40883#M28455</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2016-05-16T22:11:59Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40955#M28456</link>
      <description>&lt;P&gt;Thanks for the reply. Being new to CDH, I do have a question. In the general settings of Spark in CM I see there's an option to either turn dynamic allocation off or on. If I wanted to tweak some of the configs for dynamic allocation as listed on the project page (&lt;A href="http://spark.apache.org/docs/latest/configuration.html#dynamic-allocation)" target="_blank"&gt;http://spark.apache.org/docs/latest/configuration.html#dynamic-allocation)&lt;/A&gt; - would I do so via the Advanced Configuration snippets?&lt;/P&gt;</description>
      <pubDate>Tue, 17 May 2016 19:59:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40955#M28456</guid>
      <dc:creator>Braundmeier</dc:creator>
      <dc:date>2016-05-17T19:59:43Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40956#M28457</link>
      <description>Typically you set this per job on the command line as args to&lt;BR /&gt;spark-shell. If a setting is really something to establish as a&lt;BR /&gt;default, you can update or point to a new, different&lt;BR /&gt;spark-defaults.conf for your jobs. Advanced config snippets are for&lt;BR /&gt;services, like the Spark history server, at least to my understanding.&lt;BR /&gt;I'm not sure that would apply.&lt;BR /&gt;</description>
      <pubDate>Tue, 17 May 2016 20:08:16 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40956#M28457</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2016-05-17T20:08:16Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40957#M28458</link>
      <description>&lt;P&gt;If the problem is users leaving their shells open, I don't think I can trust them to add extra parameters to their CLI arugments to ensure they don't eat up extra resources (from their point of view, why would they care if they're using up my resources?).&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;How about changing the&amp;nbsp;&lt;SPAN&gt;spark-defaults.conf for the Gateway Default Group in CM? Would that accomplish what I'm looking for?&lt;/SPAN&gt;&lt;/P&gt;</description>
      <pubDate>Tue, 17 May 2016 20:34:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40957#M28458</guid>
      <dc:creator>Braundmeier</dc:creator>
      <dc:date>2016-05-17T20:34:31Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40958#M28459</link>
      <description>Yes, that sounds right, though I confess I haven't tried that myself.&lt;BR /&gt;Others here may have better suggestions.&lt;BR /&gt;</description>
      <pubDate>Tue, 17 May 2016 20:41:44 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40958#M28459</guid>
      <dc:creator>srowen</dc:creator>
      <dc:date>2016-05-17T20:41:44Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40997#M28460</link>
      <description>&lt;P&gt;FWIW, there is a safety valve setting in CM for spark-defaults.conf&lt;/P&gt;</description>
      <pubDate>Wed, 18 May 2016 20:09:44 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/40997#M28460</guid>
      <dc:creator>scrawford</dc:creator>
      <dc:date>2016-05-18T20:09:44Z</dc:date>
    </item>
    <item>
      <title>Re: Idle Spark Shells</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/66688#M28461</link>
      <description>&lt;P&gt;Can you expand on this? Am pretty new to spark and this is marked as the solution.&amp;nbsp;&lt;/P&gt;&lt;P&gt;Also, since dynamicAllocation can handle this why would an user not want to enable that instead?&lt;/P&gt;</description>
      <pubDate>Tue, 24 Apr 2018 18:53:37 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Idle-Spark-Shells/m-p/66688#M28461</guid>
      <dc:creator>ashwarg</dc:creator>
      <dc:date>2018-04-24T18:53:37Z</dc:date>
    </item>
  </channel>
</rss>

