<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark 2 - attemptFailuresValidityInterval issue in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-2-attemptFailuresValidityInterval-issue/m-p/66227#M76783</link>
    <description>&lt;P&gt;Sorry, this is a bug described in&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/SPARK-22876" target="_blank"&gt;SPARK-22876&lt;/A&gt;&amp;nbsp;which suggests that the current logic of&amp;nbsp;&lt;EM&gt;spark.yarn.am.attemptFailuresValidityInterva&lt;/EM&gt;&lt;SPAN&gt;&lt;EM&gt;l&lt;/EM&gt; is flawed.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;While the jira is still being worked on, looking at the comments, I don't foresee a fix anytime soon.&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Wed, 11 Apr 2018 06:08:20 GMT</pubDate>
    <dc:creator>AutoIN</dc:creator>
    <dc:date>2018-04-11T06:08:20Z</dc:date>
    <item>
      <title>Spark 2 - attemptFailuresValidityInterval issue</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-2-attemptFailuresValidityInterval-issue/m-p/66001#M76782</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;We are running a spark-submit with options:&lt;/P&gt;
&lt;P&gt;--deploy-mode cluster&lt;/P&gt;
&lt;P&gt;--conf "spark.yarn.maxAppAttempts=3"&lt;BR /&gt;--conf "spark.yarn.am.attemptFailuresValidityInterval=30s"&lt;/P&gt;
&lt;P&gt;--conf...&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;and our application throws&amp;nbsp;intentionally an exception after 70 seconds on the driver, in order to cause a manual failure.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;We expected our application, with these parameters, to run forever, because the attemptFailuresValidityInterval should reset the maxAppAttempts counter sooner than the custom exception. But after 3 failures the application stops.&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Our installation:&lt;/P&gt;
&lt;P&gt;- SPARK2-2.1.0.cloudera2&lt;BR /&gt;- CDH 5.11&lt;/P&gt;
&lt;P&gt;&amp;nbsp;&lt;/P&gt;
&lt;P&gt;Any ideas are more than welcome!&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 13:03:36 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-2-attemptFailuresValidityInterval-issue/m-p/66001#M76782</guid>
      <dc:creator>K_bd</dc:creator>
      <dc:date>2022-09-16T13:03:36Z</dc:date>
    </item>
    <item>
      <title>Re: Spark 2 - attemptFailuresValidityInterval issue</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-2-attemptFailuresValidityInterval-issue/m-p/66227#M76783</link>
      <description>&lt;P&gt;Sorry, this is a bug described in&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/SPARK-22876" target="_blank"&gt;SPARK-22876&lt;/A&gt;&amp;nbsp;which suggests that the current logic of&amp;nbsp;&lt;EM&gt;spark.yarn.am.attemptFailuresValidityInterva&lt;/EM&gt;&lt;SPAN&gt;&lt;EM&gt;l&lt;/EM&gt; is flawed.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;While the jira is still being worked on, looking at the comments, I don't foresee a fix anytime soon.&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 11 Apr 2018 06:08:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-2-attemptFailuresValidityInterval-issue/m-p/66227#M76783</guid>
      <dc:creator>AutoIN</dc:creator>
      <dc:date>2018-04-11T06:08:20Z</dc:date>
    </item>
  </channel>
</rss>

