<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: need to control spark2-history size in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/need-to-control-spark2-history-size/m-p/242328#M204131</link>
    <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/112829/nikhilbelure99.html" nodeid="112829"&gt;@John&lt;/A&gt;&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;If you see your "/spark2-history" log folder has many old orphan files then probably these are the files left over due to spark-driver failures / crashes ...etc.&lt;/P&gt;&lt;P&gt;You can check your Spark configs following parameters.&lt;/P&gt;&lt;PRE&gt;spark.history.fs.cleaner.enabled=true
spark.history.fs.cleaner.interval
spark.history.fs.cleaner.maxAge&lt;/PRE&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;NOTE: &lt;/STRONG&gt;However, there are some issues reported already for older version of Spark where the spark.history.fs.cleaner required some improvements. &lt;/P&gt;&lt;P&gt;As part of &lt;A href="https://issues.apache.org/jira/browse/SPARK-8617" target="_blank"&gt;https://issues.apache.org/jira/browse/SPARK-8617&lt;/A&gt;  fix Spark 2.2 should function properly. &lt;/P&gt;&lt;P&gt;Also please check if the ownership of the files present inside the "/spark2-history" is correctly set or not? If not then please set it correctly according to your setup.&lt;/P&gt;&lt;PRE&gt;# hdfs dfs -chown spark:hadoop  /spark2-history 
&lt;/PRE&gt;&lt;P&gt;.&lt;/P&gt;</description>
    <pubDate>Thu, 20 Jun 2019 14:23:46 GMT</pubDate>
    <dc:creator>jsensharma</dc:creator>
    <dc:date>2019-06-20T14:23:46Z</dc:date>
  </channel>
</rss>

