<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question File too large Exception in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106140#M46428</link>
    <description>&lt;P&gt;Hi &lt;/P&gt;&lt;P&gt;I am trying to process avro record using mapreduce where the key of the map is an avro record&lt;/P&gt;&lt;PRE&gt;public void map(AvroKey&amp;lt;GenericData.Record&amp;gt; key, NullWritable value, Context context)&lt;/PRE&gt;&lt;P&gt;The job fails if the number of columns to be processed in each record goes beyond a particular value.Say for example if the number of fields in each row is more than 100, my job fails.I tried to increase the map memory and java heap space in the cluster, but it didn't help.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="9447-error.png" style="width: 863px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/23537iFECF824A8326137E/image-size/medium?v=v2&amp;amp;px=400" role="button" title="9447-error.png" alt="9447-error.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Thanks in advance&lt;/P&gt;&lt;P&gt;Aparna&lt;/P&gt;</description>
    <pubDate>Mon, 19 Aug 2019 11:54:28 GMT</pubDate>
    <dc:creator>aparna24aravind</dc:creator>
    <dc:date>2019-08-19T11:54:28Z</dc:date>
    <item>
      <title>File too large Exception</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106140#M46428</link>
      <description>&lt;P&gt;Hi &lt;/P&gt;&lt;P&gt;I am trying to process avro record using mapreduce where the key of the map is an avro record&lt;/P&gt;&lt;PRE&gt;public void map(AvroKey&amp;lt;GenericData.Record&amp;gt; key, NullWritable value, Context context)&lt;/PRE&gt;&lt;P&gt;The job fails if the number of columns to be processed in each record goes beyond a particular value.Say for example if the number of fields in each row is more than 100, my job fails.I tried to increase the map memory and java heap space in the cluster, but it didn't help.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="9447-error.png" style="width: 863px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/23537iFECF824A8326137E/image-size/medium?v=v2&amp;amp;px=400" role="button" title="9447-error.png" alt="9447-error.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Thanks in advance&lt;/P&gt;&lt;P&gt;Aparna&lt;/P&gt;</description>
      <pubDate>Mon, 19 Aug 2019 11:54:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106140#M46428</guid>
      <dc:creator>aparna24aravind</dc:creator>
      <dc:date>2019-08-19T11:54:28Z</dc:date>
    </item>
    <item>
      <title>Re: File too large Exception</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106141#M46429</link>
      <description>&lt;P&gt;Hi Aparna,&lt;/P&gt;&lt;P&gt;Please go through this URL hope it will help you.&lt;/P&gt;&lt;P&gt;&lt;A href="http://stackoverflow.com/questions/25242287/filenotfoundexception-file-too-large" target="_blank"&gt;http://stackoverflow.com/questions/25242287/filenotfoundexception-file-too-large&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 17 Nov 2016 21:11:25 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106141#M46429</guid>
      <dc:creator>maheshmsh88</dc:creator>
      <dc:date>2016-11-17T21:11:25Z</dc:date>
    </item>
    <item>
      <title>Re: File too large Exception</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106142#M46430</link>
      <description>&lt;P&gt;Have you tried this in Spark?   or NiFi?&lt;/P&gt;&lt;P&gt;How much memory is configured in your app?&lt;/P&gt;&lt;P&gt;How much is configured in YARN for your job resources?&lt;/P&gt;&lt;P&gt;Can you post additional logs?   code?   submit details?&lt;/P&gt;&lt;P&gt;Why is the key an avro record and not the value?&lt;/P&gt;&lt;P&gt;You should make sure you have enough space in HDFS and also in the regular file system as some of the reduce stage will get mapped to regular disk.&lt;/P&gt;&lt;P&gt;Can you post hdfs and regular file system &lt;STRONG&gt;df&lt;/STRONG&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 17 Nov 2016 22:28:28 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106142#M46430</guid>
      <dc:creator>TimothySpann</dc:creator>
      <dc:date>2016-11-17T22:28:28Z</dc:date>
    </item>
    <item>
      <title>Re: File too large Exception</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106143#M46431</link>
      <description>&lt;P&gt;&lt;EM&gt;Hi &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Please see my inline comments.&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Have you tried this in Spark? or NiFi?&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;No&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;How much is configured in YARN for your job resources?&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Memory allocated for yarn containers in each node - 200GB&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Can you post additional logs? code? submit details?&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;I did not get any extra info other than FSError&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;Why is the key an avro record and not the value?&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;I am using AvroKeyInputFormat&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;You should make sure you have enough space in HDFS and also in the regular file system as some of the reduce stage will get mapped to regular disk.&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;I have enough space left in HDFS more precisely &lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;HDFS -only 3% is being used and&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Local FS -only 15% is being used&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;U&gt;Ulimit&lt;/U&gt;&lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;core file
size  (blocks, -c) 0&lt;/P&gt;&lt;P&gt;data seg
size  (kbytes, -d)
unlimited&lt;/P&gt;&lt;P&gt;scheduling
priority 
(-e) 0&lt;/P&gt;&lt;P&gt;file
size 
(blocks, -f) unlimited&lt;/P&gt;&lt;P&gt;pending
signals 
(-i) 1032250&lt;/P&gt;&lt;P&gt;max locked memory  (kbytes,
-l) 64&lt;/P&gt;&lt;P&gt;max memory
size  (kbytes, -m) unlimited&lt;/P&gt;&lt;P&gt;open
files 
(-n) 1024&lt;/P&gt;&lt;P&gt;pipe
size  (512
bytes, -p) 8&lt;/P&gt;&lt;P&gt;POSIX message queues  (bytes, -q)
819200&lt;/P&gt;&lt;P&gt;real-time
priority 
(-r) 0&lt;/P&gt;&lt;P&gt;stack
size 
(kbytes, -s) 10240&lt;/P&gt;&lt;P&gt;cpu
time 
(seconds, -t) unlimited&lt;/P&gt;&lt;P&gt;max user
processes 
(-u) 1024&lt;/P&gt;&lt;P&gt;virtual
memory  (kbytes, -v)
unlimited&lt;/P&gt;&lt;P&gt;file
locks 
(-x) unlimited&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;</description>
      <pubDate>Wed, 23 Nov 2016 14:14:41 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106143#M46431</guid>
      <dc:creator>aparna24aravind</dc:creator>
      <dc:date>2016-11-23T14:14:41Z</dc:date>
    </item>
    <item>
      <title>Re: File too large Exception</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106144#M46432</link>
      <description>&lt;P&gt;Hi &lt;/P&gt;&lt;P&gt;I was able to resolve the issue,the disk utilization in local directory (where logs and out files are created) in one of the node was more than the yarn.nodemanager.disk-health-checker.max-disk-utilization-per-disk-percentage setting.
I freed up some space and also set the max-disk-utilization-percentage to much higher value.&lt;/P&gt;&lt;P&gt;Thanks&lt;/P&gt;&lt;P&gt;Aparna&lt;/P&gt;</description>
      <pubDate>Thu, 05 Jan 2017 23:07:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/File-too-large-Exception/m-p/106144#M46432</guid>
      <dc:creator>aparna24aravind</dc:creator>
      <dc:date>2017-01-05T23:07:31Z</dc:date>
    </item>
  </channel>
</rss>

