<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Sqoop GC overhead limit exceeded after CDH5.2 update in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20604#M3362</link>
    <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;we updated sqoop from cdh5.0.1 to cdh5.2 and now it fails everytime with a GC overhead limit exceeded error.&lt;/P&gt;&lt;P&gt;The old version was able to import over 14GB of data over one mapper and the import fails now when a mapper gets too many rows. I checked a heap dump and the memory was completely used by over 3.5 million rows of data (-Xmx 1700M).&lt;/P&gt;&lt;P&gt;The connector is mysql-jdbc version 5.1.33 and the job imports the data as text file in a have table.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can I avoid this with a setting or is this a bug that should go to jira?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you,&lt;/P&gt;&lt;P&gt;Jürgen&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 09:10:20 GMT</pubDate>
    <dc:creator>juergen_thomann</dc:creator>
    <dc:date>2022-09-16T09:10:20Z</dc:date>
    <item>
      <title>Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20604#M3362</link>
      <description>&lt;P&gt;Hi,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;we updated sqoop from cdh5.0.1 to cdh5.2 and now it fails everytime with a GC overhead limit exceeded error.&lt;/P&gt;&lt;P&gt;The old version was able to import over 14GB of data over one mapper and the import fails now when a mapper gets too many rows. I checked a heap dump and the memory was completely used by over 3.5 million rows of data (-Xmx 1700M).&lt;/P&gt;&lt;P&gt;The connector is mysql-jdbc version 5.1.33 and the job imports the data as text file in a have table.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Can I avoid this with a setting or is this a bug that should go to jira?&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you,&lt;/P&gt;&lt;P&gt;Jürgen&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 09:10:20 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20604#M3362</guid>
      <dc:creator>juergen_thomann</dc:creator>
      <dc:date>2022-09-16T09:10:20Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20720#M3363</link>
      <description>&lt;P&gt;This appears to be a regression caused by the fix in &lt;A target="_self" href="https://issues.apache.org/jira/browse/SQOOP-1400"&gt;SQOOP-1400&lt;/A&gt;. Instead of fetching results from MySQL row-by-row, sqoop is instead attempting to load the entire result set in memory.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We worked around it by upgrading to MySQL/J Connector 5.1.33 (which you're already on), and then including&amp;nbsp;"--fetch-size -2147483648" in&amp;nbsp;our sqoop command line options list. This restores the old row-by-row behaviour (the weird fetch size is a sentinel value recognised by the MySQL JDBC driver.)&lt;/P&gt;</description>
      <pubDate>Thu, 23 Oct 2014 03:26:48 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20720#M3363</guid>
      <dc:creator>sveiss</dc:creator>
      <dc:date>2014-10-23T03:26:48Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20732#M3364</link>
      <description>&lt;P&gt;Thanks for the answer!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I also found the workaround after some time, but you were faster to post it. I'll open a Jira for it, that it will be fixed in new versions.&lt;/P&gt;</description>
      <pubDate>Thu, 23 Oct 2014 07:33:03 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/20732#M3364</guid>
      <dc:creator>juergen_thomann</dc:creator>
      <dc:date>2014-10-23T07:33:03Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/77315#M3365</link>
      <description>&lt;P&gt;Use&amp;nbsp;?dontTrackOpenResources=true&amp;amp;defaultFetchSize=1000&amp;amp;useCursorFetch=true property in Mysql Connection string. It work without changing JVM parameter.&lt;/P&gt;</description>
      <pubDate>Sun, 22 Jul 2018 16:59:32 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/77315#M3365</guid>
      <dc:creator>Snd</dc:creator>
      <dc:date>2018-07-22T16:59:32Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/77316#M3366</link>
      <description>Thanks a lot! It worked forme also</description>
      <pubDate>Sun, 22 Jul 2018 17:05:50 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/77316#M3366</guid>
      <dc:creator>Radhe</dc:creator>
      <dc:date>2018-07-22T17:05:50Z</dc:date>
    </item>
    <item>
      <title>Re: Sqoop GC overhead limit exceeded after CDH5.2 update</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/87544#M3367</link>
      <description>&lt;BLOCKQUOTE&gt;&lt;HR /&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/28062"&gt;@Snd&lt;/a&gt;&amp;nbsp;wrote:&lt;BR /&gt;&lt;P&gt;Use&amp;nbsp;?dontTrackOpenResources=true&amp;amp;defaultFetchSize=1000&amp;amp;useCursorFetch=true property in Mysql Connection string. It work without changing JVM parameter.&lt;/P&gt;&lt;HR /&gt;&lt;/BLOCKQUOTE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thank you! it's worked!!!!&lt;/P&gt;</description>
      <pubDate>Sat, 09 Mar 2019 09:43:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Sqoop-GC-overhead-limit-exceeded-after-CDH5-2-update/m-p/87544#M3367</guid>
      <dc:creator>wing</dc:creator>
      <dc:date>2019-03-09T09:43:55Z</dc:date>
    </item>
  </channel>
</rss>

