Support Questions
Find answers, ask questions, and share your expertise
Announcements
Alert: Welcome to the Unified Cloudera Community. Former HCC members be sure to read and learn how to activate your account here.

Getting OutOfMemoryError: GC overhead limit exceeded in production

Solved Go to solution
Highlighted

Getting OutOfMemoryError: GC overhead limit exceeded in production

Rising Star

My pig script is running fine while am using a low volume of dataset but its throwing OutOfMemoryError while running it on production with a large dataset.

Line 59661: 2017-01-09 13:04:41,610 FATAL [main] org.apache.hadoop.mapred.YarnChild: Error running child : java.lang.OutOfMemoryError: GC overhead limit exceeded Line 73580: 2017-01-09 12:57:44,043 FATAL [main] org.apache.hadoop.mapred.YarnChild: Error running child : java.lang.OutOfMemoryError: GC overhead limit exceeded Line 73580: 2017-01-09 12:57:44,043 FATAL [main] org.apache.hadoop.mapred.YarnChild: Error running child : java.lang.OutOfMemoryError: GC overhead limit exceeded Line 74087: Halting due to Out Of Memory Error... Line 87609: 2017-01-09 12:57:58,235 INFO [communication thread] org.apache.hadoop.mapred.Task: Communication exception: java.lang.OutOfMemoryError: GC overhead limit exceeded Line 87622: 2017-01-09 12:58:09,979 FATAL [IPC Client (378689909) connection to /166.37.225.35:40341 from job_1469684844014_1921354] org.apache.hadoop.yarn.YarnUncaughtExceptionHandler: Thread Thread[IPC Client (378689909) connection to /166.37.225.35:40341 from job_1469684844014_1921354,5,main] threw an Error. Shutting down now...

1 ACCEPTED SOLUTION

Accepted Solutions

Re: Getting OutOfMemoryError: GC overhead limit exceeded in production

Expert Contributor

Try to increase your memory. Edit:

export HADOOP_OPTS="-XX:NewRatio=12 -Xmx4096m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:+UseGCOverheadLimit -XX:+UseConcMarkSweepGC"

hive-env template in Advanced hive-env tab. The -Xmx****m parameter represents the amount of memory in MB.

1 REPLY 1

Re: Getting OutOfMemoryError: GC overhead limit exceeded in production

Expert Contributor

Try to increase your memory. Edit:

export HADOOP_OPTS="-XX:NewRatio=12 -Xmx4096m -XX:MaxHeapFreeRatio=40 -XX:MinHeapFreeRatio=15 -XX:+UseGCOverheadLimit -XX:+UseConcMarkSweepGC"

hive-env template in Advanced hive-env tab. The -Xmx****m parameter represents the amount of memory in MB.