Support Questions

Find answers, ask questions, and share your expertise

java.lang.OutOfMemoryError: GC overhead limit exceeded

avatar
Explorer

When we run the query in hive through beeline we got below error

Exception in thread "broadcast-exchange-0" java.lang.OutOfMemoryError: GC overhead limit exceeded at org.apache.spark.sql.execution.SparkPlan$$anon$1.next(SparkPlan.scala:273) at org.apache.spark.sql.execution.SparkPlan$$anon$1.next(SparkPlan.scala:269) at scala.collection.Iterator$class.foreach(Iterator.scala:893) at org.apache.spark.sql.execution.SparkPlan$$anon$1.foreach(SparkPlan.scala:269) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeCollect$1.apply(SparkPlan.scala:291) at org.apache.spark.sql.execution.SparkPlan$$anonfun$executeCollect$1.apply(SparkPlan.scala:290) at scala.collection.IndexedSeqOptimized$class.foreach(IndexedSeqOptimized.scala:33) at scala.collection.mutable.ArrayOps$ofRef.foreach(ArrayOps.scala:186) at org.apache.spark.sql.execution.SparkPlan.executeCollect(SparkPlan.scala:290) at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:75) at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1$$anonfun$apply$1.apply(BroadcastExchangeExec.scala:72) at org.apache.spark.sql.execution.SQLExecution$.withExecutionId(SQLExecution.scala:94) at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72) at org.apache.spark.sql.execution.exchange.BroadcastExchangeExec$$anonfun$relationFuture$1.apply(BroadcastExchangeExec.scala:72) at scala.concurrent.impl.Future$PromiseCompletingRunnable.liftedTree1$1(Future.scala:24) at scala.concurrent.impl.Future$PromiseCompletingRunnable.run(Future.scala:24) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617) at java.lang.Thread.run(Thread.java:745) 17/01/15 06:31:38 ERROR InsertIntoHadoopFsRelationCommand: Aborting job. org.apache.spark.SparkException: Exception thrown in awaitResult: at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:194) [12:03:22 PM] Zakariya Baduda: Caused by: java.util.concurrent.TimeoutException: Futures timed out after [300 seconds] at scala.concurrent.impl.Promise$DefaultPromise.ready(Promise.scala:219) at scala.concurrent.impl.Promise$DefaultPromise.result(Promise.scala:223) at scala.concurrent.Await$$anonfun$result$1.apply(package.scala:190) at scala.concurrent.BlockContext$DefaultBlockContext$.blockOn(BlockContext.scala:53) at scala.concurrent.Await$.result(package.scala:190) at org.apache.spark.util.ThreadUtils$.awaitResult(ThreadUtils.scala:190) ... 172 more 17/01/15 06:31:38 ERROR DefaultWriterContainer: Job job_201701150626_0000 abort

2 REPLIES 2

avatar
New Contributor

This is a relatively common error, usually caused by too many objects or large structures in memory. Try using -XX:-UseGCOverheadLimit or increasing your heap size.

avatar
Rising Star

Could you try to increase the Hiverserver2 heap size?