Support Questions

Find answers, ask questions, and share your expertise
Announcements
Celebrating as our community reaches 100,000 members! Thank you!

Who agreed with this topic

Spark app throwing java.lang.OutOfMemoryError: GC overhead limit exceeded

avatar
Expert Contributor

I got a 40 node cdh 5.1 cluster and attempting to run a simple spark app that processes about 10-15GB raw data but I keep running into this error:

 

  java.lang.OutOfMemoryError: GC overhead limit exceeded

 

Each node has 8 cores and 2GB memory. I notice the heap size on the executors is set to 512MB with total set to 2GB. Wanted to know whats the heap size needs to be set to for such data sizes?

 

Thanks for the input!

Who agreed with this topic