- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Kryo serialization failed: Buffer overflow
- Labels:
-
Apache Spark
Created on
‎08-21-2019
12:12 AM
- last edited on
‎04-04-2022
07:33 AM
by
VidyaSargur
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I am getting the org.apache.spark.SparkException: Kryo serialization failed: Buffer overflow when I am execute the collect on 1 GB of RDD(for example : My1GBRDD.collect).
When I am execution the same thing on small Rdd(600MB), It will execute successfully. The problem with above 1GB RDD.
For more details please refer the following steps which I do.
1. Create RDD of input file.
2. mapToPair on the RDD.
3. groupByKey() on the RDD.
4. collectAsMap on the RDD.
On the 4th step I got the SparkException as follows,
org.apache.spark.SparkException: Kryo serialization failed: Buffer overflow. Available: 0, required: 37
Serialization trace:
otherElements (org.apache.spark.util.collection.CompactBuffer). To avoid this, increase spark.kryoserializer.buffer.max value.
at org.apache.spark.serializer.KryoSerializerInstance.serialize(KryoSerializer.scala:350)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:393)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: com.esotericsoftware.kryo.KryoException: Buffer overflow. Available: 0, required: 37
Created ‎09-10-2020 12:53 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi ,
How did you solve this issue , i have the same.
Created ‎04-04-2022 06:43 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Increase spark.kryoserializer.buffer.max property value value according to the required size , by default it is 64 MB.
Got same Exception, ran job by increasing the value and was able to run it properly.
