<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spak structured streaming job failed in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/329927#M230546</link>
    <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/92709"&gt;@EBH&lt;/a&gt;,&amp;nbsp;Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future.&lt;/P&gt;</description>
    <pubDate>Wed, 10 Nov 2021 06:48:57 GMT</pubDate>
    <dc:creator>VidyaSargur</dc:creator>
    <dc:date>2021-11-10T06:48:57Z</dc:date>
    <item>
      <title>Spak structured streaming job failed</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/328960#M230365</link>
      <description>&lt;P&gt;Hello all,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;We have an application Spark Structured Streaming (version 3.0.1.3.0.7110.0-81) which crashed after 1hours and 15 minutes.&lt;BR /&gt;This application read data from input kafka topic, run some transfomation and write the result in an another output kafka topic.&lt;/P&gt;&lt;P&gt;The application use the mode "foreachBatch" with a trigger interval set to 30 secondes and maxOffsetsPerTrigger set to 7000.&lt;/P&gt;&lt;P&gt;We get this exception :&lt;BR /&gt;java.lang.OutOfMemoryError: Java heap space&lt;BR /&gt;at java.util.Arrays.copyOf(Arrays.java:3332)&lt;BR /&gt;at java.lang.AbstractStringBuilder.ensureCapacityInternal(AbstractStringBuilder.java:124)&lt;BR /&gt;at java.lang.AbstractStringBuilder.append(AbstractStringBuilder.java:448)&lt;BR /&gt;at java.lang.StringBuilder.append(StringBuilder.java:136)&lt;BR /&gt;at scala.collection.mutable.StringBuilder.append(StringBuilder.scala:203)&lt;BR /&gt;at scala.collection.TraversableOnce.$anonfun$addString$1(TraversableOnce.scala:369)&lt;BR /&gt;at scala.collection.TraversableOnce$$Lambda$90/910599202.apply(Unknown Source)&lt;BR /&gt;at scala.collection.immutable.List.foreach(List.scala:392)&lt;BR /&gt;at scala.collection.TraversableOnce.addString(TraversableOnce.scala:362)&lt;BR /&gt;at scala.collection.TraversableOnce.addString$(TraversableOnce.scala:358)&lt;BR /&gt;at scala.collection.AbstractTraversable.addString(Traversable.scala:108)&lt;BR /&gt;at scala.collection.TraversableOnce.mkString(TraversableOnce.scala:328)&lt;BR /&gt;at scala.collection.TraversableOnce.mkString$(TraversableOnce.scala:327)&lt;BR /&gt;at scala.collection.AbstractTraversable.mkString(Traversable.scala:108)&lt;BR /&gt;at scala.collection.TraversableOnce.mkString(TraversableOnce.scala:330)&lt;BR /&gt;at scala.collection.TraversableOnce.mkString$(TraversableOnce.scala:330)&lt;BR /&gt;at scala.collection.AbstractTraversable.mkString(Traversable.scala:108)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.codegen.CodegenContext.declareAddedFunctions(CodeGenerator.scala:541)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.codegen.GenerateSafeProjection$.create(GenerateSafeProjection.scala:197)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.codegen.GenerateSafeProjection$.create(GenerateSafeProjection.scala:39)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.codegen.CodeGenerator.generate(CodeGenerator.scala:1278)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.SafeProjection$.createCodeGeneratedObject(Projection.scala:170)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.SafeProjection$.createCodeGeneratedObject(Projection.scala:167)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.CodeGeneratorWithInterpretedFallback.createObject(CodeGeneratorWithInterpretedFallback.scala:52)&lt;BR /&gt;at org.apache.spark.sql.catalyst.expressions.SafeProjection$.create(Projection.scala:193)&lt;BR /&gt;at org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$Deserializer.apply(ExpressionEncoder.scala:180)&lt;BR /&gt;at org.apache.spark.sql.catalyst.encoders.ExpressionEncoder$Deserializer.apply(ExpressionEncoder.scala:173)&lt;BR /&gt;at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)&lt;BR /&gt;at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)&lt;BR /&gt;at scala.collection.Iterator$$anon$10.next(Iterator.scala:459)&lt;BR /&gt;at scala.collection.Iterator$$anon$12.hasNext(Iterator.scala:512)&lt;BR /&gt;at scala.collection.Iterator$$anon$10.hasNext(Iterator.scala:458)&lt;/P&gt;&lt;P&gt;We suspect a memory leak and we can see in the heap dump many objects of type java.util.concurrent.ConcurrentHashMap.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Thanks for your help / ideas.&lt;/P&gt;&lt;P&gt;Laurent&lt;/P&gt;</description>
      <pubDate>Wed, 27 Oct 2021 09:19:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/328960#M230365</guid>
      <dc:creator>EBH</dc:creator>
      <dc:date>2021-10-27T09:19:12Z</dc:date>
    </item>
    <item>
      <title>Re: Spak structured streaming job failed</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/329048#M230379</link>
      <description>&lt;P&gt;Hi&amp;nbsp;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/92709"&gt;@EBH&lt;/a&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Spark application is failed with OOM error. To understand why OOM error, we need to go through Spark event logs, application logs and spark submit command. Currently you are not shared any of the logs.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Next step is, try to increase the &lt;STRONG&gt;executor/driver memory&lt;/STRONG&gt; and set &lt;STRONG&gt;memory overhead&lt;/STRONG&gt; value as 0.1 or 0.2 % of driver/executor memory. If still issue is not resolved please raise an cloudera case we will work on this issue.&lt;/P&gt;</description>
      <pubDate>Thu, 28 Oct 2021 06:10:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/329048#M230379</guid>
      <dc:creator>RangaReddy</dc:creator>
      <dc:date>2021-10-28T06:10:12Z</dc:date>
    </item>
    <item>
      <title>Re: Spak structured streaming job failed</title>
      <link>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/329927#M230546</link>
      <description>&lt;P&gt;&lt;a href="https://community.cloudera.com/t5/user/viewprofilepage/user-id/92709"&gt;@EBH&lt;/a&gt;,&amp;nbsp;Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future.&lt;/P&gt;</description>
      <pubDate>Wed, 10 Nov 2021 06:48:57 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/Spak-structured-streaming-job-failed/m-p/329927#M230546</guid>
      <dc:creator>VidyaSargur</dc:creator>
      <dc:date>2021-11-10T06:48:57Z</dc:date>
    </item>
  </channel>
</rss>

