<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Spark Sql error : Unable to acquire 1048576 bytes of memory in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126637#M26988</link>
    <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/2126/brathi.html" nodeid="2126"&gt;@Bharat Rathi&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;I am not sure what version of Spark you are using but this sounds a lot like &lt;A href="https://issues.apache.org/jira/browse/SPARK-10309"&gt;SPARK-10309&lt;/A&gt; (a known issue in Spark 1.5). Notice that this is specifically related to Tungsten. You can try disabling Tungsten as sugested by Jit Ken Tan in the JIRA by the following: &lt;/P&gt;&lt;PRE&gt;sqlContext.setConf("spark.sql.tungsten.enabled", "false")&lt;/PRE&gt;</description>
    <pubDate>Tue, 03 May 2016 23:54:01 GMT</pubDate>
    <dc:creator>bwilson</dc:creator>
    <dc:date>2016-05-03T23:54:01Z</dc:date>
    <item>
      <title>Spark Sql error : Unable to acquire 1048576 bytes of memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126636#M26987</link>
      <description>&lt;P&gt;&lt;/P&gt;&lt;P&gt;Below is the stacktrace for error. Please suggest.&lt;/P&gt;&lt;P&gt;Driver stacktrace:
  at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$$failJobAndIndependentStages(DAGScheduler.scala:1283)
  at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
  at org.apache.spark.scheduler.DAGScheduler$$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
  at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
  at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
  at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270)
  at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
  at org.apache.spark.scheduler.DAGScheduler$$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
  at scala.Option.foreach(Option.scala:236)
  at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
  at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
  at org.apache.spark.util.EventLoop$$anon$1.run(EventLoop.scala:48)
  at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1824)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1837)
  at org.apache.spark.SparkContext.runJob(SparkContext.scala:1850)
  at org.apache.spark.sql.execution.SparkPlan.executeTake(SparkPlan.scala:215)
  at org.apache.spark.sql.execution.Limit.executeCollect(basicOperators.scala:207)
  at org.apache.spark.sql.hive.HiveContext$QueryExecution.stringResult(HiveContext.scala:591)
  at org.apache.spark.sql.hive.thriftserver.SparkSQLDriver.run(SparkSQLDriver.scala:63)
  at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.processCmd(SparkSQLCLIDriver.scala:308)
  at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:376)
  at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver$.main(SparkSQLCLIDriver.scala:226)
  at org.apache.spark.sql.hive.thriftserver.SparkSQLCLIDriver.main(SparkSQLCLIDriver.scala)
  at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
  at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
  at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
  at java.lang.reflect.Method.invoke(Method.java:498)
  at org.apache.spark.deploy.SparkSubmit$.org$apache$spark$deploy$SparkSubmit$$runMain(SparkSubmit.scala:685)
  at org.apache.spark.deploy.SparkSubmit$.doRunMain$1(SparkSubmit.scala:180)
  at org.apache.spark.deploy.SparkSubmit$.submit(SparkSubmit.scala:205)
  at org.apache.spark.deploy.SparkSubmit$.main(SparkSubmit.scala:120)
  at org.apache.spark.deploy.SparkSubmit.main(SparkSubmit.scala)
Caused by: java.io.IOException: Unable to acquire 1048576 bytes of memory
  at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.acquireNewPage(UnsafeExternalSorter.java:351)
  at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.&amp;lt;init&amp;gt;(UnsafeExternalSorter.java:138)
  at org.apache.spark.util.collection.unsafe.sort.UnsafeExternalSorter.create(UnsafeExternalSorter.java:106)
  at org.apache.spark.sql.execution.UnsafeExternalRowSorter.&amp;lt;init&amp;gt;(UnsafeExternalRowSorter.java:68)
  at org.apache.spark.sql.execution.TungstenSort.org$apache$spark$sql$execution$TungstenSort$$preparePartition$1(sort.scala:146)
  at org.apache.spark.sql.execution.TungstenSort$$anonfun$doExecute$3.apply(sort.scala:169)
  at org.apache.spark.sql.execution.TungstenSort$$anonfun$doExecute$3.apply(sort.scala:169)
  at org.apache.spark.rdd.MapPartitionsWithPreparationRDD.prepare(MapPartitionsWithPreparationRDD.scala:50)
  at org.apache.spark.rdd.ZippedPartitionsBaseRDD$$anonfun$tryPrepareParents$1.applyOrElse(ZippedPartitionsRDD.scala:83)
  at org.apache.spark.rdd.ZippedPartitionsBaseRDD$$anonfun$tryPrepareParents$1.applyOrElse(ZippedPartitionsRDD.scala:82)
  at scala.runtime.AbstractPartialFunction.apply(AbstractPartialFunction.scala:33)
  at scala.collection.TraversableLike$$anonfun$collect$1.apply(TraversableLike.scala:278)
  at scala.collection.immutable.List.foreach(List.scala:318)
  at scala.collection.TraversableLike$class.collect(TraversableLike.scala:278)
  at scala.collection.AbstractTraversable.collect(Traversable.scala:105)
  at org.apache.spark.rdd.ZippedPartitionsBaseRDD.tryPrepareParents(ZippedPartitionsRDD.scala:82)
  at org.apache.spark.rdd.ZippedPartitionsRDD2.compute(ZippedPartitionsRDD.scala:97)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
  at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
  at org.apache.spark.rdd.MapPartitionsWithPreparationRDD.compute(MapPartitionsWithPreparationRDD.scala:63)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
  at org.apache.spark.rdd.MapPartitionsRDD.compute(MapPartitionsRDD.scala:38)
  at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
  at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
  at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
  at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
  at org.apache.spark.scheduler.Task.run(Task.scala:88)
  at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
  at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)
  at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)
  at java.lang.Thread.run(Thread.java:745)&lt;/P&gt;</description>
      <pubDate>Tue, 03 May 2016 06:12:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126636#M26987</guid>
      <dc:creator>brathi</dc:creator>
      <dc:date>2016-05-03T06:12:21Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Sql error : Unable to acquire 1048576 bytes of memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126637#M26988</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/2126/brathi.html" nodeid="2126"&gt;@Bharat Rathi&lt;/A&gt;,&lt;/P&gt;&lt;P&gt;I am not sure what version of Spark you are using but this sounds a lot like &lt;A href="https://issues.apache.org/jira/browse/SPARK-10309"&gt;SPARK-10309&lt;/A&gt; (a known issue in Spark 1.5). Notice that this is specifically related to Tungsten. You can try disabling Tungsten as sugested by Jit Ken Tan in the JIRA by the following: &lt;/P&gt;&lt;PRE&gt;sqlContext.setConf("spark.sql.tungsten.enabled", "false")&lt;/PRE&gt;</description>
      <pubDate>Tue, 03 May 2016 23:54:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126637#M26988</guid>
      <dc:creator>bwilson</dc:creator>
      <dc:date>2016-05-03T23:54:01Z</dc:date>
    </item>
    <item>
      <title>Re: Spark Sql error : Unable to acquire 1048576 bytes of memory</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126638#M26989</link>
      <description>&lt;P&gt;Thanks &lt;A rel="user" href="https://community.cloudera.com/users/99/bwilson.html" nodeid="99"&gt;@Brandon Wilson&lt;/A&gt;, I got it to work by increasing the driver memory. in my case.&lt;/P&gt;</description>
      <pubDate>Thu, 05 May 2016 00:34:13 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Spark-Sql-error-Unable-to-acquire-1048576-bytes-of-memory/m-p/126638#M26989</guid>
      <dc:creator>brathi</dc:creator>
      <dc:date>2016-05-05T00:34:13Z</dc:date>
    </item>
  </channel>
</rss>

