<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question SPARK Throwing error while using pyspark on sql context in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290978#M215225</link>
    <description>&lt;P&gt;Welcome to&lt;BR /&gt;____ __&lt;BR /&gt;/ __/__ ___ _____/ /__&lt;BR /&gt;_\ \/ _ \/ _ `/ __/ '_/&lt;BR /&gt;/__ / .__/\_,_/_/ /_/\_\ version 2.3.0.2.6.5.0-292&lt;BR /&gt;/_/&lt;/P&gt;&lt;P&gt;Using Python version 2.7.14 (default, Dec 7 2017 17:05:42)&lt;BR /&gt;SparkSession available as 'spark'.&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt;&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; df=spark.sql('select * from sws_dev.vw_dlx_rpr_ordr_dtl_base limit 1').show()&lt;BR /&gt;[Stage 0:=====================&amp;gt; (18 + 28) / 46]20/03/03 07:01:08 ERROR DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/03/temp_shuffle_280c5065-f954-4ec8-b3d0-7c1f5c18b581&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/03/temp_shuffle_280c5065-f954-4ec8-b3d0-7c1f5c18b581 (Too many open files)&lt;BR /&gt;at java.io.FileOutputStream.open0(Native Method)&lt;BR /&gt;at java.io.FileOutputStream.open(FileOutputStream.java:270)&lt;BR /&gt;at java.io.FileOutputStream.&amp;lt;init&amp;gt;(FileOutputStream.java:213)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1386)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214)&lt;BR /&gt;at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)&lt;BR /&gt;at org.apache.spark.scheduler.Task.run(Task.scala:109)&lt;BR /&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)&lt;BR /&gt;at java.lang.Thread.run(Thread.java:745)&lt;BR /&gt;20/03/03 07:01:08 ERROR Executor: Exception in task 3.0 in stage 0.0 (TID 3)&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/3c/temp_shuffle_8450fcd1-d97c-4c34-ac52-196e03030bf9 (Too many open files)&lt;BR /&gt;at java.io.FileOutputStream.open0(Native Method)&lt;BR /&gt;at java.io.FileOutputStream.open(FileOutputStream.java:270)&lt;BR /&gt;at java.io.FileOutputStream.&amp;lt;init&amp;gt;(FileOutputStream.java:213)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237)&lt;BR /&gt;at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)&lt;BR /&gt;at org.apache.spark.scheduler.Task.run(Task.scala:109)&lt;BR /&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)&lt;BR /&gt;at java.lang.Thread.run(Thread.java:745)&lt;BR /&gt;20/03/03 07:01:08 ERROR Executor: Exception in task 9.0 in stage 0.0 (TID 9)&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/21/temp_shuffle_19e93f90-4de2-43c9-a715-c8668e96d793 (Too many open files)&lt;/P&gt;</description>
    <pubDate>Fri, 16 Sep 2022 14:35:34 GMT</pubDate>
    <dc:creator>Govins</dc:creator>
    <dc:date>2022-09-16T14:35:34Z</dc:date>
    <item>
      <title>SPARK Throwing error while using pyspark on sql context</title>
      <link>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290978#M215225</link>
      <description>&lt;P&gt;Welcome to&lt;BR /&gt;____ __&lt;BR /&gt;/ __/__ ___ _____/ /__&lt;BR /&gt;_\ \/ _ \/ _ `/ __/ '_/&lt;BR /&gt;/__ / .__/\_,_/_/ /_/\_\ version 2.3.0.2.6.5.0-292&lt;BR /&gt;/_/&lt;/P&gt;&lt;P&gt;Using Python version 2.7.14 (default, Dec 7 2017 17:05:42)&lt;BR /&gt;SparkSession available as 'spark'.&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt;&lt;BR /&gt;&amp;gt;&amp;gt;&amp;gt; df=spark.sql('select * from sws_dev.vw_dlx_rpr_ordr_dtl_base limit 1').show()&lt;BR /&gt;[Stage 0:=====================&amp;gt; (18 + 28) / 46]20/03/03 07:01:08 ERROR DiskBlockObjectWriter: Uncaught exception while reverting partial writes to file /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/03/temp_shuffle_280c5065-f954-4ec8-b3d0-7c1f5c18b581&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/03/temp_shuffle_280c5065-f954-4ec8-b3d0-7c1f5c18b581 (Too many open files)&lt;BR /&gt;at java.io.FileOutputStream.open0(Native Method)&lt;BR /&gt;at java.io.FileOutputStream.open(FileOutputStream.java:270)&lt;BR /&gt;at java.io.FileOutputStream.&amp;lt;init&amp;gt;(FileOutputStream.java:213)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter$$anonfun$revertPartialWritesAndClose$2.apply$mcV$sp(DiskBlockObjectWriter.scala:217)&lt;BR /&gt;at org.apache.spark.util.Utils$.tryWithSafeFinally(Utils.scala:1386)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.revertPartialWritesAndClose(DiskBlockObjectWriter.scala:214)&lt;BR /&gt;at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.stop(BypassMergeSortShuffleWriter.java:237)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:102)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)&lt;BR /&gt;at org.apache.spark.scheduler.Task.run(Task.scala:109)&lt;BR /&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)&lt;BR /&gt;at java.lang.Thread.run(Thread.java:745)&lt;BR /&gt;20/03/03 07:01:08 ERROR Executor: Exception in task 3.0 in stage 0.0 (TID 3)&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/3c/temp_shuffle_8450fcd1-d97c-4c34-ac52-196e03030bf9 (Too many open files)&lt;BR /&gt;at java.io.FileOutputStream.open0(Native Method)&lt;BR /&gt;at java.io.FileOutputStream.open(FileOutputStream.java:270)&lt;BR /&gt;at java.io.FileOutputStream.&amp;lt;init&amp;gt;(FileOutputStream.java:213)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.initialize(DiskBlockObjectWriter.scala:103)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.open(DiskBlockObjectWriter.scala:116)&lt;BR /&gt;at org.apache.spark.storage.DiskBlockObjectWriter.write(DiskBlockObjectWriter.scala:237)&lt;BR /&gt;at org.apache.spark.shuffle.sort.BypassMergeSortShuffleWriter.write(BypassMergeSortShuffleWriter.java:151)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:96)&lt;BR /&gt;at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:53)&lt;BR /&gt;at org.apache.spark.scheduler.Task.run(Task.scala:109)&lt;BR /&gt;at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:345)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1142)&lt;BR /&gt;at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:617)&lt;BR /&gt;at java.lang.Thread.run(Thread.java:745)&lt;BR /&gt;20/03/03 07:01:08 ERROR Executor: Exception in task 9.0 in stage 0.0 (TID 9)&lt;BR /&gt;java.io.FileNotFoundException: /tmp/blockmgr-c5bcbbe3-8da0-44a0-8025-1b183c81d532/21/temp_shuffle_19e93f90-4de2-43c9-a715-c8668e96d793 (Too many open files)&lt;/P&gt;</description>
      <pubDate>Fri, 16 Sep 2022 14:35:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290978#M215225</guid>
      <dc:creator>Govins</dc:creator>
      <dc:date>2022-09-16T14:35:34Z</dc:date>
    </item>
    <item>
      <title>Re: SPARK Throwing error while using pyspark on sql context</title>
      <link>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290979#M215226</link>
      <description>&lt;P&gt;Tried verbose mode and still finding this issues !!&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Wed, 04 Mar 2020 07:54:27 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290979#M215226</guid>
      <dc:creator>Govins</dc:creator>
      <dc:date>2020-03-04T07:54:27Z</dc:date>
    </item>
    <item>
      <title>Re: SPARK Throwing error while using pyspark on sql context</title>
      <link>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290980#M215227</link>
      <description>&lt;P&gt;Fixed:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;This is what i infered, while running spark the mode is made as client as you see below:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Parsed arguments:&lt;BR /&gt;master local[*]&lt;BR /&gt;deployMode null&lt;BR /&gt;executorMemory null&lt;BR /&gt;executorCores null&lt;BR /&gt;totalExecutorCores null&lt;BR /&gt;propertiesFile /usr/hdp/current/spark2-client/conf/spark-defaults.conf&lt;BR /&gt;driverMemory 4g&lt;BR /&gt;driverCores null&lt;BR /&gt;driverExtraClassPath null&lt;BR /&gt;driverExtraLibraryPath /usr/hdp/current/hadoop-client/lib/native:/usr/hdp/current/hadoop-client/lib/native/Linux-amd64-64&lt;BR /&gt;driverExtraJavaOptions null&lt;BR /&gt;supervise false&lt;BR /&gt;queue default&lt;BR /&gt;numExecutors null&lt;BR /&gt;files null&lt;BR /&gt;pyFiles null&lt;BR /&gt;archives null&lt;BR /&gt;mainClass null&lt;BR /&gt;primaryResource pyspark-shell&lt;BR /&gt;name PySparkShell&lt;BR /&gt;childArgs []&lt;BR /&gt;jars null&lt;BR /&gt;packages null&lt;BR /&gt;packagesExclusions null&lt;BR /&gt;repositories null&lt;BR /&gt;verbose true&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;When we use --master yarn this gets success !! .&lt;/P&gt;</description>
      <pubDate>Wed, 04 Mar 2020 07:55:47 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/SPARK-Throwing-error-while-using-pyspark-on-sql-context/m-p/290980#M215227</guid>
      <dc:creator>Govins</dc:creator>
      <dc:date>2020-03-04T07:55:47Z</dc:date>
    </item>
  </channel>
</rss>

