<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: PySpark - Error initializing SparkContext in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/PySpark-Error-initializing-SparkContext/m-p/51888#M23513</link>
    <description>&lt;P&gt;The problem seems to be with configuration rather than dependency, I am not sure what configuration is missing.&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here is my configuration &amp;nbsp;:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;spark-defaults.conf :&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;spark.authenticate=false&lt;BR /&gt;spark.dynamicAllocation.enabled=true&lt;BR /&gt;spark.dynamicAllocation.executorIdleTimeout=60&lt;BR /&gt;spark.dynamicAllocation.minExecutors=0&lt;BR /&gt;spark.dynamicAllocation.schedulerBacklogTimeout=1&lt;BR /&gt;spark.eventLog.dir=hdfs://dtest.turn.com:8020/user/spark/applicationHistory&lt;BR /&gt;spark.eventLog.enabled=true&lt;BR /&gt;spark.serializer=org.apache.spark.serializer.KryoSerializer&lt;BR /&gt;spark.shuffle.service.enabled=true&lt;BR /&gt;spark.shuffle.service.port=7337&lt;BR /&gt;spark.master=yarn&lt;BR /&gt;spark.yarn.jars=hdfs://dtest.turn.com:8020/user/spark/spark-2.1-bin-hadoop/*&lt;BR /&gt;spark.yarn.historyServer.address=&lt;A href="http://dtest.turn.com:18088" target="_blank"&gt;http://dtest.turn.com:18088&lt;/A&gt;&lt;BR /&gt;spark.driver.extraLibraryPath=/opt/cloudera/parcels/CDH-5.5.4-1.cdh5.5.4.p0.9/lib/hadoop/lib/native&lt;BR /&gt;spark.executor.extraLibraryPath=/opt/cloudera/parcels/CDH-5.5.4-1.cdh5.5.4.p0.9/lib/hadoop/lib/native&lt;BR /&gt;spark.yarn.am.extraLibraryPath=/opt/cloudera/parcels/CDH-5.5.4-1.cdh5.5.4.p0.9/lib/hadoop/lib/native&lt;BR /&gt;spark.yarn.config.gatewayPath=/opt/cloudera/parcels&lt;BR /&gt;spark.yarn.config.replacementPath={{HADOOP_COMMON_HOME}}/../../..&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
    <pubDate>Tue, 07 Mar 2017 18:54:31 GMT</pubDate>
    <dc:creator>Murthy</dc:creator>
    <dc:date>2017-03-07T18:54:31Z</dc:date>
  </channel>
</rss>

