<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question PYSPARK with different python versions on yarn is failing with errors. in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136080#M18985</link>
    <description>&lt;P&gt;Hi, we have hdp 2.3.4 with python 2.6.6 installed on our cluster. PYSPARK works perfectly with 2.6.6 version. We have a use case to use pandas package and for that we need python3. So we have installed python 3.4 in a different location and updated the below variables in spark-env.sh&lt;/P&gt;&lt;PRE&gt;export PYSPARK_PYTHON=/opt/rh/rh-python34/root/usr/bin/python3.4 

export LD_LIBRARY_PATH=/opt/rh/rh-python34/root/usr/lib64/ 

export PYTHONHASHSEED=0&lt;/PRE&gt;&lt;P&gt;We are using below spark-submit command to submit the jobs.&lt;/P&gt;&lt;PRE&gt;/usr/hdp/current/spark-client/bin/spark-submit --master yarn-client --num-executors 10 --conf spark.executor.memory=1g --conf spark.yarn.queue=batch --conf spark.executor.extraLibraryPath=/opt/rh/rh-python34/root/usr/lib64  src/main/python/test.py&lt;/PRE&gt;&lt;P&gt;We are getting below error when executing the job. Can you please let me know if there is any other way of accomplishing this? &lt;/P&gt;&lt;PRE&gt;py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.
: org.apache.spark.SparkException: Job aborted due to stage failure: Task 78 in stage 4.0 failed 4 times, most recent failure: Lost task 78.3 in stage 4.0 (TID 1585, lxhdpwrktst006.lowes.com): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/worker.py", line 111, in main
    process()
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/worker.py", line 106, in process
    serializer.dump_stream(func(split_index, iterator), outfile)
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/serializers.py", line 133, in dump_stream
    for obj in iterator:
  File "/usr/hdp/2.3.4.0-3485/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 1704, in add_shuffle_key
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/rdd.py", line 74, in portable_hash
    raise Exception("Randomness of hash of string should be disabled via PYTHONHASHSEED")
Exception: Randomness of hash of string should be disabled via PYTHONHASHSEED
at org.apache.spark.api.python.PythonRunner$anon$1.read(PythonRDD.scala:166)
at org.apache.spark.api.python.PythonRunner$anon$1.&amp;lt;init&amp;gt;(PythonRDD.scala:207)
at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:125)
at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:70)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
at org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:342)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:88)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$failJobAndIndependentStages(DAGScheduler.scala:1283)
at org.apache.spark.scheduler.DAGScheduler$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
at org.apache.spark.scheduler.DAGScheduler$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270)
at org.apache.spark.scheduler.DAGScheduler$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGScheduler$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at scala.Option.foreach(Option.scala:236)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
at org.apache.spark.util.EventLoop$anon$1.run(EventLoop.scala:48)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1824)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1837)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1850)
at org.apache.spark.api.python.PythonRDD$.runJob(PythonRDD.scala:393)
at org.apache.spark.api.python.PythonRDD.runJob(PythonRDD.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379)
at py4j.Gateway.invoke(Gateway.java:259)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:207)
at java.lang.Thread.run(Thread.java:745)&lt;/PRE&gt;</description>
    <pubDate>Thu, 11 Feb 2016 02:38:39 GMT</pubDate>
    <dc:creator>goutham_koneru</dc:creator>
    <dc:date>2016-02-11T02:38:39Z</dc:date>
    <item>
      <title>PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136080#M18985</link>
      <description>&lt;P&gt;Hi, we have hdp 2.3.4 with python 2.6.6 installed on our cluster. PYSPARK works perfectly with 2.6.6 version. We have a use case to use pandas package and for that we need python3. So we have installed python 3.4 in a different location and updated the below variables in spark-env.sh&lt;/P&gt;&lt;PRE&gt;export PYSPARK_PYTHON=/opt/rh/rh-python34/root/usr/bin/python3.4 

export LD_LIBRARY_PATH=/opt/rh/rh-python34/root/usr/lib64/ 

export PYTHONHASHSEED=0&lt;/PRE&gt;&lt;P&gt;We are using below spark-submit command to submit the jobs.&lt;/P&gt;&lt;PRE&gt;/usr/hdp/current/spark-client/bin/spark-submit --master yarn-client --num-executors 10 --conf spark.executor.memory=1g --conf spark.yarn.queue=batch --conf spark.executor.extraLibraryPath=/opt/rh/rh-python34/root/usr/lib64  src/main/python/test.py&lt;/PRE&gt;&lt;P&gt;We are getting below error when executing the job. Can you please let me know if there is any other way of accomplishing this? &lt;/P&gt;&lt;PRE&gt;py4j.protocol.Py4JJavaError: An error occurred while calling z:org.apache.spark.api.python.PythonRDD.runJob.
: org.apache.spark.SparkException: Job aborted due to stage failure: Task 78 in stage 4.0 failed 4 times, most recent failure: Lost task 78.3 in stage 4.0 (TID 1585, lxhdpwrktst006.lowes.com): org.apache.spark.api.python.PythonException: Traceback (most recent call last):
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/worker.py", line 111, in main
    process()
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/worker.py", line 106, in process
    serializer.dump_stream(func(split_index, iterator), outfile)
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/serializers.py", line 133, in dump_stream
    for obj in iterator:
  File "/usr/hdp/2.3.4.0-3485/spark/python/lib/pyspark.zip/pyspark/rdd.py", line 1704, in add_shuffle_key
  File "/grid10/hadoop/yarn/log/usercache/hdpbatch/appcache/application_1453942498577_5382/container_e28_1453942498577_5382_01_000058/pyspark.zip/pyspark/rdd.py", line 74, in portable_hash
    raise Exception("Randomness of hash of string should be disabled via PYTHONHASHSEED")
Exception: Randomness of hash of string should be disabled via PYTHONHASHSEED
at org.apache.spark.api.python.PythonRunner$anon$1.read(PythonRDD.scala:166)
at org.apache.spark.api.python.PythonRunner$anon$1.&amp;lt;init&amp;gt;(PythonRDD.scala:207)
at org.apache.spark.api.python.PythonRunner.compute(PythonRDD.scala:125)
at org.apache.spark.api.python.PythonRDD.compute(PythonRDD.scala:70)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
at org.apache.spark.api.python.PairwiseRDD.compute(PythonRDD.scala:342)
at org.apache.spark.rdd.RDD.computeOrReadCheckpoint(RDD.scala:300)
at org.apache.spark.rdd.RDD.iterator(RDD.scala:264)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:73)
at org.apache.spark.scheduler.ShuffleMapTask.runTask(ShuffleMapTask.scala:41)
at org.apache.spark.scheduler.Task.run(Task.scala:88)
at org.apache.spark.executor.Executor$TaskRunner.run(Executor.scala:214)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615)
at java.lang.Thread.run(Thread.java:745)
Driver stacktrace:
at org.apache.spark.scheduler.DAGScheduler.org$apache$spark$scheduler$DAGScheduler$failJobAndIndependentStages(DAGScheduler.scala:1283)
at org.apache.spark.scheduler.DAGScheduler$anonfun$abortStage$1.apply(DAGScheduler.scala:1271)
at org.apache.spark.scheduler.DAGScheduler$anonfun$abortStage$1.apply(DAGScheduler.scala:1270)
at scala.collection.mutable.ResizableArray$class.foreach(ResizableArray.scala:59)
at scala.collection.mutable.ArrayBuffer.foreach(ArrayBuffer.scala:47)
at org.apache.spark.scheduler.DAGScheduler.abortStage(DAGScheduler.scala:1270)
at org.apache.spark.scheduler.DAGScheduler$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGScheduler$anonfun$handleTaskSetFailed$1.apply(DAGScheduler.scala:697)
at scala.Option.foreach(Option.scala:236)
at org.apache.spark.scheduler.DAGScheduler.handleTaskSetFailed(DAGScheduler.scala:697)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.doOnReceive(DAGScheduler.scala:1496)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1458)
at org.apache.spark.scheduler.DAGSchedulerEventProcessLoop.onReceive(DAGScheduler.scala:1447)
at org.apache.spark.util.EventLoop$anon$1.run(EventLoop.scala:48)
at org.apache.spark.scheduler.DAGScheduler.runJob(DAGScheduler.scala:567)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1824)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1837)
at org.apache.spark.SparkContext.runJob(SparkContext.scala:1850)
at org.apache.spark.api.python.PythonRDD$.runJob(PythonRDD.scala:393)
at org.apache.spark.api.python.PythonRDD.runJob(PythonRDD.scala)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:606)
at py4j.reflection.MethodInvoker.invoke(MethodInvoker.java:231)
at py4j.reflection.ReflectionEngine.invoke(ReflectionEngine.java:379)
at py4j.Gateway.invoke(Gateway.java:259)
at py4j.commands.AbstractCommand.invokeMethod(AbstractCommand.java:133)
at py4j.commands.CallCommand.execute(CallCommand.java:79)
at py4j.GatewayConnection.run(GatewayConnection.java:207)
at java.lang.Thread.run(Thread.java:745)&lt;/PRE&gt;</description>
      <pubDate>Thu, 11 Feb 2016 02:38:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136080#M18985</guid>
      <dc:creator>goutham_koneru</dc:creator>
      <dc:date>2016-02-11T02:38:39Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136081#M18986</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt;&lt;P&gt;did you install python3 and pandas on every host that runs Spark?&lt;/P&gt;</description>
      <pubDate>Thu, 11 Feb 2016 02:40:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136081#M18986</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-11T02:40:31Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136082#M18987</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/393/aervits.html" nodeid="393"&gt;@Artem Ervits&lt;/A&gt; Yes they are installed on all the nodes in cluster. &lt;/P&gt;</description>
      <pubDate>Thu, 11 Feb 2016 02:42:56 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136082#M18987</guid>
      <dc:creator>goutham_koneru</dc:creator>
      <dc:date>2016-02-11T02:42:56Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136083#M18988</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt; what about this? &lt;A href="http://stackoverflow.com/questions/30279783/apache-spark-how-to-use-pyspark-with-python-3" target="_blank"&gt;http://stackoverflow.com/questions/30279783/apache-spark-how-to-use-pyspark-with-python-3&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 11 Feb 2016 02:45:07 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136083#M18988</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-11T02:45:07Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136084#M18989</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/393/aervits.html" nodeid="393"&gt;@Artem Ervits&lt;/A&gt; PYTHON_SPARK variable is already set in spark-env.sh and is loaded by default now. I have also tried submitting the job by mentioning it with spark-submit but same error.&lt;/P&gt;&lt;PRE&gt;PYSPARK_PYTHON=/opt/rh/rh-python34/root/usr/bin/python3 /usr/hdp/current/spark-client/bin/spark-submit --master yarn-client --num-executors 120 --conf spark.executor.memory=1g --conf spark.yarn.queue=batch --conf spark.executor.extraLibraryPath=/opt/rh/rh-python34/root/usr/lib64  src/main/python/test.py&lt;/PRE&gt;</description>
      <pubDate>Thu, 11 Feb 2016 02:54:39 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136084#M18989</guid>
      <dc:creator>goutham_koneru</dc:creator>
      <dc:date>2016-02-11T02:54:39Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136085#M18990</link>
      <description>&lt;P&gt;Tagging experts &lt;A rel="user" href="https://community.cloudera.com/users/194/jniemiec.html" nodeid="194"&gt;@Joseph Niemiec&lt;/A&gt; and &lt;A rel="user" href="https://community.cloudera.com/users/332/vshukla.html" nodeid="332"&gt;@vshukla&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 11 Feb 2016 03:19:55 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136085#M18990</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-11T03:19:55Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136086#M18991</link>
      <description>&lt;P&gt;Hi &lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt;&lt;/P&gt;&lt;P&gt;The issue here is we need to pass PYTHONHASHSEED=0 to the executors as an environment variable.&lt;/P&gt;&lt;P&gt;One way to do that is to export SPARK_YARN_USER_ENV=PYTHONHASHSEED=0 and then invoke spark-submit or pyspark.&lt;/P&gt;&lt;P&gt;With this change, my pyspark repro that used to hit this error runs successfully.&lt;/P&gt;&lt;P&gt;export PYSPARK_PYTHON=/usr/local/bin/python3.3&lt;/P&gt;&lt;P&gt;export PYTHONHASHSEED=0&lt;/P&gt;&lt;P&gt;export SPARK_YARN_USER_ENV=PYTHONHASHSEED=0&lt;/P&gt;&lt;P&gt;bin/pyspark --master yarn-client --executor-memory 512m&lt;/P&gt;&lt;P&gt;n = sc.parallelize(range(1000)).map(str).countApproxDistinct()&lt;/P&gt;</description>
      <pubDate>Sun, 14 Feb 2016 15:35:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136086#M18991</guid>
      <dc:creator>rvenkatesh</dc:creator>
      <dc:date>2016-02-14T15:35:19Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136087#M18992</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/266/rvenkatesh.html" nodeid="266"&gt;@Ram Venkatesh&lt;/A&gt; Thank you it worked. This environment variable is mentioned in 1.0.0 guide but not in 1.5.2. Thank you for pointing it out.&lt;/P&gt;</description>
      <pubDate>Sun, 14 Feb 2016 22:02:02 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136087#M18992</guid>
      <dc:creator>goutham_koneru</dc:creator>
      <dc:date>2016-02-14T22:02:02Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136088#M18993</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/332/vshukla.html" nodeid="332"&gt;@vshukla&lt;/A&gt;&lt;P&gt; &lt;A rel="user" href="https://community.cloudera.com/users/528/rsriharsha.html" nodeid="528"&gt;@Ram Sriharsha&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Sun, 14 Feb 2016 22:11:34 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136088#M18993</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-14T22:11:34Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136089#M18994</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt; can you share a link where you found this?&lt;/P&gt;</description>
      <pubDate>Sat, 27 Feb 2016 01:16:46 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136089#M18994</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-27T01:16:46Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136090#M18995</link>
      <description>&lt;P&gt;&lt;A href="http://people.apache.org/~tdas/spark-1.0.0-rc11-docs/running-on-yarn.html" target="_blank"&gt;http://people.apache.org/~tdas/spark-1.0.0-rc11-docs/running-on-yarn.html&lt;/A&gt;&lt;/P&gt;</description>
      <pubDate>Sat, 27 Feb 2016 02:52:01 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136090#M18995</guid>
      <dc:creator>goutham_koneru</dc:creator>
      <dc:date>2016-02-27T02:52:01Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136091#M18996</link>
      <description>&lt;P&gt;&lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt; thank you very much!&lt;/P&gt;</description>
      <pubDate>Sat, 27 Feb 2016 02:53:26 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136091#M18996</guid>
      <dc:creator>aervits</dc:creator>
      <dc:date>2016-02-27T02:53:26Z</dc:date>
    </item>
    <item>
      <title>Re: PYSPARK with different python versions on yarn is failing with errors.</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136092#M18997</link>
      <description>&lt;A rel="user" href="https://community.cloudera.com/users/2371/gouthamkoneru.html" nodeid="2371"&gt;@Goutham Koneru&lt;/A&gt;&lt;P&gt; &lt;A rel="user" href="https://community.cloudera.com/users/393/aervits.html" nodeid="393"&gt;@Artem Ervit&lt;/A&gt; &lt;A rel="user" href="https://community.cloudera.com/users/266/rvenkatesh.html" nodeid="266"&gt;@Ram Venkatesh&lt;/A&gt;&lt;/P&gt;&lt;P&gt;Hi, i am trying to install python 3 too in my hdp 2.5.3 cluster? how does this affect the other copoennts other than spark? is this recommended to do in production? can i use anaconda instead?&lt;/P&gt;</description>
      <pubDate>Sun, 01 Apr 2018 22:13:51 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/PYSPARK-with-different-python-versions-on-yarn-is-failing/m-p/136092#M18997</guid>
      <dc:creator>pmj</dc:creator>
      <dc:date>2018-04-01T22:13:51Z</dc:date>
    </item>
  </channel>
</rss>

