Member since
03-10-2018
5
Posts
0
Kudos Received
0
Solutions
05-30-2018
01:34 AM
Hi @Paul Hernandez First of all, my problem has solved by adding zeppelin properties like @Felix Albani show me. In my case, my cluster is based on CentOS 7. The OS has python 2.7 as default and some packages such as yum have dependency on the default python. The symlink '/bin/python' is heading this default python and if it is changed, yum is not working any more. Hope this help. SW
... View more
05-29-2018
06:11 AM
Hi @Paul Hernandez Thank you for your comment. I checked the post you told me and found it is not a good idea: changing symlink in bin/. It might make trouble to linux system.
... View more
05-29-2018
06:07 AM
@Sandeep Nemuri. Thank you very much. It helped a lot.
... View more
05-22-2018
05:37 AM
Hi. I have a question regarding yarn application. As shown in the snapshot, hundreds of undefined yarn applications are created by dr.who every time. I never created that kind of jobs and I think they are created automatically by somehow but I don't know why. Could you please advise that this happens in normal or not? If not, how can i fix it?
... View more
Labels:
- Labels:
-
Apache YARN
04-27-2018
11:11 AM
Hi. I built a cluster with HDP ambari Version
2.6.1.5 and I am using anaconda3 as my python interpreter. I have a problem of changing or alter python version for Spark2 pyspark in zeppelin When I check python version of Spark2 by pyspark, it shows as bellow which means OK to me. ____ __
/ __/__ ___ _____/ /__
_\ \/ _ \/ _ `/ __/ '_/
/__ / .__/\_,_/_/ /_/\_\ version 2.2.0.2.6.4.0-91
/_/
Using Python version 3.6.4 (default, Jan 16 2018 18:10:19)
SparkSession available as 'spark'.
>>> import sys
>>> print (sys.path)
['', '/tmp/spark-14a0fb52-5fea-4c1f-bf6b-c0bd0c37eedf/userFiles-54205d05-fbf0-4ec1-b274-4c5a2b78e840', '/usr/hdp/current/spark2-client/python/lib/py4j-0.10.4-src.zip', '/usr/hdp/current/spark2-client/python', '/root', '/root/anaconda3/lib/python36.zip', '/root/anaconda3/lib/python3.6', '/root/anaconda3/lib/python3.6/lib-dynload', '/root/anaconda3/lib/python3.6/site-packages']
>>> print (sys.version)
3.6.4 |Anaconda, Inc.| (default, Jan 16 2018, 18:10:19)
[GCC 7.2.0]
>>> exit()
When I check python version of Spark2 by zeppelin, it shows different results as below. %spark2.pyspark
print(sc.version)
import sys
print(sys.version)
print()
print(sys.path)
2.2.0.2.6.4.0-91
2.7.5 (default, Aug 4 2017, 00:39:18)
[GCC 4.8.5 20150623 (Red Hat 4.8.5-16)]
()
['/tmp', u'/tmp/spark-75f5d1d5-fefa-4dc8-bc9b-c797dec106d7/userFiles-1c25cf01-7758-49dd-a1eb-f1fbd084e9af/py4j-0.10.4-src.zip', u'/tmp/spark-75f5d1d5-fefa-4dc8-bc9b-c797dec106d7/userFiles-1c25cf01-7758-49dd-a1eb-f1fbd084e9af/pyspark.zip', u'/tmp/spark-75f5d1d5-fefa-4dc8-bc9b-c797dec106d7/userFiles-1c25cf01-7758-49dd-a1eb-f1fbd084e9af', '/usr/hdp/current/spark2-client/python/lib/py4j-0.10.4-src.zip', '/usr/hdp/current/spark2-client/python', '/usr/hdp/current/spark2-client/python/lib/py4j-0.8.2.1-src.zip', '/usr/lib64/python27.zip', '/usr/lib64/python2.7', '/usr/lib64/python2.7/plat-linux2', '/usr/lib64/python2.7/lib-tk', '/usr/lib64/python2.7/lib-old', '/usr/lib64/python2.7/lib-dynload', '/usr/lib64/python2.7/site-packages', '/usr/lib64/python2.7/site-packages/gtk-2.0', '/usr/lib/python2.7/site-packages']
I have tried to update zeppelin interpreter setting known by other questions and answers such as export PYSPARK_PYTHON = /root/anaconda3/bin/python I updated both zeppelin.env.sh and interpreter setting via zeppelin GUI but it didn't work. I think it cause because zeppelin's python path is heading /usr/lib64/python2.7 which is base for centos but I don't know how to fix it. If there is any idea of this problem, please let me know. Any of your advice would be appreciated. Thank you.
... View more
Labels:
- Labels:
-
Apache Spark