I'm not developper, I'm admin for Hadoop plateform.
We have intalled HDP 2.4.2 so with SPARK 1.6.1 my questions concerning Versionning about Python and R.
All my servers are installed with Centos 6.8 with python 2.6.6 so it is possible to use PySpark ?
My developper said his wants python 2.7.X I don't know why. If I need to install Python 2.7 or 3 is need to install on all plateform or just in one datanode or master?
SparkR needs R to install, it is not ship with Spark ?