Support Questions

Find answers, ask questions, and share your expertise
Announcements
Celebrating as our community reaches 100,000 members! Thank you!

How do I tell which version ofSpark I am running?

avatar
Expert Contributor

I have finally got HDP 2.3 to run! Now I am trying to install zeppelin via these instructions:

http://hortonworks.com/hadoop-tutorial/apache-zeppelin/

but I need to know which version of Spark I am running. How do I find this in HDP?

TIA!

1 ACCEPTED SOLUTION

avatar
Super Collaborator

7361-versioninfo.png

@ed day You can get the version information from Ambari UI. Click on Admin -> Stack and Versions and you will find the version information under Version tab.

View solution in original post

3 REPLIES 3

avatar
Super Collaborator

7361-versioninfo.png

@ed day You can get the version information from Ambari UI. Click on Admin -> Stack and Versions and you will find the version information under Version tab.

avatar
Rising Star

The most easy way just launch "spark-shell" at the command line. This will give you the active version running on your cluster:

[root@xxxxxxx ~]# spark-shell 16/09/05 17:15:15 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable Welcome to ____ __ / __/__ ___ _____/ /__ _\ \/ _ \/ _ `/ __/ '_/ /___/ .__/\_,_/_/ /_/\_\ version 1.3.1 /_/

Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_71) Type in expressions to have them evaluated. Type :help for more information.

[root@xxxxxxx ~]# spark-shell
16/09/05 17:15:15 WARN NativeCodeLoader: Unable to load native-hadoop library for your platform... using builtin-java classes where applicable
Welcome to
      ____              __
     / __/__  ___ _____/ /__
    _\ \/ _ \/ _ `/ __/  '_/
   /___/ .__/\_,_/_/ /_/\_\   version 1.3.1
      /_/

Using Scala version 2.10.4 (OpenJDK 64-Bit Server VM, Java 1.7.0_71)
Type in expressions to have them evaluated.
Type :help for more information.

avatar
New Contributor

sc.version()

or

spark -submit --version