- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Basic Question: Install Spark on HDP 2.3 using Ambari
- Labels:
-
Apache Spark
Created ‎02-12-2016 02:38 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
I have a 9-node (6 slaves, 2 masters and 1 edge node) cluster of HDP 2.3 with Ambari running. Currently only HDFS, YARN, Zookeeper, and Ambari metrics are running.
I'd like to install Spark. When I did an install of Spark 1.4.1 via Ambari, it installed a Spark history server on one node and spark client on 2 nodes. I don't see spark on the other nodes. Do I have to install Spark client on every node and set the master and slaves configuration and start spark manually?
I am not connected to the Internet and there are no proxy servers.
Created ‎02-12-2016 02:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
If you want spark client in all the nodes then you can install it using ambari.
You can start spark thrift and History server from the ambari.
You don't have to do anyting if you are using ambari to manager spark.
Created ‎02-12-2016 02:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
If you want spark client in all the nodes then you can install it using ambari.
You can start spark thrift and History server from the ambari.
You don't have to do anyting if you are using ambari to manager spark.
Created ‎02-12-2016 02:48 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created ‎02-12-2016 03:02 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Rahul Tikekar Good question..Please see this doc and http://spark.apache.org/docs/latest/running-on-yarn.html
When you submit spark job then you define master "the --master
parameter is yarn
."
In client
mode, the driver runs in the client process, and the application master is only used for requesting resources from YARN.
Created ‎02-12-2016 07:54 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created ‎02-12-2016 07:59 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
install Spark Standalone mode, you simply place a compiled version of Spark on each node on the cluster. @Rahul Tikekar
Created ‎02-12-2016 07:52 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Created ‎09-30-2016 12:17 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
i think hdp spark does not support standalone cluster mode, only yarn mode. Am i right?
,i think so, hdp spark does not support standalone mode. right? only yarn mode.
