Created 02-12-2016 02:38 PM
I have a 9-node (6 slaves, 2 masters and 1 edge node) cluster of HDP 2.3 with Ambari running. Currently only HDFS, YARN, Zookeeper, and Ambari metrics are running.
I'd like to install Spark. When I did an install of Spark 1.4.1 via Ambari, it installed a Spark history server on one node and spark client on 2 nodes. I don't see spark on the other nodes. Do I have to install Spark client on every node and set the master and slaves configuration and start spark manually?
I am not connected to the Internet and there are no proxy servers.
Created 02-12-2016 02:43 PM
If you want spark client in all the nodes then you can install it using ambari.
You can start spark thrift and History server from the ambari.
You don't have to do anyting if you are using ambari to manager spark.
Created 02-12-2016 02:43 PM
If you want spark client in all the nodes then you can install it using ambari.
You can start spark thrift and History server from the ambari.
You don't have to do anyting if you are using ambari to manager spark.
Created 02-12-2016 02:48 PM
Created 02-12-2016 03:02 PM
@Rahul Tikekar Good question..Please see this doc and http://spark.apache.org/docs/latest/running-on-yarn.html
When you submit spark job then you define master "the --master
parameter is yarn
."
In client
mode, the driver runs in the client process, and the application master is only used for requesting resources from YARN.
Created 02-12-2016 07:54 PM
Created 02-12-2016 07:59 PM
install Spark Standalone mode, you simply place a compiled version of Spark on each node on the cluster. @Rahul Tikekar
Created 02-12-2016 07:52 PM
Created 09-30-2016 12:17 AM
i think hdp spark does not support standalone cluster mode, only yarn mode. Am i right?
,i think so, hdp spark does not support standalone mode. right? only yarn mode.