Created 03-02-2016 05:34 PM
How can we configure cluster to have spark separated from other echo system components
Created 03-02-2016 05:42 PM
You can logically segregate a cluster using Yarn node labels. http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_yarn_resource_mgt/content/ch_node_labels....
You can also choose different queues for Spark and Hive. It won't necessarily prevent tasks running on same nodes but at least they won't compete for resources.
Created 03-02-2016 05:42 PM
You can logically segregate a cluster using Yarn node labels. http://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.4.0/bk_yarn_resource_mgt/content/ch_node_labels....
You can also choose different queues for Spark and Hive. It won't necessarily prevent tasks running on same nodes but at least they won't compete for resources.
Created 03-02-2016 06:29 PM
Thanks, that helps.
At the same time, can you point me to Setting Up Time-Based Queue Capacity Change.
Created 03-02-2016 06:32 PM
Its good it you separate this as a new question. Right now there is no support for time-based queue capacity change.
However, we were able to run a cron based job that refreshes queues with manual changes to capacity scheduler. However, if you do this and someone either restarts RMs and/or refreshes queues from ambari, your cron based changes will be overwritten.
Created 03-02-2016 08:24 PM
@kjilla if this is a satisfactory answer, please accept the answer.