Support Questions

Find answers, ask questions, and share your expertise

Hive on Spark with Spark Dynamic Resource Allocation


In Hive configuration we have 
spark.dynamicAllocation.enabled = true, actually checked

spark.executor.cores = 6

Does dynamic override the set number of 6? It does not appear dynamic cores are being allocated.







Spark dynamic allocation allows spark to add and remove the executors dynamically based on your application's demand. But for each executor, it will still respect your settings (e.g. cores and RAM per executor) and use a unified and fixed spec. 


Hope it helps.

Master Guru
Dynamic Allocation [1] controls the number of parallel running executor
containers in YARN, not the number of CPU vcore resource allocated to a
single executor container.

[1] -
Take a Tour of the Community
Don't have an account?
Your experience may be limited. Sign in to explore more.