Support Questions

Find answers, ask questions, and share your expertise
Celebrating as our community reaches 100,000 members! Thank you!

When Container is killed, Yarn/Spark doesn't give me a new container and my app is like a Zombie

New Contributor



I've a mistake with my app in cluster mode with Spark. My app have 2 containers : one driver and one executor.

When the executor container is killed (by an error or what else), my app doesn't give an other container or kill the current attempt to reset the job. The job is like a zombie, the driver doesn't see it and continue without any error.


I missed maybe something about the yarn conf or spark-submit's params.

Have you ever meet this situation ?


New Contributor

Tested without dynamic allocation, a new container is spawning.

Is there an option with dynamic allocation like


but should be


spark.streaming.dynamicAllocation.minExecutors=1 seems to not work