- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
spark.yarn.executor.memoryOverhead
- Labels:
-
Apache Spark
-
Apache YARN
-
Cloudera Manager
Created on ‎09-14-2017 03:59 AM - edited ‎09-16-2022 05:14 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Got below error
17/09/12 20:41:36 WARN yarn.YarnAllocator: Container killed by YARN for exceeding memory limits. 1.5 GB of 1.5 GB physical memory used. Consider boosting spark.yarn.executor.memoryOverhead.
17/09/12 20:41:39 ERROR cluster.YarnClusterScheduler: Lost executor 1 on xyz.com: remote Akka client disassociated
Please help as not able to find spark.executor.memory or spark.yarn.executor.memoryOverhead in Cloudera Manager (Cloudera Enterprise 5.4.7)
Created on ‎09-14-2017 06:40 PM - edited ‎09-14-2017 06:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
spark.executor.memory can be found in Cloudera Manager under Hive->configuration and search for Java Heap.
Spark Executor Maximum Java Heap Size
spark.executor.memory
HiveServer2 Default Group
256 MiB
Spark Driver Maximum Java Heap Size
spark.driver.memory
HiveServer2 Default Group
256 MiB
Created ‎09-14-2017 11:22 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
These can be set per job as well. Spark-submit --executor-memory
https://spark.apache.org/docs/1.6.0/submitting-applications.html
Created on ‎09-14-2017 06:40 PM - edited ‎09-14-2017 06:43 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
spark.executor.memory can be found in Cloudera Manager under Hive->configuration and search for Java Heap.
Spark Executor Maximum Java Heap Size
spark.executor.memory
HiveServer2 Default Group
256 MiB
Spark Driver Maximum Java Heap Size
spark.driver.memory
HiveServer2 Default Group
256 MiB
Created on ‎09-15-2017 04:35 AM - edited ‎09-15-2017 04:37 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Thank you.
Additional query, do you know why these spark configs are placed under hive?
Created ‎09-16-2017 10:27 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
It's a spark side configuraion. So you can always specify it via "--conf" option with spark-submit, or you can set the property globally on CM via "Spark Client Advanced Configuration Snippet (Safety Valve) for spark-conf/spark-defaults.conf", so CM will include such setting for you via spark gateway client configuration.
Created ‎01-05-2020 06:33 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi,
This parameter spark.executor.memory (or) spark.yarn.executor.memoryOverhead can be set in Spark submit command or you can set it Advanced configurations.
Thanks
AKR
