<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question YARN Node Labels - Effective Capacity is 0% on labeled partition and infinity% on DEFAULT_PARTITION in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/YARN-Node-Labels-Effective-Capacity-is-0-on-labeled/m-p/375154#M242308</link>
    <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm new to using labels on YARN nodes. I have successfully set up the labels but the scheduler is allocating all resources to the DEFAULT_PARTITION under "Effective Capacity" and 0 resources to the labeled partition.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="yarn_scheduler.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38228i5FCAF2F3A5E974E6/image-size/large?v=v2&amp;amp;px=999" role="button" title="yarn_scheduler.jpg" alt="yarn_scheduler.jpg" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="node_labels.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38227iD746BA1CC9717909/image-size/large?v=v2&amp;amp;px=999" role="button" title="node_labels.jpg" alt="node_labels.jpg" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="yarn_app.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38226iEBCCFBA655F666A5/image-size/large?v=v2&amp;amp;px=999" role="button" title="yarn_app.jpg" alt="yarn_app.jpg" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As the screenshots illustrate, the NodeManager is launching with the correct label and has the correct resources assigned to that label, however, applications will not start when assigned to that label because although the partition has resources assigned to it, the queue under the partition does not. Here's my capacity-scheduler.xml:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;&amp;lt;configuration&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.maximum-am-resource-percent&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;1.0&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.resource-calculator&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.queues&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;default,spark&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
   &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-node-label-expression&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
&amp;lt;/configuration&amp;gt;&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;And here is the relevant parts of yarn-site.xml:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.enabled&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.configuration-type&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;distributed&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.fs-store.root-dir&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;hdfs://xxx:9000/user/yarn/node-labels/&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.nodemanager.node-labels.provider&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;config&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.nodemanager.node-labels.provider.configured-node-partition&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm using Hadoop 3.3.4 built from source. In case it matters, this is in my dev environment with a single ResourceManager and NodeManager. Any suggestions are much appreciated. Thanks!&lt;/P&gt;</description>
    <pubDate>Fri, 11 Aug 2023 15:30:09 GMT</pubDate>
    <dc:creator>aarongrubb</dc:creator>
    <dc:date>2023-08-11T15:30:09Z</dc:date>
    <item>
      <title>YARN Node Labels - Effective Capacity is 0% on labeled partition and infinity% on DEFAULT_PARTITION</title>
      <link>https://community.cloudera.com/t5/Support-Questions/YARN-Node-Labels-Effective-Capacity-is-0-on-labeled/m-p/375154#M242308</link>
      <description>&lt;P&gt;Hello,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm new to using labels on YARN nodes. I have successfully set up the labels but the scheduler is allocating all resources to the DEFAULT_PARTITION under "Effective Capacity" and 0 resources to the labeled partition.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="yarn_scheduler.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38228i5FCAF2F3A5E974E6/image-size/large?v=v2&amp;amp;px=999" role="button" title="yarn_scheduler.jpg" alt="yarn_scheduler.jpg" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="node_labels.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38227iD746BA1CC9717909/image-size/large?v=v2&amp;amp;px=999" role="button" title="node_labels.jpg" alt="node_labels.jpg" /&gt;&lt;/span&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="yarn_app.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/38226iEBCCFBA655F666A5/image-size/large?v=v2&amp;amp;px=999" role="button" title="yarn_app.jpg" alt="yarn_app.jpg" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;As the screenshots illustrate, the NodeManager is launching with the correct label and has the correct resources assigned to that label, however, applications will not start when assigned to that label because although the partition has resources assigned to it, the queue under the partition does not. Here's my capacity-scheduler.xml:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;&amp;lt;configuration&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.maximum-am-resource-percent&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;1.0&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.resource-calculator&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.queues&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;default,spark&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
   &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-node-label-expression&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
&amp;lt;/configuration&amp;gt;&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;And here is the relevant parts of yarn-site.xml:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.enabled&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.configuration-type&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;distributed&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.node-labels.fs-store.root-dir&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;hdfs://xxx:9000/user/yarn/node-labels/&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.nodemanager.node-labels.provider&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;config&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.nodemanager.node-labels.provider.configured-node-partition&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;I'm using Hadoop 3.3.4 built from source. In case it matters, this is in my dev environment with a single ResourceManager and NodeManager. Any suggestions are much appreciated. Thanks!&lt;/P&gt;</description>
      <pubDate>Fri, 11 Aug 2023 15:30:09 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/YARN-Node-Labels-Effective-Capacity-is-0-on-labeled/m-p/375154#M242308</guid>
      <dc:creator>aarongrubb</dc:creator>
      <dc:date>2023-08-11T15:30:09Z</dc:date>
    </item>
    <item>
      <title>Re: YARN Node Labels - Effective Capacity is 0% on labeled partition and infinity% on DEFAULT_PARTITION</title>
      <link>https://community.cloudera.com/t5/Support-Questions/YARN-Node-Labels-Effective-Capacity-is-0-on-labeled/m-p/375361#M242420</link>
      <description>&lt;P&gt;This was caused by me overlooking "root" as an actual queue and not giving it the proper permissions for label and capacity to pass on to the child queues. The configuration in the writeup here tipped me off:&amp;nbsp;&lt;A href="https://www.ibm.com/support/pages/yarn-node-labels-label-based-scheduling-and-resource-isolation-hadoop-dev" target="_self"&gt;https://www.ibm.com/support/pages/yarn-node-labels-label-based-scheduling-and-resource-isolation-hadoop-dev&lt;/A&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Here is the full configuration that gives me the desired behaviour:&lt;BR /&gt;&lt;BR /&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;&amp;lt;configuration&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.maximum-am-resource-percent&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;1.0&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.resource-calculator&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;org.apache.hadoop.yarn.util.resource.DefaultResourceCalculator&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;*&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;100&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;100&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;100&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;100&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.queues&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;default,spark&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-node-label-expression&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=11776,vcores=4]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.default.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;node&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.accessible-node-labels.node.maximum-capacity&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;[memory=4096,vcores=1]&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.default-application-priority&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;9&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
    &amp;lt;property&amp;gt;
        &amp;lt;name&amp;gt;yarn.scheduler.capacity.root.spark.disable_preemption&amp;lt;/name&amp;gt;
        &amp;lt;value&amp;gt;true&amp;lt;/value&amp;gt;
    &amp;lt;/property&amp;gt;
&amp;lt;/configuration&amp;gt;&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Fri, 18 Aug 2023 12:04:29 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/YARN-Node-Labels-Effective-Capacity-is-0-on-labeled/m-p/375361#M242420</guid>
      <dc:creator>aarongrubb</dc:creator>
      <dc:date>2023-08-18T12:04:29Z</dc:date>
    </item>
  </channel>
</rss>

