<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: KUDU Couldn't send request to peer  Status: Remote error: Service unavailable: UpdateConsensus request on kudu.consensus.ConsensusService from IP dropped due to backpressure. The service queue is full in Support Questions</title>
    <link>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/317189#M227052</link>
    <description>&lt;P&gt;Hi!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Those warning messages about dropped RPC requests due to backpressure is a sign that particular tablet server is likely overloaded.&amp;nbsp; Consider the following remedies:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Upgrade to the recent version of Kudu (1.14 as of now). Since Kudu 1.9.0 there have been many fixes which might help to reduce memory pressure for write-intensive workloads (e.g. see&amp;nbsp;&amp;nbsp;&lt;SPAN class="s1"&gt;&lt;SPAN class="s1"&gt;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2727" target="_self"&gt;KUDU-2727&lt;/A&gt;,&lt;/SPAN&gt;&lt;/SPAN&gt;&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2929" target="_self"&gt;&lt;SPAN class="s1"&gt;KUDU-2929&lt;/SPAN&gt;&lt;/A&gt;&lt;SPAN class="s1"&gt;), read-only workloads (&lt;A href="https://issues.apache.org/jira/browse/KUDU-2836" target="_self"&gt;KUDU-2836&lt;/A&gt;), and bunch of other improvements&lt;/SPAN&gt;&lt;SPAN&gt;. BTW, if you are using CDH, then upgrading to CDH6.3.4 is a good first step in that direction: CDH6.3.4 contains fixes for KUDU-2727, KUDU-2929, KUDU-2836 (those were back-ported into CDH6.3.4).&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;Make sure the tablet replica distribution is even across tablet servers: run the 'kudu cluster rebalance' CLI tool.&lt;/LI&gt;&lt;LI&gt;If you suspect replica hot-spotting, consider re-creating the table in question to fan out the write stream across multiple tablets.&amp;nbsp; I guess reading this guide might be useful:&amp;nbsp;&lt;A href="https://kudu.apache.org/docs/schema_design.html" target="_self"&gt;https://kudu.apache.org/docs/schema_design.html&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;If nothing from the above helps, consider adding a few more tablet server nodes into your cluster.&amp;nbsp; Once new nodes are added into the cluster, don't forget to run the 'kudu cluster rebalance' CLI tool.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Kind regards,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Alexey&lt;/P&gt;</description>
    <pubDate>Thu, 27 May 2021 02:36:21 GMT</pubDate>
    <dc:creator>Alexey1c</dc:creator>
    <dc:date>2021-05-27T02:36:21Z</dc:date>
    <item>
      <title>KUDU Couldn't send request to peer  Status: Remote error: Service unavailable: UpdateConsensus request on kudu.consensus.ConsensusService from IP dropped due to backpressure. The service queue is full</title>
      <link>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/316985#M226981</link>
      <description>&lt;P&gt;KUDU version:&amp;nbsp;&amp;nbsp;&lt;SPAN&gt;1.9.0+cdh6.2.0&lt;BR /&gt;1、memory_limit_hard_bytes: 100G&lt;BR /&gt;2、memory.soft_limit_in_bytes: -1&lt;BR /&gt;3、memory.limit_in_bytes: -1&lt;BR /&gt;4、maintenance_manager_num_threads: 4&lt;BR /&gt;5、block_cache_capacity_mb : 2G&lt;BR /&gt;&lt;BR /&gt;The cluster has 4 tablet-server，and three yarn nodemanager are in the same node with tablet-server.&lt;/SPAN&gt;&lt;/P&gt;&lt;P&gt;When I running a MR job in Yarn，just hive sql， and kudu tablet server will quit random, with fowlling logs:&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;LI-CODE lang="markup"&gt;T 5e5cdb8cf25d4c93aeaf013781419109 P ac586d8c49f84c4c82770ae079256893 -&amp;gt; Peer ac44fc76284d4b959eca897309e465b0 (ch4.360kad.com:7050): Couldn't send request to peer ac44fc76284d4b959eca897309e465b0. Status: Remote error: Service unavailable: UpdateConsensus request on kudu.consensus.ConsensusService from 10.0.57.16:26274 dropped due to backpressure. The service queue is full; it has 50 items.. This is attempt 1: this message will repeat every 5th retry.&lt;/LI-CODE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="QQ图片20210524115052.png" style="width: 783px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/31274i273285471FFAFC2F/image-size/large?v=v2&amp;amp;px=999" role="button" title="QQ图片20210524115052.png" alt="QQ图片20210524115052.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="QQ图片20210524114744.jpg" style="width: 999px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/31273i8FE478F691724EB5/image-size/large?v=v2&amp;amp;px=999" role="button" title="QQ图片20210524114744.jpg" alt="QQ图片20210524114744.jpg" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;I dont know how sole this problem&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Mon, 24 May 2021 03:52:14 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/316985#M226981</guid>
      <dc:creator>JiewenZ</dc:creator>
      <dc:date>2021-05-24T03:52:14Z</dc:date>
    </item>
    <item>
      <title>Re: KUDU Couldn't send request to peer  Status: Remote error: Service unavailable: UpdateConsensus request on kudu.consensus.ConsensusService from IP dropped due to backpressure. The service queue is full</title>
      <link>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/317189#M227052</link>
      <description>&lt;P&gt;Hi!&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Those warning messages about dropped RPC requests due to backpressure is a sign that particular tablet server is likely overloaded.&amp;nbsp; Consider the following remedies:&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;Upgrade to the recent version of Kudu (1.14 as of now). Since Kudu 1.9.0 there have been many fixes which might help to reduce memory pressure for write-intensive workloads (e.g. see&amp;nbsp;&amp;nbsp;&lt;SPAN class="s1"&gt;&lt;SPAN class="s1"&gt;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2727" target="_self"&gt;KUDU-2727&lt;/A&gt;,&lt;/SPAN&gt;&lt;/SPAN&gt;&amp;nbsp;&lt;A href="https://issues.apache.org/jira/browse/KUDU-2929" target="_self"&gt;&lt;SPAN class="s1"&gt;KUDU-2929&lt;/SPAN&gt;&lt;/A&gt;&lt;SPAN class="s1"&gt;), read-only workloads (&lt;A href="https://issues.apache.org/jira/browse/KUDU-2836" target="_self"&gt;KUDU-2836&lt;/A&gt;), and bunch of other improvements&lt;/SPAN&gt;&lt;SPAN&gt;. BTW, if you are using CDH, then upgrading to CDH6.3.4 is a good first step in that direction: CDH6.3.4 contains fixes for KUDU-2727, KUDU-2929, KUDU-2836 (those were back-ported into CDH6.3.4).&lt;/SPAN&gt;&lt;/LI&gt;&lt;LI&gt;Make sure the tablet replica distribution is even across tablet servers: run the 'kudu cluster rebalance' CLI tool.&lt;/LI&gt;&lt;LI&gt;If you suspect replica hot-spotting, consider re-creating the table in question to fan out the write stream across multiple tablets.&amp;nbsp; I guess reading this guide might be useful:&amp;nbsp;&lt;A href="https://kudu.apache.org/docs/schema_design.html" target="_self"&gt;https://kudu.apache.org/docs/schema_design.html&lt;/A&gt;&lt;/LI&gt;&lt;LI&gt;If nothing from the above helps, consider adding a few more tablet server nodes into your cluster.&amp;nbsp; Once new nodes are added into the cluster, don't forget to run the 'kudu cluster rebalance' CLI tool.&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Kind regards,&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Alexey&lt;/P&gt;</description>
      <pubDate>Thu, 27 May 2021 02:36:21 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/317189#M227052</guid>
      <dc:creator>Alexey1c</dc:creator>
      <dc:date>2021-05-27T02:36:21Z</dc:date>
    </item>
    <item>
      <title>Re: KUDU Couldn't send request to peer  Status: Remote error: Service unavailable: UpdateConsensus request on kudu.consensus.ConsensusService from IP dropped due to backpressure. The service queue is full</title>
      <link>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/317196#M227057</link>
      <description>&lt;P&gt;Thanks youe reply.&amp;nbsp; Unfortunately, an upgrade is not available at this time in my company.&lt;BR /&gt;I have rebalance my tablet server and modified the config 【&lt;SPAN&gt;maintenance_manager_num_threads&lt;/SPAN&gt;】to 8 , 【&lt;SPAN&gt;block_cache_capacity_mb&lt;/SPAN&gt;】 to 512MB, 【&lt;SPAN&gt;memory_limit_hard_bytes&lt;/SPAN&gt;】to 60G。&lt;/P&gt;&lt;P&gt;AND then I try to run MR job on yarn, when get map counts with 96, and memory with 194G on yarn, kudu server is&amp;nbsp;stable. So I continue to run a few job on yarn to to observe Kudu, and kudu server is still stable. So I think it is ok and&amp;nbsp;set up scheduling tasks.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="JiewenZ_0-1622094728876.png" style="width: 400px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/31329i94EE7CA200AA1C89/image-size/medium?v=v2&amp;amp;px=400" role="button" title="JiewenZ_0-1622094728876.png" alt="JiewenZ_0-1622094728876.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;But when today, a job run 179 maps, the kudu server is random quit...&lt;/P&gt;&lt;P&gt;this is the memory detail one of tablet server&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="JiewenZ_0-1622095461519.png" style="width: 400px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/31331iFE72045C2FA20DAF/image-size/medium?v=v2&amp;amp;px=400" role="button" title="JiewenZ_0-1622095461519.png" alt="JiewenZ_0-1622095461519.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="JiewenZ_1-1622095485923.png" style="width: 400px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/31332iD3D06E26C7CEF1D2/image-size/medium?v=v2&amp;amp;px=400" role="button" title="JiewenZ_1-1622095485923.png" alt="JiewenZ_1-1622095485923.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;</description>
      <pubDate>Thu, 27 May 2021 06:04:51 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Support-Questions/KUDU-Couldn-t-send-request-to-peer-Status-Remote-error/m-p/317196#M227057</guid>
      <dc:creator>JiewenZ</dc:creator>
      <dc:date>2021-05-27T06:04:51Z</dc:date>
    </item>
  </channel>
</rss>

