<?xml version="1.0" encoding="UTF-8"?>
<rss xmlns:content="http://purl.org/rss/1.0/modules/content/" xmlns:dc="http://purl.org/dc/elements/1.1/" xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#" xmlns:taxo="http://purl.org/rss/1.0/modules/taxonomy/" version="2.0">
  <channel>
    <title>question Re: Data Nodes displaying incorrect block report in Archives of Support Questions (Read Only)</title>
    <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205681#M78880</link>
    <description>&lt;P&gt;&lt;EM&gt;&lt;A href="https://community.hortonworks.com/questions/194026/@Samant%20Thakur"&gt;@Samant Thakur&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Did you go through the links I posted above? &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Your data node and Namenode heap sizes need some tuning.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Are you seeing any data node high HEAP SIZE alert?&lt;/EM&gt;&lt;/P&gt;&lt;H3&gt;&lt;EM&gt;Estimating NameNode Heap Memory Needed&lt;A href="https://www.cloudera.com/documentation/enterprise/5-8-x/topics/admin_nn_memory_config.html#concept_oyk_bdq_mv__nn_heap_example2"&gt;&lt;/A&gt;&lt;/EM&gt;&lt;/H3&gt;&lt;P&gt;&lt;EM&gt;Memory is estimated by considering the capacity of a cluster. Values are rounded. The below cluster physically stores 4800 TB, or approximately 36 million block files (at the default block size). Replication determines how many namespace blocks represent these block files.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;At capacity, with the recommended allocation of 1 GB of memory per million blocks, The Cluster needs 12 GB of maximum heap space. &lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;200 hosts of 24 TB each = 4800 TB.&lt;/I&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;EM&gt;Blocksize=128 MB, &lt;STRONG&gt;Replication=3&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in MB: 200 * 24,000,000 MB = 4,800,000,000 MB (4800 TB)&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Disk space needed per block: 128 MB per block * 3 = &lt;STRONG&gt;384 MB&lt;/STRONG&gt; storage per block&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in blocks: 4,800,000,000 MB / 384 MB = &lt;STRONG&gt;12,000,000 blocks&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;At capacity, with the recommended allocation of 1 GB of memory per million blocks, This cluster  will need &lt;STRONG&gt;12 GB&lt;/STRONG&gt; of maximum heap space.&lt;BR /&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Hope that helps please revert!&lt;/EM&gt;&lt;/P&gt;</description>
    <pubDate>Thu, 31 May 2018 15:00:44 GMT</pubDate>
    <dc:creator>Shelton</dc:creator>
    <dc:date>2018-05-31T15:00:44Z</dc:date>
    <item>
      <title>Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205676#M78875</link>
      <description>&lt;P&gt;I am getting a strange issue with 3 out of 8 data nodes in our HDP 2.6.0 cluster. These 3 data nodes are not reporting the correct number of blocks and also not sending the block reports to name node on regular intervals.&lt;/P&gt;&lt;P&gt;Ambari reporting :&lt;/P&gt;&lt;PRE&gt;[Alert][datanode_storage] Unable to extract JSON from JMX response
&lt;/PRE&gt;&lt;P&gt;Any suggestion what is wrong with our cluster?&lt;/P&gt;&lt;P&gt;Thanks in advance for your assistance.&lt;/P&gt;&lt;BR /&gt;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/6876i2BA2825B297DD696/image-size/large?v=1.0&amp;amp;px=999" border="0" alt="namenode-ui.png" title="namenode-ui.png" /&gt;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/6878iE95E73AB044D24E1/image-size/large?v=1.0&amp;amp;px=999" border="0" alt="datanode-ui.png" title="datanode-ui.png" /&gt;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/6879i079F7C6DB7F9253D/image-size/large?v=1.0&amp;amp;px=999" border="0" alt="data-node-jmx.png" title="data-node-jmx.png" /&gt;</description>
      <pubDate>Wed, 30 May 2018 02:39:50 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205676#M78875</guid>
      <dc:creator>samant_thakur</dc:creator>
      <dc:date>2018-05-30T02:39:50Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205677#M78876</link>
      <description>&lt;P&gt;&lt;EM&gt;&lt;A href="@Samant Thakur"&gt; @Samant Thakur&lt;/A&gt;&lt;BR /&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;The JMX response typically indicates  3 things why the datanode was not accessible. &lt;/EM&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;EM&gt;Network Issue&lt;BR /&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;DataNode is down&lt;BR /&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Excessively long Garbage Collection &lt;BR /&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;This message comes from &lt;STRONG&gt;"/usr/lib/python2.6/site-packages/ambari_agent/alerts/metric_alert.py" &lt;/STRONG&gt;script and following is the logic:&lt;/EM&gt;&lt;/P&gt;&lt;PRE&gt; if isinstance(self.metric_info, JmxMetric):
      jmx_property_values, http_code = self._load_jmx(alert_uri.is_ssl_enabled, host, port, self.metric_info)
      if not jmx_property_values and http_code in [200, 307]:
        collect_result = self.RESULT_UNKNOWN
        value_list.append('HTTP {0} response (metrics unavailable)'.format(str(http_code)))
      elif not jmx_property_values and http_code not in [200, 307]:
        raise Exception("[Alert][{0}] Unable to extract JSON from JMX response".format(self.get_name()))
      else:
        value_list.extend(jmx_property_values)
        check_value = self.metric_info.calculate(value_list)
        value_list.append(check_value)&lt;/PRE&gt;&lt;P&gt;&lt;EM&gt;&lt;STRONG&gt;Network &lt;/STRONG&gt;i&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;MTU (Maximum Transmission Unit) is related to TCP/IP networking in Linux. It refers to the size (in bytes) of the largest datagram that a given layer of a communications protocol can pass at a time. It should be identical on all the nodes.MTU is set in &lt;STRONG&gt;/etc/sysconfig/network-scripts/ifcfg-ethx &lt;/STRONG&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;You can see current MTU setting with&lt;STRONG&gt; ifconfig&lt;/STRONG&gt; command under Linux: &lt;/EM&gt;&lt;/P&gt;&lt;PRE&gt;$ netstat -i &lt;/PRE&gt;&lt;P&gt;&lt;EM&gt;check the second row or &lt;/EM&gt;&lt;/P&gt;&lt;PRE&gt;$ ip link list &lt;/PRE&gt;&lt;P&gt;&lt;EM&gt;- Check the host file on those failing nodes &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;- Check if DNS server is having problems in name resolution. &lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;&lt;STRONG&gt;Run TestDFSIO performance tests &lt;/STRONG&gt;&lt;/I&gt;&lt;/P&gt;&lt;PRE&gt;yarn jar /usr/hdp/2.x.x.x.x/hadoop-mapreduce/hadoop-mapreduce-client-jobclient-*tests.jar TestDFSIO -write -nrFiles 100 -fileSize 100 TestDFSIO Read Test hadoop jar &lt;/PRE&gt;&lt;P&gt;&lt;I&gt;&lt;STRONG&gt;Iperf &lt;/STRONG&gt;&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;Is a widely used tool for network performance measurement and tuning &lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;See &lt;A href="https://community.hortonworks.com/articles/8563/typical-hdp-cluster-network-configuration-best-pra.html" target="_blank"&gt; Typical HDP Cluster Network Configuration Best Practices&lt;/A&gt; &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;&lt;STRONG&gt;Datanode is down &lt;/STRONG&gt;&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;Restart the datanode using Ambari or manually&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;&lt;STRONG&gt;Garbage Collection&lt;/STRONG&gt;&lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Running the &lt;A href="https://github.com/chewiebug/GCViewer"&gt;GCViewer&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;&lt;EM&gt;Enable GC logging for Datanode service&lt;/EM&gt;&lt;/STRONG&gt;&lt;/P&gt;&lt;OL&gt;
&lt;LI&gt;&lt;EM&gt;Open hadoop-env.sh and look for the following line&lt;/EM&gt;&lt;PRE&gt;export HADOOP_DATANODE_OPTS="-Dcom.sun.management.jmxremote -Xms2048m -Xmx2048m -Dhadoop.security.logger=ERROR,DRFAS $HADOOP_DATANODE_OPTS"&lt;/PRE&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;Insert the following into HADOOP_DATANODE_OPTS param&lt;/EM&gt;&lt;PRE&gt;-verbose:gc
-XX:+PrintGCDetails
-Xloggc:${HADOOP_LOG_DIR}/hadoop-hdfs-datanode-`date +'%Y%m%d%H%M'`.gclog
-XX:+UseGCLogFileRotation
-XX:NumberOfGCLogFiles=20&lt;/PRE&gt;
&lt;/LI&gt;
&lt;LI&gt;&lt;EM&gt;After adding the GC log pram the HADOOP_DATANODE_OPTS should look like this&lt;/EM&gt;&lt;PRE&gt;export HADOOP_DATANODE_OPTS="-Dcom.sun.management.jmxremote -Xms2048m -Xmx2048m -Dhadoop.security.logger=ERROR,DRFAS -verbose:gc -XX:+PrintGCDetails -Xloggc:${HADOOP_LOG_DIR}/hadoop-hdfs-datanode-`date +'%Y%m%d%H%M'`.gclog -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=20 $HADOOP_DATANODE_OPTS"&lt;/PRE&gt;&lt;/LI&gt;&lt;/OL&gt;&lt;P&gt;&lt;EM&gt;The log should give you a detailed info.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Hope that helps&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 30 May 2018 04:50:58 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205677#M78876</guid>
      <dc:creator>Shelton</dc:creator>
      <dc:date>2018-05-30T04:50:58Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205678#M78877</link>
      <description>&lt;P&gt;Thank you so much @&lt;A href="https://community.hortonworks.com/users/1271/sheltong.html"&gt;Geoffrey Shelton Okot&lt;/A&gt; for assistance on this. I really appreciate it.&lt;/P&gt;&lt;P&gt;1. MTU setting is same for all our data nodes. I have verified it.&lt;/P&gt;&lt;P&gt;2. I have performed testdfsio test .Pls see the attachment for test results.&lt;/P&gt;&lt;P&gt;3. Enable GC debugging.my hadoop-env template looks like below.&lt;/P&gt;&lt;P&gt;export HADOOP_DATANODE_OPTS="-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/hadoop/$USER/hs_err_pid%p.log -XX:NewSize=800m -XX:MaxNewSize=800m -XX:PermSize=128m -XX:MaxPermSize=256m -Xloggc:/var/log/hadoop/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms{{dtnode_heapsize}} -Xmx{{dtnode_heapsize}} -Dhadoop.security.logger=ERROR,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_DATANODE_OPTS} -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseParNewGC"&lt;/P&gt;&lt;P&gt;After enabling GC debugging &amp;amp; restarting name nodes and data nodes.Below alarm disappeared&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Unable to extract JSON from JMX response error &lt;/STRONG&gt;&lt;/P&gt;&lt;P&gt;But now ,I am getting below error now on problematic data node in&lt;STRONG&gt; &lt;/STRONG&gt;hadoop-hdfs-datanode-.log&lt;/P&gt;&lt;P&gt;2018-05-30 19:53:32,985 WARN  datanode.DataNode (BPServiceActor.java:offerService(673)) - IOException in offerService
java.io.EOFException: End of File Exception between local host is: "datanodehost/"; destination host is: "Namenodehost":8020; : java.io.EOFException; For more details see:  &lt;A href="http://wiki.apache.org/hadoop/EOFException" target="_blank"&gt;http://wiki.apache.org/hadoop/EOFException&lt;/A&gt;
        at sun.reflect.GeneratedConstructorAccessor15.newInstance(Unknown Source)
        at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
        at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:801)
        at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:765)
        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1558)
        at org.apache.hadoop.ipc.Client.call(Client.java:1498)
        at org.apache.hadoop.ipc.Client.call(Client.java:1398)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
        at com.sun.proxy.$Proxy15.blockReport(Unknown Source)
        at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB.blockReport(DatanodeProtocolClientSideTranslatorPB.java:211)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.blockReport(BPServiceActor.java:374)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:645)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:785)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.EOFExceptionat java.io.DataInputStream.readInt(DataInputStream.java:392)
        at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1119)
        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1014)
2018-05-30 19:53:33,100 INFO  datanode.DataNode (DataXceiver.java:writeBlock(669)) - Receiving BP-1033621575--1507285615620:blk_1461467777_387788610 src: /:42658 dest: /:50010
2018-05-30 19:53:33,878 INFO  datanode.DataNode (DataXceiver.java:writeBlock(669)) - Receiving BP-1033621575--1507285615620:blk_1461467782_387788615 src: /:43782 dest: /:50010
2018-05-30 19:53:36,197 INFO  datanode.DataNode (DataXceiver.java:writeBlock(669)) - Receiving BP-1033621575--1507285615620:blk_1368137451_294431710 src: /:52176 dest: /:50010&lt;/P&gt;&lt;P&gt;GC.log&lt;/P&gt;&lt;P&gt;9239114K(31375360K), 0.0954324 secs] [Times: user=0.75 sys=0.00, real=0.10 secs]
2018-05-30T20:37:23.000+0200: 15180.545: [GC (Allocation Failure) 2018-05-30T20:37:23.000+0200: 15180.545: [ParNew: 733378K-&amp;gt;81919K(737280K), 0.0994234 secs] 9892898K-&amp;gt;9739137K(31375360K), 0.0996623 secs] [Times: user=0.78 sys=0.01, real=0.10 secs]
2018-05-30T20:37:29.962+0200: 15187.508: [GC (Allocation Failure) 2018-05-30T20:37:29.963+0200: 15187.508: [ParNew: 727808K-&amp;gt;81689K(737280K), 0.1043798 secs] 10385026K-&amp;gt;10379938K(31375360K), 0.1046235 secs] [Times: user=0.83 sys=0.00, real=0.11 secs]
2018-05-30T20:37:33.884+0200: 15191.430: [GC (Allocation Failure) 2018-05-30T20:37:33.885+0200: 15191.430: [ParNew: 733664K-&amp;gt;81919K(737280K), 0.1201577 secs] 11031913K-&amp;gt;10881691K(31375360K), 0.1203890 secs] [Times: user=0.95 sys=0.00, real=0.12 secs]
2018-05-30T20:37:41.029+0200: 15198.574: [GC (Allocation Failure) 2018-05-30T20:37:41.029+0200: 15198.575: [ParNew: 727734K-&amp;gt;78326K(737280K), 0.1015139 secs] 11527506K-&amp;gt;11522912K(31375360K), 0.1017500 secs] [Times: user=0.81 sys=0.00, real=0.10 secs]
2018-05-30T20:37:44.780+0200: 15202.325: [GC (Allocation Failure) 2018-05-30T20:37:44.780+0200: 15202.325: [ParNew: 730789K-&amp;gt;81920K(737280K), 0.0937630 secs] 12175374K-&amp;gt;12020024K(31375360K), 0.0939903 secs] [Times: user=0.74 sys=0.00, real=0.09 secs]
2018-05-30T20:37:51.818+0200: 15209.363: [GC (Allocation Failure) 2018-05-30T20:37:51.818+0200: 15209.363: [ParNew: 723037K-&amp;gt;78409K(737280K), 0.1089323 secs] 12661141K-&amp;gt;12638859K(31375360K), 0.1091735 secs] [Times: user=0.87 sys=0.01, real=0.11 secs]
2018-05-30T20:37:55.071+0200: 15212.616: [GC (Allocation Failure) 2018-05-30T20:37:55.071+0200: 15212.616: [ParNew: 733424K-&amp;gt;81919K(737280K), 0.0912281 secs] 13293874K-&amp;gt;13139143K(31375360K), 0.0914462 secs] [Times: user=0.72 sys=0.00, real=0.09 secs]
2018-05-30T20:38:02.582+0200: 15220.127: [GC (Allocation Failure) 2018-05-30T20:38:02.582+0200: 15220.127: [ParNew: 731000K-&amp;gt;80436K(737280K), 0.1039197 secs] 13788224K-&amp;gt;13781232K(31375360K), 0.1041447 secs] [Times: user=0.82 sys=0.00, real=0.10 secs]
2018-05-30T20:38:05.811+0200: 15223.356: [GC (Allocation Failure) 2018-05-30T20:38:05.811+0200: 15223.356: [ParNew: 734976K-&amp;gt;81919K(737280K), 0.0843448 secs] 14435772K-&amp;gt;14285826K(31375360K), 0.0845672 secs] [Times: user=0.67 sys=0.00, real=0.09 secs]
2018-05-30T20:38:13.249+0200: 15230.794: [GC (Allocation Failure) 2018-05-30T20:38:13.249+0200: 15230.794: [ParNew: 725770K-&amp;gt;80833K(737280K), 0.0967994 secs] 14929677K-&amp;gt;14924119K(31375360K), 0.0970191 secs] [Times: user=0.76 sys=0.00, real=0.10 secs]
2018-05-30T20:38:16.685+0200: 15234.231: [GC (Allocation Failure) 2018-05-30T20:38:16.686+0200: 15234.231: [ParNew: 735203K-&amp;gt;81920K(737280K), 0.0984436 secs] 15578489K-&amp;gt;15419615K(31375360K), 0.0986753 secs] [Times: user=0.78 sys=0.00, real=0.10 secs]
2018-05-30T20:38:24.385+0200: 15241.930: [GC (Allocation Failure) 2018-05-30T20:38:24.385+0200: 15241.930: [ParNew: 735008K-&amp;gt;79750K(737280K), 0.0981608 secs] 16072704K-&amp;gt;16066284K(31375360K), 0.0983850 secs] [Times: user=0.78 sys=0.00, real=0.09 secs]
2018-05-30T20:38:27.513+0200: 15245.058: [GC (Allocation Failure) 2018-05-30T20:38:27.513+0200: 15245.058: [ParNew: 731825K-&amp;gt;81920K(737280K), 0.0928862 secs] 16718359K-&amp;gt;16566812K(31375360K), 0.0931079 secs] [Times: user=0.73 sys=0.00, real=0.10 secs]
2018-05-30T20:38:35.118+0200: 15252.664: [GC (Allocation Failure) 2018-05-30T20:38:35.119+0200: 15252.664: [ParNew: 728589K-&amp;gt;81823K(737280K), 0.1155139 secs] 17213482K-&amp;gt;17208899K(31375360K), 0.1157287 secs] [Times: user=0.91 sys=0.01, real=0.11 secs]
2018-05-30T20:38:39.004+0200: 15256.549: [GC (Allocation Failure) 2018-05-30T20:38:39.004+0200: 15256.549: [ParNew: 735843K-&amp;gt;81920K(737280K), 0.0939004 secs] 17862919K-&amp;gt;17682067K(31375360K), 0.0941023 secs] [Times: user=0.74 sys=0.00, real=0.10 secs]
2018-05-30T20:38:46.888+0200: 15264.433: [GC (Allocation Failure) 2018-05-30T20:38:46.888+0200: 15264.433: [ParNew: 730708K-&amp;gt;78583K(737280K), 0.0952740 secs] 18330855K-&amp;gt;18343737K(31375360K), 0.0954785 secs] [Times: user=0.75 sys=0.01, real=0.09 secs]&lt;/P&gt;&lt;P&gt;Issue still persists with data nodes.3 out of 8 data nodes are reporting very less number of blocks&lt;/P&gt;&lt;P&gt;Please assist.&lt;/P&gt;&lt;P&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 31 May 2018 01:50:13 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205678#M78877</guid>
      <dc:creator>samant_thakur</dc:creator>
      <dc:date>2018-05-31T01:50:13Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205679#M78878</link>
      <description>&lt;P&gt;&lt;EM&gt;&lt;A href="https://community.hortonworks.com/questions/194026/@Samant%20Thakur"&gt;@Samant Thakur&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;There is definitely a network problem with the 3 nodes are they same hardware? NIC and network? &lt;/EM&gt;&lt;/P&gt;&lt;PRE&gt;&lt;EM&gt;Average IO rate mb/sec: 27.063095092773438 
Average IO rate mb/sec: 19.786481857299805&lt;/EM&gt;&lt;/PRE&gt;&lt;P&gt;&lt;EM&gt; "&lt;STRONG&gt;Allocation Failure&lt;/STRONG&gt;" is a cause of GC cycle to kick. &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;A GC allocation failure means that the garbage collector could not move objects from young gen to old gen fast enough because it does not have enough memory in old gen. This can cause application slowness. &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;&lt;STRONG&gt;Whats your DataNode maximum Java heap size&lt;/STRONG&gt;? &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.hortonworks.com/questions/64677/datanode-heapsize-computation.html"&gt;&lt;EM&gt;&lt;/EM&gt;&lt;/A&gt;&lt;A href="https://community.hortonworks.com/questions/64677/datanode-heapsize-computation.html" target="_blank"&gt;https://community.hortonworks.com/questions/64677/datanode-heapsize-computation.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.hortonworks.com/questions/45381/do-i-need-to-tune-java-heap-size.html"&gt;&lt;EM&gt;&lt;/EM&gt;&lt;/A&gt;&lt;A href="https://community.hortonworks.com/questions/45381/do-i-need-to-tune-java-heap-size.html" target="_blank"&gt;https://community.hortonworks.com/questions/45381/do-i-need-to-tune-java-heap-size.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;A href="https://community.hortonworks.com/questions/78981/data-node-heap-size-warning.html"&gt;&lt;EM&gt;&lt;/EM&gt;&lt;/A&gt;&lt;A href="https://community.hortonworks.com/questions/78981/data-node-heap-size-warning.html" target="_blank"&gt;https://community.hortonworks.com/questions/78981/data-node-heap-size-warning.html&lt;/A&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Do you have a NameNode HA component? if so it may be that failover has occurred -but the client doesn't detect this and retry its operation. &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;If it isn't a production cluster can you restart all the components?&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 31 May 2018 03:27:10 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205679#M78878</guid>
      <dc:creator>Shelton</dc:creator>
      <dc:date>2018-05-31T03:27:10Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205680#M78879</link>
      <description>&lt;P&gt;Thank you! I really appreciate your time and efforts.&lt;/P&gt;&lt;P&gt;1. Data node heap size is 30 GB.My worry is that why only 3 nodes are giving the issue not others if something is wrong with configuration. what is should be ideal heap size for data nodes do you have any idea?  I did not find any formula to calculate the  heap size for data nodes.&lt;/P&gt;&lt;P&gt;2. We are using name node HA. I suspect that HA switch over might have caused this problem.I have restarted all the components.what should I check for if issue is caused by name node HA.?Name node heap size is 75 GB ..used 70%.&lt;/P&gt;</description>
      <pubDate>Thu, 31 May 2018 12:38:31 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205680#M78879</guid>
      <dc:creator>samant_thakur</dc:creator>
      <dc:date>2018-05-31T12:38:31Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205681#M78880</link>
      <description>&lt;P&gt;&lt;EM&gt;&lt;A href="https://community.hortonworks.com/questions/194026/@Samant%20Thakur"&gt;@Samant Thakur&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Did you go through the links I posted above? &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Your data node and Namenode heap sizes need some tuning.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Are you seeing any data node high HEAP SIZE alert?&lt;/EM&gt;&lt;/P&gt;&lt;H3&gt;&lt;EM&gt;Estimating NameNode Heap Memory Needed&lt;A href="https://www.cloudera.com/documentation/enterprise/5-8-x/topics/admin_nn_memory_config.html#concept_oyk_bdq_mv__nn_heap_example2"&gt;&lt;/A&gt;&lt;/EM&gt;&lt;/H3&gt;&lt;P&gt;&lt;EM&gt;Memory is estimated by considering the capacity of a cluster. Values are rounded. The below cluster physically stores 4800 TB, or approximately 36 million block files (at the default block size). Replication determines how many namespace blocks represent these block files.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;At capacity, with the recommended allocation of 1 GB of memory per million blocks, The Cluster needs 12 GB of maximum heap space. &lt;/I&gt;&lt;/P&gt;&lt;P&gt;&lt;I&gt;200 hosts of 24 TB each = 4800 TB.&lt;/I&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;EM&gt;Blocksize=128 MB, &lt;STRONG&gt;Replication=3&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in MB: 200 * 24,000,000 MB = 4,800,000,000 MB (4800 TB)&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Disk space needed per block: 128 MB per block * 3 = &lt;STRONG&gt;384 MB&lt;/STRONG&gt; storage per block&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in blocks: 4,800,000,000 MB / 384 MB = &lt;STRONG&gt;12,000,000 blocks&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;At capacity, with the recommended allocation of 1 GB of memory per million blocks, This cluster  will need &lt;STRONG&gt;12 GB&lt;/STRONG&gt; of maximum heap space.&lt;BR /&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Hope that helps please revert!&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Thu, 31 May 2018 15:00:44 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205681#M78880</guid>
      <dc:creator>Shelton</dc:creator>
      <dc:date>2018-05-31T15:00:44Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205682#M78881</link>
      <description>&lt;P&gt;@&lt;A href="https://community.hortonworks.com/users/1271/sheltong.html" rel="nofollow noopener noreferrer" target="_blank"&gt;Geoffrey Shelton Okot&lt;/A&gt; &lt;/P&gt;&lt;P&gt;Yes, I have been through the post mentioned by you. We had data nodes failure issues in past ,increase heap size fixed it but I will fine tune them. Below is heap utilization for data node (max heap 30 GB). High heap usage data nodes (marked in red) are the problematic ones.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="76545-heap-size.png" style="width: 1214px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/17078i98A420DE12B2600E/image-size/medium?v=v2&amp;amp;px=400" role="button" title="76545-heap-size.png" alt="76545-heap-size.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;&lt;/P&gt;&lt;P&gt;&lt;STRONG&gt;Hadoop env&lt;/STRONG&gt;&lt;/P&gt;&lt;PRE&gt;      SHARED_HADOOP_NAMENODE_OPTS="-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile={{hdfs_log_dir_prefix}}/$USER/hs_err_pid%p.log -XX:NewSize={{namenode_opt_newsize}} -XX:MaxNewSize={{namenode_opt_maxnewsize}} -Xloggc:{{hdfs_log_dir_prefix}}/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly -Xms{{namenode_heapsize}} -Xmx{{namenode_heapsize}} -Dhadoop.security.logger=INFO,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT"
      export HADOOP_NAMENODE_OPTS="${SHARED_HADOOP_NAMENODE_OPTS} -XX:OnOutOfMemoryError=\"/usr/hdp/current/hadoop-hdfs-namenode/bin/kill-name-node\" -Dorg.mortbay.jetty.Request.maxFormContentSize=-1 ${HADOOP_NAMENODE_OPTS}"
      
export HADOOP_DATANODE_OPTS="-server -XX:ParallelGCThreads=8 -XX:+UseConcMarkSweepGC -XX:ErrorFile=/var/log/hadoop/$USER/hs_err_pid%p.log -XX:NewSize=800m -XX:MaxNewSize=800m -Xloggc:/var/log/hadoop/$USER/gc.log-`date +'%Y%m%d%H%M'` -verbose:gc -XX:+PrintGCDetails -XX:+PrintGCTimeStamps -XX:+PrintGCDateStamps -Xms{{dtnode_heapsize}} -Xmx{{dtnode_heapsize}} -Dhadoop.security.logger=ERROR,DRFAS -Dhdfs.audit.logger=INFO,DRFAAUDIT ${HADOOP_DATANODE_OPTS} -XX:CMSInitiatingOccupancyFraction=70 -XX:+UseCMSInitiatingOccupancyOnly -XX:+UseParNewGC "

export HADOOP_SECONDARYNAMENODE_OPTS="${SHARED_HADOOP_NAMENODE_OPTS} -XX:OnOutOfMemoryError=\"/usr/hdp/current/hadoop-hdfs-secondarynamenode/bin/kill-secondary-name-node\" ${HADOOP_SECONDARYNAMENODE_OPTS}"

&lt;/PRE&gt;&lt;P&gt;You mentioned "&lt;EM&gt;A GC allocation failure means that the garbage collector could not move objects from young gen to old gen fast enough because it does not have enough memory in old gen.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;&lt;STRONG&gt;which parameter holds values for old gen?&lt;/STRONG&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;we have got 8 data nodes, CPU 2*8 ,memory 256 GB, Disk -12*6 =72 TB&lt;/P&gt;&lt;P&gt;&lt;EM&gt;8 hosts of  72 TB each = 576 TB.&lt;/EM&gt;&lt;/P&gt;&lt;UL&gt;
&lt;LI&gt;&lt;EM&gt;&lt;STRONG&gt;our cluster&lt;/STRONG&gt; Blocksize=128 MB, &lt;STRONG&gt;Replication=3&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in MB: 8* 72,000,000 MB = 576,000,000 MB (576TB)&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Disk space needed per block: 128 MB per block * 3 = &lt;STRONG&gt;384 MB&lt;/STRONG&gt; storage per block&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Cluster capacity in blocks: 576,000,000 MB / 384 MB = 1,500,000 &lt;STRONG&gt;blocks&lt;/STRONG&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;But ambari is reporting 156,710872 blocks, am I missing something here? &lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="76547-blocks.png" style="width: 469px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/17079i8CA8997957872B29/image-size/medium?v=v2&amp;amp;px=400" role="button" title="76547-blocks.png" alt="76547-blocks.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;Await for your response. Thank you so much!&lt;/P&gt;&lt;BR /&gt;&lt;IMG src="https://community.cloudera.com/t5/image/serverpage/image-id/6202iE0477C56752B30FB/image-size/large?v=1.0&amp;amp;px=999" border="0" alt="blocks.png" title="blocks.png" /&gt;</description>
      <pubDate>Sun, 18 Aug 2019 04:48:06 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205682#M78881</guid>
      <dc:creator>samant_thakur</dc:creator>
      <dc:date>2019-08-18T04:48:06Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205683#M78882</link>
      <description>&lt;P&gt;&lt;EM&gt;&lt;A href="https://community.hortonworks.com/questions/194026/@Samant%20Thakur"&gt;@Samant Thakur&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Have you configured your cluster for rack awareness?&lt;/EM&gt;&lt;/P&gt;&lt;UL&gt;
&lt;LI&gt;&lt;EM&gt;Rack awareness prevents data loss&lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Rack awareness improves network performance&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;HDFS block placement will use rack awareness for fault tolerance by placing one block replica on a different rack. This provides data availability in the event of a network switch failure or partition within the cluster.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;You will need the help of your network/data center team to share the network topology and how the nodes are spread out in the racks.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;You can use Ambari UI  --&amp;gt; Hosts to set the rack topology after knowing the subnets and DC setup. To understand better see &lt;A href="https://docs.hortonworks.com/HDPDocuments/HDP2/HDP-2.6.4/bk_hdfs-administration/content/ch_configuring_rack_awareness_hdp.html" target="_blank"&gt;HDP rack awareness&lt;/A&gt;  also see &lt;A href="https://community.hortonworks.com/articles/43057/rack-awareness-1.html" target="_blank"&gt;HCC  rack-awareness-series-1&lt;/A&gt; and &lt;A href="https://community.hortonworks.com/articles/43164/rack-awareness-series-2.html" target="_blank"&gt;HCC rack-awareness-series-2&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;Hope that helps&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Mon, 04 Jun 2018 06:06:19 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205683#M78882</guid>
      <dc:creator>Shelton</dc:creator>
      <dc:date>2018-06-04T06:06:19Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205684#M78883</link>
      <description>&lt;P&gt;@&lt;A href="https://community.hortonworks.com/users/1271/sheltong.html" rel="nofollow noopener noreferrer" target="_blank"&gt;Geoffrey Shelton Okot&lt;/A&gt; ,Thank you so much for getting back to me.&lt;/P&gt;&lt;P&gt;We don't have rack awareness enabled on our DR cluster as it's 8 data nodes cluster only. we do have rack awareness in our production cluster.&lt;/P&gt;&lt;P&gt;We can enable rack awareness later but my first priority is to get back the blocks on data nodes as faulty data nodes are not sending any block report to name node. Here is current status as of today.&lt;/P&gt;&lt;P&gt;&lt;span class="lia-inline-image-display-wrapper lia-image-align-inline" image-alt="76584-data-node-06052018.png" style="width: 1130px;"&gt;&lt;img src="https://community.cloudera.com/t5/image/serverpage/image-id/17077iFBDCB7FB0219B905/image-size/medium?v=v2&amp;amp;px=400" role="button" title="76584-data-node-06052018.png" alt="76584-data-node-06052018.png" /&gt;&lt;/span&gt;&lt;/P&gt;&lt;P&gt;I am still getting the EOFException error on problematic data nodes other data nodes are not giving this error.&lt;/P&gt;&lt;P&gt;I checked with our network team &amp;amp; they said all the data nodes are connected to same NIC and there is no packet loss.&lt;/P&gt;&lt;P&gt;Hardware team found some correctable memory errors but nothing major.&lt;/P&gt;&lt;P&gt;Is there any maximum number of blocks retention limits for a particular data node? I meant that is there any possibility that max. number of blocks retention limit has been exceeded for problematic data nodes &amp;amp; because of that they stopped sending the block report to name node due to some capacity/resource constraints?   Please guide.Do I need to report this as a bug to apache foundation?&lt;/P&gt;&lt;PRE&gt;java.io.EOFException: End of File Exception between local host is: "DATANODE HOST"; destination host is: "NAMENDOE HOST":8020; : java.io.EOFException; For more details see:  &lt;A href="http://wiki.apache.org/hadoop/EOFException" target="_blank" rel="nofollow noopener noreferrer"&gt;http://wiki.apache.org/hadoop/EOFException&lt;/A&gt;
        at sun.reflect.GeneratedConstructorAccessor14.newInstance(Unknown Source)
        at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(DelegatingConstructorAccessorImpl.java:45)
        at java.lang.reflect.Constructor.newInstance(Constructor.java:423)
        at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.java:801)
        at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:765)
        at org.apache.hadoop.ipc.Client.getRpcResponse(Client.java:1558)
        at org.apache.hadoop.ipc.Client.call(Client.java:1498)
        at org.apache.hadoop.ipc.Client.call(Client.java:1398)
        at org.apache.hadoop.ipc.ProtobufRpcEngine$Invoker.invoke(ProtobufRpcEngine.java:233)
        at com.sun.proxy.$Proxy15.blockReport(Unknown Source)
        at org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolClientSideTranslatorPB.blockReport(DatanodeProtocolClientSideTranslatorPB.java:211)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.blockReport(BPServiceActor.java:374)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.offerService(BPServiceActor.java:645)
        at org.apache.hadoop.hdfs.server.datanode.BPServiceActor.run(BPServiceActor.java:785)
        at java.lang.Thread.run(Thread.java:745)
Caused by: java.io.EOFException
        at java.io.DataInputStream.readInt(DataInputStream.java:392)
        at org.apache.hadoop.ipc.Client$Connection.receiveRpcResponse(Client.java:1119)
        at org.apache.hadoop.ipc.Client$Connection.run(Client.java:1014)
&lt;/PRE&gt;</description>
      <pubDate>Sun, 18 Aug 2019 04:47:53 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205684#M78883</guid>
      <dc:creator>samant_thakur</dc:creator>
      <dc:date>2019-08-18T04:47:53Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205685#M78884</link>
      <description>&lt;P&gt;&lt;EM&gt; &lt;A href="https://community.hortonworks.com/questions/194026/@Samant%20Thakur"&gt;@Samant Thakur&lt;/A&gt;&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;When a Hadoop framework creates a new block, it places the first replica on the local node. And place the second one in a different rack, and the third one is on a different node on the local node. During block replicating, if the number of existing replicas is one, place the second on a different rack. When the number of existing replicas are two, if the two replicas are in the same rack, place the third one on a different rack. &lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;The main purpose of Rack awareness is to:&lt;/EM&gt;&lt;/P&gt;&lt;UL&gt;&lt;LI&gt;&lt;EM&gt; Improve data reliability and data availability. &lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Better cluster performance. &lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Prevents data loss if the entire rack fails. &lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;To improve network bandwidth. &lt;/EM&gt;&lt;/LI&gt;&lt;LI&gt;&lt;EM&gt;Keep the bulk flow in-rack when possible. &lt;BR /&gt;&lt;/EM&gt;&lt;/LI&gt;&lt;/UL&gt;&lt;P&gt;&lt;EM&gt;If your production and this problematic cluster have the same &lt;STRONG&gt;Ambari/HDP&lt;/STRONG&gt; version then, you can't call it a bug but client specific problem.&lt;/EM&gt;&lt;/P&gt;&lt;P&gt;&lt;EM&gt;I would still insist you enable &lt;STRONG&gt;rack awareness&lt;/STRONG&gt; and monitor over 24hr to see the change in the alerts. Have you tried running a cluster balancing utility? &lt;/EM&gt;&lt;/P&gt;&lt;PRE&gt;&lt;EM&gt;$ hadoop balancer&lt;/EM&gt;&lt;/PRE&gt;&lt;P&gt;&lt;EM&gt;HTH&lt;/EM&gt;&lt;/P&gt;</description>
      <pubDate>Wed, 06 Jun 2018 20:29:12 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/205685#M78884</guid>
      <dc:creator>Shelton</dc:creator>
      <dc:date>2018-06-06T20:29:12Z</dc:date>
    </item>
    <item>
      <title>Re: Data Nodes displaying incorrect block report</title>
      <link>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/301001#M78885</link>
      <description>&lt;DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;FONT size="2"&gt;&lt;SPAN&gt;&lt;STRONG&gt;Services Reporting Alerts&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;UNKNOWN&lt;/TD&gt;&lt;TD&gt;[AMBARI_METRICS]&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;FONT size="2"&gt;&lt;SPAN&gt;&lt;STRONG&gt;AMBARI_METRICS&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;UNKNOWN&lt;/TD&gt;&lt;TD&gt;Metrics Collector - HBase Master CPU Utilization&lt;DIV&gt;&lt;FONT size="1"&gt;&lt;SPAN&gt;[Alert][ams_metrics_collector_hbase_master_cpu] Unable to extract JSON from JMX response&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Is this big issue? anything worry about it?&lt;/P&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;P&gt;&lt;SPAN&gt;another alert is&amp;nbsp;&lt;/SPAN&gt;&lt;/P&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;FONT size="2"&gt;&lt;SPAN&gt;&lt;STRONG&gt;Services Reporting Alerts&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;OK&lt;/TD&gt;&lt;TD&gt;[AMBARI_METRICS]&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;DIV&gt;&lt;FONT size="2"&gt;&lt;SPAN&gt;&lt;STRONG&gt;AMBARI_METRICS&lt;/STRONG&gt;&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/DIV&gt;&lt;DIV&gt;&lt;TABLE&gt;&lt;TBODY&gt;&lt;TR&gt;&lt;TD&gt;OK&lt;/TD&gt;&lt;TD&gt;Metrics Collector - HBase Master CPU Utilization&lt;DIV&gt;&lt;FONT size="1"&gt;&lt;SPAN&gt;12 CPU, load 12.8%&lt;/SPAN&gt;&lt;/FONT&gt;&lt;/DIV&gt;&lt;/TD&gt;&lt;/TR&gt;&lt;/TBODY&gt;&lt;/TABLE&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;Please advice me I am waiting for you kind response.&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;P&gt;&amp;nbsp;&lt;/P&gt;&lt;/DIV&gt;&lt;/DIV&gt;</description>
      <pubDate>Thu, 06 Aug 2020 14:18:43 GMT</pubDate>
      <guid>https://community.cloudera.com/t5/Archives-of-Support-Questions/Data-Nodes-displaying-incorrect-block-report/m-p/301001#M78885</guid>
      <dc:creator>sohelsazid</dc:creator>
      <dc:date>2020-08-06T14:18:43Z</dc:date>
    </item>
  </channel>
</rss>

