Member since
07-30-2020
219
Posts
46
Kudos Received
60
Solutions
My Accepted Solutions
| Title | Views | Posted |
|---|---|---|
| 4886 | 11-20-2024 11:11 PM | |
| 2924 | 09-26-2024 05:30 AM | |
| 2459 | 10-26-2023 08:08 AM | |
| 4193 | 09-13-2023 06:56 AM | |
| 4503 | 08-25-2023 06:04 AM |
01-25-2023
09:26 PM
If the same Node is getting down every time, it's worth checking the Memory utilization at the OS end. You can check the /var/log/messages of the NN host when the NN went down and check if the process is getting killed by an oom.
... View more
01-23-2023
05:55 AM
@RammiSE Try the below : ./hbase hbck -j /tmp/hbase-hbck2-1.2.0.jar assigns -o f0b4865fe8ea07321ed8eb237a592c10
... View more
01-23-2023
12:18 AM
@RammiSE you will need to assign the respective namespace region ID by checking the Hbase Master log using the hbck2 jar
... View more
01-22-2023
11:54 PM
Hi @RammiSE , Based on the exception, the hbase:namespace table is not online. You will need to assign the namespace region to bring up the Hbase Master. https://docs.cloudera.com/documentation/enterprise/6/6.3/topics/admin_hbase_hbck.html ~~~ Caused by: java.io.IOException: Timedout 300000ms waiting for namespace table to be assigned and enabled: tableName=hbase:namespace, state=ENABLED
... View more
01-11-2023
07:59 AM
@Love_Cat Are you setting this up for the 1st time or was it working fine earlier? As per the log, it seems the job is not able to locate the hbase meta location. Do you have Hbase Gateway role added to the node on which the sqoop job is running?
... View more
12-29-2022
12:04 AM
1 Kudo
@mabilgen you have 143 million blocks in the cluster and the NN heap is 95GB. This is why the NN is not holding up. You would need to bring the total block count to 90 million for NN to start working properly as the NN expects atleast 150GB of heap for 143 million blocks to work smoothly.
... View more
12-28-2022
10:04 AM
1 Kudo
HI @mabilgen , The main problem that you have on this cluster is lack of RAM on the host which is limited to 128GB. The Namenodes while startup will consume the allocated heap of 98GB that leaves 30GB of memory for any other processes running on this host. When other processes also utilise this remaining 30GB of memory, you are seeing huge JVM pauses as the garbage collector is trying to de-reference the objects to free up the memory but this takes too much time that the Namenode gives up on the Journalnode Quorum and fails over. As a Thumb rule, you should allocate 1GB of heap for 1 million blocks. So if there are more than 98million blocks on this cluster, then the current NN heap is not sufficient. 1) Try to lower the Total block count on the cluster by deleting any unwanted files or old snapshots. 2) If feasible, add more Physical RAM to the host Any amount of tunings won't be helpful in this situation as the jvm pauses are too big to be managed by tunings. As such, you would need to either perform cleanup of HDFS or Add more RAM to NN hosts or Move Namenode to another Node which has higher RAM.
... View more
12-28-2022
06:17 AM
HI @Sadique1 There is no specific option to decommission a disk from Datanode. The best option would be to decommission the Datanode followed by removing the disk and recommission it back. https://community.cloudera.com/t5/Community-Articles/Decommission-and-Reconfigure-Data-Node-Disks/ta-p/248262 The downside of decommission and recommission is that it will need to copy all the blocks to another host for all disks, not just the one you are removing. OR Considering that you don't have files with rep factor=1(fsck will tell you if any such files exist.), You could remove the disk from 1 DN and restart it, and wait for the NN to recover the under replicated blocks. Then repeat for each node. Note : If there are files with rep factor as 1 on those Datanode disks, then you will be getting missing blocks. Was your question answered? Please take some time to click on “Accept as Solution” below this post. If you find a reply useful, say thanks by clicking on the thumbs up button.
... View more
12-27-2022
11:19 AM
Hi @younes You will need to use the CM api to get this done. Check if the below helps : https://community.cloudera.com/t5/Support-Questions/hadoop-amp-amp-eco-system-cmdline-start-stop/m-p/359732#M238177
... View more
12-07-2022
10:31 PM
1 Kudo
Hi @jgabrey-1216863216 , This has been fixed in CDP 7.1.7 SP1 CHF20 (p1063). You can refer the below doc : https://docs.cloudera.com/cdp-private-cloud-base/7.1.7/runtime-release-notes/topics/chf-pvcb-sp1-overview.html#ariaid-title2
... View more