Member since
07-30-2020
219
Posts
45
Kudos Received
60
Solutions
My Accepted Solutions
Title | Views | Posted |
---|---|---|
429 | 11-20-2024 11:11 PM | |
486 | 09-26-2024 05:30 AM | |
1081 | 10-26-2023 08:08 AM | |
1852 | 09-13-2023 06:56 AM | |
2126 | 08-25-2023 06:04 AM |
10-09-2023
05:51 AM
Can you try to run the below command on all 3 Zookeeper instances? echo "stat" | nc localhost 2181 | grep Mode
... View more
10-04-2023
11:30 PM
@Noel_0317 The directory /hadoop/dfs/name/ might be your Namenode data directory that contains the metadata in the form of fsimage and edits. So won't recommend deleting it if that's the case. You can confirm if this directory is indeed the NN data directory by checking the HDFS configuration. If the cluster is working and still taking writes, you can verify if the Namenode Data dir has been changed to a different mount point if the latest data available on it is from July.
... View more
09-30-2023
03:58 AM
1 Kudo
Hi @Noel_0317 I am sharing few hdfs commands to be checked on file system level of hdfs. hdfs dfs -df /hadoop - Shows the capacity, free and used space of the filesystem. hdfs dfs -df -h /hadoop - Shows the capacity, free and used space of the filesystem. -h parameter Formats the sizes of files in a human-readable fashion. hdfs dfs -du /hadoop/file - Show the amount of space, in bytes, used by the files that match the specified file pattern. hdfs dfs -du -s /hadoop/file - Rather than showing the size of each individual file that matches the pattern, shows the total (summary) size. hdfs dfs -du -h /hadoop/file - Show the amount of space, in bytes, used by the files that match the specified file pattern. Formats the sizes of files in a human-readable. fashion Let me know if this helps.
... View more
09-20-2023
06:48 AM
@newtocm, Have any of the replies helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future.
... View more
08-31-2023
10:32 PM
@Crash, Have the replies helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future. d?
... View more
08-31-2023
09:52 PM
@Yasine, I'm happy to see you resolved your issue. Please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future.
... View more
08-23-2023
10:57 PM
Is anyone aware if there are any plans for hflush implementation in EC in future versions?
... View more
08-21-2023
02:52 AM
@Ben1996, Has the reply helped resolve your issue? If so, please mark the appropriate reply as the solution, as it will make it easier for others to find the answer in the future.
... View more
08-08-2023
07:49 AM
When im trying to install in my centos linux machine im facing issues like I have installed Phoenix in linux server but getting below error Caused by: java.lang.IllegalStateException: Multiple resources found for 'META-INF/native/liborg_apache_hbase_thirdparty_netty_transport_native_epoll_x86_64.so' with different content: [jar:file:/home/hadoop/hbase/hbase-2.5.5/lib/hbase-shaded-netty-4.1.4.jar!/META-INF/native/liborg_apache_hbase_thirdparty_netty_transport_native_epoll_x86_64.so, jar:file:/home/hadoop/hbase/hbase-2.5.5/lib/phoenix-client-hbase-2.5-5.1.3.jar!/META-INF/native/liborg_apache_hbase_thirdparty_netty_transport_native_epoll_x86_64.so] So is there any alternate way to install Apache Phoenix in centos 7
... View more
08-04-2023
12:36 AM
@Noel_0317 Do you have rack awareness configured for the Datanodes? Also, check for any disk-level issues on the datanode. Try enabling Debug for block placement : log4j.logger.org.apache.hadoop.hdfs.server.blockmanagement.BlockPlacementPolicy=DEBUG
log4j.logger.org.apache.hadoop.hdfs.protocol.BlockStoragePolicy=DEBUG
... View more