Created on 12-01-2016 03:29 AM - edited 08-17-2019 07:37 AM
Background
Customer attached bigger disks to expand the data node storage. If one disk physical fail, can also use this solution.
Step by Step
# for disk in /hadoop/disk-sd{d..j}/Hadoop > do > mkdir ${disk}/yarn/log && chown yarn:hadoop ${disk}/yarn/log > mkdir ${disk}/yarn/local && chown yarn:hadoop ${disk}/yarn/local > done
$ hdfs fsck / | egrep -v '^\.+$' | grep -v eplica FSCK started by hdfs (auth:KERBEROS_SSL) from /192.168.141.39 for path / at Tue Nov 29 10:42:34 UTC 20161.............................................................................................Status: HEALTHY Total size: 769817156313 B (Total open files size: 75189484 B) Total dirs: 4934 Total files: 23693 Total symlinks: 0 (Files currently being written: 30) Total blocks (validated): 27536 (avg. block size 27956753 B) (Total open file blocks (not validated): 24) Corrupt blocks: 0 Number of data-nodes: 7 Number of racks: 1FSCK ended at Tue Nov 29 10:42:34 UTC 2016 in 433 millisecondsThe filesystem under path '/' is HEALTHY
Created on 04-10-2017 08:48 AM
Can you please tell why do we need to stop all services on the host while only DataNode and NodeManager is affected? Why do we also need to stop RegionServer and other services?
Created on 04-17-2017 10:22 AM
After stopping DataNode, Hbase RegionServer still holds FDs from mounts and it needs to be stopped before you can unmount the volumes.
User | Count |
---|---|
758 | |
379 | |
316 | |
309 | |
270 |