I have a disk running full on one of my Data node:
[ayguha@dh03 hadoop]$ sudo du -h --max-depth=1 674G ./hdfs 243G ./yarn 916G .
[xx@dh03 local]$ sudo du -h --max-depth=1 1.4G ./filecache 3.2G ./usercache 68K ./nmPrivate 242G .
There are over 1k tmp files accumulating in /data/hadoop/yarn/local
[ayguha@dh03 local]$ ls -l *.tmp | wc -l 1055 ./optimized-preview-record-buffer-2808068b-4d54-492e-a31a-385065d25a408826610818023522318.tmp ./preview-record-buffer-24a7477f-01f0-427e-a032-54866df48b197825057363055390034.tmp ./preview-record-buffer-b22020bb-6ec2-4f73-9d65-65dbba50136e527236496621902098.tmp [ayguha@dh03 local]$ find ./*preview-record-buffer* -type f -mtime +90 | wc -l 973
There are near 1k files that are older than 3 months . Is it safe to delete these files ?
ENV: Hadoop 220.127.116.11.4.0.0-169 HDP 2.4
I did look at the Link you posted and decided to delete the file.
CAUTION: For some reason a few hours later there were inconsistencies in the cluster . One of the data nodes (D5) were clean up was done had corruption in the way containers were processed. Some jobs for which containers were lunched in D5 executed to completion successfully and some other jobs failed due to Vertex failed error. We could not find any errors in RM log/Datanode Log/Node Manager Log
We had to remove D5 off the cluster and reinstall node manager to set things right.