I've noticed in Ambari under HDFS metrics that we have 2 blocks with corrupt replicas. Running " hdfs fsck / " shows no corrupt blocks and system is healthy. Running "hdfs dfsadmin-report" shows 2 corrupt replicas (same as Ambari dashboard)
I've restarted Ambari metrics & Ambari Agents on all nodes + Ambari-server as noted in one of the threads i came across but still - problem remains.
This will be a list the affected files, and the output will not be a bunch of dots, the output should include something like this with all your affected files.
/path/to/filename.file_extension: CORRUPT blockpool BP-1016133662-10.29.100.41-1415825958975 block blk_1073904305
/path/to/filename.file_extension: MISSING 1 blocks of total size 15620361 B
The next step would be to determine the importance of the file, can it just be removed and copied back into place, or is there sensitive data that needs to be regenerated? you have a replication factor of 1 so analyze well.
Remove the corrupted file(s)
This command will move the corrupted file to the trash incase you realise the files is importantyou still have an option of recovering it .
$ hdfs dfs -rm /path/to/filename.file_extension
When you use skip the trash to permanently delete if you are sure you really don't need that file.