Yes, the block count alert serves to indicate an early warning to a growing
number of small files issue. While your DN can handle a lot of blocks in
general, going too high will cause perf. issues. Small files will also
cause a lot of processing overheads and slow down workloads generally.
If you use CM Enterprise, you can use the Reports feature to find small
files and who/where its top culprits are:
http://www.cloudera.com/documentation/enterprise/latest/topics/cm_dg_disk_usage_reports.htmlOtherwise, a script to analyse a single hadoop fs -ls -R / output may also
suffice.