I'm looking for a smart solution for monitoring hdfs folder quotas.
I know how to get the quota of a single directory (hadoop fs -count -q /path/to/directory), and can also do a recursive script, but on very large scale hdfs it is not efficient.
Has anyone used or know of a smart \ efficient solution for this ?
Or a way to show all folders that have quotas ?
Thanks in advance