Typically you either need to scale out due to HDFS disk usage, or you need to scale out for computational reasons.
If I have 10 or so datanodes and they are each allocated 80% of the system memory for YARN, would them all running 100% of their YARN allocation for a majority of the day indicate that I need to scale out for computational reasons? Currently only my HDFS is at 60% utilization.
I am primarily running Tez jobs, CPU doesn't seem to be hit as much, but my YARN memory allocation is constantly 100% and I have users complaining about slow running jobs. I assume this is because they have to wait for other jobs to free up resources for them to get their job to run.
Are there any things I could look for in this situation?
Running Ambari 2.5.1 and HDP 2.6.1.