We are using Hortonworks HDP 2.1 (HDFS 2.4), with replication factor 3. We have recently decommissioned a datanode and that left a lot of under replicated blocks in the cluster.
Cluster is now trying to satisfy the replication factor by distributing under replicated blocks among other nodes.
Appreciate your help on this
1 . First you need to run hadoop fsck / to check the under-replicated blocks. Then you can run hadoop -setrep 2 to the files which are under-replicated. This will stop the process.
2. Yes it will remove the third copy.
So eventually, -setrep 2 stops the process and spawn another process (deleting the third copy) right? Is there any way to stop the cluster removing the third copy? I'm trying to reduce the cpu utilization.