- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
Need suggestion to stop or Decommission Datanode?
- Labels:
-
Apache Hadoop
Created ‎04-24-2019 03:55 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Hi , We are using 12 Datanodes in our cluster. And we have 3 replication factor.
Currently we are doing cost optimization and our total cluster is 35% used. We want to remove one datanode from our cluster, but we want to see first how our cluster performs on 11 datanodes. if it doesn't perform well, we need to restart that datanode again.
To experiment this- Do I need to Stop the datanode or Decomission datanode?
Please need suggestion.
Created on ‎04-24-2019 04:40 PM - edited ‎08-17-2019 03:47 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
It's well documented in by hortonworks Once you launch the decommissioning the blocks on that node will be distributed to the other node remaining nodes
If the replication factor is higher than the number of existing data nodes after the removal, the removal process is not going to succeed!
Created on ‎04-24-2019 04:40 PM - edited ‎08-17-2019 03:47 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
It's well documented in by hortonworks Once you launch the decommissioning the blocks on that node will be distributed to the other node remaining nodes
If the replication factor is higher than the number of existing data nodes after the removal, the removal process is not going to succeed!
Created ‎04-25-2019 02:48 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Geoffrey Shelton Okot Thank you for your response. Instead of Decommissioning the node Can I stop it?
What is the difference? Whats the best way to stop it or Decommission?
Created ‎04-25-2019 04:38 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
Surely you could just t^do that but what happens to the replicas stored on that particular data node? Somehow your cluster has to reconstruct those replicas if you had a replicator factor of more than 1.
Here I was talking about planned maintenance! Just switching it off will also force your cluster to do the same thing in the background with alerts and ONLY when the replicas have been reconstructed will those alerts go away.
There is a performance cost for both decommissioning and just unplugging the data node.
Created ‎04-26-2019 01:26 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Geoffrey Shelton Okot Thank you. I think Decommissioning is a best way to do this. Do I need to turn on maintainance mode while I do decommissioning datanode? Also Do I need to stop the services like - NodeManager, Ambari metrics for that datanode after decommissioning?
