- Subscribe to RSS Feed
- Mark Question as New
- Mark Question as Read
- Float this Question for Current User
- Bookmark
- Subscribe
- Mute
- Printer Friendly Page
adding additional volume to data nodes.
- Labels:
-
Apache Hadoop
Created ‎11-05-2018 11:37 AM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
We are running hadoop HA cluster using AWS EC2 instances with 17 Data ndoes (All instances are M4.4xlarge including name nodes). All the DN's are configured with 16TB (EBS st1) volumes for hdfs.
Now we are running out of HDFS storage and looking to extend the storage. Since 16TB is max limit for st1 EBS we cannot extend the existing volume.
Trying to add additional 16TB volumes to few data nodes and update "DataNode directories" in ambari with this new volume path.
Will this approach impact any performance issue with cluster ? Any other things need be considered in this approach ?
Created ‎11-09-2018 05:18 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Sajesh PP - As you increase the size of a data node you can run into performance problems such as to much read/write activity on a single data node. If this occurs it is better to add new data nodes with additional storage.
Created ‎11-09-2018 05:18 PM
- Mark as New
- Bookmark
- Subscribe
- Mute
- Subscribe to RSS Feed
- Permalink
- Report Inappropriate Content
@Sajesh PP - As you increase the size of a data node you can run into performance problems such as to much read/write activity on a single data node. If this occurs it is better to add new data nodes with additional storage.
