We have lots of applications who wants to use Hadoop. I am doing a capacity planning for the data they bring in
I need recommendations for data nodes, task per node and memory required in hdp environment to process the data
lets say i have 10 TB of data and this for one year, so how do i calculate the above things
Hardware i have is: 48 CPU cores, 251.6 GB of RAM and 23 TB of disk space in data nodes and there are 50 data nodes