28806
DISCUSSIONS
102204
MEMBERS
3161
ARTICLES
Created 10-08-2013 11:01 PM
Hi
I am new toHadoop.
I would like to know the decision factors considered for setting up hadoop cluster for processing large volume of data.
I have read that the defalut block size is 64Mb or 128Mb. On what factor it was decied?
Number of mappers are decided by the number of input split, the size of slit being the blocksize.
Also I have seen several answers like number of reducers is directly proportional to number of reducer slots in the cluster, another
answer some mathematical calcualtion like 4 cores machine then you can have (0.75*no cores)no. of MR slots.