Pro Hadoop pp 127-175 | Cite as

MapReduce Details for Multimachine Clusters


Organizations run Hadoop Core to provide MapReduce services for their processing needs. They may have datasets that can’t fit on a single machine, have time constraints that are impossible to satisfy with a small number of machines, or need to rapidly scale the computing power applied to a problem due to varying input set sizes. You will have your own unique reasons for running MapReduce applications.


File System Input File Sequence File Member Variable Hadoop Framework 


Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Copyright information

© Jason Venner 2009

Personalised recommendations