Advertisement

Pro Hadoop pp 127-175 | Cite as

MapReduce Details for Multimachine Clusters

Abstract

Organizations run Hadoop Core to provide MapReduce services for their processing needs. They may have datasets that can’t fit on a single machine, have time constraints that are impossible to satisfy with a small number of machines, or need to rapidly scale the computing power applied to a problem due to varying input set sizes. You will have your own unique reasons for running MapReduce applications.

Keywords

File System Input File Sequence File Member Variable Hadoop Framework 
These keywords were added by machine and not by the authors. This process is experimental and the keywords may be updated as the learning algorithm improves.

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

Copyright information

© Jason Venner 2009

Personalised recommendations