O'Reilly logo
  • Libin Thomas thinks this is interesting:

Each map task in Hadoop is broken into the following phases: record reader, mapper, combiner, and partitioner. The output of the map tasks, called the intermediate keys and values, are sent to the reducers. The reduce tasks are broken into the following phases: shuffle, sort,

From

Cover of MapReduce Design Patterns

Note

mapreduce lifecycle