Hadoop Map-Reduce Explained with an Example

0
This article represents key steps of Hadoop Map-Reduce Jobs using a word count example. Please feel free to comment/suggest if I missed to mention one or more important points. Also, sorry for the typos.

Following are the key steps of how Hadoop MapReduce works in a word count problem:

  • Input is fed to a program, say a RecordReader, that reads data line-by-line or record-by-record.
  • Mapping process starts which includes following steps:
    • Combining: Combines the data (word) with its count such as 1
    • Partitioning: Creates one partition for each word occurence
    • Shuffling: Move words to right partition
    • Sorting: Sort the partition by word
  • Last step is Reducing which comes up with the result such as word count for each occurence of word.

Following diagram represents above steps.

how does map reduce work
Following diagram depicts another view on how map-reduce works:
map_reduce_example
In above diagram, one could see that, primarily, there are three key phases of a map-reduce job:
  • Map: This phase processes data in form of key-value pairs
  • Partitioning/Shuffling/Sorting: This groups similar keys together and sort them
  • Reduce: This places final result with the key.
Ajitesh Kumar

Ajitesh Kumar

Ajitesh has been recently working in the area of AI and machine learning. Currently, his research area includes Safe & Quality AI. In addition, he is also passionate about various different technologies including programming languages such as Java/JEE, Javascript and technologies such as Blockchain, mobile computing, cloud-native technologies, application security, cloud computing platforms, big data etc.

He has also authored the book, Building Web Apps with Spring 5 and Angular.
Ajitesh Kumar

Leave A Reply

Time limit is exhausted. Please reload the CAPTCHA.