Thursday, 14 May 2015

Hadoop mapreduce basics

Below are the basic points to be remebered about mapreduce..

            map: (K1,V1) → list(K2,V2)
             reduce: (K2,list(V2)) → list(K3,V3)
  1. MapReduce works in 2 phase . Map phase and Reduce phase.
  2. Mapreduce is a data preparation phase and  for filteration of unwanted records.
  3. Reducers have an Iterator which keeps on iterating the values associated with Key.
  4. Rewriting on the same directory will raise exception.
  5. 1 reducer = 1 output file.
  6. we can set number of reducers in mapreduce program.
  7. Input is divided into Splits.and should be equal to the block size
  8. if the reducer is not defined the out put will be written to hdfs.
  9. Read Data Localisation.
  10. Map write output is local not hdfs and in shuffle and sort phase it is sent to reducers.
  11. combiner works well in association and commutative processes.
  12. Combiners cutdown shuffle btw networks in map and reduce phase.
  13. Hadoop Streaming to run in different languages.   

No comments:

Post a Comment