What combiners is and when you should use a combiner in a MapReduce Job?
What is shuffling in mapreduce?
What combiners are and when you should utilize a combiner in a map reduce job?
What is a "reducer" in Hadoop?
Where sorting is done on mapper node or reducer node in MapReduce?
What is the Job interface in MapReduce framework?
How would you tackle calculating the number of unique visitors for each hour by mining a huge apache log? You can use post processing on the output of the mapreduce job.
What is the difference between map and reduce?
Explain what does the conf.setmapper class do?
Explain about the partitioning, shuffle and sort phase
How does fault tolerance work in mapreduce?
What is identity mapper and chain mapper?
Can we submit the mapreduce job from slave node?
What MapReduce framework consists of?
what are the basic parameters of a Mapper?