How to change a number of mappers running on a slave in MapReduce?
What is the relationship between Job and Task in Hadoop?
What is the sequence of execution of map, reduce, recordreader, split, combiner, partitioner?
How many Reducers run for a MapReduce job?
How would you tackle calculating the number of unique visitors for each hour by mining a huge apache log? You can use post processing on the output of the mapreduce job.
How does Hadoop Classpath plays a vital role in stopping or starting in Hadoop daemons?
What counter in Hadoop MapReduce?
In Hadoop what is InputSplit?
Mention what is the next step after mapper or maptask?
What happens if the quantity of the reducer is 0 in mapreduce?
How does inputsplit in mapreduce determines the record boundaries correctly?
How will you submit extra files or data ( like jars, static files, etc. ) For a mapreduce job during runtime?
Mention what are the main configuration parameters that user need to specify to run mapreduce job?
What are the main configuration parameters in a MapReduce program?
What are the the issues associated with the map and reduce slots based mechanism in mapReduce?