Why do we use HDFS for applications having large data sets and not when there are lot of small files?
1 2111Post New Apache Hadoop Questions
What are the modules that constitute the Apache Hadoop 2.0 framework?
How can I restart namenode?
Suppose Hadoop spawned 100 tasks for a job and one of the task failed. What will Hadoop do?
Does this lead to security issues?
Explain the master class and the output class do?
Explain what if rack 2 and datanode fails?
On what basis name node distribute blocks across the data nodes?
Is hadoop a memory?
How blocks are distributed among all data nodes for a particular chunk of data?
What does ‘jps’ command do?
What is the functionality of jobtracker in hadoop?
How is hadoop different from other data processing tools?
what should be the ideal replication factor in hadoop?
Explain the features of pseudo mode?
Why we cannot do aggregation (addition) in a mapper? Why we require reducer for that?