How does hdfs give great throughput?
Answer / Narendar Kumar
HDFS achieves high throughput by using a data distribution strategy, where each file is split into blocks and distributed across multiple nodes in the cluster. This allows for parallel processing of I/O operations.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is a rack awareness algorithm and why is it used in hadoop?
How to restart NameNode or all the daemons in Hadoop HDFS?
On what basis name node distribute blocks across the data nodes in HDFS?
List the various HDFS daemons in HDFS cluster?
What is the difference between Input Split and an HDFS Block?
How much Metadata will be created on NameNode in Hadoop?
What is hdfs block size?
What alternate way does HDFS provides to recover data in case a Namenode, without backup, fails and cannot be recovered?
What is the difference between NAS and HDFS?
Why do we need hdfs?
Can multiple clients write into an HDFS file concurrently?
Since the data is replicated thrice in hdfs, does it mean that any calculation done on one node will also be replicated on the other two?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)