Why ‘Reading‘ is done in parallel and ‘Writing‘ is not in HDFS?
Answer / Kum. Gunjan Singh
In HDFS, reading is done in parallel because it can read data blocks from multiple DataNodes simultaneously. This improves the read performance. On the other hand, writing is a sequential process because when a block is being written, no other nodes should modify it until the write operation is completed to ensure data integrity.
| Is This Answer Correct ? | 0 Yes | 0 No |
Which classes are used by the hive to read and write hdfs files?
How HDFS client divide the file into the block while storing inside HDFS?
How data or file is read in Hadoop HDFS?
What do you mean by metadata in Hadoop?
Does HDFS allow a client to read a file which is already opened for writing?
Explain the hdfs architecture?
What is the module in HDFS?
If data is present in HDFS and RF is defined, then how can we change Replication Factor?
What is secondary namenode?
Why HDFS?
When NameNode enter in Safe Mode?
Explain the difference between an hdfs block and input split?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)