Why do we need hdfs?
Answer / Sandip Kumar Srivastava
HDFS (Hadoop Distributed File System) is needed for big data processing because it allows for the storage and processing of large datasets across multiple nodes. Its fault tolerance, scalability, and high-throughput capabilities make it ideal for handling the complex requirements of big data.
| Is This Answer Correct ? | 0 Yes | 0 No |
Explain hdfs?
What is a task tracker?
Write command to copy a file from HDFS to linux(local).
Is namenode also a commodity?
Explain the process that overwrites the replication factors in HDFS?
Why does hive not store metadata information in hdfs?
How to copy a file into HDFS with a different block size to that of existing block size configuration?
What is Fault Tolerance in HDFS?
How does hdfs give great throughput?
How does data transfer happen from hdfs to hive?
Does HDFS allow a client to read a file which is already opened for writing?
How to Delete directory from HDFS?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)