How will you perform the inter cluster data copying work in hdfs?
Answer / Arvind Kumar Sinha
Inter-cluster data copying in HDFS can be achieved using DistCP or Data Transfer Protocol (DTP). DistCP is a distributed file copy utility for Hadoop, while DTP allows high-speed transfer of large amounts of data between HDFS clusters.
| Is This Answer Correct ? | 0 Yes | 0 No |
How one can change Replication factor when Data is already stored in HDFS
What is NameNode and DataNode in HDFS?
Explain how are file systems checked in hdfs?
Hdfs stores data using commodity hardware which has higher chances of failures. So, how hdfs ensures the fault tolerance capability of the system?
What alternate way does HDFS provides to recover data in case a Namenode, without backup, fails and cannot be recovered?
How to create directory in HDFS?
What is a difference between an input split and hdfs block?
Explain the key features of hdfs?
Explain what is difference between an input split and hdfs block?
While processing data from hdfs, does it execute code near data?
What should be the HDFS Block size to get maximum performance from Hadoop cluster?
What is Fault Tolerance in Hadoop HDFS?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)