Hadoop uses replication to achieve fault tolerance. How is this achieved in Apache Spark?
Answer / Akhilesh Kumari
Apache Spark doesn't use explicit replication for fault tolerance like Hadoop does. Instead, it uses resilient distributed datasets (RDDs) which are immutable distributed collections of data that can tolerate failures by default.
| Is This Answer Correct ? | 0 Yes | 0 No |
Can you define pagerank?
what do you mean by the worker node?
When should you use spark cache?
What is an "Accumulator"?
On what all basis can you differentiate rdd, dataframe, and dataset?
How is transformation on rdd different from action?
Is spark written in scala?
Does diesel engine have spark plug?
Is it possible to run Apache Spark without Hadoop?
What are broadcast variables in spark?
What is driver and executor in spark?
Explain various level of persistence in Apache Spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)