Answer Posted / Isha Vishnoi
Apache Spark achieves fault tolerance through resilient distributed datasets (RDDs). RDDs are immutable distributed collections of data partitioned across nodes in a cluster. When an action is triggered on an RDD, the system checks the status of each node to ensure that all tasks have completed successfully. If any task fails, Spark will automatically re-execute the failed task on another available node, ensuring that the final result remains consistent.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers