Answer Posted / Anshi Gupta
Shuffle is a process in Apache Spark where data is redistributed among executors for grouping, sorting, or joining. It can be computationally expensive due to the large amount of data movement involved, but Spark has optimizations like sort merge phase and shuffle spill to mitigate performance issues.
| Is This Answer Correct ? | 0 Yes | 0 No |
Post New Answer View All Answers