Can you explain accumulators in apache spark?
Answer / Surendra Pal
"Accumulators are variables that can be updated by the tasks running on an executor during a Spark job. They allow a simple form of side effect computation where a value is collected across multiple data partitions and aggregated into one final result. Accumulators can be used to track counters, or for custom computations like finding the maximum or minimum value in a dataset.".
| Is This Answer Correct ? | 0 Yes | 0 No |
Why apache spark is faster than hadoop?
What is spark used for?
Define sparksession in apache spark? Why is it needed?
What is the role of Driver program in Spark Application?
Explain the key features of Spark.
What is catalyst framework in spark?
How do I start a spark server?
What is spark client?
Define the run-time architecture of Spark?
Which is better hadoop or spark?
What is sc parallelize in spark?
Are spark dataframes distributed?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)