What are accumulators in Apache Spark?
Answer / Vivek Kumar Mishra
Accumulators in Apache Spark are variables that can be updated by tasks running on executors during the execution of an action. They provide a way to aggregate results across multiple tasks without the need for shuffling data. Accumulators can be used for various purposes, such as tracking the number of records processed or maintaining a moving average. Users can define their own accumulator classes using Scala or Java.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is a pipelinedrdd?
Can you do real-time processing with Spark SQL?
What is javardd?
Who invented the first spark plug?
Are spark dataframes distributed?
Define paired RDD in Apache Spark?
What is the default partition in spark?
What does a Spark Engine do?
What is aggregatebykey spark?
What is Starvation scenario in spark streaming?
Which all languages Apache Spark supports?
What happens to rdd when one of the nodes on which it is distributed goes down?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)