When creating an RDD, what goes on internally?
Answer / Manjeet Singh
"When creating an RDD in Apache Spark, the driver program creates a logical directed acyclic graph (DAG) of computations. Each node in the DAG represents a transformation or action on the data, while edges indicate the dependencies between them. The RDD lineage is maintained to track each RDD's creation and all its transformations."
| Is This Answer Correct ? | 0 Yes | 0 No |
Explain fullOuterJoin() operation in Apache Spark?
What apache spark is used for?
Name types of Cluster Managers in Spark.
Is it possible to run Spark and Mesos along with Hadoop?
What is write ahead log(journaling) in Spark?
Does spark use yarn?
Compare Hadoop and Spark?
What is tungsten engine in spark?
What is lineage graph in spark?
What is cluster mode in spark?
What is Spark SQL?
What is the bottom layer of abstraction in the Spark Streaming API ?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)