How rdd persist the data?
Answer / Yesh Pal Singh
"RDDs (Resilient Distributed Datasets) in Apache Spark do not store data persistently unless explicitly told to do so. Persisting an RDD makes it available for future computations and reduces the amount of data that needs to be processed repeatedly. The persistence level can be controlled using methods like cache(), remember(), and checkpoint().".
| Is This Answer Correct ? | 0 Yes | 0 No |
Explain fullOuterJoin() operation in Apache Spark?
What are the ways in which one can know that the given operation is transformation or action?
If map reduce is inferior to spark then is there any benefit of learning it?
What is heap memory in spark?
What is a dataframe spark?
What is the difference between Spark Transform in DStream and map ?
Which are the methods to create rdd in spark?
Who invented the first spark plug?
What is a "Spark Executor"?
Explain the difference between Spark SQL and Hive.
What is executor memory in a spark application?
Compare Hadoop and Spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)