Explain various level of persistence in Apache Spark?
Answer / Ankit Multani
There are three levels of persistence in Apache Spark: (1) Memory-only, where data is stored only in memory and not on disk, providing fast access but with limited durability. (2) DisktOnly, where data is saved to disk but not kept in memory after processing, offering improved durability at the cost of slower access times. (3) MemoryAndDisks, which combines both memory-only and disktOnly persistence, allowing for efficient caching and querying of data.
| Is This Answer Correct ? | 0 Yes | 0 No |
Explain Spark leftOuterJoin() and rightOuterJoin() operation?
On which all platform can Apache Spark run?
Which is the best spark certification?
How do I optimize my spark code?
Explain pipe() operation in Apache Spark?
What is difference between cache and persist in spark?
what do you mean by the worker node?
What is rdd map?
What is the difference between persist() and cache()?
What is the spark driver?
Is spark sql faster than hive?
What is dag – directed acyclic graph?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)