Which storage level does the cache () function use?
Answer / Parvendra Singh
The cache() function in Apache Spark uses the MEMORY_ONLY storage level by default. This means that the RDD or DataFrame will be stored entirely in memory, which can improve the performance for subsequent actions on the same dataset.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is coalesce in spark sql?
Describe Spark SQL?
Explain api create or replace tempview()?
How Spark uses Akka?
How can we create RDD in Apache Spark?
Explain how RDDs work with Scala in Spark
What does a Spark Engine do?
What is Resilient Distributed Dataset (RDD) in Apache Spark? How does it make spark operator rich?
What are the various libraries available on top of Apache Spark?
Can I run Apache Spark without Hadoop?
What is meant by spark in big data?
Do I need to know scala to learn spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)