By Default, how many partitions are created in RDD in Apache Spark?
Answer / Deepak Diwakar
By default, when an RDD is created from a local file using `textFile()` or `wholeTextFiles()`, Spark creates 20 partitions for the RDD. When reading data from HDFS, Spark creates as many partitions as there are blocks in the file on the HDFS cluster.
| Is This Answer Correct ? | 0 Yes | 0 No |
How do I get apache spark on windows 10?
What do you mean by Speculative execution in Apache Spark?
How do I optimize my spark code?
Why do people use spark?
Is it possible to run Apache Spark on Apache Mesos?
What is the FlatMap Transformation in Apache Spark RDD?
What is a worker node in Apache Spark?
Which are the various data sources available in spark sql?
What is faster than apache spark?
What is shark?
Apache Spark is a good fit for which type of machine learning techniques?
How do I clear my spark cache?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)