How many ways can you create rdd in spark?
Answer / Abhay Deep Srivastava
There are multiple ways to create RDDs (Resilient Distributed Datasets) in Spark, including parallelizing an existing collection, reading data from files (such as text files or binary files), and using APIs for external data sources like HDFS, Cassandra, or Twitter.
| Is This Answer Correct ? | 0 Yes | 0 No |
What are the differences between Caching and Persistence method in Apache Spark?
Can you define rdd lineage?
Is spark secure?
Is it possible to run Spark and Mesos along with Hadoop?
What is javardd?
List out the ways of creating RDD in Apache Spark?
Is a distributed machine learning framework on top of spark?
What is Starvation scenario in spark streaming?
Define Partition and Partitioner in Apache Spark?
How can you minimize data transfers when working with Spark?
Is scala required for spark?
Is there any benefit of learning MapReduce, then?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)