List out the ways of creating RDD in Apache Spark?
Answer / Gyanesh Kumar
The following are the ways to create RDDs in Apache Spark:
1. Creating an RDD from a local collection using parallelize() method.
2. Reading data from an external file or HDFS file system.
3. Using SQLContext or DataFrame API to convert DataFrames and Datasets into RDDs.
| Is This Answer Correct ? | 0 Yes | 0 No |
When should you use spark cache?
What are Actions?
What do you use spark for?
What is PageRank in Spark?
What is mlib?
What is Apache Spark and what are the benefits of Spark over MapReduce?
What advantages does Spark offer over Hadoop MapReduce?
What is the use of spark sql?
What is executor cores in spark?
What is catalyst query optimizer in apache spark?
What is worker node in Apache Spark cluster?
What database does spark use?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)