How do we create rdds in spark?
Answer / Satya Swarup Anand
RDD (Resilient Distributed Datasets) can be created in Spark using several methods. Some common ways to create RDDs include loading data from an external storage system like HDFS or local file systems, creating RDDs from another RDD using transformations such as map, filter, and union, or using Spark Shell's built-in functions like textFile.
| Is This Answer Correct ? | 0 Yes | 0 No |
Define various running modes of apache spark?
What is distributed cache in spark?
Do I need to learn scala for spark?
How is streaming implemented in spark? Explain with examples.
Why is rdd immutable?
What exactly is spark?
Explain about the core components of a distributed Spark application?
What is setmaster in spark?
Define paired RDD in Apache Spark?
Is there any API available for implementing graphs in Spark?
Does google use spark?
What is scala and spark?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)