How do you process big data with spark?
Answer / Raza Hussain Khan
To process big data with Apache Spark, you first create a Spark Context or Session, then load your data into a Resilient Distributed Dataset (RDD), and apply transformations and actions to it. Transformations return new datasets, and actions return a value.
| Is This Answer Correct ? | 0 Yes | 0 No |
Discuss writeahead logging in Apache Spark Streaming?
What is the difference between hadoop and spark?
What is tungsten engine in spark?
Define "Transformations" in Spark
What do you understand by Lazy Evaluation?
What is the standalone mode in spark cluster?
What is meant by rdd lazy evaluation?
Which are the methods to create rdd in spark?
What is the advantage of a Parquet file?
What is RDD?
What is Apache Spark? What is the reason behind the evolution of this framework?
What are the ways in which one can know that the given operation is transformation or action?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)