How can apache spark be used alongside hadoop?
Answer / Chandrabhan Kushwaha
Apache Spark can be integrated with Hadoop using Hadoop Distributed File System (HDFS) for data storage and MapReduce as a batch processing engine. Spark can read data from HDFS directly or use YARN as a resource manager to run Spark applications on a Hadoop cluster.
| Is This Answer Correct ? | 0 Yes | 0 No |
How do I clear my spark cache?
How can you compare Hadoop and Spark in terms of ease of use?
What is spark yarn executor memoryoverhead?
What is full form of rdd?
Is it possible to use Apache Spark for accessing and analyzing data stored in Cassandra databases?
What is spark training?
Can you explain spark sql?
Can you define yarn?
What is the difference between dataset and dataframe in spark?
Define the term ‘sparse vector.’
Why scala is used in spark?
What is the disadvantage of spark sql?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)