When running Spark applications, is it necessary to install Spark on all the nodes of YARN cluster?
Answer / Mukesh Singh Laspal
No, it's not necessary to install Spark on all nodes in a YARN cluster. Instead, you only need to install Spark on the nodes that will run the application's executor processes. The resource manager (usually a YARN NodeManager) and the application master manage the distribution of tasks among these executors.
| Is This Answer Correct ? | 0 Yes | 0 No |
What is salting in spark?
Does hadoop install spark?
How rdd can be created in spark?
What happens if rdd partition is lost due to worker node failure?
What is DStream in Apache Spark Streaming?
Hadoop uses replication to achieve fault tolerance. How is this achieved in Apache Spark?
Explain the terms Spark Partitions and Partitioners?
What is the difference between Caching and Persistence in Apache Spark?
What is Map() operation in Apache Spark?
Can you do real-time processing with Spark SQL?
Name the two types of shared variable available in Apache Spark?
List the functions of Spark SQL?
Apache Hadoop (394)
MapReduce (354)
Apache Hive (345)
Apache Pig (225)
Apache Spark (991)
Apache HBase (164)
Apache Flume (95)
Apache Impala (72)
Apache Cassandra (392)
Apache Mahout (35)
Apache Sqoop (82)
Apache ZooKeeper (65)
Apache Ambari (93)
Apache HCatalog (34)
Apache HDFS Hadoop Distributed File System (214)
Apache Kafka (189)
Apache Avro (26)
Apache Presto (15)
Apache Tajo (26)
Hadoop General (407)