0 votes
in Apache Spark by
Is it necessary to install Spark on all nodes of the YARN cluster?

1 Answer

0 votes
by

No. It doesn't seem necessary to install Spark on all YARN cluster nodes because Spark runs on top of the YARN. Apache Spark runs independently from its installation. Spark provides some options to use YARN when dispatching jobs to the cluster, rather than its built-in manager or Mesos. Besides this, there are also some configurations to run YARN, such as master, deploy-mode, driver-memory, executor-memory, executor-cores, and queue.

Related questions

+1 vote
asked Sep 16, 2022 in Spark Preliminaries by sharadyadav1986
0 votes
asked Jul 23, 2020 in IONIC by Hodge
...