0 votes
in Spark Preliminaries by
When can you use Apache Spark? What is better about Spark than MapReduce?

1 Answer

0 votes
by

Spark moves pretty quickly. Hadoop MapReduce is ten times slower in memory than other programming frameworks. It uses RAM in the right way so that it works faster.

In Map Reduce Paradigm, you write a lot of Map-Reduce tasks and then use the Oozie/shell script to link these tasks together. This process takes a long time, and the role of map-reducing is slow.

Changing production from one MR job to another MR job can sometimes require writing more code because Oozie may need to be more.

Spark lets you do everything from a single application or console and get the results immediately. It's pretty easy to switch between "doing something locally" and "Running something on a cluster." This means that the creator has less background change and can work faster.

Related questions

+1 vote
asked Sep 16, 2022 in Spark Preliminaries by sharadyadav1986
0 votes
asked Aug 25, 2022 in Apache Spark by sharadyadav1986
...