in Spark Sql by
Q:
How Apache Spark uses Akka in Spark?

1 Answer

0 votes
by

The main purpose of using Akka by the Spark is to schedule. After the process of registering all the workers requested for the task to the master. All the tasks are assigned by the master. The messaging between the masters and workers can be performed with Akka.

Click here to read more about Loan/Mortgage
Click here to read more about Insurance

Related questions

0 votes
asked Mar 14, 2020 in Spark Sql by rajeshsharma
0 votes
asked Mar 14, 2020 in Spark Sql by rajeshsharma
+1 vote
asked Mar 9, 2020 in Spark Sql by SakshiSharma
0 votes
asked Jan 13, 2020 in Big Data | Hadoop by sharadyadav1986
0 votes
asked Mar 14, 2020 in Spark Sql by rajeshsharma
...