0 votes
in Apache Spark by
What is RDD in Spark?

1 Answer

0 votes
by

RDD in Apache Spark stands for Resilient Distributed Datasets. It is a fundamental data structure of Spark that acts as an immutable distributed collection of objects. Each dataset in RDD is divided into logical partitions, which are computed on different cluster nodes. RDDs can contain any Python, Java, or Scala objects, including user-defined classes.

Related questions

0 votes
asked Mar 14, 2020 in Spark Sql by rajeshsharma
0 votes
asked Mar 8, 2020 in Spark Sql by rahuljain1
...