in Big Data | Hadoop by
Q:
What are the two ways to create RDD in Spark?

▼ Show 1 Answer

0 votes
by

We can create RDD in Spark in following two ways:

1. Internal: We can parallelize an existing collection of data within our Spark Driver program and create a RDD out of it.

2. External: We can also create RDD by referencing a Dataset in an external data

 

source like AWS S3, HDFS, HBASE etc.

 

Learn More with Madanswer

Related questions

0 votes
asked Jun 8, 2020 in Spark Sql by DavidAnderson
...