Apache Spark's core data structure, used for distributed data processing, is called what?

  • RDD (Resilient Distributed Dataset)
  • Dataframe
  • HDFS (Hadoop Distributed File System)
  • NoSQL
Apache Spark uses RDD (Resilient Distributed Dataset) as its core data structure for distributed data processing. RDDs are immutable, fault-tolerant collections of data that can be processed in parallel.
Add your answer
Loading...

Leave a comment

Your email address will not be published. Required fields are marked *