The process of persisting intermediate data in memory to avoid recomputation in Apache Spark is called ________.

  • Caching
  • Checkpointing
  • Repartitioning
  • Serialization
In Apache Spark, the process of persisting intermediate data in memory to avoid recomputation is known as caching. This technique enhances performance by storing RDDs or DataFrames in memory for reuse in subsequent operations, reducing the need for recomputation.
Add your answer
Loading...

Leave a comment

Your email address will not be published. Required fields are marked *