Scenario: You are tasked with processing a large batch of log data stored in HDFS and generating summary reports. Which Hadoop component would you use for this task, and why?

  • Apache Hadoop MapReduce
  • Apache Kafka
  • Apache Pig
  • Apache Sqoop
Apache Hadoop MapReduce is ideal for processing large batch data stored in HDFS and generating summary reports. It provides a scalable and fault-tolerant framework for parallel processing of distributed data.
Add your answer
Loading...

Leave a comment

Your email address will not be published. Required fields are marked *