When planning for disaster recovery, how should a Hadoop administrator prioritize data in different HDFS directories?
- Prioritize based on access frequency
- Prioritize based on creation date
- Prioritize based on file size
- Prioritize based on replication factor
A Hadoop administrator should prioritize data in different HDFS directories based on the replication factor. Critical data should have a higher replication factor to ensure availability and fault tolerance in the event of node failures.
Loading...
Related Quiz
- In YARN, ____ is a critical process that optimizes the use of resources across the cluster.
- In a scenario involving time-series data storage, what HBase feature would be most beneficial?
- To troubleshoot connectivity issues between nodes, a Hadoop administrator should check the ____ configurations.
- What makes Apache Flume highly suitable for event-driven data ingestion into Hadoop?
- In Hive, the storage of metadata is managed by which component?