Bird
0
0

What is a typical performance issue caused by storing numerous tiny files in HDFS?

easy📝 Conceptual Q1 of 15
Hadoop - Performance Tuning
What is a typical performance issue caused by storing numerous tiny files in HDFS?
AImproved fault tolerance automatically
BFaster data retrieval due to parallel file access
CReduced disk space consumption
DIncreased NameNode memory usage and slower metadata operations
Step-by-Step Solution
Solution:
  1. Step 1: Understand HDFS metadata handling

    Each file, regardless of size, requires metadata storage in the NameNode's memory.
  2. Step 2: Impact of many small files

    Having many small files increases metadata load, causing higher memory usage and slower file system operations.
  3. Final Answer:

    Increased NameNode memory usage and slower metadata operations -> Option D
  4. Quick Check:

    Small files increase metadata overhead [OK]
Quick Trick: Many small files overload NameNode memory [OK]
Common Mistakes:
  • Assuming small files improve read speed
  • Believing small files reduce disk usage
  • Thinking fault tolerance improves automatically

Want More Practice?

15+ quiz questions · All difficulty levels · Free

Free Signup - Practice All Questions
More Hadoop Quizzes