Bird
0
0

You want to optimize storage by compressing Parquet files with Snappy codec and also ensure schema evolution compatibility. Which approach is best?

hard📝 Application Q9 of 15
Hadoop - Performance Tuning
You want to optimize storage by compressing Parquet files with Snappy codec and also ensure schema evolution compatibility. Which approach is best?
AWrite Parquet files with snappy compression and use Avro schema for evolution
BWrite Parquet files with gzip compression and no schema
CWrite ORC files with snappy compression and no schema
DWrite Avro files with snappy compression and no schema
Step-by-Step Solution
Solution:
  1. Step 1: Understand compression and schema evolution

    Snappy compression is fast and commonly used with Parquet. Avro schemas help with schema evolution.
  2. Step 2: Evaluate options

    Write Parquet files with snappy compression and use Avro schema for evolution combines Parquet with snappy and Avro schema for evolution. Others either lack schema or use wrong formats.
  3. Final Answer:

    Write Parquet files with snappy compression and use Avro schema for evolution -> Option A
  4. Quick Check:

    Snappy + Avro schema = best for Parquet evolution [OK]
Quick Trick: Use snappy compression and Avro schema for Parquet evolution [OK]
Common Mistakes:
  • Ignoring schema for evolution
  • Using gzip instead of snappy
  • Confusing ORC and Avro formats

Want More Practice?

15+ quiz questions · All difficulty levels · Free

Free Signup - Practice All Questions
More Hadoop Quizzes