0
0
Hadoopdata~5 mins

Block storage and replication in Hadoop - Cheat Sheet & Quick Revision

Choose your learning style9 modes available
Recall & Review
beginner
What is block storage in Hadoop?
Block storage in Hadoop means breaking large files into smaller fixed-size pieces called blocks. These blocks are stored across different machines to manage big data efficiently.
Click to reveal answer
beginner
Why does Hadoop replicate blocks?
Hadoop replicates blocks to make sure data is safe and available even if some machines fail. Replication copies blocks to multiple machines to avoid data loss.
Click to reveal answer
intermediate
What is the default block size in Hadoop HDFS?
The default block size in Hadoop HDFS is usually 128 MB. This size helps balance between managing many blocks and efficient data processing.
Click to reveal answer
intermediate
How does replication factor affect storage in Hadoop?
Replication factor is the number of copies of each block. A higher replication factor means more copies, which increases storage use but improves fault tolerance.
Click to reveal answer
advanced
Explain how Hadoop ensures data reliability using block storage and replication.
Hadoop splits files into blocks and stores them on different machines. It makes multiple copies (replicas) of each block. If one machine fails, other copies keep data safe and accessible.
Click to reveal answer
What is the main purpose of block replication in Hadoop?
ATo speed up file deletion
BTo improve data safety and availability
CTo reduce the file size
DTo compress data blocks
What happens to a large file in Hadoop before storage?
AIt is split into smaller blocks
BIt is stored as one big file
CIt is converted to text
DIt is deleted
What is the typical default block size in Hadoop HDFS?
A10 MB
B1 GB
C64 KB
D128 MB
If the replication factor is 3, how many copies of each block exist?
A3
B2
C1
D4
Which of these is NOT a benefit of block replication?
AData fault tolerance
BIncreased storage space
CFaster data deletion
DData availability
Describe how Hadoop uses block storage and replication to protect data.
Think about how breaking files and copying blocks helps keep data safe.
You got /4 concepts.
    Explain the role of the replication factor in Hadoop's data storage system.
    Consider how more copies affect storage space and data safety.
    You got /4 concepts.