Jan 11, 2020 in Big Data | Hadoop
Q: What do you know about Block and Block scanner in HDFS?

1 Answer

0 votes
Jan 11, 2020

A large file in HDFS is broken into multiple parts and each part is stored on a different Block. By default a Block is of 64 MB capacity in HDFS.

Block Scanner is a program that every Data node in HDFS runs periodically to verify the checksum of every block stored on the data node.

 

The purpose of a Block Scanner is to detect any data corruption errors on Data node.

Related questions

0 votes
Mar 25, 2020 in SAP
0 votes
Apr 15, 2020 in Robotic Process Automation
...