+1 vote
in Big Data | Hadoop by
What do you know about Block and Block scanner in HDFS?

1 Answer

0 votes
by

A large file in HDFS is broken into multiple parts and each part is stored on a different Block. By default a Block is of 64 MB capacity in HDFS.

Block Scanner is a program that every Data node in HDFS runs periodically to verify the checksum of every block stored on the data node.

 

The purpose of a Block Scanner is to detect any data corruption errors on Data node.

Related questions

+1 vote
asked Jan 26, 2020 in Big Data | Hadoop by rajeshsharma
0 votes
asked Jan 26, 2020 in Big Data | Hadoop by rajeshsharma
...