Jan 11 in Big Data | Hadoop
Q: What do you know about Block and Block scanner in HDFS?

1 Answer

Jan 11

A large file in HDFS is broken into multiple parts and each part is stored on a different Block. By default a Block is of 64 MB capacity in HDFS.

Block Scanner is a program that every Data node in HDFS runs periodically to verify the checksum of every block stored on the data node.


The purpose of a Block Scanner is to detect any data corruption errors on Data node.

Click here to read more about Loan/Mortgage
Click here to read more about Insurance

Related questions

Mar 25 in SAP
Apr 15 in Robotic Process Automation
Feb 23 in Big Data | Hadoop