Posted On: Feb 22, 2018
Hadoop Distributed File System (HDFS) stores documents as information blocks and circulates these blocks over the whole cluster. As HDFS was intended to be fault tolerant and to keep running on ware equipment, blocks are replicated in various circumstances to guarantee high data accessibility.
Furthermore Yes, I can change the block size of HDFS records by changing the default size parameter show in hdfs-site.xml. But after changing, I have to restart the cluster for this property change to take effect.
Never Miss an Articles from us.