DATABASE FUNDAMENTALS
BASICS OF BIG DATA
Question
[CLICK ON ANY CHOICE TO KNOW THE RIGHT ANSWER]
|
|
HDFS
|
|
PIG
|
|
HIVE
|
|
HADOOP
|
Detailed explanation-1: -HDFS provides fault tolerance by replicating the data blocks and distributing it among different DataNodes across the cluster.
Detailed explanation-2: -HDFS is highly fault tolerant. It uses replica process to handle faults. This means client data is repeated many times (default replica factor is 3) on different DataNode in the HDFS cluster. So that in case of any DataNode goes down, the data will be accessed from other DataNodes.
Detailed explanation-3: -Fault tolerance in Hadoop HDFS refers to the working strength of a system in unfavorable conditions and how that system can handle such a situation. HDFS is highly fault-tolerant. Before Hadoop 3, it handles faults by the process of replica creation.
Detailed explanation-4: -The necessity for re-replication may arise due to many reasons: a DataNode may become unavailable, a replica may become corrupted, a hard disk on a DataNode may fail, or the replication factor of a file may be increased.