Download as docx, pdf, or txt
Download as docx, pdf, or txt
You are on page 1of 1

Q.

Ans: User applications access the file system using the HDFS client, a code library that exports the HDFS
file system inter-face. When an application reads a file, the HDFS client first asks the NameNode for the
list of DataNodes that host replicas of the blocks of the file. It then contacts a DataNode directly and
requests the transfer of the desired block. When a client writes, it first asks the NameNode to choose
DataNodes to host replicas of the first block of the file. The client organizes a pipeline from node-to-
node and sends the data. When the first block is filled, the client requests new DataNodes to be chosen
to host replicas of the next block. A new pipeline is organized, and the client sends the further bytes of
the file. Each choice of DataNodes is likely to be different.

Q.2

Google File System is a proprietary distributed file system and is exclusive for Google Inc. and developed
to accommodate Google’s expanding data processing requirements. GFS provides fault tolerance,
reliability, scalability, availability, and performance to large networks and connected nodes. GFS is made
up of several storage systems built from low-cost commodity hardware components.

How is Hadoop different from GFS?

The concept of Hadoop is basically derived from google file system hence the Hadoop distributed file
system is based on the Google file system. MapReduce is the programming framework used by Google.
Hadoop Distributed File System and MapReduce are the components of Hadoop project owned by
Apache. Hadoop MapReduce is based on the idea of the Google MapReduce.

You might also like