HADOOP AND ITS ARCHITECTURE
Click Here
Hadoop is an
open source framework (OOF) written in java.
It allows to distributed processing of large amount of data across
clusters of computers using simple programming models. A Hadoop frame-worked
application works in an environment that provides distributed storage and
computation across clusters of computers. Hadoop is designed to scale up from
single server to thousands of machines, each offering local computation and
storage.
Hadoop
Architecture:
It has 4
level Architecture or we can say that it has four modules.
1 1. Hadoop Common: This part of
programming module . it consists of Java Libraries and utilities required by
other hadoop modules. These libraries provides file Management or File system
and OS and it also contain required java files and script which are required to
start hadoop.
2. Hadoop YARN: This is a framework used
to manage the cluster resources and job scheduling.
3. Hadoop Distributed File System
(HDFS): It is system used to manage data in hadoop . it based on the google
file system and provides a distributed file system that is designed to run
thousands of small computer machines in a reliable, fault-tolerant manner.
4. Hadoop MapReduce: It is based on YARN
for parallel processing of large amount of data. It is software framework which
process big amount of data in parallel on large cluster of commodity hardware
in a reliable, fault-tolerant manner.
Hadoop
Architecture Diagram