Lets Start With Hadoop.

           HADOOP AND ITS ARCHITECTURE


Introduction and Big Data:
Click Here

Hadoop is an open source framework (OOF) written in java.  It allows to distributed processing of large amount of data across clusters of computers using simple programming models. A Hadoop frame-worked application works in an environment that provides distributed storage and computation across clusters of computers. Hadoop is designed to scale up from single server to thousands of machines, each offering local computation and storage.

Hadoop Architecture:

It has 4 level Architecture or we can say that it has four modules. 

1     1. Hadoop Common: This part of programming module . it consists of Java         Libraries and utilities required by other hadoop modules. These                     libraries provides file Management or File system and OS and it also             contain required java files and script which are required to start                     hadoop.

       2. Hadoop YARN: This is a framework used to manage the cluster                       resources and job scheduling.

       3. Hadoop Distributed File System (HDFS): It is system used to manage             data in hadoop . it based on the google file system and provides a                   distributed file system that is designed to run thousands of small                   computer machines in a reliable, fault-tolerant manner.


       4. Hadoop MapReduce: It is based on YARN for parallel processing of               large amount of data. It is software framework which process big                   amount of data in parallel on large cluster of commodity hardware in a         reliable, fault-tolerant manner.
        
       Hadoop Architecture Diagram
       
       

       



Share this

Related Posts

Previous
Next Post »

Powered by Blogger.

Popular Posts