HADOOP COMPONENTS & FAULT TOLERANCE

##plugins.themes.bootstrap3.article.main##

Shyamala S Pushparani M

Abstract

In this modern trend of information technology and computer science storing and processing a data is very important. Nowadays even a terabytes and petabytes of data is not enough for storing large amount of database. Hadoop is designed to store large chunks of data sets reliably. Hadoop is an open source software which supports parallel and distributed data processing. It is highly scalable compute platform. Hadoop enables users to store and process large amount of data which is not possible while using less scalable techniques. Hadoop also provide faults tolerance mechanism by which system continues to function correctly even after some components fail’s working properly. Faults tolerance is mainly achieved using data duplication and making copies of same data sets in two or more data nodes. In this paper we describe the major components of Hadoop along with how fault tolerance is achieved by means of data duplication, check point and recovery

##plugins.themes.bootstrap3.article.details##

Section
Articles