This blog is to give brief introduction about Hadoop for those who know next to nothing about this technology. Big Data is at the foundation of all the megatrends that are happening today, from social to the cloud to mobile devices to gaming. This blog will help to build the foundation to take the next step in learning this interesting technology. Let's get started:
1. What's Big Data?
Ever since the enhancement of technology, data has been growing every day. Everyone owns gadgets nowadays. Every smart device generates data. One of the prominent sources of data is social media. We, being social animal love to share our thoughts, feelings with others and social media is the right platform for the interaction with other all around the world.
The following image shows data generated by users on the social media every 60 seconds. Data has been exponentially getting generated through these sources.
The following image show the Global Mobile Data Traffic prediction by Cisco till 2020.
Hence, Big Data is:
The term used for a collection of data sets so large and complex that it becomes difficult to process using on-hand database management tools or traditional data processing applications.
IBM data scientist break big data into 4 dimensions: Volume (Scale of data), Variety (Different forms of data), Velocity (Analysis of streaming data) and Veracity (Uncertainty of data).
Depending on the industry and organization, big data encompasses information from various internal and external sources such as transactions, social media, enterprise content, sensors and mobile devices etc. Companies can leverage data to meet their customer needs, optimize their products, services and operations. This massive amount of data can also be used by the companies to find new sources of the revenue.
2. How did Hadoop come into picture?
These massive amounts of data generated is difficult to store and process using traditional database system. Traditional database management system is used to store and process relational and structured data only. However, in todays world there are lots of unstructured data getting generated like images, audio files, videos; hence traditional system will fail to store and process these kinds of data. Effective solution for this problem is Hadoop.
Hadoop is a framework to process Big Data. It is a framework that allows to store and process large data sets in parallel and distributed fashion.
Hadoop Core Components:
There are two main components of Hadoop: HDFS and MapReduce
Hadoop Distributed File System (HDFS) takes care of storage part of Hadoop architecture.
MapReduce is a processing model and software framework for writing applications which can run on Hadoop. These programs of MapReduce are capable of processing Big Data in parallel on large clusters of computational nodes.
3. What's HDFS and what are its core components?
HDFS stores files across many nodes in a cluster.
Hadoop follows Master-Slave architecture and hence HDFS being its core component also follows the same architecture.
NameNode and DataNode are the core components of HDFS:
Final fsimage in the Secondary NameNode allows faster failover as it prevents edit logs in the NameNode from getting too huge.New edit log file in the NameNode contains all the modifications/changes that happen during the checkinpointing.
Each file is stored on HDFS as blocks. The default size of each block is 128 MB in Apache Hadoop 2.x (64 MB in Apache Hadoop 1.x)
After file is divided into data blocks as shown in the below figure, these data blocks will be then distributed across all the Data Nodes present in the Hadoop cluster.
1> Fault Tolerance – Each data blocks are replicated thrice ((everything is stored on three machines/DataNodes by default) in the cluster. This helps to protect the data against DataNode (machine) failure.
2> Space – Just add more datanodes and re-balance the size if you need more disk space.
3> Scalability – Unlike traditional database system that can't scale to process large datasets; HDFS is highly scalable because it can store and distribute very large datasets across many nodes that can operate in parallel.
4> Flexibility – It can store any kind of data, whether its structured, semi-structured or unstructured.
5> Cost-effective – HDFS has direct attached storage and shares the cost of the network and computers it runs on with the MapReduce. It's also an open source software.
In the next blog we shall discuss about MapReduce, another core component of Hadoop. Stay tuned.