Hadoop Tutorial for Big Data Enthusiasts – The Optimal way of Learning Hadoop

Hadoop Tutorial – One of the most searched terms on the internet today. Do you know the reason? It is because Hadoop is the major part or framework of Big Data.

If you don’t know anything about Big Data then you are in major trouble. But don’t worry I have something for you which is completely FREE 520+ Big Data Tutorials This free tutorial series will make you a master of Big Data in just few weeks. Also, I have explained a little about Big Data in this blog.

“Hadoop is a technology to store massive datasets on a cluster of cheap machines in a distributed manner”. It was originated by Doug Cutting and Mike Cafarella.

Doug Cutting’s kid named Hadoop to one of his toy that was a yellow elephant. Doug then used the name for his open source project because it was easy to spell, pronounce, and not used elsewhere.

Interesting, right?

Keeping you updated with latest technology trends, Join DataFlair on Telegram

Hadoop Tutorial

Now, let’s begin our interesting Hadoop tutorial with the basic introduction to Big Data.

What is Big Data?

Big Data refers to the datasets too large and complex for traditional systems to store and process. The major problems faced by Big Data majorly falls under three Vs. They are volume, velocity, and variety.

Do you know – Every minute we send 204 million emails, generate 1.8 million Facebook likes, send 278 thousand Tweets, and up-load 200,000 photos to Facebook. 

Volume: The data is getting generated in order of Tera to petabytes. The largest contributor of data is social media. For instance, Facebook generates 500 TB of data every day. Twitter generates 8TB of data daily.

Velocity: Every enterprise has its own requirement of the time frame within which they have process data. Many use cases like credit card fraud detection have only a few seconds to process the data in real-time and detect fraud. Hence there is a need of framework which is capable of high-speed data computations.

Variety: Also the data from various sources have varied formats like text, XML, images, audio, video, etc. Hence the Big Data technology should have the capability of performing analytics on a variety of data.

Big Data Quotes

Hope you have checked the Free Big Data DataFlair Tutorial Series. Here is one more interesting article for you – Top Big Data Quotes by the Experts

Why Hadoop is Invented?

Let us discuss the shortcomings of the traditional approach which led to the invention of Hadoop – 

1. Storage for Large Datasets

The conventional RDBMS is incapable of storing huge amounts of Data. The cost of data storage in available RDBMS is very high. As it incurs the cost of hardware and software both.

2. Handling data in different formats

The RDBMS is capable of storing and manipulating data in a structured format. But in the real world we have to deal with data in a structured, unstructured and semi-structured format.

 3. Data getting generated with high speed:

The data in oozing out in the order of tera to peta bytes daily. Hence we need a system to process data in real-time within a few seconds. The traditional RDBMS fail to provide real-time processing at great speeds.

If these professionals can make a switch to Big Data, so can you:
Rahul Doddamani Story - DataFlair
Rahul Doddamani
Java → Big Data Consultant, JDA
Follow on
Mritunjay Singh Success Story - DataFlair
Mritunjay Singh
PeopleSoft → Big Data Architect, Hexaware
Follow on
Rahul Doddamani Success Story - DataFlair
Rahul Doddamani
Big Data Consultant, JDA
Follow on
I got placed, scored 100% hike, and transformed my career with DataFlair
Enroll now
Richa Tandon Success Story - DataFlair
Richa Tandon
Support → Big Data Engineer, IBM
Follow on
DataFlair Web Services
You could be next!
Enroll now

What is Hadoop?

Hadoop is the solution to above Big Data problems. It is the technology to store massive datasets on a cluster of cheap machines in a distributed manner. Not only this it provides Big Data analytics through distributed computing framework.

It is an open-source software developed as a project by Apache Software Foundation. Doug Cutting created Hadoop. In the year 2008 Yahoo gave Hadoop to Apache Software Foundation. Since then two versions of Hadoop has come. Version 1.0 in the year 2011 and version 2.0.6 in the year 2013. Hadoop comes in various flavors like Cloudera, IBM BigInsight, MapR and Hortonworks.

Prerequisites to Learn Hadoop

  • Familiarity with some basic Linux Command – Hadoop is set up over Linux Operating System preferable Ubuntu. So one must know certain basic Linux commands. These commands are for uploading the file in HDFS, downloading the file from HDFS and so on.
  • Basic Java concepts – Folks want to learn Hadoop can get started in Hadoop while simultaneously grasping basic concepts of Java. We can write map and reduce functions in Hadoop using other languages too. And these are Python, Perl, C, Ruby, etc. This is possible via streaming API. It supports reading from standard input and writing to standard output. Hadoop also has high-level abstractions tools like Pig and Hive which do not require familiarity with Java.

Big Data Hadoop Tutorial Video

Wait! understand the concepts of Hadoop through this awesome video – 

Hope the above Big Data Hadoop Tutorial video helped you. Let us see further.

Hadoop consists of three core components –

  • Hadoop Distributed File System (HDFS) – It is the storage layer of Hadoop.
  • Map-Reduce – It is the data processing layer of Hadoop.
  • YARN – It is the resource management layer of Hadoop.

Core Components of Hadoop

Let us understand these Hadoop components in detail.

1. HDFS

Short for Hadoop Distributed File System provides for distributed storage for Hadoop. HDFS has a master-slave topology.

Hadoop working

Master is a high-end machine where as slaves are inexpensive computers. The Big Data files get divided into the number of blocks. Hadoop stores these blocks in a distributed fashion on the cluster of slave nodes. On the master, we have metadata stored.

HDFS has two daemons running for it. They are :

NameNode : NameNode performs following functions –

  • NameNode Daemon runs on the master machine.
  • It is responsible for maintaining, monitoring and managing DataNodes.
  • It records the metadata of the files like the location of blocks, file size, permission, hierarchy etc.
  • Namenode captures all the changes to the metadata like deletion, creation and renaming of the file in edit logs.
  • It regularly receives heartbeat and block reports from the DataNodes.

DataNode: The various functions of DataNode are as follows –

  • DataNode runs on the slave machine.
  • It stores the actual business data.
  • It serves the read-write request from the user.
  • DataNode does the ground work of creating, replicating and deleting the blocks on the command of NameNode.
  • After every 3 seconds, by default, it sends heartbeat to NameNode reporting the health of HDFS.

Explore the top features of HDFS that a Hadoop developer must know

Erasure Coding in HDFS

Till Hadoop 2.x replication is the only method for providing fault tolerance. Hadoop 3.0 introduces one more method called erasure coding. Erasure coding provides the same level of fault tolerance but with lower storage overhead.

Erasure coding is usually used in RAID (Redundant Array of Inexpensive Disks) kind of storage. RAID provides erasure coding via striping. In this, it divides the data into smaller units like bit/byte/block and stores the consecutive units on different disks. Hadoop calculates parity bits for each of these cell (units). We call this process as encoding. On the event of loss of certain cells, Hadoop computes these by decoding. Decoding is a process in which lost cells gets recovered from remaining original and parity cells.

Erasure coding is mostly used for warm or cold data which undergo less frequent I/O access. The replication factor of Erasure coded file is always one. we cannot change it by -setrep command. Under erasure coding storage overhead is never more than 50%.

Under conventional Hadoop storage replication factor of 3 is default. It means 6 blocks will get replicated into 6*3 i.e. 18 blocks. This gives a storage overhead of 200%. As opposed to this in Erasure coding technique there are 6 data blocks and 3 parity blocks. This gives storage overhead of 50%.

The File System Namespace

HDFS supports hierarchical file organization. One can create, remove, move or rename a file. NameNode maintains file system Namespace. NameNode records the changes in the Namespace. It also stores the replication factor of the file.

2. MapReduce 

It is the data processing layer of Hadoop. It processes data in two phases.

They are:-

Map Phase- This phase applies business logic to the data. The input data gets converted into key-value pairs.

Reduce Phase- The Reduce phase takes as input the output of Map Phase. It applies aggregation based on the key of the key-value pairs.

Hadoop MapReduce Working

You must check this MapReduce tutorial to start your learning. 

Map-Reduce works in the following way:

  • The client specifies the file for input to the Map function. It splits it into tuples
  • Map function defines key and value from the input file. The output of the map function is this key-value pair.
  • MapReduce framework sorts the key-value pair from map function.
  • The framework merges the tuples having the same key together.
  • The reducers get these merged key-value pairs as input.
  • Reducer applies aggregate functions on key-value pair.
  • The output from the reducer gets written to HDFS.

3. YARN

Short for Yet Another Resource Locator has the following components:-

Resource Manager

How resource manager works

  • Resource Manager runs on the master node.
  • It knows where the location of slaves (Rack Awareness).
  • It is aware about how much resources each slave have.
  • Resource Scheduler is one of the important service run by the Resource Manager.
  • Resource Scheduler decides how the resources get assigned to various tasks.
  • Application Manager is one more service run by Resource Manager.
  • Application Manager negotiates the first container for an application.
  • Resource Manager keeps track of the heart beats from the Node Manager.
Node Manager

How node manager works

  • It runs on slave machines.
  • It manages containers. Containers are nothing but a fraction of Node Manager’s resource capacity
  • Node manager monitors resource utilization of each container.
  • It sends heartbeat to Resource Manager.
Job Submitter

Job submitter in Yarn

The application startup process is as follows:-

  • The client submits the job to Resource Manager.
  • Resource Manager contacts Resource Scheduler and allocates container.
  • Now Resource Manager contacts the relevant Node Manager to launch the container.
  • Container runs Application Master.

The basic idea of YARN was to split the task of resource management and job scheduling. It has one global Resource Manager and per-application Application Master. An application can be either one job or DAG of jobs.

The Resource Manager’s job is to assign resources to various competing applications. Node Manager runs on the slave nodes. It is responsible for containers, monitoring resource utilization and informing about the same to Resource Manager.

The job of Application master is to negotiate resources from the Resource Manager. It also works with NodeManager to execute and monitor the tasks.

Wait before scrolling further! This is the time to read about the top 15 Hadoop Ecosystem components

Why Hadoop?

Let us now understand why Big Data Hadoop is very popular, why Apache Hadoop capture more than 90% of the big data market.

Apache Hadoop is not only a storage system but is a platform for data storage as well as processing. It is scalable (as we can add more nodes on the fly), Fault-tolerant (Even if nodes go down, data processed by another node).
Following characteristics of Hadoop make it a unique platform:

  • Flexibility to store and mine any type of data whether it is structured, semi-structured or unstructured. It is not bounded by a single schema.
  • Excels at processing data of complex nature. Its scale-out architecture divides workloads across many nodes. Another added advantage is that its flexible file-system eliminates ETL bottlenecks.
  • Scales economically, as discussed it can deploy on commodity hardware. Apart from this its open-source nature guards against vendor lock.

What is Hadoop Architecture?

Hadoop Quiz

After understanding what is Apache Hadoop, let us now understand the Hadoop Architecture in detail.

Hadoo Works

How Hadoop Works

Hadoop works in master-slave fashion. There is a master node and there are n numbers of slave nodes where n can be 1000s. Master manages, maintains and monitors the slaves while slaves are the actual worker nodes. In Hadoop architecture, the Master should deploy on good configuration hardware, not just commodity hardware. As it is the centerpiece of Hadoop cluster.

Master stores the metadata (data about data) while slaves are the nodes which store the data. Distributedly data stores in the cluster. The client connects with the master node to perform any task. Now in this Hadoop tutorial for beginners, we will discuss different features of Hadoop in detail.

Hadoop Features

Here are the top Hadoop features that make it popular –

1. Reliability

In the Hadoop cluster, if any node goes down, it will not disable the whole cluster. Instead, another node will take the place of the failed node. Hadoop cluster will continue functioning as nothing has happened. Hadoop has built-in fault tolerance feature.

2. Scalable

Hadoop gets integrated with cloud-based service. If you are installing Hadoop on the cloud you need not worry about scalability. You can easily procure more hardware and expand your Hadoop cluster within minutes.

3. Economical

Hadoop gets deployed on commodity hardware which is cheap machines. This makes Hadoop very economical. Also as Hadoop is an open system software there is no cost of license too.

4. Distributed Processing

In Hadoop, any job submitted by the client gets divided into the number of sub-tasks. These sub-tasks are independent of each other. Hence they execute in parallel giving high throughput.

5. Distributed Storage

Hadoop splits each file into the number of blocks. These blocks get stored distributedly on the cluster of machines.

6. Fault Tolerance

Hadoop replicates every block of file many times depending on the replication factor. Replication factor is 3 by default. In Hadoop suppose any node goes down then the data on that node gets recovered. This is because this copy of the data would be available on other nodes due to replication. Hadoop is fault tolerant.

Are you looking for more Features? Here are the additional Hadoop Features that make it special. 

Hadoop Flavors

This section of the Hadoop Tutorial talks about the various flavors of Hadoop.

  • Apache – Vanilla flavor, as the actual code is residing in Apache repositories.
  • Hortonworks – Popular distribution in the industry.
  • Cloudera – It is the most popular in the industry.
  • MapR – It has rewritten HDFS and its HDFS is faster as compared to others.
  • IBM – Proprietary distribution is known as Big Insights.

All the databases have provided native connectivity with Hadoop for fast data transfer. Because, to transfer data from Oracle to Hadoop, you need a connector.

All flavors are almost same and if you know one, you can easily work on other flavors as well.

Hadoop Future Scope

There is going to be a lot of investment in the Big Data industry in coming years. According to a report by FORBES, 90% of global organizations will be investing in Big Data technology. Hence the demand for Hadoop resources will also grow. Learning Apache Hadoop will give you accelerated growth in career. It also tends to increase your pay package.

There is a lot of gap between the supply and demand of Big Data professional. The skill in Big Data technologies continues to be in high demand. This is because companies grow as they try to get the most out of their data. Therefore, their salary package is quite high as compared to professionals in other technology.

The managing director of Dice, Alice Hills has said that Hadoop jobs have seen 64% increase from the previous year. It is evident that Hadoop is ruling the Big Data market and its future is bright. The demand for Big Data Analytics professional is ever increasing. As it is a known fact that data is nothing without power to analyze it.

You must check Expert’s Prediction for the Future of Hadoop

Summary – Hadoop Tutorial

On concluding this Hadoop tutorial, we can say that Apache Hadoop is the most popular and powerful big data tool. Big Data stores huge amount of data in the distributed manner and processes the data in parallel on a cluster of nodes. It provides the world’s most reliable storage layer- HDFS. Batch processing engine MapReduce and Resource management layer- YARN.

On summarizing this Hadoop Tutorial, I want to give you a quick revision of all the topics we have discussed

  • The concept of Big Data
  • Reason for Hadoop’s Invention
  • Prerequisites to learn Hadoop
  • Introduction to Hadoop
  • Core components of Hadoop
  • Why Hadoop
  • Hadoop Architecture
  • Features of Hadoop
  • Hadoop Flavours
  • Future Scope of Hadoop

Hope this Hadoop Tutorial helped you. If you face any difficulty while understanding Hadoop concept, comment below.

This is the right time to start your Hadoop learning with industry experts

59 Responses

  1. ramesh says:

    good content

  2. Duke says:

    Very informative article, good explaination of the Hadoop Framework.

  3. Whitelimo says:

    very good article, i blind with hadoop before

Leave a Reply

Your email address will not be published. Required fields are marked *

This site is protected by reCAPTCHA and the Google Privacy Policy and Terms of Service apply.