SlideShare a Scribd company logo
TIB Academy,
5/3, Varathur Road, Kundalahalli Gate,
Bangalore-560066.
+91-9513332301 / 02 www.tibacademy.in
 Open source software framework designed
for storage and processing of large scale
data on clusters of commodity hardware
 Created by Doug Cutting and Mike Carafella
in 2005.
 Cutting named the program after his son’s
toy elephant.
 Data-intensive text processing
 Assembly of large genomes
 Graph mining
 Machine learning and data mining
 Large scale social network analysis
Hadoop tutorial for beginners-tibacademy.in
• Contains Libraries and other
modules
Hadoop
Common
• Hadoop Distributed File SystemHDFS
• Yet Another Resource
Negotiator
Hadoop
YARN
• A programming model for large
scale data processing
Hadoop
MapReduce
Hadoop tutorial for beginners-tibacademy.in
 What were the limitations of earlier large-
scale computing?
 What requirements should an alternative
approach have?
 How does Hadoop address those
requirements?
 Historically computation was processor-
bound
› Data volume has been relatively small
› Complicated computations are performed on that
data
 Advances in computer technology has
historically centered around improving the
power of a single machine
Hadoop tutorial for beginners-tibacademy.in
 Moore’s Law
› The number of transistors on a dense integrated
circuit doubles every two years
 Single-core computing can’t scale with
current computing needs
 Power consumption limits the speed
increase we get from transistor density
 Allows developers
to use multiple
machines for a
single task
 Programming on a distributed system is
much more complex
› Synchronizing data exchanges
› Managing a finite bandwidth
› Controlling computation timing is complicated
“You know you have a distributed system when
the crash of a computer you’ve never
heard of stops you from getting any work
done.” –Leslie Lamport
 Distributed systems must be designed with
the expectation of failure
 Typically divided into Data Nodes and
Compute Nodes
 At compute time, data is copied to the
Compute Nodes
 Fine for relatively small amounts of data
 Modern systems deal with far more data
than was gathering in the past
 Facebook
› 500 TB per day
 Yahoo
› Over 170 PB
 eBay
› Over 6 PB
 Getting the data to the processors becomes
the bottleneck
 Must support partial
failure
 Must be scalable
 Failure of a single component must not cause
the failure of the entire system only a
degradation of the application performance
 Failure should not
result in the loss of
any data
 If a component fails, it should be able to
recover without restarting the entire system
 Component failure or recovery during a job
must not affect the final output
 Increasing resources should increase load
capacity
 Increasing the load on the system should
result in a graceful decline in performance
for all jobs
› Not system failure
 Based on work done by Google in the early
2000s
› “The Google File System” in 2003
› “MapReduce: Simplified Data Processing on
Large Clusters” in 2004
 The core idea was to distribute the data as it
is initially stored
› Each node can then perform computation on the
data it stores without moving the data for the
initial processing
 Applications are written in a high-level
programming language
› No network programming or temporal dependency
 Nodes should communicate as little as possible
› A “shared nothing” architecture
 Data is spread among the machines in advance
› Perform computation where the data is already
stored as often as possible
 When data is loaded onto the system it is
divided into blocks
› Typically 64MB or 128MB
 Tasks are divided into two phases
› Map tasks which are done on small portions of data
where the data is stored
› Reduce tasks which combine data to produce the
final output
 A master program allocates work to individual
nodes
 Failures are detected by the master program
which reassigns the work to a different node
 Restarting a task does not affect the nodes
working on other portions of the data
 If a failed node restarts, it is added back to the
system and assigned new tasks
 The master can redundantly execute the same
task to avoid slow running nodes
Hadoop tutorial for beginners-tibacademy.in
 Responsible for storing data on the cluster
 Data files are split into blocks and distributed
across the nodes in the cluster
 Each block is replicated multiple times
 HDFS is a file system written in Java based
on the Google’s GFS
 Provides redundant storage for massive
amounts of data
 HDFS works best with a smaller number of
large files
› Millions as opposed to billions of files
› Typically 100MB or more per file
 Files in HDFS are write once
 Optimized for streaming reads of large files
and not random reads
 Files are split into blocks
 Blocks are split across many machines at load
time
› Different blocks from the same file will be stored on
different machines
 Blocks are replicated across multiple machines
 The NameNode keeps track of which blocks
make up a file and where they are stored
 Default replication is 3-fold
 When a client wants to retrieve data
› Communicates with the NameNode to determine
which blocks make up a file and on which data
nodes those blocks are stored
› Then communicated directly with the data nodes
to read the data
Hadoop tutorial for beginners-tibacademy.in
 A method for distributing computation across
multiple nodes
 Each node processes the data that is stored at
that node
 Consists of two main phases
› Map
› Reduce
 Automatic parallelization and distribution
 Fault-Tolerance
 Provides a clean abstraction for
programmers to use
 Reads data as key/value pairs
› The key is often discarded
 Outputs zero or more key/value pairs
 Output from the mapper is sorted by key
 All values with the same key are guaranteed
to go to the same machine
 Called once for each unique key
 Gets a list of all values associated with a key
as input
 The reducer outputs zero or more final
key/value pairs
› Usually just one output per input key
Hadoop tutorial for beginners-tibacademy.in
Hadoop tutorial for beginners-tibacademy.in
 NameNode
› Holds the metadata for the HDFS
 Secondary NameNode
› Performs housekeeping functions for the NameNode
 DataNode
› Stores the actual HDFS data blocks
 JobTracker
› Manages MapReduce jobs
 TaskTracker
› Monitors individual Map and Reduce tasks
 Stores the HDFS file system information in a
fsimage
 Updates to the file system (add/remove blocks)
do not change the fsimage file
› They are instead written to a log file
 When starting the NameNode loads the fsimage
file and then applies the changes in the log file
 NOT a backup for the NameNode
 Periodically reads the log file and applies the
changes to the fsimage file bringing it up to
date
 Allows the NameNode to restart faster when
required
 JobTracker
› Determines the execution plan for the job
› Assigns individual tasks
 TaskTracker
› Keeps track of the performance of an individual
mapper or reducer
Hadoop tutorial for beginners-tibacademy.in
 MapReduce is very powerful, but can be
awkward to master
 These tools allow programmers who are
familiar with other programming styles to
take advantage of the power of MapReduce
 Hive
› Hadoop processing with SQL
 Pig
› Hadoop processing with scripting
 Cascading
› Pipe and Filter processing model
 HBase
› Database model built on top of Hadoop
 Flume
› Designed for large scale data movement

More Related Content

What's hot (20)

PPSX
Hadoop-Quick introduction
Sandeep Singh
 
PPTX
عصر کلان داده، چرا و چگونه؟
datastack
 
PDF
getFamiliarWithHadoop
AmirReza Mohammadi
 
PPTX
Hadoop: The elephant in the room
cacois
 
PPTX
Hadoop Fundamentals
its_skm
 
PDF
Hadoop, HDFS and MapReduce
fvanvollenhoven
 
PPTX
Hadoop introduction
musrath mohammad
 
PPTX
Apache Hadoop India Summit 2011 talk "Hadoop Map-Reduce Programming & Best Pr...
Yahoo Developer Network
 
PDF
Hadoop Adventures At Spotify (Strata Conference + Hadoop World 2013)
Adam Kawa
 
PPTX
Big data Hadoop presentation
Shivanee garg
 
ODP
Hadoop seminar
KrishnenduKrishh
 
PPTX
2. hadoop fundamentals
Lokesh Ramaswamy
 
PPTX
Big data & hadoop
Abhi Goyan
 
PDF
Apache Hadoop - Big Data Engineering
BADR
 
PPTX
Apache hadoop technology : Beginners
Shweta Patnaik
 
PPT
Cloud Computing: Hadoop
darugar
 
PPTX
Introduction to Apache Hadoop
Christopher Pezza
 
PDF
Hadoop-2.6.0 Slides
kul prasad subedi
 
Hadoop-Quick introduction
Sandeep Singh
 
عصر کلان داده، چرا و چگونه؟
datastack
 
getFamiliarWithHadoop
AmirReza Mohammadi
 
Hadoop: The elephant in the room
cacois
 
Hadoop Fundamentals
its_skm
 
Hadoop, HDFS and MapReduce
fvanvollenhoven
 
Hadoop introduction
musrath mohammad
 
Apache Hadoop India Summit 2011 talk "Hadoop Map-Reduce Programming & Best Pr...
Yahoo Developer Network
 
Hadoop Adventures At Spotify (Strata Conference + Hadoop World 2013)
Adam Kawa
 
Big data Hadoop presentation
Shivanee garg
 
Hadoop seminar
KrishnenduKrishh
 
2. hadoop fundamentals
Lokesh Ramaswamy
 
Big data & hadoop
Abhi Goyan
 
Apache Hadoop - Big Data Engineering
BADR
 
Apache hadoop technology : Beginners
Shweta Patnaik
 
Cloud Computing: Hadoop
darugar
 
Introduction to Apache Hadoop
Christopher Pezza
 
Hadoop-2.6.0 Slides
kul prasad subedi
 

Viewers also liked (8)

PPT
Hadoop tutorial
Aamir Ameen
 
PDF
Hadoop Tutorial | What is Hadoop | Hadoop Project on Reddit | Edureka
Edureka!
 
PPTX
Hadoop or Spark: is it an either-or proposition? By Slim Baltagi
Slim Baltagi
 
PPTX
java tutorial for beginner - Free Download
TIB Academy
 
PDF
Introduction to Big Data Hadoop Training Online by www.itjobzone.biz
ITJobZone.biz
 
PDF
React Components Lifecycle | React Tutorial for Beginners | ReactJS Training ...
Edureka!
 
PDF
Big Data Tutorial For Beginners | What Is Big Data | Big Data Tutorial | Hado...
Edureka!
 
PPTX
AI and Machine Learning Demystified by Carol Smith at Midwest UX 2017
Carol Smith
 
Hadoop tutorial
Aamir Ameen
 
Hadoop Tutorial | What is Hadoop | Hadoop Project on Reddit | Edureka
Edureka!
 
Hadoop or Spark: is it an either-or proposition? By Slim Baltagi
Slim Baltagi
 
java tutorial for beginner - Free Download
TIB Academy
 
Introduction to Big Data Hadoop Training Online by www.itjobzone.biz
ITJobZone.biz
 
React Components Lifecycle | React Tutorial for Beginners | ReactJS Training ...
Edureka!
 
Big Data Tutorial For Beginners | What Is Big Data | Big Data Tutorial | Hado...
Edureka!
 
AI and Machine Learning Demystified by Carol Smith at Midwest UX 2017
Carol Smith
 
Ad

Similar to Hadoop tutorial for beginners-tibacademy.in (20)

PPTX
Big Data Unit 4 - Hadoop
RojaT4
 
PPTX
Apache Hadoop Big Data Technology
Jay Nagar
 
PPTX
Bigdata and Hadoop Introduction
umapavankumar kethavarapu
 
PDF
cloud computing notes for enginnering students
onkaps18
 
PPTX
Introduction to Hadoop and Big Data
Joe Alex
 
PPT
Hadoop mapreduce and yarn frame work- unit5
RojaT4
 
PPT
Hadoop
Girish Khanzode
 
PPTX
Big Data and Hadoop
Flavio Vit
 
PPT
hadoop
swatic018
 
PPT
hadoop
swatic018
 
PDF
Hadoop ecosystem; J.Ayeesha parveen 2 nd M.sc., computer science Bon Secours...
AyeeshaParveen
 
PDF
Introduction to Hadoop Administration
Ramesh Pabba - seeking new projects
 
PDF
Introduction to Hadoop Administration
Ramesh Pabba - seeking new projects
 
DOCX
Hadoop Seminar Report
Atul Kushwaha
 
PPTX
Managing Big data with Hadoop
Nalini Mehta
 
PDF
Hadoop ecosystem J.AYEESHA PARVEEN II-M.SC.,COMPUTER SCIENCE, BON SECOURS CO...
AyeeshaParveen
 
PPTX
HADOOP.pptx
Bharathi567510
 
PPTX
Hadoop
RittikaBaksi
 
PPT
Lecture Slide - Introduction to Hadoop, HDFS, MapR.ppt
SuchithraaPalani
 
PPTX
Hadoop.pptx
sonukumar379092
 
Big Data Unit 4 - Hadoop
RojaT4
 
Apache Hadoop Big Data Technology
Jay Nagar
 
Bigdata and Hadoop Introduction
umapavankumar kethavarapu
 
cloud computing notes for enginnering students
onkaps18
 
Introduction to Hadoop and Big Data
Joe Alex
 
Hadoop mapreduce and yarn frame work- unit5
RojaT4
 
Big Data and Hadoop
Flavio Vit
 
hadoop
swatic018
 
hadoop
swatic018
 
Hadoop ecosystem; J.Ayeesha parveen 2 nd M.sc., computer science Bon Secours...
AyeeshaParveen
 
Introduction to Hadoop Administration
Ramesh Pabba - seeking new projects
 
Introduction to Hadoop Administration
Ramesh Pabba - seeking new projects
 
Hadoop Seminar Report
Atul Kushwaha
 
Managing Big data with Hadoop
Nalini Mehta
 
Hadoop ecosystem J.AYEESHA PARVEEN II-M.SC.,COMPUTER SCIENCE, BON SECOURS CO...
AyeeshaParveen
 
HADOOP.pptx
Bharathi567510
 
Hadoop
RittikaBaksi
 
Lecture Slide - Introduction to Hadoop, HDFS, MapR.ppt
SuchithraaPalani
 
Hadoop.pptx
sonukumar379092
 
Ad

More from TIB Academy (17)

PDF
AWS Training Institute in Bangalore | Best AWS Course In Bangalore
TIB Academy
 
PDF
MySQL training in Bangalore | Best MySQL Course in Bangalore
TIB Academy
 
PDF
CCNA Training in Bangalore | Best Networking course in Bangalore
TIB Academy
 
PDF
Core Java Training in Bangalore | Best Core Java Class in Bangalore
TIB Academy
 
PDF
Advance Java Training in Bangalore | Best Java Training Institute
TIB Academy
 
PPTX
Best Hadoop Training in Bangalore - TIB Academy
TIB Academy
 
PPTX
Selenium training for beginners
TIB Academy
 
PPTX
Python Training
TIB Academy
 
PPTX
TIB Academy provides best Oracal DBA classes in Bangalore
TIB Academy
 
PPTX
Aws tutorial for beginners- tibacademy.in
TIB Academy
 
PPT
C C++ tutorial for beginners- tibacademy.in
TIB Academy
 
PPTX
Java tutorial for beginners-tibacademy.in
TIB Academy
 
PPT
Android tutorial for beginners-traininginbangalore.com
TIB Academy
 
PPT
SoapUI Training in Bangalore
TIB Academy
 
PPT
R programming
TIB Academy
 
PPT
Spring-training-in-bangalore
TIB Academy
 
PPTX
Salesforce Certification
TIB Academy
 
AWS Training Institute in Bangalore | Best AWS Course In Bangalore
TIB Academy
 
MySQL training in Bangalore | Best MySQL Course in Bangalore
TIB Academy
 
CCNA Training in Bangalore | Best Networking course in Bangalore
TIB Academy
 
Core Java Training in Bangalore | Best Core Java Class in Bangalore
TIB Academy
 
Advance Java Training in Bangalore | Best Java Training Institute
TIB Academy
 
Best Hadoop Training in Bangalore - TIB Academy
TIB Academy
 
Selenium training for beginners
TIB Academy
 
Python Training
TIB Academy
 
TIB Academy provides best Oracal DBA classes in Bangalore
TIB Academy
 
Aws tutorial for beginners- tibacademy.in
TIB Academy
 
C C++ tutorial for beginners- tibacademy.in
TIB Academy
 
Java tutorial for beginners-tibacademy.in
TIB Academy
 
Android tutorial for beginners-traininginbangalore.com
TIB Academy
 
SoapUI Training in Bangalore
TIB Academy
 
R programming
TIB Academy
 
Spring-training-in-bangalore
TIB Academy
 
Salesforce Certification
TIB Academy
 

Recently uploaded (20)

PDF
Geographical Diversity of India 100 Mcq.pdf/ 7th class new ncert /Social/Samy...
Sandeep Swamy
 
PDF
0725.WHITEPAPER-UNIQUEWAYSOFPROTOTYPINGANDUXNOW.pdf
Thomas GIRARD, MA, CDP
 
PPTX
HUMAN RESOURCE MANAGEMENT: RECRUITMENT, SELECTION, PLACEMENT, DEPLOYMENT, TRA...
PRADEEP ABOTHU
 
PDF
CONCURSO DE POESIA “POETUFAS – PASSOS SUAVES PELO VERSO.pdf
Colégio Santa Teresinha
 
PDF
The History of Phone Numbers in Stoke Newington by Billy Thomas
History of Stoke Newington
 
PDF
Chapter-V-DED-Entrepreneurship: Institutions Facilitating Entrepreneurship
Dayanand Huded
 
PPTX
How to Convert an Opportunity into a Quotation in Odoo 18 CRM
Celine George
 
PPTX
Cultivation practice of Litchi in Nepal.pptx
UmeshTimilsina1
 
PPTX
MENINGITIS: NURSING MANAGEMENT, BACTERIAL MENINGITIS, VIRAL MENINGITIS.pptx
PRADEEP ABOTHU
 
PPTX
Neurodivergent Friendly Schools - Slides from training session
Pooky Knightsmith
 
PPTX
How to Create Odoo JS Dialog_Popup in Odoo 18
Celine George
 
PPTX
GRADE-3-PPT-EVE-2025-ENG-Q1-LESSON-1.pptx
EveOdrapngimapNarido
 
PDF
DIGESTION OF CARBOHYDRATES,PROTEINS,LIPIDS
raviralanaresh2
 
PPTX
care of patient with elimination needs.pptx
Rekhanjali Gupta
 
PDF
The Different Types of Non-Experimental Research
Thelma Villaflores
 
PPTX
Universal immunization Programme (UIP).pptx
Vishal Chanalia
 
PPTX
Identifying elements in the story. Arrange the events in the story
geraldineamahido2
 
PPTX
PPT-Q1-WK-3-ENGLISH Revised Matatag Grade 3.pptx
reijhongidayawan02
 
PPTX
Growth and development and milestones, factors
BHUVANESHWARI BADIGER
 
PPTX
How to Handle Salesperson Commision in Odoo 18 Sales
Celine George
 
Geographical Diversity of India 100 Mcq.pdf/ 7th class new ncert /Social/Samy...
Sandeep Swamy
 
0725.WHITEPAPER-UNIQUEWAYSOFPROTOTYPINGANDUXNOW.pdf
Thomas GIRARD, MA, CDP
 
HUMAN RESOURCE MANAGEMENT: RECRUITMENT, SELECTION, PLACEMENT, DEPLOYMENT, TRA...
PRADEEP ABOTHU
 
CONCURSO DE POESIA “POETUFAS – PASSOS SUAVES PELO VERSO.pdf
Colégio Santa Teresinha
 
The History of Phone Numbers in Stoke Newington by Billy Thomas
History of Stoke Newington
 
Chapter-V-DED-Entrepreneurship: Institutions Facilitating Entrepreneurship
Dayanand Huded
 
How to Convert an Opportunity into a Quotation in Odoo 18 CRM
Celine George
 
Cultivation practice of Litchi in Nepal.pptx
UmeshTimilsina1
 
MENINGITIS: NURSING MANAGEMENT, BACTERIAL MENINGITIS, VIRAL MENINGITIS.pptx
PRADEEP ABOTHU
 
Neurodivergent Friendly Schools - Slides from training session
Pooky Knightsmith
 
How to Create Odoo JS Dialog_Popup in Odoo 18
Celine George
 
GRADE-3-PPT-EVE-2025-ENG-Q1-LESSON-1.pptx
EveOdrapngimapNarido
 
DIGESTION OF CARBOHYDRATES,PROTEINS,LIPIDS
raviralanaresh2
 
care of patient with elimination needs.pptx
Rekhanjali Gupta
 
The Different Types of Non-Experimental Research
Thelma Villaflores
 
Universal immunization Programme (UIP).pptx
Vishal Chanalia
 
Identifying elements in the story. Arrange the events in the story
geraldineamahido2
 
PPT-Q1-WK-3-ENGLISH Revised Matatag Grade 3.pptx
reijhongidayawan02
 
Growth and development and milestones, factors
BHUVANESHWARI BADIGER
 
How to Handle Salesperson Commision in Odoo 18 Sales
Celine George
 

Hadoop tutorial for beginners-tibacademy.in

  • 1. TIB Academy, 5/3, Varathur Road, Kundalahalli Gate, Bangalore-560066. +91-9513332301 / 02 www.tibacademy.in
  • 2.  Open source software framework designed for storage and processing of large scale data on clusters of commodity hardware  Created by Doug Cutting and Mike Carafella in 2005.  Cutting named the program after his son’s toy elephant.
  • 3.  Data-intensive text processing  Assembly of large genomes  Graph mining  Machine learning and data mining  Large scale social network analysis
  • 5. • Contains Libraries and other modules Hadoop Common • Hadoop Distributed File SystemHDFS • Yet Another Resource Negotiator Hadoop YARN • A programming model for large scale data processing Hadoop MapReduce
  • 7.  What were the limitations of earlier large- scale computing?  What requirements should an alternative approach have?  How does Hadoop address those requirements?
  • 8.  Historically computation was processor- bound › Data volume has been relatively small › Complicated computations are performed on that data  Advances in computer technology has historically centered around improving the power of a single machine
  • 10.  Moore’s Law › The number of transistors on a dense integrated circuit doubles every two years  Single-core computing can’t scale with current computing needs
  • 11.  Power consumption limits the speed increase we get from transistor density
  • 12.  Allows developers to use multiple machines for a single task
  • 13.  Programming on a distributed system is much more complex › Synchronizing data exchanges › Managing a finite bandwidth › Controlling computation timing is complicated
  • 14. “You know you have a distributed system when the crash of a computer you’ve never heard of stops you from getting any work done.” –Leslie Lamport  Distributed systems must be designed with the expectation of failure
  • 15.  Typically divided into Data Nodes and Compute Nodes  At compute time, data is copied to the Compute Nodes  Fine for relatively small amounts of data  Modern systems deal with far more data than was gathering in the past
  • 16.  Facebook › 500 TB per day  Yahoo › Over 170 PB  eBay › Over 6 PB  Getting the data to the processors becomes the bottleneck
  • 17.  Must support partial failure  Must be scalable
  • 18.  Failure of a single component must not cause the failure of the entire system only a degradation of the application performance  Failure should not result in the loss of any data
  • 19.  If a component fails, it should be able to recover without restarting the entire system  Component failure or recovery during a job must not affect the final output
  • 20.  Increasing resources should increase load capacity  Increasing the load on the system should result in a graceful decline in performance for all jobs › Not system failure
  • 21.  Based on work done by Google in the early 2000s › “The Google File System” in 2003 › “MapReduce: Simplified Data Processing on Large Clusters” in 2004  The core idea was to distribute the data as it is initially stored › Each node can then perform computation on the data it stores without moving the data for the initial processing
  • 22.  Applications are written in a high-level programming language › No network programming or temporal dependency  Nodes should communicate as little as possible › A “shared nothing” architecture  Data is spread among the machines in advance › Perform computation where the data is already stored as often as possible
  • 23.  When data is loaded onto the system it is divided into blocks › Typically 64MB or 128MB  Tasks are divided into two phases › Map tasks which are done on small portions of data where the data is stored › Reduce tasks which combine data to produce the final output  A master program allocates work to individual nodes
  • 24.  Failures are detected by the master program which reassigns the work to a different node  Restarting a task does not affect the nodes working on other portions of the data  If a failed node restarts, it is added back to the system and assigned new tasks  The master can redundantly execute the same task to avoid slow running nodes
  • 26.  Responsible for storing data on the cluster  Data files are split into blocks and distributed across the nodes in the cluster  Each block is replicated multiple times
  • 27.  HDFS is a file system written in Java based on the Google’s GFS  Provides redundant storage for massive amounts of data
  • 28.  HDFS works best with a smaller number of large files › Millions as opposed to billions of files › Typically 100MB or more per file  Files in HDFS are write once  Optimized for streaming reads of large files and not random reads
  • 29.  Files are split into blocks  Blocks are split across many machines at load time › Different blocks from the same file will be stored on different machines  Blocks are replicated across multiple machines  The NameNode keeps track of which blocks make up a file and where they are stored
  • 31.  When a client wants to retrieve data › Communicates with the NameNode to determine which blocks make up a file and on which data nodes those blocks are stored › Then communicated directly with the data nodes to read the data
  • 33.  A method for distributing computation across multiple nodes  Each node processes the data that is stored at that node  Consists of two main phases › Map › Reduce
  • 34.  Automatic parallelization and distribution  Fault-Tolerance  Provides a clean abstraction for programmers to use
  • 35.  Reads data as key/value pairs › The key is often discarded  Outputs zero or more key/value pairs
  • 36.  Output from the mapper is sorted by key  All values with the same key are guaranteed to go to the same machine
  • 37.  Called once for each unique key  Gets a list of all values associated with a key as input  The reducer outputs zero or more final key/value pairs › Usually just one output per input key
  • 40.  NameNode › Holds the metadata for the HDFS  Secondary NameNode › Performs housekeeping functions for the NameNode  DataNode › Stores the actual HDFS data blocks  JobTracker › Manages MapReduce jobs  TaskTracker › Monitors individual Map and Reduce tasks
  • 41.  Stores the HDFS file system information in a fsimage  Updates to the file system (add/remove blocks) do not change the fsimage file › They are instead written to a log file  When starting the NameNode loads the fsimage file and then applies the changes in the log file
  • 42.  NOT a backup for the NameNode  Periodically reads the log file and applies the changes to the fsimage file bringing it up to date  Allows the NameNode to restart faster when required
  • 43.  JobTracker › Determines the execution plan for the job › Assigns individual tasks  TaskTracker › Keeps track of the performance of an individual mapper or reducer
  • 45.  MapReduce is very powerful, but can be awkward to master  These tools allow programmers who are familiar with other programming styles to take advantage of the power of MapReduce
  • 46.  Hive › Hadoop processing with SQL  Pig › Hadoop processing with scripting  Cascading › Pipe and Filter processing model  HBase › Database model built on top of Hadoop  Flume › Designed for large scale data movement

Editor's Notes

  • #15: Example of failure issues. Linux lab is distributed file system, if the file server fails, what happens.
  • #24: Example of map and reduce
  • #30: Default replication is 3-fold