SlideShare a Scribd company logo
2
Most read
3
Most read
20
Most read
CLASSIFICATION USING
BACK-PROPAGATION
Backpropagation Algorithms
 The back-propagation learning algorithm is one of the most important
developments in neural networks.
 Backpropagation is the generalization of the Widrow-Hoff learning
rule to multiple-layer networks and nonlinear differentiable transfer
functions.
 This learning algorithm is applied to multilayer feed-forward networks
consisting of processing elements with continuous differentiable
activation functions.
 The networks associated with back-propagation algorithm are also
called back-propagation networks(BPNs).
Backpropagation Algorithms
 The Aim Of The Neural Network Is To Train The Net Ot
Achieve A Balance Between The Net’s Ability To
Respond(memorization) And Its Ability To Give
Resasonable Responses To The Input That Is Similar But
Not Identical To The One That Is Use In Trianing
(Generalization).
Architecture
This section presents the architecture of the network that is most
commonly used with the backpropagation algorithm –
the multilayer feedforward network
Architecture
Feedforward Network
Feedforward networks often have one or more hidden layers of sigmoid neurons followed
by an output layer of linear neurons.
Multiple layers of neurons with nonlinear transfer functions allow the network to learn
nonlinear and linear relationships between input and output vectors.
The linear output layer lets the network produce values outside the range -1 to +1. On the
other hand, if you want to constrain the outputs of a network (such as between 0 and 1),
then the output layer should use a sigmoid transfer function (such as logsig).
Learning Algorithm:
Backpropagation
The following slides describes teaching process of multi-layer neural network
employing backpropagation algorithm. To illustrate this process the three layer neural
network with two inputs and one output,which is shown in the picture below, is used:
Learning Algorithm:
Backpropagation
Each neuron is composed of two units. First unit adds products of weights coefficients and
input signals. The second unit realise nonlinear function, called neuron transfer (activation)
function. Signal e is adder output signal, and y = f(e) is output signal of nonlinear element.
Signal y is also output signal of neuron.
Learning Algorithm:
Backpropagation
To teach the neural network we need training data set. The training data set consists of input
signals (x1 and x2 ) assigned with corresponding target (desired output) z.
The network training is an iterative process. In each iteration weights coefficients of nodes
are modified using new data from training data set. Modification is calculated using
algorithm described below:
Each teaching step starts with forcing both input signals from training set. After this stage we
can determine output signals values for each neuron in each network layer.
Learning Algorithm:
Backpropagation
Pictures below illustrate how signal is propagating through the network,
Symbols w(xm)n represent weights of connections between network input xm and
neuron n in input layer. Symbols yn represents output signal of neuron n.
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Propagation of signals through the hidden layer. Symbols wmn represent weights
of connections between output of neuron m and input of neuron n in the next
layer.
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Learning Algorithm:
Backpropagation
Propagation of signals through the output layer.
Learning Algorithm:
Backpropagation
In the next algorithm step the output signal of the network y is compared
with the desired output value (the target), which is found in training data
set. The difference is called error signal d of output layer neuron
Learning Algorithm:
Backpropagation
The idea is to propagate error signal d (computed in single teaching step)
back to all neurons, which output signals were input for discussed
neuron.
Learning Algorithm:
Backpropagation
The idea is to propagate error signal d (computed in single teaching step)
back to all neurons, which output signals were input for discussed
neuron.
Learning Algorithm:
Backpropagation
The weights' coefficients wmn used to propagate errors back are equal to
this used during computing output value. Only the direction of data flow
is changed (signals are propagated from output to inputs one after the
other). This technique is used for all network layers. If propagated errors
came from few neurons they are added. The illustration is below:
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Learning Algorithm:
Backpropagation
When the error signal for each neuron is computed, the weights
coefficients of each neuron input node may be modified. In formulas
below df(e)/de represents derivative of neuron activation function (which
weights are modified).
Backpropagation
applications
 They have been successful on a wide array of real-world data,
including
 handwritten character recognition,
 pathology and laboratory medicine,
 and training a computer to pronounce English text.
Backpropagation
terminologies
 EPOCH
 EPOCHUPDATION
 TERMINATING CONDITION

More Related Content

What's hot (20)

PPTX
Introduction to Linear Discriminant Analysis
Jaclyn Kokx
 
ODP
Machine Learning with Decision trees
Knoldus Inc.
 
PPT
2.5 backpropagation
Krish_ver2
 
PPTX
Multilayer perceptron
omaraldabash
 
PDF
Decision trees in Machine Learning
Mohammad Junaid Khan
 
PPT
Back propagation
Nagarajan
 
PDF
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
Edureka!
 
PDF
Support Vector Machines ( SVM )
Mohammad Junaid Khan
 
PPTX
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
Simplilearn
 
PPTX
Machine learning clustering
CosmoAIMS Bassett
 
PPTX
Data mining Measuring similarity and desimilarity
Rushali Deshmukh
 
PPTX
Supervised learning and Unsupervised learning
Usama Fayyaz
 
PDF
Deep Feed Forward Neural Networks and Regularization
Yan Xu
 
PPTX
Feature Selection in Machine Learning
Upekha Vandebona
 
PPTX
Classification Algorithm.
Megha Sharma
 
PDF
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
PPT
Adaline madaline
Nagarajan
 
PDF
Classification Based Machine Learning Algorithms
Md. Main Uddin Rony
 
PDF
Naive Bayes
CloudxLab
 
Introduction to Linear Discriminant Analysis
Jaclyn Kokx
 
Machine Learning with Decision trees
Knoldus Inc.
 
2.5 backpropagation
Krish_ver2
 
Multilayer perceptron
omaraldabash
 
Decision trees in Machine Learning
Mohammad Junaid Khan
 
Back propagation
Nagarajan
 
Recurrent Neural Networks (RNN) | RNN LSTM | Deep Learning Tutorial | Tensorf...
Edureka!
 
Support Vector Machines ( SVM )
Mohammad Junaid Khan
 
Convolutional Neural Network - CNN | How CNN Works | Deep Learning Course | S...
Simplilearn
 
Machine learning clustering
CosmoAIMS Bassett
 
Data mining Measuring similarity and desimilarity
Rushali Deshmukh
 
Supervised learning and Unsupervised learning
Usama Fayyaz
 
Deep Feed Forward Neural Networks and Regularization
Yan Xu
 
Feature Selection in Machine Learning
Upekha Vandebona
 
Classification Algorithm.
Megha Sharma
 
Machine Learning: Introduction to Neural Networks
Francesco Collova'
 
Adaline madaline
Nagarajan
 
Classification Based Machine Learning Algorithms
Md. Main Uddin Rony
 
Naive Bayes
CloudxLab
 

Similar to Classification using back propagation algorithm (20)

PPTX
Unit ii supervised ii
Indira Priyadarsini
 
PPT
Lec 6-bp
Taymoor Nazmy
 
PPT
Back propagation
DrBaljitSinghKhehra
 
PPTX
Classification_by_back_&propagation.pptx
SadiaSaleem301
 
PPT
nural network ER. Abhishek k. upadhyay
abhishek upadhyay
 
PPT
Intro to Deep learning - Autoencoders
Akash Goel
 
DOCX
Backpropagation
ariffast
 
PDF
Classification by back propagation, multi layered feed forward neural network...
bihira aggrey
 
PPTX
Back Propagation-11-11-2qwasdddddd024.pptx
vinodkumarthatipamul
 
PPT
this is a Ai topic neural network ML_Lecture_4.ppt
ry54321288
 
PPTX
Training Neural Networks.pptx
ksghuge
 
PPTX
CS767_Lecture_05.pptx
ShujatHussainGadi
 
PPTX
back propagation1_presenation_lab 6.pptx
someyamohsen2
 
PDF
Mlp trainning algorithm
Hưng Đặng
 
PPTX
Artificial neural networks
ShwethaShreeS
 
PPTX
Feed forward back propogation algorithm .pptx
neelamsanjeevkumar
 
PPT
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
 
PPTX
Backpropagation.pptx
VandanaVipparthi
 
Unit ii supervised ii
Indira Priyadarsini
 
Lec 6-bp
Taymoor Nazmy
 
Back propagation
DrBaljitSinghKhehra
 
Classification_by_back_&propagation.pptx
SadiaSaleem301
 
nural network ER. Abhishek k. upadhyay
abhishek upadhyay
 
Intro to Deep learning - Autoencoders
Akash Goel
 
Backpropagation
ariffast
 
Classification by back propagation, multi layered feed forward neural network...
bihira aggrey
 
Back Propagation-11-11-2qwasdddddd024.pptx
vinodkumarthatipamul
 
this is a Ai topic neural network ML_Lecture_4.ppt
ry54321288
 
Training Neural Networks.pptx
ksghuge
 
CS767_Lecture_05.pptx
ShujatHussainGadi
 
back propagation1_presenation_lab 6.pptx
someyamohsen2
 
Mlp trainning algorithm
Hưng Đặng
 
Artificial neural networks
ShwethaShreeS
 
Feed forward back propogation algorithm .pptx
neelamsanjeevkumar
 
Back_propagation_algorithm.Back_propagation_algorithm.Back_propagation_algorithm
sureshkumarece1
 
Backpropagation.pptx
VandanaVipparthi
 
Ad

Recently uploaded (20)

PDF
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
PPTX
2025 CGI Congres - Surviving agile v05.pptx
Derk-Jan de Grood
 
PPTX
Mechanical Design of shell and tube heat exchangers as per ASME Sec VIII Divi...
shahveer210504
 
PDF
MAD Unit - 1 Introduction of Android IT Department
JappanMavani
 
PDF
REINFORCEMENT LEARNING IN DECISION MAKING SEMINAR REPORT
anushaashraf20
 
PPT
Testing and final inspection of a solar PV system
MuhammadSanni2
 
PDF
AN EMPIRICAL STUDY ON THE USAGE OF SOCIAL MEDIA IN GERMAN B2C-ONLINE STORES
ijait
 
PDF
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
PPTX
GitOps_Without_K8s_Training_detailed git repository
DanialHabibi2
 
PPTX
What is Shot Peening | Shot Peening is a Surface Treatment Process
Vibra Finish
 
PPTX
Knowledge Representation : Semantic Networks
Amity University, Patna
 
PPTX
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
PPTX
澳洲电子毕业证澳大利亚圣母大学水印成绩单UNDA学生证网上可查学历
Taqyea
 
PDF
Pressure Measurement training for engineers and Technicians
AIESOLUTIONS
 
PPTX
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
PDF
Basic_Concepts_in_Clinical_Biochemistry_2018كيمياء_عملي.pdf
AdelLoin
 
PDF
Halide Perovskites’ Multifunctional Properties: Coordination Engineering, Coo...
TaameBerhe2
 
PPTX
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
PDF
WD2(I)-RFQ-GW-1415_ Shifting and Filling of Sand in the Pond at the WD5 Area_...
ShahadathHossain23
 
PDF
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
Water Industry Process Automation & Control Monthly July 2025
Water Industry Process Automation & Control
 
2025 CGI Congres - Surviving agile v05.pptx
Derk-Jan de Grood
 
Mechanical Design of shell and tube heat exchangers as per ASME Sec VIII Divi...
shahveer210504
 
MAD Unit - 1 Introduction of Android IT Department
JappanMavani
 
REINFORCEMENT LEARNING IN DECISION MAKING SEMINAR REPORT
anushaashraf20
 
Testing and final inspection of a solar PV system
MuhammadSanni2
 
AN EMPIRICAL STUDY ON THE USAGE OF SOCIAL MEDIA IN GERMAN B2C-ONLINE STORES
ijait
 
Viol_Alessandro_Presentazione_prelaurea.pdf
dsecqyvhbowrzxshhf
 
GitOps_Without_K8s_Training_detailed git repository
DanialHabibi2
 
What is Shot Peening | Shot Peening is a Surface Treatment Process
Vibra Finish
 
Knowledge Representation : Semantic Networks
Amity University, Patna
 
OCS353 DATA SCIENCE FUNDAMENTALS- Unit 1 Introduction to Data Science
A R SIVANESH M.E., (Ph.D)
 
澳洲电子毕业证澳大利亚圣母大学水印成绩单UNDA学生证网上可查学历
Taqyea
 
Pressure Measurement training for engineers and Technicians
AIESOLUTIONS
 
fatigue in aircraft structures-221113192308-0ad6dc8c.pptx
aviatecofficial
 
Basic_Concepts_in_Clinical_Biochemistry_2018كيمياء_عملي.pdf
AdelLoin
 
Halide Perovskites’ Multifunctional Properties: Coordination Engineering, Coo...
TaameBerhe2
 
Final Major project a b c d e f g h i j k l m
bharathpsnab
 
WD2(I)-RFQ-GW-1415_ Shifting and Filling of Sand in the Pond at the WD5 Area_...
ShahadathHossain23
 
methodology-driven-mbse-murphy-july-hsv-huntsville6680038572db67488e78ff00003...
henriqueltorres1
 
Ad

Classification using back propagation algorithm

  • 2. Backpropagation Algorithms  The back-propagation learning algorithm is one of the most important developments in neural networks.  Backpropagation is the generalization of the Widrow-Hoff learning rule to multiple-layer networks and nonlinear differentiable transfer functions.  This learning algorithm is applied to multilayer feed-forward networks consisting of processing elements with continuous differentiable activation functions.  The networks associated with back-propagation algorithm are also called back-propagation networks(BPNs).
  • 3. Backpropagation Algorithms  The Aim Of The Neural Network Is To Train The Net Ot Achieve A Balance Between The Net’s Ability To Respond(memorization) And Its Ability To Give Resasonable Responses To The Input That Is Similar But Not Identical To The One That Is Use In Trianing (Generalization).
  • 4. Architecture This section presents the architecture of the network that is most commonly used with the backpropagation algorithm – the multilayer feedforward network
  • 5. Architecture Feedforward Network Feedforward networks often have one or more hidden layers of sigmoid neurons followed by an output layer of linear neurons. Multiple layers of neurons with nonlinear transfer functions allow the network to learn nonlinear and linear relationships between input and output vectors. The linear output layer lets the network produce values outside the range -1 to +1. On the other hand, if you want to constrain the outputs of a network (such as between 0 and 1), then the output layer should use a sigmoid transfer function (such as logsig).
  • 6. Learning Algorithm: Backpropagation The following slides describes teaching process of multi-layer neural network employing backpropagation algorithm. To illustrate this process the three layer neural network with two inputs and one output,which is shown in the picture below, is used:
  • 7. Learning Algorithm: Backpropagation Each neuron is composed of two units. First unit adds products of weights coefficients and input signals. The second unit realise nonlinear function, called neuron transfer (activation) function. Signal e is adder output signal, and y = f(e) is output signal of nonlinear element. Signal y is also output signal of neuron.
  • 8. Learning Algorithm: Backpropagation To teach the neural network we need training data set. The training data set consists of input signals (x1 and x2 ) assigned with corresponding target (desired output) z. The network training is an iterative process. In each iteration weights coefficients of nodes are modified using new data from training data set. Modification is calculated using algorithm described below: Each teaching step starts with forcing both input signals from training set. After this stage we can determine output signals values for each neuron in each network layer.
  • 9. Learning Algorithm: Backpropagation Pictures below illustrate how signal is propagating through the network, Symbols w(xm)n represent weights of connections between network input xm and neuron n in input layer. Symbols yn represents output signal of neuron n.
  • 12. Learning Algorithm: Backpropagation Propagation of signals through the hidden layer. Symbols wmn represent weights of connections between output of neuron m and input of neuron n in the next layer.
  • 15. Learning Algorithm: Backpropagation Propagation of signals through the output layer.
  • 16. Learning Algorithm: Backpropagation In the next algorithm step the output signal of the network y is compared with the desired output value (the target), which is found in training data set. The difference is called error signal d of output layer neuron
  • 17. Learning Algorithm: Backpropagation The idea is to propagate error signal d (computed in single teaching step) back to all neurons, which output signals were input for discussed neuron.
  • 18. Learning Algorithm: Backpropagation The idea is to propagate error signal d (computed in single teaching step) back to all neurons, which output signals were input for discussed neuron.
  • 19. Learning Algorithm: Backpropagation The weights' coefficients wmn used to propagate errors back are equal to this used during computing output value. Only the direction of data flow is changed (signals are propagated from output to inputs one after the other). This technique is used for all network layers. If propagated errors came from few neurons they are added. The illustration is below:
  • 20. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 21. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 22. Learning Algorithm: Backpropagation When the error signal for each neuron is computed, the weights coefficients of each neuron input node may be modified. In formulas below df(e)/de represents derivative of neuron activation function (which weights are modified).
  • 23. Backpropagation applications  They have been successful on a wide array of real-world data, including  handwritten character recognition,  pathology and laboratory medicine,  and training a computer to pronounce English text.