SlideShare a Scribd company logo
1 of 70
Back-Propagation Algorithm
Basic Neuron Model In A
Feedforward Network
ā€¢ Inputs xi
arrive through
pre-synaptic connections
ā€¢ Synaptic efficacy is
modeled using real
weights wi
ā€¢ The response of the
neuron is a nonlinear
function f of its weighted
inputs
Task
Plot the following type of Neural activation functions.
1(a) Threshold Function
Ļ†(v)= +1 for vā‰„0
0 for v<0
1(b) Threshold Function
Ļ†(v)= +1 for vā‰„0
-1 otherwise
2 Piecewise linear Function
Ļ†(v)= 1 for vā‰„+1/2
v for +1/2>v>-1/2
0 for vā‰¤-1/2
3(a) Sigmoid Function
Ļ†(v)=1/(1+ exp(-Ī»v))
3(b) Sigmoid Function
Ļ†(v)=2/(1+ exp(-Ī»v))
3(c) Sigmoid Function
Ļ†(v)=tanh(Ī»v)
For 3 vary the value of ā€˜Ī»ā€™ and show the changes in the graph.
Multiple Input Neuron
Single Layer Artificial Neural Networks
Layer of Neurons
Multilayer Network
Banana & Apple Sorter
Prototype Vectors
Banana Apple Problem
Illustration of a Neural
Network
Different networks
ā˜»Perceptron
ā€“ Feedforward Network, Linear Decision Boundary, One Neuron for
Each Decision
ā˜»Hamming Network
ā˜»Hopfield Network
- Dynamic Associative Memory Network
ā˜»Error Back Propagation network
ā˜»Radial basis network
ā˜»ART
ā˜»Brain in a box neural network
ā˜»Cellular neural Network
ā˜»Neocognitron
ā˜»Functional
Network Topology
Feedforward
Inputs
Outputs
Inputs
Feedback
Outputs
1970s
The Backpropagation algorithm was first proposed by
Paul Werbos in the 1970's. However, it was
rediscoved in 1986 by Rumelhart and McClelland &
became widely used.
It took 30 years before the error backpropagation (or
in short: backprop) algorithm popularized.
Differences In Networks
Feedforward Networks
ā€¢ Solutions are known
ā€¢ Weights are learned
ā€¢ Evolves in the weight
space
ā€¢ Used for:
ā€“ Prediction
ā€“ Classification
ā€“ Function
approximation
Feedback Networks
ā€¢ Solutions are
unknown
ā€¢ Weights are
prescribed
ā€¢ Evolves in the state
space
ā€¢ Used for:
ā€“ Constraint satisfaction
ā€“ Optimization
ā€“ Feature matching
Architecture
A Back Prop network has atleast 3 layers of units:
an input layer, at least one intermediate hidden layer, &
an output layer. Connection weights in a Back Prop
network are one way. Units are connected in a feed-
forward fashion with input units fully connected to units
in the hidden layer & hidden units fully connected to units
in the output layer. When a Back Prop network is cycled,
an input pattern is propagated forward to the output units
through the intervening input-to-hidden and hidden-to-
output weights.
Inputs To Neurons
ā€¢ Arise from other neurons or from outside
the network
ā€¢ Nodes whose inputs arise outside the
network are called input nodes and simply
copy values
ā€¢ An input may excite or inhibit the response
of the neuron to which it is applied,
depending upon the weight of the
connection
Fully connected network
Weights
ā€¢ Represent synaptic efficacy and may be
excitatory or inhibitory
ā€¢ Normally, positive weights are considered
as excitatory while negative weights are
thought of as inhibitory
ā€¢ Learning is the process of modifying the
weights in order to produce a network that
performs some function
Finding net
Output
ā€¢ The response function is normally nonlinear
ā€¢ Samples include
ā€“ Sigmoid
ā€“ Piecewise linear
x
e
xf Ī»āˆ’
+
=
1
1
)(
ļ£³
ļ£²
ļ£±
<
ā‰„
=
Īø
Īø
xif
xifx
xf
,0
,
)(
Back propagation Networks
I1
I2
1
Hidden Layer
H1
H2
O1
O2
Output Layer
Wi,j
Wj,k
1ā€™s - bias
āˆ‘
+
= āˆ’
j
jxj Hw
e
xO
,
1
1
)(
I3
1
āˆ‘
+
= āˆ’
i
ixi Iw
e
xH
,
1
1
)(
Weight updation
Backpropagation Preparation
ā€¢ Training Set
A collection of input-output patterns that are
used to train the network
ā€¢ Testing Set
A collection of input-output patterns that are
used to assess network performance
ā€¢ Learning Rate-Ī·
A scalar parameter, analogous to step size in
numerical integration, used to set the rate of
adjustments
Learning
ā€¢ Learning occurs during a training phase in which each input
pattern in a training set is applied to the input units and then
propagated forward.
ā€¢ The pattern of activation arriving at the output layer is then
compared with the correct output pattern to calculate an
error signal.
ā€¢ The error signal for each such target output pattern is then
back propagated from the outputs to the inputs in order to
appropriately adjust the weights in each layer of the network.
Learning
ā€¢ The process goes on for several cycles till the error
reduces to a predefined limit.
ā€¢ After a BackProp network has learned the correct
classification for a set of inputs, it can be tested on a
second set of inputs to see how well it classifies
untrained patterns.
ā€¢ Thus, an important consideration in applying
BackProp learning is how well the network
generalizes.
The basic principles of the back propagation algorithm are:
(1) the error of the output signal of a neuron is used to
adjust its weights such that the error decreases, and (2)
the error in hidden layers is estimated proportional to the
weighted sum of the (estimated) errors in the layer
above.
Patterns
ļ®Training patterns (70%)
ļ®Testing patterns (30%)
During the training, the data is presented to the network
several thousand times. For each data sample, the
current output of the network is calculated and compared
to the "true" target value. The error signal dj of neuron j
is computed from the difference between the target and
the calculated output. For hidden neurons, this difference
is estimated by the weighted error signals of the layer
above. The error terms are then used to adjust the
weights wij of the neural network.
A Pseudo-Code Algorithm
ā€¢ Randomly choose the initial weights
ā€¢ While error is too large
ā€“ For each training pattern (presented in random order)
ā€¢ Apply the inputs to the network
ā€¢ Calculate the output for every neuron from the input layer,
through the hidden layer(s), to the output layer
ā€¢ Calculate the error at the outputs
ā€¢ Use the output error to compute error signals for pre-output
layers
ā€¢ Use the error signals to compute weight adjustments
ā€¢ Apply the weight adjustments
ā€“ Periodically evaluate the network performance
Network Error
ā€¢ Total-Sum-Squared-Error (TSSE)
ā€¢ Root-Mean-Squared-Error (RMSE)
āˆ‘ āˆ‘ āˆ’=
patterns outputs
actualdesiredTSSE 2
)(
2
1
outputspatterns
TSSE
RMSE
*##
*2
=
Apply Inputs From A Pattern
ā€¢ Apply the value of
each input parameter
to each input node
ā€¢ Input nodes computer
only the identity
function
Feedforward
Inputs
Outputs
Calculate Outputs For Each
Neuron Based On The Pattern
ā€¢ The output from neuron j
for pattern p is Opj where
and
k ranges over the input
indices and Wjk is the
weight on the connection
from input k to neuron j
Feedforward
Inputs
Outputs
jnetjpj
e
netO Ī»āˆ’
+
=
1
1
)(
āˆ‘+=
k
jkpkbiasj WOWbiasnet *
Calculate The Error Signal For
Each Output Neuron
ā€¢ The output neuron error signal Ī“pj is given
by Ī“pj=(Tpj-Opj) Opj (1-Opj)
ā€¢ Tpj is the target value of output neuron j for
pattern p
ā€¢ Opj is the actual output value of output
neuron j for pattern p
Calculate The Error Signal For
Each Hidden Neuron
ā€¢ The hidden neuron error signal Ī“pj is given
by
where Ī“pk is the error signal of a post-
synaptic neuron k and Wkj is the weight of
the connection from hidden neuron j to the
post-synaptic neuron k
kj
k
pkpjpjpj WOO āˆ‘āˆ’= Ī“Ī“ )1(
Calculate And Apply Weight
Adjustments
ā€¢ Compute weight adjustments āˆ†Wji at time
t by
āˆ†Wji(t)= Ī· Ī“pj Opi
ā€¢ Apply weight adjustments according to
Wji(t+1) = Wji(t) + āˆ†Wji(t)
ā€¢ Some add a momentum term Ī±āˆ—āˆ†Wji(t-1)
ā€¢ Thus, the network adjusts its weights after each data
sample. This learning process is in fact a gradient
descent in the error surface of the weight space - with all
its drawbacks. The learning algorithm is slow and prone
to getting stuck in a local minimum.
Simulation Issues
ļ‚§ How to Select Initial Weights
ļ‚§ Local Minima
ļ‚§ Solutions to Local minima
ļ‚§ Rate of Learning
ļ‚§ Stopping Criterion
ļ‚§ Initialization
ā€¢ For the standard back propagation algorithm, the initial
weights of the multi-layer perceptron have to be
relatively small. They can, for instance, be selected
randomly from a small interval around zero. During
training they are slowly adapted. Starting with small
weights is crucial, because large weights are rigid and
cannot be changed quickly.
Sequential & Batch modes
For a given training set ,back-propagation learning
proceeds in two basic ways:
1. Sequential Mode
2. Batch Mode
Sequential mode
ā€¢ The sequential mode of back-propagation learning is also
referred to as on-line, pattern or stochastic mode.
ā€¢ To be specific, consider an epoch consisting of N training ex.
Arranged in the order (x(1),d(1)),ā€¦,(x(N),d(N)).
ā€¢ The first ex. pair (x(1),d(1))in the epoch is presented to the
network,& the sequence of forward & backward computations
described previously is performed, resulting in certain adjustments
to the synaptic weights & bias level of the network.
ā€¢ The second ex. (x(N),d(N)) in the epoch is presented,& the
sequence of forward & backward computations is repeated,
resulting in the further adjustments to the synaptic weights & bias
levels. This process is continued until the last example pair
(x(N),d(N)) in the epoch is accounted for.
Batch Propagation
ā€¢ In this mode of back-propagation learning weight
updating is performed after the presentation of all the
training examples that constitute an epoch.
ā€¢ For a particular epoch, the cost function is the average
squared error, reproduced here in composite form is
defined as:-
Ī¾av = (1/2N )Ī£ Ī£ ej
2
(n) for n=1 to N
for j ā‚¬ C
ā€¢ Let N denote the total no. of patterns contained in the
training set. The average squared error energy is
obtained by summing Ī¾(n) over all n and then
normalizing with respect to the set size N, as shown by :-
ā€¢ Ī¾av = 1/N Ī£ Ī¾(n) for n=1 to N
Stopping Criteria
ā€¢ The back-propagation algorithm cannot be shown to converge .
ā€¢ To formulate a criterion, it is logically to think in terms of the
unique properties of a local or global minimum.
ā€¢ The back-propagation algorithm is considered to have
converged when the Euclidean norm of the gradient vector reaches
a sufficient small gradient threshold.
ā€¢ The back-propagation algorithm is considered to have converged
when the absolute rate of change in the average squared error pre
epoch is sufficiently small.
ā€¢ The drawback of this convergence criterion is that, for
successful trials, learning time may be long.
ā€¢ The back-propagation algorithm makes adjustments by
computing the derivative, or slope of the network error
with respect to each neuronā€™s output. It attempts to
minimize the overall error by descending this slope to the
minimum value for every weight. It advances one step
down the slope each epoch. If the network takes steps
that are too large, it may pass the global minimum. If it
takes steps that are small, it may settle on local minima,
or take an inordinate amount of time to arrive at the
global minimum. The ideal step size for a given problem
requires detailed, high-order derivative analysis, a task
not performed by the algorithm.
Minima
ā€¢ Local minima
ā€¢ Global minima
Local Minima
ļ®For simple 2 layer networks (without a hidden layer), the
error surface is bowl shaped and using gradient-descent to
minimize error is not a problem; the network will always
find an errorless solution (at the bottom of the bowl). Such
errorless solutions are called global minima.
ļ®However, extra hidden layer implies complex surfaces.
Since some minima are deeper than others, it is possible
that gradient descent may not find a global minima.
Instead, the network may fall into local minima which
represent suboptimal solutions.
ā€¢ The algorithm cycles through the training samples as:-
ā€¢ Initialization
ā€¢ Presentation of training Examples
ā€¢ Forward Computation
Initialization
ā€¢ Assuming that no prior information is available, pick the
synaptic weights and thresholds from a uniform
distribution whose mean is zero & whose variance is
chosen to make the standard deviation of the induced
local fields of the neurons lie at the transition between
the linear and saturated parts of the sigmoid activation
function.
Presentation of training Examples
Present the network with an epoch of training examples.
For each example in the set order in same fashion,
perform the sequence of forward and backward
computation as described below.
Solutions to Local minima
ļ®Usual solution : More hidden layers. Logic -
Although additional hidden units increase the
complexity of the error surface, the extra
dimensionalilty increases the number of possible
escape routes.
ļ®Our solution ā€“ Tunneling
Rate of Learning
ļ®If the learning rate Ī· is very small, then the
algorithm proceeds slowly, but accurately follows
the path of steepest descent in weight space.
ļ®If Ī· is large, the algorithm may oscillate.
ļ®A simple method of effectively increasing the rate of
learning is to modify the delta rule by including a
momentum term:
Ī”wji
(n) = Ī± Ī”wji
(n-1) + Ī· Ī“j
(n)yi
(n)
where Ī± is a positive constant termed the momentum
constant. This is called the generalized delta rule.
ļ‚§ The effect is that if the basic delta rule is consistently
pushing a weight in the same direction, then it gradually
gathers "momentum" in that direction.
Forward Computation
An Example: Exclusive ā€œORā€
ā€¢ Training set
ā€“ ((0.1, 0.1), 0.1)
ā€“ ((0.1, 0.9), 0.9)
ā€“ ((0.9, 0.1), 0.9)
ā€“ ((0.9, 0.9), 0.1)
ā€¢ Testing set
ā€“ Use at least 121 pairs equally spaced on the
unit square and plot the results
ā€“ Omit the training set (if desired)
An Example (continued):
Network Architectureinputs
output(s)
An Example (continued):
Network Architecture
Sample
input
0.1
0.9
Target
output
0.9
1
1
1
Feedforward Network Training by
Backpropagation: Process
Summary
ā€¢ Select an architecture
ā€¢ Randomly initialize weights
ā€¢ While error is too large
ā€“ Select training pattern and feedforward to find
actual network output
ā€“ Calculate errors and backpropagate error
signals
ā€“ Adjust weights
ā€¢ Evaluate performance using the test set
An Example (continued):
Network Architecture
Sample
input
0.1
0.9
Actual
output
???
1
1
1
??
??
??
??
??
??
??
??
??
Target
output
0.9
Feedforward Network Training by
Backpropagation: Process
Summary
ā€¢ Select an architecture
ā€¢ Randomly initialize weights
ā€¢ While error is too large
ā€“ Select training pattern and feedforward to find
actual network output
ā€“ Calculate errors and backpropagate error
signals
ā€“ Adjust weights
ā€¢ Evaluate performance using the test set
Backpropagation
ā€¢Very powerful - can learn any function, given enough
hidden units! With enough hidden units, we can
generate any function.
ā€¢Have the same problems of Generalization vs.
Memorization. With too many units, we will tend to
memorize the input and not generalize well. Some
schemes exist to ā€œpruneā€ the neural network.
BackProp networks are not limited in its use because
they can adapt their weights to acquire new
knowledge. BackProp networks learn by example,
and can be used to make predictions.
Write a program to train and simulate neural
network for following network
ā€“ Input Nodes = 2 &
Output Nodes = 1
ā€“ Input Nodes = 3 and
Output nodes = 1
Inputs Outputs
A B Y
0 0 0
0 1 1
1 0 1
1 1 0
Inputs Outputs
A B C Y
0 0 0 0
0 0 1 0
0 1 0 0
0 1 1 0
1 0 0 1
1 0 1 1
1 1 0 1
1 1 1 1
ā€¢ Artificial Neural Network
ā€“ Simon Haykin
ā€¢ Artificial Neural Network
ā€“ Jacek Zurada

More Related Content

What's hot

Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksFrancesco Collova'
Ā 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward NetworksTamer Ahmed Farrag, PhD
Ā 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural NetworksNAGUR SHAREEF SHAIK
Ā 
Support vector machine
Support vector machineSupport vector machine
Support vector machineMusa Hawamdah
Ā 
Deep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsDeep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsBuhwan Jeong
Ā 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural networkNagarajan
Ā 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural NetworkAtul Krishna
Ā 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)EdutechLearners
Ā 
Classification using back propagation algorithm
Classification using back propagation algorithmClassification using back propagation algorithm
Classification using back propagation algorithmKIRAN R
Ā 
Deep belief network.pptx
Deep belief network.pptxDeep belief network.pptx
Deep belief network.pptxSushilAcharya18
Ā 
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...Mohammed Bennamoun
Ā 
Neural Networks
Neural NetworksNeural Networks
Neural NetworksNikitaRuhela
Ā 
Perceptron and Sigmoid Neurons
Perceptron and Sigmoid NeuronsPerceptron and Sigmoid Neurons
Perceptron and Sigmoid NeuronsShajun Nisha
Ā 
Principles of soft computing-Associative memory networks
Principles of soft computing-Associative memory networksPrinciples of soft computing-Associative memory networks
Principles of soft computing-Associative memory networksSivagowry Shathesh
Ā 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networkmustafa aadel
Ā 

What's hot (20)

Neural network
Neural networkNeural network
Neural network
Ā 
Machine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural NetworksMachine Learning: Introduction to Neural Networks
Machine Learning: Introduction to Neural Networks
Ā 
04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks04 Multi-layer Feedforward Networks
04 Multi-layer Feedforward Networks
Ā 
Perceptron & Neural Networks
Perceptron & Neural NetworksPerceptron & Neural Networks
Perceptron & Neural Networks
Ā 
Support vector machine
Support vector machineSupport vector machine
Support vector machine
Ā 
Deep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsDeep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applications
Ā 
Gradient descent method
Gradient descent methodGradient descent method
Gradient descent method
Ā 
Introduction Of Artificial neural network
Introduction Of Artificial neural networkIntroduction Of Artificial neural network
Introduction Of Artificial neural network
Ā 
Artificial Neural Network
Artificial Neural NetworkArtificial Neural Network
Artificial Neural Network
Ā 
Perceptron (neural network)
Perceptron (neural network)Perceptron (neural network)
Perceptron (neural network)
Ā 
Classification using back propagation algorithm
Classification using back propagation algorithmClassification using back propagation algorithm
Classification using back propagation algorithm
Ā 
Deep belief network.pptx
Deep belief network.pptxDeep belief network.pptx
Deep belief network.pptx
Ā 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
Ā 
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...
Artificial Neural Network Lecture 6- Associative Memories & Discrete Hopfield...
Ā 
Lstm
LstmLstm
Lstm
Ā 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
Ā 
Perceptron and Sigmoid Neurons
Perceptron and Sigmoid NeuronsPerceptron and Sigmoid Neurons
Perceptron and Sigmoid Neurons
Ā 
CNN Tutorial
CNN TutorialCNN Tutorial
CNN Tutorial
Ā 
Principles of soft computing-Associative memory networks
Principles of soft computing-Associative memory networksPrinciples of soft computing-Associative memory networks
Principles of soft computing-Associative memory networks
Ā 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
Ā 

Viewers also liked

Back propagation
Back propagationBack propagation
Back propagationNagarajan
Ā 
The Back Propagation Learning Algorithm
The Back Propagation Learning AlgorithmThe Back Propagation Learning Algorithm
The Back Propagation Learning AlgorithmESCOM
Ā 
Perceptron
PerceptronPerceptron
PerceptronNagarajan
Ā 
Kohonen self organizing maps
Kohonen self organizing mapsKohonen self organizing maps
Kohonen self organizing mapsraphaelkiminya
Ā 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation NetworkAkshay Dhole
Ā 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQESCOM
Ā 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQESCOM
Ā 
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IINeural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IIEdgar Carrillo
Ā 
Neural Networks: Self-Organizing Maps (SOM)
Neural Networks:  Self-Organizing Maps (SOM)Neural Networks:  Self-Organizing Maps (SOM)
Neural Networks: Self-Organizing Maps (SOM)Mostafa G. M. Mostafa
Ā 
Back propagation network
Back propagation networkBack propagation network
Back propagation networkHIRA Zaidi
Ā 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORKESCOM
Ā 
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin RSelf-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin Rshanelynn
Ā 
Adaline madaline
Adaline madalineAdaline madaline
Adaline madalineNagarajan
Ā 
How to calculate back propagation
How to calculate back propagationHow to calculate back propagation
How to calculate back propagationShinagawa Seitaro
Ā 
Machine Learning for Data Mining
Machine Learning for Data MiningMachine Learning for Data Mining
Machine Learning for Data MiningBhuban Roy
Ā 
Pattern Recognition - Designing a minimum distance class mean classifier
Pattern Recognition - Designing a minimum distance class mean classifierPattern Recognition - Designing a minimum distance class mean classifier
Pattern Recognition - Designing a minimum distance class mean classifierNayem Nayem
Ā 

Viewers also liked (20)

Back propagation
Back propagationBack propagation
Back propagation
Ā 
The Back Propagation Learning Algorithm
The Back Propagation Learning AlgorithmThe Back Propagation Learning Algorithm
The Back Propagation Learning Algorithm
Ā 
Hopfield Networks
Hopfield NetworksHopfield Networks
Hopfield Networks
Ā 
Perceptron
PerceptronPerceptron
Perceptron
Ā 
Kohonen self organizing maps
Kohonen self organizing mapsKohonen self organizing maps
Kohonen self organizing maps
Ā 
Counter propagation Network
Counter propagation NetworkCounter propagation Network
Counter propagation Network
Ā 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQ
Ā 
Learning Vector Quantization LVQ
Learning Vector Quantization LVQLearning Vector Quantization LVQ
Learning Vector Quantization LVQ
Ā 
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo IINeural networks Self Organizing Map by Engr. Edgar Carrillo II
Neural networks Self Organizing Map by Engr. Edgar Carrillo II
Ā 
Neural Networks: Self-Organizing Maps (SOM)
Neural Networks:  Self-Organizing Maps (SOM)Neural Networks:  Self-Organizing Maps (SOM)
Neural Networks: Self-Organizing Maps (SOM)
Ā 
Backpropagation algo
Backpropagation  algoBackpropagation  algo
Backpropagation algo
Ā 
Back propagation network
Back propagation networkBack propagation network
Back propagation network
Ā 
Counterpropagation NETWORK
Counterpropagation NETWORKCounterpropagation NETWORK
Counterpropagation NETWORK
Ā 
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin RSelf-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Self-Organising Maps for Customer Segmentation using R - Shane Lynn - Dublin R
Ā 
Adaline madaline
Adaline madalineAdaline madaline
Adaline madaline
Ā 
Self-organizing map
Self-organizing mapSelf-organizing map
Self-organizing map
Ā 
How to calculate back propagation
How to calculate back propagationHow to calculate back propagation
How to calculate back propagation
Ā 
Machine Learning for Data Mining
Machine Learning for Data MiningMachine Learning for Data Mining
Machine Learning for Data Mining
Ā 
Pattern Recognition - Designing a minimum distance class mean classifier
Pattern Recognition - Designing a minimum distance class mean classifierPattern Recognition - Designing a minimum distance class mean classifier
Pattern Recognition - Designing a minimum distance class mean classifier
Ā 
Seminar Presentation
Seminar PresentationSeminar Presentation
Seminar Presentation
Ā 

Similar to nural network ER. Abhishek k. upadhyay

Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...bihira aggrey
Ā 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networksarjitkantgupta
Ā 
Backpropagation.pptx
Backpropagation.pptxBackpropagation.pptx
Backpropagation.pptxVandanaVipparthi
Ā 
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryHands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryAhmed Yousry
Ā 
Unit 2 ml.pptx
Unit 2 ml.pptxUnit 2 ml.pptx
Unit 2 ml.pptxPradeeshSAI
Ā 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural networkIshaneeSharma
Ā 
Reinforcement Learning and Artificial Neural Nets
Reinforcement Learning and Artificial Neural NetsReinforcement Learning and Artificial Neural Nets
Reinforcement Learning and Artificial Neural NetsPierre de Lacaze
Ā 
Neural network basic and introduction of Deep learning
Neural network basic and introduction of Deep learningNeural network basic and introduction of Deep learning
Neural network basic and introduction of Deep learningTapas Majumdar
Ā 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNNaren Chandra Kattla
Ā 
08 neural networks
08 neural networks08 neural networks
08 neural networksankit_ppt
Ā 
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Simplilearn
Ā 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyayabhishek upadhyay
Ā 
Introduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptxIntroduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptxPoonam60376
Ā 

Similar to nural network ER. Abhishek k. upadhyay (20)

Lec 6-bp
Lec 6-bpLec 6-bp
Lec 6-bp
Ā 
Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...Classification by back propagation, multi layered feed forward neural network...
Classification by back propagation, multi layered feed forward neural network...
Ā 
Artificial neural networks
Artificial neural networksArtificial neural networks
Artificial neural networks
Ā 
Backpropagation.pptx
Backpropagation.pptxBackpropagation.pptx
Backpropagation.pptx
Ā 
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousryHands on machine learning with scikit-learn and tensor flow by ahmed yousry
Hands on machine learning with scikit-learn and tensor flow by ahmed yousry
Ā 
Unit 2 ml.pptx
Unit 2 ml.pptxUnit 2 ml.pptx
Unit 2 ml.pptx
Ā 
Unit iii update
Unit iii updateUnit iii update
Unit iii update
Ā 
Artificial neural network
Artificial neural networkArtificial neural network
Artificial neural network
Ā 
Reinforcement Learning and Artificial Neural Nets
Reinforcement Learning and Artificial Neural NetsReinforcement Learning and Artificial Neural Nets
Reinforcement Learning and Artificial Neural Nets
Ā 
6
66
6
Ā 
Neural network basic and introduction of Deep learning
Neural network basic and introduction of Deep learningNeural network basic and introduction of Deep learning
Neural network basic and introduction of Deep learning
Ā 
Electricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANNElectricity Demand Forecasting Using ANN
Electricity Demand Forecasting Using ANN
Ā 
08 neural networks
08 neural networks08 neural networks
08 neural networks
Ā 
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Deep Learning Interview Questions And Answers | AI & Deep Learning Interview ...
Ā 
nural network ER. Abhishek k. upadhyay
nural network ER. Abhishek  k. upadhyaynural network ER. Abhishek  k. upadhyay
nural network ER. Abhishek k. upadhyay
Ā 
Neural Networks
Neural NetworksNeural Networks
Neural Networks
Ā 
Unit ii supervised ii
Unit ii supervised iiUnit ii supervised ii
Unit ii supervised ii
Ā 
CS767_Lecture_05.pptx
CS767_Lecture_05.pptxCS767_Lecture_05.pptx
CS767_Lecture_05.pptx
Ā 
Introduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptxIntroduction to Perceptron and Neural Network.pptx
Introduction to Perceptron and Neural Network.pptx
Ā 
Neural networks
Neural networksNeural networks
Neural networks
Ā 

More from abhishek upadhyay

Nural network ER. Abhishek k. upadhyay Learning rules
Nural network ER. Abhishek  k. upadhyay Learning rulesNural network ER. Abhishek  k. upadhyay Learning rules
Nural network ER. Abhishek k. upadhyay Learning rulesabhishek upadhyay
Ā 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyayabhishek upadhyay
Ā 
Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyayabhishek upadhyay
Ā 
bi copter Major project report ER.Abhishek upadhyay b.tech (ECE)
bi copter  Major project report ER.Abhishek upadhyay b.tech (ECE)bi copter  Major project report ER.Abhishek upadhyay b.tech (ECE)
bi copter Major project report ER.Abhishek upadhyay b.tech (ECE)abhishek upadhyay
Ā 
(1) nanowire battery gerling (4)
(1) nanowire battery gerling (4)(1) nanowire battery gerling (4)
(1) nanowire battery gerling (4)abhishek upadhyay
Ā 
moving message display of lcd
 moving message display of lcd moving message display of lcd
moving message display of lcdabhishek upadhyay
Ā 

More from abhishek upadhyay (13)

Nural network ER. Abhishek k. upadhyay Learning rules
Nural network ER. Abhishek  k. upadhyay Learning rulesNural network ER. Abhishek  k. upadhyay Learning rules
Nural network ER. Abhishek k. upadhyay Learning rules
Ā 
Nural network ER. Abhishek k. upadhyay
Nural network ER. Abhishek  k. upadhyayNural network ER. Abhishek  k. upadhyay
Nural network ER. Abhishek k. upadhyay
Ā 
Nural network ER.Abhishek k. upadhyay
Nural network  ER.Abhishek k. upadhyayNural network  ER.Abhishek k. upadhyay
Nural network ER.Abhishek k. upadhyay
Ā 
bi copter Major project report ER.Abhishek upadhyay b.tech (ECE)
bi copter  Major project report ER.Abhishek upadhyay b.tech (ECE)bi copter  Major project report ER.Abhishek upadhyay b.tech (ECE)
bi copter Major project report ER.Abhishek upadhyay b.tech (ECE)
Ā 
A project report on
A project report onA project report on
A project report on
Ā 
Oc ppt
Oc pptOc ppt
Oc ppt
Ā 
lcd
lcdlcd
lcd
Ā 
abhishek
abhishekabhishek
abhishek
Ā 
mmu
mmummu
mmu
Ā 
(1) nanowire battery gerling (4)
(1) nanowire battery gerling (4)(1) nanowire battery gerling (4)
(1) nanowire battery gerling (4)
Ā 
moving message display of lcd
 moving message display of lcd moving message display of lcd
moving message display of lcd
Ā 
Bluetooth
BluetoothBluetooth
Bluetooth
Ā 
Khetarpal
KhetarpalKhetarpal
Khetarpal
Ā 

Recently uploaded

Indian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptIndian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptMadan Karki
Ā 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx959SahilShah
Ā 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvLewisJB
Ā 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...asadnawaz62
Ā 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionMebane Rash
Ā 
Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...121011101441
Ā 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionDr.Costas Sachpazis
Ā 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm Systemirfanmechengr
Ā 
Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleAlluxio, Inc.
Ā 
Introduction-To-Agricultural-Surveillance-Rover.pptx
Introduction-To-Agricultural-Surveillance-Rover.pptxIntroduction-To-Agricultural-Surveillance-Rover.pptx
Introduction-To-Agricultural-Surveillance-Rover.pptxk795866
Ā 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024Mark Billinghurst
Ā 
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfCCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfAsst.prof M.Gokilavani
Ā 
Earthing details of Electrical Substation
Earthing details of Electrical SubstationEarthing details of Electrical Substation
Earthing details of Electrical Substationstephanwindworld
Ā 
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfg
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfgUnit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfg
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfgsaravananr517913
Ā 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfme23b1001
Ā 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AIabhishek36461
Ā 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...Chandu841456
Ā 

Recently uploaded (20)

Indian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.pptIndian Dairy Industry Present Status and.ppt
Indian Dairy Industry Present Status and.ppt
Ā 
Application of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptxApplication of Residue Theorem to evaluate real integrations.pptx
Application of Residue Theorem to evaluate real integrations.pptx
Ā 
Work Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvvWork Experience-Dalton Park.pptxfvvvvvvv
Work Experience-Dalton Park.pptxfvvvvvvv
Ā 
complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...complete construction, environmental and economics information of biomass com...
complete construction, environmental and economics information of biomass com...
Ā 
US Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of ActionUS Department of Education FAFSA Week of Action
US Department of Education FAFSA Week of Action
Ā 
Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...Instrumentation, measurement and control of bio process parameters ( Temperat...
Instrumentation, measurement and control of bio process parameters ( Temperat...
Ā 
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective IntroductionSachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Sachpazis Costas: Geotechnical Engineering: A student's Perspective Introduction
Ā 
Class 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm SystemClass 1 | NFPA 72 | Overview Fire Alarm System
Class 1 | NFPA 72 | Overview Fire Alarm System
Ā 
Correctly Loading Incremental Data at Scale
Correctly Loading Incremental Data at ScaleCorrectly Loading Incremental Data at Scale
Correctly Loading Incremental Data at Scale
Ā 
young call girls in Green ParkšŸ” 9953056974 šŸ” escort Service
young call girls in Green ParkšŸ” 9953056974 šŸ” escort Serviceyoung call girls in Green ParkšŸ” 9953056974 šŸ” escort Service
young call girls in Green ParkšŸ” 9953056974 šŸ” escort Service
Ā 
Introduction-To-Agricultural-Surveillance-Rover.pptx
Introduction-To-Agricultural-Surveillance-Rover.pptxIntroduction-To-Agricultural-Surveillance-Rover.pptx
Introduction-To-Agricultural-Surveillance-Rover.pptx
Ā 
IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024IVE Industry Focused Event - Defence Sector 2024
IVE Industry Focused Event - Defence Sector 2024
Ā 
POWER SYSTEMS-1 Complete notes examples
POWER SYSTEMS-1 Complete notes  examplesPOWER SYSTEMS-1 Complete notes  examples
POWER SYSTEMS-1 Complete notes examples
Ā 
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdfCCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
CCS355 Neural Networks & Deep Learning Unit 1 PDF notes with Question bank .pdf
Ā 
Earthing details of Electrical Substation
Earthing details of Electrical SubstationEarthing details of Electrical Substation
Earthing details of Electrical Substation
Ā 
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfg
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfgUnit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfg
Unit7-DC_Motors nkkjnsdkfnfcdfknfdgfggfg
Ā 
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptxExploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Exploring_Network_Security_with_JA3_by_Rakesh Seal.pptx
Ā 
Electronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdfElectronically Controlled suspensions system .pdf
Electronically Controlled suspensions system .pdf
Ā 
Past, Present and Future of Generative AI
Past, Present and Future of Generative AIPast, Present and Future of Generative AI
Past, Present and Future of Generative AI
Ā 
An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...An experimental study in using natural admixture as an alternative for chemic...
An experimental study in using natural admixture as an alternative for chemic...
Ā 

nural network ER. Abhishek k. upadhyay

  • 2. Basic Neuron Model In A Feedforward Network ā€¢ Inputs xi arrive through pre-synaptic connections ā€¢ Synaptic efficacy is modeled using real weights wi ā€¢ The response of the neuron is a nonlinear function f of its weighted inputs
  • 3.
  • 4. Task Plot the following type of Neural activation functions. 1(a) Threshold Function Ļ†(v)= +1 for vā‰„0 0 for v<0 1(b) Threshold Function Ļ†(v)= +1 for vā‰„0 -1 otherwise 2 Piecewise linear Function Ļ†(v)= 1 for vā‰„+1/2 v for +1/2>v>-1/2 0 for vā‰¤-1/2 3(a) Sigmoid Function Ļ†(v)=1/(1+ exp(-Ī»v)) 3(b) Sigmoid Function Ļ†(v)=2/(1+ exp(-Ī»v)) 3(c) Sigmoid Function Ļ†(v)=tanh(Ī»v) For 3 vary the value of ā€˜Ī»ā€™ and show the changes in the graph.
  • 6. Single Layer Artificial Neural Networks
  • 9. Banana & Apple Sorter
  • 12. Illustration of a Neural Network
  • 13. Different networks ā˜»Perceptron ā€“ Feedforward Network, Linear Decision Boundary, One Neuron for Each Decision ā˜»Hamming Network ā˜»Hopfield Network - Dynamic Associative Memory Network ā˜»Error Back Propagation network ā˜»Radial basis network ā˜»ART ā˜»Brain in a box neural network ā˜»Cellular neural Network ā˜»Neocognitron ā˜»Functional
  • 15. 1970s The Backpropagation algorithm was first proposed by Paul Werbos in the 1970's. However, it was rediscoved in 1986 by Rumelhart and McClelland & became widely used. It took 30 years before the error backpropagation (or in short: backprop) algorithm popularized.
  • 16.
  • 17. Differences In Networks Feedforward Networks ā€¢ Solutions are known ā€¢ Weights are learned ā€¢ Evolves in the weight space ā€¢ Used for: ā€“ Prediction ā€“ Classification ā€“ Function approximation Feedback Networks ā€¢ Solutions are unknown ā€¢ Weights are prescribed ā€¢ Evolves in the state space ā€¢ Used for: ā€“ Constraint satisfaction ā€“ Optimization ā€“ Feature matching
  • 18. Architecture A Back Prop network has atleast 3 layers of units: an input layer, at least one intermediate hidden layer, & an output layer. Connection weights in a Back Prop network are one way. Units are connected in a feed- forward fashion with input units fully connected to units in the hidden layer & hidden units fully connected to units in the output layer. When a Back Prop network is cycled, an input pattern is propagated forward to the output units through the intervening input-to-hidden and hidden-to- output weights.
  • 19. Inputs To Neurons ā€¢ Arise from other neurons or from outside the network ā€¢ Nodes whose inputs arise outside the network are called input nodes and simply copy values ā€¢ An input may excite or inhibit the response of the neuron to which it is applied, depending upon the weight of the connection
  • 21. Weights ā€¢ Represent synaptic efficacy and may be excitatory or inhibitory ā€¢ Normally, positive weights are considered as excitatory while negative weights are thought of as inhibitory ā€¢ Learning is the process of modifying the weights in order to produce a network that performs some function
  • 23. Output ā€¢ The response function is normally nonlinear ā€¢ Samples include ā€“ Sigmoid ā€“ Piecewise linear x e xf Ī»āˆ’ + = 1 1 )( ļ£³ ļ£² ļ£± < ā‰„ = Īø Īø xif xifx xf ,0 , )(
  • 24. Back propagation Networks I1 I2 1 Hidden Layer H1 H2 O1 O2 Output Layer Wi,j Wj,k 1ā€™s - bias āˆ‘ + = āˆ’ j jxj Hw e xO , 1 1 )( I3 1 āˆ‘ + = āˆ’ i ixi Iw e xH , 1 1 )(
  • 26. Backpropagation Preparation ā€¢ Training Set A collection of input-output patterns that are used to train the network ā€¢ Testing Set A collection of input-output patterns that are used to assess network performance ā€¢ Learning Rate-Ī· A scalar parameter, analogous to step size in numerical integration, used to set the rate of adjustments
  • 27. Learning ā€¢ Learning occurs during a training phase in which each input pattern in a training set is applied to the input units and then propagated forward. ā€¢ The pattern of activation arriving at the output layer is then compared with the correct output pattern to calculate an error signal. ā€¢ The error signal for each such target output pattern is then back propagated from the outputs to the inputs in order to appropriately adjust the weights in each layer of the network.
  • 28. Learning ā€¢ The process goes on for several cycles till the error reduces to a predefined limit. ā€¢ After a BackProp network has learned the correct classification for a set of inputs, it can be tested on a second set of inputs to see how well it classifies untrained patterns. ā€¢ Thus, an important consideration in applying BackProp learning is how well the network generalizes.
  • 29. The basic principles of the back propagation algorithm are: (1) the error of the output signal of a neuron is used to adjust its weights such that the error decreases, and (2) the error in hidden layers is estimated proportional to the weighted sum of the (estimated) errors in the layer above.
  • 31. During the training, the data is presented to the network several thousand times. For each data sample, the current output of the network is calculated and compared to the "true" target value. The error signal dj of neuron j is computed from the difference between the target and the calculated output. For hidden neurons, this difference is estimated by the weighted error signals of the layer above. The error terms are then used to adjust the weights wij of the neural network.
  • 32. A Pseudo-Code Algorithm ā€¢ Randomly choose the initial weights ā€¢ While error is too large ā€“ For each training pattern (presented in random order) ā€¢ Apply the inputs to the network ā€¢ Calculate the output for every neuron from the input layer, through the hidden layer(s), to the output layer ā€¢ Calculate the error at the outputs ā€¢ Use the output error to compute error signals for pre-output layers ā€¢ Use the error signals to compute weight adjustments ā€¢ Apply the weight adjustments ā€“ Periodically evaluate the network performance
  • 33. Network Error ā€¢ Total-Sum-Squared-Error (TSSE) ā€¢ Root-Mean-Squared-Error (RMSE) āˆ‘ āˆ‘ āˆ’= patterns outputs actualdesiredTSSE 2 )( 2 1 outputspatterns TSSE RMSE *## *2 =
  • 34.
  • 35. Apply Inputs From A Pattern ā€¢ Apply the value of each input parameter to each input node ā€¢ Input nodes computer only the identity function Feedforward Inputs Outputs
  • 36. Calculate Outputs For Each Neuron Based On The Pattern ā€¢ The output from neuron j for pattern p is Opj where and k ranges over the input indices and Wjk is the weight on the connection from input k to neuron j Feedforward Inputs Outputs jnetjpj e netO Ī»āˆ’ + = 1 1 )( āˆ‘+= k jkpkbiasj WOWbiasnet *
  • 37. Calculate The Error Signal For Each Output Neuron ā€¢ The output neuron error signal Ī“pj is given by Ī“pj=(Tpj-Opj) Opj (1-Opj) ā€¢ Tpj is the target value of output neuron j for pattern p ā€¢ Opj is the actual output value of output neuron j for pattern p
  • 38. Calculate The Error Signal For Each Hidden Neuron ā€¢ The hidden neuron error signal Ī“pj is given by where Ī“pk is the error signal of a post- synaptic neuron k and Wkj is the weight of the connection from hidden neuron j to the post-synaptic neuron k kj k pkpjpjpj WOO āˆ‘āˆ’= Ī“Ī“ )1(
  • 39. Calculate And Apply Weight Adjustments ā€¢ Compute weight adjustments āˆ†Wji at time t by āˆ†Wji(t)= Ī· Ī“pj Opi ā€¢ Apply weight adjustments according to Wji(t+1) = Wji(t) + āˆ†Wji(t) ā€¢ Some add a momentum term Ī±āˆ—āˆ†Wji(t-1)
  • 40. ā€¢ Thus, the network adjusts its weights after each data sample. This learning process is in fact a gradient descent in the error surface of the weight space - with all its drawbacks. The learning algorithm is slow and prone to getting stuck in a local minimum.
  • 41.
  • 42. Simulation Issues ļ‚§ How to Select Initial Weights ļ‚§ Local Minima ļ‚§ Solutions to Local minima ļ‚§ Rate of Learning ļ‚§ Stopping Criterion ļ‚§ Initialization
  • 43. ā€¢ For the standard back propagation algorithm, the initial weights of the multi-layer perceptron have to be relatively small. They can, for instance, be selected randomly from a small interval around zero. During training they are slowly adapted. Starting with small weights is crucial, because large weights are rigid and cannot be changed quickly.
  • 44. Sequential & Batch modes For a given training set ,back-propagation learning proceeds in two basic ways: 1. Sequential Mode 2. Batch Mode
  • 45. Sequential mode ā€¢ The sequential mode of back-propagation learning is also referred to as on-line, pattern or stochastic mode. ā€¢ To be specific, consider an epoch consisting of N training ex. Arranged in the order (x(1),d(1)),ā€¦,(x(N),d(N)). ā€¢ The first ex. pair (x(1),d(1))in the epoch is presented to the network,& the sequence of forward & backward computations described previously is performed, resulting in certain adjustments to the synaptic weights & bias level of the network. ā€¢ The second ex. (x(N),d(N)) in the epoch is presented,& the sequence of forward & backward computations is repeated, resulting in the further adjustments to the synaptic weights & bias levels. This process is continued until the last example pair (x(N),d(N)) in the epoch is accounted for.
  • 46. Batch Propagation ā€¢ In this mode of back-propagation learning weight updating is performed after the presentation of all the training examples that constitute an epoch. ā€¢ For a particular epoch, the cost function is the average squared error, reproduced here in composite form is defined as:- Ī¾av = (1/2N )Ī£ Ī£ ej 2 (n) for n=1 to N for j ā‚¬ C
  • 47. ā€¢ Let N denote the total no. of patterns contained in the training set. The average squared error energy is obtained by summing Ī¾(n) over all n and then normalizing with respect to the set size N, as shown by :- ā€¢ Ī¾av = 1/N Ī£ Ī¾(n) for n=1 to N
  • 48. Stopping Criteria ā€¢ The back-propagation algorithm cannot be shown to converge . ā€¢ To formulate a criterion, it is logically to think in terms of the unique properties of a local or global minimum. ā€¢ The back-propagation algorithm is considered to have converged when the Euclidean norm of the gradient vector reaches a sufficient small gradient threshold. ā€¢ The back-propagation algorithm is considered to have converged when the absolute rate of change in the average squared error pre epoch is sufficiently small. ā€¢ The drawback of this convergence criterion is that, for successful trials, learning time may be long.
  • 49.
  • 50. ā€¢ The back-propagation algorithm makes adjustments by computing the derivative, or slope of the network error with respect to each neuronā€™s output. It attempts to minimize the overall error by descending this slope to the minimum value for every weight. It advances one step down the slope each epoch. If the network takes steps that are too large, it may pass the global minimum. If it takes steps that are small, it may settle on local minima, or take an inordinate amount of time to arrive at the global minimum. The ideal step size for a given problem requires detailed, high-order derivative analysis, a task not performed by the algorithm.
  • 51.
  • 53. Local Minima ļ®For simple 2 layer networks (without a hidden layer), the error surface is bowl shaped and using gradient-descent to minimize error is not a problem; the network will always find an errorless solution (at the bottom of the bowl). Such errorless solutions are called global minima. ļ®However, extra hidden layer implies complex surfaces. Since some minima are deeper than others, it is possible that gradient descent may not find a global minima. Instead, the network may fall into local minima which represent suboptimal solutions.
  • 54. ā€¢ The algorithm cycles through the training samples as:- ā€¢ Initialization ā€¢ Presentation of training Examples ā€¢ Forward Computation
  • 55. Initialization ā€¢ Assuming that no prior information is available, pick the synaptic weights and thresholds from a uniform distribution whose mean is zero & whose variance is chosen to make the standard deviation of the induced local fields of the neurons lie at the transition between the linear and saturated parts of the sigmoid activation function.
  • 56. Presentation of training Examples Present the network with an epoch of training examples. For each example in the set order in same fashion, perform the sequence of forward and backward computation as described below.
  • 57. Solutions to Local minima ļ®Usual solution : More hidden layers. Logic - Although additional hidden units increase the complexity of the error surface, the extra dimensionalilty increases the number of possible escape routes. ļ®Our solution ā€“ Tunneling
  • 58. Rate of Learning ļ®If the learning rate Ī· is very small, then the algorithm proceeds slowly, but accurately follows the path of steepest descent in weight space. ļ®If Ī· is large, the algorithm may oscillate.
  • 59. ļ®A simple method of effectively increasing the rate of learning is to modify the delta rule by including a momentum term: Ī”wji (n) = Ī± Ī”wji (n-1) + Ī· Ī“j (n)yi (n) where Ī± is a positive constant termed the momentum constant. This is called the generalized delta rule. ļ‚§ The effect is that if the basic delta rule is consistently pushing a weight in the same direction, then it gradually gathers "momentum" in that direction.
  • 61. An Example: Exclusive ā€œORā€ ā€¢ Training set ā€“ ((0.1, 0.1), 0.1) ā€“ ((0.1, 0.9), 0.9) ā€“ ((0.9, 0.1), 0.9) ā€“ ((0.9, 0.9), 0.1) ā€¢ Testing set ā€“ Use at least 121 pairs equally spaced on the unit square and plot the results ā€“ Omit the training set (if desired)
  • 62. An Example (continued): Network Architectureinputs output(s)
  • 63. An Example (continued): Network Architecture Sample input 0.1 0.9 Target output 0.9 1 1 1
  • 64. Feedforward Network Training by Backpropagation: Process Summary ā€¢ Select an architecture ā€¢ Randomly initialize weights ā€¢ While error is too large ā€“ Select training pattern and feedforward to find actual network output ā€“ Calculate errors and backpropagate error signals ā€“ Adjust weights ā€¢ Evaluate performance using the test set
  • 65. An Example (continued): Network Architecture Sample input 0.1 0.9 Actual output ??? 1 1 1 ?? ?? ?? ?? ?? ?? ?? ?? ?? Target output 0.9
  • 66. Feedforward Network Training by Backpropagation: Process Summary ā€¢ Select an architecture ā€¢ Randomly initialize weights ā€¢ While error is too large ā€“ Select training pattern and feedforward to find actual network output ā€“ Calculate errors and backpropagate error signals ā€“ Adjust weights ā€¢ Evaluate performance using the test set
  • 67. Backpropagation ā€¢Very powerful - can learn any function, given enough hidden units! With enough hidden units, we can generate any function. ā€¢Have the same problems of Generalization vs. Memorization. With too many units, we will tend to memorize the input and not generalize well. Some schemes exist to ā€œpruneā€ the neural network.
  • 68. BackProp networks are not limited in its use because they can adapt their weights to acquire new knowledge. BackProp networks learn by example, and can be used to make predictions.
  • 69. Write a program to train and simulate neural network for following network ā€“ Input Nodes = 2 & Output Nodes = 1 ā€“ Input Nodes = 3 and Output nodes = 1 Inputs Outputs A B Y 0 0 0 0 1 1 1 0 1 1 1 0 Inputs Outputs A B C Y 0 0 0 0 0 0 1 0 0 1 0 0 0 1 1 0 1 0 0 1 1 0 1 1 1 1 0 1 1 1 1 1
  • 70. ā€¢ Artificial Neural Network ā€“ Simon Haykin ā€¢ Artificial Neural Network ā€“ Jacek Zurada