19_Learning.ppt

gnans Kgnanshek
gnans Kgnanshekprofessor à apec
Learning in Neural Networks
Neurons and the Brain
Neural Networks
Perceptrons
Multi-layer Networks
Applications
The Hopfield Network
Neural Networks
A model of reasoning based on the human brain
complex networks of simple computing elements
capable of learning from examples
 with appropriate learning methods
collection of simple elements performs high-level
operations
Neural Networks and the Brain (Cont.)
 The human brain incorporates nearly 10 billion
neurons and 60 trillion connections between them.
 Our brain can be considered as a highly complex,
non-linear and parallel information-processing
system.
 Learning is a fundamental and essential
characteristic of biological neural networks.
Artificial Neuron (Perceptron) Diagram
 weighted inputs are summed up by the input function
 the (nonlinear) activation function calculates the activation
value, which determines the output
[Russell & Norvig, 1995]
Common Activation Functions
Stept(x) = 1 if x >= t, else 0
Sign(x) = +1 if x >= 0, else –1
Sigmoid(x) = 1/(1+e-x)
[Russell & Norvig, 1995]
Neural Networks and Logic Gates
simple neurons can act as logic gates
 appropriate choice of activation function, threshold, and
weights
 step function as activation function
[Russell & Norvig, 1995]
Network Structures
layered structures
 networks are arranged into layers
 interconnections mostly between two layers
 some networks may have feedback connections
Perceptrons
 single layer, feed-
forward network
 historically one of the
first types of neural
networks
 late 1950s
 the output is
calculated as a step
function applied to
the weighted sum of
inputs
 capable of learning
simple functions
 linearly separable
[Russell & Norvig, 1995]
[Russell & Norvig, 1995]
Perceptrons and Linear Separability
perceptrons can deal with linearly separable
functions
some simple functions are not linearly separable
 XOR function
0,0
0,1
1,0
1,1
0,0
0,1
1,0
1,1
AND XOR
Perceptrons and Linear Separability
 linear separability can be extended to more than two dimensions
 more difficult to visualize
[Russell & Norvig, 1995]
How does the perceptron learn its
classification tasks?
This is done by making small adjustments in the
weights
 to reduce the difference between the actual and desired
outputs of the perceptron.
The initial weights are randomly assigned
 usually in the range [0.5, 0.5], or [0, 1]
Then the they are updated to obtain the output
consistent with the training examples.
Perceptrons and Learning
perceptrons can learn from examples through a
simple learning rule. For each example row
(iteration), do the following:
 calculate the error of a unit Erri as the difference between
the correct output Ti and the calculated output Oi
Erri = Ti - Oi
 adjust the weight Wj of the input Ij such that the error
decreases
Wij = Wij +  *Iij * Errij
  is the learning rate, a positive constant less than unity.
 this is a gradient descent search through the weight space
Example of
perceptron
learning: the
logical
operation
AND
Inputs
x1 x2
0
0
1
1
0
1
0
1
0
0
0
Epoch
Desired
output
Yd
1
Initial
weights
w1 w2
1
0.3
0.3
0.3
0.2
0.1
0.1
0.1
0.1
0
0
1
0
Actual
output
Y
Error
e
0
0
1
1
Final
weights
w1 w2
0.3
0.3
0.2
0.3
0.1
0.1
0.1
0.0
0
0
1
1
0
1
0
1
0
0
0
2
1
0.3
0.3
0.3
0.2
0
0
1
1
0
0
1
0
0.3
0.3
0.2
0.2
0.0
0.0
0.0
0.0
0
0
1
1
0
1
0
1
0
0
0
3
1
0.2
0.2
0.2
0.1
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0.0
0
0
1
0
0
0
1
1
0.2
0.2
0.1
0.2
0.0
0.0
0.0
0.1
0
0
1
1
0
1
0
1
0
0
0
4
1
0.2
0.2
0.2
0.1
0.1
0.1
0.1
0.1
0
0
1
1
0
0
1
0
0.2
0.2
0.1
0.1
0.1
0.1
0.1
0.1
0
0
1
1
0
1
0
1
0
0
0
5
1
0.1
0.1
0.1
0.1
0.1
0.1
0.1
0.1
0
0
0
1
0
0
0
0.1
0.1
0.1
0.1
0.1
0.1
0.1
0.1
0
Threshold:  = 0.2; learning rate:  = 0.1
Two-dimensional plots of basic logical
operations
x1
x2
1
(a) AND (x1  x2)
1
x1
x2
1
1
(b) OR (x1  x2)
x1
x2
1
1
(c) Exclusive-OR
(x1  x2)
0
0 0
A perceptron can learn the operations AND and
OR, but not Exclusive-OR.
Multi-Layer Neural Networks
The network consists of an input layer of source
neurons, at least one middle or hidden layer of
computational neurons, and an output layer of
computational neurons.
The input signals are propagated in a forward
direction on a layer-by-layer basis
 feedforward neural network
the back-propagation learning algorithm can be used
for learning in multi-layer networks
Diagram Multi-Layer Network
two-layer network
 input units Ik
 usually not counted as a
separate layer
 hidden units aj
 output units Oi
usually all nodes of one
layer have weighted
connections to all nodes
of the next layer
Ik
aj
Oi
Wji
Wkj
Input
layer
First
hidden
layer
Second
hidden
layer
Output
layer
O
u
t
p
u
t
S
i
g
n
a
l
s
I
n
p
u
t
S
i
g
n
a
l
s
Multilayer perceptron with two hidden
layers
Back-Propagation Algorithm
Learning in a multilayer network proceeds the same
way as for a perceptron.
A training set of input patterns is presented to the
network.
The network computes its output pattern, and if there
is an error  or in other words a difference between
actual and desired output patterns  the weights are
adjusted to reduce this error.
 proceeds from the output layer to the hidden layer(s)
 updates the weights of the units leading to the layer
Back-Propagation Algorithm
In a back-propagation neural network, the learning
algorithm has two phases.
First, a training input pattern is presented to the
network input layer. The network propagates the
input pattern from layer to layer until the output
pattern is generated by the output layer.
If this pattern is different from the desired output, an
error is calculated and then propagated backwards
through the network from the output layer to the
input layer. The weights are modified as the error is
propagated.
Three-layer Feed-Forward Neural Network
( trained using back-propagation algorithm)
Input
layer
xi
x1
x2
xn
1
2
i
n
Output
layer
1
2
k
l
yk
y1
y2
yl
Input signals
Error signals
wjk
Hidden
layer
wij
1
2
j
m
Three-layer network for solving the Exclusive-
OR operation
y5
5
x1 3
1
x2
Input
layer
Output
layer
Hidden layer
4
2
3
w13
w24
w23
w24
w35
w45
4
5
1
1
1
Final results of three-layer network learning
Inputs
x1 x2
1
0
1
0
1
1
0
0
0
1
1
Desired
output
yd
0
0.0155
Actual
output
y5
Y
Error
e
Sum of
squared
errors
e
0.9849
0.9849
0.0175
0.0155
0.0151
0.0151
0.0175
0.0010
Network for solving the Exclusive-OR operation
y5
5
x1 3
1
x2 4
2
+1.0
1
1
1
+1.0
+1.0
+1.0
+1.5
+1.0
+1.0
+0.5
+0.5
(a) Decision boundary constructed by hidden neuron 3;
(b) Decision boundary constructed by hidden neuron 4;
(c) Decision boundaries constructed by the complete
three-layer network
x1
x2
1
(a)
1
x2
1
1
(b)
0
0
x1 + x2 – 1.5 = 0 x1 + x2 – 0.5 = 0
x1 x1
x2
1
1
(c)
0
Decision boundaries
Capabilities of Multi-Layer Neural
Networks
expressiveness
 weaker than predicate logic
 good for continuous inputs and outputs
computational efficiency
 training time can be exponential in the number of inputs
 depends critically on parameters like the learning rate
 local minima are problematic
 can be overcome by simulated annealing, at additional cost
generalization
 works reasonably well for some functions (classes of
problems)
 no formal characterization of these functions
Capabilities of Multi-Layer Neural
Networks (cont.)
sensitivity to noise
 very tolerant
 they perform nonlinear regression
transparency
 neural networks are essentially black boxes
 there is no explanation or trace for a particular answer
 tools for the analysis of networks are very limited
 some limited methods to extract rules from networks
prior knowledge
 very difficult to integrate since the internal representation
of the networks is not easily accessible
Applications
domains and tasks where neural networks are
successfully used
 recognition
 control problems
 series prediction
 weather, financial forecasting
 categorization
 sorting of items (fruit, characters, …)
 Neural networks were designed on analogy with
the brain.
 The brain’s memory, however, works by
association.
 For example, we can recognise a familiar face even in an
unfamiliar environment within 100-200 ms.
 We can also recall a complete sensory experience,
including sounds and scenes, when we hear only a few bars
of music.
 The brain routinely associates one thing with
another.
The Hopfield Network
 Multilayer neural networks trained with the back-
propagation algorithm are used for pattern
recognition problems.
 However, to emulate the human memory’s
associative characteristics we need a different type
of network: a recurrent neural network.
 A recurrent neural network has feedback loops
from its outputs to its inputs.
Single-layer n-neuron Hopfield network
xi
x1
x2
xn
I
n
p
u
t
S
i
g
n
a
l
s
yi
y1
y2
yn
1
2
i
n
O
u
t
p
u
t
S
i
g
n
a
l
s
 The stability of recurrent networks was solved only in
1982, when John Hopfield formulated the physical
principle of storing information in a dynamically
stable network.
Chapter Summary
learning is very important for agents to improve their
decision-making process
 unknown environments, changes, time constraints
most methods rely on inductive learning
 a function is approximated from sample input-output pairs
 neural networks consist of simple interconnected
computational elements
multi-layer feed-forward networks can learn any
function
 provided they have enough units and time to learn
1 sur 31

Recommandé

ACUMENS ON NEURAL NET AKG 20 7 23.pptx par
ACUMENS ON NEURAL NET AKG 20 7 23.pptxACUMENS ON NEURAL NET AKG 20 7 23.pptx
ACUMENS ON NEURAL NET AKG 20 7 23.pptxgnans Kgnanshek
2 vues128 diapositives
lecture07.ppt par
lecture07.pptlecture07.ppt
lecture07.pptbutest
12.9K vues73 diapositives
Artificial neural networks par
Artificial neural networks Artificial neural networks
Artificial neural networks ShwethaShreeS
122 vues24 diapositives
Soft Computing-173101 par
Soft Computing-173101Soft Computing-173101
Soft Computing-173101AMIT KUMAR
3.1K vues19 diapositives
SOFT COMPUTERING TECHNICS -Unit 1 par
SOFT COMPUTERING TECHNICS -Unit 1SOFT COMPUTERING TECHNICS -Unit 1
SOFT COMPUTERING TECHNICS -Unit 1sravanthi computers
1.5K vues77 diapositives
Perceptron par
PerceptronPerceptron
PerceptronNagarajan
29.5K vues49 diapositives

Contenu connexe

Similaire à 19_Learning.ppt

Supervised Learning par
Supervised LearningSupervised Learning
Supervised Learningbutest
9.8K vues73 diapositives
ANN.pptx par
ANN.pptxANN.pptx
ANN.pptxAROCKIAJAYAIECW
7 vues16 diapositives
ANN.ppt par
ANN.pptANN.ppt
ANN.pptUI19EC21Talluriissak
14 vues57 diapositives
Neural Networks Ver1 par
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1ncct
2.2K vues56 diapositives
Neural networks par
Neural networksNeural networks
Neural networksBasil John
1.5K vues57 diapositives
10-Perceptron.pdf par
10-Perceptron.pdf10-Perceptron.pdf
10-Perceptron.pdfESTIBALYZJIMENEZCAST
2 vues40 diapositives

Similaire à 19_Learning.ppt(20)

Supervised Learning par butest
Supervised LearningSupervised Learning
Supervised Learning
butest9.8K vues
Neural Networks Ver1 par ncct
Neural  Networks  Ver1Neural  Networks  Ver1
Neural Networks Ver1
ncct2.2K vues
Neural networks par Basil John
Neural networksNeural networks
Neural networks
Basil John1.5K vues
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience par hirokazutanaka
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
JAISTサマースクール2016「脳を知るための理論」講義04 Neural Networks and Neuroscience
hirokazutanaka1.2K vues
Artificial neural network by arpit_sharma par Er. Arpit Sharma
Artificial neural network by arpit_sharmaArtificial neural network by arpit_sharma
Artificial neural network by arpit_sharma
Er. Arpit Sharma217 vues
Artificial Neural Network_VCW (1).pptx par pratik610182
Artificial Neural Network_VCW (1).pptxArtificial Neural Network_VCW (1).pptx
Artificial Neural Network_VCW (1).pptx
pratik61018213 vues
Artificial Neural Networks-Supervised Learning Models par DrBaljitSinghKhehra
Artificial Neural Networks-Supervised Learning ModelsArtificial Neural Networks-Supervised Learning Models
Artificial Neural Networks-Supervised Learning Models
Artificial Neural Networks-Supervised Learning Models par DrBaljitSinghKhehra
Artificial Neural Networks-Supervised Learning ModelsArtificial Neural Networks-Supervised Learning Models
Artificial Neural Networks-Supervised Learning Models
Economic Load Dispatch (ELD), Economic Emission Dispatch (EED), Combined Econ... par cscpconf
Economic Load Dispatch (ELD), Economic Emission Dispatch (EED), Combined Econ...Economic Load Dispatch (ELD), Economic Emission Dispatch (EED), Combined Econ...
Economic Load Dispatch (ELD), Economic Emission Dispatch (EED), Combined Econ...
cscpconf58 vues
Artifical Neural Network and its applications par Sangeeta Tiwari
Artifical Neural Network and its applicationsArtifical Neural Network and its applications
Artifical Neural Network and its applications
Sangeeta Tiwari1.6K vues

Plus de gnans Kgnanshek

EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptx par
EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptxEDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptx
EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptxgnans Kgnanshek
8 vues26 diapositives
types of research.pptx par
types of research.pptxtypes of research.pptx
types of research.pptxgnans Kgnanshek
3 vues22 diapositives
CS8601 4 MC NOTES.pdf par
CS8601 4 MC NOTES.pdfCS8601 4 MC NOTES.pdf
CS8601 4 MC NOTES.pdfgnans Kgnanshek
1 vue84 diapositives
CS8601 3 MC NOTES.pdf par
CS8601 3 MC NOTES.pdfCS8601 3 MC NOTES.pdf
CS8601 3 MC NOTES.pdfgnans Kgnanshek
3 vues72 diapositives
CS8601 2 MC NOTES.pdf par
CS8601 2 MC NOTES.pdfCS8601 2 MC NOTES.pdf
CS8601 2 MC NOTES.pdfgnans Kgnanshek
1 vue55 diapositives
CS8601 1 MC NOTES.pdf par
CS8601 1 MC NOTES.pdfCS8601 1 MC NOTES.pdf
CS8601 1 MC NOTES.pdfgnans Kgnanshek
6 vues51 diapositives

Plus de gnans Kgnanshek(18)

EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptx par gnans Kgnanshek
EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptxEDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptx
EDC SLIDES PRESENTATION PRINCIPAL ON WEDNESDAY.pptx
unit-1 MANAGEMENT AND ORGANIZATIONS.pptx par gnans Kgnanshek
unit-1 MANAGEMENT AND ORGANIZATIONS.pptxunit-1 MANAGEMENT AND ORGANIZATIONS.pptx
unit-1 MANAGEMENT AND ORGANIZATIONS.pptx
gnans Kgnanshek10 vues

Dernier

Retail Store Scavenger Hunt.pptx par
Retail Store Scavenger Hunt.pptxRetail Store Scavenger Hunt.pptx
Retail Store Scavenger Hunt.pptxjmurphy154
52 vues10 diapositives
12.5.23 Poverty and Precarity.pptx par
12.5.23 Poverty and Precarity.pptx12.5.23 Poverty and Precarity.pptx
12.5.23 Poverty and Precarity.pptxmary850239
162 vues30 diapositives
MercerJesse3.0.pdf par
MercerJesse3.0.pdfMercerJesse3.0.pdf
MercerJesse3.0.pdfjessemercerail
92 vues6 diapositives
Jibachha publishing Textbook.docx par
Jibachha publishing Textbook.docxJibachha publishing Textbook.docx
Jibachha publishing Textbook.docxDrJibachhaSahVetphys
54 vues14 diapositives
Java Simplified: Understanding Programming Basics par
Java Simplified: Understanding Programming BasicsJava Simplified: Understanding Programming Basics
Java Simplified: Understanding Programming BasicsAkshaj Vadakkath Joshy
625 vues155 diapositives
NodeJS and ExpressJS.pdf par
NodeJS and ExpressJS.pdfNodeJS and ExpressJS.pdf
NodeJS and ExpressJS.pdfArthyR3
47 vues17 diapositives

Dernier(20)

Retail Store Scavenger Hunt.pptx par jmurphy154
Retail Store Scavenger Hunt.pptxRetail Store Scavenger Hunt.pptx
Retail Store Scavenger Hunt.pptx
jmurphy15452 vues
12.5.23 Poverty and Precarity.pptx par mary850239
12.5.23 Poverty and Precarity.pptx12.5.23 Poverty and Precarity.pptx
12.5.23 Poverty and Precarity.pptx
mary850239162 vues
NodeJS and ExpressJS.pdf par ArthyR3
NodeJS and ExpressJS.pdfNodeJS and ExpressJS.pdf
NodeJS and ExpressJS.pdf
ArthyR347 vues
ANGULARJS.pdf par ArthyR3
ANGULARJS.pdfANGULARJS.pdf
ANGULARJS.pdf
ArthyR349 vues
Narration lesson plan par TARIQ KHAN
Narration lesson planNarration lesson plan
Narration lesson plan
TARIQ KHAN69 vues
11.30.23A Poverty and Inequality in America.pptx par mary850239
11.30.23A Poverty and Inequality in America.pptx11.30.23A Poverty and Inequality in America.pptx
11.30.23A Poverty and Inequality in America.pptx
mary85023986 vues
Career Building in AI - Technologies, Trends and Opportunities par WebStackAcademy
Career Building in AI - Technologies, Trends and OpportunitiesCareer Building in AI - Technologies, Trends and Opportunities
Career Building in AI - Technologies, Trends and Opportunities
WebStackAcademy41 vues
Creative Restart 2023: Leonard Savage - The Permanent Brief: Unearthing unobv... par Taste
Creative Restart 2023: Leonard Savage - The Permanent Brief: Unearthing unobv...Creative Restart 2023: Leonard Savage - The Permanent Brief: Unearthing unobv...
Creative Restart 2023: Leonard Savage - The Permanent Brief: Unearthing unobv...
Taste53 vues
EILO EXCURSION PROGRAMME 2023 par info33492
EILO EXCURSION PROGRAMME 2023EILO EXCURSION PROGRAMME 2023
EILO EXCURSION PROGRAMME 2023
info33492181 vues
JRN 362 - Lecture Twenty-Three (Epilogue) par Rich Hanley
JRN 362 - Lecture Twenty-Three (Epilogue)JRN 362 - Lecture Twenty-Three (Epilogue)
JRN 362 - Lecture Twenty-Three (Epilogue)
Rich Hanley41 vues
JQUERY.pdf par ArthyR3
JQUERY.pdfJQUERY.pdf
JQUERY.pdf
ArthyR3103 vues

19_Learning.ppt

  • 1. Learning in Neural Networks Neurons and the Brain Neural Networks Perceptrons Multi-layer Networks Applications The Hopfield Network
  • 2. Neural Networks A model of reasoning based on the human brain complex networks of simple computing elements capable of learning from examples  with appropriate learning methods collection of simple elements performs high-level operations
  • 3. Neural Networks and the Brain (Cont.)  The human brain incorporates nearly 10 billion neurons and 60 trillion connections between them.  Our brain can be considered as a highly complex, non-linear and parallel information-processing system.  Learning is a fundamental and essential characteristic of biological neural networks.
  • 4. Artificial Neuron (Perceptron) Diagram  weighted inputs are summed up by the input function  the (nonlinear) activation function calculates the activation value, which determines the output [Russell & Norvig, 1995]
  • 5. Common Activation Functions Stept(x) = 1 if x >= t, else 0 Sign(x) = +1 if x >= 0, else –1 Sigmoid(x) = 1/(1+e-x) [Russell & Norvig, 1995]
  • 6. Neural Networks and Logic Gates simple neurons can act as logic gates  appropriate choice of activation function, threshold, and weights  step function as activation function [Russell & Norvig, 1995]
  • 7. Network Structures layered structures  networks are arranged into layers  interconnections mostly between two layers  some networks may have feedback connections
  • 8. Perceptrons  single layer, feed- forward network  historically one of the first types of neural networks  late 1950s  the output is calculated as a step function applied to the weighted sum of inputs  capable of learning simple functions  linearly separable [Russell & Norvig, 1995]
  • 9. [Russell & Norvig, 1995] Perceptrons and Linear Separability perceptrons can deal with linearly separable functions some simple functions are not linearly separable  XOR function 0,0 0,1 1,0 1,1 0,0 0,1 1,0 1,1 AND XOR
  • 10. Perceptrons and Linear Separability  linear separability can be extended to more than two dimensions  more difficult to visualize [Russell & Norvig, 1995]
  • 11. How does the perceptron learn its classification tasks? This is done by making small adjustments in the weights  to reduce the difference between the actual and desired outputs of the perceptron. The initial weights are randomly assigned  usually in the range [0.5, 0.5], or [0, 1] Then the they are updated to obtain the output consistent with the training examples.
  • 12. Perceptrons and Learning perceptrons can learn from examples through a simple learning rule. For each example row (iteration), do the following:  calculate the error of a unit Erri as the difference between the correct output Ti and the calculated output Oi Erri = Ti - Oi  adjust the weight Wj of the input Ij such that the error decreases Wij = Wij +  *Iij * Errij   is the learning rate, a positive constant less than unity.  this is a gradient descent search through the weight space
  • 13. Example of perceptron learning: the logical operation AND Inputs x1 x2 0 0 1 1 0 1 0 1 0 0 0 Epoch Desired output Yd 1 Initial weights w1 w2 1 0.3 0.3 0.3 0.2 0.1 0.1 0.1 0.1 0 0 1 0 Actual output Y Error e 0 0 1 1 Final weights w1 w2 0.3 0.3 0.2 0.3 0.1 0.1 0.1 0.0 0 0 1 1 0 1 0 1 0 0 0 2 1 0.3 0.3 0.3 0.2 0 0 1 1 0 0 1 0 0.3 0.3 0.2 0.2 0.0 0.0 0.0 0.0 0 0 1 1 0 1 0 1 0 0 0 3 1 0.2 0.2 0.2 0.1 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0.0 0 0 1 0 0 0 1 1 0.2 0.2 0.1 0.2 0.0 0.0 0.0 0.1 0 0 1 1 0 1 0 1 0 0 0 4 1 0.2 0.2 0.2 0.1 0.1 0.1 0.1 0.1 0 0 1 1 0 0 1 0 0.2 0.2 0.1 0.1 0.1 0.1 0.1 0.1 0 0 1 1 0 1 0 1 0 0 0 5 1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0 0 0 1 0 0 0 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0.1 0 Threshold:  = 0.2; learning rate:  = 0.1
  • 14. Two-dimensional plots of basic logical operations x1 x2 1 (a) AND (x1  x2) 1 x1 x2 1 1 (b) OR (x1  x2) x1 x2 1 1 (c) Exclusive-OR (x1  x2) 0 0 0 A perceptron can learn the operations AND and OR, but not Exclusive-OR.
  • 15. Multi-Layer Neural Networks The network consists of an input layer of source neurons, at least one middle or hidden layer of computational neurons, and an output layer of computational neurons. The input signals are propagated in a forward direction on a layer-by-layer basis  feedforward neural network the back-propagation learning algorithm can be used for learning in multi-layer networks
  • 16. Diagram Multi-Layer Network two-layer network  input units Ik  usually not counted as a separate layer  hidden units aj  output units Oi usually all nodes of one layer have weighted connections to all nodes of the next layer Ik aj Oi Wji Wkj
  • 18. Back-Propagation Algorithm Learning in a multilayer network proceeds the same way as for a perceptron. A training set of input patterns is presented to the network. The network computes its output pattern, and if there is an error  or in other words a difference between actual and desired output patterns  the weights are adjusted to reduce this error.  proceeds from the output layer to the hidden layer(s)  updates the weights of the units leading to the layer
  • 19. Back-Propagation Algorithm In a back-propagation neural network, the learning algorithm has two phases. First, a training input pattern is presented to the network input layer. The network propagates the input pattern from layer to layer until the output pattern is generated by the output layer. If this pattern is different from the desired output, an error is calculated and then propagated backwards through the network from the output layer to the input layer. The weights are modified as the error is propagated.
  • 20. Three-layer Feed-Forward Neural Network ( trained using back-propagation algorithm) Input layer xi x1 x2 xn 1 2 i n Output layer 1 2 k l yk y1 y2 yl Input signals Error signals wjk Hidden layer wij 1 2 j m
  • 21. Three-layer network for solving the Exclusive- OR operation y5 5 x1 3 1 x2 Input layer Output layer Hidden layer 4 2 3 w13 w24 w23 w24 w35 w45 4 5 1 1 1
  • 22. Final results of three-layer network learning Inputs x1 x2 1 0 1 0 1 1 0 0 0 1 1 Desired output yd 0 0.0155 Actual output y5 Y Error e Sum of squared errors e 0.9849 0.9849 0.0175 0.0155 0.0151 0.0151 0.0175 0.0010
  • 23. Network for solving the Exclusive-OR operation y5 5 x1 3 1 x2 4 2 +1.0 1 1 1 +1.0 +1.0 +1.0 +1.5 +1.0 +1.0 +0.5 +0.5
  • 24. (a) Decision boundary constructed by hidden neuron 3; (b) Decision boundary constructed by hidden neuron 4; (c) Decision boundaries constructed by the complete three-layer network x1 x2 1 (a) 1 x2 1 1 (b) 0 0 x1 + x2 – 1.5 = 0 x1 + x2 – 0.5 = 0 x1 x1 x2 1 1 (c) 0 Decision boundaries
  • 25. Capabilities of Multi-Layer Neural Networks expressiveness  weaker than predicate logic  good for continuous inputs and outputs computational efficiency  training time can be exponential in the number of inputs  depends critically on parameters like the learning rate  local minima are problematic  can be overcome by simulated annealing, at additional cost generalization  works reasonably well for some functions (classes of problems)  no formal characterization of these functions
  • 26. Capabilities of Multi-Layer Neural Networks (cont.) sensitivity to noise  very tolerant  they perform nonlinear regression transparency  neural networks are essentially black boxes  there is no explanation or trace for a particular answer  tools for the analysis of networks are very limited  some limited methods to extract rules from networks prior knowledge  very difficult to integrate since the internal representation of the networks is not easily accessible
  • 27. Applications domains and tasks where neural networks are successfully used  recognition  control problems  series prediction  weather, financial forecasting  categorization  sorting of items (fruit, characters, …)
  • 28.  Neural networks were designed on analogy with the brain.  The brain’s memory, however, works by association.  For example, we can recognise a familiar face even in an unfamiliar environment within 100-200 ms.  We can also recall a complete sensory experience, including sounds and scenes, when we hear only a few bars of music.  The brain routinely associates one thing with another. The Hopfield Network
  • 29.  Multilayer neural networks trained with the back- propagation algorithm are used for pattern recognition problems.  However, to emulate the human memory’s associative characteristics we need a different type of network: a recurrent neural network.  A recurrent neural network has feedback loops from its outputs to its inputs.
  • 30. Single-layer n-neuron Hopfield network xi x1 x2 xn I n p u t S i g n a l s yi y1 y2 yn 1 2 i n O u t p u t S i g n a l s  The stability of recurrent networks was solved only in 1982, when John Hopfield formulated the physical principle of storing information in a dynamically stable network.
  • 31. Chapter Summary learning is very important for agents to improve their decision-making process  unknown environments, changes, time constraints most methods rely on inductive learning  a function is approximated from sample input-output pairs  neural networks consist of simple interconnected computational elements multi-layer feed-forward networks can learn any function  provided they have enough units and time to learn