Title: FeedForward Neural Networks
1Feed-Forward Neural Networks
2Content
- Introduction
- Multilayer Perceptron
- Back Propagation Learning Algorithm
3Feed-Forward Neural Networks
4Artificial Neural Networks
- To simulate human brain behavior
- A new generation of information processing system.
5Applications
- Pattern Matching
- Pattern Recognition
- Associate Memory (Content Addressable Memory)
- Function Approximation
- Learning
- Optimization
- Vector Quantization
- Data Clustering
6Applications
Traditional Computers are inefficient on these
tasks, although their computation speed is fast.
- Pattern Matching
- Pattern Recognition
- Associate Memory (Content Addressable Memory)
- Function Approximation
- Learning
- Optimization
- Vector Quantization
- Data Clustering
7Processing Units of an ANN
- The Configuration of ANNs
- Consist of a large number of interconnected
processing elements called neurons. - A human brain consists of 1011 neurons of many
different types. - How ANN works?
- Collective behavior.
8The Biologic Neurons
9The Artificial Neurons
10The Artificial Neurons
wij positive excitatory negative
inhibitory zero no connection ?i bias
Proposed by McCulloch and Pitts 1943 called M-P
neurons
11What Can a Neuron Do?
- A hard limiter.
- A binary threshold unit.
- Hyperspace separation.
0
1
12What Can an ANN Do?
- A neurally inspired mathematical model.
- Consists a large number of highly interconnected
PEs. - Its connections (weights) holds knowledge.
- The response of PE depends only on local
information. - Its collective behavior demonstrates the
computation power. - With learning, recalling and, generalization
capability.
13Basis Entities of an ANN
- Models of Neurons or PEs.
- Models of synaptic interconnections and
structures. - Training or learning rules
14Feed-Forward Neural Networks
15Single-Layer Perceptron
Training Set
16Single-Layer Perceptron
Training Set
What it can?
What it cannot?
17Multilayer Perceptron
Output Layer
Hidden Layer
Input Layer
18Multilayer Perceptron
Where the knowledge from?
Classification
Output
Analysis
Learning
Input
19How an MLP Works?
Example
- Not linearly separable.
- Is a single layer perceptron workable?
XOR
20How an MLP Works?
Example
00
01
11
21How an MLP Works?
Example
00
01
11
22How an MLP Works?
Example
00
01
11
23How an MLP Works?
Example
24Parity Problem
Is the problem linearly separable?
25Parity Problem
x3
P1
P2
x2
P3
x1
26Parity Problem
111
011
001
000
27Parity Problem
111
011
001
000
28Parity Problem
111
P4
011
001
000
29Parity Problem
P4
30General Problem
31General Problem
32Hyperspace Partition
33Region Encoding
001
000
010
100
101
110
111
34Hyperspace Partition Region Encoding Layer
35Region Identification Layer
36Region Identification Layer
37Region Identification Layer
38Region Identification Layer
39Region Identification Layer
40Region Identification Layer
41Region Identification Layer
42Classification
0
?1
1
43Feed-Forward Neural Networks
- Back Propagation Learning algorithm
44Activation Function Sigmoid
Remember this
45Supervised Learning
Training Set
Output Layer
Hidden Layer
Input Layer
46Supervised Learning
Training Set
Sum of Squared Errors
Goal
Minimize
47Back Propagation Learning Algorithm
- Learning on Output Neurons
- Learning on Hidden Neurons
48Learning on Output Neurons
?
?
49Learning on Output Neurons
depends on the activation function
50Learning on Output Neurons
Using sigmoid,
51Learning on Output Neurons
Using sigmoid,
52Learning on Output Neurons
53Learning on Output Neurons
How to train the weights connecting to output
neurons?
54Learning on Hidden Neurons
?
?
55Learning on Hidden Neurons
56Learning on Hidden Neurons
?
57Learning on Hidden Neurons
58Learning on Hidden Neurons
59Back Propagation
60Back Propagation
61Back Propagation
62Learning Factors
- Initial Weights
- Learning Constant (?)
- Cost Functions
- Momentum
- Update Rules
- Training Data and Generalization
- Number of Layers
- Number of Hidden Nodes
63Reading Assignments
- Shi Zhong and Vladimir Cherkassky, Factors
Controlling Generalization Ability of MLP
Networks. In Proc. IEEE Int. Joint Conf. on
Neural Networks, vol. 1, pp. 625-630, Washington
DC. July 1999. (http//www.cse.fau.edu/zhong/pubs
.htm) - Rumelhart, D. E., Hinton, G. E., and Williams, R.
J. (1986b). "Learning Internal Representations by
Error Propagation," in Parallel Distributed
Processing Explorations in the Microstructure of
Cognition, vol. I, D. E. Rumelhart, J. L.
McClelland, and the PDP Research Group. MIT
Press, Cambridge (1986). - (http//www.cnbc.cmu.edu/plaut/85-419/papers/Rum
elhartETAL86.backprop.pdf).