Title: Lecture 6. Perceptron Simplest Neural Network
1Lecture 6. Perceptron Simplest Neural Network
Saint-Petersburg State Polytechnic
University Distributed Intelligent Systems
Dept ARTIFICIAL NEURAL NETWORKS
- Prof. Dr. Viacheslav P. Shkodyrev
- EMail shkodyrev_at_imop.spbstu.ru
Perceptron is one of the first and simplest
artificial neural networks, which was presented
in the middle of 50-th years. It was a first
mathematical model, which demonstrates new
paradigms of machine learning computational
environment, and a threshold logic units model of
classification tasks.
2 Perceptron was proposed by Rosenblatt (1958)
as the first model for learning with a teacher,
i.e. supervised learning. This model is an
enhancement of threshold logic units (TLUs)
which used for the classification of patterns
said to be linearly separable. How to formalize
and interpret this model?
- Objective of Lecture 6
- This lecture introduces the simplest class
of neural networks perseptron and its
application to pattern classification. It is
possible to interpret the functionality of
perceptron and threshold logic units model
geometrically via a separating hyper plane. In
this lecture we will define what we mean by a
perceptron learning rule, explain the P networks
and its training algorithms, discuss the
limitation of the P networks. You will learn - What is a single-layer perceptron via
threshold logic units model - Perceptrons as Linear and Non-Linear
Classifiers via threshold logic theory - Multi-Layer Perceptrons networks
- Perceptron Learning Rules
3It was P which presents at first time a new
paradigms of training computation algorithms.
We solve a classification task when we
assign any image, represented by a feature
vector, to one of two classes, which we shall
denote by F of A, so that class A corresponds to
the character a and class F corresponds to b.
Using the perceptron training algorithm, we
may to classify two linearly separable classes
4Architecture of SLP
The single-layer perceptron was the first
simplest model that generates a great interest
due to its ability to generalize from its
training vectors and learn from initially
randomly distributed connections.
Input pattern x
5Geometric interpretation of Threshold Logic Units
6Exclusive-Or (XOR) gate Problem Minsly,
Papert, 1969
Solution of the Exclusive-Or (XOR) gate Problem
x2
Linear separable surface can not to solve the
Exclusive-Or gate classification tasks
XOR-Logic
Overcoming of problem is multi-layers network
7With an one-input two-layers Perceptron we have
close separable area with catting out boundary
u2(1)w2(1)x b
u1(1)
w1(1)
u1(1)w1(1)x b
w1(2)
u (2)
y
x
w2(2)
w2(1)
u2(1)
Close separable boundary at 1-Dimension
space of x
where
8With a two-inputs two-layers Perceptron net we
can realize a convex separable surface
Layer 2
Layer1
Convex separable boundary at 2-Dimension space
9With a three-layers two-inputs Perceptron net we
can realize a non-convex separable surface
Complex concave separable surface
where
10 The aim of learning is to minimize the
instantaneous squared error of the output signal.
Solution of Task
where
Widrow-Hoff learning rule (delta rule on
base of state error minimization
Rosenblatts learning rule on base of quantized
error minimization
Modified Rosenblatts learning rule on base of
non-quantized error minimization
11The first original perceptron learning rule for
adjusting the weights was developed by Rosenblatt.
We determine the cost function via quantized
error e
Then weights change value is
where
or
- is a vector of quantized error with element ej.
12Applying the algebraic transformation
We get the final equation
We determine the modified cost function via
quantized error e
13 Learning of a SLP illustrates a supervised
learning rule which aims to assign the input
patterns x1, x2, ,xp to one of the
prespecified classes or categories with desired
response if perceptron outputs for every classes
we know in advanced the desired response.
14The Rosenblatts learning rule realises the
weights change value as
15- Marvin L., Minsly, S.Pafert Perceptrons
Expanded Edition Introduction to Computation
Geometry. - Haykin S. Neural Networks A Comprehensive
Foundation. Mac-Millan. N.York.1994. - Laurence Fausett Fundamentals of Neural
Networks Architecture, Algorithms, and
Applications . Prentice Hall, 1994. - Cichocki A. Unbehauen R. Neural Networks for
Optimization and Signal Processing, Wiley, 1993. -