Neural Network Toolbox - PowerPoint PPT Presentation

1 / 20
About This Presentation
Title:

Neural Network Toolbox

Description:

Martin T. Hagan, Howard B. Demuth & Mark Beale, 1996, Neural Network Design, ... This can be hard-limit, linear, threshold linear, log-sigmoid, or various other. ... – PowerPoint PPT presentation

Number of Views:826
Avg rating:3.0/5.0
Slides: 21
Provided by: harry8
Category:

less

Transcript and Presenter's Notes

Title: Neural Network Toolbox


1
Neural Network Toolbox
  • COMM2M
  • Harry R. Erwin, PhD
  • University of Sunderland

2
Resources
  • Martin T. Hagan, Howard B. Demuth Mark Beale,
    1996, Neural Network Design, Martin/Hagan
    (Distributed by the University of Colorado)
  • http//www.mathworks.com/access/helpdesk/help/pdf_
    doc/nnet/nnet.pdf
  • This can be found in the COMM2M Lectures folder
    as NNET.PDF.

3
Neural Networks in MATLAB
  • The MATLAB neural network toolkit does not model
    biological neural networks.
  • Instead, it models simple abstractions
    corresponding to biological models, typically
    trained using some sort of supervised learning,
    although unsupervised learning and direct design
    are also supported.

4
Notation
  • It helps if you have some understanding of
    mathematical notation and systems analysis.
  • Synaptic weights are typically represented in
    matrices wi,j. Sparse matrices (I.e., mostly
    zero) are the most biologically realistic.
  • Biases are used to control spiking probabilities
    or rates relative to some nominal monotonic
    function of membrane potential at the soma (cell
    body).

5
How Does This Relate to Biological Neural
Networks?
  • The inputs correspond to action potentials (or AP
    rates) received by the dendritic tree.
  • The weights correspond to
  • Conductance density in the post-synaptic membrane
  • Signal strength reduction between the synapse and
    the cell soma
  • The output corresponds to action potentials or
    spiking rates at the axonal hillock in the soma.
  • The neurons are phasic, not tonic.

6
Assumptions
  • Linearity is important. The membrane potential at
    the soma is a weighted linear sum of the
    activations at the various synapses.
  • The weightings reflect both synaptic conductances
    and the transmission loss between the synapse and
    the soma.
  • Time is usually quantized.

7
Topics Covered in the Users Guide
  • Neuron models.
  • Perceptrons
  • Linear filters
  • Back-propagation
  • Control systems
  • Radial basis networks
  • Self-organizing and LVQ function networks
  • Recurrent networks
  • Adaptive filters

8
Neuron models.
  • Scalar input with bias. Membrane potential at the
    soma is the scalar input plus the bias.
  • Output is computed by applying a monotonic
    transfer function to the scalar input. This can
    be hard-limit, linear, threshold linear,
    log-sigmoid, or various other.
  • A layer is a layer of transfer functions. This
    often corresponds to a layer of cells, but local
    non-linearities can create multi-layer cells.

9
Network Architectures
  • Neural network architectures usually consist of
    multiple layers of cells.
  • A common architecture consists of three layers
    (input, hidden, and output).
  • This has at least a notional correspondence to
    how neocortex is organized in your brain.
  • Dynamics of these networks can be analyzed
    mathematically.

10
Perceptrons
  • Perceptron neurons perform hard limited (hardlim)
    transformations on linear combinations of their
    inputs.
  • The hardlim transformation means that a
    perceptron classifies vector inputs into two
    subsets separated by a plane (linearly
    separable). The bias moves the plane away from
    the origin.
  • Smooth transformations can be used.
  • A perceptron architecture is a single layer of
    neurons.

11
Learning Rules
  • A learning rule is a procedure for modifying the
    weights of a neural network
  • Based on examples and desired outputs in
    supervised learning,
  • Based on input values only in unsupervised
    learning.
  • Perceptrons are trained using supervised
    learning. Convergence rate of training is good.

12
Linear filters
  • Neural networks similar to perceptrons, but with
    linear transfer functions, are called linear
    filters.
  • Limited to linearly separable problems.
  • Error is minimized using the Widrow-Hoff
    algorithm.
  • Minsky and Papert criticized this approach.

13
Back-propagation
  • Invented by Paul Werbos (now at NSF).
  • Allows multi-layer perceptrons with non-linear
    differentiable transfer functions to be trained.
  • The magic is that errors can be propagated
    backward through the network to control weight
    adjustment.

14
Control systems
  • Neural networks can be used in predictive
    control.
  • Basically, the neural network is trained to
    represent forward dynamics of the plant. Having
    that neural network model allows the system to
    predict the effects of control value changes.
  • Useful in real computer applications.
  • Would be very useful in modeling reinforcement
    learning in biological networks if we could
    identify how forward models are learned and
    stored in the brain.

15
Radial basis networks
  • Perceptron networks work for linearly separable
    data. Suppose the data is locally patchy instead.
    RBF networks were invented for that.
  • The input to the transfer function is the bias
    times the distance between a preferred vector of
    the cell and the input vector.
  • The transfer function is e-nn, for n being the
    input.
  • Two-layer networks. Usually trained by exact
    design or by adding neurons until the error falls
    below a threshold.

16
Self-organizing and LVQ function networks
  • The neurons move around in input space until the
    set of inputs is uniformly covered.
  • They can also be trained in a supervised manner
    (LVQ).
  • Kohonen networks.

17
Recurrent networks
  • Elman and Hopfield networks
  • These model how neocortical principal cells are
    believed to function.
  • Elman networks are two-layer back-propagation
    networks with recurrence. Can learn temporal
    patterns. Whether theyre sufficiently general to
    model how the brain does the same thing is a
    research question.
  • Hopfield networks give you autoassociative memory.

18
Adaptive filters
  • Similar to perceptrons with linear transfer
    functions. Limited to linearly separable
    problems.
  • Powerful learning rule.
  • Used in signal processing and control systems.

19
Conclusions
  • Neural networks are powerful but sophisticated
    (gorilla in a dinner jacket).
  • Theyre also a good deal simpler than
    biologically neural networks.
  • One of the things to do is to learn how to use
    the MATLAB toolbox functions, but another is how
    to extend the toolbox.

20
Tutorial
  • Poirazi, Brannon, and Mel, 2003, Pyramidal
    Neuron as a Two-Layer Neural Network, Neuron,
    37989-999, March 27, 2003, suggests that
    cortical pyramidal cells can be validly modeled
    as two-layer neural networks.
  • Tutorial assignment, investigate that, using some
    variant of back-propagation to train the network
    to recognize digits. Remember the weights of the
    individual branches are constant its only the
    synaptic weights that are trained.
  • A test and training dataset is provided
    (Numbers.ppt).
Write a Comment
User Comments (0)
About PowerShow.com