Neural%20Networks - PowerPoint PPT Presentation

About This Presentation
Title:

Neural%20Networks

Description:

Neural Networks Multi-stage regression/classification model output function PPR hidden layer bias unit synaptic weight activation function also known as ridge ... – PowerPoint PPT presentation

Number of Views:223
Avg rating:3.0/5.0
Slides: 15
Provided by: acil150
Category:

less

Transcript and Presenter's Notes

Title: Neural%20Networks


1
Neural Networks
  • Multi-stage regression/classification model

output function
PPR
hidden layer
bias unit
synaptic weight
activation function
also known as ridge functions in PPR
PPR
2
Activation Function
  • Gaussian radial basis ? radial basis function
    network
  • Sigmoid
  • differentiable
  • almost linear around 0

large s
small s
3
Output Function
  • Regression
  • Classification
  • Weihao Any other justification to make output
    layer sum-to-one using, say, softmax function in
    Eq. 11.6?
  • Answer Think of NN as a function approximator.

4
Fitting NN (contd)
  • Gradient descent (for regression)

5
Back-propagation(aka delta rule)
  • Forward pass

given/computed values
6
Back-propagation (contd)
  • Backward pass

given/computed values
7
Back-propagation (contd)
  • Error surface and learning rate
  • ?opt(T) optimal learning rate at weights T
  • Assuming quadratic error surface
  • will NOT converge

8
Back-propagation (contd)
  • Batch learning vs. online learning
  • Often too slow
  • Newton method not attractive (2nd derivative too
    costly)
  • Use conjugate gradients, variable metric methods,
    etc. (Ch. 10, Numerical Recipes in C
    http//www.library.cornell.edu/nr/bookcpdf.html)

9
Back-propagation (contd)
regularization!
  • Prevent over-fitting
  • Start from zero weights
  • Introducing non-linearity when necessary
  • Early stopping
  • Smaller/adaptive learning rates
  • Convergence guaranteed if

10
Back-propagation (contd)
  • Joy since all parameters for starting are
    close to 0, how could different starting points
    ended in models differ that much?
  • Answer non-linearity.
  • Joy To prevent fitting is there any way to
    train the model to the global minimum point and
    then "prune" it?
  • Answer Global minimum is elusive. But the people
    have tried the idea of pruning, in weight decay
    (later) and optimal brain damageY. LeCun, J.
    S. Denker, and S. A. Solla. Optimal brain damage.
    In D. S. Touretzky, editor, Advances in Neural
    Information Processing Systems II, pages
    598--605. Morgan Kaufmann, San Mateo, CA, 1990.
    Less salient connections can be removed
    (pruned) saliency ? magnitude!

11
Historical Background
  • McCulloch Pitts, 1941 behavior of simple
    neural networks.
  • A. Turing, 1948 B-type unorganized machine
    consisting of networks of NAND gates.
  • Rosenblatt, 1958 two-layer perceptrons.
  • Minsky Papert, 1969 (Perceptrons) showing XOR
    problems for perceptrons connectionism winter
    came.
  • Rumelhart, Hinton Williams, 1986 first
    well-known introduction of back-propagation
    algorithm connectionism revived.

12
Model Complexity of NN
  • Fan How can we measure the complexity of a NN?
    If one NN has many layers but few nodes, another
    has many nodes but few layers, which one is more
    complex? Kevyn how can we derive the
    effective of degrees of freedom based on the
    number of iterations we have performed?
  • Answer remember in Ch. 5 we haveand we define
    - can we do the same for
    NN?For regression, andbut we have
    non-linearity inside B!Maybe use Taylor
    expansion of sigmoid and proceed? (so were
    basically approximating NN using linear models)

13
NN Universal Approximator?
  • Kolmogorov proved any continuous function g(x)
    defined on the unit hypercube In can be
    represented as for properly chosen and
    .(A. N. Kolmogorov. On the representation of
    continuous functions of several variables by
    superposition of continuous functions of one
    variable and addition. Doklady Akademiia Nauk
    SSSR, 114(5)953-956, 1957)

14
NN vs. PPR
  • NN parametric version of PPR
  • less complex s implies more terms (20-100 vs.
    5-10)
Write a Comment
User Comments (0)
About PowerShow.com