Nonlinear Empirical Models - PowerPoint PPT Presentation

1 / 33
About This Presentation
Title:

Nonlinear Empirical Models

Description:

structure motivated by physiological structure of brain ... threshold behaviour - e.g., sigmoid function, ordinary differential equation ... – PowerPoint PPT presentation

Number of Views:109
Avg rating:3.0/5.0
Slides: 34
Provided by: jamesmc9
Category:

less

Transcript and Presenter's Notes

Title: Nonlinear Empirical Models


1
Nonlinear Empirical Models
2
Neural Network Models of Process Behaviour
  • generally modeling input-output behaviour
  • empirical models - no attempt to model physical
    structure
  • estimated from plant data

3
Neural Networks...
  • structure motivated by physiological structure of
    brain
  • individual nodes or cells - neurons -sometimes
    called perceptrons
  • neuron characteristics - notion of firing or
    threshold behaviour

4
Stages of Neural Network Model Development
  • data collection - training set, validation set
  • specification / initialization - structure of
    network, initial values
  • learning or training - estimation of parameters
  • validation - ability to predict new data set
    collected under same conditions

5
Data Collection
  • expected range and point of operation
  • size of input perturbation signal
  • type of input perturbation signal - random input
    sequence?- number of levels (two or more?)
  • validation data set

6
Model Structure
  • numbers and types of nodes
  • input, hidden, output
  • depends on type of neural network- e.g.,
    Feedforward Neural Network- e.g., Recurrent
    Neural Network
  • types of neuron functions - threshold behaviour -
    e.g., sigmoid function, ordinary differential
    equation

7
Learning (Training)
  • estimation of network parameters - weights,
    thresholds and bias terms
  • nonlinear optimization problem
  • objective function - typically sum of squares of
    output prediction error
  • optimization algorithm - gradient-based method or
    variation

8
Validation
  • use estimated NN model to predict outputs for new
    data set
  • if prediction unacceptable, re-train NN model
    with modifications - e.g., number of neurons
  • diagnostics - sum of squares of prediction
    error- R2 - coefficient of determination

9
Feedforward Neural Networks
  • signals flow forward from input through hidden
    nodes to output- no internal feedback
  • input nodes - receive external inputs (e.g.,
    controls) and scale to 0,1 range
  • hidden nodes - collect weighted sums of inputs
    from other nodes and act on the sum with a
    nonlinear function

10
Feedforward Neural Networks (FNN)
  • output nodes - similar to hidden nodes BUT they
    produce signals leaving the network (outputs)
  • FNN has one input layer, one output layer, and
    can have many hidden layers

11
FNN - Neuron Model
  • ith neuron in layer l1

threshold value
weight
state of neuron
activation function
12
FNN parameters
  • weights wl1ij - weight on output from jth
    neuron in layer l entering neuron i in layer l1
  • threshold - determines value of function when
    inputs to neuron are zero
  • bias - provision for additional constants to be
    added

13
FNN Activation Function
  • typically sigmoidal function

14
FNN Structure
hidden layer
input layer
output layer
15
Mathematical Basis
  • approximation of functions
  • e.g., Cybenko, 1989 - J. of Mathematics of
    Control, Signals and Systems
  • approximation to arbitrary degree given
    sufficiently large number of nodes - sigmoidal

16
Training FNNs
  • calculate sum of squares of output prediction
    error
  • take current iterates of parameters, calculate
    forward and calculate E
  • update estimates of weights working backwards -
    backpropagation

17
Estimation
  • typically using a gradient-based optimization
    method
  • make adjustments proportional to
  • issues - highly over-parameterized models -
    potential for singularity
  • e.g., Levenberg-Marquardt algo.

18
How to use FNN for modeling dynamic behaviour?
  • structure of FNN suggests static model
  • model dynamic model as nonlinear difference
    equation
  • essentially a NARMAX model

19
Linear discrete time transfer function
  • transfer function
  • equivalent difference equation

20
FNN Structure - 1st order linear example
hidden layer
input layer
output layer
yk
yk1
uk
uk-1
21
FNN model for 1st order linear example
  • essentially modelling algebraic relationship
    between past and present inputs and outputs
  • nonlinear activation function not required
  • weights required - correspond to coefficients in
    discrete transfer function

22
Applications of FNNs
  • process modeling - bioreactors, pulp and paper,
  • nonlinear control
  • data reconciliation
  • fault detection
  • some industrial applications - many academic
    (simulation) studies

23
Typical dimensions
  • Dayal et al., 1994 - 3-state jacketted CSTR as a
    basis
  • 700 data points in training set
  • 6 inputs, 1 hidden layer with 6 nodes, 1 output

24
Advantages of Neural Net Models
  • limited process knowledge required - but be
    careful (e.g., Dayal et al. paper)
  • flexible - can model difficult relationships
    directly (e.g., inverse of a nonlinear control
    problem)

25
Disadvantages
  • potential for large computational requirements -
    implications for real-time application
  • highly over-parameterized
  • limited insight into process structure
  • amount of data required
  • limited to range of data collection

26
Recurrent Neural Networks
  • neurons contain differential equation model - 1st
    order linear nonlinearity
  • contain feedback and feedforward components
  • can represent continuous dynamics
  • e.g., You and Nikolaou, 1993

27
Nonlinear Empirical Model Representations
  • Volterra Series (continuous and discrete)
  • Nonlinear Auto-Regressive Moving Average with
    Exogenous Inputs (NARMAX)
  • Cascade Models

28
Volterra Series Models
  • higher-order convolution models
  • continuous

29
Volterra Series Model
  • discrete time

30
Volterra Series models...
  • can be estimated directly from data or derived
    from state space models
  • causality - limits of sum or integration
  • functions hi - referred to as the ith order
    kernel
  • applications - typically second-order(e.g.,
    Pearson et al., 1994 - binder)

31
NARMAX models
  • nonlinear difference equation models
  • typical form
  • dependence on lagged ys - autoregressive
  • dependence on lagged us - moving average

32
NARMAX examples
  • with products, cross-products
  • 2nd order Volterra model
  • as NARMAX model in u only, with second order
    terms

33
Nonlinear Cascade Models
  • made from serial and parallel arrangements of
    static nonlinear and linear dynamic elements
  • e.g., 1st order linear dynamic element fed into a
    squaring element
  • obtain products of lagged inputs
  • cf. second order Volterra term
Write a Comment
User Comments (0)
About PowerShow.com