Time Series Models - PowerPoint PPT Presentation

1 / 76
About This Presentation
Title:

Time Series Models

Description:

Models Topics Stochastic processes Stationarity White noise Random walk Moving average processes Autoregressive processes More general processes Stochastic Processes ... – PowerPoint PPT presentation

Number of Views:56
Avg rating:3.0/5.0
Slides: 77
Provided by: econKuleu1
Category:

less

Transcript and Presenter's Notes

Title: Time Series Models


1
Time SeriesModels
2
Topics
  • Stochastic processes
  • Stationarity
  • White noise
  • Random walk
  • Moving average processes
  • Autoregressive processes
  • More general processes

3
Stochastic Processes
1
4
Stochastic processes
  • Time series are an example of a stochastic or
    random process
  • A stochastic process is 'a statistical phenomenen
    that evolves in timeaccording to probabilistic
    laws'
  • Mathematically, a stochastic process is an
    indexed collection of random variables

5
Stochastic processes
  • We are concerned only with processes indexed by
    time, either discrete time or continuous time
    processes such as

6
Inference
  • We base our inference usually on a single
    observation or realization of the process over
    some period of time, say 0, T (a continuous
    interval of time) or at a sequence of time points
    0, 1, 2, . . . T

7
Specification of a process
  • To describe a stochastic process fully, we must
    specify the all finite dimensional distributions,
    i.e. the joint distribution of of the random
    variables for any finite set of times

8
Specification of a process
  • A simpler approach is to only specify the
    momentsthis is sufficient if all the joint
    distributions are normal
  • The mean and variance functions are given by

9
Autocovariance
  • Because the random variables comprising the
    process are not independent, we must also specify
    their covariance

10
Stationarity
2
11
Stationarity
  • Inference is most easy, when a process is
    stationaryits distribution does not change over
    time
  • This is strict stationarity
  • A process is weakly stationary if its mean and
    autocovariance functions do not change over time

12
Weak stationarity
  • The autocovariance depends only on the time
    difference or lag between the two time points
    involved

13
Autocorrelation
  • It is useful to standardize the autocovariance
    function (acvf)
  • Consider stationary case only
  • Use the autocorrelation function (acf)

14
Autocorrelation
  • More than one process can have the same acf
  • Properties are

15
White Noise
3
16
White noise
  • This is a purely random process, a sequence of
    independent and identically distributed random
    variables
  • Has constant mean and variance
  • Also

17
Random Walk
3
18
Random walk
  • Start with Zt being white noise or purely
    random
  • Xt is a random walk if

19
Random walk
  • The random walk is not stationary
  • First differences are stationary

20
Moving Average Processes
4
21
Moving average processes
  • Start with Zt being white noise or purely
    random, mean zero, s.d. ??Z
  • Xt is a moving average process of order q
    (written MA(q)) if for some constants ?0, ?1, . .
    . ?q we have
  • Usually ?0 1

22
Moving average processes
  • The mean and variance are given by
  • The process is weakly stationary because the mean
    is constant and the covariance does not depend on
    t

23
Moving average processes
  • If the Zt's are normal then so is the process,
    and it is then strictly stationary
  • The autocorrelation is

24
Moving average processes
  • Note the autocorrelation cuts off at lag q
  • For the MA(1) process with ?0 1

25
Moving average processes
  • In order to ensure there is a unique MA process
    for a given acf, we impose the condition of
    invertibility
  • This ensures that when the process is written in
    series form, the series converges
  • For the MA(1) process Xt Zt ?Zt - 1, the
    condition is ?lt 1

26
Moving average processes
  • For general processes introduce the backward
    shift operator B
  • Then the MA(q) process is given by

27
Moving average processes
  • The general condition for invertibility is that
    all the roots of the equation ???????? lie
    outside the unit circle (have modulus less than
    one)

28
Autoregressive Processes
4
29
Autoregressive processes
  • Assume Zt is purely random with mean zero and
    s.d. ?z
  • Then the autoregressive process of order p or
    AR(p) process is

30
Autoregressive processes
  • The first order autoregression is
  • Xt ?Xt - 1 Zt
  • Provided ?lt1 it may be written as an infinite
    order MA process
  • Using the backshift operator we have
  • (1 ?B)Xt Zt

31
Autoregressive processes
  • From the previous equation we have

32
Autoregressive processes
  • Then E(Xt) 0, and if ?lt1

33
Autoregressive processes
  • The AR(p) process can be written as

34
Autoregressive processes
  • This is for
  • for some ?1, ?2, . . .
  • This gives Xt as an infinite MA process, so it
    has mean zero

35
Autoregressive processes
  • Conditions are needed to ensure that various
    series converge, and hence that the variance
    exists, and the autocovariance can be defined
  • Essentially these are requirements that the ?i
    become small quickly enough, for large i

36
Autoregressive processes
  • The ?i may not be able to be found however.
  • The alternative is to work with the ?i
  • The acf is expressible in terms of the roots ?i,
    i1,2, ...p of the auxiliary equation

37
Autoregressive processes
  • Then a necessary and sufficient condition for
    stationarity is that for every i, ?ilt1
  • An equivalent way of expressing this is that the
    roots of the equation
  • must lie outside the unit circle

38
ARMA processes
  • Combine AR and MA processes
  • An ARMA process of order (p,q) is given by

39
ARMA processes
  • Alternative expressions are possible using the
    backshift operator

40
ARMA processes
  • An ARMA process can be written in pure MA or pure
    AR forms, the operators being possibly of
    infinite order
  • Usually the mixed form requires fewer parameters

41
ARIMA processes
  • General autoregressive integrated moving average
    processes are called ARIMA processes
  • When differenced say d times, the process is an
    ARMA process
  • Call the differenced process Wt. Then Wt is an
    ARMA process and

42
ARIMA processes
  • Alternatively specify the process as
  • This is an ARIMA process of order (p,d,q)

43
ARIMA processes
  • The model for Xt is non-stationary because the AR
    operator on the left hand side has d roots on the
    unit circle
  • d is often 1
  • Random walk is ARIMA(0,1,0)
  • Can include seasonal termssee later

44
Non-zero mean
  • We have assumed that the mean is zero in the
    ARIMA models
  • There are two alternatives
  • mean correct all the Wt terms in the model
  • incorporate a constant term in the model

45
The Box-JenkinsApproach
46
Topics
  • Outline of the approach
  • Sample autocorrelation partial autocorrelation
  • Fitting ARIMA models
  • Diagnostic checking
  • Example
  • Further ideas

47
Outline of theBox-Jenkins Approach
1
48
Box-Jenkins approach
  • The approach is an iterative one involving
  • model identification
  • model fitting
  • model checking
  • If the model checking reveals that there are
    problems, the process is repeated

49
Models
  • Models to be fitted are from the ARIMA class of
    models (or SARIMA class if the data are seasonal)
  • The major tools in the identification process are
    the (sample) autocorrelation function and partial
    autocorrelation function

50
Autocorrelation
  • Use the sample autocovariance and sample variance
    to estimate the autocorrelation
  • The obvious estimator of the autocovariance is

51
Autocovariances
  • The sample autocovariances are not unbiased
    estimates of the autocovariancesbias is of order
    1/N
  • Sample autocovariances are correlated, so may
    display smooth ripples at long lags which are not
    in the actual autocovariances

52
Autocovariances
  • Can use a different divisor (N-k instead of N) to
    decrease biasbut may increase mean square error
  • Can use jacknifing to reduce bias (to order 1/N
    )divide the sample in half and estimate using
    the whole and both halves

2
53
Autocorrelation
  • More difficult to obtain properties of sample
    autocorrelation
  • Generally still biased
  • When process is white noise
  • E(rk) ??1/N
  • Var( rk) ??1/N
  • Correlations are normal for N large

54
Autocorrelation
  • Gives a rough test of whether an autocorrelation
    is non-zero
  • If rkgt2/?(N) suspect the autocorrelation at
    that lag is non-zero
  • Note that when examining many autocorrelations
    the chance of falsly identifying a non-zero one
    increases
  • Consider physical interpretation

55
Partial autocorrelation
  • Broadly speaking the partial autocorrelation is
    the correlation between Xt and Xtk with the
    effect of the intervening variables removed
  • Sample partial autocorrelations are found from
    sample autocorrelations by solving a set of
    equations known as the Yule-Walker equations

56
Model identification
  • Plot the autocorrelations and partial
    autocorrelations for the series
  • Use these to try and identify an appropriate
    model
  • Consider stationary series first

57
Stationary series
  • For a MA(q) process the autocorrelation is zero
    at lags greater than q, partial autocorrelations
    tail off in exponential fashion
  • For an AR(p) process the partial autocorrelation
    is zero at lags greater than p, autocorrelations
    tail off in exponential fashion

58
Stationary series
  • For mixed ARMA processes, both the acf and pacf
    will have large values up to q and p
    respectively, then tail off in an exponential
    fashion
  • See graphs in MW, pp. 136137
  • Try fitting a model and examine the residuals is
    the approach used

59
Non-stationary series
  • The existence of non-stationarity is indicated by
    an acf which is large at long lags
  • Induce stationarity by differencing
  • Differencing once is generally sufficient, twice
    may be needed
  • Overdifferencing introduces autocorrelation
    should be avoided

60
Estimation
2
61
Estimation
  • We will always fit the model using Minitab
  • AR models may be fitted by least squares, or by
    solving the Yule-Walker equations
  • MA models require an iterative procedure
  • ARMA models are like MA models

62
Minitab
  • Minitab uses an iterative least squares approach
    to fitting ARMA models
  • Standard errors can be calculated for the
    parameter estimates so confidence intervals and
    tests of significance can be carried out

63
Diagnostic Checking
3
64
Diagnostic checking
  • Based on residuals
  • Residuals should be Normally distributed have
    zero mean, by uncorrelated, and should have
    minimum variance or dispersion

65
Procedures
  • Plot residuals against time
  • Draw histogram
  • Obtain normal scores plot
  • Plot acf and pacf of residuals
  • Plot residuals against fitted values
  • Note that residuals are not uncorrelated, but are
    approximately so at long lags

66
Procedures
  • Portmanteau test
  • Overfitting

67
Portmanteau test
  • Box and Peirce proposed a statistic which tests
    the magnitudes of the residual autocorrelations
    as a group
  • Their test was to compare Q below with the
    Chi-Square with K p q d.f. when fitting an
    ARMA(p, q) model

68
Portmanteau test
  • Box Ljung discovered that the test was not good
    unless n was very large
  • Instead use modified Box-Pierce or
    Ljung-Box-Pierce statisticreject model if Q is
    too large

69
Overfitting
  • Suppose we think an AR(2) model is appropriate.
    We fit an AR(3) model.
  • The estimate of the additional parameter should
    not be significantly different from zero
  • The other parameters should not change much
  • This is an example of overfitting

70
Further Ideas
4
71
Other identification tools
  • Chatfield(1979), JRSS A among others has
    suggested the use of the inverse autocorrelation
    to assist with identification of a suitable model
  • Abraham Ledolter (1984) Biometrika show that
    although this cuts off after lag p for the AR(p)
    model it is less effective than the partial
    autocorrelation for detecting the AR order

72
AIC
  • The Akaike Information Criterion is a function of
    the maximum likelihood plus twice the number of
    parameters
  • The number of parameters in the formula penalizes
    models with too many parameters

73
Parsimony
  • Once principal generally accepted is that models
    should be parsimonioushaving as few parameters
    as possible
  • Note that any ARMA model can be represented as a
    pure AR or pure MA model, but the number of
    parameters may be infinite

74
Parsimony
  • AR models are easier to fit so there is a
    temptation to fit a less parsimonious AR model
    when a mixed ARMA model is appropriate
  • Ledolter Abraham (1981) Technometrics show
    that fitting unnecessary extra parameters, or an
    AR model when a MA model is appropriate, results
    in loss of forecast accuracy

75
Exponential smoothing
  • Most exponential smoothing techniques are
    equivalent to fitting an ARIMA model of some sort
  • Winters' multiplicative seasonal smoothing has no
    ARIMA equivalent
  • Winters' additive seasonal smoothing has a very
    non-parsimonious ARIMA equivalent

76
Exponential smoothing
  • For example simple exponential smoothing is the
    optimal method of fitting the ARIMA (0, 1, 1)
    process
  • Optimality is obtained by taking the smoothing
    parameter ?? to be 1 ? when the model is
Write a Comment
User Comments (0)
About PowerShow.com