Model Selection Techniques for Repeated Measures Covariance Structures - PowerPoint PPT Presentation

1 / 54
About This Presentation
Title:

Model Selection Techniques for Repeated Measures Covariance Structures

Description:

Used in repeated measures to model the dependencies among observations taken on the same unit. ... of the 6 time vectors for the ARH(1) data fit with a ... – PowerPoint PPT presentation

Number of Views:338
Avg rating:3.0/5.0
Slides: 55
Provided by: ebarrymose
Category:

less

Transcript and Presenter's Notes

Title: Model Selection Techniques for Repeated Measures Covariance Structures


1
Model Selection Techniques for Repeated Measures
Covariance Structures
  • E. Barry MoserLouisiana State University
  • Raúl E. Macchiavelli
  • University of Puerto Rico

2

Outline
  • Introduction
  • Covariance and Inverse Covariance Matrices and
    Structures
  • Likelihood and Likelihood Ratio Test
    Decomposition
  • Graphical Fingerprints
  • Examples

3
Covariance Structures
  • Used in repeated measures to model the
    dependencies among observations taken on the same
    unit.
  • The choice of structure affects both power and
    validity of test procedures about the mean.
  • In general, if the structure is simpler than
    necessary, tests may be invalid.
  • On the other hand, unnecessarily complex
    structures may decrease the power.

4
Consider n independent normal observations, each
of them representing T repeated measures
Define
Assume a linear model
5
We are usually interested in making inferences
about ?, while ? contains the nuisance
parameters.
Define the concentration matrix as ??-1
We can see that the concentration matrix is in
the canonical parameterization for the
likelihood
6
Compound Symmetry
  • Equi-correlated times

7
Structures with zeroes in
  • First order autoregressive, AR(1)

8
Structures with zeroes in
  • First order heteroscedastic autoregressive,
    ARH(1), same as AR(1) but
  • First order antedependence, ANTE(1)

9
Structures with zeroes in
  • Banded Toeplitz, TOEP(q)
  • Heteroscedastic banded Toeplitz, TOEPH(q)
  • Banded general covariance, UN(q)

10
Importance of Zeros
  • Structural zeroes -gt covariance matrix
  • Marginal independence
  • Structural zeroes -gt concentration matrix
  • Conditional independence

11
Making inferences about the mean (Diggle, 1988)
  • Fit a mean model (overfit is preferable to
    underfit avoid spurious correlations).
  • Obtain initial covariance structures
  • Relevant theory
  • Graphics
  • Choose an appropriate structure using formal
    statistical techniques.
  • Make inferences about the mean.

12
Methods for choosing covariances
  • Sequence of likelihood ratio tests
  • Penalized likelihood criteria
  • Graphical models
  • Methods used in linear structural equations
    (Jöreskögs LISREL)
  • Goodness of fit indices
  • Covariance residuals

13
Likelihood ratio test
  • Consider a likelihood ratio test comparing a
    proposed structure with the general unstructured
    alternative

where
and
are residuals of structured and unstructured
models, respectively.
14
  • Applying the spectral decomposition to
  • the LRT we can write

and
where
is a diagonal matrix of diagonal elements
of
and
and
are the jth eigen-
value and vector, respectively, of
the correlation matrix.
15
This results in the decomposition of the LRT as
where
16
Decomposition emphasis
  • Provides contributions to the LRT according to
    each subject
  • Provides contributions to the LRT according to
    the components of time
  • Since the LRT has an asymptotic expected value of
    df (difference in number of parameters), divide
    each component dij by df to spot large individual
    contributions, or sum up by subject or dimension.

17
Alternative decomposition of the LRT
Let
then
which can be further decomposed into
where
and
are variances from
and
respectively, and
and
are eigenvalues from
and
18
Decomposition emphasis
  • Variances at each time are directly compared
  • Correlation structures are compared by comparing
    the variances of the principal components of the
    correlation matrix estimates
  • Contributions by the residuals are kept separate

19
Graphical fingerprints
  • Principal components of several of the
    correlation structures have characteristic
    patterns
  • Consider Compound Symmetry

20
Graphical fingerprints
  • AR(1) - eigenvalues

21
Graphical fingerprints
  • AR(1) - eigenvectors

22
Graphical fingerprints
  • Plot the time vectors in the space of the
    principal components from the UN correlation
    matrix and examine the resulting pattern.
  • Rotate the UN correlation pattern to best match
    (Procrustes rotation) the fitted structured
    correlation model and overlay the vector plots.

23
AR(1) fingerprint
24
AR(1) fingerprint
25
Examples
  • Simulated CS structure
  • Simulated ARH(1) structure
  • Pasture forage experiment

26
Simulated CS structure
Likelihood ratio test (df19) decomposition for
the compound symmetry data fitted with a compound
symmetry model.
 
27
Simulated CS structure
28
Simulated CS structure
29
Simulated ARH(1) structure
Likelihood ratio test (df19) decomposition for
the ARH(1) data fitted with a homogeneous
compound symmetry (CS) model .
30
Simulated ARH(1) structure
Likelihood ratio test (df14) decomposition for
the ARH(1) data fitted with a heterogeneous
compound symmetry (CSH) model .
31
Simulated ARH(1) structure
Likelihood ratio test (df19) decomposition for
the ARH(1) data fitted with a homogeneous
first-order autoregressive (AR(1)) model.
32
Simulated ARH(1) structure
Likelihood ratio test (df14) decomposition for
the ARH(1) data fitted with a heterogeneous
first-order autoregressive (ARH(1)) model.
33
Endpoints of the 6 time vectors for the ARH(1)
data fit with a heterogeneous compound symmetry
(CSH) model plotted in the space of the first two
principal axes.
34
Endpoints of the 6 time vectors for the ARH(1)
data fit with a homogeneous first-order
autoregressive (AR(1)) model plotted in the space
of the first two principal axes.
35
Endpoints of the 6 time vectors for the ARH(1)
data fit with a homogeneous first-order
autoregressive (AR(1)) model plotted in the space
of the last two principal axes.
36
Pasture forage experiment
  • Comparison of minimum and non-tillage with
    conventional methods of signal-grass pasture
    establishment
  • Randomized complete block design with 3
    replicates of 5 treatments
  • 5 repeated measurements of average plot coverage
    taken at monthly intervals

37
Pasture forage experiment
  • Treatments
  • Minimum tillage then seeded
  • Minimum tillageherbicideminimum tillage again
    at 45 days then seeded
  • Minimum tillageherbicidedisking at 45 days then
    seeded
  • Non-tillageherbicide then seeded at 45 days
  • Conventional planting using disking and harrowing
    then seeded

38
Proc MIXED Code - UN
  • Proc Mixed DataForage MethodML
  • Class Block Treatment Time
  • Model Coverage Block Treatment Time /
    OutPredResidUN
  • Repeated Time / SubjectPlot TypeUN R RCorr
  • Run

39
Unstructured Correlation Matrix
40
Inverse of UnstructuredCorrelation Matrix
41
Proc MIXED Code ARH(1)
  • Proc Mixed DataForage MethodML
  • Class Block Treatment Time
  • Model Coverage Block Treatment Time /
    OutPredResidARH1
  • Repeated Time / SubjectPlot TypeARH(1) R
    RCorr
  • Run

42
Likelihood ratio test (df9) decomposition for
the pasture forage data fitted with a
heterogeneous first-order autoregressive (ARH(1))
model.
43
Likelihood ratio test (df9) decomposition of
variances (V) for the pasture forage data fitted
with a heterogeneous first-order autoregressive
(ARH(1)) model.
44
Likelihood ratio test (df9) decomposition of
eigenvalues (l) of correlation matrix for the
pasture forage data fitted with a heterogeneous
first-order autoregressive (ARH(1)) model.
45
Endpoints of the 6 time vectors for the pasture
forage data fit with a heterogeneous first-order
autoregressive (ARH(1)) model plotted in the
space of the first two principal axes.
46
Likelihood ratio test (df3) decomposition for
the pasture forage data fitted with a two-factor
factor analytic (FA(2)) model.
47
Likelihood ratio test (df3) decomposition of
variances (V) for the pasture forage data fitted
with a two-factor factor analytic (FA(2)) model.
48
Likelihood ratio test (df3) decomposition of
eigenvalues (l) of correlation matrix for the
pasture forage data fitted with a two-factor
factor analytic (FA(2)) model.
49
Endpoints of the 6 time vectors for the pasture
forage data fit with a two-factor factor analytic
(FA(2)) model plotted in the space of the first
two principal axes.
50
(No Transcript)
51
Conclusions
  • Decomposition of the LRT focuses attention
    separately upon the variance structure,
    correlation structure, and residuals.
  • Graphical fingerprints yield a visual measure of
    fit indicating problems with the fitted
    correlation structure, and may help identify
    outlier time points.

52
Future
  • Missing data likelihood
  • Subject and time-point outliers/diagnostics
  • Fixed-effect tests
  • Mixed models
  • Categorical repeated measures data models

53
Selected references
  • Diggle, P. (1988) An approach to the analysis of
    repeated measurements. Biometrics 44, 959-71.
  • Wolfinger, R. (1996) Heterogeneous
    variance-covariance structures for repeated
    measures. Journal of Agricultural, Biological,
    and Environmental Statistics 1, 205-230.

54
Selected references (cont)
  • Macchiavelli, R. and E.B. Moser (1997) Analysis
    of repeated measures with ante-dependence models.
    Biometrical Journal 39 (3), 339-350.
Write a Comment
User Comments (0)
About PowerShow.com