Title: Messung und statistische Analyse von Kundenzufriedenheit
1Messung und statistische Analyse von
Kundenzufriedenheit
- KF Qualitätsmanagement
- Vertiefungskurs V
2Outline
- Customer satisfaction measurement
- The Structural Equation Model (SEM)
- Estimation of SEMs
- Evaluation of SEMs
- Practice of SEM-Analysis
3The ACSI Model
Ref. http//www.theacsi.org/model.htm
4ACSI-Model Latent Variables
- Customer Expectations combine customers
experiences and information about it via media,
advertising, salespersons, and word-of-mouth from
other customers - Perceived Quality overall quality, reliability,
the extent to which a product/service meets the
customers needs - Customer Satisfaction overall satisfaction,
fulfillment of expectations, comparison with
ideal - Perceived Value overall price given quality and
overall quality given price - Customer Complaints percentage of respondents
who reported a problem - Customer Loyalty likelihood to purchase at
various price points
5 Baseline Q2 1995 Q2 1996 Q2 1997 Q2 1998 Q2 1999 Q2 2000 Q2 2001 Q2 2002 Q2 2003 Q22004 Changes Changes
MANUFACTURING/DURABLES 79.2 79.8 78.8 78.4 77.9 77.3 79.4 78.7 79.0 79.2 78.3 -1.1 -1.1
Personal Computers 78 75 73 70 71 72 74 71 71 72 74 2.8 -5.1
Apple Computer, Inc. 77 75 76 70 69 72 75 73 73 77 81 5.2 5.2
Dell Inc. NM NM NM 72 74 76 80 78 76 78 79 1.3 9.7
Gateway, Inc. NM NM NM NM 76 76 78 73 72 69 74 7.2 -2.6
All Others NM 70 73 72 69 69 68 67 70 69 71 2.9 1.4
Hewlett-Packard Company HP 78 80 77 75 72 74 74 73 71 70 71 1.4 -9.0
Hewlett-Packard Company Compaq 78 77 74 67 72 71 71 69 68 68 69 1.5 -11.5
6The European Customer Satisfaction Index (ECSI)
Ref. http//www.swics.ch/ecsi/index.html
7ACSIe-Model for Food Retail
Emotional Factor
Hackl et al. (2000) Latent variables and path
coefficients
Perceived Quality
0,33
0,35
Custo- mer Satis- faction
0,37
0,36
0,73
(-0,01)
0,34
Loyalty
Expec- tations
0,53
Value
0,34
(0,06)
8Austrian Food Retail Market
- Pilot for an Austrian National CS Index (Zuba,
1997) - Data collection December 1996 by Dr Fessel GfK
(professional market research agency) - 839 interviews, 327 complete observations
- Austria-wide active food retail chains (1996
50 from the 10.5 BEUR market) - Billa well-assorted medium-sized outlets
- Hofer limited range at good prices
- Merkur large-sized supermarkets with
comprehensive range - Meinl top in quality and service
9The Data
Indicators Latent
total expected quality (EGESQ), expected compliance with demands (EANFO), expected shortcomings (EMANG) Expectations (E)
total perceived quality (OGESQ), perceived compliance with needs (OANFO), perceived shortcomings (OMANG) Perceived Quality (Q)
value for price (VAPRI), price for value (PRIVA) Value (P)
total satisfaction (CSTOT), fulfilled expectations (ERWAR), comparison with ideal (IDEAL) Customer Sa-tisfaction (CS)
number of oral complaints (NOBES), number of written complaints (NOBRI) Voice (V)
repurchase probability (WIEDE), tolerance against price-change (PRVER) Loyalty (L)
10The Emotional Factor
- Principal component analysis of satisfaction
drivers - staff (availability, politeness)
- outlet (make-up, presentation of merchandise,
cleanliness) - range (freshness and quality, richness)
- price-value ratio (value for price, price for
value) - customer orientation (access to outlet, shopping
hours, queuing time for checkout, paying modes,
price information, sales, availability of sales) - identifies (Zuba, 1997)
- staff, outlet, range Emotional factor
- price-value ratio Value
- customer orientation Cognitive factor
11Structural Equation Models
- Combine three concepts
- Latent variables
- Pearson (1904), psychometrics
- Factor analysis model
- Path analysis
- Wright (1934), biometrics
- Technique to analyze systems of relations
- Simultaneous regression models
- Econometrics
12Customer Satisfaction
- Is the result of the customers comparison of
- his/her expectations with
- his/her experiences
- has consequences on
- loyalty
- future profits of the supplier
13Expectation vs. Experience
- Expectation reflects
- customers needs
- offer on the market
- image of the supplier
- etc.
- Experiences include
- perceived performance/quality
- subjective assessment
- etc.
14CS-Model Path Diagram
Expecta- tions
Custo- mer Satis- faction
Loyalty
Perceived Quality
15A General CS-Model
Voice
Expecta- tions
Custo- mer Satis- faction
Loyalty
Perceived Quality
Profits
16CS-Model Structure
EX expectation PQ perceived quality CS
customer satisfaction LY loyalty
to from EX PQ CS LY
EX X X 0
PQ 0 X 0
CS 0 0 X
LY 0 0 0
Recursive structure triangular form of
relations
17CS-Model Equations
PQ a1 g11EX z1 CS a2 b21PQ g21EX
z2 LY a3 b32CS z3
Simultaneous equations model in latent
variables Exogenous EX Endogenous PQ, CS,
LY Error terms (noises) z1, z2, z3
18Simple Linear Regression
- Model Y a gX z
- Observations (xi, yi), i1,,n
- Fitted Model Y a cX
- OLS-estimates a, c
- minimize the sum of squared residuals
- sxy sample-covariance of X and Y
19Criteria of Model Fit
- R2 coefficient of determination
- the squared correlation between Y and Y
R2 ryy2 - t-Test Test of H0 g0 against H1g?0
- tc/s.e.(c)
- s.e.(c) standard error of c
- F-Test Test of H0 R20 against H1 R2?0
- follows for large n the F-distribution with n-2
and 2 df
20Multiple Linear Regression
- Model Y a X1g1 ... Xkgk z a xg
z - Observations (xi1,, xik, yi), i1,,n
- In Matrix-Notation y a Xg z
- y, z n-vectors, g k-vector, X nxk-matrix
- Fitted Model y a Xc
- OLS-estimates a, c
-
- R2 ryy2
- F-Test
- t-Test
21Simultaneous Equations Models
- A 2-equations model
- PQ a1 g11EX z1
- CS a2 b21PQ g21EX z2
- In matrix-notation Y BY GX z
- with
path coefficients
22Simultaneous Equations Models
- Model Y BY GX z
- Y, z m-vectors,
- B (mxm)-matrix
- G (mxK)-matrix,
- X K-vector
- Problems
- Simultaneous equation bias OLS-estimates of
coefficients are not consistent - Identifiability Can coefficients be consistently
estimated?
Some assumptions z E(z)0, Cov(z) S
Exogeneity Cov(X,z) 0
23Path Analytic Model
PQ g11EX z1 CS b21PQ g21EX z2
EX
d1
z2
g21
CS
Var(d1) sEX2
g11
PQ
b21
z1
24Path Analysis
- Wright (1921, 1934)
- A multivariate technique
- Model Variables may be
- structurally related
- structurally unrelated, but correlated
- Decomposition of covariances allows to write
covariances as functions of structural parameters
- Definition of direct and indirect effects
25Example
sCS,EX g21s2EX b21sPQ,EX
g21s2EX g11b21s2EX
EX
d1
z2
g21
CS
g11
PQ
b21
with standardized variable EX
rCS,EX g21 g11b21
z1
26Direct and Indirect Effects
- rCS,EX g21 g11b21
- Direct effect coefficient that links independent
with dependent variable e.g., g21 is direct
effect of EX on CS - Indirect effect effect of one variable on
another via one or more intervening variable(s),
e.g., g11b21 - Total indirect effect sum of indirect effects
between two variables - Total effect sum of direct and total indirect
effects between two variables
27Decomposition of Covariance syx
variable on path from X to Y
?YI path coefficient of variable I to Y
28First Law of Path Analysis
- Decomposition of covariance sxy between Y and X
- Assumptions
- Exogenous (X) and endogenous variables (Y) have
mean zero - Errors or noises (z)
- have mean zero and equal variances across
observations - are uncorrelated across observations
- are uncorrelated with exogenous variables
- are uncorrelated across equations
29Identification
- PQ g11EX z1 Y1 g11X z1
- CS b21PQ g21EX z2 Y2 b21Y1 g21X z2
- In matrix-notation Y BY GX z
- Number of parameters p6
- Model is identified, if all parameters can be
expressed as functions of variances/covariances
of observed variables
30Identification, contd
- Y1 g11X z1
- Y2 b21Y1 g21X z2
- s1X g11 sX2
- s2X b21s1X g21sX2
- s21 b21s12 g21s1X
- sX2 sX2
- sy12 g11s1Xs12
- sy22 b21s21 g21s2Xs22
p6 first 3 equations allow unique solution
for path coefficients, last three for variances
of d and z
31Condition for Identification
- Just-identified all parameters can be uniquely
derived from functions of variances/covariances - Over-identified at least one parameter is not
uniquely determined - Under-identified insufficient number of
variances/covariances - Necessary, but not sufficient condition for
identification number of variances/covariances
at least as large as number of parameters - A general and operational rule for checking
identification has not been found
32Latent variables and Indicators
- Latent variables (LVs) or constructs or factors
are unobservable, but - We might find indicators or manifest variables
(MVs) for the LVs that can be used as measures of
the latent variable - Indicators are imperfect measures of the latent
variable
33Indicators for Expectation
From Swedish CSB Questionnaire, Banks Private
Customers
d1
E1
EX
d2
E2
E1, E2, E3 block of LVs for Expectation
d3
E3
E1 When you became a customer of AB-Bank, you
probably knew something about them. How would you
grade your expectations on a scale of 1 (very
low) to 10 (very high)? E2 Now think about the
different services they offer, such as bank
loans, rates, Rate your expectations on a scale
of 1 to 10? E3 Finally rate your overall
expectations on a scale of 1 to 10?
34Notation
d1
l1
X1l1xd1 X2l2xd2 X3l3xd3
X1
x
l2
d2
X2
l3
d3
X3
reflective indicators
- x latent variable, factor
- Xi indicators, manifest
- variables
- li factor loadings
- i measurement errors,
- noise
Some properties LV unit variance noise di
has mean zero, variance si2, uncorrela- ted
with other noises
35Notation
d1
l1
X1l1xd1 X2l2xd2 X3l3xd3 X Lx d
X1
x
l2
d2
X2
l3
d3
X3
In matrix-notation with vectors X, L, and
d e.g., X (X1, X2, X3)
- x latent variable, factor
- Xi indicators, manifest
- variables
- li factor loadings
- i measurement error,
- noise
36CS-Model Path Diagram
d1
E1
EX
d2
z2
E2
e4
g21
d3
C1
CS
E3
e5
g11
C2
e1
e6
Q1
C3
PQ
b21
e2
Q2
e3
Q3
z1
37SEM-Model Path Diagram
d1
X1
x
d2
z2
X2
e4
g21
d3
Y4
h2
X3
e5
g11
Y5
e1
e6
Y1
Y6
h1
b21
e2
Y2
e3
Y3
z1
38SEM-Model Notation
Inner relations, inner model
Outer relations, measurement model
X, d 3-component vector Y, e 6-component vector
X Lxxd, Y Lyhe
39Statistical Assumptions
- Error terms of inner model (z) have
- zero means
- constant variances across observations
- are uncorrelated across observations
- are uncorrelated with exogenous variables
- Error terms of measurement models (d, e) have
- zero means
- constant variances across observations
- are uncorrelated across observations
- are uncorrelated with latent variables and with
each other - Latent variables are standardized
40Covariance Matrix of Manifest Variables
- Unrestricted covariance matrix (order K kxky)
- S Var(X,Y)
- Model-implied covariance matrix
-
41Estimation of the Parameters
- Covariance fitting methods
- search for values of parameters q so that the
model-implied covariance matrix fits the observed
unrestricted covariance matrix of the MVs - LISREL (LInear Structural RELations) Jöreskog
(1973), Keesling (1972), Wiley (1973) - Software LISREL by Jöreskog Sörbom
- PLS techniques
- partition of q in estimable subsets of
parameters - iterative optimizations provide successive
approximations for LV scores and parameters - Wold (1973, 1980)
42Discrepancy Function
- The discrepancy or fitting function
- F(SS) F(S S(q))
- is a measure of the distance between the
model-implied covariance-matrix S(q) and the
estimated unrestricted covariance-matrix S - Properties of the discrepancy function
- F(SS) 0
- F(SS) 0 if SS
-
43Covariance Fitting (LISREL)
- Estimates of the parameters are derived by
- F(SS(q)) q min
- Minimization of (K number of indicators)
- F(SS) logS logS trace (SS-1) K
- gives ML-estimates, if the manifest variables
are independently, multivariate normally
distributed - Iterative Algorithm (Newton-Raphson type)
- Identification
- Choice of starting values is crucial
- Other choices of F result in estimation methods
like OLS and GLS ADF (asymptotically
distribution free)
44PLS Techniques
- Estimates factor scores for latent variables
- Estimates structural parameters (path
coefficients, loading coefficients), based on
estimated factor scores, using the principle of
least squares - Maximizes the predictive accuracy
- Predictor specification, viz. that E(hx)
equals the systematic part of the model, implies
E(zx)0 the error term has (conditional) mean
zero - No distributional assumptions beyond those on 1st
and 2nd order moments
45The PLS-Algorithm
- Step 1 Estimation of factor scores
- Outer approximation
- Calculation of inner weights
- Inner approximation
- Calculation of outer weights
- Step 2 Estimation of path and loading
coefficients by minimizing Var(z) and Var(d) - Step 3 Estimation of location parameters
(intercepts) - Bo from h Bo Bh Gx z
- Lo from X Lo Lxx d
46Estimation of Factor Scores
- Factor hi realizations Yin, n1,,N
- Yin(o) outer approximation of Yin
- Yin(i) inner approximation of Yin
- Indicator Yij observations yijn j1,,Ji
n1,,N - Outer approximation Yin(o)Sjwijyijn s.t.
Var(Yi(o))1 - Inner weights vihsign(rih), if hi and hh
adjacent otherwise vih0 rihcorr(hi,hh)
(centroid weighting) - Inner approximation Yin(i)ShvihYhn(o) s.t.
Var(Yi(i))1 - Outer weights wijcorr(Yij,Yi(i))
- Start choose arbitrary values for wij
- Repeat 1. through 4. until outer weights converge
47Example
d1
E1
EX
d2
z2
E2
e4
g21()
d3
C1
CS
E3
e5
g11()
C2
e1
e6
Q1
C3
PQ
b21()
e2
Q2
e3
Q3
z1
48Example, contd
- Starting values wEX,1,,wEX,3,wPQ,1,,wPQ,3,wCS,1,
,wCS,3 - Outer approximation
- EXn(o) SjwEX,jEjn similar PQn(o), CSn(o)
- standardized
- Inner approximation
- EXn(i) PQn(o) CSn(o)
- PQn(i) EXn(o) CSn(o)
- CSn(i) EXn(o) PQn(o)
- standardized
- Outer weights
- wEX,j corr(Ej,EX(i)), j1,,3 similar wPQ,j,
wCS,j
49Choice of Inner Weights
- Centroid weighting scheme Yin(i)ShvihYhn(o)
- vijsign(rih), if hi and hh adjacent, vij0
otherwise - with rihcorr(hi,hh) these weights are obtained
if vih are chosen to be 1 or -1 and Var(Yi(i))
is maximized - Weighting schemes
hh predecessor hh successor
centroid sign(rih) sign(rih)
factor, PC rih rih
path bih rih
bih coefficient in regression of hi on hh
50Measurement Model Examples
- Latent variables from Swedish CSB Model
- Expectation
- E1 new customer feelings
- E2 special products/services expectations
- E3 overall expectation
- Perceived Quality
- Q1 range of products/services
- Q2 quality of service
- Q3 clarity of information on products/services
- Q4 opening hours and appearance of location
- Q5 etc.
51Measurement Models
- Reflective model each indicator is reflecting
the latent variable (example 1) - Yij lijhi eij
- Yij is called a reflective or effect indicator
(of hi) - Formative model (example 2)
- hi py'Yi di
- py is a vector of ki weights Yij are called
formative or cause indicators - Hybrid or MIMIC model (for multiple indicators
and multiple causes) - Choice between formative and reflective depends
on the substantive theory - Formative models often used for exogenous,
reflective and MIMIC models for endogenous
variables
52Estimation of Outer Weights
- Mode A estimation of Yi(o) reflective
measurement model - weight wij is coefficient from simple regression
of Yi(i) on Yij wij corr(Yij,Yi(i)) - Mode B estimation of Yi(o) formative
measurement model - weight wij is coefficient of Yij from multiple
regression of Yi(i) on Yij, j1,,Ji - multicollinearity?!
- MIMIC model
53Properties of Estimators
- A general proof for convergence of the
PLS-algorithm does not exists practitioners
experience no problems - Factor scores are inconsistent but consistent at
large consistency is achieved with increasing
sample size and block size - Loading coefficients are inconsistent and seem to
be overestimated - Path coefficients are inconsistent and seem to be
underestimated
54ACSI Model Results
Perceived Quality
Voice
-0,38 -0,29
0,78 0,47
Custo- mer Satis- faction
0,90 0,73
0,17 (0,06)
0,95 0,53
0,57 0,35
(-0,15) 0,12
Loyalty
Expec- tations
Value
0,40 0,35
-0,24 (0,06)
EQS-estimates PLS-estimates
55Evaluation of SEM-Models
- Depends on estimation method
- Covariance-fitting methods distributional
assumptions, optimal parameter estimates, factor
indeterminacy - PLS path modeling non-parametric, optimal
prediction accuracy, LV scores - Step 1 Inspection of estimation results (R2,
parameter estimates, standard errors, LV scores,
residuals, etc.) - Step 2 Assessment of fit
- Covariance-fitting methods global measures
- PLS path modeling partial fitting measures
56Inspection of Results
- Covariance-fitting methods global optimization
- Model parameters and their standard errors do
they confirm theory? - Correlation residuals sij-sij(q)
- Graphical methods
- PLS techniques iterative optimization of outer
models and inner model - Model parameters
- Resampling procedures like blindfolding or
jackknifing give standard errors of model
parameters - LV scores
- Graphical methods
57Fit Indices
- Covariance-fitting methods covariance fit
measures such as - Chi-square statistics
- Goodness of Fit Index (GFI), AGFI
- Normed Fit Index (NFI), NNFI, CFI
- Etc.
- Basis is the discrepancy function
- PLS path modeling prediction-based measures
- Communality
- Redundancy
- Stone-Geissers Q2
58Chi-square Statistic
- Test of H0 S S(q) against non-specified
alternative - Test-statistic X2(N-1)F(SS( ))
- If model is just identified (cp) X20
cK(K1)/2, p number of parameters in q - Under usual regularity conditions (normal
distribution, ML-estimation), X2 is
asymptotically ?2(c-p)-distributed - Non-significant X2 indicate the over-identified
model does not differ from a just-identified
version - Problem X2 increases with increasing N
- Some prefer X2/(c-p) to X2 (has reduced
sensitivity to sample size) rule of thumb
X2/(c-p) lt 3 is acceptable
59Goodness of Fit Indices
- Goodness of Fit Index (Jöreskog Sörbom)
- Portion of observed covariances explained by the
model-implied covariances - How much better fits the model as compared to no
model at all - Ranges from 0 (poor fit) to 1 (perfect fit)
- Rule of thumb GFI gt 0.9
- AGFI penalizes model complexity
60Other Fit Indices
- Normed Fit Index, NFI (Bentler Bonett)
- Similar to GFI, but compares with a baseline
model, typically the independence model
(indicators are uncorrelated) - Ranges from 0 (poor fit) to 1 (perfect fit)
- Rule of thumb NFI gt 0.9
- Comparative Fit Index, CFI (Bentler)
- Less depending of sample size than NFI
- Non-Normed Fit Index, NNFI (Bentler Bonett)
- Also known as Tucker-Lewis Index
- Adjusted for model complexity
- Root mean squared error of approximation, RMSEA
(Steiger)
61Assessment of PLS Results
- Not a single but many optimization steps not a
global measure but many measures of various
aspects of results - Indices for assessing the predictive relevance
- Portions of explained variance (R2)
- Communality, redundancy, etc.
- Stone-Geissers Q2
- Reliability indices
- NFI, assuming normality of indicators
- Allows comparisons with covariance-fitting results
62Some Indices
- Assessment of diagonal fit (proportion of
explained variances) - SMC (squared multiple correlation coefficient)
R2 (average) proportion of the variance of LVs
that is explained by other LVs concerns the
inner model - Communality H2 (average) proportion of the
variance of indicators that is explained by the
LVs directly connected to it concerns the outer
model - Redundancy F2 (average) proportion of the
variance of indicators that is explained by
predictor LVs of its own LV - r2 proportion of explained variance of
indicators
63Some Indices, contd
- Assessment of non-diagonal fit
- Explained indicator covariances
- rs 1- c/s
- with c rms(C), s rms(S) C estimate of
Cov(e) - Explained latent variable correlation
- rr 1- q/r
- with q rms(Q), r rms(Cov(Y)) Q estimate of
Cov(z) - reY rms (Cov(e,Y)), e outer residuals
- reu rms (Cov(e,u)), u inner residuals
- rms(A) (SiSj aij2)1/2 root mean squared
covariances (diagonal elements of symmetric A
excluded from summation)
64Stone-Geissers Q2
- Similar to R2
-
- E sum of squared prediction errors O sum of
squared deviations from mean - Prediction errors from resampling (blindfolding,
jackknifing) - E.g., communality of Yij, an indicator of hi
65Lohmöllers Advice
- Check fit of outer model
- Low unexplained portion of indicator variances
and covariances - High communalities in reflective blocks, low
residual covariances - Residual covariances between blocks close to zero
- Covariances between outer residuals and latent
variables close to zero - Check fit of inner model
- Low unexplained portion of latent variable
indicator variances and covariances - Check fit of total model
- High redundancy coefficient
- Low covariances of inner and outer residuals
66ACSI Model Results
Perceived Quality
Voice
-0,38 -0,29
0,78 0,47
Custo- mer Satis- faction
0,90 0,73
0,17 (0,06)
0,95 0,53
0,57 0,35
(-0,15) 0,12
Loyalty
Expec- tations
Value
0,40 0,35
-0,24 (0,06)
EQS-estimates PLS-estimates
67Diagnostics EQS
ACSI ACSIe
c2 247.5 378.7
df 81 173
NNFI 0.898 0.930
RMSEA 0.079 0.060
68Diagnostics PLS (centroid weighting)
ACSI ACSI e Hui Schenk
R2 0.29 0.35 0.43 0.40
Q2 0.36 0.41 0.58 0.49
rr 0.47 0.55 0.58 0.59
H2 0.71 0.64 0.64 0.64
F2 0.22 0.24 0.30 0.26
r2 0.63 0.63 0.57 0.60
reY 0.26 0.24 0.19 0.09
reu 0.19 0.17 0.16 0.08
69Practice of SEM Analysis
- Theoretical basis
- Data
- Scaling metric or nominal (in LISREL not
standard) - Sample-size a good choice is 10p (p number of
parameters) lt5p cases might result in unstable
estimates large number of cases will result in
large values of X2 - Reflective indicators are assumed to be
uni-dimensional it is recommended to use
principal axis extraction, Cronbachs alpha and
similar to confirm the suitability of data - Model
- Identification must be checked for covariance
fitting methods - Indicators for LV can be formative or reflective
formative indicators not supported in LISREL
70Practice of SEM Analalysis contd
- Model
- LISREL allows for more general covariance
structures e.g., correlation of measurement
errors - Estimation
- Repeat estimation with varying starting values
- Diagnostic checks
- Use graphical tools like plots of residuals etc.
- Check each measurement model
- Check each structural equation
- Lohmöllers advice
- Model trimming
- Stepwise model building (Hui, 1982 Schenk, 2001)
71LISREL vs PLS
- Models
- PLS assumes recursive inner structure
- PLS allows for higher complexity w.r.t. B, G, and
L LISREL w.r.t. Y and Q - Estimation method
- Distributional assumptions in PLS not needed
- Formative measurement model in PLS
- Factor scores in PLS
- PLS biased estimates, consistency at large
- LISREL ML-theory
- In PLS diagnostics much richer
- Empirical facts
- LISREL needs in general larger samples
- LISREL needs more computation
72The Extended Model
Emotional Factor
(0,20) 0,33
Perceived Quality
0,31 0,35
0,58 0,37
Custo- mer Satis- faction
0,55 0,36
0,85 0,53
0,87 0,73
(-0,14) (-0,01)
0,48 0,34
Loyalty
Expec- tations
Value
0,41 0,34
(-0,14) (0,06)
EQS-estimates PLS-estimates
73Diagnostics EQS
ACSI ACSI e
c2 247.5 378.7
df 81 173
NNFI 0.898 0.930
RMSEA 0.079 0.060
74Diagnostics PLS (centroid weighting)
ACSI ACSI e Hui Schenk
R2 0.29 0.35 0.43 0.40
Q2 0.36 0.41 0.58 0.49
rr 0.47 0.55 0.58 0.59
H2 0.71 0.64 0.64 0.64
F2 0.22 0.24 0.30 0.26
r2 0.63 0.63 0.57 0.60
reY 0.26 0.24 0.19 0.09
reu 0.19 0.17 0.16 0.08
75Model Building Huis Approach
Emotional Factor
Perceived Quality
0,61
0,43
0,31
-0,18
Custo- mer Satis- faction
0,10
0,35
0,36
0,42
0,33
Expec- tations
-0,18
0,17
Loyalty
Value
0,63
0,21
0,23
0,12
76Model Building Schenks Approach
Emotional Factor
Perceived Quality
0,32
0,35
Custo- mer Satis- faction
0,31
0,32
0,73
0,34
Expec- tations
0,60
Value
77The end
78Data-driven Specification
- No solid a priori knowledge about relations among
variables - Stepwise regression
- Search of the best model
- Forward selection
- Backward elimination
- Problem omitted variable bias
- General to specific modeling
79Stepwise SE Model Building
- Hui (1982) models with interdependent inner
relations - Schenk (2001) guaranties causal structure, i.e.,
triangular matrix B of path coefficients in the
inner model - ? B ? ?
80Stepwise SE Model Building
- Huis algorithm
- Stage 1
- Calculate case values Yij for LVs ?i as principal
component of corresponding block, calculate R
Corr(Y) - Choose for each endogenous LV the one with
highest correlation to form a simple regression - Repeat until a stable model is reached
- PLS-estimate the model, calculate case values,
and recalculate R - Drop from each equation LVs with t-value tlt1,65
- Add in each equation the LV with highest partial
correlation with dependent LV
81Stepwise SE Model Building
- Huis algorithm, contd
- Stage 2
- Use rank condition for checking identifiability
of each equation - Use 2SLS for estimating the path coefficients in
each equation
82Huis vs. Schenks Algorithm
- Huis algorithm is not restricted to a causal
structure allows cycles and an arbitrary
structure of matrix B - Schenks algorithm
- uses an iterative procedure similar to that used
by Hui - makes use of a priori information about the
structure of the causal chain connecting the
latent variables - latent variables are to be sorted
83Stepwise SE Model Building
- Schenks algorithm
- Calculate case values Yij for LVs ?i as principal
component of corresponding block, calculate R
Corr(Y) - Choose pair of LVs with highest correlation
- Repeat until a stable model is reached
- PLS-estimate the model, calculate case values,
and recalculate R - Drop LVs with non-significant t-value
- Add LV with highest correlation with already
included LVs
84Data, special CS dimensions
Staff 2 availability1 (PERS), politeness1 (FREU)
Outlet 3 make-up1 (GEST), presentation of mer-chandise1 (PRAE), cleanliness1 (SAUB)
Range 2 freshness and quality (QUAL), richness (VIEL)
Customer-orientation 7 access to outlet (ERRE), shopping hours (OEFF), queuing time for checkout1 (WART), paying modes1 (ZAHL), price information1 (PRAU), sales (SOND), availability of sales (VERF)
1 Dimension of Emotional Factor
85References
- C. Fornell (1992), A National Customer
Satisfaction Barometer The Swedish Experience.
Journal of Marketing, (56), 6-21. - C. Fornell and Jaesung Cha (1994), Partial Least
Squares, pp. 52-78 in R.P. Bagozzi (ed.),
Advanced Methods of Marketing Research.
Blackwell. - J.B. Lohmöller (1989), Latent variable path
modeling with partial least squares.
Physica-Verlag. - H. Wold (1982), Soft modeling. The basic design
and some extensions, in Vol.2 of Jöreskog-Wold
(eds.), Systems under Indirect Observation.
North-Holland. - H. Wold (1985), Partial Least Squares, pp.
581-591 in S. Kotz, N.L. Johnson (eds.),
Encyclopedia of Statistical Sciences, Vol. 6.
Wiley.