Title: Model validity, testing and analysis
1Model validity, testing and analysis
Yaman Barlas Bogaziçi University Industrial
Engineering Department 34342 Bebek Istanbul,
Turkey ybarlas_at_boun.edu.tr http//www.ie.boun.edu.
tr/barlas SESDYN Group http//www.ie.boun.edu.t
r/labs/sesdyn/
2Conceptual and Philosophical Foundations
- Model Validity and Types of Models
- Statistical Forecasting models (black box)
- Descriptive Policy models (transparent)
- Philosophical Aspects
- - Philosophy of Science
- - Logical Empiricim and Absolute Truth
- - Conversational justification relative truth
(purpose) - - Statistical significance testing
- (Barlas and Carpenter 1990 and Barlas 1996)
3Two aspects of model validity
- Structure Validity
- Primary importance
- Special place in System Dynamics
- Behavior Validity
- Role in system dynamics
- The special type of behavior validity in system
dynamics - Ex ante versus ex post prediction
- (Barlas 1996 and 1989)
4Overall Nature and Selected Tests ofFormal Model
Validation
5Logical Sequence of Formal Steps ofModel
Validation
6Structure Validity
- (Simulation Verification)
- Direct Structure Tests
- Crucial, yet highly qualitative and informal
- Distributed through the entire modeling
methodology - Indirect Structure Tests (Structure-oriented
behavior) - Crucial and partly quantitative and formal
- Tool SiS software
7Indirect Structure Testing Software SiS
- Based on automated dynamic pattern recognition
- Extreme condition pattern testing
- Also in parameter calibration and policy design
- (Kanar 1999 Kanar and Barlas 1999 Bog et al
2004)
8Indirect Structure Testing Software (SiS)
Basic Dynamic Patterns
9Indirect Structure Testing Software (SiS)
List of dynamic behavior pattern classes
10Software Implementation
General Picture of the Processes in Validity
Testing mode
General Picture of the Processes in Parameter
Calibration mode
11Sample Model Used with SiS
12Validity Testing with Default Parameters
Simulation Output (with default base parameters)
Likelihood Values of simulation behavior
correctly classified as the GR2DB pattern
13Validity Testing by Setting Parameters
Fig1 Simulation Output (with base parameters)
Fig2 Simulation Output (with changed
parameters)
Likelihood Values of simulation behavior in Fig2
compared to the NEXGR pattern
14Parameter Calibration with Specified Pattern
Simulation Output (with base parameters)
The ranges and number of values tried for each
parameter
15Result of the Parameter Calibration
Simulation Output as Desired (after automated
parameter calibration)
- Best parameter set is 41
- Best Likelihood Result 1.2119776136254248
- Best Parameter Set
- 1. advertising effectiveness 0.25
- 2. customer sales effectiveness 6.0
- 3. sales size 1.0
16Parameter Calibration with Input Data
A view of the SiS interface during parameter
calibration
17Result of the Parameter Calibration
Fig1 Simulation Output (with base parameters)
Fig2 Simulation Output (after parameter
calibration to match the input pattern)
- Best parameter set is 21
- Best Likelihood Result 3.7109428620957883
- Best Parameter Set
- 1. advertising effectiveness 5.0
- 2. customer sales effectiveness 0.0
18Behavior Validity
- Two types of patterns
- Steady state
- Transient
- Major pattern components
- Trend, periods, amplitudes, ...
19Behavior Validity Testing Software BTS II
20Behavior Validity Testing Software BTS II
21BTS II ToolsTrend Regression
Model y(t) a b t a 1.4272937 b 0.9913937
22BTS II ToolsMoments
- Moment Calculations
- Of Data Points 100
- 1st Moment (Mean) 1.4272937
- 2nd Moment (Variance) 2.7107011
23BTS II ToolsAutocorrelation
24BTS II ToolsAutocorrelation Test
25BTS II ToolsSpectral Density Function
dominant period1 20 Value 16.1181481405124 do
minant period2 8 Value 0.373946663988869
26BTS II ToolsCross correlation
Max CrossCorrelation 0.7367365 at lag 0
27BTS II ToolsAmplitude Estimation
Model y(t) a b sin ( 2 p t / period c
) a 1.4272937 b 1.9958872 c
0.3500578 Amplitude Estimate 3.9917744
28BTS II ToolsDiscrepancy Coefficient
- Of Data Points 100
-
- U 0.0363687
- U1 0.0231044
- U2 0.0054147
- U3 0.9714809
29BTS II ToolsTrend in Amplitude
30BTS II ToolsTrend in Amplitude
constant 7.4321903 phase angle 3.1273996 trend
of amplitude const of amplitude
10.1432480 slope of amplitude
12.562881
31Uses of BTS II and SiS in Model Analysis
- Analysis Understanding the dynamic properties of
the model - BTS II can assist in quantifying, measuring and
assessing dynamic pattern components - SiS can assist in deeper structural analysis
(related to qualitative pattern modes)
32Uses of BTS II and SiS in Policy Design
- BTS II can assist in numerical performance
improvement policies - SiS can assist in more structural dynamic pattern
improvement - Parameter calibration can be extended to cover
automated policy design
33Implementation Issues
- More tools
- User friendliness
- More thorough (field) testing of the tools
- Better integration with simulation software
- ...
34Policy Implementation Issues
- Validity of the policy recommendation
- (Robustness, timing, duration, transition...)
- Finally, validity of the implementation itself
- Validated model means just a reliable laboratory
implementation validity does not automatically
follow it is a whole area in itself
35Concluding Observations
- Validity as a process, rather than an outcome
- Continuous (prolonged) validity testing
- Validation, analysis and policy design all
integrated - From validity towards quality
- Quality built-in versus inspected-in
- Group model building
- Testing by interactive gaming
36Back to philosophy...
- A gradual, continuous, multi-method, qualitative
and quantitative, formal and informal process of
establishing confidence in a model. We should use
any formal test/tool compatible with this
philosophy, but never assume that tools
themselves would be sufficient without proper
philosophy
37References
- Akkermans, H.A., and J.A.M. Vennix. 1997.
Clients' opinions on group model-building an
exploratory study. System Dynamics Review 13(1)
3-31 - Andersen, D.F. 1980. How Differences in Analytic
Paradigms Can Lead to Differences in Policy
Decisions. In Elements of the System Dynamics
Method, ed. J. Randers. Portland, OR
Productivity Press - Andersen, D.F., G.P. Richardson and J.A.M.
Vennix. 1997. Group model building adding more
science to the craft. System Dynamics Review.
13(2) 187-201 - Ansoff, H.I and D.P. Slevin. 1968. An
Appreciation of Industrial Dynamics. Management
Science 14 383-397. - Barlas, Y. and A. Erdem. 1994. Output Behavior
Validation in System Dynamics Simulation. In
Proceedings of the European Simulation Symposium
(Istanbul, Turkey) 81-84. - Barlas, Y. 1992. Comments on On the Very Idea of
a System Dynamics Model of Kuhnian Science.
System Dynamics Review 8(1) 43-47. - Barlas, Y. 1990. An Autocorrelation Function
Test For Output Validation. Simulation
55(1)7-16. - Barlas, Y. and S. Carpenter. 1990. Philosophical
Roots of Model Validation Two Paradigms. System
Dynamics Review 6(2)148-166. - Barlas, Y. 1989a. Multiple Tests for Validation
of System Dynamics Type of Simulation Models.
European Journal of Operational Research
42(1)59-87. - Barlas, Y. 1989b. Tests of Model Behavior That
Can Detect Structural Flaws Demonstration With
Simulation Experiments. In Computer-Based
Management of Complex Systems International
System Dynamics Conference. P.M.Milling and
E.O.K.Zahn, eds. Berlin Springer-Verlag. - Barlas, Y. 1985. Validation of System Dynamics
Models With a Sequential Procedure Involving
Multiple Quantitative Methods. Unpublished Ph.D.
Dissertation. Georgia Institute of Technology.
Atlanta, GA - Bell, J.A and M.F. Bell. 1980. System Dynamics
and Scientific Method. In Elements of the System
Dynamics Method, ed. J. Randers. Portland, OR
Productivity Press - Carson, E.R.and R.L.Flood. 1990. Model
Validation Philosophy, Methodology and
Examples. Trans Inst MC.12(4) 178-185.
38- Coyle, G. And D. Exelby. 2000. The validation of
commercial system dynamics models. System
Dynamics Review 16(1) 27-41 - Dangerfield, B. 1994. The System Dynamics
Modeling Process and DYSMAP2. In Modeling For
Learning Organizations. Morecroft, J.D.W and J.D.
Sterman, eds. Portland, OR Productivity Press - Eberlein, R.L and D.W. Peterson. 1994.
Understanding Models with VENSIM. In Modeling
For Learning Organizations. Morecroft, J.D.W and
J.D. Sterman, eds. Portland, OR Productivity
Press - European Journal of Operational Research. 1993.
Special Issue on Model Validation. 66(2). - Forrester J.W. and P.M.Senge. 1980. Tests For
Building Confidence in System Dynamics Models.
In System Dynamics. Legasto, A.A., J.W. Forrester
and J.M. Lyneis, eds. Amsterdam North-Holland - Forrester, J.W., G.W. Low and N.J. Mass.1974.
The Debate on World Dynamics A Response to
Nordhaus. Policy Sciences 5 169-190. - Forrester, J.W. 1973. Confidence in Models of
Social Behavior with Emphasis on System Dynamics
Models. System Dynamics Group Working Paper.
Sloan School of Management, MIT, Cambridge, MA. - Forrester, J.W. 1968. A Response to Ansoff and
Slevin. Management Science 14 601-618. - Forrester, J.W. 1961. Industrial Dynamics.
Portland, OR Productivity Press - Graham, A.K., J.D.W. Morecroft, P.M. Senge and
J.D. Sterman. 1994. Model-supported Case
Studies in Management Education. In Modeling For
Learning Organizations. Morecroft, J.D.W and J.D.
Sterman, eds. Portland, OR Productivity Press - Kleijnen, J.P.C. 1995. Verification and
Validation of Simulation Models. European
Journal of Operational Research 82 145-162. - Lane, D.C. 1995. The Folding Star A Comparative
Reframing and Extension of Validity Concepts in
System Dynamics. In Proceedings of
International System Dynamics Conference (Tokyo,
Japan). Vol.1 111-130. - Lane, D.C. 2001. Rerum cognoscere causas Part
II - Opportunities generated by the
agency/structure debate and suggestions for
clarifying the social theoretic position of
system dynamics. System Dynamics Review 17(4)
293-309 - Lane, D.C. 2001. Rerum cognoscere causas Part I
- How do the ideas of system dynamics relate to
traditional social theories and the
voluntarism/determinism debate? System Dynamics
Review 17(2) 97-118
39- Lyneis, J.M., K.S. Reichelt and T. Sjoblom. 1994.
Professional DYNAMO Simulation Software to
Facilitate Management Learning and Decision
Making. In Modeling For Learning Organizations.
Morecroft, J.D.W and J.D. Sterman, eds. Portland,
OR Productivity Press - Mass, N. and P.M. Senge. 1980. Alternative Tests
for Selecting Model Variables. In Elements of
the System Dynamics Method, ed. J. Randers.
Portland, OR Productivity Press - Meadows, D. 1989. Gaming to Implement System
Dynamics Models. In Computer-Based Management of
Complex Systems International System Dynamics
Conference. P.M.Milling and E.O.K.Zahn, eds.
Berlin Springer-Verlag. - Meadows, D. H. 1980. The Unavoidable A Priori.
In Elements of the System Dynamics Method, ed. J.
Randers. Portland, OR Productivity Press - Miller, R.G. 1981. Simultaneous Statistical
Inference. New York, N.Y Springer-Verlag - Mitroff, I. 1969. Fundamental Issues in the
Simulation of Human Behavior. Management
Science. 15 635-649. - Morecroft, J.D.W. and J.D. Sterman, eds. 1994.
Modeling For Learning Organizations. Portland,
OR Productivity Press - Morrison, D.E. and R.E. Henkel, eds. 1970. The
Significance Test Controversy. Chicago, IL
Aldine Press - Nordhaus, W.D. 1973. World Dynamics Measurement
Without Data. Economic Journal 83 1156-1183. - Naylor, T.H. and J.M. Finger. 1968. Verification
of Computer Simulation Models. Management
Science. 14 92-101. - Oral, M. and O. Kettani. 1993. The Facets of the
Modeling and Validation Process in Operations
Research. European Journal of Operational
Research 66(2) 216-234. - Peterson, D.W. and R.L. Eberlein. 1994. Reality
Check A Bridge Between Systems Thinking and
System Dynamics. System Dynamics Review 10(2-3)
159-174 - Radzicki, M.J. 1992. Comments on On the Very
Idea of a System Dynamics Model of Kuhnian
Science. System Dynamics Review 8(1) 49-53. - Radzicki, M.J. 1990. Methodologia Oeconomiae et
Systematis Dynamis. System Dynamics Review 6(2)
123-147. - Richardson, G.P. 1991. What Are We Publishing? A
View from the Editors Desk. System Dynamics
Review 7(1) 61-67. - Richardson, G.P. and A.L.Pugh III. 1981.
Introduction To System Dynamics Modeling With
DYNAMO. Portland, OR Productivity Press.
40- Roberts, N., D. Andersen, R. Deal, M. Garet, W.
Shaffer. 1983. Introduction to Computer
Simulation A System Dynamics Approach. Portland,
OR Productivity Press - Rouwette, E.A.J.A., J.A.M. Vennix and T.
Mullekom. 2002. Group model building
effectiveness a review of assessment studies.
System Dynamics Review 18(1) 5-45 - Scholl, G.J. 1995. Benchmarking the System
Dynamics Community Research Results. System
Dynamics Review 11(2) 139-155. - Schruben, L.W. 1980. Establishing the
Credibility of Simulations. Simulation. 34(3)
101-105. - Senge, P.M. 1977. Statistical Estimation of
Feedback Models. Simulation. 28 177-184. - Schlesinger, S. et al. 1979. Terminology for
Model Credibility. Simulation. 32(3)103-104. - Sterman, J.D. 1992. Comments on On the Very Idea
of a System Dynamics Model of Kuhnian Science.
System Dynamics Review 8(1) 35-42. - Sterman, J.D. 1989. Modeling Managerial
Behavior Misperceptions of Feedback in a Dynamic
Decision Making Experiment. Management Science
35(3) 321-39. - Sterman, J.D. 1987. Testing Behavioral
Simulation Models by Direct Experiment.
Management Science 33(12) 1572-1592. - Sterman, J.D. 1985. The Growth of Knowledge
Testing a Theory of Scientific Revolutions with a
Formal Model. Technological Forecasting and
Social Change 28(2) 93-122. - Sterman, J. D. 1984. Appropriate Summary
Statistics for Evaluating the Historical Fit of
System Dynamics Models. Dynamica. 10(2)51-66. - Vennix, J.A.M. 1999. Group model-building
tackling messy problems. System Dynamics Review.
15(4) 379-401 - Vennix, J.A.M., H.A. Akkermans and E.A.J.A.
Rouwette. 1996. Group model-building to
facilitate organizational change an exploratory
study. System Dynamics Review. 12(1) 39-58 - Wittenberg, J. 1992. On the Very Idea of a
System Dynamics Model of Kuhnian Science. System
Dynamics Review 8(1) 21-33. - Wright, R.D. 1972. Validating Dynamic Models An
Evaluation of Tests of Predictive Power. In
Proceedings of Summer Computer Simulation
Conference. 1286-96. - Zellner, A. 1980. Comment on Forresters
Information Sources for Modeling the National
Economy. Journal of the American Statistical
Association 75 567-569.