Title: Problem Intervention Investigation, Formulation, Design, Implementation
1Problem InterventionInvestigation, Formulation,
Design, Implementation Evaluation
2Overview
- Approaches to organisational problems scientist
vs (?) practitioner - Deciding appropriate interventions - collecting
data about the problem - Intervention designs
- Evaluation
3Dealing with Organisational Problems The
Role(s) of the Research Practitioner
4Scientist vs Practitioner Approaches to dealing
with organisational problems
Case study example an organisation wants to
design a new structured interview for its
graduate training scheme, which will include some
measure of their leadership potential. Difference
s between the scientist and the practitioner
in how they tackle this ? Occupational
psychologists as Scientist-Practitioners
5Defining the problem - methods of investigation
- Prerequisites
- Appropriate to research objective
- Elicit an appropriate form of data
- Feasible (time, resource, organisational
constraints) - Piloted
- Ethical
- Agreed and accepted by the organisation
- Used appropriately
- Are experienced and confident with
6Methods of Investigation (see also M46 M21
notes for Data Collection Methods employed in
organisations)
- Interviews
- amount of structure
- specific types
- Q-Sort (Stephenson, 1936)
- Rep Grid (Kelly, 1955)
- Critical Incident Technique (Flanagan, 1954)
- Focus groups
- Tips for successful interviewing
- Advantages ?
- Disadvantages ?
7- Psychometrics
- Measures of a psychological construct which have
been designed to have psychometric rigor
(validity and reliability) for that context, and
responses usually expressed as standardised
scores. - Examples
- Team Working Team Climate Inventory
- Employee Stress Occupational Stress Indicator
(OSI), Spielberger State-Trait Anxiety Inventory
(STAI), - Workplace Reaction Job Involvement Scale,
Organisational Commitment Questionnaire, Job
Satisfaction Index - Cognitive Ability Graduate Managerial
Assessments (GMA) Watson-Glaser Critical
Thinking, Ravens progressive matrices - Personality 16PFi, NEO-PIR, OPQ
- Interest Vocational Preference Inventory (VPI),
Strong Vocational Interest Blank. - .. Other common techniques observation,
questionnaires, diaries.
8Interventions
- an ongoing system of relationships for the
purpose of helping (Argyris, 1970) - a planned programmatic activities pursued by
both clients and consultants in the pursuit of
organisational development (French, Bell
Zawacki, 1994) - Blake Mouton (1994) 9 categories of OD
interventions (e.g. theory intervention, cultural
intervention, procedural intervention,
experimental intervention) - French et al (1995) differentiate according to
focus of interventions (individual, dyads, teams,
intergroup, organisations)
9Models of Intervention
- 1. The Intervention Model (Argyris, 1971)
- Assumptions organisations cannot diagnose /or
solve own problems - Activities of consultant data collection gtgtgt
develop options - 2. The Planned Change Model (Lippitt, Watson
Westley, 1958 Schein, 1972). - Assumption organisation will detect problem
- Activities of consultant diagnostic activity
(as contracted), specific interventions and
evaluation.
10- 3. The Action Research Model
- Assumption the diagnostic process is integral
to the intervention. - Activities of consultant iterative process of
intervention-evaluation - 4. Purchase of Expertise Model
- Buyer (organisation) defines need, decides
organisation does not have time/resources - 5. Doctor-Patient Model
- Consultant gives organisation health check
treats any symptoms of ill health - 6. Process Consultation Model (Schein, 1969
1988) - Consultant tries to give client org insight into
process events collaboratively decide
intervention
11Interventions
- Organisation Employee
- Focused Focused
12Intervention Evaluation
- systematic collection of descriptive and
judgmental information necessary to make
effective training decisions related to the
selection, adoption, value and modification of
various instructional objectives (Goldstein,
1991) - 1. Why is evaluation necessary?
- 2. Who should do the evaluation
- 3. What aspects should be evaluated, and when
should this be done? - 4. What kinds of measurements (approaches and
methods) need to be used?
131. Why is evaluation necessary?
- Patrick (1992)
- Training programme improvement
- Feedback to- and decisions about trainees
- Administrative and organisational decisions
- Validation of the training intervention
- N.B. Summative vs Formative evaluation (Scriven,
1967). - 2. Who should carry out the evaluation?
- Person running the training course?
- Trainees?
- Line manager? N.B. Issues of bias, sensitivity
- HR/Personnel? of feedback, accuracy,
- Others? sufficient knowledge, etc...
143. What aspects should be evaluated, and when?
- Whitelaw (1972) Hamblin (1974) Kirkpatrick
(1967) - Level 1 reactions of trainees to content and
methods - Level 2 learning outcomes attained
- Level 3 job behaviour in the workplace
- Level 4 Effect on trainees department /
local org effects - Level 5 Ultimate effect on organisation
- N.B. Criterion development
- criterion referenced vs norm-referenced.
- Criteria relevance (Goldstein, 1991)
- criteria deficiency
- criteria contamination
- Phillips (1990). Evaluation usually limited to
trainee reactions
154. Evaluation Methods (Approaches Techniques)
- A. Trainee Reactions
- Questionnaire/Post-course survey of(what
dimensions)? - B. Learning Outcomes Attained
- Post-training test
- Measure of retention of training
- C. Job Behaviour
- Techniques to measure on-site performance ?
Objective subjective performance criteria
164. Evaluation Methods (Approaches Techniques) II
- D. Department / Local Organisational Effects
- Quality/Quantity of production
- Safety N.B. Issues?
- Turnover (in team)
- Team climate/attitudes towards work
- E. Ultimate Organisational Effects
- Usually financial effect (utility analysis), but
is maximising profits always an organisations
main objective?
17N.B. Formative Evaluation
- Why focus on the processes of the intervention?
- Explains why certain results were obtained
- Helps develop an explanatory basis for the
development of an intervention - Determine the dynamics/subtleties of the
intervention
18Evaluation Designs
- One group, post-measure only
- X T2
- One group, pre- and post measures
- T1 X T2
- Two groups, pre- and post measures, experimental
control - Re T1 X T2
- Rc T1 T2
- Two groups, pre- and post measures,
non-equivalent control - Re T1 X T2
- Ro T1 T2
19Threats to accurate evaluation
- Internal validity
- the extent to which training has brought about
new learning - vs
- External validity
- the extent to which training will generalise to
subsequent groups of trainees in same/different
organisations
20Conclusions
- Organisations are in a constant state of flux
- Employees have needs that are dictated by
personal growth, career development, work-life
interactions, etc. - Adequate analysis of employee needs provides the
basis for interventions - Interventions need to be evaluated