Title: Educational Research: Instruments (
1Educational Research Instruments (caveat
emptor)
EDU 8603 Educational Research Richard M. Jacobs,
OSA, Ph.D.
2Instruments
- tools researchers use to collect data for
research studies (alternatively called tests)
3The types of instruments
1. Cognitive Instruments
2. Affective Instruments
3. Projective Instruments
41. Cognitive instruments...
- Measure an individuals attainment in academic
areas typically used to diagnose strengths and
weaknesses
5Types of cognitive instruments...
provide information about how well the test
takers have learned what they have been taught in
school
achievement is determined by comparing it to the
norm, the performance of a national group of
similar students who have taken the same test
6measure the intellect and abilities not normally
taught and often are used to predict future
performance
typically provide an overall score, a verbal
score, and a quantitative score
72. Affective instruments...
- Measure characteristics of individuals along a
number of dimensions and to assess feelings,
values, and attitudes toward self, others, and a
variety of other activities, institutions, and
situations
8Types of affective instruments...
self-reports of an individuals beliefs,
perceptions, or feelings about self, others, and
a variety of activities, institutions, and
situations
frequently use Likert, semantic differential,
Thurstone , or Guttman scales
9measure the relative strength of an individuals
valuing of theoretical, economic, aesthetic,
social, political, and religious values
10an individuals self-report measuring how
behaviors characteristic of defined personality
traits describe that individual
113. Projective instruments...
- Measure a respondents feelings or thoughts to an
ambiguous stimulus
12Primary type of projective test...
participants react to a stimulus such as a
picture, inkblot or word onto which they project
a description
13Selecting an instrument...
1. determine precisely the type of instrument
needed
2. identify and locate appropriate instruments
3. compare and analyze instruments
4. select best instrument
14Instrument sources
Burros Mental Measurements Yearbook
Tests in Print
PRO-ED Publications Test Critiques Compendium
ETS Test Collection Database
ERIC/AE Test Review Locator
ERIC/Burros Test Publisher Directory
15Rules governing the selection instruments...
1. the highest validity
2. the highest reliability
3. the greatest ease of administration, scoring,
and interpretation
4. test takers lack of familiarity with
instrument
5. avoids potentially controversial matters
16Administering the instrument...
1. make arrangements in advance
2. ensure ideal testing environment
3. be prepared for all probable contingencies
17Two issues in using instruments...
1. Validity the degree to which the instrument
measures what it purports to measure
2. Reliability the degree to which the
instrument consistently measures what it purports
to measure
18Types of validity...
1. Content validity
2. Criterion-related validity
3. Construct validity
191. Content validity the degree to which an
instrument measures an intended content area
20forms of content validity
sampling validity does the instrument reflect
the total content area?
item validity are the items included on the
instrument relevant to the measurement of the
intended content area?
212. Criterion-related validity an individual
takes two forms of an instrument which are then
correlated to discriminate between those
individuals who possess a certain characteristic
from those who do not
22forms of criterion-related validity
concurrent validity the degree to which scores
on one test correlate to scores on another test
when both tests are administered in the same time
frame
predictive validity the degree to which a test
can predict how well individual will do in a
future situation
233. Construct validity a series of studies
validate that the instrument really measures what
it purports to measure
24Types of reliability...
1. Stability
2. Equivalence
3. Internal consistency
251. Stability (test-retest) the degree to which
two scores on the same instrument are consistent
over time
262. Equivalence (equivalent forms) the degree
to which identical instruments (except for the
actual items included) yield identical scores
273. Internal consistency (split-half reliability
with Spearman-Brown correction formula ,
Kuder-Richardson and Cronbacks Alpha
reliabilities, scorer/rater reliability) the
degree to which one instrument yields consistent
results
28Terms associated with instruments...
29Data
the pieces of information researchers collect
through instruments to examine a topic or
hypothesis
30Constructs
abstractions of behavioral factors that cannot
be observed directly and which researchers invent
to explain behavior
31Variable
a construct that can take on two or more values
or scores
32Raw scores
the number of items an individual scored on an
instrument
33Measurement scales
the representation of variables so that they can
be quantified
34Measurement scales...
Qualitative (categorical)
1. nominal variables
Quantitative (continuous)
2. ordinal variables
3. interval variables
4. ratio variables
351. nominal (categorical) classifies persons or
objects into two or more categories
362. ordinal (order) classifies persons or
objects and ranks them in terms of the degree to
which those persons or objects possess a
characteristic of interest
373. interval ranks, orders, and classifies
persons or objects according to equal differences
with no true zero point
384. ratio ranks, orders, classifies persons or
objects according to equal differences with a
true zero point
39Norm reference
provides an indication about how one individual
performed on an instrument compared to the other
students performing on the same instrument
40Criterion reference
involves a comparison against predetermined
levels of performance
41Self reference
involves measuring how an individuals
performance changes over time
42Operationalize
the process of defining behavioral processes
that can be observed
43Standard error of measurement
an estimate of how often a researcher can expect
errors of a given size on an instrument
44Mini-Quiz
a large standard error of measurement indicates
a high degree of reliability
false
45a large standard error of measurement indicates
low reliability
true
46most affective tests are projective
false
47the primary source of test information for
educational researchers is the Burros Mental
Measurements Yearbook
true
48research hypotheses are usually stated in terms
of variables
true
49similar to a Thurstone scale, a Guttman scale
attempts to determine whether an attitude is
unidimensional
true
50validity requires the collection of evidence to
support the desired interpretation
true
51researchers should first consider developing an
instrument rather than utilizing a published
instrument
false
52a researchers goal is to achieve perfect
predictive validity
false
53predictive validity is extremely important for
instruments that are used to classify or select
individuals
true
54a high validity coefficient is closer to 1.00
than 0.00
true
55norm reference and criterion reference are
synonymous terms
false
56criterion related refers to correlating one
instrument with a second instrument the second
instrument is the criterion against with the
validity of the second instrument is judged
false
57a valid test is always reliable but a reliable
test is not always valid
true
58it is difficult to state appropriate reliability
coefficients because reliability, like validity,
is dependent upon the group being tested, i.e.,
groups with different characteristics will
produce different reliabilities
true
59content validity is not compromised if the
instrument covers topics not taught
false
60the tendency of an individual to respond
continually in a particular way
response set
61a study which consists of two quantitative
variables
correlational
62a study which consists of one categorical and
one quantitative variable
experimental or causal-comparative
63a study which consists of two or more
categorical variables
correlational or descriptive
64data collection methods which emphasize student
processes or products
performance
65data collection methods including
multiple-choice, true-false, and matching
selection
66data collection methods in which students fill
in the blank, provide a short answer, or write an
essay
supply
67an instrument administered, scored, and
interpreted in the same way no matter where or
when it is administered
standardized
68the term that includes the general process of
collecting, synthesizing, and interpreting
information, whether formal or informal
assessment
69a formal, systematic, usually paper-and-pencil
procedure for gathering information about
peoples cognitive and affective characteristics
test
70the degree to which individuals seek out or
participate in particular activities, objects,
and ideas
interests
71also called temperament, the characteristics
representing an individuals typical behaviors
and describes what individual do in their natural
life circumstances
personality
72things individuals feel favorable or unfavorable
about the tendency to accept or reject groups,
ideas, or objects
attitudes
73deeply held beliefs about ideas, persons, or
objects
values
74requires administering the predictor instruments
to a different sample from the same population
and developing a new equation
cross-validation
75Minnesota Multiphasic Personality Inventory
personality inventory
76Stanford-Binet
achievement test
77Strong Campbell
interest inventory
78SRA Survey of Basic Skills
achievement test
79Weschler Intelligence Scales
aptitude test
80Gates-McGinitie Reading Test
achievement test
81Otis-Lennon School Ability Test
aptitude test
82Kuder Occupational
interest inventory
83Rorschach Inkblot Test
projective
84Meyers-Briggs Type Indicator
personality inventory
85Iowa Test of Basic Skills
achievement test
86Thematic Apperception Test
projective
87compares the content of the test to the domain
being measured
content
88Graduate Record Examination
predictive
89correlates scores from one instrument to scores
on a criterion measure, either at the same or
different time
criterion-related
90amasses convergent, divergent, and
content-related evidence to determine that the
presumed construct is what is being measured
construct
91- Which type of reliability
scores on one instrument are consistent over time
stability (test-retest)
92- Which type of reliability
the extent to which independent scorers or a
single scorer over time agree on the scoring of
an open-ended instrument
scorer/rater
93- Which type of reliability
scores correlate between similar version of an
instrument given at different times
equivalence and stability
94- Which type of reliability
scores correlate between two versions of a test
that are intended to be equivalent
equivalence (alternate forms)
95- Which type of reliability
the extent to which items included on an
instrument are similar to one another in content
internal consistency
96- Which type of response scale
an individual gives a quantitative rating to a
topic where each position on the continuum has an
associated score value
semantic differential
97- Which type of response scale
value points are assigned to a participants
responses to a series of statements
Likert
98- Which type of response scale
participants select from a list of statements
that represent differing points of view from
those which participations agree
Thurstone
99This module has focused on...
instruments
which describes the procedures researchers use
to select individuals to participate in a study
100The next module will focus on...
qualitative research
...the tools researchers use to gather data for a
study