Title: Summative Assessment: Rubrics and Tests
1Summative AssessmentRubrics and Tests
- Effective Teaching and Learning
- Baker College
2Outcomes
- Apply a systematic process for creating a test
blueprint - Identify attributes of effective test questions
- Explain the advantages and disadvantages of
different types of test questions - Assess the quality of tests and test items
- Create samples of effective questions
3What type of assessment?
- Procedural knowledge
- Declarative knowledge
4Test Writing Process
- Planning the test
- Writing test items
- Selecting test items
- Formatting the test
- Assessing the test
- Revising the test
- Using the test
- After the test
5Planning the Test
- Content Blueprint
- Learning outcomes
- Weight
- Length
- Types of items
- Number of items
6Test Blueprint
7Types of Items
- Recognition
- True-false
- Multiple-choice
- Multiple-answer
- Matching
- Ordering
- Recall
- Short Answer
- Completion
- Essay
8Average Response Time
9Writing Test Items
- Simple and direct wording
- Avoid jargon
- Avoid trivia items
- Match items to learning outcomes
- Each item has an agreed upon correct answer
- Write more questions than you will need
10Multiple-choice Items
- Stem
- Direct question
- Incomplete statement
- Responses
- One correct answer
- Multiple distracters
11Stem
- Clearly worded
- One idea
- Avoid the use of negatives
- Enough information to answer the question
- Direct questions preferred
- Blanks at the end of the stem
- Include words repeated in all responses
12Responses
- 3-5 per item
- Avoid all of the above and none of the above
- Grammatically correct with stem
- Similar length and structure
- Avoid absolute words
- Listed in a logical order
- Mutually exclusive and not overlapping
13Distracters
- Plausible
- Common misconceptions
- Logical misinterpretations
- Clichés
- Partial answers
- Technical terms or jargon
14Example
- What is the minimum number of responses for a
multiple-choice item? - A) 2
- B) 3
- C) 4
- D) 5
15Application Example
- What problem exists in the following
multiple-choice stem - ________ is the most common type of test item.
- Absolute words should be avoided in the stem.
- The stem contains more than one idea or concept.
- Not enough information is presented to answer the
question. - The fill-in the blank should come at the end.
16Analysis and Evaluation Example Stem
An instructor was asked the following question
"Briefly list and explain how you develop a
test. As an answer, this instructor wrote the
following I begin by going through the chapter
and writing questions based on the material in
the text and my lectures. Then I decide how many
questions I want and select the best questions
from the list that I have developed. I format the
test and add instructions. After a few days, I
review the questions and make any revisions that
need to be done and remove any jargon or wording
lifted directly from the text. Then I use the
test in class. Based on how the class does, I may
make changes for the next time I teach the class.
17Analysis and Evaluation Example Responses
- Based on the process described in Effective
Classroom Tests, how would you judge this
answer? - A) EXCELLENT (all steps in the right order with
correct, clear, and complete descriptions) - B) GOOD (all stages correct in the right order,
but the descriptions are not as complete as they
should be). - C) MEDIOCRE (one or two stages are missing, OR
the stages are in the wrong order, OR the
explanations are not complete, OR the
explanations are irrelevant) - D) UNACCEPTABLE (one or more stages are missing
AND the explanations are not complete AND/OR they
are irrelevant)
18Poor Question 1
- Good multiple choice items
- A) are easy to write
- B) can only test memorized content
- C) are better than essay items
- D) there is no such thing
- E) can test a wide range of content
19Poor Question 2
- Which of the following characteristics is not
true of completion test items but is an important
distinguishing attribute of matching tests,
multiple-choice questions, and true-false items? - A) They are objective test items.
- B) They require knowledge recognition but not
production. - C) Much more difficult to construct.
20Poor Question 3
- Which of the following statements is FALSE?
- A) Misfeasance is the improperly doing of an
illegal act. - B) Nonfeasance is improperly doing a legal act.
- C) Nonfeasance is the failure to do an act that
one must do legally. - D) Misfeasance is the failure to PROPERLY do an
act that one has a duty to perform. - E) None of the above.
21Poor Question 4
- __________ is/are the best method to determine if
students have learned something. - A) Comprehensive Exam
- B) Homework Assignments
- C) Pop Quizzes
- D) Research Paper
22Selecting Test Items
- Outcome Weight x Questions by Type Questions
of Each Type for Outcome
23Formatting the Test
- Group items by type
- Sort items by increasing difficulty
- Add instructions
- Review layout and pagination
- Write answer key
24Assessing the Test
- Self
- 2-3 days after writing the test
- Clarity
- Clues in items to other items
- Non-expert
- Clarity
- Contextual clues
- Peer
- Content
- Weighting to outcomes
- Answer key
- Students
- Clarity
- Content
25Test Taking Procedures
- Use of notes or other materials
- Time limits
26After the Test
- Item Analysis
- Areas for review
- Test revisions
27Activity
- Write two test questions on any topic
- One question should be an example of a good test
item - One question should be an example of a poorly
written test item
28Share
- Share your two questions with a partner
- Can they determine which is good and which is
bad? - Can they explain what makes one poorly written?
- As a team, how can you fix the poorly written
questions?
29Intermission
30Outcomes
- Determine what characteristics are important in
evaluating student work - Evaluate rubrics, analytic scales, and other
evaluation methods - Describe the contents of a good rubric
- Identify rubrics already in use at Baker College
- Begin work on a rubric for a class
31What is a Rubric?
- A rubric is a scoring tool or guide that lists
the specific criteria and the ranges for multiple
levels of achievement for a piece of work or
performance. A rubric consists of a set of
well-defined factors and criteria describing the
dimensions of an assignment to be assessed or
evaluated.
32Parts of a Rubric
- Scale (columns)
- Dimensions (rows)
- Criteria descriptions (cells)
33Benefits of Rubrics
- Communicates the instructors expectations
- Streamlines the process for feedback to the
student - Facilitates equitable grading
- Standardizes assessment across different
instructors
34Uses for Rubrics
- Papers
- Presentations
- Projects
- Essays
- Homework
- Case Studies
- Participation/Class Discussion
- Portfolios
35Types of Rubrics
- Analytic
- Page 11
- Holistic
- Page 12 and 13
- Check List
- Page 14
- Scoring Guide
- Page 15
36Creating a Rubric
- Identify components/outcomes of the assignment
- Determine a scale
- Add criteria
- Assign points
- Set component weights (optional)
- Assess the rubric
- Test and revise
37Activity
- Split into groups of 3-4
- Determine team roles
- Select an assignment that needs a rubric
- Can be a specific assignment, such as a research
paper for ENG 102 - Can be of a more general nature such as a class
presentation
38Step 1 Identify Components
- List 5 major objectives/outcomes of the
assignment - Write these items as the row headers of the sheet
provided
39Step 2 Determine a Scale
- Aim for 3-5 levels
- Can use an odd or even number of items
- Use the headings on the next slide for ideas
- Write these as column headings on the sheet
provided
40Potential Column Headings
- Outstanding Accomplished Proficient
Developing Beginning - Accomplished Average Developing Beginning
- Excellent Good Needs Improvement
Unsatisfactory - Exceptional Acceptable Marginal
Unacceptable - Expert Practitioner Apprentice Novice
- Professional Adequate Needs Work Youre
Fired - Exceeds Expectation On Target Beginning
- Exemplary Competent Developing
- High Medium Low
- Outstanding Proficient Shows Potential
41Step 3 Add Criteria
- Create descriptions for each level of performance
for each criteria in the cells of the rubric - Bullet points
- Paragraphs
- Write these criteria in the cells of the sheet
provided
42Step 4 Assign Points
- Assign points for each level of performance
- Can use either of the following
- Discrete values (5, 4, 3, 2, 1)
- Ranges (10-9) for each level
- Indicate the point value on the sheet provided
- Normally placed with the scale
43Step 5 Set Component Weights
- Allows for different levels of importance
- Spelling/grammar more or less important than
content? - Determine if weights are necessary for your
rubric - Assign weights accordingly
- See example on page 17-18 of the handout
44Step 6 Assess the Rubric
- Assess your rubric using a metarubric
- See examples on page 11-15 of your handout
- Conduct a peer review
- Ask one or two other instructors to review your
rubric - Provide time for student review
- Allow students to ask questions and make comments
45Group Project
- Trade rubrics with another group
- Assess the rubric using a metarubric from page
11-15
46Discussion
- What metarubric(s) did you use? Why?
- What did you see on the other teams rubric that
you liked? - Could you understand the assignment easily by
reviewing the rubric?
47Step 7 Implement and refine
- Refine your rubric based on feedback from other
instructors and students - Make notes each time you use the rubric for
continuous improvement purposes - Share with others
48Rubric Reliability Validity
- Reliability
- the likelihood that a given measurement
procedure will yield the same description of a
given phenomena if the measurement is repeated. - Validity
- the extent to which a specific measurement
provides data that relate to commonly accepted
meanings of a particular concept. - Babbie, 1986
49Reliability Requires
- Instructor should reach same conclusion each time
- Different instructors should reach similar
conclusion (interrater reliability)
50Interrater Reliability
- Independently score a set of student samples
- Review responses for consistent and inconsistent
responses - Discuss and reconcile inconsistencies
- Repeat with second group of samples
- Maki, 2004, p. 127
51Validity Requires
- Reliability
- Comprehensiveness
- Cover all outcomes
- Economy
- Space is usually limited, so be selective about
what goes into the rubric - Balanced scoring and weighting
52Discussion and Questions