Title: AutoTutor: Integrating Learning with Emotions
1AutoTutor Integrating Learning with Emotions
- Art Graesser, Bethany McDaniel and Sidney DMello
2Overview
- Project goals (Art)
- Methods of data collection (Bethany)
- Affective states during learning (Sidney)
- Occurrence of affective states
- Inter-judge agreement
- Detection of Ekmans facial actions (Bethany)
- Dialogue acts and emotions (Sidney)
3AutoTutor Highlights
- Art Graesser (PI)
- Zhiqiang Cai
- Stan Franklin
- Barry Gholson
- Max Louwerse
- Natalie Person
- Roz Picard (MIT)
- Vasile Rus
- Patrick Chipman
- Scotty Craig
- Sidney DMello
- Tanner Jackson
- Brandon King
- Bethany McDaniel
- Jeremiah Sullins
- Kristy Tapp
- Adam Wanderman
- Amy Witherspoon
- Learn by conversation in natural language
- Subject matters
- Computer literacy
- Conceptual physics
- Critical thinking (now)
- Tracks and adapts to cognition, emotions, and
abilities of learner. - Improves learning nearly a letter grade compared
to reading textbooks (.8
sigma effect size)
4AutoTutor 1998
(Graesser, Wiemer-Hastings, Wiemer-Hastings,
Kreuz, 1999)
5(No Transcript)
6Expectations and misconceptions in Sun Earth
problem
- EXPECTATIONS
- The sun exerts a gravitational force on the
earth. - The earth exerts a gravitational force on the
sun. - The two forces are a third-law pair.
- The magnitudes of the two forces are the same.
- MISCONCEPTIONS
- Only the larger object exerts a force.
- The force of earth on sun may be less than that
of sun on earth.
7Expectation and Misconception-Tailored Dialog
(Pervasive in AutoTutor unskilled human tutors)
- Tutor asks question that requires explanatory
reasoning - Student answers with fragments of information,
distributed over multiple turns - Tutor analyzes the fragments of the explanation
- Compares to a list of expectations (good
sentences) - Compares to a list of misconceptions (bad
answers) - Tutor posts goals performs dialog acts (hints,
prompts) to improve explanation - Fills in missing expectations (one at a time)
- Corrects expected misconceptions (immediately)
- Tutor handles periodic sub-dialogues
- Student questions
- Student meta-communicative acts
(e.g., What did you say?)
8Dialog Moves
- Positive immediate feedback Yeah Right!
- Neutral immediate feedback Okay Uh huh
- Negative immediate feedback No Not quite
- Pump for more information What else?
- Hint What about the earths gravity?
- Prompt for specific information The earth
exerts a gravitational force on what? - Assert The earth exerts a gravitational force
on the sun. - Correct The smaller object also exerts a force.
- Repeat So, once again,
- Summarize So to recap,
- Answer student question
9Hint-Prompt-Assertion Cycles to Cover One
Expectation
Hint
- Cycle fleshes out one expectation at a time
- Exit cycle when
- LSA-cosine(S, E ) gt T
- S student input
- E expectation
- T threshold
Prompt
Assertion
Hint
Prompt
Assertion
10How might AutoTutor be Responsive to the
Learners Affect States
- If learner is frustrated, then AutoTutor gives a
hint. - If bored, then some engaging razzle dazzle
- If flow/absorbed, then lay low
- If confused, then intelligently manage optimal
confusion
11AutoTutor tracking learners emotions
(Memphis team MIT team with Roz Picard)
Confusion
Excitement
Boredom
Flow
Eureka
Frustration
12- Methods of data collection
13Visual IBM Blue eyes camera
Posture Body Pressure Measurement System
AutoTutor
Pressure force sensitive mouse and keyboard
AutoTutor text dialog
14Facial ExpressionsThe IBM Blue Eyes Camera
Red Eye Effect
IBM Blue Eyes Camera
Eyebrow Templates
15Posture PatternsThe Body Pressure Measurement
System
16Gold Standard Study
- Session one
- Participants (N28) interact with AutoTutor
- Collect data with sensors
- BPMS
- Blue eyes camera
- AutoTutor logs
- Participants view their video and give ratings
- Session two (one week later)
- Participants view another participants video and
give affect indications every 20 seconds - Expert judges (N2) give affect ratings
- Learning measures - 32 multiple choice questions
(pretest-posttest)
17Reliability of Emotion Judgments
18Interrater Reliability Judges
- F(5,135) 38.94, MSe .046, p lt .01
- Trained Judges gt All Raters,
- Self- Peer lt All Raters
- Peer-Judge 1 gt Peer-Judge 2
- Peer-Judge 1 gt Self-Judge 1
19Interrater ReliabilityEmotions
F(6,162) 23.13, MSe .058, p lt .01 Delight gt
Confusion gt Boredom gt (Flow Frustration
Neutral Surprise)
20Interrater ReliabilityJudgments Type
21Number of Observations
22Interrater ReliabilityTrained Judge Recoding
Overall Kappa .49 N
1133
23Comparison with other Affect Judgment Efforts
24Emotion Distribution
25Conclusions
- Trained judges who are experienced in coding
facial actions provide affective judgments that
are more reliable and that match the learners
self reports than the judgments of untrained
peers. - The judgments by peers have very little
correspondence to the self reports of learners. - Third, an emotion labeling task is more difficult
if judges are asked to make emotion ratings at
regularly polled timestamps, when compared to
spontaneous judgments. - Different types of emotions are elicited at
different judgment points.
26- Detection of Ekmans Facial Actions
27Detection of Ekmans Facial Actions
- Facial Action Coding System
- (Ekman, 2003 Ekman Friesen, 1978)
- Each Action Unit (AU) represents muscular
activity which produces a change in facial
appearance
Neutral
AU 1 Inner Brow Raise
AU 7 Lid Tightener
28Discriminability of Action Units
- How well each AU can be detected
- Motion How detectable the change is from the
Neutral position. - Edge How clearly defined a line or object on the
face is in relation to the surrounding area. - Texture The level of graininess for the general
area. The degree of variation of the intensity of
the surface, quantifying properties such as
smoothness, coarseness and regularity.
29 Grading Discriminability
- 2 expert judges
- trained on the Facial Action Coding System
- Rate Motion, Edge, and Texture on a 1-6 scale
- 1Very Difficult, 6Very Easy
- Averaged the 3 scores for each expert
- Score between 3-18
- 3Very Difficult to detect, 18 Very Easy to
detect
30Probabilities of the occurrence of AUs across
emotions and discriminability
31Probabilities of the occurrence of AUs across
emotions and discriminability
32Probabilities of the occurrence of AUs across
emotions and discriminability
33Probabilities of the occurrence of AUs across
emotions and discriminability
34Facial Action Coding System
Neutral
AU 1 Inner Brow Raise
35Facial Action Coding System
Neutral
AU 2 Outer Brow Raise
36Facial Action Coding System
Neutral
AU 4 Brow Lower
37Facial Action Coding System
Neutral
AU 7 Lid Tightener
38Facial Action Coding System
Neutral
AU 12 Lip corner puller
39Facial Action Coding System
Neutral
AU 14 Dimpler
40Predicting Emotions from Conversational Cues
41Auto Tutors Text Dialog
Student answers
LSA matches
42AUTO TUTORS CONVERSATIONAL CHANNELS
TEMPORAL
VERBOSITY
ANS. QUALITY
ADVANCER
FEEDBACK
LOCAL GOOD LOCAL BAD GLOBAL GOOD GLOBAL BAD DEL
LCL GOOD DEL LCL BAD DEL GLG GOOD DEL GLG BAD
PUMP HINT PROMPT CORRECTION ASSERTION SUMMARY
NEGATIVE NEUTRAL NEG NEUTRAL NEUTRAL POS POSITIVE
REAL TIME SUBTOPIC TURN RESPONSE
WORDS CHARS SPEECH ACT
DIRECTNESS
FEEDBACK
43Collinearity Analysis
44Multiple Regression Analyses
p lt .10 p lt .05
45Predictors of Affective States
46How might AutoTutor be Responsive to the
Learners Affect States
- If learner is frustrated, then AutoTutor gives a
hint. - If bored, then some engaging razzle dazzle
- If flow/absorbed, then lay low
- If confused, then intelligently manage optimal
confusion