Ben Berman - PowerPoint PPT Presentation

1 / 26
About This Presentation
Title:

Ben Berman

Description:

Cognitive Limitations and Vulnerabilities on the Flight Deck: What Can We Do? Ben Berman NASA Ames Research Center/San Jose State University Key Dismukes – PowerPoint PPT presentation

Number of Views:43
Avg rating:3.0/5.0
Slides: 27
Provided by: humansyst
Category:
Tags: ben | berman

less

Transcript and Presenter's Notes

Title: Ben Berman


1
Cognitive Limitations and Vulnerabilities on the
Flight Deck What Can We Do?
Ben Berman NASA Ames Research Center/San Jose
State University Key Dismukes NASA Ames Research
Center Loukia Loukopoulos NASA Ames Research
Center/San Jose State University Human Factors
in Aviation Conference San Antonio, Texas March
6, 2007
2
Most Airline Accidents Attributed to Crew Error
  • What does this mean?
  • Why do highly skilled pilots make fatal errors?
  • How should we think about the role of errors in
    accidents?
  • Draw upon cognitive science research on skilled
    performance of human operators

3
Broken Bolt vs.Unflipped Switch
  • Physical causes leave characteristic traces
  • No such thing for human performance
  • Why drops out, for good reason
  • Can never know with certainty why accident crew
    made specific errors--but can determine why the
    population of pilots is vulnerable
  • Put why back into the analysis with informed
    perspective

4
Approach
  • Reviewed NTSB reports of the 19 U.S. airline
    accidents between 1991-2001 attributed primarily
    to crew error
  • Followed the approach of NTSB (1994)
  • Asked Why might any airline crew in situation
    of accident crew knowing only what they knew
    be vulnerable?

5
Hindsight Bias
  • Knowing the outcome of an accident flight reveals
    what the crew should have done differently
  • But accident crew does not know the outcome
  • They respond to situation as they perceive it at
    the moment
  • Presumption has to be that more underlies an
    error than lack of skill or motivation

6
Training, experience, personal goals
  • Social/Organizational
  • Influences
  • Formal procedures policies
  • Explicit goals rewards
  • Implicit goals rewards
  • Actual norms for line operations

Immediate demands of situation tasks being
performed
Human cognition characteristics limitations
Crew responses to situation
7
American 903May 12, 1997 West Palm Beach,
Florida
  • Airbus 300/600 in its initial descent for Miami
  • Airplane leveled, slowed, and began a holding
    pattern at 16,000 feet
  • Uncontrolled roll and pitch oscillations steep
    descent recovered with one serious injury/damage
    to the tail that was discovered much later
  • Probable cause The flight crews failure to
    maintain adequate airspeed during level-off which
    led to an inadvertent stall, and their subsequent
    failure to use proper stall recovery techniques
  • Contributing factor The flight crews failure
    to properly use the autothrottle

8
American 903Sequence of Events
  • Autothrottle became disconnected during descent
    from cruise--not clear why, and not noticed or
    remembered by the crew
  • Autopilot captured 16,000 feet and pitched up to
    maintain altitude, so airspeed decreased
  • Vibration and buffeting began that the crew
    attributed to turbulence
  • Airplane stalled and departed controlled flight
  • Crew recovered below 13,000 feet, proceeded to
    safe landing in Miami

9
American 903Dim Cues, Missing Cues, and Hints
  • Autothrottle disconnection did not have salient
    annunciation
  • Dim cue (compared to autopilot, for example)
  • Crews often do not notice unexpected mode changes
  • Autothrottles had tripped off on other flights
  • Hint about accidents from everyday operations
  • Crew might have disconnected autothrottle without
    thinking about it or remembering
  • Automaticity

10
American 903Out of the Loop, Missing Cues, and
the Wrong Procedure
  • Neither pilot noticed the airspeed loss until
    well below maneuvering speed
  • Vulnerability from being out of the active
    control loop
  • Crew expected aircraft to slow, just not that
    much
  • Stall warning did not activate prior to stall
  • Dependence on missing cue
  • Crew performed wind shear recovery rather than
    stall recovery
  • Primed to think about thunderstorms
  • Training for wind shear and min-altitude approach
    to stall recovery
  • Again, absence of timely stall warning

11
American 903Lessons
  • Tendency toward degraded monitoring of normally
    reliable automation is rooted in basic human
    cognitive vulnerabilities
  • Simply cautioning pilots to monitor more
    carefully will not greatly reduce vulnerability,
    by itself.
  • Snowball effect of one error producing overload
    that undermines error-trapping and results in
    more errors
  • Missing cues can be extremely hazardous
  • Concept of reliability on the flight deck
  • Variability of performance
  • Two pilots, redundant procedures, several warning
    systems, and an airplane goal is to keep errors
    from sneaking through all those layers of cheese

12
Each Accident Has Unique Surface Features and
Combinations of Factors
  • Countermeasures to surface features of past
    accidents will not prevent future accidents
  • Must examine deep structure of accidents to find
    common factors

13
Cross-Cutting Factors Contributing to Crew Errors
  • Situations requiring rapid response
  • Challenges of managing concurrent tasks
  • Equipment failure and design flaws
  • Misleading or missing cues normally present
  • Stress
  • Shortcomings in training and/or guidance
  • Social/organizational issues
  • Plan continuation bias

14

Cross-Cutting Factors
  • Situations requiring rapid response
  • Nearly 2/3 of 19 accidents
  • Examples upset attitudes, false stick shaker
    activation after rotation, anomalous airspeed
    indications at rotation, autopilot-induced
    oscillation at Decision Height, pilot-induced
    oscillation during flare
  • Very rare occurrences, but high risk
  • Surprise is a factor
  • Inadequate time to think through situation
  • Automatic response required from pilot

15

Cross-Cutting Factors
  • Challenges of managing concurrent tasks
  • Workload high in some accidents (e.g., Little
    Rock, 1999)
  • Overloaded crews failed to recognize situation
    getting out of hand (snowball effect) losing
    the ability to recognize that they were
    overloaded
  • Monitoring and cross-checking suffered
  • Crews became reactive instead of
    proactive/strategic
  • But adequate time available for all tasks in
    many accidents
  • Inherent cognitive limitations in switching
    attention preoccupation with one task of many
    forgetting to resume interrupted or deferred
    tasks

16

Cross-Cutting Factors
  • Stress
  • Stress is normal physiological/behavioral
    response to threat
  • Acute stress hampers performance
  • Narrows attention (tunneling)
  • Reduces working memory capacity
  • Combination of surprise, stress, time pressure,
    and concurrent task demands can be lethal setup

17

Cross-Cutting Factors
  • Social/Organizational Issues
  • Actual norms may deviate from Flight Operations
    Manual
  • Accident crew judgment decision-making may not
    differ from non-accident crews in similar
    situations
  • Lincoln Lab study Penetration of storm cells on
    approach not uncommon
  • Other flights may have landed or taken off
    without difficulty a minute or two before
    accident flight
  • Little data available on extent to which accident
    crews actions are typical/atypical
  • Competing pressures not often acknowledged
  • Implicit messages from company may conflict with
    formal guidance
  • e.g. on-time performance vs. conservative
    response to ambiguous situations
  • Pilots may not be consciously aware of influence
    of internalized competing goal

18

Cross-Cutting Factors
  • Plan continuation bias (e.g., Burbank, 2000)
  • Unconscious cognitive bias to continue original
    plan in spite of changing conditions
  • Appears stronger as one nears completion of
    activity (e.g., approach to landing)
  • Why are crews reluctant to go-around?
  • Bias may prevent noticing subtle cues indicating
    original conditions have changed
  • Reactive responding is easier than proactive
    thinking
  • Default plan always worked before

19

Implications and CountermeasuresThe Individual
  • Vulnerabilities are inherent in our human
    cognition
  • Limitations, biases
  • Effects extend to error trapping mechanisms
  • Monitoring subject to task shedding under
    workload
  • Checklists automatized, subject to
    slips/omissions
  • Recognize realistic rather than theoretical
    performance of humans in generating errors as
    they work, and in catching errors

20

Implications and CountermeasuresThe Aviation
System
  • Accept the variability of human performance
  • Because someone makes an error does not make them
    deficient or complacent
  • Errors are inevitable
  • Recognize that errors are probabilistic
  • Societys demand for reliability is nearly
    unlimited
  • Focus on net system reliability, made up of
  • Reliability of each error trapping mechanism
  • Independence of error trapping mechanisms to
    prevent the holes in the swiss cheese from lining
    up (e.g., landing checklist initiation cued by
    callout to extend the gear)

21

Implications and CountermeasuresThe Aviation
System (contd)
  • Causal attribution to individual can distract
    attention from underlying cause and prevention
  • Remediation lies with the system
  • Guard against dont do that anymore
    prescriptions
  • No evidence of Bad Apple theory
  • All other remediation requires more difficult
    changes in the system

22
Net System Reliability Okay, what can we do?
  • Treat monitoring and crosschecking as being as
    important as anything else we do
  • Suggestions for training checking
  • Cut down on negative training (e.g., continuing
    unstabilized approaches in sim makes plan
    continuation bias stronger on the line
  • Teach pilots to monitor, give them the
    opportunity to practice in the sim
  • Build at least one realistic challenge and
    decision into every sim session
  • Teach how to manage workload, practice in the sim

23
What can we do?(continued)
  • Procedures design review
  • Match human characteristics and limitations
  • Independence
  • Equipment design review
  • Match human characteristics and limitations
  • Safety Change
  • Heed the hints--the next accident is trying to
    warn you about itself right now
  • AA903, AA1340, CAL1943
  • How to separate the wheat from the chaff?

24
  • Dismukes, R. K., Berman, B.A., Loukopoulos, L.
    L. The Limits of Expertise Rethinking
    Pilot Error and the Causes of Airline Accidents.
  • www.ashgate.com
  • More information on NASA Human Factors Research
  • http//human-factors.arc.nasa.gov/his/flightcognit
    ion/
  • This research was partially funded by NASAs
    Aviation Safety Program and by the FAA (Eleana
    Edens, Program Manager).

25

What else can we do?
26
Write a Comment
User Comments (0)
About PowerShow.com