Abstract

While there is value in thinking about forms of automation failure, the primary role of human factors is to specify how the interface conveys automation state and behavior: when the automation fails and when the automation “performs as designed” but still results in safety events. How do we make automation an effective crewmember? I describe what the flightcrew, ideally, should understand about autoflight state or behavior, using aviation examples to illustrate how that information was not successfully conveyed to the flightcrew, leaving the flightcrew with an incomplete and/or incorrect understanding. Specifically, the interface needs to address automation state, its current targets or objectives, its limitations in achieving those targets, whether it is approaching an operational boundary, data validity, broader checks on what is operationally reasonable, and how to intervene.

Full Text
Published version (Free)

Talk to us

Join us for a 30 min session where you can share your feedback and ask us any queries you have

Schedule a call