Abstract
While there is value in thinking about forms of automation failure, the primary role of human factors is to specify how the interface conveys automation state and behavior: when the automation fails and when the automation “performs as designed” but still results in safety events. How do we make automation an effective crewmember? I describe what the flightcrew, ideally, should understand about autoflight state or behavior, using aviation examples to illustrate how that information was not successfully conveyed to the flightcrew, leaving the flightcrew with an incomplete and/or incorrect understanding. Specifically, the interface needs to address automation state, its current targets or objectives, its limitations in achieving those targets, whether it is approaching an operational boundary, data validity, broader checks on what is operationally reasonable, and how to intervene.
Talk to us
Join us for a 30 min session where you can share your feedback and ask us any queries you have
More From: Journal of Cognitive Engineering and Decision Making
Disclaimer: All third-party content on this website/platform is and will remain the property of their respective owners and is provided on "as is" basis without any warranties, express or implied. Use of third-party content does not indicate any affiliation, sponsorship with or endorsement by them. Any references to third-party content is to identify the corresponding services and shall be considered fair use under The CopyrightLaw.