Self-adjusting autonomous systems (SAS) are spreading from well-defined control activities, such as manufacturing, to complex activities with multi-faceted human interactions and decision making, such as those involved in piloting an aircraft. SAS’ ability to solve large problems of certain types far exceeds that of humans’: problems with millions of variables and constraints are tractable for machines, while human decision making is far more limited in scope. However, an automated solution is only as good as the problem statement, including its completeness. Until SAS are proven and perceived to be as or more adaptable than humans, and resilient in the face of unanticipated (and therefore not included in decision making models) faults and variable conditions, humans will have to remain in ultimate control of decision making, while supported by machine-based information and advice. H-M interaction in many domains has numerous well-known and well-documented difficulties, including lack or excess of trust, both of which can lead to serious problems, especially in time-critical and safety-critical situations where human decision makers quickly become overwhelmed with information. In these situations, humans become either reluctant to take advice from machines or lose situational awareness and basic skills via overreliance on machines. Our overarching objective is to develop a concept and an associated software-enabled mechanism that will support real-time decision making in time-critical and safety-critical activities.More »
This project benefits all NASA's missions (airspace, aerospace, space, and science) that involve human-machine interactions.More »
|Organizations Performing Work||Role||Type||Location|
|Langley Research Center (LaRC)||Lead Organization||NASA Center||Hampton, VA|
A final report document may be available for this project. If you would like to request it, please contact us.