Augmented Reality systems come with many benefits derived by co-locating information with a user's environment through the use of one or more output modalities such as visual, auditory and tactile. In the case of future human spaceflight programs involving deep space missions, the ground operations infrastructure currently utilized for support of LEO missions will be less accessible or unavailable. This will place far higher emphasis on the importance of automated and intelligent tools for tasking, advising, and monitoring autonomous crew activity. Augmented Reality systems will play a key role in achieving crew autonomy. With this fusion of real and virtual perception, however, comes challenges to ensure that the information is presented in a way that the user can effective consume it meet the goals of the situation. The proposed innovation is a configurable and extensible Augmented Reality Adaptable Information Manager (AR-AIM) that provides an infrastructure for integration of disparate sensor inputs, task definitions, consideration factors, prioritization algorithms, and output modalities to achieve adaptive augmented reality. Specifically, it will include a diverse combination of capabilities needed to make user interfaces that dynamically change to increase user precision and reduce stress on the user caused by the system. By being implemented as a reusable library, the AR-AIM controller suite can be leveraged across many systems, use cases, and domains. Features will include: - FOV Management - Occlusion Engine - Optimal Text Placement - Complexity Scaling - Multiple Modality Support - Data Interfaces - User Contextual Adaptations - Stress predictor - Multi-user Integration
More »