An updated, advanced head-mounted display is providing test pilots with unprecedented analysis and evaluation capabilities for aircraft and pilot performance, tasks that until now have been largely evaluated subjectively. The Fused Reality system combines real-world images from a video camera with computer-generated virtual images to create a highly immersive environment for complex tasks, such as landing, formation flying, and aerial refueling. Originally developed under a NASA Small Business Innovation Research (SBIR) contract and flight-tested on a Calspan Learjet, this tool has been significantly improved, new tasks and capabilities have been added, and is now a portable, standalone system that can be easily integrated on a wide range of flight test aircraft. Work to date: This project enabled Armstrong researchers to add functionality to the system first developed under an SBIR contract. The Fused Reality tool provides pilots with a high-fidelity in-flight simulator that can be used for training or repeatable handling quality evaluations. A standalone navigation system now provides aircraft state data, eliminating the need to obtain this data from aircraft instrumentation and allowing the system to be installed on virtually any aircraft. In-flight uses have been expanded from aerial refueling to include training tasks for formation flight, runway approach, and landing. The Fused Reality system has been integrated into the National Test Pilot School's Gippsland GA-8 Airvan research aircraft to support flight test and evaluation of the system. Eight functional flights were flown on the NTPS GA-8 Airvan. These flights identified several improvements and system fixes which were corrected prior to the next flight test. In January 2015, 4 evaluation flight of FR were flown by NASA test pilots. The system worked perfectly and the pilots were able to provide handling qualities ratings and comments for the test aircraft. No delays or other system issues were reported by the pilots. It was felt that pilot ratings were representative of the aircraft performance and were not influenced by the Fused Reality System. Looking ahead: Researchers are discussing applications with a diverse set of potential users. The Fused Reality system will be further evaluated in a test program with the USAF Test Pilot School in March 2016. In this test program, the aircraft performance will be rated using traditional real world tasks such as offset landing and formation flight. These two tasks will then be evaluated using similar virtual tasks generated by Fused Reality. It is predicted that the rating will match using both methods further confirming the ability of FR to generate handling qualities tasks that reflect true aircraft performance. Partners: Systems Technology, Inc. and National Test Pilots School Benefits: Portable: The system can be quickly installed on any aircraft. Versatile: Supported tasks include training, pilot/task evaluation, and post-flight analysis of aircraft and pilot performance. Powerful: Pilot workload measures can be computed both in real time and post-flight and then compared and subjectively scored. Applications: Test pilot training Astronaut training with robotic devices for exploration missions Enhancement of unmanned aircraft system ground stations for improved operator situational awareness In-flight aircraft carrier landing training Commercial space astronaut trainingMore »
This technology has potential to benefit funded NASA missions in two distinct ways. For in-flight evaluation for new aircraft configurations and advanced flight controls, Fused Reality generates realistic, immersive and repeatable handling qualities tasks that are safer and more affordable than traditional handling qualities tasks which require additional aircraft. These handling qualities tasks are used by test pilots to evaluate and rate new aircraft configurations and advanced flight control laws. Fused Reality can also be used for astronaut and pilot training at NASA. Currently flying tasks like formation flight, aerial refueling and landing can be taught safely using Fused Reality. In addition, potential astronaut training tasks are also being identified. Fused Reality is also being investigated for remote operation of UAVs and planetary exploration rovers.More »
|Organizations Performing Work||Role||Type||Location|
|Armstrong Flight Research Center (AFRC)||Lead Organization||NASA Center||Edwards, California|
|National Test Pilot School||Supporting Organization||Academia||Mojave, California|
|Systems Technology, Inc.||Supporting Organization||Industry||Hawthorne, California|
High-acceleration spaceflight missions require the use of diaphragms to control propellants. Future comet retrieval or rendezvous, on-orbit servicing, and similar missions cannot rely on small accelerations to control propellant. Modeling of the dynamics of mixed liquid and deformable solid motion remains computationally expensive, therefore simplifications may be beneficial. Purdue University has developed one such simplification for design that relies on a non-dimensional parameter for diaphragm and liquid interactions. In parabolic flight demonstrations, the researchers sought to demonstrate that the non-dimensional parameter can be used to determine whether the propellant system can be easily modeled as a compliant and stiff system or whether the system requires a more complex coupled liquid-structure interaction model. Flights in 2011 produced data inefficiently, leaving the primary question unanswered. Subsequent flights in 2016 aimed to generate new data and scale results to advance the design capability. Researchers expect the data acquired will enable improved modeling choices and advancement of propellant diaphragm technology.