Effective automation is critical in achieving the capacity and safety goals of the Next Generation Air Traffic System. Unfortunately creating integration and validation tools for such automation is difficult as the interactions between automation and their human counterparts is complex and unpredictable. This validation becomes even more difficult as we integrate wide-reaching technologies that affect the behavior of different decision makers in the system such as pilots, controllers and airlines. While overt short-term behavior changes can be explicitly modeled with traditional agent modeling systems, subtle behavior changes caused by the integration of new technologies may snowball into larger problems and be very hard to detect. To overcome these obstacles, we show how integration of new technologies can be validated by learning behavior models based on goals. In this framework, human participants are not modeled explicitly. Instead, their goals are modeled and through reinforcement learning their actions are predicted. The main advantage to this approach is that modeling is done within the context of the entire system allowing for accurate modeling of all participants as they interact as a whole. In addition such an approach allows for efficient trade studies and feasibility testing on a wide range of automation scenarios. The goal of this paper is to test that such an approach is feasible. To do this we implement this approach using a simple discrete-state learning system on a scenario where 50 aircraft need to self-navigate using Automatic Dependent Surveillance-Broadcast (ADS-B) information. In this scenario, we show how the approach can be used to predict the ability of pilots to adequately balance aircraft separation and fly efficient paths. We present results with several levels of complexity and airspace congestion.
[1]
Maarten Sierhuis,et al.
Agent-Based Modeling of Collaboration and Work Practices Onboard the International Space Station
,
2002,
SSRN Electronic Journal.
[2]
D. Stahl,et al.
On Players' Models of Other Players: Theory and Experimental Evidence
,
1995
.
[3]
Michael I. Jordan,et al.
Reinforcement Learning Algorithm for Partially Observable Markov Decision Problems
,
1994,
NIPS.
[4]
Russell Bent,et al.
Counter-Factual Reinforcement Learning: How to Model Decision-Makers That Anticipate the Future
,
2012,
Decision Making and Imperfection.
[5]
Russell Bent,et al.
Cyber-Physical Security: A Game Theory Model of Humans Interacting Over Control Systems
,
2013,
IEEE Transactions on Smart Grid.
[6]
Kagan Tumer,et al.
Distributed agent-based air traffic flow management
,
2007,
AAMAS '07.
[7]
Marco Wiering,et al.
Reinforcement Learning
,
2014,
Adaptation, Learning, and Optimization.
[8]
Guillaume Brat,et al.
Using Game Theoretic Models to Predict Pilot Behavior in NextGen Merging and Landing Scenario
,
2012
.
[9]
David H. Wolpert,et al.
Game Theoretic Modeling of Pilot Behavior During Mid-Air Encounters
,
2011,
ArXiv.
[10]
Thomas B Sheridan,et al.
Final Report and Recommendations for Research on Human-Automation Interaction in the Next Generation Air Transportation System
,
2006
.