Learning Indirect Actions in Complex Domains: Action Suggestions for Air Traffic Control

Providing intelligent algorithms to manage the ever-increasing flow of air traffic is critical to the efficiency and economic viability of air transportation systems. Yet, current automated solutions leave existing human controllers "out of the loop" rendering the potential solutions both technically dangerous (e.g. inability to react to suddenly developing conditions) and politically charged (e.g. role of air traffic controllers in a fully automated system). Instead, this paper outlines a distributed agent-based solution where agents provide suggestions to human controllers. Though conceptually pleasing, this approach introduces two critical research issues. First, the agent actions are now filtered through interactions with other agents, human controllers and the environment before leading to a system state. This indirect action-to-effect process creates a complex learning problem. Second, even in the best case, not all air traffic controllers will be willing or able to follow the agents' suggestions. This partial participation effect will require the system to be robust to the number of controllers that follow the agent suggestions. In this paper, we present an agent reward structure that allows agents to learn good actions in this indirect environment, and explore the ability of those suggestion agents to achieve good system level performance. We present a series of experiments based on real historical air traffic data combined with simulation of air traffic flow around the New York city area. Results show that the agents can improve system-wide performance by up to 20% over that of human controllers alone, and that these results degrade gracefully when the number of human controllers that follow the agents' suggestions declines.

[1]  Kagan Tumer,et al.  Efficient Evaluation Functions for Multi-rover Systems , 2004, GECCO.

[2]  Kai Nagel,et al.  Multi-modal traffic in TRANSIMS , 2001 .

[3]  Kagan Tumer,et al.  Collectives and Design Complex Systems , 2004 .

[4]  Kapil Sheth,et al.  Aggregate Flow Model for Air-Traffic Management , 2004 .

[5]  Kai Nagel,et al.  Towards truly agent-based traffic and mobility simulations , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..

[6]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[7]  Nicholas R. Jennings,et al.  On agent-based software engineering , 2000, Artif. Intell..

[8]  Kagan Tumer,et al.  Regulating air traffic flow with coupled agents , 2008, AAMAS.

[9]  David McNally,et al.  Concept and Laboratory Analysis of Trajectory-Based Automation for Separation Assurance , 2007 .

[10]  Alexandre M. Bayen,et al.  Lagrangian Delay Predictive Model for Sector-Based Air Traffic Flow , 2005 .

[11]  S. Shankar Sastry,et al.  Conflict resolution for air traffic management: a study in multiagent hybrid systems , 1998, IEEE Trans. Autom. Control..

[12]  David Sislák,et al.  Autonomous agents for air-traffic deconfliction , 2006, AAMAS '06.

[13]  James K. Archibald,et al.  A cooperative multi-agent approach to free flight , 2005, AAMAS '05.

[14]  Ana L. C. Bazzan,et al.  Agents in Traffic Modelling - From Reactive to Social Behaviour , 1999, KI.

[15]  Peter Dayan,et al.  Q-learning , 1992, Machine Learning.

[16]  Dimitris Bertsimas,et al.  The Air Traffic Flow Management Problem with Enroute Capacities , 1998, Oper. Res..

[17]  M. Christodoulou,et al.  Nonlinear mixed integer programming for aircraft collision avoidance in free flight , 2004, Proceedings of the 12th IEEE Mediterranean Electrotechnical Conference (IEEE Cat. No.04CH37521).

[18]  Kagan Tumer,et al.  Distributed agent-based air traffic flow management , 2007, AAMAS '07.

[19]  Andres G. Zellweger,et al.  Modeling Distributed Human Decision Making in Traffic Flow Management Operations , 2001 .

[20]  Maarten Sierhuis,et al.  To BDI, or not to BDI: design choices in an agent-based traffic flow management simulation , 2008, SpringSim '08.

[21]  Kapil Sheth,et al.  FACET: Future ATM Concepts Evaluation Tool , 2001 .

[22]  Kagan Tumer,et al.  Optimal Payoff Functions for Members of Collectives , 2001, Adv. Complex Syst..

[23]  Antonio Bicchi,et al.  Conflict resolution problems for air traffic management systems solved with mixed integer programming , 2002, IEEE Trans. Intell. Transp. Syst..

[24]  Peter Dayan,et al.  Technical Note: Q-Learning , 2004, Machine Learning.

[25]  Axel van Lamsweerde,et al.  Learning machine learning , 1991 .

[26]  Frank Dignum,et al.  Achieving cooperation among selfish agents in the air traffic management domain using signed money , 2007, AAMAS '07.

[27]  Craig Bonaceto,et al.  Naturalistic Decision Making in the Air Traffic Control Tower: Combining Approaches to Support Changes in Procedures , 2005 .

[28]  Kagan Tumer,et al.  Efficient Evaluation Functions for Evolving Coordination , 2008, Evolutionary Computation.

[29]  Ana L. C. Bazzan,et al.  Case Studies on the Braess Paradox: Simulating Route Recommendation and Learning in Abstract and Microscopic Models , 2005 .

[30]  Zbigniew Michalewicz,et al.  Handbook of Evolutionary Computation , 1997 .