Model identification and optimization for operational simulation

This paper describes initial research to define and demonstrate an integrated set of algorithms for conducting high-level Operational Simulations. In practice, an Operational Simulation would be used during an ongoing military mission to monitor operations, update state information, compare actual versus planned states, and suggest revised alternative Courses of Action. Significant technical challenges to this realization result from the size and complexity of the problem domain, the inherent uncertainty of situation assessments, and the need for immediate answers. Taking a top-down approach, we initially define the problem with respect to high-level military planning. By narrowing the state space we are better able to focus on model, data, and algorithm integration issues without getting sidetracked by issues specific to any single application or implementation. We propose three main functions in the planning cycle: situation assessment, parameter update, and plan assessment and prediction. Situation assessment uses hierarchical Bayes Networks to estimate initial state probabilities. A parameter update function based on Hidden Markov Models then produces revised state probabilities and state transition probabilities - model identification. Finally, the plan assessment and prediction function uses these revised estimates for simulation-based prediction as well as for determining optimal policies via Markov Decision Processes and simulation-optimization heuristics.

[1]  J. Mack THE ENEMY SYSTEM , 1988, The Lancet.

[2]  M. Dresher The Mathematics of Games of Strategy: Theory and Applications , 1981 .

[3]  John B. Moore,et al.  Hidden Markov Models: Estimation and Control , 1994 .

[4]  Timothy E. Busch,et al.  Modeling of air operations for course-of-action determination , 2002, SPIE Defense + Commercial Sensing.

[5]  J. B. Cruz,et al.  Moving horizon Nash strategies for a military air operation , 2002 .

[6]  Lee W. Wagenhals Operational concept for effects-based course-of-action development and evaluation in a war game , 2002, SPIE Defense + Commercial Sensing.

[7]  Yolanda Carson,et al.  Simulation optimization: methods and applications , 1997, WSC '97.

[8]  Frederic A. Miercort,et al.  On weapons scores and force strengths , 1995 .

[9]  D. Rubin,et al.  Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .

[10]  Dimitri P. Bertsekas,et al.  Dynamic Programming and Optimal Control, Two Volume Set , 1995 .

[11]  Douglas A. Popken,et al.  An Investigation of System Identification Techniques for Simulation Model Abstraction , 2000 .

[12]  Douglas A. Popken,et al.  A hybrid system-identification method for forecasting telecommunications product demands , 2002 .

[13]  Michail G. Lagoudakis,et al.  Value Function Approximation in Zero-Sum Markov Games , 2002, UAI.

[14]  J. Pearl Causality: Models, Reasoning and Inference , 2000 .

[15]  Douglas A. Popken Model abstraction results using state-space system identifications , 2000, Defense, Security, and Sensing.

[16]  John M. D. Hill,et al.  Concept-Level Anticipatory Planning , 2002 .

[17]  Paul K. Davis,et al.  Synthetic cognitive modeling of adversaries for effects-based planning , 2002, SPIE Defense + Commercial Sensing.

[18]  J. Surdu,et al.  SIMULATIONS IN COMMAND POSTS OF THE FUTURE , 2001 .

[19]  H. S. Morse,et al.  The DARPA JFACC program: modeling and control of military operations , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).

[20]  John McCarthy,et al.  What Computers Still Can't Do , 1996, Artif. Intell..

[21]  K. Ito,et al.  Stochastic games and inverse Lyapunov methods in air operations , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).

[22]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[23]  Accelerated randomized stochastic optimization , 2003 .

[24]  Leo Breiman,et al.  Classification and Regression Trees , 1984 .

[25]  John N. Tsitsiklis,et al.  Rollout Algorithms for Combinatorial Optimization , 1997, J. Heuristics.

[26]  Alexander H. Levis,et al.  Creating executable models of influence nets with colored Petri nets , 1998, International Journal on Software Tools for Technology Transfer.

[27]  Ivar Jacobson,et al.  The Unified Modeling Language User Guide , 1998, J. Database Manag..

[28]  Alan R. Washburn,et al.  The LP/POMDP marriage: Optimization with imperfect information , 2000 .

[29]  William T. Scherer,et al.  SMG: a new simulation/optimization approach for large-scale problems , 1999, WSC '99.

[30]  Leslie Pack Kaelbling,et al.  Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..

[31]  Alex F. Sisti,et al.  Dynamic situation assessment and prediction (DSAP) , 2003, SPIE Defense + Commercial Sensing.

[32]  W.M. McEneaney,et al.  Stochastic game approach to air operations , 2004, IEEE Transactions on Aerospace and Electronic Systems.

[33]  Timothy E. Busch,et al.  Theoretical underpinnings of predictive battlespace awareness , 2003, SPIE Defense + Commercial Sensing.

[34]  I. Tunay,et al.  Game-theoretic linear quadratic method for air mission control , 2000, Proceedings of the 39th IEEE Conference on Decision and Control (Cat. No.00CH37187).