Computing Solutions in Infinite-Horizon Discounted Adversarial Patrolling Games

Stackelberg games form the core of a number of tools deployed for computing optimal patrolling strategies in adversarial domains, such as the US Federal Air Marshall Service and the US Coast Guard. In traditional Stackelberg security game models the attacker knows only the probability that each target is covered by the defender, but is oblivious to the detailed timing of the coverage schedule. In many real-world situations, however, the attacker can observe the current location of the defender and can exploit this knowledge to reason about the defender's future moves. We show that this general modeling framework can be captured using adversarial patrolling games (APGs) in which the defender sequentially moves between targets, with moves constrained by a graph, while the attacker can observe the defender's current location and his (stochastic) policy concerning future moves. We offer a very general model of infinite-horizon discounted adversarial patrolling games. Our first contribution is to show that defender policies that condition only on the previous defense move (i.e., Markov stationary policies) can be arbitrarily sub-optimal for general APGs. We then offer a mixed-integer nonlinear programming (MINLP) formulation for computing optimal randomized policies for the defender that can condition on history of bounded, but arbitrary, length, as well as a mixed-integer linear programming (MILP) formulation to approximate these, with provable quality guarantees. Additionally, we present a non-linear programming (NLP) formulation for solving zero-sum APGs. We show experimentally that MILP significantly outperforms the MINLP formulation, and is, in turn, significantly outperformed by the NLP specialized to zero-sum games.

[1]  Jack Edmonds,et al.  Maximum matching and a polyhedron with 0,1-vertices , 1965 .

[2]  Garth P. McCormick,et al.  Computability of global solutions to factorable nonconvex programs: Part I — Convex underestimating problems , 1976, Math. Program..

[3]  J. Filar,et al.  Competitive Markov Decision Processes , 1996 .

[4]  Vincent Conitzer,et al.  Computing the optimal strategy to commit to , 2006, EC '06.

[5]  Playing games for security: an efficient exact algorithm for solving Bayesian Stackelberg games , 2008, AAMAS 2008.

[6]  Sarit Kraus,et al.  Playing games for security: an efficient exact algorithm for solving Bayesian Stackelberg games , 2008, AAMAS.

[7]  Manish Jain,et al.  Computing optimal randomized resource allocations for massive security games , 2009, AAMAS 2009.

[8]  Nicola Basilico,et al.  Extending Algorithms for Mobile Robot Patrolling in the Presence of Adversaries to More Realistic Settings , 2009, 2009 IEEE/WIC/ACM International Joint Conference on Web Intelligence and Intelligent Agent Technology.

[9]  Nicola Basilico,et al.  Leader-follower strategies for robotic patrolling in environments with arbitrary topologies , 2009, AAMAS.

[10]  Manish Jain,et al.  Software Assistants for Randomized Patrol Planning for the LAX Airport Police and the Federal Air Marshal Service , 2010, Interfaces.

[11]  Nicola Basilico,et al.  Asynchronous Multi-Robot Patrolling against Intrusions in Arbitrary Topologies , 2010, AAAI.

[12]  Nicola Basilico,et al.  A Game-Theoretical Model Applied to an Active Patrolling Camera , 2010, 2010 International Conference on Emerging Security Technologies.

[13]  Mark Newman,et al.  Networks: An Introduction , 2010 .

[14]  Vincent Conitzer,et al.  Stackelberg vs. Nash in Security Games: An Extended Investigation of Interchangeability, Equivalence, and Uniqueness , 2011, J. Artif. Intell. Res..

[15]  Nicola Basilico,et al.  Automated Abstractions for Patrolling Security Games , 2011, AAAI.

[16]  Branislav Bosanský,et al.  Computing time-dependent policies for patrolling games with mobile targets , 2011, AAMAS.

[17]  Bo An,et al.  GUARDS and PROTECT: next generation applications of security games , 2011, SECO.

[18]  Vincent Conitzer,et al.  Computing Optimal Strategies to Commit to in Stochastic Games , 2012, AAAI.

[19]  Nicola Basilico,et al.  Patrolling security games: Definition and algorithms for solving large instances with single patroller and single intruder , 2012, Artif. Intell..

[20]  Yevgeniy Vorobeychik,et al.  Computing Stackelberg Equilibria in Discounted Stochastic Games , 2012, AAAI.