Robust Solutions in Stackelberg Games : Addressing Boundedly Rational Human Preference Models

Stackelberg games represent an important class of games in which one player, the leader, commits to a strategy and the remaining players, the followers, make their decision with knowledge of the leader’s commitment. Existing algorithms for Bayesian Stackelberg games find optimal solutions while modeling uncertainty over follower types with an a-priori probability distribution. Unfortunately, in real-world applications, the leader may also face uncertainty over the follower’s response which makes the optimality guarantees of these algorithms fail. Such uncertainty arises because the follower’s specific preferences or the follower’s observations of the leader’s strategy may not align with the rational strategy, and it is not amenable to a-priori probability distributions. These conditions especially hold when dealing with human subjects. To address these uncertainties while providing quality guarantees, we propose three new robust algorithms based on mixed-integer linear programs (MILPs) for Bayesian Stackelberg games. A key result of this paper is a detailed experimental analysis that demonstrates that these new MILPs deal better with human responses: a conclusion based on 800 games with 57 human subjects as followers. We also provide run-time results on these MILPs.

[1]  Jean Cardinal,et al.  Pricing of Geometric Transportation Networks , 2009, CCCG.

[2]  Sarit Kraus,et al.  An efficient heuristic approach for security against multiple adversaries , 2007, AAMAS '07.

[3]  B. Stengel,et al.  Leadership with commitment to mixed strategies , 2004 .

[4]  Ariel Orda,et al.  Achieving network optima using Stackelberg routing strategies , 1997, TNET.

[5]  Laurent El Ghaoui,et al.  Robustness in Markov Decision Problems with Uncertain Transition Matrices , 2003, NIPS.

[6]  H. Simon,et al.  Rational choice and the structure of the environment. , 1956, Psychological review.

[7]  Vincent Conitzer,et al.  Computing the optimal strategy to commit to , 2006, EC '06.

[8]  Herbert A. Simon,et al.  The Sciences of the Artificial , 1970 .

[9]  Sarit Kraus,et al.  Playing games for security: an efficient exact algorithm for solving Bayesian Stackelberg games , 2008, AAMAS.

[10]  M. Friedman The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance , 1937 .

[11]  Dimitris Bertsimas,et al.  Robust game theory , 2006, Math. Program..

[12]  Gerald G. Brown,et al.  Defending Critical Infrastructure , 2006, Interfaces.

[13]  A. Rubinstein Modeling Bounded Rationality , 1998 .

[14]  Drew Fudenberg,et al.  Game theory (3. pr.) , 1991 .

[15]  Yoav Shoham,et al.  If multi-agent learning is the answer, what is the question? , 2007, Artif. Intell..

[16]  Fernando Ordóñez,et al.  Robust Wardrop Equilibrium , 2007, NET-COOP.

[17]  R. Selten,et al.  A Generalized Nash Solution for Two-Person Bargaining Games with Incomplete Information , 1972 .