Online Learning and Planning in Resource Conservation Games
暂无分享,去创建一个
Milind Tambe | Albert Xin Jiang | William B. Haskell | Yundi Qian | Milind Tambe | A. Jiang | W. Haskell | Yundi Qian
[1] Joel Veness,et al. Monte-Carlo Planning in Large POMDPs , 2010, NIPS.
[2] Sarit Kraus,et al. Playing games for security: an efficient exact algorithm for solving Bayesian Stackelberg games , 2008, AAMAS.
[3] Trey Smith,et al. Probabilistic planning for robotic exploration , 2007 .
[4] Vincent Conitzer,et al. Learning and Approximating the Optimal Strategy to Commit To , 2009, SAGT.
[5] Milind Tambe,et al. Online planning for optimal protector strategies in resource conservation games , 2014, AAMAS.
[6] O. H. Brownlee,et al. ACTIVITY ANALYSIS OF PRODUCTION AND ALLOCATION , 1952 .
[7] Milind Tambe,et al. A unified method for handling discrete and continuous uncertainty in Bayesian Stackelberg games , 2012, AAMAS.
[8] Robert J. Aumann,et al. Repeated Games with Incomplete Information , 1995 .
[9] Csaba Szepesvári,et al. Bandit Based Monte-Carlo Planning , 2006, ECML.
[10] Peter Dayan,et al. Q-learning , 1992, Machine Learning.
[11] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.
[12] Milind Tambe. Security and Game Theory: EFFICIENT ALGORITHMS FOR MASSIVE SECURITY GAMES , 2011 .
[13] Milind Tambe,et al. Security and Game Theory - Algorithms, Deployed Systems, Lessons Learned , 2011 .
[14] Vincent Conitzer,et al. AWESOME: A general multiagent learning algorithm that converges in self-play and learns a best response against stationary opponents , 2003, Machine Learning.
[15] Vladik Kreinovich,et al. Security games with interval uncertainty , 2013, AAMAS.
[16] David Hsu,et al. SARSOP: Efficient Point-Based POMDP Planning by Approximating Optimally Reachable Belief Spaces , 2008, Robotics: Science and Systems.
[17] Gerald Tesauro,et al. Playing repeated Stackelberg games with unknown opponents , 2012, AAMAS.