Decentralized MDPs with sparse interactions
暂无分享,去创建一个
[1] V. Kaul,et al. Planning , 2012 .
[2] Kevin Leyton-Brown,et al. Action-Graph Games , 2011, Games Econ. Behav..
[3] U. Rieder,et al. Markov Decision Processes , 2010 .
[4] Martin Allen,et al. Complexity of Decentralized Control: Special Cases , 2009, NIPS.
[5] Milind Tambe,et al. Exploiting Coordination Locales in Distributed POMDPs via Social Model Shaping , 2009, ICAPS.
[6] Dimitri P. Bertsekas,et al. Neuro-Dynamic Programming , 2009, Encyclopedia of Optimization.
[7] Martin Allen,et al. Agent interactions in decentralized environments , 2009 .
[8] Martin Allen,et al. Interactions in Decentralized Environments , 2009 .
[9] S. Simić. On a global upper bound for Jensen's inequality , 2008 .
[10] Francisco S. Melo,et al. Interaction-driven Markov games for decentralized multiagent planning under uncertainty , 2008, AAMAS.
[11] Shlomo Zilberstein,et al. Formal models and algorithms for decentralized decision making under uncertainty , 2008, Autonomous Agents and Multi-Agent Systems.
[12] Kevin Leyton-Brown,et al. Computing Pure Nash Equilibria in Symmetric Action Graph Games , 2007, AAAI.
[13] Manuela M. Veloso,et al. Exploiting factored representations for decentralized execution in multiagent teams , 2007, AAMAS '07.
[14] Manuela Veloso,et al. Execution-time communication decisions for coordination of multi-agent teams , 2007 .
[15] David S. Leslie,et al. Generalised weakened fictitious play , 2006, Games Econ. Behav..
[16] Kevin Leyton-Brown,et al. A Polynomial-Time Algorithm for Action Graph Games , 2006, AAAI.
[17] Francisco S. Melo,et al. Transition Entropy in Partially Observable Markov Decision Processes , 2006, IAS.
[18] Milind Tambe,et al. Hybrid BDI-POMDP Framework for Multiagent Teaming , 2011, J. Artif. Intell. Res..
[19] Victor R. Lesser,et al. Analyzing myopic approaches for multi-agent communication , 2005, IEEE/WIC/ACM International Conference on Intelligent Agent Technology.
[20] Richard S. Sutton,et al. Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.
[21] Nikos A. Vlassis,et al. Utile Coordination: Learning Interdependencies Among Cooperative Agents , 2005, CIG.
[22] P. J. Gmytrasiewicz,et al. A Framework for Sequential Planning in Multi-Agent Settings , 2005, AI&M.
[23] Victor R. Lesser,et al. Decentralized Markov decision processes with event-driven interactions , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..
[24] Jeff G. Schneider,et al. Approximate solutions for partially observable stochastic games with common payoffs , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..
[25] Kevin Leyton-Brown,et al. Computing Nash Equilibria of Action-Graph Games , 2004, UAI.
[26] Nikos A. Vlassis,et al. Sparse cooperative Q-learning , 2004, ICML.
[27] Claudia V. Goldman,et al. Solving Transition Independent Decentralized Markov Decision Processes , 2004, J. Artif. Intell. Res..
[28] Claudia V. Goldman,et al. Decentralized Control of Cooperative Systems: Categorization and Complexity Analysis , 2004, J. Artif. Intell. Res..
[29] Claudia V. Goldman,et al. Transition-independent decentralized markov decision processes , 2003, AAMAS '03.
[30] D. Aberdeen,et al. A ( Revised ) Survey of Approximate Methods for Solving Partially Observable Markov Decision Processes , 2003 .
[31] Milind Tambe,et al. The Communicative Multiagent Team Decision Problem: Analyzing Teamwork Theories and Models , 2011, J. Artif. Intell. Res..
[32] Shobha Venkataraman,et al. Context-specific multiagent coordination and planning with factored MDPs , 2002, AAAI/IAAI.
[33] Michael L. Littman,et al. Graphical Models for Game Theory , 2001, UAI.
[34] Sridhar Mahadevan,et al. Hierarchical multi-agent reinforcement learning , 2001, AGENTS '01.
[35] Carlos Guestrin,et al. Multiagent Planning with Factored MDPs , 2001, NIPS.
[36] Neil Immerman,et al. The Complexity of Decentralized Control of Markov Decision Processes , 2000, UAI.
[37] Anne Condon,et al. On the Undecidability of Probabilistic Planning and Infinite-Horizon Partially Observable Markov Decision Problems , 1999, AAAI/IAAI.
[38] Manuela M. Veloso,et al. Team-partitioned, opaque-transition reinforcement learning , 1999, AGENTS '99.
[39] Kee-Eung Kim,et al. Solving Very Large Weakly Coupled Markov Decision Processes , 1998, AAAI/IAAI.
[40] Craig Boutilier,et al. The Dynamics of Reinforcement Learning in Cooperative Multiagent Systems , 1998, AAAI/IAAI.
[41] Leslie Pack Kaelbling,et al. Planning and Acting in Partially Observable Stochastic Domains , 1998, Artif. Intell..
[42] Richard S. Sutton,et al. Introduction to Reinforcement Learning , 1998 .
[43] Manuela M. Veloso,et al. Team-Partitioned, Opaque-Transition Reinforced Learning , 1998, RoboCup.
[44] A. Cassandra,et al. Exact and approximate algorithms for partially observable markov decision processes , 1998 .
[45] Satinder P. Singh,et al. How to Dynamically Merge Markov Decision Processes , 1997, NIPS.
[46] Munindar P. Singh,et al. Readings in agents , 1997 .
[47] Craig Boutilier,et al. Planning, Learning and Coordination in Multiagent Decision Processes , 1996, TARK.
[48] Leslie Pack Kaelbling,et al. Learning Policies for Partially Observable Environments: Scaling Up , 1997, ICML.
[49] Anthony R. Cassandra,et al. Optimal Policies for Partially Observable Markov Decision Processes , 1994 .
[50] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[51] Ming Tan,et al. Multi-Agent Reinforcement Learning: Independent versus Cooperative Agents , 1997, ICML.
[52] C. Watkins. Learning from delayed rewards , 1989 .
[53] John N. Tsitsiklis,et al. The Complexity of Markov Decision Processes , 1987, Math. Oper. Res..
[54] Edward J. Sondik,et al. The Optimal Control of Partially Observable Markov Processes over a Finite Horizon , 1973, Oper. Res..
[55] L. Shapley,et al. Stochastic Games* , 1953, Proceedings of the National Academy of Sciences.