Planning with continuous resources in agent systems
暂无分享,去创建一个
[1] Milind Tambe,et al. Dangers in Multiagent Rescue using DEFACTO , 2009 .
[2] Milind Tambe,et al. RIAACT: a robust approach to adjustable autonomy for human-multiagent teams , 2008, AAMAS.
[3] Makoto Yokoo,et al. Not all agents are equal: scaling up distributed POMDPs for agent networks , 2008, AAMAS.
[4] Sridhar Mahadevan,et al. Proto-value Functions: A Laplacian Framework for Learning Representation and Control in Markov Decision Processes , 2007, J. Mach. Learn. Res..
[5] Marek Petrik,et al. Anytime Coordination Using Separable Bilinear Programs , 2007, AAAI.
[6] Makoto Yokoo,et al. Letting loose a SPIDER on a network of POMDPs: generating quality guaranteed policies , 2007, AAMAS '07.
[7] Shlomo Zilberstein,et al. Memory-Bounded Dynamic Programming for DEC-POMDPs , 2007, IJCAI.
[8] Milind Tambe,et al. A Fast Analytical Algorithm for Solving Markov Decision Processes with Real-Valued Resources , 2007, IJCAI.
[9] Marek Petrik,et al. An Analysis of Laplacian Methods for Value Function Approximation in MDPs , 2007, IJCAI.
[10] Jianhui Wu,et al. Coordinated Plan Management Using Multiagent MDPs , 2006, AAAI Spring Symposium: Distributed Plan and Schedule Management.
[11] Abdel-Illah Mouaddib,et al. An Iterative Algorithm for Solving Constrained Decentralized Markov Decision Processes , 2006, AAAI.
[12] Sebastian Thrun,et al. Winning the DARPA Grand Challenge , 2006, PKDD.
[13] Mor Harchol-Balter,et al. Closed form solutions for mapping general distributions to quasi-minimal PH distributions , 2006, Perform. Evaluation.
[14] Makoto Yokoo,et al. Winning back the CUP for distributed POMDPs: planning over continuous belief spaces , 2006, AAMAS '06.
[15] Edmund H. Durfee,et al. Symmetric Primal-Dual Approximate Linear Programming for Factored MDPs , 2006, ISAIM.
[16] Edmund H. Durfee,et al. Stationary Deterministic Policies for Constrained MDPs with Multiple Rewards, Costs, and Discount Factors , 2005, IJCAI.
[17] Ronen I. Brafman,et al. Planning with Continuous Resources in Stochastic Domains , 2005, IJCAI.
[18] Abdel-Illah Mouaddib,et al. A polynomial algorithm for decentralized Markov decision processes with temporal constraints , 2005, AAMAS '05.
[19] Milind Tambe,et al. Exploiting belief bounds: practical POMDPs for personal assistant agents , 2005, AAMAS '05.
[20] Makoto Yokoo,et al. Networked Distributed POMDPs: A Synergy of Distributed Constraint Optimization and POMDPs , 2005, IJCAI.
[21] Håkan L. S. Younes. Planning and Execution with Phase Transitions , 2005, AAAI.
[22] Lihong Li,et al. Lazy Approximation for Solving Continuous Finite-Horizon MDPs , 2005, AAAI.
[23] R. Sargent,et al. Mission planning and target tracking for autonomous instrument placement , 2005, 2005 IEEE Aerospace Conference.
[24] John P. Lewis,et al. The DEFACTO System: Coordinating Human-Agent Teams for the Future of Disaster Response , 2005, Multi-Agent Programming.
[25] Shlomo Zilberstein,et al. Dynamic Programming for Partially Observable Stochastic Games , 2004, AAAI.
[26] Makoto Yokoo,et al. Communications for improving policy computation in distributed POMDPs , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..
[27] Sarit Kraus,et al. Towards a formalization of teamwork with resource constraints , 2004, Proceedings of the Third International Joint Conference on Autonomous Agents and Multiagent Systems, 2004. AAMAS 2004..
[28] Milos Hauskrecht,et al. Solving Factored MDPs with Continuous and Discrete Variables , 2004, UAI.
[29] Yishay Mansour,et al. A Sparse Sampling Algorithm for Near-Optimal Planning in Large Markov Decision Processes , 1999, Machine Learning.
[30] Timothy W. McLain,et al. Multiple UAV cooperative search under collision avoidance and limited range communication constraints , 2003, 42nd IEEE International Conference on Decision and Control (IEEE Cat. No.03CH37475).
[31] Michail G. Lagoudakis,et al. Least-Squares Policy Iteration , 2003, J. Mach. Learn. Res..
[32] David V. Pynadath,et al. Taming Decentralized POMDPs: Towards Efficient Policy Computation for Multiagent Settings , 2003, IJCAI.
[33] Claudia V. Goldman,et al. Transition-independent decentralized markov decision processes , 2003, AAMAS '03.
[34] Claudia V. Goldman,et al. Optimizing information exchange in cooperative multi-agent systems , 2003, AAMAS '03.
[35] Makoto Yokoo,et al. An asynchronous complete method for distributed constraint optimization , 2003, AAMAS '03.
[36] Daniel N. Nikovski,et al. Non-Linear Stochastic Control in Continuous State Spaces by Exact Integration in Bellman's Equations , 2003 .
[37] Milind Tambe,et al. The Communicative Multiagent Team Decision Problem: Analyzing Teamwork Theories and Models , 2011, J. Artif. Intell. Res..
[38] David E. Smith,et al. Planning Under Continuous Time and Resource Uncertainty: A Challenge for AI , 2002, AIPS Workshop on Planning for Temporal Domains.
[39] Craig Boutilier,et al. Stochastic dynamic programming with factored representations , 2000, Artif. Intell..
[40] Neil Immerman,et al. The Complexity of Decentralized Control of Markov Decision Processes , 2000, UAI.
[41] Michael L. Littman,et al. Exact Solutions to Time-Dependent MDPs , 2000, NIPS.
[42] Andrew Y. Ng,et al. Policy Invariance Under Reward Transformations: Theory and Application to Reward Shaping , 1999, ICML.
[43] David J. C. Mackay,et al. Introduction to Monte Carlo Methods , 1998, Learning in Graphical Models.
[44] Victor R. Lesser,et al. Designing a Family of Coordination Algorithms , 1997, ICMAS.
[45] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[46] Marcel F. Neuts,et al. Matrix-Geometric Solutions in Stochastic Models , 1981 .
[47] Raymond A. Marie,et al. Calculating equilibrium probabilities for &lgr;(n)/Ck/1/N queues , 1980 .
[48] D. Rubin,et al. Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .
[49] K. Mani Chandy,et al. Approximate Analysis of Central Server Models , 1975, IBM J. Res. Dev..
[50] R. Bellman. Dynamic Programming , 1957, Science.
[51] R. Howard. Dynamic Programming and Markov Processes , 1960 .
[52] D. Cox. A use of complex probabilities in the theory of stochastic processes , 1955, Mathematical Proceedings of the Cambridge Philosophical Society.