RIAACT: a robust approach to adjustable autonomy for human-multiagent teams
暂无分享,去创建一个
[1] Milind Tambe,et al. Towards Adjustable Autonomy for the Real World , 2002, J. Artif. Intell. Res..
[2] Martin L. Puterman,et al. Markov Decision Processes: Discrete Stochastic Dynamic Programming , 1994 .
[3] Marek Petrik,et al. An Analysis of Laplacian Methods for Value Function Approximation in MDPs , 2007, IJCAI.
[4] Michael L. Littman,et al. Exact Solutions to Time-Dependent MDPs , 2000, NIPS.
[5] References , 1971 .
[6] Milind Tambe,et al. Exploiting belief bounds: practical POMDPs for personal assistant agents , 2005, AAMAS '05.
[7] Hector J. Levesque,et al. Intention is Choice with Commitment , 1990, Artif. Intell..
[8] Lihong Li,et al. Lazy Approximation for Solving Continuous Finite-Horizon MDPs , 2005, AAAI.
[9] Milind Tambe,et al. A Fast Analytical Algorithm for Solving Markov Decision Processes with Real-Valued Resources , 2007, IJCAI.
[10] Milind Tambe,et al. Using multiagent teams to improve the training of incident commanders , 2006, AAMAS '06.
[11] Reid G. Simmons,et al. Coordinated Multiagent Teams and Sliding Autonomy for Large-Scale Assembly , 2006, Proceedings of the IEEE.
[12] Milind Tambe,et al. Hybrid BDI-POMDP Framework for Multiagent Teaming , 2011, J. Artif. Intell. Res..
[13] U. Rieder,et al. Markov Decision Processes , 2010 .
[14] Jeffrey D. Anderson,et al. Managing autonomy in robot teams: Observations from four experiments , 2007, 2007 2nd ACM/IEEE International Conference on Human-Robot Interaction (HRI).