暂无分享,去创建一个
[1] Wojciech M. Czarnecki,et al. Grandmaster level in StarCraft II using multi-agent reinforcement learning , 2019, Nature.
[2] A. Colman. Cooperation, psychological game theory, and limitations of rationality in social interaction , 2003, Behavioral and Brain Sciences.
[3] Pieter Abbeel,et al. Emergence of Grounded Compositional Language in Multi-Agent Populations , 2017, AAAI.
[4] Timothy W. Finin,et al. KQML as an agent communication language , 1994, CIKM '94.
[5] Bart De Schutter,et al. Multi-agent Reinforcement Learning: An Overview , 2010 .
[6] Hirotaka Osawa,et al. Constructing a Human-like agent for the Werewolf Game using a psychological model based multiple perspectives , 2016, 2016 IEEE Symposium Series on Computational Intelligence (SSCI).
[7] M. R. Genesereth,et al. Knowledge Interchange Format Version 3.0 Reference Manual , 1992, LICS 1992.
[8] P. D. O'Brien,et al. FIPA — Towards a Standard for Software Agents , 1998 .
[9] Yoshua Bengio,et al. A Neural Probabilistic Language Model , 2003, J. Mach. Learn. Res..
[10] Keith Decker,et al. Mixed-Autonomy Traffic Control with Proximal Policy Optimization , 2019, 2019 IEEE Vehicular Networking Conference (VNC).
[11] Shimon Whiteson,et al. Learning to Communicate with Deep Multi-Agent Reinforcement Learning , 2016, NIPS.
[12] Tetsuro Tanaka,et al. Human-Side Strategies in the Werewolf Game Against the Stealth Werewolf Strategy , 2016, Computers and Games.
[13] Irwin King,et al. Mathematical Modeling of Social Games , 2009, 2009 International Conference on Computational Science and Engineering.
[14] Laura Graesser,et al. Emergent Linguistic Phenomena in Multi-Agent Communication Games , 2019, EMNLP.
[15] Yoav Shoham,et al. Multi-Agent Reinforcement Learning:a critical survey , 2003 .
[16] Qi Sun,et al. Centralized Cooperation for Connected and Automated Vehicles at Intersections by Proximal Policy Optimization , 2020, IEEE Transactions on Vehicular Technology.
[17] Angeliki Lazaridou,et al. Emergent Multi-Agent Communication in the Deep Learning Era , 2020, ArXiv.
[18] Ronald Fagin,et al. Reasoning about knowledge , 1995 .
[19] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[20] Joelle Pineau,et al. On the Pitfalls of Measuring Emergent Communication , 2019, AAMAS.
[21] M. Kuperman,et al. Social games in a social network. , 2000, Physical review. E, Statistical, nonlinear, and soft matter physics.
[22] Dong-Jun Wu,et al. Software agents for knowledge management: coordination in multi-agent supply chains and auctions , 2001, Expert Syst. Appl..
[23] Yaoyiran Li,et al. Emergent Communication Pretraining for Few-Shot Machine Translation , 2020, COLING.
[24] Tomoyuki Kaneko,et al. Application of Deep Reinforcement Learning in Werewolf Game Agents , 2018, 2018 Conference on Technologies and Applications of Artificial Intelligence (TAAI).
[25] Joel Z. Leibo,et al. Autocurricula and the Emergence of Innovation from Social Interaction: A Manifesto for Multi-Agent Intelligence Research , 2019, ArXiv.
[26] Igor Mordatch,et al. Emergent Tool Use From Multi-Agent Autocurricula , 2019, ICLR.
[27] Hirotaka Osawa,et al. Investigation of the effects of nonverbal information on werewolf , 2014, 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE).
[28] Hirotaka Osawa,et al. Werewolf Game Modeling Using Action Probabilities Based on Play Log Analysis , 2016, Computers and Games.
[29] James A. Reggia,et al. Progress in the Simulation of Emergent Communication and Language , 2003, Adapt. Behav..
[30] Dorian Kodelja,et al. Multiagent cooperation and competition with deep reinforcement learning , 2015, PloS one.
[31] Traian Rebedea,et al. Playing Atari Games with Deep Reinforcement Learning and Human Checkpoint Replay , 2016, ArXiv.
[32] Ben Poole,et al. Categorical Reparameterization with Gumbel-Softmax , 2016, ICLR.
[33] Jürgen Schmidhuber,et al. Deep learning in neural networks: An overview , 2014, Neural Networks.
[34] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[35] Chris Martens,et al. Keeping the Story Straight: A Comparison of Commitment Strategies for a Social Deduction Game , 2018, AIIDE.
[36] Sarvapali D. Ramchurn,et al. Agent-based micro-storage management for the Smart Grid , 2010, AAMAS.
[37] George Legrady,et al. Visual Diagnostics for Deep Reinforcement Learning Policy Development , 2018, ArXiv.
[38] Makoto Yokoo,et al. Adopt: asynchronous distributed constraint optimization with quality guarantees , 2005, Artif. Intell..
[39] Guy Lever,et al. Deterministic Policy Gradient Algorithms , 2014, ICML.
[40] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[41] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[42] Stephen Clark,et al. Emergent Communication through Negotiation , 2018, ICLR.
[43] Sebastian Thrun,et al. Multi-robot SLAM with Sparse Extended Information Filers , 2003, ISRR.
[44] Ruslan Salakhutdinov,et al. On Emergent Communication in Competitive Multi-Agent Teams , 2020, AAMAS.
[45] Stergios I. Roumeliotis,et al. Multi-robot SLAM with Unknown Initial Correspondence: The Robot Rendezvous Case , 2006, 2006 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[46] Yi Wu,et al. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.
[47] Bart De Schutter,et al. A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).
[48] Sarah Wiseman,et al. What Data do Players Rely on in Social Deduction Games? , 2019, CHI PLAY.
[49] Nicolas Le Roux,et al. Understanding the impact of entropy on policy optimization , 2018, ICML.
[50] Yee Whye Teh,et al. The Concrete Distribution: A Continuous Relaxation of Discrete Random Variables , 2016, ICLR.
[51] José M. F. Moura,et al. Natural Language Does Not Emerge ‘Naturally’ in Multi-Agent Dialog , 2017, EMNLP.
[52] Yingqian Zhang,et al. Multiagent task allocation in social networks , 2011, Autonomous Agents and Multi-Agent Systems.