暂无分享,去创建一个
Hankz Hankui Zhuo | Junge Zhang | Haiyin Piao | Deheng Ye | Chao Yu | Zifan Wu | H. Zhuo | Deheng Ye | Junge Zhang | Zifan Wu | Haiyin Piao | Chao Yu
[1] Chongjie Zhang,et al. Convergence of Multi-Agent Learning with a Finite Step Size in General-Sum Games , 2019, AAMAS.
[2] Guy Lever,et al. Value-Decomposition Networks For Cooperative Multi-Agent Learning Based On Team Reward , 2018, AAMAS.
[3] Mykel J. Kochenderfer,et al. Cooperative Multi-agent Control Using Deep Reinforcement Learning , 2017, AAMAS Workshops.
[4] Yung Yi,et al. QTRAN: Learning to Factorize with Transformation for Cooperative Multi-Agent Reinforcement Learning , 2019, ICML.
[5] Pablo Hernandez-Leal,et al. A Survey of Learning in Multiagent Environments: Dealing with Non-Stationarity , 2017, ArXiv.
[6] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[7] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[8] John Langford,et al. Approximately Optimal Approximate Reinforcement Learning , 2002, ICML.
[9] Shimon Whiteson,et al. Counterfactual Multi-Agent Policy Gradients , 2017, AAAI.
[10] Shimon Whiteson,et al. QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning , 2018, ICML.
[11] Peter A. Beling,et al. Value-Decomposition Multi-Agent Actor-Critics , 2021, AAAI.
[12] Anamika Sharma,et al. A distributed reinforcement learning based sensor node scheduling algorithm for coverage and connectivity maintenance in wireless sensor network , 2020, Wirel. Networks.
[13] Shimon Whiteson,et al. Learning to Communicate with Deep Multi-Agent Reinforcement Learning , 2016, NIPS.
[14] Filippos Christianos,et al. Dealing with Non-Stationarity in Multi-Agent Deep Reinforcement Learning , 2019, ArXiv.
[15] Beining Han,et al. Off-Policy Multi-Agent Decomposed Policy Gradients , 2020, ICLR.
[16] Herke van Hoof,et al. Addressing Function Approximation Error in Actor-Critic Methods , 2018, ICML.
[17] Yi Wu,et al. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.
[18] Sergey Levine,et al. High-Dimensional Continuous Control Using Generalized Advantage Estimation , 2015, ICLR.
[19] Shimon Whiteson,et al. The StarCraft Multi-Agent Challenge , 2019, AAMAS.
[20] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[21] Robert Babuska,et al. A Survey of Actor-Critic Reinforcement Learning: Standard and Natural Policy Gradients , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).
[22] Frans A. Oliehoek,et al. A Concise Introduction to Decentralized POMDPs , 2016, SpringerBriefs in Intelligent Systems.
[23] V. Climenhaga. Markov chains and mixing times , 2013 .
[24] Xin Wang,et al. Distributed Multiagent Coordinated Learning for Autonomous Driving in Highways Based on Dynamic Coordination Graphs , 2020, IEEE Transactions on Intelligent Transportation Systems.
[25] Dorian Kodelja,et al. Multiagent cooperation and competition with deep reinforcement learning , 2015, PloS one.
[26] Craig Boutilier,et al. The Dynamics of Reinforcement Learning in Cooperative Multiagent Systems , 1998, AAAI/IAAI.
[27] Pan Zhou,et al. Multi-Agent Deep Reinforcement Learning for Urban Traffic Light Control in Vehicular Networks , 2020, IEEE Transactions on Vehicular Technology.
[28] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[29] Hao Wu,et al. Mastering Complex Control in MOBA Games with Deep Reinforcement Learning , 2019, AAAI.
[30] Shimon Whiteson,et al. MAVEN: Multi-Agent Variational Exploration , 2019, NeurIPS.
[31] Qiang Fu,et al. Towards Playing Full MOBA Games with Deep Reinforcement Learning , 2020, NeurIPS.
[32] Haibo He,et al. Multi-Agent Trust Region Policy Optimization , 2020, IEEE transactions on neural networks and learning systems.
[33] Yu Wang,et al. The Surprising Effectiveness of PPO in Cooperative Multi-Agent Games , 2021, NeurIPS.
[34] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.