Value-Decomposition Multi-Agent Actor-Critics
暂无分享,去创建一个
[1] Masayoshi Tomizuka,et al. Interaction-aware Decision Making with Adaptive Strategies under Merging Scenarios , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[2] Gorjan Alagic,et al. #p , 2019, Quantum information & computation.
[3] 장윤희,et al. Y. , 2003, Industrial and Labor Relations Terms.
[4] Jun Wang,et al. Multi-Agent Reinforcement Learning , 2020, Deep Reinforcement Learning.
[5] Shimon Whiteson,et al. Counterfactual Multi-Agent Policy Gradients , 2017, AAAI.
[6] Yoshua Bengio,et al. Incorporating Functional Knowledge in Neural Networks , 2009, J. Mach. Learn. Res..
[7] P. Alam,et al. H , 1887, High Explosives, Propellants, Pyrotechnics.
[8] Fei Sha,et al. Actor-Attention-Critic for Multi-Agent Reinforcement Learning , 2018, ICML.
[9] Hoong Chuin Lau,et al. Credit Assignment For Collective Multiagent RL With Global Rewards , 2018, NeurIPS.
[10] Moshe Dor,et al. אבן, and: Stone , 2017 .
[11] Guy Lever,et al. Value-Decomposition Networks For Cooperative Multi-Agent Learning Based On Team Reward , 2018, AAMAS.
[12] Mykel J. Kochenderfer,et al. Cooperative Multi-agent Control Using Deep Reinforcement Learning , 2017, AAMAS Workshops.
[13] G. Fitzgerald,et al. 'I. , 2019, Australian journal of primary health.
[14] Yung Yi,et al. QTRAN: Learning to Factorize with Transformation for Cooperative Multi-Agent Reinforcement Learning , 2019, ICML.
[15] G. G. Stokes. "J." , 1890, The New Yale Book of Quotations.
[16] P. Alam,et al. R , 1823, The Herodotus Encyclopedia.
[17] Miss A.O. Penney. (b) , 1974, The New Yale Book of Quotations.
[18] Stephen C. Adams,et al. Counterfactual Multi-Agent Reinforcement Learning with Graph Convolution Communication , 2020, ArXiv.
[19] Shimon Whiteson,et al. Learning to Communicate with Deep Multi-Agent Reinforcement Learning , 2016, NIPS.
[20] Thomas de Quincey. [C] , 2000, The Works of Thomas De Quincey, Vol. 1: Writings, 1799–1820.
[21] Quoc V. Le,et al. HyperNetworks , 2016, ICLR.
[22] Zongqing Lu,et al. Learning Attentional Communication for Multi-Agent Cooperation , 2018, NeurIPS.
[23] Peter Stone,et al. Deep Recurrent Q-Learning for Partially Observable MDPs , 2015, AAAI Fall Symposia.
[24] Joelle Pineau,et al. TarMAC: Targeted Multi-Agent Communication , 2018, ICML.
[25] Yishay Mansour,et al. Policy Gradient Methods for Reinforcement Learning with Function Approximation , 1999, NIPS.
[26] Shimon Whiteson,et al. Stabilising Experience Replay for Deep Multi-Agent Reinforcement Learning , 2017, ICML.
[27] Alex Graves,et al. Asynchronous Methods for Deep Reinforcement Learning , 2016, ICML.
[28] Drew Wicke,et al. Multiagent Soft Q-Learning , 2018, AAAI Spring Symposia.
[29] P. Alam. ‘A’ , 2021, Composites Engineering: An A–Z Guide.
[30] Dorian Kodelja,et al. Multiagent cooperation and competition with deep reinforcement learning , 2015, PloS one.
[31] Benjamin Y. Choo,et al. Health-aware hierarchical control for smart manufacturing using reinforcement learning , 2017, 2017 IEEE International Conference on Prognostics and Health Management (ICPHM).
[32] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[33] Kagan Tumer,et al. Optimal Payoff Functions for Members of Collectives , 2001, Adv. Complex Syst..
[34] Nicolas Usunier,et al. Episodic Exploration for Deep Deterministic Policies: An Application to StarCraft Micromanagement Tasks , 2016, ArXiv.
[35] Yi Wu,et al. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.
[36] Shimon Whiteson,et al. The StarCraft Multi-Agent Challenge , 2019, AAMAS.
[37] Bart De Schutter,et al. A Comprehensive Survey of Multiagent Reinforcement Learning , 2008, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).
[38] P. Alam. ‘O’ , 2021, Composites Engineering: An A–Z Guide.
[39] Shimon Whiteson,et al. QMIX: Monotonic Value Function Factorisation for Deep Multi-Agent Reinforcement Learning , 2018, ICML.
[40] Danna Zhou,et al. d. , 1840, Microbial pathogenesis.
[41] Peng Peng,et al. Multiagent Bidirectionally-Coordinated Nets: Emergence of Human-level Coordination in Learning to Play StarCraft Combat Games , 2017, 1703.10069.
[42] J. Koenderink. Q… , 2014, Les noms officiels des communes de Wallonie, de Bruxelles-Capitale et de la communaute germanophone.
[43] Yoshua Bengio,et al. Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.
[44] Jun Wang,et al. Multiagent Bidirectionally-Coordinated Nets for Learning to Play StarCraft Combat Games , 2017, ArXiv.
[45] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.