The Body is Not a Given: Joint Agent Policy Learning and Morphology Evolution
暂无分享,去创建一个
Guy Lever | Pushmeet Kohli | Chrisantha Fernando | Siqi Liu | Thore Graepel | Yoram Bachrach | Nicolas Heess | Dylan Banarse | Yoram Bachrach | N. Heess | Siqi Liu | Pushmeet Kohli | Siqi Liu | Guy Lever | T. Graepel | Chrisantha Fernando | D. Banarse | Y. Bachrach | Dylan Banarse
[1] L. Shapley,et al. Stochastic Games* , 1953, Proceedings of the National Academy of Sciences.
[2] L. S. Shapley,et al. 17. A Value for n-Person Games , 1953 .
[3] G. Simpson. THE BALDWIN EFFECT , 1953 .
[4] P. Dubey. On the uniqueness of the Shapley value , 1975 .
[5] A. Elo. The rating of chessplayers, past and present , 1978 .
[6] A. Roth,et al. The Shapley—Shubik and Banzhaf power indices as probabilities , 1988 .
[7] G. Jantzen. 1988 , 1988, The Winning Cars of the Indianapolis 500.
[8] Faruk Gul. Bargaining Foundations of Shapley Value , 1989 .
[9] V. Feltkamp. Alternative axiomatic characterizations of the Shapley and Banzhaf values , 1995 .
[10] Karl Sims,et al. Evolving virtual creatures , 1994, SIGGRAPH.
[11] Irmengard Rauch. 1994 , 1994, Semiotica.
[12] Michael L. Littman,et al. Markov Games as a Framework for Multi-Agent Reinforcement Learning , 1994, ICML.
[13] L. Darrell Whitley,et al. Lamarckian Evolution, The Baldwin Effect and Function Optimization , 1994, PPSN.
[14] Neil Immerman,et al. The Complexity of Decentralized Control of Markov Decision Processes , 2000, UAI.
[15] M. Gribaudo,et al. 2002 , 2001, Cell and Tissue Research.
[16] B. Weber,et al. Evolution and Learning: The Baldwin Effect Reconsidered , 2003 .
[17] Shlomo Zilberstein,et al. Dynamic Programming for Partially Observable Stochastic Games , 2004, AAAI.
[18] Eytan Ruppin,et al. Feature Selection Based on the Shapley Value , 2005, IJCAI.
[19] Ely Porat,et al. Power and stability in connectivity games , 2008, AAMAS.
[20] Evangelos Markakis,et al. Approximating power indices: theoretical and empirical analysis , 2010, Autonomous Agents and Multi-Agent Systems.
[21] Ely Porat,et al. Path disruption games , 2010, AAMAS.
[22] Moshe Tennenholtz,et al. Solving Cooperative Reliability Games , 2011, UAI.
[23] Hod Lipson,et al. Automatic Design and Manufacture of Soft Robots , 2012, IEEE Transactions on Robotics.
[24] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[25] Borys Wróbel,et al. Co-evolution of morphology and control of soft-bodied multicellular animats , 2012, GECCO '12.
[26] Risto Miikkulainen,et al. Open-ended behavioral complexity for evolved virtual creatures , 2013, GECCO '13.
[27] Jan Peters,et al. Reinforcement learning in robotics: A survey , 2013, Int. J. Robotics Res..
[28] Hod Lipson,et al. Unshackling evolution , 2014 .
[29] Pushmeet Kohli,et al. The cost of principles: analyzing power in compatibility weighted voting games , 2014, AAMAS.
[30] Dan Lessin and Don Fussell and Risto Miikkulainen,et al. Adopting Morphology to Multiple Tasks in Evolved Virtual Creatures , 2014, ALIFE.
[31] Yuval Tassa,et al. Learning Continuous Control Policies by Stochastic Value Gradients , 2015, NIPS.
[32] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[33] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[34] Jianfeng Gao,et al. Deep Reinforcement Learning for Dialogue Generation , 2016, EMNLP.
[35] Hod Lipson,et al. On the Difficulty of Co-Optimizing Morphology and Control in Evolved Virtual Creatures , 2016, ALIFE.
[36] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[37] Yair Zick,et al. Algorithmic Transparency via Quantitative Input Influence: Theory and Experiments with Learning Systems , 2016, 2016 IEEE Symposium on Security and Privacy (SP).
[38] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[39] Marc G. Bellemare,et al. Safe and Efficient Off-Policy Reinforcement Learning , 2016, NIPS.
[40] A. E. Eiben,et al. Analysis of Lamarckian evolution in morphologically evolving robots , 2017, ECAL.
[41] Yuval Tassa,et al. Emergence of Locomotion Behaviours in Rich Environments , 2017, ArXiv.
[42] Max Jaderberg,et al. Population Based Training of Neural Networks , 2017, ArXiv.
[43] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[44] Denis Gifford,et al. 1953 , 2018, The British Film Catalogue.
[45] Karol Hausman,et al. Learning an Embedding Space for Transferable Robot Skills , 2018, ICLR.
[46] Oriol Vinyals,et al. Hierarchical Representations for Efficient Architecture Search , 2017, ICLR.
[47] Pieter Abbeel,et al. Continuous Adaptation via Meta-Learning in Nonstationary and Competitive Environments , 2017, ICLR.
[48] Joel Z. Leibo,et al. Human-level performance in first-person multiplayer games with population-based deep reinforcement learning , 2018, ArXiv.
[49] Martin A. Riedmiller,et al. Learning by Playing - Solving Sparse Reward Tasks from Scratch , 2018, ICML.
[50] A. E. Eiben,et al. Analysing the Relative Importance of Robot Brains and Bodies , 2018 .
[51] Sergey Levine,et al. Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations , 2017, Robotics: Science and Systems.