暂无分享,去创建一个
Sergey Levine | Claire J. Tomlin | Somil Bansal | Roberto Calandra | S. Levine | C. Tomlin | R. Calandra | Somil Bansal
[1] Harold J. Kushner,et al. A New Method of Locating the Maximum Point of an Arbitrary Multipeak Curve in the Presence of Noise , 1964 .
[2] Jonas Mockus,et al. On Bayesian Methods for Seeking the Extremum , 1974, Optimization Techniques.
[3] Richard S. Sutton,et al. Dyna, an integrated architecture for learning, planning, and reacting , 1990, SGAR.
[4] Shlomo Zilberstein,et al. Models of Bounded Rationality , 1995 .
[5] H. Simon. Models of Bounded Rationality: Empirically Grounded Economic Reason , 1997 .
[6] J. Kocijan,et al. Gaussian process model based predictive control , 2004, Proceedings of the 2004 American Control Conference.
[7] Michael A. Osborne,et al. Gaussian Processes for Global Optimization , 2008 .
[8] Carl E. Rasmussen,et al. Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.
[9] Jan Peters,et al. Model Learning with Local Gaussian Process Regression , 2009, Adv. Robotics.
[10] P. Dayan,et al. States versus Rewards: Dissociable Neural Prediction Error Signals Underlying Model-Based and Model-Free Reinforcement Learning , 2010, Neuron.
[11] Carl E. Rasmussen,et al. PILCO: A Model-Based and Data-Efficient Approach to Policy Search , 2011, ICML.
[12] Jan Peters,et al. Model learning for robot control: a survey , 2011, Cognitive Processing.
[13] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[14] Jan Peters,et al. A Survey on Policy Search for Robotics , 2013, Found. Trends Robotics.
[15] Alan Fern,et al. Using trajectory data to improve bayesian optimization for reinforcement learning , 2014, J. Mach. Learn. Res..
[16] Yunpeng Pan,et al. Probabilistic Differential Dynamic Programming , 2014, NIPS.
[17] Sergey Levine,et al. Learning Neural Network Policies with Guided Policy Search under Unknown Dynamics , 2014, NIPS.
[18] Jonas Buchli,et al. Learning of closed-loop motion control , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[19] Yuval Tassa,et al. Learning Continuous Control Policies by Stochastic Value Gradients , 2015, NIPS.
[20] Carl E. Rasmussen,et al. Gaussian Processes for Data-Efficient Learning in Robotics and Control , 2015, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[21] Peter Englert,et al. Sparse Gaussian process regression for compliant, real-time robot control , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[22] Jan Peters,et al. Bayesian optimization for learning gaits under uncertainty , 2015, Annals of Mathematics and Artificial Intelligence.
[23] Nando de Freitas,et al. Taking the Human Out of the Loop: A Review of Bayesian Optimization , 2016, Proceedings of the IEEE.
[24] Sergey Levine,et al. Continuous Deep Q-Learning with Model-based Acceleration , 2016, ICML.
[25] Sergey Levine,et al. Goal-driven dynamics learning via Bayesian optimization , 2017, 2017 IEEE 56th Annual Conference on Decision and Control (CDC).
[26] Gaurav S. Sukhatme,et al. Combining Model-Based and Model-Free Updates for Trajectory-Centric Reinforcement Learning , 2017, ICML.
[27] Priya L. Donti,et al. Task-based End-to-end Model Learning , 2017, ArXiv.
[28] Ai Poh Loh,et al. Model-based contextual policy search for data-efficient generalization of robot skills , 2017, Artif. Intell..
[29] Priya L. Donti,et al. Task-based End-to-end Model Learning in Stochastic Optimization , 2017, NIPS.
[30] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).