暂无分享,去创建一个
Marcin Andrychowicz | Wojciech Zaremba | Pieter Abbeel | Peter Welinder | Lerrel Pinto | P. Abbeel | Marcin Andrychowicz | Wojciech Zaremba | Lerrel Pinto | P. Welinder
[1] Zdravko Balorda,et al. Reducing uncertainty of objects by robot pushing , 1990, Proceedings., IEEE International Conference on Robotics and Automation.
[2] Boris Polyak,et al. Acceleration of stochastic approximation by averaging , 1992 .
[3] Andrew W. Moore,et al. Reinforcement Learning: A Survey , 1996, J. Artif. Intell. Res..
[4] John N. Tsitsiklis,et al. Actor-Critic Algorithms , 1999, NIPS.
[5] Vijay Kumar,et al. Robotic grasping and contact: a review , 2000, Proceedings 2000 ICRA. Millennium Conference. IEEE International Conference on Robotics and Automation. Symposia Proceedings (Cat. No.00CH37065).
[6] Toru Omata,et al. Fast dextrous re-grasping with optimal contact forces and contact sensor-based impedance control , 2001, Proceedings 2001 ICRA. IEEE International Conference on Robotics and Automation (Cat. No.01CH37164).
[7] T. Başar,et al. A New Approach to Linear Filtering and Prediction Problems , 2001 .
[8] Sanjoy Dasgupta,et al. Off-Policy Temporal Difference Learning with Function Approximation , 2001, ICML.
[9] Andrew Y. Ng,et al. Learning omnidirectional path following using dimensionality reduction , 2007, Robotics: Science and Systems.
[10] Yo-Sung Ho,et al. Hole filling method using depth based in-painting for view synthesis in free viewpoint television and 3-D video , 2009, 2009 Picture Coding Symposium.
[11] Brett Browning,et al. A survey of robot learning from demonstration , 2009, Robotics Auton. Syst..
[12] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[13] Siddhartha S. Srinivasa,et al. A Framework for Push-Grasping in Clutter , 2011, Robotics: Science and Systems.
[14] Na-Eun Yang,et al. Depth hole filling using the depth distribution of neighboring regions of depth holes in the Kinect sensor , 2012, 2012 IEEE International Conference on Signal Processing, Communication and Computing (ICSPCC 2012).
[15] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[16] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[17] Trevor Darrell,et al. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation , 2013, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[18] Guy Lever,et al. Deterministic Policy Gradient Algorithms , 2014, ICML.
[19] Tom Schaul,et al. Universal Value Function Approximators , 2015, ICML.
[20] Sergey Levine,et al. Trust Region Policy Optimization , 2015, ICML.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] Javier García,et al. A comprehensive survey on safe reinforcement learning , 2015, J. Mach. Learn. Res..
[23] Jonathan P. How,et al. Efficient reinforcement learning for robots using informative simulated priors , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[24] Shane Legg,et al. Human-level control through deep reinforcement learning , 2015, Nature.
[25] Yuval Tassa,et al. Continuous control with deep reinforcement learning , 2015, ICLR.
[26] Peter I. Corke,et al. Vision-Based Reaching Using Modular Deep Networks: from Simulation to the Real World , 2016, ArXiv.
[27] Wojciech Zaremba,et al. Transfer from Simulation to Real World through Learning Deep Inverse Dynamics Model , 2016, ArXiv.
[28] Vladlen Koltun,et al. Playing for Data: Ground Truth from Computer Games , 2016, ECCV.
[29] Martial Hebert,et al. Improved Learning of Dynamics Models for Control , 2016, ISER.
[30] Michael Milford,et al. Modular Deep Q Networks for Sim-to-real Transfer of Visuo-motor Policies , 2016, ICRA 2017.
[31] Stephen James,et al. 3D Simulation for Robot Arm Control with Deep Q-Learning , 2016, ArXiv.
[32] Demis Hassabis,et al. Mastering the game of Go with deep neural networks and tree search , 2016, Nature.
[33] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[34] Martín Abadi,et al. TensorFlow: Large-Scale Machine Learning on Heterogeneous Distributed Systems , 2016, ArXiv.
[35] Sergey Levine,et al. End-to-End Training of Deep Visuomotor Policies , 2015, J. Mach. Learn. Res..
[36] Abhinav Gupta,et al. The Curious Robot: Learning Visual Representations via Physical Interactions , 2016, ECCV.
[37] Abhinav Gupta,et al. Supersizing self-supervision: Learning to grasp from 50K tries and 700 robot hours , 2015, 2016 IEEE International Conference on Robotics and Automation (ICRA).
[38] Kate Saenko,et al. Learning a visuomotor controller for real world robotic grasping using simulated depth images , 2017, CoRL.
[39] Pieter Abbeel,et al. Probabilistically safe policy transfer , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[40] Kate Saenko,et al. Learning a visuomotor controller for real world robotic grasping using easily simulated depth images , 2017, ArXiv.
[41] Xinyu Liu,et al. Dex-Net 2.0: Deep Learning to Plan Robust Grasps with Synthetic Point Clouds and Analytic Grasp Metrics , 2017, Robotics: Science and Systems.
[42] Marcin Andrychowicz,et al. Hindsight Experience Replay , 2017, NIPS.
[43] Razvan Pascanu,et al. Sim-to-Real Robot Learning from Pixels with Progressive Nets , 2016, CoRL.
[44] Wojciech Zaremba,et al. Domain randomization for transferring deep neural networks from simulation to the real world , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[45] Greg Turk,et al. Preparing for the Unknown: Learning a Universal Policy with Online System Identification , 2017, Robotics: Science and Systems.
[46] Abhinav Gupta,et al. Learning to push by grasping: Using multiple tasks for effective learning , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[47] Yi Wu,et al. Multi-Agent Actor-Critic for Mixed Cooperative-Competitive Environments , 2017, NIPS.
[48] Andrew J. Davison,et al. Transferring End-to-End Visuomotor Control from Simulation to Real World for a Multi-Stage Task , 2017, CoRL.
[49] Balaraman Ravindran,et al. EPOpt: Learning Robust Neural Network Policies Using Model Ensembles , 2016, ICLR.
[50] Abhinav Gupta,et al. Robust Adversarial Reinforcement Learning , 2017, ICML.
[51] Danica Kragic,et al. Deep predictive policy training using reinforcement learning , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[52] Ziyan Wu,et al. DepthSynth: Real-Time Realistic Synthetic Data Generation from CAD Models for 2.5D Recognition , 2017, 2017 International Conference on 3D Vision (3DV).
[53] James Davidson,et al. Supervision via competition: Robot adversaries for learning tasks , 2016, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[54] Sergey Levine,et al. (CAD)$^2$RL: Real Single-Image Flight without a Single Real Image , 2016, Robotics: Science and Systems.
[55] Shimon Whiteson,et al. Counterfactual Multi-Agent Policy Gradients , 2017, AAAI.
[56] Sergey Levine,et al. Learning hand-eye coordination for robotic grasping with deep learning and large-scale data collection , 2016, Int. J. Robotics Res..