State-Only Imitation Learning for Dexterous Manipulation
暂无分享,去创建一个
Xiaolong Wang | Lerrel Pinto | Jitendra Malik | Ilija Radosavovic | Ilija Radosavovic | J. Malik | Lerrel Pinto | Xiaolong Wang
[1] Jitendra Malik,et al. Learning Navigation Subroutines by Watching Videos , 2019, ArXiv.
[2] Marcin Andrychowicz,et al. Solving Rubik's Cube with a Robot Hand , 2019, ArXiv.
[3] Geoffrey J. Gordon,et al. A Reduction of Imitation Learning and Structured Prediction to No-Regret Online Learning , 2010, AISTATS.
[4] Sham M. Kakade,et al. A Natural Policy Gradient , 2001, NIPS.
[5] Andrew Y. Ng,et al. Pharmacokinetics of a novel formulation of ivermectin after administration to goats , 2000, ICML.
[6] A. Meltzoff. Understanding the Intentions of Others: Re-Enactment of Intended Acts by 18-Month-Old Children. , 1995, Developmental psychology.
[7] S. Chiba,et al. Dynamic programming algorithm optimization for spoken word recognition , 1978 .
[8] Hao Su,et al. State Alignment-based Imitation Learning , 2019, ICLR.
[9] Claude Sammut,et al. A Framework for Behavioural Cloning , 1995, Machine Intelligence 15.
[10] Alec Radford,et al. Proximal Policy Optimization Algorithms , 2017, ArXiv.
[11] Jakub W. Pachocki,et al. Learning dexterous in-hand manipulation , 2018, Int. J. Robotics Res..
[12] Peter Bakker,et al. Robot see, robot do: An overview of robot imitation , 1996 .
[13] Sergey Levine,et al. Deep Dynamics Models for Learning Dexterous Manipulation , 2019, CoRL.
[14] Jitendra Malik,et al. Learning Navigation Subroutines from Egocentric Videos , 2019, CoRL.
[15] Alexei A. Efros,et al. Curiosity-Driven Exploration by Self-Supervised Prediction , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[16] Yannick Schroecker,et al. Imitating Latent Policies from Observation , 2018, ICML.
[17] Sergey Levine,et al. Time-Contrastive Networks: Self-Supervised Learning from Video , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[18] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[19] Yuval Tassa,et al. MuJoCo: A physics engine for model-based control , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[20] C. Karen Liu,et al. Dexterous manipulation using both palm and fingers , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).
[21] Dean Pomerleau,et al. ALVINN, an autonomous land vehicle in a neural network , 2015 .
[22] Dieter Fox,et al. DexPilot: Vision-Based Teleoperation of Dexterous Robotic Hand-Arm System , 2019, 2020 IEEE International Conference on Robotics and Automation (ICRA).
[23] Xin Zhang,et al. End to End Learning for Self-Driving Cars , 2016, ArXiv.
[24] Vikash Kumar,et al. MuJoCo HAPTIX: A virtual reality system for hand manipulation , 2015, 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids).
[25] Sergey Levine,et al. (CAD)$^2$RL: Real Single-Image Flight without a Single Real Image , 2016, Robotics: Science and Systems.
[26] Ronald J. Williams,et al. Simple Statistical Gradient-Following Algorithms for Connectionist Reinforcement Learning , 2004, Machine Learning.
[27] Mo Yu,et al. Hybrid Reinforcement Learning with Expert State Sequences , 2019, AAAI.
[28] Pieter Abbeel,et al. Autonomous Helicopter Aerobatics through Apprenticeship Learning , 2010, Int. J. Robotics Res..
[29] Jitendra Malik,et al. Learning to Poke by Poking: Experiential Learning of Intuitive Physics , 2016, NIPS.
[30] M. Kawato,et al. A hierarchical neural-network model for control and learning of voluntary movement , 2004, Biological Cybernetics.
[31] Byron Boots,et al. Provably Efficient Imitation Learning from Observation Alone , 2019, ICML.
[32] Pieter Abbeel,et al. Benchmarking Deep Reinforcement Learning for Continuous Control , 2016, ICML.
[33] Peter Stone,et al. Behavioral Cloning from Observation , 2018, IJCAI.
[34] Sheldon Andrews,et al. Goal directed multi-finger manipulation: Control policies and analysis , 2013, Comput. Graph..
[35] Robyn M. Holmes,et al. Cultural Learning , 2020, Cultural Psychology.
[36] Christus,et al. A General Method Applicable to the Search for Similarities in the Amino Acid Sequence of Two Proteins , 2022 .
[37] Yuval Tassa,et al. Simulation tools for model-based robotics: Comparison of Bullet, Havok, MuJoCo, ODE and PhysX , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[38] Jitendra Malik,et al. SFV , 2018, ACM Trans. Graph..
[39] Jitendra Malik,et al. Zero-Shot Visual Imitation , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[40] Nando de Freitas,et al. Playing hard exploration games by watching YouTube , 2018, NeurIPS.
[41] Vikash Kumar,et al. Fast, strong and compliant pneumatic actuation for dexterous tendon-driven hands , 2013, 2013 IEEE International Conference on Robotics and Automation.
[42] Sergey Levine,et al. Deep Reinforcement Learning in a Handful of Trials using Probabilistic Dynamics Models , 2018, NeurIPS.
[43] S. Srinivasa,et al. Push-grasping with dexterous hands: Mechanics and a method , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.
[44] Stefan Schaal,et al. Is imitation learning the route to humanoid robots? , 1999, Trends in Cognitive Sciences.
[45] Sergey Levine,et al. DeepMimic , 2018, ACM Trans. Graph..
[46] Abhinav Gupta,et al. The Curious Robot: Learning Visual Representations via Physical Interactions , 2016, ECCV.
[47] Nando de Freitas,et al. Reinforcement and Imitation Learning for Diverse Visuomotor Skills , 2018, Robotics: Science and Systems.
[48] Martin A. Riedmiller,et al. Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards , 2017, ArXiv.
[49] Peter Stone,et al. RIDM: Reinforced Inverse Dynamics Modeling for Learning from a Single Observed Demonstration , 2019, IEEE Robotics and Automation Letters.
[50] Stefano Ermon,et al. Generative Adversarial Imitation Learning , 2016, NIPS.
[51] Sergey Levine,et al. Learning Robust Rewards with Adversarial Inverse Reinforcement Learning , 2017, ICLR 2017.
[52] Sergey Levine,et al. Imitation from Observation: Learning to Imitate Behaviors from Raw Video via Context Translation , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[53] Wojciech Zaremba,et al. Domain randomization for transferring deep neural networks from simulation to the real world , 2017, 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[54] Sergey Levine,et al. Neural Network Dynamics for Model-Based Deep Reinforcement Learning with Model-Free Fine-Tuning , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[55] Peter Stone,et al. Generative Adversarial Imitation from Observation , 2018, ArXiv.
[56] Jitendra Malik,et al. Combining self-supervised learning and imitation for vision-based rope manipulation , 2017, 2017 IEEE International Conference on Robotics and Automation (ICRA).
[57] Wojciech Zaremba,et al. Transfer from Simulation to Real World through Learning Deep Inverse Dynamics Model , 2016, ArXiv.
[58] Abhinav Gupta,et al. Multiple Interactions Made Easy (MIME): Large Scale Demonstrations Data for Imitation , 2018, CoRL.
[59] Pieter Abbeel,et al. Apprenticeship learning via inverse reinforcement learning , 2004, ICML.
[60] Stefan Schaal,et al. 2008 Special Issue: Reinforcement learning of motor skills with policy gradients , 2008 .
[61] Sergey Levine,et al. Learning Complex Dexterous Manipulation with Deep Reinforcement Learning and Demonstrations , 2017, Robotics: Science and Systems.