暂无分享,去创建一个
Katsushi Ikeuchi | Hideki Koike | Kazuhiro Sasabuchi | Naoki Wake | Daichi Saito | K. Ikeuchi | H. Koike | Naoki Wake | Kazuhiro Sasabuchi | Daichi Saito
[1] Silvio Savarese,et al. Learning task-oriented grasping for tool manipulation from simulated self-supervision , 2018, Robotics: Science and Systems.
[2] Aaron M. Dollar,et al. Analysis of Human Grasping Behavior: Object Characteristics and Grasp Type , 2014, IEEE Transactions on Haptics.
[3] Honglak Lee,et al. Deep learning for detecting robotic grasps , 2013, Int. J. Robotics Res..
[4] P. Corke,et al. On the choice of grasp type and location when handing over an object , 2019, Science Robotics.
[5] Katsushi Ikeuchi,et al. Task-Oriented Motion Mapping on Robots of Various Configuration Using Body Role Division , 2020, IEEE Robotics and Automation Letters.
[6] Aaron M. Dollar,et al. Analysis of Human Grasping Behavior: Correlating Tasks, Objects and Grasps , 2014, IEEE Transactions on Haptics.
[7] Huseyin Atakan Varol,et al. Human grasping database for activities of daily living with depth, color and kinematic data streams , 2018, Scientific Data.
[8] Francesc Moreno-Noguer,et al. GanHand: Predicting Human Grasp Affordances in Multi-Object Scenes , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Katsushi Ikeuchi,et al. Contact Web Status Presentation for Freehand Grasping in MR-based Robot-teaching , 2021, HRI.
[10] Sergey Levine,et al. QT-Opt: Scalable Deep Reinforcement Learning for Vision-Based Robotic Manipulation , 2018, CoRL.
[11] M. Kiefer,et al. Action observation can prime visual object recognition , 2009, Experimental Brain Research.
[12] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[13] Stefano Caselli,et al. Grasp recognition in virtual reality for robot pregrasp planning by demonstration , 2006, Proceedings 2006 IEEE International Conference on Robotics and Automation, 2006. ICRA 2006..
[14] Katsushi Ikeuchi,et al. Toward automatic robot instruction from perception-mapping human grasps to manipulator grasps , 1997, IEEE Trans. Robotics Autom..
[15] Jun Takamatsu,et al. A Learning-from-Observation Framework: One-Shot Robot Teaching for Grasp-Manipulation-Release Household Operations , 2020, ArXiv.
[16] Charles C. Kemp,et al. ContactDB: Analyzing and Predicting Grasp Contact via Thermal Imaging , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[17] Yoichi Sato,et al. A scalable approach for understanding the visual structures of hand grasps , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[18] Vincent Lepetit,et al. HOnnotate: A Method for 3D Annotation of Hand and Object Poses , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Shanxin Yuan,et al. First-Person Hand Action Benchmark with RGB-D Videos and 3D Hand Pose Annotations , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[21] Sinisa Todorovic,et al. A Multi-scale CNN for Affordance Segmentation in RGB Images , 2016, ECCV.
[22] Francesc Moreno-Noguer,et al. Learning Depth-Aware Deep Representations for Robotic Perception , 2017, IEEE Robotics and Automation Letters.
[23] Danica Kragic,et al. Learning to Estimate Pose and Shape of Hand-Held Objects from RGB Images , 2019, 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).
[24] Danica Kragic,et al. The GRASP Taxonomy of Human Grasp Types , 2016, IEEE Transactions on Human-Machine Systems.
[25] Kazuhiro Sasabuchi,et al. Grasp-type Recognition Leveraging Object Affordance , 2020, ArXiv.
[26] Aaron M. Dollar,et al. The Yale human grasping dataset: Grasp, object, and task data in household and machine shop environments , 2015, Int. J. Robotics Res..
[27] Siddhartha S. Srinivasa,et al. The YCB object and Model set: Towards common benchmarks for manipulation research , 2015, 2015 International Conference on Advanced Robotics (ICAR).
[28] Darwin G. Caldwell,et al. AffordanceNet: An End-to-End Deep Learning Approach for Object Affordance Detection , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[29] Iori Yanokura,et al. Verbal Focus-of-Attention System for Learning-from-Demonstration , 2020, ArXiv.
[30] J. Gibson,et al. The Senses Considered As Perceptual Systems , 1967 .
[31] Yoichi Sato,et al. Understanding hand-object manipulation by modeling the contextual relationship between actions, grasp types and object attributes , 2018, ArXiv.
[32] Sergey Levine,et al. Using Simulation and Domain Adaptation to Improve Efficiency of Deep Robotic Grasping , 2018, 2018 IEEE International Conference on Robotics and Automation (ICRA).
[33] Yi Li,et al. Robot Learning Manipulation Action Plans by "Watching" Unconstrained Videos from the World Wide Web , 2015, AAAI.
[34] Kris M. Kitani,et al. How do we use our hands? Discovering a diverse set of common grasps , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[35] Yu Sun,et al. Robot grasp planning based on demonstrated grasp strategies , 2015, Int. J. Robotics Res..
[36] Danica Kragic,et al. Task-Based Robot Grasp Planning Using Probabilistic Inference , 2015, IEEE Transactions on Robotics.
[37] Shuang Cong,et al. Robotic Grasping of Novel Objects from RGB-D Images by Using Multi-Level Convolutional Neural Networks , 2018, 2018 IEEE International Conference on Information and Automation (ICIA).
[38] Danica Kragic,et al. Data-Driven Grasp Synthesis—A Survey , 2013, IEEE Transactions on Robotics.
[39] Danica Kragic,et al. Affordance detection for task-specific grasping using deep learning , 2017, 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids).
[40] Ashutosh Saxena,et al. Efficient grasping from RGBD images: Learning using a new rectangle representation , 2011, 2011 IEEE International Conference on Robotics and Automation.
[41] Danica Kragic,et al. Learning Task-Oriented Grasping From Human Activity Datasets , 2019, IEEE Robotics and Automation Letters.
[42] Joseph Redmon,et al. Real-time grasp detection using convolutional neural networks , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).
[43] Deva Ramanan,et al. Understanding Everyday Hands in Action from RGB-D Images , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[44] Cordelia Schmid,et al. Learning Joint Reconstruction of Hands and Manipulated Objects , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[45] Manfred Lau,et al. Tactile mesh saliency , 2016, ACM Trans. Graph..
[46] R. Howe,et al. Human grasp choice and robotic grasp analysis , 1990 .
[47] Lea Fleischer,et al. The Senses Considered As Perceptual Systems , 2016 .
[48] Peter Corke,et al. Closing the Loop for Robotic Grasping: A Real-time, Generative Grasp Synthesis Approach , 2018, Robotics: Science and Systems.
[49] Yu Sun,et al. Grasp planning based on strategy extracted from demonstration , 2014, 2014 IEEE/RSJ International Conference on Intelligent Robots and Systems.