Interactive Open-Ended Object, Affordance and Grasp Learning for Robotic Manipulation

Service robots are expected to autonomously and efficiently work in human-centric environments. For this type of robots, object perception and manipulation are challenging tasks due to need for accurate and real-time response. This paper presents an interactive open-ended learning approach to recognize multiple objects and their grasp affordances concurrently. This is an important contribution in the field of service robots since no matter how extensive the training data used for batch learning, a robot might always be confronted with an unknown object when operating in human-centric environments. The paper describes the system architecture and the learning and recognition capabilities. Grasp learning associates grasp configurations (i.e., end-effector positions and orientations) to grasp affordance categories. The grasp affordance category and the grasp configuration are taught through verbal and kinesthetic teaching, respectively. A Bayesian approach is adopted for learning and recognition of object categories and an instance-based approach is used for learning and recognition of affordance categories. An extensive set of experiments has been performed to assess the performance of the proposed approach regarding recognition accuracy, scalability and grasp success rate on challenging datasets and real-world scenarios.

[1]  Walter Daelemans,et al.  Memory-Based Language Processing , 2009, Studies in natural language processing.

[2]  Geoffrey A. Hollinger,et al.  HERB: a home exploring robotic butler , 2010, Auton. Robots.

[3]  Armando J. Pinho,et al.  Experience-Based Robot Task Learning and Planning with Goal Inference , 2016, ICAPS.

[4]  Luís Seabra Lopes,et al.  Object Learning and Grasping Capabilities for Robotic Home Assistants , 2016, RoboCup.

[5]  Leonidas J. Guibas,et al.  FPNN: Field Probing Neural Networks for 3D Data , 2016, NIPS.

[6]  Yiannis Aloimonos,et al.  Affordance detection of tool parts from geometric features , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[7]  Gi Hyun Lim,et al.  3D object perception and perceptual learning in the RACE project , 2016, Robotics Auton. Syst..

[8]  Advait Jain,et al.  EL-E: an assistive mobile manipulator that autonomously fetches objects from flat surfaces , 2010, Auton. Robots.

[9]  L. Seabra Lopes,et al.  How many words can my robot learn?: An approach and experiments with one-class learning , 2007 .

[10]  Danica Kragic,et al.  Affordance detection for task-specific grasping using deep learning , 2017, 2017 IEEE-RAS 17th International Conference on Humanoid Robotics (Humanoids).

[11]  Tae-Kyun Kim,et al.  Perceiving, Learning, and Recognizing 3D Objects: An Approach to Cognitive Service Robots , 2018, AAAI.

[12]  Luís Seabra Lopes,et al.  Using spoken words to guide open-ended category formation , 2011, Cognitive Processing.

[13]  Alexander Herzog,et al.  Template-based learning of grasp selection , 2012, 2012 IEEE International Conference on Robotics and Automation.

[14]  Hamidreza Kasaei,et al.  OrthographicNet: A Deep Learning Approach for 3D Object Recognition in Open-Ended Domains , 2019, ArXiv.

[15]  Darwin G. Caldwell,et al.  AffordanceNet: An End-to-End Deep Learning Approach for Object Affordance Detection , 2017, 2018 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Oliver Kroemer,et al.  Generalization of human grasping for multi-fingered robot hands , 2012, 2012 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[17]  Morgan Quigley,et al.  ROS: an open-source Robot Operating System , 2009, ICRA 2009.

[18]  Luís Seabra Lopes,et al.  Learning to grasp familiar objects using object view recognition and template matching , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[19]  Hamidreza Mohades Kasaei,et al.  OrthographicNet: A Deep Transfer Learning Approach for 3-D Object Recognition in Open-Ended Domains , 2019, IEEE/ASME Transactions on Mechatronics.

[20]  Alexander Herzog,et al.  Learning of grasp selection based on shape-templates , 2014, Auton. Robots.

[21]  Jörn Malzahn,et al.  WALK‐MAN: A High‐Performance Humanoid Platform for Realistic Environments , 2017, J. Field Robotics.

[22]  Dieter Fox,et al.  Patch Volumes: Segmentation-Based Consistent Mapping with RGB-D Cameras , 2013, 2013 International Conference on 3D Vision.

[23]  J. Kevin O'Regan,et al.  Is There Something Out There? Inferring Space from Sensorimotor Dependencies , 2003, Neural Computation.

[24]  Michael I. Jordan,et al.  Latent Dirichlet Allocation , 2001, J. Mach. Learn. Res..

[25]  Luís Seabra Lopes,et al.  An experimental protocol for the evaluation of open-ended category learning algorithms , 2015, 2015 IEEE International Conference on Evolving and Adaptive Intelligent Systems (EAIS).

[26]  Luís Seabra Lopes,et al.  GOOD: A global orthographic object descriptor for 3D object recognition and manipulation , 2016, Pattern Recognit. Lett..

[27]  Gi Hyun Lim,et al.  Towards lifelong assistive robotics: A tight coupling between object perception and manipulation , 2018, Neurocomputing.

[28]  Danica Kragic,et al.  Learning a dictionary of prototypical grasp-predicting parts from grasping experience , 2013, 2013 IEEE International Conference on Robotics and Automation.

[29]  Bharath Hariharan,et al.  Low-Shot Visual Recognition by Shrinking and Hallucinating Features , 2016, 2017 IEEE International Conference on Computer Vision (ICCV).

[30]  Tamim Asfour,et al.  Integrated Grasp and motion planning , 2010, 2010 IEEE International Conference on Robotics and Automation.

[31]  Gi Hyun Lim,et al.  Concurrent learning of visual codebooks and object categories in open-ended domains , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[32]  Dejan Pangercic,et al.  Robotic roommates making pancakes , 2011, 2011 11th IEEE-RAS International Conference on Humanoid Robots.

[33]  Luís Seabra Lopes,et al.  Hierarchical Object Representation for Open-Ended Object Category Learning and Recognition , 2016, NIPS.

[34]  Jianxiong Xiao,et al.  3D ShapeNets: A deep representation for volumetric shapes , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[35]  Nikolaos G. Tsagarakis,et al.  Detecting object affordances with Convolutional Neural Networks , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[36]  PerronninFlorent,et al.  Good Practice in Large-Scale Learning for Image Classification , 2014 .

[37]  Luís Seabra Lopes,et al.  Coping with Context Change in Open-Ended Object Recognition without Explicit Context Information , 2018, IROS.

[38]  Dieter Fox,et al.  A large-scale hierarchical multi-view RGB-D object dataset , 2011, 2011 IEEE International Conference on Robotics and Automation.

[39]  Andrew E. Johnson,et al.  Using Spin Images for Efficient Object Recognition in Cluttered 3D Scenes , 1999, IEEE Trans. Pattern Anal. Mach. Intell..