Action-grounded surface geometry and volumetric shape feature representations for object affordance prediction

Many 3D feature descriptors have been developed over the years to solve problems that require the representation of object shape, e. g. object recognition or pose estimation, but comparatively few have been developed specifically to tackle the problem of object affordance learning, a domain where the interaction between action parameters and sensory features play a crucial role. In previous work, we introduced a feature descriptor that divided an object point cloud into coarse-grained cells, derived simple features from each of the cells, and grounded those features with respect to a reference frame defined by a pushing action. We also compared this action-grounded descriptor to an equivalent non-action-grounded descriptor coupled with action features in a push affordance classification task and established that the action-grounded encoding can provide improved performance. In this paper, we investigate modifying more well-established 3D shape descriptors based on surface geometry, in particular the Viewpoint Feature Histogram (VFH), such that they are action-grounded in a similar manner, compare them to volumetric octree-based representations, and conclude that having multi-scaled representations in which parts at each scale can be referenced with respect to each other may be a crucial component in action-grounded affordance learning.

[1]  Radu Bogdan Rusu,et al.  3D is here: Point Cloud Library (PCL) , 2011, 2011 IEEE International Conference on Robotics and Automation.

[2]  Nico Blodow,et al.  CAD-model recognition and 6DOF pose estimation using 3D cues , 2011, 2011 IEEE International Conference on Computer Vision Workshops (ICCV Workshops).

[3]  Ales Ude,et al.  Self-Supervised Online Learning of Basic Object Push Affordances , 2015 .

[4]  Nico Blodow,et al.  Aligning point cloud views using persistent feature histograms , 2008, 2008 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[5]  Alexander Stoytchev,et al.  Some Basic Principles of Developmental Robotics , 2009, IEEE Transactions on Autonomous Mental Development.

[6]  Radu Bogdan Rusu,et al.  Semantic 3D Object Maps for Everyday Manipulation in Human Living Environments , 2010, KI - Künstliche Intelligenz.

[7]  E. Reed The Ecological Approach to Visual Perception , 1989 .

[8]  Emre Ugur,et al.  Traversability: A Case Study for Learning and Perceiving Affordances in Robots , 2010, Adapt. Behav..

[9]  James M. Rehg,et al.  Learning contact locations for pushing and orienting unknown objects , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).

[10]  Giorgio Metta,et al.  Multi-model approach based on 3D functional features for tool affordance learning in robotics , 2015, 2015 IEEE-RAS 15th International Conference on Humanoid Robots (Humanoids).

[11]  G. Metta,et al.  Exploring affordances and tool use on the iCub , 2013, 2013 13th IEEE-RAS International Conference on Humanoid Robots (Humanoids).

[12]  Ales Ude,et al.  Action-grounded push affordance bootstrapping of unknown objects , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[13]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[14]  Danica Kragic,et al.  A Sensorimotor Learning Framework for Object Categorization , 2016, IEEE Transactions on Cognitive and Developmental Systems.

[15]  Michael Beetz,et al.  Detecting and segmenting objects for mobile manipulation , 2009, 2009 IEEE 12th International Conference on Computer Vision Workshops, ICCV Workshops.

[16]  Giulio Sandini,et al.  Learning about objects through action - initial steps towards artificial cognition , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[17]  Ales Ude,et al.  Learning primitive actions through object exploration , 2008, Humanoids 2008 - 8th IEEE-RAS International Conference on Humanoid Robots.

[18]  Emre Ugur,et al.  Goal emulation and planning in perceptual space using learned affordances , 2011, Robotics Auton. Syst..

[19]  Mårten Björkman,et al.  Learning to disambiguate object hypotheses through self-exploration , 2014, 2014 IEEE-RAS International Conference on Humanoid Robots.

[20]  Manuel Lopes,et al.  Learning Object Affordances: From Sensory--Motor Coordination to Imitation , 2008, IEEE Transactions on Robotics.

[21]  Nico Blodow,et al.  Combined 2D–3D categorization and classification for multimodal perception systems , 2011, Int. J. Robotics Res..

[22]  Andrew E. Johnson,et al.  Spin-Images: A Representation for 3-D Surface Matching , 1997 .

[23]  Ales Ude,et al.  Comparison of action-grounded and non-action-grounded 3-D shape features for object affordance classification , 2015, 2015 International Conference on Advanced Robotics (ICAR).

[24]  Gary R. Bradski,et al.  Fast 3D recognition and pose using the Viewpoint Feature Histogram , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[25]  Nico Blodow,et al.  Fast Point Feature Histograms (FPFH) for 3D registration , 2009, 2009 IEEE International Conference on Robotics and Automation.

[26]  Federico Tombari,et al.  Unique Signatures of Histograms for Local Surface Description , 2010, ECCV.