A Sensorimotor Learning Framework for Object Categorization

This paper presents a framework that enables a robot to discover various object categories through interaction. The categories are described using action-effect relations, i.e., sensorimotor contingencies rather than more static shape or appearance representation. The framework provides a functionality to classify objects and the resulting categories, associating a class with a specific module. We demonstrate the performance of the framework by studying a pushing behavior in robots, encoding the sensorimotor contingencies and their predictability with Gaussian Processes. We show how entropy-based action selection can improve object classification and how functional categories emerge from the similarities of effects observed among the objects. We also show how a multidimensional action space can be realized by parameterizing pushing using both position and velocity.

[1]  Danica Kragic,et al.  Visual object-action recognition: Inferring object affordances from human demonstration , 2011, Comput. Vis. Image Underst..

[2]  Silvio Savarese,et al.  3D generic object categorization, localization and pose estimation , 2007, 2007 IEEE 11th International Conference on Computer Vision.

[3]  Paul F. M. J. Verschure,et al.  The acquisition of intentionally indexed and object centered affordance gradients: A biomimetic controller and mobile robotics benchmark , 2011, 2011 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Danica Kragic,et al.  Active 3D scene segmentation and detection of unknown objects , 2010, 2010 IEEE International Conference on Robotics and Automation.

[5]  Sinan Kalkan,et al.  Learning Affordances for Categorizing Objects and Their Properties , 2010, 2010 20th International Conference on Pattern Recognition.

[6]  Markus Vincze,et al.  Affordance based Part Recognition for Grasping and Manipulation , 2011 .

[7]  D M Wolpert,et al.  Multiple paired forward and inverse models for motor control , 1998, Neural Networks.

[8]  Bernt Schiele,et al.  Analyzing appearance and contour based methods for object categorization , 2003, 2003 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2003. Proceedings..

[9]  Pierre-Yves Oudeyer,et al.  Object Learning Through Active Exploration , 2014, IEEE Transactions on Autonomous Mental Development.

[10]  Gunnar Farnebäck,et al.  Two-Frame Motion Estimation Based on Polynomial Expansion , 2003, SCIA.

[11]  Carl E. Rasmussen,et al.  Gaussian processes for machine learning , 2005, Adaptive computation and machine learning.

[12]  Alexander Stoytchev,et al.  Some Basic Principles of Developmental Robotics , 2009, IEEE Transactions on Autonomous Mental Development.

[13]  Alexander Stoytchev,et al.  Behavior-Grounded Representation of Tool Affordances , 2005, Proceedings of the 2005 IEEE International Conference on Robotics and Automation.

[14]  Eren Erdal Aksoy,et al.  Learning the semantics of object–action relations by observation , 2011, Int. J. Robotics Res..

[15]  Christopher W. Geib,et al.  The meaning of action: a review on action recognition and mapping , 2007, Adv. Robotics.

[16]  Rolf Pfeifer,et al.  How the body shapes the way we think - a new view on intelligence , 2006 .

[17]  Markus Vincze,et al.  Supervised learning of hidden and non-hidden 0-order affordances and detection in real scenes , 2012, 2012 IEEE International Conference on Robotics and Automation.

[18]  Robert C. Bolles,et al.  Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography , 1981, CACM.

[19]  Alexander Maye,et al.  A discrete computational model of sensorimotor contingencies for object perception and control of behavior , 2011, 2011 IEEE International Conference on Robotics and Automation.

[20]  Emre Ugur,et al.  Goal emulation and planning in perceptual space using learned affordances , 2011, Robotics Auton. Syst..

[21]  Mark Steedman,et al.  Object-Action Complexes: Grounded abstractions of sensory-motor processes , 2011, Robotics Auton. Syst..

[22]  P. Verschure,et al.  A biologically based model for the integration of sensory–motor contingencies in rules and plans: A prefrontal cortex based extension of the Distributed Adaptive Control architecture , 2011, Brain Research Bulletin.

[23]  Dirk Kraft,et al.  A Survey of the Ontogeny of Tool Use: From Sensorimotor Experience to Planning , 2013, IEEE Transactions on Autonomous Mental Development.

[24]  Jivko Sinapov,et al.  Vibrotactile Recognition and Categorization of Surfaces by a Humanoid Robot , 2011, IEEE Transactions on Robotics.

[25]  Ales Ude,et al.  Action-grounded push affordance bootstrapping of unknown objects , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[26]  Mitsuo Kawato,et al.  MOSAIC Model for Sensorimotor Learning and Control , 2001, Neural Computation.

[27]  E. Reed The Ecological Approach to Visual Perception , 1989 .

[28]  J. Sinapov,et al.  Detecting the functional similarities between tools using a hierarchical representation of outcomes , 2008, 2008 7th IEEE International Conference on Development and Learning.

[29]  Giulio Sandini,et al.  Learning about objects through action - initial steps towards artificial cognition , 2003, 2003 IEEE International Conference on Robotics and Automation (Cat. No.03CH37422).

[30]  Jivko Sinapov,et al.  From Acoustic Object Recognition to Object Categorization by a Humanoid Robot , 2009 .

[31]  Guillaume Bouchard,et al.  Hierarchical part-based visual object categorization , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[32]  Danica Kragic,et al.  Interactive object classification using sensorimotor contingencies , 2013, 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[33]  David L. Faigman,et al.  Human category learning. , 2005, Annual review of psychology.

[34]  Jivko Sinapov,et al.  A Behavior-Grounded Approach to Forming Object Categories: Separating Containers From Noncontainers , 2012, IEEE Transactions on Autonomous Mental Development.

[35]  Jitendra Malik,et al.  Normalized Cuts and Image Segmentation , 2000, IEEE Trans. Pattern Anal. Mach. Intell..

[36]  Giorgio Metta,et al.  Better Vision through Manipulation , 2003, Adapt. Behav..

[37]  Manuel Lopes,et al.  Learning Object Affordances: From Sensory--Motor Coordination to Imitation , 2008, IEEE Transactions on Robotics.

[38]  Markus Vincze,et al.  Parallel Deep Learning with Suggestive Activation for Object Category Recognition , 2013, ICVS.

[39]  A. Noë,et al.  A sensorimotor account of vision and visual consciousness. , 2001, The Behavioral and brain sciences.

[40]  Gerald E. Loeb,et al.  Bayesian Exploration for Intelligent Identification of Textures , 2012, Front. Neurorobot..

[41]  Connor Schenck,et al.  Grounding semantic categories in behavioral interactions: Experiments with 100 objects , 2014, Robotics Auton. Syst..

[42]  Jan Peters,et al.  Model learning for robot control: a survey , 2011, Cognitive Processing.

[43]  Ian R. Fasel,et al.  InfoMax Control for Acoustic Exploration of Objects by a Mobile Robot , 2011, Lifelong Learning.