The learning of adjectives and nouns from affordance and appearance features

We study how a robot can link concepts represented by adjectives and nouns in language with its own sensorimotor interactions. Specifically, an iCub humanoid robot interacts with a group of objects using a repertoire of manipulation behaviors. The objects are labeled using a set of adjectives and nouns. The effects induced on the objects are labeled as affordances, and classifiers are learned to predict the affordances from the appearance of an object. We evaluate three different models for learning adjectives and nouns using features obtained from the appearance and affordances of an object, through cross-validated training as well as through testing on novel objects. The results indicate that shape-related adjectives are best learned using features related to affordances, whereas nouns are best learned using appearance features. Analysis of the feature relevancy shows that affordance features are more relevant for adjectives, and appearance features are more relevant for nouns. We show that adjective predictions can be used to solve the odd-one-out task on a number of examples. Finally, we link our results with studies from psychology, neuroscience and linguistics that point to the differences between the development and representation of adjectives and nouns in humans.

[1]  Mark Steedman,et al.  Object-Action Complexes: Grounded abstractions of sensory-motor processes , 2011, Robotics Auton. Syst..

[2]  Bradley G. Goodyear,et al.  Neural Correlates of Concreteness in Semantic Categorization , 2007, Journal of Cognitive Neuroscience.

[3]  M. Goodale,et al.  Separate visual pathways for perception and action , 1992, Trends in Neurosciences.

[4]  Angelo Cangelosi,et al.  An Embodied Model for Sensorimotor Grounding and Grounding Transfer: Experiments With Epigenetic Robots , 2006, Cogn. Sci..

[5]  Antonio Torralba,et al.  Exploiting hierarchical context on a large database of object categories , 2010, 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[6]  Angelo Cangelosi,et al.  Grounding language in action and perception: from cognitive agents to humanoid robots. , 2010, Physics of life reviews.

[7]  Markus Graf,et al.  The role of action representations in visual object recognition , 2006, Experimental Brain Research.

[8]  Stevan Harnad The Symbol Grounding Problem , 1999, ArXiv.

[9]  James M. Rehg,et al.  Learning Visual Object Categories for Robot Affordance Prediction , 2010, Int. J. Robotics Res..

[10]  Alexandre Bernardino,et al.  A computational model of object affordances , 2010 .

[11]  Jivko Sinapov,et al.  The odd one out task: Toward an intelligence test for robots , 2010, 2010 IEEE 9th International Conference on Development and Learning.

[12]  Linda B. Smith,et al.  Learning Adjectives in the Real World: How Learning Nouns Impedes Learning Adjectives , 2007 .

[13]  Igor Kononenko,et al.  Estimating Attributes: Analysis and Extensions of RELIEF , 1994, ECML.

[14]  Chen Yu,et al.  On the Integration of Grounding Language and Learning Objects , 2004, AAAI.

[15]  E. Gibson Perceptual Learning in Development: Some Basic Concepts , 2000 .

[16]  Rüdiger Dillmann,et al.  The KIT object models database: An object model database for object recognition, localization and manipulation in service robotics , 2012, Int. J. Robotics Res..

[17]  Li Jianc,et al.  A FMRI Study on Modifiers of Noun by Nouns,Verbs and Adjectives in Chinese , 2006 .

[18]  R. Nicoletti,et al.  Simon-Like and Functional Affordance Effects with Tools: The Effects of Object Perceptual Discrimination and Object Action State , 2010, Quarterly journal of experimental psychology.

[19]  Mark C. Baker Lexical Categories: Verbs, Nouns and Adjectives , 2003 .

[20]  Michael Beetz,et al.  ORO, a knowledge management module for cognitive architectures in robotics , 2010, IROS 2010.

[21]  Martin V. Butz,et al.  Simultaneously emerging Braitenberg codes and compositionality , 2011, Adapt. Behav..

[22]  Radu Bogdan Rusu,et al.  3D is here: Point Cloud Library (PCL) , 2011, 2011 IEEE International Conference on Robotics and Automation.

[23]  George S. Cree,et al.  Evocation of functional and volumetric gestural knowledge by objects and words , 2008, Cognition.

[24]  Manuel Lopes,et al.  Learning Object Affordances: From Sensory--Motor Coordination to Imitation , 2008, IEEE Transactions on Robotics.

[25]  Michael P. Kaschak,et al.  Grounding language in action , 2002, Psychonomic bulletin & review.

[26]  Luís Seabra Lopes,et al.  Using spoken words to guide open-ended category formation , 2011, Cognitive Processing.

[27]  Sinan Kalkan,et al.  Verb concepts from affordances , 2014 .

[28]  Angelo Cangelosi,et al.  The power of words , 2011, 2011 IEEE International Conference on Development and Learning (ICDL).

[29]  Haris Dindo,et al.  A probabilistic approach to learning a visually grounded language model through human-robot interaction , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[30]  Brian Scassellati,et al.  Robotic vocabulary building using extension inference and implicit contrast , 2009, Artificial Intelligence.

[31]  Domenico Parisi,et al.  THE ROLE OF PERCEPTION AND ACTION IN OBJECT CATEGORISATION , 2002 .

[32]  Sinan Kalkan,et al.  Learning Affordances for Categorizing Objects and Their Properties , 2010, 2010 20th International Conference on Pattern Recognition.

[33]  A. Cangelosi,et al.  The processing of verbs and nouns in neural networks: Insights from synthetic brain imaging , 2004, Brain and Language.

[34]  A. Glenberg,et al.  Processing Abstract Language Modulates Motor System Activity , 2008, Quarterly journal of experimental psychology.

[35]  A. Glenberg,et al.  Action-based language: A theory of language acquisition, comprehension, and production , 2012, Cortex.

[36]  E. Warrington,et al.  Abstract and concrete concepts have structurally different representational frameworks. , 2005, Brain : a journal of neurology.

[37]  Chen Wu,et al.  Using context with statistical relational models: object recognition from observing user activity in home environment , 2009, UCVP '09.

[38]  Angelo Cangelosi,et al.  The Adaptive Advantage of Symbolic Theft Over Sensorimotor Toil: Grounding Language in Perceptual Categories , 2001 .

[39]  Maya Cakmak,et al.  To Afford or Not to Afford: A New Formalization of Affordances Toward Affordance-Based Robot Control , 2007, Adapt. Behav..

[40]  John A. Bullinaria,et al.  Connectionist Models of Cognition and Perception. , 2002 .

[41]  Michael Beetz,et al.  ORO, a knowledge management platform for cognitive architectures in robotics , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[42]  Antonio Chella,et al.  Grounded Human-Robot Interaction , 2009, AAAI Fall Symposium: Biologically Inspired Cognitive Architectures.

[43]  Steven A. Jax,et al.  Response interference between functional and structural actions linked to the same familiar object , 2010, Cognition.

[44]  Eren Erdal Aksoy,et al.  Categorizing object-action relations from semantic scene graphs , 2010, 2010 IEEE International Conference on Robotics and Automation.

[45]  Alexandre Bernardino,et al.  Language Bootstrapping: Learning Word Meanings From Perception–Action Association , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[46]  G. Lakoff,et al.  The Brain's concepts: the role of the Sensory-motor system in conceptual knowledge , 2005, Cognitive neuropsychology.

[47]  Giulio Sandini,et al.  The iCub humanoid robot: an open platform for research in embodied cognition , 2008, PerMIS.

[48]  E. Reed The Ecological Approach to Visual Perception , 1989 .

[49]  Giulio Sandini,et al.  In Press, Ieee Transactions on Autonomous Mental Development , 2010 .

[50]  D. Marocco,et al.  Manipulating Objects and Telling Words: A Study on Concrete and Abstract Words Acquisition , 2011, Front. Psychology.

[51]  L. Steels Evolving grounded communication for robots , 2003, Trends in Cognitive Sciences.

[52]  Anna M. Borghi,et al.  OBJECT CONCEPTS AND EMBODIMENT : WHY SENSORIMOTOR AND COGNITIVE PROCESSES CANNOT BE SEPARATED , 2006 .

[53]  Bernhard Hommel,et al.  A computational model of perception and action for cognitive robotics , 2011, Cognitive Processing.

[54]  Trevor Darrell,et al.  Robotic learning of haptic adjectives through physical interaction , 2015, Robotics Auton. Syst..

[55]  Sinan Kalkan,et al.  Affordances and Emergence of Concepts , 2010, EpiRob.

[56]  A. Glenberg,et al.  Symbol Grounding and Meaning: A Comparison of High-Dimensional and Embodied Theories of Meaning , 2000 .

[57]  Giorgio Metta,et al.  Better Vision through Manipulation , 2003, Adapt. Behav..

[58]  Rolf A. Zwaan,et al.  Seeing, acting, understanding: motor resonance in language comprehension. , 2006, Journal of experimental psychology. General.

[59]  Kevin Gold,et al.  Toward Fast Mapping for Robot Adjective Learning , 2010, AAAI Fall Symposium: Dialog with Robots.

[60]  Rolf A. Zwaan,et al.  Embodied Language: A Review of the Role of the Motor System in Language Comprehension , 2008, Quarterly journal of experimental psychology.

[61]  Peter Carbonetto,et al.  Why can’t José read? The problem of learning semantic associations in a robot environment , 2003, HLT-NAACL 2003.

[62]  L. Barsalou,et al.  Whither structured representation? , 1999, Behavioral and Brain Sciences.

[63]  C. Schenck,et al.  Interactive Object Recognition Using Proprioceptive Feedback , 2009 .

[64]  Galit W. Sassoon,et al.  Adjectival vs. nominal categorization processes: The Rule vs. Similarity hypothesis , 2011 .