Interactive Robot Learning of Gestures, Language and Affordances

A growing field in robotics and Artificial Intelligence (AI) research is human-robot collaboration, whose target is to enable effective teamwork between humans and robots. However, in many situations human teams are still superior to human-robot teams, primarily because human teams can easily agree on a common goal with language, and the individual members observe each other effectively, leveraging their shared motor repertoire and sensorimotor resources. This paper shows that for cognitive robots it is possible, and indeed fruitful, to combine knowledge acquired from interacting with elements of the environment (affordance exploration) with the probabilistic observation of another agent's actions. We propose a model that unites (i) learning robot affordances and word descriptions with (ii) statistical recognition of human gestures with vision sensors. We discuss theoretical motivations, possible implementations, and we show initial results which highlight that, after having acquired knowledge of its surrounding environment, a humanoid robot can generalize this knowledge to the case when it observes another agent (human partner) performing the same motor actions previously executed during training.

[1]  Manuel Lopes,et al.  Learning Object Affordances: From Sensory--Motor Coordination to Imitation , 2008, IEEE Transactions on Robotics.

[2]  James J. Gibson,et al.  The Ecological Approach to Visual Perception: Classic Edition , 2014 .

[3]  Alexandre Bernardino,et al.  From human instructions to robot actions: Formulation of goals, affordances and probabilistic planning , 2016, 2016 IEEE International Conference on Robotics and Automation (ICRA).

[4]  G. Knoblich,et al.  Predicting the Effects of Actions: Interactions of Perception and Action , 2001, Psychological science.

[5]  Mark Steedman,et al.  Object-Action Complexes: Grounded abstractions of sensory-motor processes , 2011, Robotics Auton. Syst..

[6]  Ugo Pattacini,et al.  Grounding speech utterances in robotics affordances: An embodied statistical language model , 2016, 2016 Joint IEEE International Conference on Development and Learning and Epigenetic Robotics (ICDL-EpiRob).

[7]  G. Rizzolatti,et al.  Neurophysiological mechanisms underlying the understanding and imitation of action , 2001, Nature Reviews Neuroscience.

[8]  C. Urgesi,et al.  Action anticipation and motor resonance in elite basketball players , 2008, Nature Neuroscience.

[9]  Angelo Cangelosi,et al.  Affordances in Psychology, Neuroscience, and Robotics: A Survey , 2018, IEEE Transactions on Cognitive and Developmental Systems.

[10]  Alexandre Bernardino,et al.  Learning intermediate object affordances: Towards the development of a tool concept , 2014, 4th International Conference on Development and Learning and on Epigenetic Robotics.

[11]  Alexandre Bernardino,et al.  Learning visual affordances of objects and tools through autonomous robot exploration , 2014, 2014 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC).

[12]  Alexandre Bernardino,et al.  Robot anticipation of human intentions through continuous gesture recognition , 2013, 2013 International Conference on Collaboration Technologies and Systems (CTS).

[13]  Alexandre Bernardino,et al.  Language Bootstrapping: Learning Word Meanings From Perception–Action Association , 2012, IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics).

[14]  Minho Lee,et al.  Understanding human intention by connecting perception and action learning in artificial agents , 2017, Neural Networks.

[15]  Angelo Cangelosi,et al.  Why Are There Developmental Stages in Language Learning? A Developmental Robotics Model of Language Development. , 2017, Cognitive science.

[16]  Luke S. Zettlemoyer,et al.  Learning from Unscripted Deictic Gesture and Language for Human-Robot Interactions , 2014, AAAI.

[17]  R. Miall,et al.  A system in the human brain for predicting the actions of others , 2004, Nature Neuroscience.

[18]  Yun Peng,et al.  Belief Update in Bayesian Networks Using Uncertain Evidence , 2006, 2006 18th IEEE International Conference on Tools with Artificial Intelligence (ICTAI'06).