Real-time natural language corrections for assistive robotic manipulators

We propose a generalizable natural language interface that allows users to provide corrective instructions to an assistive robotic manipulator in real-time. This work is motivated by the desire to improve collaboration between humans and robots in a home environment. Allowing human operators to modify properties of how their robotic counterpart achieves a goal on-the-fly increases the utility of the system by incorporating the strengths of the human partner (e.g. visual acuity and environmental knowledge). This work is applicable to users with and without disability. Our natural language interface is based on the distributed correspondence graph, a probabilistic graphical model that assigns semantic meaning to user utterances in the context of the robot’s environment and current behavior. We then use the desired corrections to alter the behavior of the robotic manipulator by treating the modifications as constraints on the motion generation (planning) paradigm. In this paper, we highlight four dimensions along which a user may wish to correct the behavior of his or her assistive manipulator. We develop our language model using data collected from Amazon Mechanical Turk to capture a comprehensive sample of terminology that people use to describe desired corrections. We then develop an end-to-end system using open-source speech-to-text software and a Kinova Robotics MICO robotic arm. To demonstrate the efficacy of our approach, we run a pilot study with users unfamiliar with robotic systems and analyze points of failure and future directions.

[1]  Jeffrey Dean,et al.  Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.

[2]  Michael A. Goodrich,et al.  Expressing homotopic requirements for mobile robot navigation through natural language instructions , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[3]  Siddhartha S. Srinivasa,et al.  People helping robots helping people: Crowdsourcing for grasping novel objects , 2010, 2010 IEEE/RSJ International Conference on Intelligent Robots and Systems.

[4]  Matthew R. Walter,et al.  Listen, Attend, and Walk: Neural Mapping of Navigational Instructions to Action Sequences , 2015, AAAI.

[5]  Pieter Abbeel,et al.  Cloth grasp point detection based on multiple-view geometric cues with application to robotic towel folding , 2010, 2010 IEEE International Conference on Robotics and Automation.

[6]  Benjamin Kuipers,et al.  Walk the Talk: Connecting Language, Knowledge, and Action in Route Instructions , 2006, AAAI.

[7]  Sonia Chernova,et al.  Robot Web Tools: Efficient messaging for cloud robotics , 2015, 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[8]  Matthew R. Walter,et al.  Learning models for following natural language directions in unknown environments , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[9]  Christian Mandel,et al.  Comparison of Wheelchair User Interfaces for the Paralysed: Head-Joystick vs. Verbal Path Selection from an offered Route-Set , 2007, EMCR.

[10]  Stevan Harnad The Symbol Grounding Problem , 1999, ArXiv.

[11]  Nicholas Roy,et al.  Efficient Grounding of Abstract Spatial Concepts for Natural Language Interaction with Robot Manipulators , 2016, Robotics: Science and Systems.

[12]  Aude Billard,et al.  A survey of Tactile Human-Robot Interactions , 2010, Robotics Auton. Syst..

[13]  Dae-Jin Kim,et al.  An empirical study with simulated ADL tasks using a vision-guided assistive robot arm , 2009, 2009 IEEE International Conference on Rehabilitation Robotics.

[14]  Matthew R. Walter,et al.  A multimodal interface for real-time soldier-robot teaming , 2016, SPIE Defense + Security.

[15]  Marc Toussaint,et al.  Understanding the geometry of workspace obstacles in Motion Optimization , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[16]  Jayesh K. Gupta,et al.  PlanIt: A crowdsourcing approach for learning to plan paths from large scale preference feedback , 2014, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[17]  Hadas Kress-Gazit,et al.  A model for verifiable grounding and execution of complex natural language instructions , 2016, 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS).

[18]  Heinz Wörn,et al.  A tactile language for intuitive human-robot communication , 2007, Humanoids.

[19]  Matthew R. Walter,et al.  Approaching the Symbol Grounding Problem with Probabilistic Graphical Models , 2011, AI Mag..

[20]  Dong-Soo Kwon,et al.  Integration of a Rehabilitation Robotic System (KARES II) with Human-Friendly Man-Machine Interaction Units , 2004, Auton. Robots.

[21]  Scott Niekum,et al.  Learning grounded finite-state representations from unstructured demonstrations , 2015, Int. J. Robotics Res..

[22]  Brett Browning,et al.  Policy Feedback for the Refinement of Learned Motion Control on a Mobile Robot , 2012, Int. J. Soc. Robotics.

[23]  Jean Oh,et al.  Inferring Maps and Behaviors from Natural Language Instructions , 2015, ISER.

[24]  Stefanie Tellex,et al.  Toward Information Theoretic Human-Robot Dialog , 2012, Robotics: Science and Systems.

[25]  Eric L. Sauser,et al.  Tactile guidance for policy refinement and reuse , 2010, 2010 IEEE 9th International Conference on Development and Learning.

[26]  A. Graser,et al.  Low level control in a semi-autonomous rehabilitation robotic system via a Brain-Computer Interface , 2007, 2007 IEEE 10th International Conference on Rehabilitation Robotics.

[27]  A. Graser,et al.  Rehabilitation robot FRIEND II - the general concept and current implementation , 2005, 9th International Conference on Rehabilitation Robotics, 2005. ICORR 2005..

[28]  Wolfram Burgard,et al.  Robot, organize my shelves! Tidying up objects by predicting user preferences , 2015, 2015 IEEE International Conference on Robotics and Automation (ICRA).

[29]  Slav Petrov,et al.  Globally Normalized Transition-Based Neural Networks , 2016, ACL.

[30]  Matthew R. Walter,et al.  Understanding Natural Language Commands for Robotic Navigation and Mobile Manipulation , 2011, AAAI.

[31]  Darwin G. Caldwell,et al.  Reinforcement Learning in Robotics: Applications and Real-World Challenges , 2013, Robotics.

[32]  A. Graser,et al.  Brain-Computer Interface for high-level control of rehabilitation robotic systems , 2007, 2007 IEEE 10th International Conference on Rehabilitation Robotics.

[33]  M Busnel,et al.  The robotized workstation "MASTER" for users with tetraplegia: description and evaluation. , 1999, Journal of rehabilitation research and development.

[34]  Stefanie Tellex,et al.  A natural language planner interface for mobile manipulators , 2014, 2014 IEEE International Conference on Robotics and Automation (ICRA).

[35]  Jan Peters,et al.  Policy Search for Motor Primitives in Robotics , 2008, NIPS 2008.

[36]  Matthew R. Walter,et al.  Efficient Natural Language Interfaces for Assistive Robots , 2014, IROS 2014.

[37]  Maya Cakmak,et al.  Adaptive Coordination Strategies for Human-Robot Handovers , 2015, Robotics: Science and Systems.

[38]  Luke S. Zettlemoyer,et al.  Learning to Parse Natural Language Commands to a Robot Control System , 2012, ISER.