Learning to Parse Natural Language to a Robot Execution System

The ability to interpret, or ground, natural language commands so they can be executed by a robot is important for enabling untrained users to interact with robots. Logic-based formal representations have been applied very successfully to robots’ world and action models taking perceptual and grounding uncertainty into account; however, these approaches have historically either used formal target representations with severely limited expressiveness or written the formal representations manually rather than learning them. As a consequence, they are not able to parse previously unseen NL commands to more complex robot control systems. In this paper, we investigate how recently developed parser learning techniques can be applied to mapping natural language commands to a formal representation expressive enough for modeling robot control systems. Specifically, we aim to learn a parser mapping natural language indoor route instructions into a LISP-like control language. We test our approach using a simulator executing RCL programs on sets of natural language route instructions given by non-experts. 1 Motivation, Problem Statement and Related Work The ability to interpret, or ground, natural language (NL) commands so they can be executed by a robot is important for enabling untrained users to interact with robots. The problem of grounding NL commands encompasses two key components: First, parsing natural language into a formal representation capable of representing a robot and its operation in an environment; and, second, mapping the formal representation to actions and perceptions in the real world. Logic-based formal representations have been applied very successfully to robot control systems, with manual grounding of commands for navigation tasks [2, 5, 4, 9]. A more flexible approach is to learn grounding relations, rather than inputting them manually. One line of work shows how parsed natural

[1]  Matthew R. Walter,et al.  Approaching the Symbol Grounding Problem with Probabilistic Graphical Models , 2011, AI Mag..

[2]  Raymond J. Mooney,et al.  Learning to Interpret Natural Language Navigation Instructions from Observations , 2011, Proceedings of the AAAI Conference on Artificial Intelligence.

[3]  Mark Steedman,et al.  Inducing Probabilistic CCG Grammars from Logical Form with Higher-Order Unification , 2010, EMNLP.

[4]  Dieter Fox,et al.  Following directions using statistical machine translation , 2010, 2010 5th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[5]  A. Haas,et al.  Learning to Follow Navigational Route Instructions , 2009, IJCAI.

[6]  Matthias Scheutz,et al.  What to do and how to do it: Translating natural language directives into temporal and dynamic logic representation for goal management and action execution , 2009, 2009 IEEE International Conference on Robotics and Automation.

[7]  Nicholas Roy,et al.  Where to go: Interpreting natural directions using global inference , 2009, 2009 IEEE International Conference on Robotics and Automation.

[8]  Alexander Ferrein,et al.  Logic-based robot control in highly dynamic domains , 2008, Robotics Auton. Syst..

[9]  Hadas Kress-Gazit,et al.  Translating Structured English to Robot Controllers , 2008, Adv. Robotics.

[10]  Dieter Fox,et al.  Voronoi Random Fields: Extracting Topological Structure of Indoor Environments via Place Labeling , 2007, IJCAI.

[11]  D. Fox,et al.  Integrated Plan-based Control of Autonomous Service Robots in Human Environments , 2001 .

[12]  Wolfram Burgard,et al.  Experiences with an Interactive Museum Tour-Guide Robot , 1999, Artif. Intell..

[13]  Wolfram Burgard,et al.  GOLEX - Bridging the Gap between Logic (GOLOG) and a Real Robot , 1998, KI.

[14]  Rachid Alami,et al.  PRS: a high level supervision and control language for autonomous mobile robots , 1996, Proceedings of IEEE International Conference on Robotics and Automation.