Personal Robot Training via Natural-Language Instructions.

Future domestic robots will need to adapt to the special needs of their users and to their environment. Programming by natural language will be a key method enabling computer language-naive users to instruct their robots. Its main advantages over other learning methods are speed of acquisition and ability to build high level symbolic rules into the robot. This paper describes the design of a practical system that uses unconstrained speech to teach a vision-based robot how to navigate in a miniature town. The robot knows a set of primitive navigation procedures that the user can refer to when giving route instructions. A particularity of this project is that the primitive procedures are determined by analysing a corpus of route instructions. It is found that primitives natural to the user, such as “turn left after the church” are very complex procedures for the robot, involving visual scene analysis and local route planning. Thus, to enable natural user-robot interaction, a high-level of intelligence needs to be built into “primitive” robot procedures. Another finding is that the set of primitive procedures is likely not to be closed. Thus, on time to time, a user is likely to refer to a procedure that is not preprogrammed in the robot. How best to handle this is currently investigated. In general, the use of Instruction-Based Learning (IBL) imposes a number of constraints on the design of robotics systems and knowledge representation. These issues and proposed solutions are described in the paper.

[1]  António J. S. Teixeira,et al.  Human-robot interaction through spoken language dialogue , 2000, Proceedings. 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000) (Cat. No.00CH37113).

[2]  Angelo Cangelosi,et al.  The Adaptive Advantage of Symbolic Theft Over Sensorimotor Toil: Grounding Language in Perceptual Categories , 2001 .

[3]  M. Denis The description of routes : A cognitive approach to the production of spatial discourse , 1997 .

[4]  Allen Newell,et al.  SOAR: An Architecture for General Intelligence , 1987, Artif. Intell..

[5]  Alex Pentland,et al.  Learning words from sights and sounds: a computational model , 2002, Cogn. Sci..

[6]  Uwe Reyle,et al.  From discourse to logic , 1993 .

[7]  Jun Tani,et al.  Model-based learning for mobile robot navigation from the dynamical systems perspective , 1996, IEEE Trans. Syst. Man Cybern. Part B.

[8]  Patrick Suppes,et al.  Language and Learning for Robots , 1994 .

[9]  Satoru Hayamizu,et al.  Socially Embedded Learning of the Office-Conversant Mobil Robot Jijo-2 , 1997, IJCAI.

[10]  Johan Bos,et al.  A model of dia-logue moves and information state revision , 1999 .

[11]  K. MacDorman,et al.  Grounding Symbols through Sensorimotor Integration , 1999 .

[12]  Luc Steels,et al.  The Origins of Syntax in Visually Grounded Robotic Agents , 1997, IJCAI.

[13]  John E. Laird,et al.  Flexibly Instructable Agents , 1995, J. Artif. Intell. Res..

[14]  Mark C. Torrance,et al.  Natural communication with robots , 1994 .

[15]  Rainer Bischoff,et al.  Natural Communication and Interaction with Humanoid Robots , 1999 .

[16]  R. A. Brooks,et al.  Intelligence without Representation , 1991, Artif. Intell..