A point-and-click interface for the real world: Laser designation of objects for mobile manipulation

We present a novel interface for human-robot interaction that enables a human to intuitively and unambiguously select a 3D location in the world and communicate it to a mobile robot. The human points at a location of interest and illuminates it (“clicks it”) with an unaltered, off-the-shelf, green laser pointer. The robot detects the resulting laser spot with an omnidirectional, catadioptric camera with a narrow-band green filter. After detection, the robot moves its stereo pan/tilt camera to look at this location and estimates the location's 3D position with respect to the robot's frame of reference. Unlike previous approaches, this interface for gesture-based pointing requires no instrumentation of the environment, makes use of a non-instrumented everyday pointing device, has low spatial error out to 3 meters, is fully mobile, and is robust enough for use in real-world applications. We demonstrate that this human-robot interface enables a person to designate a wide variety of everyday objects placed throughout a room. In 99.4% of these tests, the robot successfully looked at the designated object and estimated its 3D position with low average error. We also show that this interface can support object acquisition by a mobile manipulator. For this application, the user selects an object to be picked up from the floor by “clicking” on it with the laser pointer interface. In 90% of these trials, the robot successfully moved to the designated object and picked it up off of the floor.

[1]  Magdalena D. Bugajska,et al.  Building a Multimodal Human-Robot Interface , 2001, IEEE Intell. Syst..

[2]  Matthew Turk,et al.  Multimodal Human-Computer Interaction , 2005 .

[3]  Brian Scassellati Mechanisms of Shared Attention for a Humanoid Robot , 1998 .

[4]  Douglas C. Engelbart,et al.  Display-Selection Techniques for Text Manipulation , 1967 .

[5]  Henrik I. Christensen,et al.  PDA interface for a field robot , 2003, Proceedings 2003 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2003) (Cat. No.03CH37453).

[6]  Julius Ziegler,et al.  Tracking of the Articulated Upper Body on Multi-View Stereo Image Sequences , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[7]  A. Waibel,et al.  MULTIMODAL HUMAN-COMPUTER INTERACTION , 1993 .

[8]  L. Natale,et al.  A Sensitive Approach to Grasping , 2005 .

[9]  Robert D. Howe,et al.  Towards grasping in unstructured environments: grasper compliance and configuration optimization , 2005, Adv. Robotics.

[10]  F. Guterl Design case history: Apple's Macintosh: A small team of little-known designers, challenged to produce a low-cost, exceptionally easy-to-use personal computer, turns out a technical milestone , 1984, IEEE Spectrum.

[11]  Shree K. Nayar,et al.  Catadioptric omnidirectional camera , 1997, Proceedings of IEEE Computer Society Conference on Computer Vision and Pattern Recognition.

[12]  Jiawen Chen,et al.  Pervasive Pose-Aware Applications and Infrastructure , 2003, IEEE Computer Graphics and Applications.

[13]  Cynthia Breazeal,et al.  Working with robots and objects: revisiting deictic reference for achieving spatial common ground , 2006, HRI '06.

[14]  Steven A. Shafer,et al.  XWand: UI for intelligent spaces , 2003, CHI '03.

[15]  Shumeet Baluja,et al.  Non-Intrusive Gaze Tracking Using Artificial Neural Networks , 1993, NIPS.

[16]  Gregory D. Abowd,et al.  A 2-Way Laser-Assisted Selection Scheme for Handhelds in a Physical Environment , 2003, UbiComp.

[17]  William Harwin,et al.  Devices for assisting manipulation: a summary of user task priorities , 1994 .

[18]  Ashutosh Saxena,et al.  Learning to Grasp Novel Objects Using Vision , 2006, ISER.

[19]  Zunaid Kazi,et al.  Knowledge Driven Planning and Multimodal Control of a Telerobot , 1998, Robotica.

[20]  Andrew Wilson,et al.  Pointing in Intelligent Environments with the WorldCursor , 2003, INTERACT.

[21]  Rainer Stiefelhagen,et al.  Real-Time Recognition of 3D-Pointing Gestures for Human-Machine-Interaction , 2003, DAGM-Symposium.