Multimodal interaction with an autonomous forklift

We describe a multimodal framework for interacting with an autonomous robotic forklift. A key element enabling effective interaction is a wireless, handheld tablet with which a human supervisor can command the forklift using speech and sketch. Most current sketch interfaces treat the canvas as a blank slate. In contrast, our interface uses live and synthesized camera images from the forklift as a canvas, and augments them with object and obstacle information from the world. This connection enables users to "draw on the world," enabling a simpler set of sketched gestures. Our interface supports commands that include summoning the forklift and directing it to lift, transport, and place loads of palletized cargo. We describe an exploratory evaluation of the system designed to identify areas for detailed study. Our framework incorporates external signaling to interact with humans near the vehicle. The robot uses audible and visual annunciation to convey its current state and intended actions. The system also provides seamless autonomy handoff: any human can take control of the robot by entering its cabin, at which point the forklift can be operated manually until the human exits.

[1]  Randall Davis,et al.  Sketch Understanding in Design: Overview of Work at the MIT AI Lab , 2002 .

[2]  Robin R. Murphy,et al.  Moonlight in Miami: a field study of human-robot interaction in the context of an urban search and rescue disaster response training exercise , 2004 .

[3]  Raffaello D'Andrea,et al.  Coordinating Hundreds of Cooperative, Autonomous Vehicles in Warehouses , 2007, AI Mag..

[4]  Hande Kaymaz-Keskinpala,et al.  PDA-based human-robotic interface , 2003, SMC'03 Conference Proceedings. 2003 IEEE International Conference on Systems, Man and Cybernetics. Conference Theme - System Security and Assurance (Cat. No.03CH37483).

[5]  Jean Scholtz,et al.  Awareness in human-robot interactions , 2003, SMC'03 Conference Proceedings. 2003 IEEE International Conference on Systems, Man and Cybernetics. Conference Theme - System Security and Assurance (Cat. No.03CH37483).

[6]  François Michaud,et al.  Egocentric and exocentric teleoperation interface using real-time, 3D video projection , 2009, 2009 4th ACM/IEEE International Conference on Human-Robot Interaction (HRI).

[7]  Magdalena D. Bugajska,et al.  Building a Multimodal Human-Robot Interface , 2001, IEEE Intell. Syst..

[8]  Terrence Fong,et al.  Multi-robot remote driving with collaborative control , 2003, IEEE Trans. Ind. Electron..

[9]  James R. Glass A probabilistic framework for segment-based speech recognition , 2003, Comput. Speech Lang..

[10]  I. Lee Hetherington,et al.  PocketSUMMIT: small-footprint continuous speech recognition , 2007, INTERSPEECH.

[11]  Takafumi Matsumaru Mobile Robot with Preliminary-announcement and Indication Function of Forthcoming Operation using Flat-panel Display , 2007, Proceedings 2007 IEEE International Conference on Robotics and Automation.

[12]  Michael A. Goodrich,et al.  Ecological Interfaces for Improving Mobile Robot Teleoperation , 2007, IEEE Transactions on Robotics.

[13]  Charles E. Thorpe,et al.  PdaDriver: A Handheld System for Remote Driving , 2003 .

[14]  Takeo Igarashi,et al.  Sketch and run: a stroke-based interface for home robots , 2009, CHI.

[15]  Tara N. Sainath,et al.  A voice-commandable robotic forklift working alongside humans in minimally-prepared outdoor environments , 2010, 2010 IEEE International Conference on Robotics and Automation.

[16]  Bo Yu,et al.  A domain-independent system for sketch recognition , 2003, GRAPHITE '03.

[17]  Derek Anderson,et al.  Using a hand-drawn sketch to control a team of robots , 2007, Auton. Robots.

[18]  Hande Kaymaz-Keskinpala,et al.  Objective data analysis for a PDA-based human robotic interface , 2004, 2004 IEEE International Conference on Systems, Man and Cybernetics (IEEE Cat. No.04CH37583).

[19]  Randall Davis,et al.  Recognition of Hand Drawn Chemical Diagrams , 2007, AAAI.

[20]  Sharon L. Oviatt,et al.  Unification-based Multimodal Integration , 1997, ACL.

[21]  Takafumi Matsumaru,et al.  Mobile Robot with Eyeball Expression as the Preliminary-Announcement and Display of the Robot’s Following Motion , 2005, Auton. Robots.

[22]  Phil Jones,et al.  Improving traffic behaviour and safety through urban design , 2005 .

[23]  Randall Davis,et al.  Speech and sketching: an empirical study of multimodal interaction , 2007, SBIM '07.

[24]  Rainer Stiefelhagen,et al.  Implementation and evaluation of a constraint-based multimodal fusion system for speech and 3D pointing gestures , 2004, ICMI '04.

[25]  Tracy Anne Hammond,et al.  PaleoSketch: accurate primitive sketch recognition and beautification , 2008, IUI '08.