Interactive Multimodal Robot Programming

As robots enter the human environment and come into contact with inexperienced users, they need to be able to interact with users in a multimodal fashion—keyboard and mouse are no longer acceptable as the only input modalities. In this paper we introduce a novel approach for programming robots interactively through a multimodal interface. The key characteristic of this approach is that the user can provide feedback interactively at any time—during both the programming and the execution phase. The framework takes a three-step approach to the problem: multimodal recognition, intention interpretation, and prioritized task execution. The multimodal recognition module translates hand gestures and spontaneous speech into a structured symbolic data stream without abstracting away the user’s intent. The intention interpretation module selects the appropriate primitives to generate a task based on the user’s input, the system’s current state, and robot sensor data. Finally, the prioritized task execution module selects and executes skill primitives based on the system’s current state, sensor inputs, and prior tasks. The framework is demonstrated by interactively controlling and programming a vacuum-cleaning robot. The demonstrations are used to exemplify the interactive programming and the plan recognition aspect of the research.

[1]  R. Stallman EMACS the extensible, customizable self-documenting display editor , 1981, SIGPLAN SIGOA Symposium on Text Manipulation.

[2]  Jean Ponce,et al.  Computer Vision: A Modern Approach , 2002 .

[3]  Horst-Michael Groß,et al.  Neural Architecture for Gesture-Based Human-Machine-Interaction , 1997, Gesture Workshop.

[4]  Claus Bahlmann,et al.  Measuring HMM similarity with the Bayes probability of error and its application to online handwriting recognition , 2001, Proceedings of Sixth International Conference on Document Analysis and Recognition.

[5]  George Musser Robots That Suck , 2003 .

[6]  Herman Roodenburg,et al.  A Cultural History of Gesture , 1992 .

[7]  Illah Nourbakhsh,et al.  Path planning for the Cye personal robot , 2000, Proceedings. 2000 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS 2000) (Cat. No.00CH37113).

[8]  Richard M. Voyles,et al.  Gesture-based programming for robotics: human-augmented software adaptation , 1999, IEEE Intell. Syst..

[9]  Pradeep K. Khosla,et al.  A software architecture-based human-machine interface for reconfigurable sensor-based control systems , 1993, Proceedings of 8th IEEE International Symposium on Intelligent Control.

[10]  Terrence Fong,et al.  Novel interfaces for remote driving: gesture, haptic, and PDA , 2001, SPIE Optics East.

[11]  Hara Prasad Das,et al.  Telemanipulator and Telepresence Technologies , 1995 .

[12]  Masahiro Fujita,et al.  Motion entertainment by a small humanoid robot based on OPEN-R , 2001, Proceedings 2001 IEEE/RSJ International Conference on Intelligent Robots and Systems. Expanding the Societal Role of Robotics in the the Next Millennium (Cat. No.01CH37180).

[13]  Sebastian Thrun,et al.  A Gesture Based Interface for Human-Robot Interaction , 2000, Auton. Robots.

[14]  Alexander Zelinsky,et al.  An algorithm for real-time stereo vision implementation of head pose and gaze direction measurement , 2000, Proceedings Fourth IEEE International Conference on Automatic Face and Gesture Recognition (Cat. No. PR00580).

[15]  H. Kimura,et al.  Acquiring hand-action models in task and behavior levels by a learning robot through observing human demonstrations , 2000 .

[16]  Hiroaki Kitano,et al.  Development of an Autonomous Quadruped Robot for Robot Entertainment , 1998, Auton. Robots.

[17]  David Kortenkamp,et al.  Recognizing and Interpreting Gestures on a Mobile Robot , 1996, AAAI/IAAI, Vol. 2.

[18]  Kanti V. Mardia,et al.  Statistics of Directional Data , 1972 .

[19]  Pradeep K. Khosla,et al.  Tropism-based cognition for the interpretation of context-dependent gestures , 1997, Proceedings of International Conference on Robotics and Automation.

[20]  Lawrence R. Rabiner,et al.  A tutorial on hidden Markov models and selected applications in speech recognition , 1989, Proc. IEEE.

[21]  Wolfram Burgard,et al.  MINERVA: a second-generation museum tour-guide robot , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[22]  Kosei Kitagaki,et al.  A Telerobotics System using Planning Functions Based on Manipulation Skills and Teaching-by-Demonastration Technique in VR , 2000 .

[23]  D. Kortenkamp,et al.  Distributed, autonomous control of Space habitats , 2001, 2001 IEEE Aerospace Conference Proceedings (Cat. No.01TH8542).

[24]  Alex Pentland,et al.  Real-Time American Sign Language Recognition Using Desk and Wearable Computer Based Video , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[25]  Katsushi Ikeuchi,et al.  Task-model based human robot cooperation using vision , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[26]  David R. Barstow,et al.  Interactive programming environments , 1984, SOEN.

[27]  Narendra Ahuja,et al.  Face Detection and Gesture Recognition for Human-Computer Interaction , 2001, The International Series in Video Computing.

[28]  Mei-Yuh Hwang,et al.  The SPHINX-II speech recognition system: an overview , 1993, Comput. Speech Lang..

[29]  Vladimir Pavlovic,et al.  Visual Interpretation of Hand Gestures for Human-Computer Interaction: A Review , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[30]  T. Takamori,et al.  Multi-modal human robot interaction for map generation , 2002, Proceedings of the 41st SICE Annual Conference. SICE 2002..

[31]  Kazuhiko Kawamura,et al.  Towards a Unified Framework for Human-Humanoid Interaction , 2000 .

[32]  電子情報通信学会 The Transactions of the Institute of Electronics, Information and Communication Engineers , 1987 .

[33]  Ipke Wachsmuth,et al.  Gesture and Sign Language in Human-Computer Interaction , 1998, Lecture Notes in Computer Science.

[34]  Yoji Yamada,et al.  Human error recovery for a human/robot parts conveyance system , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[35]  Marjorie Skubic,et al.  Communicating with Teams of Cooperative Robots , 2002 .

[36]  Takuichi Nishimura,et al.  Adaptation to gesture performers by an on-line teaching system for spotting recognition of gestures from a time-varying image , 2000, Systems and Computers in Japan.

[37]  Alex Pentland,et al.  Understanding purposeful human motion , 1999, Proceedings IEEE International Workshop on Modelling People. MPeople'99.

[38]  Thea Iberall,et al.  Dextrous robot hands , 1990 .

[39]  Yangsheng Xu,et al.  Online, interactive learning of gestures for human/robot interfaces , 1996, Proceedings of IEEE International Conference on Robotics and Automation.

[40]  K. Mardia Statistics of Directional Data , 1972 .

[41]  Katsushi Ikeuchi,et al.  Toward automatic robot instruction from perception-mapping human grasps to manipulator grasps , 1997, IEEE Trans. Robotics Autom..

[42]  Yoshiaki Shirai,et al.  Interactive gesture interface for intelligent wheelchairs , 2000, 2000 IEEE International Conference on Multimedia and Expo. ICME2000. Proceedings. Latest Advances in the Fast Changing World of Multimedia (Cat. No.00TH8532).

[43]  Y. Matsumotot,et al.  Development of intelligent wheelchair system with face and gaze based interface , 2001, Proceedings 10th IEEE International Workshop on Robot and Human Interactive Communication. ROMAN 2001 (Cat. No.01TH8591).

[44]  Paul E. Rybski,et al.  Interactive task training of a mobile robot through human gesture recognition , 1999, Proceedings 1999 IEEE International Conference on Robotics and Automation (Cat. No.99CH36288C).

[45]  Magdalena D. Bugajska,et al.  Building a Multimodal Human-Robot Interface , 2001, IEEE Intell. Syst..

[46]  Marjorie Skubic,et al.  Using spatial language in a human-robot dialog , 2002, Proceedings 2002 IEEE International Conference on Robotics and Automation (Cat. No.02CH37292).

[47]  H. Harry Asada,et al.  Hand-in-glove human-machine interface and interactive control: task process modeling using dual Petri nets , 1998, Proceedings. 1998 IEEE International Conference on Robotics and Automation (Cat. No.98CH36146).

[48]  Kazuo Tanie,et al.  Human-machine interaction through an intelligent user interface based on contention architecture , 1996, Proceedings 5th IEEE International Workshop on Robot and Human Communication. RO-MAN'96 TSUKUBA.

[49]  Katsushi Ikeuchi,et al.  Toward an assembly plan from observation. I. Task recognition with polyhedral objects , 1994, IEEE Trans. Robotics Autom..

[50]  Pradeep K. Khosla,et al.  Manipulation task primitives for composing robot skills , 1997, Proceedings of International Conference on Robotics and Automation.

[51]  Sharon L. Oviatt,et al.  Taming recognition errors with a multimodal interface , 2000, CACM.

[52]  Michael Vande Weghe,et al.  An architecture for gesture-based control of mobile robots , 1999, Proceedings 1999 IEEE/RSJ International Conference on Intelligent Robots and Systems. Human and Environment Friendly Robots with High Intelligence and Emotional Quotients (Cat. No.99CH36289).

[53]  Lynne E. Parker,et al.  Multi-Robot Systems: From Swarms to Intelligent Automata , 2002, Springer Netherlands.