Multimodal fusion for interaction systems

Researchers in computer science and computer engineering devote now a significant part of their efforts in communication and interaction between human and machine. Indeed, with the advent of real -time multimodal and multimedia processing, computer is no longer seen only as a calculation tool, but as a machine of communication processing, a machine that accompanies, assists or promotes many activities in daily life. A multimodal interface allows a more flexible and natural interaction between a user and a co mputing system. It extends the capabilities of this system to better match the natural communication means of human beings. In such interactive system, fusion engines are the fundamental components that interpret input events whose meaning can vary accord ing to a given context. Fusion of events from various communication sources, such as speech, pen, text, gesture, etc. allow the richness of human -machine interaction This research will allow a better understanding of the multimodal fusion and interaction, by the construction of a fusion engine using technologies of semantic web domain. The aim is to develop an expert fusion system for multimodal human machine interaction that will lead to design a monitoring tool for normal persons, seniors and handicaps to ensure their support, at home or outside.

[1]  Philip R. Cohen,et al.  QuickSet: multimodal interaction for distributed applications , 1997, MULTIMEDIA '97.

[2]  Tyson R. Henry,et al.  Integrating gesture and snapping into a user interface toolkit , 1990, UIST '90.

[3]  Michael Johnston,et al.  Unification-based Multimodal Parsing , 1998, ACL.

[4]  Alexander H. Waibel,et al.  Towards Unrestricted Lip Reading , 2000, Int. J. Pattern Recognit. Artif. Intell..

[5]  Jennifer C. Lai,et al.  Examining modality usage in a conversational multimodal application for mobile e-mail access , 2007, Int. J. Speech Technol..

[6]  Dominique Béroule Management of time distortions through rough coincidence detection , 1989, EUROSPEECH.

[7]  Philippe A. Palanque,et al.  Design, specification, and verification of interactive systems , 2001, SOEN.

[8]  Joëlle Coutaz,et al.  MATIS: A multimodal airline travel information system , 1993 .

[9]  Richard A. Bolt,et al.  “Put-that-there”: Voice and gesture at the graphics interface , 1980, SIGGRAPH '80.

[10]  Gregory J. Wolff,et al.  Lipreading by Neural Networks: Visual Preprocessing, Learning, and Sensory Integration , 1993, NIPS.

[11]  Roope Raisamo,et al.  Testing usability of multimodal applications with visually impaired children , 2006, IEEE MultiMedia.

[12]  Monica Reggiani,et al.  A multimodal user interface for remote object exploration in teleoperation systems , 2002 .

[13]  Denis Lalanne,et al.  HephaisTK: a toolkit for rapid prototyping of multimodal interfaces , 2009, ICMI-MLMI '09.

[14]  A BoltRichard,et al.  Put-that-there , 1980 .

[15]  Eric Barboni,et al.  Model-Based Engineering of Widgets, User Applications and Servers Compliant with ARINC 661 Specification , 2006, DSV-IS.

[16]  Harriet J. Nock,et al.  Assessing face and speech consistency for monologue detection in video , 2002, MULTIMEDIA '02.

[17]  Byeong-Seok Shin,et al.  Wearable Multimodal Interface for Helping Visually Handicapped Persons , 2006, ICAT.

[18]  Wolfgang Wahlster,et al.  SmartKom: Foundations of Multimodal Dialogue Systems , 2006, SmartKom.

[19]  Andreas Holzinger,et al.  Accessible Multimodal Web Pages with Sign Language Translations for Deaf and Hard of Hearing Users , 2009, 2009 20th International Workshop on Database and Expert Systems Application.

[20]  Sharon L. Oviatt,et al.  Multimodal Integration - A Statistical View , 1999, IEEE Trans. Multim..

[21]  Wayne Robbins Multimedia Synchronization , 2008, Encyclopedia of Multimedia.

[22]  Yarden Katz,et al.  Pellet: A practical OWL-DL reasoner , 2007, J. Web Semant..