ABSTRACT Multimodal systems allow humans to interact with machines through multiple modalities such as speech, gesture and gaze. Different multimodal systems that have been developed so far will be discussed in this paper. These include put that there,Cubricon,Xtra, Quickset,RIA with MIND The growing interest in multimodal interface design is inspired in large part by the goals of supporting more transparent, flexible, efficient, and powerfully expressive means of human–computer interaction than in the past. Multimodal interfaces are expected to support a wider range of diverse applications, be usable by a broader spectrum of the average population, and function more reliably under realistic and challenging usage conditions. We also describe a diverse collection of state-of-the-art multimodal systems that process users‟ spoken and gestural input. These applications range from map-based and virtual reality systems for engaging in simulations and training, to field medic systems for mobile use in noisy environments, to web-based transactions and standard text-editing applications that will reshape daily computing and have a significant commercial impact. To realize successful multimodal systems of the future, many key research challenges remain to be addressed. Among these challenges are the development of cognitive theories to guide multimodal system design, and the development of effective natural language processing, dialogue processing, and error-handling techniques. In addition, new multimodal systems will be needed that can function more robustly and adaptively, and with support for collaborative multiperson use. Gesture interpretation can be seen as a way for computers to begin to understand human
[1]
Sharon L. Oviatt,et al.
Designing the User Interface for Multimodal Speech and Pen-Based Gesture Applications: State-of-the-Art Systems and Future Research Directions
,
2000,
Hum. Comput. Interact..
[2]
Jarmo Verho,et al.
A Wearable, Wireless Gaze Tracker with Integrated Selection Command Source for Human‐Computer Interaction
,
2011,
IEEE Transactions on Information Technology in Biomedicine.
[3]
Yael Edan,et al.
A Real-Time Hand Gesture System based on Evolutionary Search
,
2005
.
[4]
Shimei Pan,et al.
Context-based multimodal input understanding in conversational systems
,
2002,
Proceedings. Fourth IEEE International Conference on Multimodal Interfaces.
[5]
Shimei Pan,et al.
Mind: A Context-Based Multimodal Interpretation Framework in Conversational Systems
,
2005
.
[6]
Yan Meng,et al.
Autonomous Self-Reconfiguration of Modular Robots by Evolving a Hierarchical Mechanochemical Model
,
2011,
IEEE Computational Intelligence Magazine.
[7]
Marcelo Worsley,et al.
Multimodal interactive spaces: MagicTV and magicMAP
,
2010,
2010 IEEE Spoken Language Technology Workshop.