User interface is a mean to an end —its primary goal is to capture user intent and communicate the results of the requested computation. On today’s devices, user interaction can be achieved through a multiplicity of interaction modalities including speech and visual interfaces. As we evolve toward an increasingly connected world where we access and interact with applications through multiple devices, it becomes crucial that the various access paths to the underlying content be synchronized. This synchronization ensures that the user interacts with the same underlying content independent of the interaction modality — despite the difference in presentation that each modality might impose. It also ensures that the effect of user interaction in any given modality is reflected consistently across all available modalities. We describe an application framework that enables tightly synchronized multimodal user interaction. This framework derives its power from representing the application model in a modality-independent manner, and by traversing this model to produce the various synchronized multimodal views. As the user interaction proceeds, we maintain our current position in the model and update the application data as determined by user intent, then reflect these updates in the various views being presented. We conclude the paper by outlining an example that demonstrates this tightly synchronized multimodal interaction, and describe some of the future challenges in building such multimodal frameworks.
[1]
Stéphane H. Maes,et al.
Multi-modal interaction in the age of information appliances
,
2000,
2000 IEEE International Conference on Multimedia and Expo. ICME2000. Proceedings. Latest Advances in the Fast Changing World of Multimedia (Cat. No.00TH8532).
[2]
T. V. Raman.
Conversational gestures for direct manipulation on the audio desktop
,
1998,
Assets '98.
[3]
Stephen Travis Pope,et al.
A cookbook for using the model-view controller user interface paradigm in Smalltalk-80
,
1988
.
[4]
T. V. Raman.
Auditory User Interfaces: Toward the Speaking Computer
,
1997
.
[5]
J. Meigs,et al.
WHO Technical Report
,
1954,
The Yale Journal of Biology and Medicine.