This paper presents a synchronization module in multimodal dialogue system architecture based on the model-view-controller (MVC) pattern for human-computer interaction. The MVC pattern is based on a clear separation of objects into three categories, i.e. model for defining and maintaining data, view for rendering interactions based on the data, and controller for coordinating actions and events that affect the model and view(s). As part of our layered multimodal dialog system architecture, this synchronization module in our approach controls the synchronization of multiple modalities, such as speech, mouse and keyboard, by interpreting XML document that incorporates SMIL and EMMA. It isolates dialog model from complex presentations associated with different channels and user interfaces through the adoption of a generic object binding mechanism. These flexibilities lead to enhanced design freedom in multimodal dialog system architecture that supports client based, sever based and distributed solutions
[1]
Jonathan Engelsma,et al.
An architecture for seamless access to distributed multimodal services
,
2005,
INTERSPEECH.
[2]
Olivier Boëffard,et al.
DORIS, a multiagent/IP platform for multimodal dialogue applications
,
2004,
INTERSPEECH.
[3]
Sharon L. Oviatt,et al.
Ten myths of multimodal interaction
,
1999,
Commun. ACM.
[4]
Feng Liu,et al.
An Extensible Three-tier XML Dialogue System Architecture for Multimodal Interaction and Automated Agent Services
,
2005,
IMSA.
[5]
Nicole Beringer,et al.
Integrated development and on-the-fly simulation of multimodal dialogs
,
2005,
INTERSPEECH.
[6]
Norbert Reithinger,et al.
An integration framework for a mobile multimodal dialogue system accessing the semantic web
,
2005,
INTERSPEECH.
[7]
Anurag Kumar Gupta,et al.
Integration patterns during multimodal interaction
,
2004,
INTERSPEECH.