Ambiguity detection in multimodal systems

Multimodal systems support users to communicate in a natural way according to their needs. However, the naturalness of the interaction implies that it is hard to find one and only one interpretation of the users' input. Consequently the necessity to define methods for users' input interpretation and ambiguity detection is arising. This paper proposes a theoretical approach based on a Constraint Multiset Grammar combined with Linear Logic, for representing and detecting ambiguities, and in particular semantic ambiguities, produced by the user's input. It considers user's input as a set of primitives defined as terminal elements of the grammar, composing multimodal sentences. The Linear Logic is used to define rules that allow detecting ambiguities connected to the semantics of the user's input. In particular, the paper presents the main features of the user's input and connections between the elements belonging to a multimodal sentence, and it enables to detect ambiguities that can arise during their interpretation process.

[1]  Michael Collins,et al.  Three Generative, Lexicalised Models for Statistical Parsing , 1997, ACL.

[2]  Patrizia Grifoni,et al.  A Hybrid Grammar-Based Approach to Multimodal Languages Specification , 2007, OTM Workshops.

[3]  Kim Marriott,et al.  A survey of visual language specification and recognition , 1998 .

[4]  Alexander H. Waibel,et al.  Growing Semantic Grammars , 1998, COLING-ACL.

[5]  Joyce Yue Chai,et al.  Cognitive Principles in Robust Multimodal Interpretation , 2006, J. Artif. Intell. Res..

[6]  Erik Kamsties,et al.  From Contract Drafting to Software Specification: Linguistic Sources of Ambiguity , 2003 .

[7]  Patrick Lincoln,et al.  Linear logic , 1992, SIGA.

[8]  King-Sun Fu,et al.  Error-Correcting Isomorphisms of Attributed Relational Graphs for Pattern Analysis , 1979, IEEE Transactions on Systems, Man, and Cybernetics.

[9]  Kim Marriott,et al.  Automatic construction of user interfaces from constraint multiset grammars , 1995, Proceedings of Symposium on Visual Languages.

[10]  M. Johnston,et al.  Combining stochastic and grammar-based language processing with finite-state edit machines , 2005, IEEE Workshop on Automatic Speech Recognition and Understanding, 2005..

[11]  Michael Johnston,et al.  Finite-state multimodal integration and understanding , 2005, Natural Language Engineering.

[12]  Fernando Ferri,et al.  An Approach for Managing Ambiguities in Multimodal Interaction , 2007, OTM Workshops.

[13]  Bob Carpenter,et al.  The logic of typed feature structures , 1992 .

[14]  Michelle X. Zhou,et al.  A probabilistic approach to reference resolution in multimodal user interfaces , 2004, IUI '04.