Evaluation of multimodal fusion methods using integration patterns modeling

Multimodal interaction has been proven as a key to more natural and flexible human-computer interaction. Research of multimodal fusion algorithms still remains highly dependent on usability testing as the only option for their proper evaluation. Usability testing is, however, a very demanding task with many drawbacks hindering further research of advanced methods and more robust algorithms in the domain. In the present research, a novel approach of automated evaluation and testing of decision-level fusion methods is introduced. The key feature behind the approach is modeling of users' integration patterns derived from a series of existing empirical studies. The proposed solution provides researchers with ability to evaluate their fusion algorithms automatically, and more importantly, without the need of expensive usability testing. Moreover, the method delivers other considerable advantages as a reproducibility of tests, high-level parameterization, and possibility to design a large scale of user models.

[1]  Sharon L. Oviatt,et al.  Toward a theory of organized multimodal integration patterns during human-computer interaction , 2003, ICMI '03.

[2]  Michael Johnston,et al.  Articles: Robust Understanding in Multimodal Interfaces , 2009, CL.

[3]  Philip R. Cohen,et al.  QuickSet: multimodal interaction for distributed applications , 1997, MULTIMEDIA '97.

[4]  Sharon L. Oviatt,et al.  Multimodal interfaces for dynamic interactive maps , 1996, CHI.

[5]  Sharon L. Oviatt,et al.  Perceptual user interfaces: multimodal interfaces that process what comes naturally , 2000, CACM.

[6]  Michael Johnston,et al.  Finite-state multimodal integration and understanding , 2005, Natural Language Engineering.

[7]  Sharon L. Oviatt,et al.  STAMP: a suite of tools for analyzing multimodal system processing , 1998, ICSLP.

[8]  Sharon L. Oviatt,et al.  Combining User Modeling and Machine Learning to Predict Users' Multimodal Integration Patterns , 2006, MLMI.

[9]  Sharon L. Oviatt,et al.  When do we interact multimodally?: cognitive load and multimodal communication patterns , 2004, ICMI '04.

[10]  Sharon L. Oviatt,et al.  Individual differences in multimodal integration patterns: what are they and why do they exist? , 2005, CHI.

[11]  James A. Larson,et al.  Guidelines for multimodal user interface design , 2004, CACM.

[12]  Sharon L. Oviatt,et al.  A rapid semi-automatic simulation technique for investigating interactive speech and handwriting , 1992, ICSLP.

[13]  Sharon L. Oviatt,et al.  Mutual disambiguation of recognition errors in a multimodel architecture , 1999, CHI '99.

[14]  Denis Lalanne,et al.  Benchmarking fusion engines of multimodal interactive systems , 2009, ICMI-MLMI '09.

[15]  Carlos Duarte,et al.  An Evaluation Framework for Assessing and Optimizing Multimodal Fusion Engines Performance , 2012, 2012 Sixth International Conference on Complex, Intelligent, and Software Intensive Systems.