Issues in measuring the benefits of multimodal interfaces

Multimedia interfaces are rapidly evolving to facilitate human/machine communication. Most of the technologies on which they are based are, as yet, imperfect. But, the interfaces do begin to allow information exchange in ways familiar and comfortable to the human-principally through natural actions in the sensory dimensions of sight, sound and touch. Further, as digital networking becomes ubiquitous, the opportunity grows for collaborative work through conferenced computing. In this context the machine takes on the role of mediator in human/machine/human communication-the ideal being to extend the intellectual abilities of humans through access to distributed information resources and collective decision making. The challenge is how to design machine mediation so that it extends, not impedes, human abilities. This report describes evolving work to incorporate multimodal interfaces into a networked system for collaborative distributed computing. It also addresses strategies for quantifying the synergies that may be gained.

[1]  Kankanahalli Srinivas,et al.  Infrastructure support for multimedia communications: a survey , 1993, [1993] Proceedings Second Workshop on Enabling Technologies@m_Infrastructure for Collaborative Enterprises.

[2]  A. Waibel,et al.  MULTIMODAL INTERPRETER SPEECH GESTURE WRITING DIALOG PROCESSOR MULTIMODAL LEARNING INTERFACES , 1995 .

[3]  W. Wahister One word says more than a thousand pictures: on the automatic verbalization of the results of image sequence analysis system , 1987 .

[4]  Joëlle Coutaz,et al.  Towards automatic evaluation of multimodal user interfaces , 1993, Knowl. Based Syst..

[5]  Vladimir Pavlovic,et al.  A Multimodal framework for Interacting with Virtual Environments , 1996 .

[6]  Caroline Henton,et al.  Saying and seeing it with feeling: techniques for synthesizing visible, emotional speech , 1994, SSW.

[7]  James L. Flanagan,et al.  Technologies for multimedia communications , 1994, Proc. IEEE.

[8]  Stephen Gale,et al.  Human Aspects of Interactive Multimedia Communication , 1990, Interact. Comput..

[9]  Paul McKevitt Integration of Natural Language and Vision Processing: Computational Models and Systems , 1995 .

[10]  Paul Richard,et al.  Multimodal virtual reality: Input-output devices, system integration, and human factors , 1996, Int. J. Hum. Comput. Interact..

[11]  C. Benoît,et al.  A set of French visemes for visual speech synthesis , 1994 .

[12]  Starr Roxanne Hiltz,et al.  The Effects of Distributed Group Support and Process Structuring on Software Requirements Development Teams: Results on Creativity and Quality , 1995, J. Manag. Inf. Syst..

[13]  Arne Jönsson,et al.  Wizard of Oz studies -- why and how , 1993, Knowl. Based Syst..

[14]  Dominic W. Massaro,et al.  Synthesis of visible speech , 1990 .

[15]  Casimir A. Kulikowski,et al.  Iconic Reporting: A New Way of Communicating Radiological Findings , 1995 .

[16]  Qiang Lin,et al.  Infrastructure Support for Multimedia Communication: A Survey , 1994, Int. J. Cooperative Inf. Syst..

[17]  R. Stephenson A and V , 1962, The British journal of ophthalmology.

[18]  J. Davenport Editor , 1960 .

[19]  Judith S. Olson,et al.  Small Group Design Meetings: An Analysis of Collaboration , 1992, Hum. Comput. Interact..

[20]  Steve Whittaker,et al.  Shared Workspaces: How Do They Work and When Are They Useful? , 1993, Int. J. Man Mach. Stud..

[21]  Wolfgang Wahlster,et al.  One word says more than a thousand pictures , 1989 .

[22]  Arne Jönsson,et al.  Wizard of Oz studies: why and how , 1993, IUI '93.

[23]  Judith S. Olson,et al.  Groupwork close up: a comparison of the group design process with and without a simple group editor , 1993, TOIS.

[24]  Richard A. Bolt,et al.  “Put-that-there”: Voice and gesture at the graphics interface , 1980, SIGGRAPH '80.