This paper describes the architecture of the MMI 2 system (Man-Machine Interface for MultiModal Interaction with knowledge based systems). The objective of the MMI 2 project is to develop a toolkit and a method for using the toolkit to construct knowledge based systems (KBS) in different domains which have multimodal interfaces and undertake cooperative dialogue with users. Within the project, a step towards the development of this toolkit is the development of a demonstrator KBS in the domain of computer network design. The project is structured so that the demonstration KBS will act as a shell which contains knowledge about the computer network design domain for the demonstration application, but which can also be used later for other domains when the requisite knowledge about those domains is encoded. The MMI2 project makes a strong commitment to “multimodal” rather than to “multimedia” interaction in the interface. The distinction intended is that a multi-media system is one which uses different presentation media (e.g. text, raster graphics, video, speech) without a commitment to the underlying representation of the information presented. For reasons of efficiency of both storage and processing, individual specialised representations are usually used for information which is intended to be presented in each individual medium, and information can only be presented in a single medium. A “multimodal” system is one which includes several input and output media, but is committed to a single internal representation language for the information. This permits the same information to be presented in any mode, chosen purely by rules which select that mode for that user at that point in task performance as being both sufficiently expressive and most efficient. The modes available in the MMI 2 demonstrator are: for input: English, French and Spanish natural languages, gesture, direct manipulation of graphics, command language; and for output: English, French and Spanish natural languages, graphics (CAD diagrams and business graphics), and non-verbal audio. The representation language used for all information within the system is a promiscuously reified typed logic called the Common Meaning Representation (CMR).
[1]
John R. Lee,et al.
User Interface Management and Design
,
1991,
EurographicSeminars.
[2]
Bart Demoen,et al.
LOKI: A logic Oriented Approach to data and knowledge bases supporting Natural Language Interaction
,
1988
.
[3]
Jerry R. Hobbs.
Ontological Promiscuity
,
1985,
ACL.
[4]
G. E. Pfaff,et al.
User Interface Management Systems
,
1985,
Eurographic Seminars.
[5]
Jock D. Mackinlay,et al.
Automating the design of graphical presentations of relational information
,
1986,
TOGS.
[6]
Françoise Darses.
Constraints in design: Towards a methodology of psychological analysis based on AI formalisms
,
1990,
INTERACT.
[7]
Samir Dami,et al.
An expert-system for French analysis within a multi-mode dialogue to be connected
,
1991,
RIAO.
[8]
Timothy W. Finin,et al.
A general user modeling system
,
1986
.
[9]
P. Falzon,et al.
Architecture of a Multimodal Dialogue Interface for Knowledge-Based Systems
,
1990
.
[10]
Anthony Conway,et al.
Enhanced interaction styles for user interfaces
,
1991,
IEEE Computer Graphics and Applications.
[11]
Karen Sparck Jones.
Realism About User Modeling
,
1989
.
[12]
James A. Pittman,et al.
Recognizing handwritten text
,
1991,
CHI.
[13]
Robert van Liere,et al.
User Interface Management Systems
,
1987,
Advances in Computer Graphics.
[14]
Timothy W. Finin,et al.
Acquiring a model of the user's beliefs from a cooperative advisory dialog
,
1988
.
[15]
Juan Carlos,et al.
Un parser basado en la sintaxis de dependencias
,
1991
.
[16]
D. Bouwhuis,et al.
The Structure of Multimodal Dialogue
,
1989
.