Evidential Reasoning for Multimodal Fusion in Human Computer Interaction

Fusion of information from multiple modalities in Human Computer Interfaces (HCI) has gained a lot of attention in recent years, and has far reaching implications in many areas of human-machine interaction. However, a major limitation of current HCI fusion systems is that the fusion process tends to ignore the semantic nature of modalities, which may reinforce, complement or contradict each other over time. Also, most systems are not robust in representing the ambiguity inherent in human gestures. In this work, we investigate an evidential reasoning based approach for intelligent multimodal fusion, and apply this algorithm to a proposed multimodal system consisting of a Hand Gesture sensor and a Brain Computing Interface (BCI). There are three major contributions of this work to the area of human computer interaction. First, we propose an algorithm for reconstruction of the 3D hand pose given a 2D input video. Second, we develop a BCI using Steady State Visually Evoked Potentials, and show how a multimodal system consisting of the two sensors can improve the efficiency and the complexity of the system, while retaining the same levels of accuracy. Finally, we propose a semantic fusion algorithm based on Transferable Belief Models, which can successfully fuse information from these two sensors, to form meaningful concepts and resolve ambiguity. We also analyze this system for robustness under various operating scenarios.

[1]  Simon P. Kelly,et al.  Visual spatial attention control in an independent brain-computer interface , 2005, IEEE Transactions on Biomedical Engineering.

[2]  Tosiyasu L. Kunii,et al.  Model-based analysis of hand posture , 1995, IEEE Computer Graphics and Applications.

[3]  G. Klir Measure of Uncertainty and Information , 2006 .

[4]  Rainer Stiefelhagen,et al.  Implementation and evaluation of a constraint-based multimodal fusion system for speech and 3D pointing gestures , 2004, ICMI '04.

[5]  Michael Johnston,et al.  Integrating multimodal language processing with speech recognition , 2000, INTERSPEECH.

[6]  Xiaorong Gao,et al.  Design and implementation of a brain-computer interface with high transfer rates , 2002, IEEE Transactions on Biomedical Engineering.

[7]  M. Mehta,et al.  MULTIMODAL INPUT FUSION IN HUMAN-COMPUTER INTERACTION On the Example of the NICE Project , 2003 .

[8]  Francesco Carlo Morabito,et al.  Kurtosis, Renyi's Entropy and Independent Component Scalp Maps for the Automatic Artifact Rejection from EEG data , 2008 .

[9]  Richard M. Dansereau,et al.  Robust joint audio-video talker localization in video conferencing using reliability information-II: Bayesian network fusion , 2005, IEEE Transactions on Instrumentation and Measurement.

[10]  Philippe Smets,et al.  The Combination of Evidence in the Transferable Belief Model , 1990, IEEE Trans. Pattern Anal. Mach. Intell..

[11]  Philippe Smets,et al.  BELIEF FUNCTIONS AND THE TRANSFERABLE BELIEF MODEL , 2000 .

[12]  Glenn Shafer,et al.  A Mathematical Theory of Evidence , 2020, A Mathematical Theory of Evidence.

[13]  M. Ramasubba Reddy,et al.  Design and implementation of High Performance Visual Stimulator for Brain Computer Interfaces , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[14]  P. de Chazal,et al.  Parametric models and spectral analysis for classification in brain-computer interfaces , 2002, 2002 14th International Conference on Digital Signal Processing Proceedings. DSP 2002 (Cat. No.02TH8628).

[15]  Anders Green,et al.  Designing for learnability in human-robot communication , 2003, IEEE Trans. Ind. Electron..

[16]  Jayantha Katupitiya,et al.  Curvature based hand shape recognition for a virtual wheelchair control interface , 2004, IEEE International Conference on Robotics and Automation, 2004. Proceedings. ICRA '04. 2004.

[17]  Chalapathy Neti,et al.  Multistage information fusion for audio-visual speech recognition , 2004, 2004 IEEE International Conference on Multimedia and Expo (ICME) (IEEE Cat. No.04TH8763).

[18]  M Kuba,et al.  Model of visually evoked cortical potentials. , 2002, Physiological research.

[19]  S.Q. Xie,et al.  Computer-brain interface , 2005, Proceedings. 2005 First International Conference on Neural Interface and Control, 2005..

[20]  A. Kramarenko,et al.  Brief Communication VALIDITY OF SPECTRAL ANALYSIS OF EVOKED POTENTIALS IN BRAIN RESEARCH , 2002, The International journal of neuroscience.

[21]  N. Birbaumer,et al.  BCI2000: a general-purpose brain-computer interface (BCI) system , 2004, IEEE Transactions on Biomedical Engineering.

[22]  Robin Hanson,et al.  Bayesian Classification with Correlation and Inheritance , 1991, IJCAI.

[23]  P. Djurić,et al.  Model selection by cross-validation , 1990, IEEE International Symposium on Circuits and Systems.

[24]  G. Pfurtscheller,et al.  Brain-Computer Interfaces for Communication and Control. , 2011, Communications of the ACM.

[25]  Shahzad Malik,et al.  Visual touchpad: a two-handed gestural input device , 2004, ICMI '04.

[26]  Edward Y. Chang,et al.  Optimal multimodal fusion for multimedia data analysis , 2004, MULTIMEDIA '04.

[27]  Mandyam D. Srinath,et al.  Multichannel fusion models for the parametric classification of differential brain activity , 2005, IEEE Transactions on Biomedical Engineering.

[28]  J. Shao Linear Model Selection by Cross-validation , 1993 .

[29]  Po-Lei Lee,et al.  Visual evoked potential actuated brain computer interface: a brain-actuated cursor system , 2005 .

[30]  Mohamed S. Kamel,et al.  Understanding hand gestures using approximate graph matching , 2005, IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans.

[31]  Paulo R. S. Mendonça,et al.  Model-Based Hand Tracking Using an Unscented Kalman Filter , 2001, BMVC.

[32]  Steven K. Feiner,et al.  Mutual disambiguation of 3D multimodal interaction in augmented and virtual reality , 2003, ICMI '03.

[33]  John F. Sowa,et al.  Knowledge representation: logical, philosophical, and computational foundations , 2000 .

[34]  Mounia Lalmas,et al.  Using Dempster-Shafer's Theory of Evidence to Combine Aspects of Information Use , 2004, Journal of Intelligent Information Systems.

[35]  George J. Klir 8 MEASURES OF UNCERTAINTY AND INFORMATION , .

[36]  Lotfi A. Zadeh,et al.  A Simple View of the Dempster-Shafer Theory of Evidence and Its Implication for the Rule of Combination , 1985, AI Mag..

[37]  R. Herpers,et al.  Fast hand gesture recognition for real-time teleconferencing applications , 2001, Proceedings IEEE ICCV Workshop on Recognition, Analysis, and Tracking of Faces and Gestures in Real-Time Systems.

[38]  Pramod K. Varshney,et al.  An adaptive multimodal biometric management algorithm , 2005, IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews).

[39]  Tim Morris,et al.  Hand Segmentation from Live Video , 2002 .

[40]  Yang Liu,et al.  A robust hand tracking for gesture-based interaction of wearable computers , 2004, Eighth International Symposium on Wearable Computers.

[41]  Jerome J. Braun Dempster-Shafer theory and Bayesian reasoning in multisensor data fusion , 2000, SPIE Defense + Commercial Sensing.

[42]  G.E. Birch,et al.  A general framework for brain-computer interface design , 2003, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[43]  Emerson Cabrera Paraiso,et al.  An intelligent speech interface for personal assistants in R&D projects , 2005, Proceedings of the Ninth International Conference on Computer Supported Cooperative Work in Design, 2005..

[44]  James C. Bezdek,et al.  Uncertainty measures for evidential reasoning II: A new measure of total uncertainty , 1993, Int. J. Approx. Reason..

[45]  Yang Chen,et al.  An evidential reasoning approach to composite combat identification (CCID) , 2004, 2004 IEEE Aerospace Conference Proceedings (IEEE Cat. No.04TH8720).

[46]  H. Lüders,et al.  American Electroencephalographic Society Guidelines for Standard Electrode Position Nomenclature , 1991, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[47]  Mohamed S. Kamel,et al.  Knowledge-based disambiguation of hand gestures , 2002, IEEE International Conference on Systems, Man and Cybernetics.

[48]  John K. Tsotsos,et al.  SAVI: an actively controlled teleconferencing system , 2001, Image Vis. Comput..

[49]  Xiaohong Yuan,et al.  Engine fault diagnosis based on multi-sensor information fusion using Dempster-Shafer evidence theory , 2007, Inf. Fusion.

[50]  Shangkai Gao,et al.  A practical VEP-based brain-computer interface. , 2006, IEEE transactions on neural systems and rehabilitation engineering : a publication of the IEEE Engineering in Medicine and Biology Society.