Exploring Large Virtual Environments by Thoughts Using a BrainComputer Interface Based on Motor Imagery and High-Level Commands

Braincomputer interfaces (BCI) are interaction devices that enable users to send commands to a computer by using brain activity only. In this paper, we propose a new interaction technique to enable users to perform complex interaction tasks and to navigate within large virtual environments (VE) by using only a BCI based on imagined movements (motor imagery). This technique enables the user to send high-level mental commands, leaving the application in charge of most of the complex and tedious details of the interaction task. More precisely, it is based on points of interest and enables subjects to send only a few commands to the application in order to navigate from one point of interest to the other. Interestingly enough, the points of interest for a given VE can be generated automatically thanks to the processing of this VE geometry. As the navigation between two points of interest is also automatic, the proposed technique can be used to navigate efficiently by thoughts within any VE. The input of this interaction technique is a newly-designed self-paced BCI which enables the user to send three different commands based on motor imagery. This BCI is based on a fuzzy inference system with reject options. In order to evaluate the efficiency of the proposed interaction technique, we compared it with the state of the art method during a task of virtual museum exploration. The state of the art method uses low-level commands, which means that each mental state of the user is associated with a simple command such as turning left or moving forward in the VE. In contrast, our method based on high-level commands enables the user to simply select its destination, leaving the application performing the necessary movements to reach this destination. Our results showed that with our interaction technique, users can navigate within a virtual museum almost twice as fast as with low-level commands, and with nearly half the commands, meaning with less stress and more comfort for the user. This suggests that our technique enables efficient use of the limited capacity of current motor imagery-based BCI in order to perform complex interaction tasks in VE, opening the way to promising new applications.

[1]  Anil K. Jain,et al.  Feature Selection: Evaluation, Application, and Small Sample Performance , 1997, IEEE Trans. Pattern Anal. Mach. Intell..

[2]  H. Lüders,et al.  American Electroencephalographic Society Guidelines for Standard Electrode Position Nomenclature , 1991, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[3]  J. Knott,et al.  Regarding the American Electroencephalographic Society guidelines for standard electrode position nomenclature: a commentary on the proposal to change the 10-20 electrode designators. , 1993, Journal of clinical neurophysiology : official publication of the American Electroencephalographic Society.

[4]  Gernot R. Müller-Putz,et al.  Self-Paced (Asynchronous) BCI Control of a Wheelchair in Virtual Environments: A Case Study with a Tetraplegic , 2007, Comput. Intell. Neurosci..

[5]  L. Fabien,et al.  Studying the Use of Fuzzy Inference Systems for Motor Imagery Classification , 2007, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[6]  Harold Mouchère,et al.  Pattern rejection strategies for the design of self-paced EEG-based Brain-Computer Interfaces , 2008, 2008 19th International Conference on Pattern Recognition.

[7]  Gary E. Birch,et al.  Towards Development of a 3-State Self-Paced Brain-Computer Interface , 2007, Comput. Intell. Neurosci..

[8]  Rabab K. Ward,et al.  ResearchArticle Towards Development of a 3-State Self-Paced Brain-Computer Interface , 2007 .

[9]  Martin Hachet,et al.  Navidget for Easy 3D Camera Positioning from 2D Inputs , 2008, 2008 IEEE Symposium on 3D User Interfaces.

[10]  G. Pfurtscheller,et al.  Information transfer rate in a five-classes brain-computer interface , 2001, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[11]  Gert Pfurtscheller,et al.  Self-paced exploration of the Austrian National Library through thought , 2007 .

[12]  F. Lotte,et al.  Self-Paced Brain-Computer Interaction with Virtual Worlds: A Quantitative and Qualitative Study "Out of the Lab" , 2008 .

[13]  Mel Slater,et al.  Brain Computer Interface for Virtual Reality Control , 2009, ESANN.

[14]  Silvia Pfleger,et al.  Advances in Human-Computer Interaction , 1995, Research Reports Esprit.

[15]  Josef Kittler,et al.  Floating search methods for feature selection with nonmonotonic criterion functions , 1994, Proceedings of the 12th IAPR International Conference on Pattern Recognition, Vol. 3 - Conference C: Signal Processing (Cat. No.94CH3440-5).

[16]  Sheel Aditya,et al.  Brain-Computer Interface (BCI) Based Musical Composition , 2010, 2010 International Conference on Cyberworlds.

[17]  Christa Neuper,et al.  Walking by Thinking: The Brainwaves Are Crucial, Not the Muscles! , 2006, PRESENCE: Teleoperators and Virtual Environments.

[18]  J.D. Bayliss,et al.  Use of the evoked potential P3 component for control in a virtual apartment , 2003, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[19]  J R Wolpaw,et al.  Spatial filter selection for EEG-based communication. , 1997, Electroencephalography and clinical neurophysiology.

[20]  Ricardo Ron-Angevin,et al.  Brain–computer interface: Changes in performance using virtual reality techniques , 2009, Neuroscience Letters.

[21]  Francisco Velasco-Álvarez,et al.  A two-class brain computer interface to freely navigate through virtual worlds / Ein Zwei-Klassen-Brain-Computer-Interface zur freien Navigation durch virtuelle Welten , 2009, Biomedizinische Technik. Biomedical engineering.

[22]  Iñaki Iturrate,et al.  A Noninvasive Brain-Actuated Wheelchair Based on a P300 Neurophysiological Protocol and Automated Navigation , 2009, IEEE Transactions on Robotics.

[23]  Gert Pfurtscheller,et al.  Navigating Virtual Reality by Thought: What Is It Like? , 2007, PRESENCE: Teleoperators and Virtual Environments.

[24]  B. Graimann,et al.  Why Use A BCI If You Are Healthy ? , 2000 .

[25]  G. Pfurtscheller,et al.  Brain–Computer Communication: Motivation, Aim, and Impact of Exploring a Virtual Apartment , 2007, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[26]  Kenneth A. Kooi,et al.  American electroencephalographic society , 1964 .

[27]  José del R. Millán,et al.  Noninvasive brain-actuated control of a mobile robot by human EEG , 2004, IEEE Transactions on Biomedical Engineering.

[28]  G. Pfurtscheller,et al.  Continuous EEG classification during motor imagery-simulation of an asynchronous BCI , 2004, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[29]  Horst Bischof,et al.  Toward Self-Paced Brain–Computer Communication: Navigation Through Virtual Worlds , 2008, IEEE Transactions on Biomedical Engineering.

[30]  Gert Pfurtscheller,et al.  Motor imagery and direct brain-computer communication , 2001, Proc. IEEE.

[31]  Geoffrey R. Norman,et al.  Comprar Biostatistics: The Bare Essentials with CDROM | David L. Streiner | 9781550093476 | BC Decker , 2007 .

[32]  Geoffrey R. Norman,et al.  Biostatistics: The Bare Essentials , 1993 .

[33]  Klaus-Robert Müller,et al.  The non-invasive Berlin Brain–Computer Interface: Fast acquisition of effective performance in untrained subjects , 2007, NeuroImage.

[34]  Anatole Lécuyer,et al.  FuRIA: An Inverse Solution Based Feature Extraction Algorithm Using Fuzzy Set Theory for Brain–Computer Interfaces , 2009, IEEE Transactions on Signal Processing.

[35]  Anatole Lécuyer,et al.  Classifying EEG for brain computer interfaces using Gaussian processes , 2008, Pattern Recognit. Lett..

[36]  Fabrice Lamarche,et al.  TopoPlan: a topological path planner for real time human navigation under floor and ceiling constraints , 2009, Comput. Graph. Forum.

[37]  Hideaki Touyama Brain-CAVE Interface Based on Steady-State Visual Evoked Potential , 2008 .

[38]  Sung Chan Jun,et al.  Feasibility of approaches combining sensor and source features in brain–computer interface , 2012, Journal of Neuroscience Methods.

[39]  Michitaka Hirose,et al.  Brain-Computer Interfaces, Virtual Reality, and Videogames , 2008, Computer.

[40]  John R. Smith,et al.  Steady-State VEP-Based Brain-Computer Interface Control in an Immersive 3D Gaming Environment , 2005, EURASIP J. Adv. Signal Process..

[41]  G. Pfurtscheller,et al.  Brain-Computer Interfaces for Communication and Control. , 2011, Communications of the ACM.

[42]  SteedAnthony,et al.  Navigating virtual reality by thought , 2007 .

[43]  Christian Laugier,et al.  Controlling a Wheelchair Indoors Using Thought , 2007, IEEE Intelligent Systems.

[44]  K.-R. Muller,et al.  Optimizing Spatial filters for Robust EEG Single-Trial Analysis , 2008, IEEE Signal Processing Magazine.