Free Virtual Navigation Using Motor Imagery Through an Asynchronous BrainComputer Interface

In this paper, an asynchronous braincomputer interface is presented that enables the control of a wheelchair in virtual environments using only one motor imagery task. The control is achieved through a graphical intentional control interface with three navigation commands (move forward, turn right, and turn left) which are displayed surrounding a circle. A bar is rotating in the center of the circle, so it points successively to the three possible commands. The user can, by motor imagery, extend this bar length to select the command at which the bar is pointing. Once a command is selected, the virtual wheelchair moves in a continuous way, so the user controls the length of the advance or the amplitude of the turns. Users can voluntarily switch from this interface to a noncontrol interface (and vice versa) when they do not want to generate any command. After performing a cue-based feedback training, three subjects carried out an experiment in which they had to navigate through the same fixed path to reach an objective. The results obtained support the viability of the system.

[1]  Gernot R. Müller-Putz,et al.  Self-Paced (Asynchronous) BCI Control of a Wheelchair in Virtual Environments: A Case Study with a Tetraplegic , 2007, Comput. Intell. Neurosci..

[2]  José del R. Millán,et al.  Continuous Brain-Actuated Control of an Intelligent Wheelchair by Human EEG , 2008 .

[3]  G. Pfurtscheller,et al.  Information transfer rate in a five-classes brain-computer interface , 2001, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[4]  G. Pfurtscheller,et al.  How many people are able to operate an EEG-based brain-computer interface (BCI)? , 2003, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[5]  Francisco Velasco-Álvarez,et al.  Asynchronous Brain-Computer Interface to Navigate in Virtual Environments Using One Motor Imagery , 2009, IWANN.

[6]  José del R. Millán,et al.  Brain-actuated interaction , 2004, Artif. Intell..

[7]  Gert Pfurtscheller,et al.  Self-paced exploration of the Austrian National Library through thought , 2007 .

[8]  Klaus-Robert Müller,et al.  Machine learning for real-time single-trial EEG-analysis: From brain–computer interfacing to mental state monitoring , 2008, Journal of Neuroscience Methods.

[9]  G. Pfurtscheller,et al.  Rapid prototyping of an EEG-based brain-computer interface (BCI) , 2001, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[10]  Javier Minguez,et al.  Human brain-teleoperated robot between remote places , 2009, 2009 IEEE International Conference on Robotics and Automation.

[11]  S. Voloshynovskiy,et al.  EEG-Based Synchronized Brain-Computer Interfaces: A Model for Optimizing the Number of Mental Tasks , 2007, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[12]  Francisco Velasco-Álvarez,et al.  A two-class brain computer interface to freely navigate through virtual worlds / Ein Zwei-Klassen-Brain-Computer-Interface zur freien Navigation durch virtuelle Welten , 2009, Biomedizinische Technik. Biomedical engineering.

[13]  John Q. Gan,et al.  Asynchronous BCI Control of a Robot Simulator with Supervised Online Training , 2007, IDEAL.

[14]  Christa Neuper,et al.  A scanning protocol for a sensorimotor rhythm-based brain–computer interface , 2009, Biological Psychology.

[15]  Horst Bischof,et al.  Toward Self-Paced Brain–Computer Communication: Navigation Through Virtual Worlds , 2008, IEEE Transactions on Biomedical Engineering.

[16]  José del R. Millán,et al.  An Introduction to Brain-Computer Interfacing , 2007 .

[17]  M. Nuttin,et al.  A brain-actuated wheelchair: Asynchronous and non-invasive Brain–computer interfaces for continuous control of robots , 2008, Clinical Neurophysiology.

[18]  Christa Neuper,et al.  Motor imagery and ERD , 1999 .

[19]  J. Azorín,et al.  Interfaces y Sistemas en Rehabilitación y Compensación Funcional para la Autonomía Personal y la Terapia Clínica , 2011 .