An error-aware gaze-based keyboard by means of a hybrid BCI system

Gaze-based keyboards offer a flexible way for human-computer interaction in both disabled and able-bodied people. Besides their convenience, they still lead to error-prone human-computer interaction. Eye tracking devices may misinterpret user’s gaze resulting in typesetting errors, especially when operated in fast mode. As a potential remedy, we present a novel error detection system that aggregates the decision from two distinct subsystems, each one dealing with disparate data streams. The first subsystem operates on gaze-related measurements and exploits the eye-transition pattern to flag a typo. The second, is a brain-computer interface that utilizes a neural response, known as Error-Related Potentials (ErrPs), which is inherently generated whenever the subject observes an erroneous action. Based on the experimental data gathered from 10 participants under a spontaneous typesetting scenario, we first demonstrate that ErrP-based Brain Computer Interfaces can be indeed useful in the context of gaze-based typesetting, despite the putative contamination of EEG activity from the eye-movement artefact. Then, we show that the performance of this subsystem can be further improved by considering also the error detection from the gaze-related subsystem. Finally, the proposed bimodal error detection system is shown to significantly reduce the typesetting time in a gaze-based keyboard.

[1]  Nitesh V. Chawla,et al.  SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..

[2]  Manuel J. A. Eugster,et al.  Natural brain-information interfaces: Recommending information by relevance inferred from human brain signals , 2016, Scientific Reports.

[3]  Reza Fazel-Rezai,et al.  A Review of Hybrid Brain-Computer Interface Systems , 2013, Adv. Hum. Comput. Interact..

[4]  Michael X Cohen,et al.  Analyzing Neural Time Series Data: Theory and Practice , 2014 .

[5]  Dan Wu,et al.  Combining Spatial Filters for the Classification of Single-Trial EEG in a Finger Movement Task , 2007, IEEE Transactions on Biomedical Engineering.

[6]  Yiannis Kompatsiaris,et al.  A Collaborative Representation Approach to Detecting Error-Related Potentials in SSVEP-BCIs , 2017, ACM Multimedia.

[7]  M. Nicolelis,et al.  Brain-Machine Interfaces: From Basic Science to Neuroprostheses and Neurorehabilitation. , 2017, Physiological reviews.

[8]  D. F. Hays,et al.  Table of Integrals, Series, and Products , 1966 .

[9]  Gerhard Tröster,et al.  Eye Movement Analysis for Activity Recognition Using Electrooculography , 2011, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[10]  Boris M. Velichkovsky,et al.  Towards gaze-mediated interaction: Collecting solutions of the "Midas touch problem" , 1997, INTERACT.

[11]  Luca T. Mainardi,et al.  Online Detection of P300 and Error Potentials in a BCI Speller , 2010, Comput. Intell. Neurosci..

[12]  Wolfgang Rosenstiel,et al.  Online Adaptation of a c-VEP Brain-Computer Interface(BCI) Based on Error-Related Potentials and Unsupervised Learning , 2012, PloS one.

[13]  Lucas C. Parra,et al.  Recipes for the linear analysis of EEG , 2005, NeuroImage.

[14]  Boris Reuderink,et al.  Distinguishing between target and nontarget fixations in a visual search task using fixation-related potentials. , 2013, Journal of vision.

[15]  E. W. Sellers,et al.  Toward enhanced P300 speller performance , 2008, Journal of Neuroscience Methods.

[16]  M. Matteucci,et al.  The Utility Metric: A Novel Method to Assess the Overall Performance of Discrete Brain–Computer Interfaces , 2010, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[17]  Wolfgang Rosenstiel,et al.  Online use of error-related potentials in healthy users and people with severe motor impairment increases performance of a P300-BCI , 2012, Clinical Neurophysiology.

[18]  Dawei Song,et al.  On the duality of specific early and late fusion strategies , 2014, 17th International Conference on Information Fusion (FUSION).

[19]  B. Hjorth EEG analysis based on time domain properties. , 1970, Electroencephalography and clinical neurophysiology.

[20]  Steffen Staab,et al.  eyeGUI: A Novel Framework for Eye-Controlled User Interfaces , 2016, NordiCHI.

[21]  Thierry Baccino,et al.  Eye-Fixation-Related Potentials: Insight into Parafoveal Processing , 2005 .

[22]  Yijun Wang,et al.  A high-speed BCI based on code modulation VEP , 2011, Journal of neural engineering.

[23]  H. Ritter,et al.  Toward FRP-Based Brain-Machine Interfaces—Single-Trial Classification of Fixation-Related Potentials , 2016, PloS one.

[24]  Iñaki Iturrate,et al.  Using frequency-domain features for the generalization of EEG error-related potentials among different tasks , 2013, 2013 35th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC).

[25]  Arne Robben,et al.  Towards the detection of error-related potentials and its integration in the context of a P300 speller brain-computer interface , 2012, Neurocomputing.

[26]  Alan Kennedy,et al.  Book Review: Eye Tracking: A Comprehensive Guide to Methods and Measures , 2016, Quarterly journal of experimental psychology.

[27]  P. König,et al.  Combining EEG and eye tracking: identification, characterization, and correction of eye movement artifacts in electroencephalographic data , 2012, Front. Hum. Neurosci..

[28]  Veit Stuphorn,et al.  Role of supplementary eye field in saccade initiation: executive, not direct, control. , 2010, Journal of neurophysiology.

[29]  Ricardo Chavarriaga,et al.  Errare machinale est: the use of error-related potentials in brain-machine interfaces , 2014, Front. Neurosci..

[30]  Marina Schmid,et al.  An Introduction To The Event Related Potential Technique , 2016 .

[31]  A. A. Fedorova,et al.  EEG Negativity in Fixations Used for Gaze-Based Control: Toward Converting Intentions into Actions with an Eye-Brain-Computer Interface , 2016, Front. Neurosci..

[32]  Bin He,et al.  Noninvasive Electroencephalogram Based Control of a Robotic Arm for Reach and Grasp Tasks , 2016, Scientific Reports.

[33]  Chandan Kumar,et al.  GazeTheWeb: A Gaze-Controlled Web Browser , 2017, W4A.

[34]  Chiharu Ishii,et al.  Control of an electric wheelchair with a brain-computer interface headset , 2016, 2016 International Conference on Advanced Mechatronic Systems (ICAMechS).