Multi-command Chest Tactile Brain Computer Interface for Small Vehicle Robot Navigation

The presented study explores the extent to which tactile stimuli delivered to five chest positions of a healthy user can serve as a platform for a brain computer interface (BCI) that could be used in an interactive application such as robotic vehicle operation. The five chest locations are used to evoke tactile brain potential responses, thus defining a tactile brain computer interface (tBCI). Experimental results with five subjects performing online tBCI provide a validation of the chest location tBCI paradigm, while the feasibility of the concept is illuminated through information-transfer rates. Additionally an offline classification improvement with a linear SVM classifier is presented through the case study.

[1]  Hitoshi Ogawa,et al.  Multi-command Tactile Brain Computer Interface: A Feasibility Study , 2013, HAID.

[2]  J. Randall Flanagan,et al.  Coding and use of tactile signals from the fingertips in object manipulation tasks , 2009, Nature Reviews Neuroscience.

[3]  Fred Plum,et al.  [The diagnosis of stupor and coma]. , 2015, Brain and nerve = Shinkei kenkyu no shinpo.

[4]  B. Ripley,et al.  Pattern Recognition , 1968, Nature.

[5]  G. Pfurtscheller,et al.  Steady-state somatosensory evoked potentials: suitable brain signals for brain-computer interfaces? , 2006, IEEE Transactions on Neural Systems and Rehabilitation Engineering.

[6]  Danilo P. Mandic,et al.  Multi-command Tactile and Auditory Brain Computer Interface based on Head Position Stimulation , 2013, ArXiv.

[7]  N. Birbaumer,et al.  An auditory oddball brain–computer interface for binary choices , 2010, Clinical Neurophysiology.

[8]  Peter Desain,et al.  Introducing the tactile speller: an ERP-based brain–computer interface for communication , 2012, Journal of neural engineering.

[9]  B. Blankertz,et al.  A New Auditory Multi-Class Brain-Computer Interface Paradigm: Spatial Hearing as an Informative Cue , 2010, PloS one.

[10]  Chih-Jen Lin,et al.  LIBLINEAR: A Library for Large Linear Classification , 2008, J. Mach. Learn. Res..

[11]  Victor V. Kryssanov,et al.  Vibrotactile stimulus frequency optimization for the haptic BCI prototype , 2012, The 6th International Conference on Soft Computing and Intelligent Systems, and The 13th International Symposium on Advanced Intelligence Systems.

[12]  Andrzej Cichocki,et al.  Spatial auditory paradigms for brain computer/machine interfacing , 2009 .

[13]  J. Wolpaw,et al.  Brain-Computer Interfaces: Principles and Practice , 2012 .

[14]  Sergios Theodoridis,et al.  Pattern Recognition, Fourth Edition , 2008 .

[15]  Dean J Krusienski,et al.  A comparison of classification techniques for the P300 Speller , 2006, Journal of neural engineering.

[16]  Tobias Kaufmann,et al.  Comparison of tactile, auditory, and visual modality for brain-computer interface use: a case study with a patient in the locked-in state , 2013, Front. Neurosci..

[17]  Gerwin Schalk,et al.  A Practical Guide to Brain–Computer Interfacing with BCI2000: General-Purpose Software for Brain-Computer Interface Research, Data Acquisition, Stimulus Presentation, and Brain Monitoring , 2010 .

[18]  Jonathan R. Wolpaw,et al.  Brain–Computer InterfacesPrinciples and Practice , 2012 .

[19]  Jan B. F. van Erp,et al.  A Tactile P300 Brain-Computer Interface , 2010, Front. Neurosci..

[20]  Valer Jurcak,et al.  10/20, 10/10, and 10/5 systems revisited: Their validity as relative head-surface-based positioning systems , 2007, NeuroImage.

[21]  Ernst Fernando Lopes Da Silva Niedermeyer,et al.  Electroencephalography, basic principles, clinical applications, and related fields , 1982 .

[22]  C. Rosenberg,et al.  Electroencephalography: Basic Principles, Clinical Applications, and Related Fields, 3rd Ed. , 1994 .