Deep Learning the EEG Manifold for Phonological Categorization from Active Thoughts

Speech-related Brain Computer Interfaces (BCI) aim primarily at finding an alternative vocal communication pathway for people with speaking disabilities. As a step towards full decoding of imagined speech from active thoughts, we present a BCI system for subject-independent classification of phonological categories exploiting a novel deep learning based hierarchical feature extraction scheme. To better capture the complex representation of high-dimensional electroencephalography (EEG) data, we compute the joint variability of EEG electrodes into a channel cross-covariance matrix. We then extract the spatio-temporal information encoded within the matrix using a mixed deep neural network strategy. Our model framework is composed of a convolutional neural network (CNN), a long-short term network (LSTM), and a deep autoencoder. We train the individual networks hierarchically, feeding their combined outputs in a final gradient boosting classification step. Our best models achieve an average accuracy of 77.9% across five different binary classification tasks, providing a significant 22.5% improvement over previous methods. As we also show visually, our work demonstrates that the speech imagery EEG possesses significant discriminative information about the intended articulatory movements responsible for natural speech synthesis.

[1]  Makoto Sato,et al.  Spatial filtering and single-trial classification of EEG during vowel speech imagery , 2009, i-CREATe.

[2]  Siyi Deng,et al.  EEG classification of imagined syllable rhythm using Hilbert spectrum methods , 2010, Journal of neural engineering.

[3]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[4]  Yu Zhang,et al.  Analysis and classification of speech imagery EEG for BCI , 2013, Biomed. Signal Process. Control..

[5]  Makoto Sato,et al.  Single-trial classification of vowel speech imagery using common spatial patterns , 2009, Neural Networks.

[6]  Pramit Saha,et al.  Hierarchical Deep Feature Learning For Decoding Imagined Speech From EEG , 2019, AAAI.

[7]  Parisa Ghane Silent speech recognition in EEG-based Brain Computer Interface , 2015 .

[8]  Yann LeCun,et al.  Generalization and network design strategies , 1989 .

[9]  José del R. Millán,et al.  Word pair classification during imagined speech using direct brain recordings , 2016 .

[10]  Ugur Halici,et al.  A novel deep learning approach for classification of EEG motor imagery signals , 2017, Journal of neural engineering.

[11]  R. Netsell SPEECH MOTOR CONTROL AND SELECTED NEUROLOGIC DISORDERS , 1982 .

[12]  Omar Farooq,et al.  Vowel classification using wavelet decomposition during speech imagery , 2016, 2016 3rd International Conference on Signal Processing and Integrated Networks (SPIN).

[13]  G. Pfurtscheller,et al.  Prosthetic Control by an EEG-based Brain-Computer Interface (BCI) , 2001 .

[14]  Christa Neuper,et al.  EEG-Based Brain-Computer Interface , 2006, Neuroergonomics.

[15]  Tanja Schultz,et al.  Automatic Speech Recognition from Neural Signals: A Focused Review , 2016, Front. Neurosci..

[16]  Jun Qin,et al.  Neural networks based EEG-Speech Models , 2016, ArXiv.

[17]  Michael D'Zmura,et al.  Toward EEG Sensing of Imagined Speech , 2009, HCI.

[18]  B. V. K. Vijaya Kumar,et al.  Imagined Speech Classification with EEG Signals for Silent Communication: A Preliminary Investigation into Synthetic Telepathy , 2010, 2010 4th International Conference on Bioinformatics and Biomedical Engineering.

[19]  Snehanshu Saha,et al.  A Communication Paradigm Using Subvocalized Speech: Translating Brain Signals into Speech , 2016, Augmented Human Research.

[20]  Syed M. Saddique,et al.  EEG Based Brain Computer Interface , 2009, J. Softw..

[21]  Frank Rudzicz,et al.  Classifying phonological categories in imagined and articulated speech , 2015, 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[22]  Boreom Lee,et al.  EEG classification in a single-trial basis for vowel speech perception using multivariate empirical mode decomposition , 2014, Journal of neural engineering.

[23]  Lina Yao,et al.  Converting Your Thoughts to Texts: Enabling Brain Typing via Deep Feature Learning of EEG Signals , 2017, 2018 IEEE International Conference on Pervasive Computing and Communications (PerCom).

[24]  Luis Villaseñor Pineda,et al.  Sonification and textification: Proposing methods for classifying unspoken words from EEG signals , 2017, Biomed. Signal Process. Control..