Toward EEG Sensing of Imagined Speech

Might EEG measured while one imagines words or sentences provide enough information for one to identify what is being thought? Analysis of EEG data from an experiment in which two syllables are spoken in imagination in one of three rhythms shows that information is present in EEG alpha, beta and theta bands. Envelopes are used to compute filters matched to a particular experimental condition; the filters' action on data from a particular trial lets one determine the experimental condition used for that trial with appreciably greater-than-chance performance. Informative spectral features within bands lead us to current work with EEG spectrograms.

[1]  Adrian D. C. Chan,et al.  Multiexpert automatic speech recognition using acoustic and myoelectric signals , 2006, IEEE Transactions on Biomedical Engineering.

[2]  N. Bu,et al.  Phoneme Classification for Speech Synthesiser using Differential EMG Signals between Muscles , 2005, 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference.

[3]  B. Hudgins,et al.  Hidden Markov model classification of myoelectric signals in speech , 2001, IEEE Engineering in Medicine and Biology Magazine.

[4]  P Suppes,et al.  Brain-wave recognition of sentences. , 1998, Proceedings of the National Academy of Sciences of the United States of America.

[5]  Gabriel Curio,et al.  Differential effects of overt, covert and replayed speech on vowel-evoked responses of the human auditory cortex , 1999, Neuroscience Letters.

[6]  M S Morse,et al.  Research summary of a scheme to ascertain the availability of speech information in the myoelectric signals of neck and head muscles using surface electrodes. , 1986, Computers in biology and medicine.

[7]  P Suppes,et al.  Brain wave recognition of words. , 1997, Proceedings of the National Academy of Sciences of the United States of America.

[8]  D. Poeppel,et al.  The cortical organization of speech processing , 2007, Nature Reviews Neuroscience.

[9]  E Ahissar,et al.  Speech comprehension is correlated with temporal response patterns recorded from auditory cortex , 2001, Proceedings of the National Academy of Sciences of the United States of America.

[10]  D. Poeppel,et al.  Speech perception at the interface of neurobiology and linguistics , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[11]  Kim Binsted,et al.  Web Browser Control Using EMG Based Sub Vocal Speech Recognition , 2005, Proceedings of the 38th Annual Hawaii International Conference on System Sciences.

[12]  E. Donchin,et al.  Talking off the top of your head: toward a mental prosthesis utilizing event-related brain potentials. , 1988, Electroencephalography and clinical neurophysiology.

[13]  D. Poeppel,et al.  Dorsal and ventral streams: a framework for understanding aspects of the functional anatomy of language , 2004, Cognition.

[14]  D. D. Lee,et al.  Sub auditory speech recognition based on EMG signals , 2003, Proceedings of the International Joint Conference on Neural Networks, 2003..

[15]  J. Fermaglich Electric Fields of the Brain: The Neurophysics of EEG , 1982 .

[16]  Tanja Schultz,et al.  Towards continuous speech recognition using surface electromyography , 2006, INTERSPEECH.

[17]  W. Levelt,et al.  The spatial and temporal signatures of word production components , 2004, Cognition.

[18]  E. Dewan Occipital Alpha Rhythm Eye Position and Lens Accommodation , 1967, Nature.

[19]  Tanja Schultz,et al.  Sub-word unit based non-audible speech recognition using surface electromyography , 2006, INTERSPEECH.

[20]  E. Fetz Volitional control of neural activity: implications for brain–computer interfaces , 2007, The Journal of physiology.

[21]  Lena Maier-Hein,et al.  Articulatory Feature Classification using Surface Electromyography , 2006, 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings.

[22]  L. Maier-Hein,et al.  Session independent non-audible speech recognition using surface electromyography , 2005, IEEE Workshop on Automatic Speech Recognition and Understanding, 2005..

[23]  D. Poeppel,et al.  Phase Patterns of Neuronal Responses Reliably Discriminate Speech in Human Auditory Cortex , 2007, Neuron.

[24]  M. Merzenich,et al.  Modulation of the Auditory Cortex during Speech: An MEG Study , 2002, Journal of Cognitive Neuroscience.

[25]  Charles Jorgensen,et al.  Small Vocabulary Recognition Using Surface Electromyography in an Acoustically Harsh Environment , 2005 .