Sparse Autoencoders for Word Decoding from Magnetoencephalography

An open problem in the processing and analysis of neuroimaging data is finding the optimal set of features from which to decode or study underlying neural processes. Handcrafted features sets have been examined in the past with varying degrees of success. Here, we explore sparse autoencoders, a method to extract high-level feature representations of neural data in an unsupervised way, without any manual feature engineering. We pair this unsupervised feature creation step with a L2 regularized regression, which is robust to overfitting. We apply this method to the task of single word decoding from Magnetoencephalography (MEG) data which is known for its high degree of noise and complexity. We show that, for some subjects, sparse autoencoding is advantageous, though it does not produce an overall increase in decoding accuracy. Our results represent the first step towards a neural-network system that could join feature extraction and decoding into one powerful joint system.

[1]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[2]  S. Taulu,et al.  The Signal Space Separation method , 2004, physics/0401166.

[3]  Geoffrey E. Hinton,et al.  Reducing the Dimensionality of Data with Neural Networks , 2006, Science.

[4]  R. Ilmoniemi,et al.  Signal-space projection method for separating MEG or EEG into components , 1997, Medical and Biological Engineering and Computing.

[5]  S. Taulu,et al.  Spatiotemporal signal space separation method for rejecting nearby interference in MEG measurements , 2006, Physics in medicine and biology.

[6]  Riitta Hari,et al.  Removal of magnetoencephalographic artifacts with temporal signal‐space separation: Demonstration with single‐trial auditory‐evoked responses , 2009, Human brain mapping.

[7]  Honglak Lee,et al.  An Analysis of Single-Layer Networks in Unsupervised Feature Learning , 2011, AISTATS.

[8]  Tom M. Mitchell,et al.  Decoding Word Semantics from Magnetoencephalography Time Series Transformations , 2012 .

[9]  Riitta Salmelin,et al.  Tracking neural coding of perceptual and semantic features of concrete nouns , 2012, NeuroImage.

[10]  Wojciech Zaremba,et al.  Modeling the variability of EEG/MEG data through statistical machine learning , 2012 .

[11]  Yann LeCun,et al.  Saturating Auto-Encoders , 2013, ICLR 2013.

[12]  Marc'Aurelio Ranzato,et al.  Building high-level features using large scale unsupervised learning , 2011, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[13]  Yann LeCun,et al.  Saturating auto-encoders: International Conference on Learning Representations, ICLR 2013 , 2013, ICLR 2013.

[14]  Anthony Maida,et al.  Natural Image Bases to Represent Neuroimaging Data , 2013, ICML.