Kernel based machine learning framework for neural decoding

Brain machine interfaces (BMI) have attracted intensive attention as a promising technology to aid disabled humans. However, the neural system is highly distributed, dynamic and complex system containing millions of neurons that are functionally interconnected. How to best interface the neural system with human engineeried technology is a critical and challenging problem. These issues motivate our research in neural decoding that is a significant step to realize useful BMI. In this dissertation, we aim to design a kernel-based machine learning framework to address a set of challenges in characterizing neural activity, decoding the information of sensory or behavioral states and controlling neural spatiotemporal patterns. Our contributions can be summarized as follows. First, we propose a nonlinear adaptive spike train decoder based on the kernel least mean square (KLMS) algorithm applied directly on the space of spike trains. Instead of using a binned representation of spike trains, we transform the vector of spike times into a function in reproducing kernel Hilbert space (RKHS), where the inner product of two sets of spike times is defined by the Schoenberg kernel, which encapsulates the statistical description of the point process that generates the spike trains, and bypasses the curse of dimensionality-resolution of the other spike representations. The simulation results indicate that our decoder has advantages in both computation time and accuracy, when the application requires fine time resolution. Secondly, the precise control of the firing pattern in a population of neurons via applied electrical stimulation is a challenge due to the sparseness of spiking responses and neural system plasticity. In this work, we propose a multiple-input-multiple-output (MIMO) adaptive inverse control scheme that operates on spike trains in a RKHS. The control scheme uses an inverse controller to approximate the neural circuit’s inverse. The proposed control system takes advantage of the precise timing of the neural events using the Schoenberg kernel based decoding methodology we proposed before. During operation, the adaptation of the controller minimizes a difference defined in the spike train RKHS between the system output and the target response and keeps the inverse controller close to the inverse of the current neural circuit, which enables adapting to neural perturbations. The results on a realistic synthetic neural circuit show that the inverse controller based on the Schoenberg kernel can successfully drive the elicited responses close to the original target responses even when significant perturbations occur. Thirdly, the spike train variability causes fluctuations in the neural decoder. Local field potentials (LFPs) are an alternate manifestations of neural activity with a more common continuous amplitude representation and longer spatiotemporal scales that can be recorded simultaneously from the same electrode array and contain complementary information about stimuli or behavior. We propose a tensor product kernel based decoder for multiscale neural activity, which allows modeling the sample from different sources individually and mapping them onto the same RKHS defined by the tensor product of the individual kernels for each source. A single linear model is adapted as done before to identify the nonlinear mapping from the multiscale neural responses to the stimuli. It enables us decoding of more complete and accurate information from heterogeneous multiscale neural activity with only with an implicit assumption of independence on their relationship. The decoding results in the rat sensory stimulation experiment show that the decoder outperforms the decoders with either single-type neural activities. In addition the multiscale decoding methodology is also used in the adaptive inverse control mode. Due to the accuracy and robustness of the decoder, the control diagram with open-loop mode is applied to controlthe spatiotemporal pattern of neural response in the rat somatosensory cortex with micro-stimulation in order to emulate the tactile sensation and obtained promising results. Finally, we quantify and comparatively validate the temporal functional connectivity between neurons by measuring the statistical dependence between their firing patterns. Temporal functional connectivity provides a quantifiable representation of the transient joint information of multi-channel neural activity, which is important to completely characterize the neural state but is normally overlooked in the temporal decoding due to its complexity. The functional connectivity pattern is represented by a graph/matrix, which is again not a conventional input for machine learning algorithms. Therefore, we propose two approaches to decode the stimulation information from the neural assembly pattern. One is to use graph theory to extract topology feature vector as the model input, which makes conventional machine learning approach applicable but dismisses the information of structural details. Therefore, we also proposed a matrix kernel that is able to map the connectivity matrix into RKHS and enable kernel based machine learning approaches directly to operate on the connectivity matrix, which bypasses the information reduction induced by the feature extraction.

[1]  Cedric E. Ginestet,et al.  Statistical parametric network analysis of functional connectivity dynamics during a working memory task , 2011, NeuroImage.

[2]  Eric T. Shea-Brown,et al.  Optimal Inputs for Phase Models of Spiking Neurons , 2006 .

[3]  Ian Daly,et al.  Brain computer interface control via functional connectivity dynamics , 2012, Pattern Recognit..

[4]  Miguel A. L. Nicolelis,et al.  Principles of neural ensemble physiology underlying the operation of brain–machine interfaces , 2009, Nature Reviews Neuroscience.

[5]  Bernhard Schölkopf,et al.  Nonlinear Component Analysis as a Kernel Eigenvalue Problem , 1998, Neural Computation.

[6]  J. Chapin,et al.  Proprioceptive and cutaneous representations in the rat ventral posterolateral thalamus. , 2008, Journal of neurophysiology.

[7]  F. Varela,et al.  Measuring phase synchrony in brain signals , 1999, Human brain mapping.

[8]  David M. Santucci,et al.  Learning to Control a Brain–Machine Interface for Reaching and Grasping by Primates , 2003, PLoS biology.

[9]  Bernhard Schölkopf,et al.  A tutorial on support vector regression , 2004, Stat. Comput..

[10]  Alexander J. Smola,et al.  Learning with Kernels: support vector machines, regularization, optimization, and beyond , 2001, Adaptive computation and machine learning series.

[11]  Henry Markram,et al.  Real-Time Computing Without Stable States: A New Framework for Neural Computation Based on Perturbations , 2002, Neural Computation.

[12]  Peter E. Latham,et al.  Pairwise Maximum Entropy Models for Studying Large Biological Systems: When They Can Work and When They Can't , 2008, PLoS Comput. Biol..

[13]  José Carlos Príncipe,et al.  A Reproducing Kernel Hilbert Space Framework for Spike Train Signal Processing , 2009, Neural Computation.

[14]  José Carlos Príncipe,et al.  A comparison of binless spike train measures , 2010, Neural Computing and Applications.

[15]  Weifeng Liu,et al.  Correntropy: Properties and Applications in Non-Gaussian Signal Processing , 2007, IEEE Transactions on Signal Processing.

[16]  Peter Dayan,et al.  Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems , 2001 .

[17]  Weifeng Liu,et al.  An Information Theoretic Approach of Designing Sparse Kernel Adaptive Filters , 2009, IEEE Transactions on Neural Networks.

[18]  Martin Bogdan,et al.  Real-Time Adaptive Microstimulation Increases Reliability of Electrically Evoked Cortical Potentials , 2011, IEEE Transactions on Biomedical Engineering.

[19]  Stphane Mallat,et al.  A Wavelet Tour of Signal Processing, Third Edition: The Sparse Way , 2008 .

[20]  Andrew S. Whitford,et al.  Cortical control of a prosthetic arm for self-feeding , 2008, Nature.

[21]  G. Buzsáki,et al.  Neuronal Oscillations in Cortical Networks , 2004, Science.

[22]  Shalom Darmanjian,et al.  Design and analysis of generative models for brain machine interfaces , 2009 .

[23]  J. Kurths,et al.  Phase synchronization: from theory to data analysis , 2003 .

[24]  Yoram Singer,et al.  Spikernels: Predicting Arm Movements by Embedding Population Spike Rate Patterns in Inner-Product Spaces , 2005, Neural Computation.

[25]  Miguel A. L. Nicolelis,et al.  Real-time control of a robot arm using simultaneously recorded neurons in the motor cortex , 1999, Nature Neuroscience.

[26]  Antonio G. Zippo,et al.  Predicting Spike Occurrence and Neuronal Responsiveness from LFPs in Primary Somatosensory Cortex , 2012, PloS one.

[27]  Shie Mannor,et al.  The kernel recursive least-squares algorithm , 2004, IEEE Transactions on Signal Processing.

[28]  Sohan Seth,et al.  Neuronal functional connectivity dynamics in cortex: An MSC-based analysis , 2010, 2010 Annual International Conference of the IEEE Engineering in Medicine and Biology.

[29]  Il Park,et al.  Capturing spike train similarity structure: A point process divergence approach , 2010 .

[30]  K. Pearson,et al.  On the Theory of Contingency , 1930 .

[31]  Mehdi Aghagolzadeh,et al.  Synergistic Coding by Cortical Neural Ensembles , 2010, IEEE Transactions on Information Theory.

[32]  F. Mormann,et al.  Mean phase coherence as a measure for phase synchronization and its application to the EEG of epilepsy patients , 2000 .

[33]  M. Roulston Estimating the errors on measured entropy and mutual information , 1999 .

[34]  Miguel A. L. Nicolelis,et al.  Brain–machine interfaces to restore motor function and probe neural circuits , 2003, Nature Reviews Neuroscience.

[35]  Alexa Riehle,et al.  Spike synchronization and firing rate in a population of motor cortical neurons in relation to movement direction and reaction time , 2003, Biological Cybernetics.

[36]  John C. Platt A Resource-Allocating Network for Function Interpolation , 1991, Neural Computation.

[37]  Rodrigo Quian Quiroga,et al.  Nonlinear multivariate analysis of neurophysiological signals , 2005, Progress in Neurobiology.

[38]  Youxian Sun,et al.  Online SVM regression algorithm-based adaptive inverse control , 2007, Neurocomputing.

[39]  Vasilis Z. Marmarelis,et al.  Nonlinear Dynamic Modeling of Physiological Systems , 2004 .

[40]  Sonja Grün,et al.  Unitary Events in Multiple Single-Neuron Spiking Activity: I. Detection and Significance , 2002, Neural Computation.

[41]  B. Widrow,et al.  Adaptive inverse control , 1987, Proceedings of 8th IEEE International Symposium on Intelligent Control.

[42]  Pamela Reinagel,et al.  Decoding visual information from a population of retinal ganglion cells. , 1997, Journal of neurophysiology.

[43]  Paul H. E. Tiesinga,et al.  A New Correlation-Based Measure of Spike Timing Reliability , 2002, Neurocomputing.

[44]  Robert E Kass,et al.  Statistical issues in the analysis of neuronal data. , 2005, Journal of neurophysiology.

[45]  Frank D. Wood,et al.  Characterizing neural dependencies with copula models , 2008, NIPS.

[46]  Nicholas G. Hatsopoulos,et al.  Brain-machine interface: Instant neural control of a movement signal , 2002, Nature.

[47]  Jürgen Kurths,et al.  Detection of n:m Phase Locking from Noisy Data: Application to Magnetoencephalography , 1998 .

[48]  R. S. Johansson,et al.  Roles of glabrous skin receptors and sensorimotor memory in automatic control of precision grip when lifting rougher or more slippery objects , 2004, Experimental Brain Research.

[49]  A. Vespignani,et al.  The architecture of complex weighted networks. , 2003, Proceedings of the National Academy of Sciences of the United States of America.

[50]  Danielle Smith Bassett,et al.  Small-World Brain Networks , 2006, The Neuroscientist : a review journal bringing neurobiology, neurology and psychiatry.

[51]  Alexander S. Ecker,et al.  Feature Selectivity of the Gamma-Band of the Local Field Potential in Primate Primary Visual Cortex , 2008, Front. Neurosci..

[52]  Badong Chen,et al.  Quantized Kernel Least Mean Square Algorithm , 2012, IEEE Transactions on Neural Networks and Learning Systems.

[53]  S. David,et al.  Influence of context and behavior on stimulus reconstruction from neural activity in primary auditory cortex. , 2009, Journal of neurophysiology.

[54]  S Micera,et al.  Neuro-fuzzy decoding of sensory information from ensembles of simultaneously recorded dorsal root ganglion neurons for functional electrical stimulation applications , 2011, Journal of neural engineering.

[55]  H. Ramlau-Hansen Smoothing Counting Process Intensities by Means of Kernel Functions , 1983 .

[56]  C. Stam,et al.  Small‐world properties of nonlinear brain activity in schizophrenia , 2009, Human brain mapping.

[57]  William Bialek,et al.  Reading a Neural Code , 1991, NIPS.

[58]  R Quian Quiroga,et al.  Performance of different synchronization measures in real data: a case study on electroencephalographic signals. , 2002, Physical review. E, Statistical, nonlinear, and soft matter physics.

[59]  J. Martinerie,et al.  The brainweb: Phase synchronization and large-scale integration , 2001, Nature Reviews Neuroscience.

[60]  J. B. Ranck,et al.  Which elements are excited in electrical stimulation of mammalian central nervous system: A review , 1975, Brain Research.

[61]  Edward T. Bullmore,et al.  Reproducibility of graph metrics of human brain functional networks , 2009, NeuroImage.

[62]  Rafael Yuste,et al.  Designing optimal stimuli to control neuronal spike timing. , 2011, Journal of neurophysiology.

[63]  O. Sporns,et al.  Complex brain networks: graph theoretical analysis of structural and functional systems , 2009, Nature Reviews Neuroscience.

[64]  José Carlos Príncipe,et al.  2011 Ieee International Workshop on Machine Learning for Signal Processing an Adaptive Decoder from Spike Trains to Micro-stimulation Using Kernel Least-mean-squares (klms) , 2022 .

[65]  Yoram Singer,et al.  Spikernels: Embedding Spiking Neurons in Inner-Product Spaces , 2002, NIPS.

[66]  Nello Cristianini,et al.  An Introduction to Support Vector Machines and Other Kernel-based Learning Methods , 2000 .

[67]  F. James Rohlf,et al.  Biometry: The Principles and Practice of Statistics in Biological Research , 1969 .

[68]  N A Fitzsimmons,et al.  Primate Reaching Cued by Multichannel Spatiotemporal Cortical Microstimulation , 2007, The Journal of Neuroscience.

[69]  Rong Jin,et al.  Identifying Functional Connectivity in Large-Scale Neural Ensemble Recordings: A Multiscale Data Mining Approach , 2009, Neural Computation.

[70]  Austin J. Brockmeier,et al.  Optimizing microstimulation using a reinforcement learning framework , 2011, 2011 Annual International Conference of the IEEE Engineering in Medicine and Biology Society.

[71]  José Carlos Príncipe,et al.  Coadaptive Brain–Machine Interface via Reinforcement Learning , 2009, IEEE Transactions on Biomedical Engineering.

[72]  Dragan F. Dimitrov,et al.  Cortical Representation of Ipsilateral Arm Movements in Monkey and Man , 2009, The Journal of Neuroscience.

[73]  C. Galletti,et al.  Contribution of visual and proprioceptive information to the precision of reaching movements , 2010, Experimental Brain Research.

[74]  N. Aronszajn Theory of Reproducing Kernels. , 1950 .

[75]  M M Mesulam,et al.  Large‐scale neurocognitive networks and distributed processing for attention, language, and memory , 1990, Annals of neurology.

[76]  H. Khalil,et al.  Model-based analysis and control of a network of basal ganglia spiking neurons in the normal and Parkinsonian states , 2011, Journal of neural engineering.

[77]  Lehel Csató,et al.  Sparse On-Line Gaussian Processes , 2002, Neural Computation.

[78]  Eric T. Shea-Brown,et al.  Toward closed-loop optimization of deep brain stimulation for Parkinson's disease: concepts and lessons from a computational model , 2007, Journal of neural engineering.