Nonlinear Modeling of the Early and Mid-Level Visual System

The work presented in this thesis is toward the goal of extracting structure and meaning from neuroscientific data. Data in visual neuroscience is becoming increasingly high dimensional and the stimulus-response relationships can be highly nonlinear. Data in visual neuroscience is also somewhat noisy due to the imprecise separation of signals from multiple neurons on an electrode, nonstationary effects in the brain, and inherent noise in the brain; neurons rarely respond identically to identical stimuli. Finding nonlinear relationships between a high dimensional stimulus and neural responses in the presence of substantial noise is a challenging nonlinear regression problem. This thesis presents effective techniques for solving this problem and creating highly predictive models of neural function. I first introduce linearized regression, a technique for modeling nonlinear responses using linear regression on a nonlinear transformation of the stimulus. Next I demonstrate a method for efficiently finding Volterra series representations of nonlinear neural responses. Finally, I demonstrate that deep neural networks can provide accurate and interpretable models of the neural computations in visual cortex.

[1]  Geoffrey E. Hinton,et al.  Learning internal representations by error propagation , 1986 .

[2]  H. Zou,et al.  Regularization and variable selection via the elastic net , 2005 .

[3]  David J. Field,et al.  Sparse coding with an overcomplete basis set: A strategy employed by V1? , 1997, Vision Research.

[4]  J. P. Jones,et al.  An evaluation of the two-dimensional Gabor filter model of simple receptive fields in cat striate cortex. , 1987, Journal of neurophysiology.

[5]  Martin Fodslette Møller,et al.  A scaled conjugate gradient algorithm for fast supervised learning , 1993, Neural Networks.

[6]  P. J. Huber Robust Estimation of a Location Parameter , 1964 .

[7]  A. Soni Control-Relevant System Identification using Nonlinear Volterra and Volterra-Laguerre Models , 2006 .

[8]  J D Victor,et al.  Striate cortex extracts higher-order spatial correlations from visual textures. , 1994, Proceedings of the National Academy of Sciences of the United States of America.

[9]  Marc Teboulle,et al.  A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..

[10]  J. Friedman Stochastic gradient boosting , 2002 .

[11]  P. McCullagh,et al.  Generalized Linear Models , 1972, Predictive Analytics.

[12]  R. Stein Some models of neuronal variability. , 1967, Biophysical journal.

[13]  Trevor Hastie,et al.  Regularization Paths for Generalized Linear Models via Coordinate Descent. , 2010, Journal of statistical software.

[14]  Jacek M. Zurada,et al.  Generalized Core Vector Machines , 2006, IEEE Transactions on Neural Networks.

[15]  R. Freeman,et al.  Oblique effect: a neural basis in the visual cortex. , 2003, Journal of neurophysiology.

[16]  Vasilis Z. Marmarelis,et al.  Volterra models and three-layer perceptrons , 1997, IEEE Trans. Neural Networks.

[17]  Razvan Pascanu,et al.  Learned-Norm Pooling for Deep Feedforward and Recurrent Neural Networks , 2013, ECML/PKDD.

[18]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[19]  M. Yuan,et al.  Model selection and estimation in regression with grouped variables , 2006 .

[20]  D. Snodderly,et al.  Spatial organization of receptive fields of V1 neurons of alert monkeys: comparison with responses to gratings. , 2002, Journal of neurophysiology.

[21]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[22]  E. Süli,et al.  Numerical Solution of Ordinary Differential Equations , 2021, Foundations of Space Dynamics.

[23]  D. G. Albrecht,et al.  Striate cortex of monkey and cat: contrast response function. , 1982, Journal of neurophysiology.

[24]  Tara N. Sainath,et al.  Improving deep neural networks for LVCSR using rectified linear units and dropout , 2013, 2013 IEEE International Conference on Acoustics, Speech and Signal Processing.

[25]  S. Appelle Perception and discrimination as a function of stimulus orientation: the "oblique effect" in man and animals. , 1972, Psychological bulletin.

[26]  B M Dow,et al.  The mapping of visual space onto foveal striate cortex in the macaque monkey , 1985, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[27]  Yann LeCun,et al.  Convolutional neural networks applied to house numbers digit classification , 2012, Proceedings of the 21st International Conference on Pattern Recognition (ICPR2012).

[28]  Léon Bottou,et al.  Stochastic Gradient Descent Tricks , 2012, Neural Networks: Tricks of the Trade.

[29]  I. Hunter,et al.  The identification of nonlinear biological systems: Volterra kernel approaches , 1996, Annals of Biomedical Engineering.

[30]  Geoffrey E. Hinton,et al.  Phoneme recognition using time-delay neural networks , 1989, IEEE Trans. Acoust. Speech Signal Process..

[31]  Keiji Tanaka,et al.  Neuronal selectivities to complex object features in the ventral visual pathway of the macaque cerebral cortex. , 1994, Journal of neurophysiology.

[32]  Michael J. Korenberg,et al.  Parallel cascade identification and kernel estimation for nonlinear systems , 2006, Annals of Biomedical Engineering.

[33]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[34]  L. Cooper,et al.  When Networks Disagree: Ensemble Methods for Hybrid Neural Networks , 1992 .

[35]  B. Efron The jackknife, the bootstrap, and other resampling plans , 1987 .

[36]  David Mackay,et al.  Probable networks and plausible predictions - a review of practical Bayesian methods for supervised neural networks , 1995 .

[37]  Guillermo Sapiro,et al.  Online Learning for Matrix Factorization and Sparse Coding , 2009, J. Mach. Learn. Res..

[38]  Tony J. Dodd,et al.  A NEW SOLUTION TO VOLTERRA SERIES ESTIMATION , 2002 .

[39]  Yoshua Bengio,et al.  On the Expressive Power of Deep Architectures , 2011, ALT.

[40]  Nitish Srivastava,et al.  Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..

[41]  Marc'Aurelio Ranzato,et al.  Unsupervised Learning of Invariant Feature Hierarchies with Applications to Object Recognition , 2007, 2007 IEEE Conference on Computer Vision and Pattern Recognition.

[42]  Christof Koch,et al.  Shunting Inhibition Does Not Have a Divisive Effect on Firing Rates , 1997, Neural Computation.

[43]  Nitish Srivastava,et al.  Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.

[44]  Bogdan E. Popescu,et al.  Gradient Directed Regularization for Linear Regression and Classi…cation , 2004 .

[45]  Yoram Singer,et al.  Adaptive Subgradient Methods for Online Learning and Stochastic Optimization , 2011, J. Mach. Learn. Res..

[46]  J. Gallant,et al.  A Three-Dimensional Spatiotemporal Receptive Field Model Explains Responses of Area MT Neurons to Naturalistic Movies , 2011, The Journal of Neuroscience.

[47]  J. Friedman Greedy function approximation: A gradient boosting machine. , 2001 .

[48]  Semir Zeki,et al.  Effect of background colors on the tuning of color-selective cells in monkey area V4. , 2006, Journal of neurophysiology.

[49]  J. Gallant,et al.  Natural Stimulus Statistics Alter the Receptive Field Structure of V1 Neurons , 2004, The Journal of Neuroscience.

[50]  J. Gallant,et al.  Spectral receptive field properties explain shape selectivity in area V4. , 2006, Journal of neurophysiology.

[51]  Mijung Park,et al.  Bayesian inference for low rank spatiotemporal neural receptive fields , 2013, NIPS.