Neural system identification for large populations separating "what" and "where"

Neuroscientists classify neurons into different types that perform similar computations at different locations in the visual field. Traditional methods for neural system identification do not capitalize on this separation of 'what' and 'where'. Learning deep convolutional feature spaces that are shared among many neurons provides an exciting path forward, but the architectural design needs to account for data limitations: While new experimental techniques enable recordings from thousands of neurons, experimental time is limited so that one can sample only a small fraction of each neuron's response space. Here, we show that a major bottleneck for fitting convolutional neural networks (CNNs) to neural data is the estimation of the individual receptive field locations, a problem that has been scratched only at the surface thus far. We propose a CNN architecture with a sparse readout layer factorizing the spatial (where) and feature (what) dimensions. Our network scales well to thousands of neurons and short recordings and can be trained end-to-end. We evaluate this architecture on ground-truth data to explore the challenges and limitations of CNN-based system identification. Moreover, we show that our network model outperforms current state-of-the art system identification models of mouse primary visual cortex.

[1]  Ryan J. Prenger,et al.  Nonlinear V1 responses to natural scenes revealed by neural network analysis , 2004, Neural Networks.

[2]  Matthias Bethge,et al.  The functional diversity of retinal ganglion cells in the mouse , 2015, Nature.

[3]  Ryan J. Prenger,et al.  The Berkeley Wavelet Transform: A Biologically Inspired Orthogonal Wavelet Transform , 2008, Neural Computation.

[4]  Eero P. Simoncelli,et al.  Testing pseudo-linear models of responses to natural scenes in primate retina , 2016, bioRxiv.

[5]  William F. Kindel,et al.  Using deep learning to reveal the neural code for images in primary visual cortex , 2017, ArXiv.

[6]  Eero P. Simoncelli,et al.  A Convolutional Subunit Model for Neuronal Responses in Macaque V1 , 2015, The Journal of Neuroscience.

[7]  Eero P. Simoncelli,et al.  Spatiotemporal Elements of Macaque V1 Receptive Fields , 2005, Neuron.

[8]  M. Bethge,et al.  Inhibition decorrelates visual feature representations in the inner retina , 2017, Nature.

[9]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[10]  Ghislain St-Yves,et al.  The feature-weighted receptive field: an interpretable encoding model for complex feature spaces , 2017, NeuroImage.

[11]  Nikolaus Kriegeskorte,et al.  Explaining the hierarchy of visual representational geometries by remixing of features from many computational vision models , 2014 .

[12]  Brian Lau,et al.  Computational subunits of visual cortical neurons revealed by artificial neural networks , 2002, Proceedings of the National Academy of Sciences of the United States of America.

[13]  Ghislain St-Yves,et al.  The feature-weighted receptive field: an interpretable encoding model for complex feature spaces , 2017 .

[14]  Jonathan W. Pillow,et al.  Capturing the Dynamical Repertoire of Single Neurons with Generalized Linear Models , 2016, Neural Computation.

[15]  Liam Paninski,et al.  Multilayer Recurrent Network Models of Primate Retinal Ganglion Cell Responses , 2016, ICLR.

[16]  David S. Shrom Functional diversity of retinal ganglion cells , 2007 .

[17]  J. P. Jones,et al.  The two-dimensional spatial structure of simple receptive fields in cat striate cortex. , 1987, Journal of neurophysiology.

[18]  J. Gallant,et al.  Complete functional characterization of sensory neurons by system identification. , 2006, Annual review of neuroscience.

[19]  Leon A. Gatys,et al.  Deep convolutional models improve predictions of macaque V1 responses to natural images , 2019, PLoS Comput. Biol..

[20]  James A. Bednar,et al.  Model Constrained by Visual Hierarchy Improves Prediction of Neural Responses to Natural Scenes , 2016, PLoS Comput. Biol..

[21]  M. Meister,et al.  Neural Circuit Inference from Function to Structure , 2017, Current Biology.

[22]  E H Adelson,et al.  Spatiotemporal energy models for the perception of motion. , 1985, Journal of the Optical Society of America. A, Optics and image science.

[23]  J. Touryan,et al.  Spatial Structure of Complex Cell Receptive Fields Measured with Natural Images , 2005, Neuron.

[24]  Tatyana O. Sharpee,et al.  Cross-orientation suppression in visual area V2 , 2017, Nature Communications.

[25]  R. Desimone,et al.  Predicting responses of nonlinear neurons in monkey striate cortex to complex patterns , 1992, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[26]  Yuwei Cui,et al.  Inferring Nonlinear Neuronal Computation Based on Physiologically Plausible Inputs , 2013, PLoS Comput. Biol..

[27]  Surya Ganguli,et al.  Deep Learning Models of the Retinal Response to Natural Scenes , 2017, NIPS.

[28]  Pavan Ramkumar,et al.  Modern machine learning outperforms GLMs at predicting spikes , 2017, bioRxiv.

[29]  Andrew Zisserman,et al.  Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.

[30]  D. Hubel,et al.  Receptive fields, binocular interaction and functional architecture in the cat's visual cortex , 1962, The Journal of physiology.

[31]  Ha Hong,et al.  Performance-optimized hierarchical models predict neural responses in higher visual cortex , 2014, Proceedings of the National Academy of Sciences.

[32]  Tim Gollisch,et al.  Eye Smarter than Scientists Believed: Neural Computations in Circuits of the Retina , 2010, Neuron.

[33]  Nicole C. Rust,et al.  Do We Know What the Early Visual System Does? , 2005, The Journal of Neuroscience.

[34]  Sergey Ioffe,et al.  Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.