EFFICIENT AND SCALABLE BIOLOGICALLY PLAUSIBLE SPIKING NEURAL NETWORKS WITH LEARNING APPLIED TO VISION

concept of her. Collins and Jin [109] show that a ‘grandmother cell’ type representation could be information theoretically efficient provided it is accompanied by distributed coding type cells. Maass [110] shows that WTA is quite powerful compared to threshold and sigmoidal gates often used in traditional neural networks. It is shown that any Boolean function can be computed using a single k-WTA unit [110]. This is very interesting, as at least two-layered perceptron circuits are needed to compute complicated functions. They also showed that any continuous function can be approximated by a single soft WTA unit (A soft winner take all operation assumes values depending on the rank of corresponding input in linear order). Another advantage is the fact that approximate WTA computation can be done very fast (linear-size) in analog VLSI chips [111]. Thus, complex feed-forward multi-layered perceptron circuits can be replaced by a single competitive WTA stage leading to low power analog VLSI chips [110]. There have been many implementations of winner take all (WTA) computations in recurrent networks in the literature [112, 113]. Also there have been many analog VLSI implementations of these circuit [113, 114]. The WTA model implemented here is influenced by the WTA implementation on recurrent networks by Oster and Liu [113]. In their implementation, the neuron that receives spikes with the shortest inter-spike interval is the winner. But it is not 49 clear in their implementation how (starting from random weights) a new neuron can learn a new category. A modified version of winner take all (WTA) with Hebbian learning is implemented here to demonstrate how different neurons can learn different categories. WTA is applied on both the learning layers while training and is switched off while testing. The WTA is implemented as follows: 1) At every time step, find the post neuron with the least spike time difference tpost1tpost2. Note that these last two post-synaptic neuron spike times are easily available. This neuron is declared as the winner. 2) The winner inhibits the other neurons from firing by sending an inhibitory pulse to others. If the winner neuron has not learned any feature, it learns the new feature by the above Hebbian learning method. The neuron remains the winner unless another neuron has a lower spike time interval or a new image is presented. The above learning approach is followed except in the last (uppermost) learning layer where the winner is declared according to the supplied category/label information about the input image rather than the spike time difference. Thus the overall approach is like a semi-supervised approach, with unsupervised learning in the lower learning layer and supervised learning in the last learning layer. We are assuming that all membrane potentials are discharged at the onset of a stimulus. This can be achieved, for example, by a decay in the membrane potential. There have been several implementations of sparse coding schemes too. Földiák [105] shows how a layer of neurons can learn to sparse code by using Hebbian excitatory connections between the input and output units and anti-Hebbian learning among the output connections. Olshausen and Field [115] showed that minimizing an objective function of high sparseness and low reconstruction error on a set of natural images yields a set of basis functions similar to the Gabor like receptive fields of simple cells in primary visual cortex. One interesting study is by 50 Einhauser et al. [116], who developed a neural network model which could develop receptive field properties similar to the simple and complex cells found in visual cortex. The network could learn from natural stimuli obtained by mounting a camera to a cat’s head to approximate input to a cat’s visual system. Though, they did not use spiking neural network. 4.3 Neuronal and Synaptic Genesis Neurogenesis and synaptogenesis (birth of neurons or synapses) has been shown to occur in the brain in certain settings [117, 118]. It is also known that it does affect some learning and memory tasks [119-121]. Many studies have indicated that the brain is much more plastic than previously thought [122]. For example, brain scans of people who loose their limbs in accidents show that parts of the brain maps corresponding to the lost limb are taken over by the surrounding brain maps. This plasticity and change can be made possible by birth and death of neurons and connections. For efficient simulations it will be important to be able to add neurons and synapses, and to also remove them when necessary. We found that our network will benefit from modeling these processes. For illustration, we trained a network with the network architecture shown in the right plot of Figure 3-9 using our learning approach. The goal was to recognize handwritten digits from 0 to 9. More details about the network structure and training problem can be found in Section 6.4.1 Our focus here is on the first learning layer in the architecture which was trained in an unsupervised way. Initially the synaptic weights were set to random values in this layer. Figure 4-2 shows 2D-arrays of synapse values after training for this learning layer plotted as 8x8 projections onto the previous layer. The synaptic weights are plotted as intensities with white representing the highest synaptic strength and black the lowest. These arrays represent features that were learned. Note that some of them remained unchanged (remain random starting values) during training and could have as well been 51 removed from the simulation for efficiency. Or, better yet, we could have started with fewer synapses and added them as needed. This would ensure that unnecessary connections are not there and will save computer memory. Figure 4-2: Images of arrays of final synapse weights plotted as 8x8 projections to previous layer in a layer of network after learning. White represents the highest synaptic strength and black represents the lowest.

[1]  W. Levy,et al.  Temporal contiguity requirements for long-term associative potentiation/depression in the hippocampus , 1983, Neuroscience.

[2]  P. Földiák,et al.  Forming sparse representations by local anti-Hebbian learning , 1990, Biological Cybernetics.

[3]  Koen E. A. van de Sande,et al.  Evaluating Color Descriptors for Object and Scene Recognition , 2010, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[4]  Wolfgang Maass,et al.  STDP enables spiking neurons to detect hidden causes of their inputs , 2009, NIPS.

[5]  David G. Lowe,et al.  Multiclass Object Recognition with Sparse, Localized Features , 2006, 2006 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'06).

[6]  Kunihiko Fukushima,et al.  Neocognitron for handwritten digit recognition , 2003, Neurocomputing.

[7]  Thomas Serre,et al.  Object recognition with features inspired by visual cortex , 2005, 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition (CVPR'05).

[8]  D. Hubel,et al.  Receptive fields and functional architecture of monkey striate cortex , 1968, The Journal of physiology.

[9]  C. Blakemore,et al.  Development of orientation columns in cat striate cortex revealed by 2-deoxyglucose autoradiography , 1983, Nature.

[10]  Ankur Gupta,et al.  Biologically-inspired spiking neural networks with Hebbian learning for vision processing , 2008 .

[11]  Thomas Serre,et al.  Robust Object Recognition with Cortex-Like Mechanisms , 2007, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[12]  D. Johnston,et al.  Regulation of Synaptic Efficacy by Coincidence of Postsynaptic APs and EPSPs , 1997 .

[13]  Vivien A. Casagrande,et al.  Biophysics of Computation: Information Processing in Single Neurons , 1999 .

[14]  J. Pincus The Brain That Changes Itself: Stories of Personal Triumph From the Frontiers of Brain Science , 2008 .

[15]  Hongjun Song,et al.  Adult neurogenesis in the mammalian central nervous system. , 2005, Annual review of neuroscience.

[16]  Guo-Qiang Bi,et al.  Synaptic modification in neural circuits: a timely action. , 2002, BioEssays : news and reviews in molecular, cellular and developmental biology.

[17]  David J. Field,et al.  What Is the Goal of Sensory Coding? , 1994, Neural Computation.

[18]  L. Trussell,et al.  Cell-specific, spike timing–dependent plasticities in the dorsal cochlear nucleus , 2004, Nature Neuroscience.

[19]  N. Logothetis,et al.  Cortical mechanisms of sensory learning and object recognition , 2008, Philosophical Transactions of the Royal Society B: Biological Sciences.

[20]  David J. Field,et al.  Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.

[21]  J. Fagot,et al.  Cross-modal integration and conceptual categorization in baboons , 2001, Behavioural Brain Research.

[22]  Y. Dan,et al.  Spike Timing-Dependent Plasticity of Neural Circuits , 2004, Neuron.

[23]  Yoshiyasu Takefuji,et al.  Neural network parallel computing , 1992, The Kluwer international series in engineering and computer science.

[24]  Edmund T. Rolls,et al.  Invariant visual object recognition: A model, with lighting invariance , 2006, Journal of Physiology-Paris.

[25]  D. Hubel,et al.  Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.

[26]  Donald A. Sofge,et al.  Handbook of Intelligent Control: Neural, Fuzzy, and Adaptive Approaches , 1992 .

[27]  Leslie G. Ungerleider,et al.  Contribution of striate inputs to the visuospatial functions of parieto-preoccipital cortex in monkeys , 1982, Behavioural Brain Research.

[28]  Thomas Serre,et al.  A quantitative theory of immediate visual recognition. , 2007, Progress in brain research.

[29]  David I. Perrett,et al.  Neurophysiology of shape processing , 1993, Image Vis. Comput..

[30]  H. Spekreijse,et al.  FigureGround Segregation in a Recurrent Network Architecture , 2002, Journal of Cognitive Neuroscience.

[31]  Stephen Grossberg,et al.  The ART of adaptive pattern recognition by a self-organizing neural network , 1988, Computer.

[32]  E. Rolls,et al.  View-invariant representations of familiar objects by neurons in the inferior temporal visual cortex. , 1998, Cerebral cortex.

[33]  Arnaud Delorme,et al.  Face identification using one spike per neuron: resistance to image degradations , 2001, Neural Networks.

[34]  Wolfgang Maass,et al.  Neural Computation with Winner-Take-All as the Only Nonlinear Operation , 1999, NIPS.

[35]  C. Koch,et al.  Invariant visual representation by single neurons in the human brain , 2005, Nature.

[36]  Anil K. Jain,et al.  Object detection using gabor filters , 1997, Pattern Recognit..

[37]  K. D. Punta,et al.  An ultra-sparse code underlies the generation of neural sequences in a songbird , 2002 .

[38]  Eugene M. Izhikevich,et al.  Which model to use for cortical spiking neurons? , 2004, IEEE Transactions on Neural Networks.

[39]  N. Logothetis,et al.  Shape representation in the inferior temporal cortex of monkeys , 1995, Current Biology.

[40]  Li I. Zhang,et al.  A critical window for cooperation and competition among developing retinotectal synapses , 1998, Nature.

[41]  N. Logothetis,et al.  View-dependent object recognition by monkeys , 1994, Current Biology.

[42]  Christoph Kayser,et al.  Learning the invariance properties of complex cells from their responses to natural stimuli , 2002, The European journal of neuroscience.

[43]  T. Poggio,et al.  A model of V4 shape selectivity and invariance. , 2007, Journal of neurophysiology.

[44]  David G. Lowe,et al.  Object recognition from local scale-invariant features , 1999, Proceedings of the Seventh IEEE International Conference on Computer Vision.

[45]  E. Rolls,et al.  A Neurodynamical cortical model of visual attention and invariant object recognition , 2004, Vision Research.

[46]  P. O. Bishop,et al.  Orientation specificity and response variability of cells in the striate cortex. , 1973, Vision research.

[47]  Edmund T. Rolls,et al.  Functions of the Primate Temporal Lobe Cortical Visual Areas in Invariant Visual Object and Face Recognition , 2000, Neuron.

[48]  L. Abbott,et al.  Competitive Hebbian learning through spike-timing-dependent synaptic plasticity , 2000, Nature Neuroscience.

[49]  N. Daw,et al.  Kittens reared in a unidirectional environment: evidence for a critical period. , 1976, The Journal of physiology.

[50]  Kunihiko Fukushima,et al.  Neocognitron: A self-organizing neural network model for a mechanism of pattern recognition unaffected by shift in position , 1980, Biological Cybernetics.

[51]  Stefan Wermter,et al.  Spike-timing-dependent synaptic plasticity: from single spikes to spike trains , 2004, Neurocomputing.

[52]  M.H. Hassoun,et al.  Fundamentals of Artificial Neural Networks , 1996, Proceedings of the IEEE.

[53]  Wulfram Gerstner,et al.  SPIKING NEURON MODELS Single Neurons , Populations , Plasticity , 2002 .

[54]  L. Saksida,et al.  A Functional Role for Adult Hippocampal Neurogenesis in Spatial Pattern Separation , 2009, Science.

[55]  A P Georgopoulos,et al.  On the relations between the direction of two-dimensional arm movements and cell discharge in primate motor cortex , 1982, The Journal of neuroscience : the official journal of the Society for Neuroscience.

[56]  John Lazzaro,et al.  Winner-Take-All Networks of O(N) Complexity , 1988, NIPS.

[57]  S. Thorpe,et al.  Seeking Categories in the Brain , 2001, Science.

[58]  J. Carey Brain Facts: A Primer on the Brain and Nervous System. , 1990 .

[59]  Lyle N. Long,et al.  Deep Blue Cannot Play Checkers: The Need for Generalized Intelligence for Mobile Robots , 2010, J. Robotics.

[60]  J. P. Jones,et al.  An evaluation of the two-dimensional Gabor filter model of simple receptive fields in cat striate cortex. , 1987, Journal of neurophysiology.

[61]  T Natschläger,et al.  Spatial and temporal pattern analysis via spiking neurons. , 1998, Network.

[62]  C. Blakemore,et al.  Innate and environmental factors in the development of the kitten's visual cortex. , 1975, The Journal of physiology.

[63]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.

[64]  D. Jin,et al.  Grandmother cells and the storage capacity of the human brain , 2006, q-bio/0603014.

[65]  Kunihiko Fukushima,et al.  Neocognitron: A hierarchical neural network capable of visual pattern recognition , 1988, Neural Networks.

[66]  Stefan Wermter,et al.  Temporal sequence detection with spiking neurons: towards recognizing robot language instructions , 2006, Connect. Sci..

[67]  Stefan Wermter,et al.  Spike-Timing Dependent Competitive Learning of Integrate-and-Fire Neurons with Active Dendrites , 2002, ICANN.

[68]  Lyle N. Long,et al.  Scalable Massively Parallel Artificial Neural Networks , 2005, J. Aerosp. Comput. Inf. Commun..

[69]  李幼升,et al.  Ph , 1989 .

[70]  Tomaso Poggio,et al.  Fast Readout of Object Identity from Macaque Inferior Temporal Cortex , 2005, Science.

[71]  P. Taupin Adult neurogenesis and neuroplasticity. , 2006, Restorative neurology and neuroscience.

[72]  G. Bi,et al.  Synaptic Modifications in Cultured Hippocampal Neurons: Dependence on Spike Timing, Synaptic Strength, and Postsynaptic Cell Type , 1998, The Journal of Neuroscience.

[73]  Nicholas T. Carnevale,et al.  Simulation of networks of spiking neurons: A review of tools and strategies , 2006, Journal of Computational Neuroscience.

[74]  Shih-Chii Liu,et al.  A winner-take-all spiking network with spiking inputs , 2004, Proceedings of the 2004 11th IEEE International Conference on Electronics, Circuits and Systems, 2004. ICECS 2004..

[75]  Stephen Grossberg,et al.  Adaptive Resonance Theory , 2010, Encyclopedia of Machine Learning.

[76]  Xiaohui Xie,et al.  Learning in neural networks by reinforcement of irregular spiking. , 2004, Physical review. E, Statistical, nonlinear, and soft matter physics.

[77]  James L. McClelland,et al.  Parallel distributed processing: explorations in the microstructure of cognition, vol. 1: foundations , 1986 .

[78]  Ankur Gupta Detecting load conditions in human walking using expectation maximization and neural networks , 2009, 2009 International Joint Conference on Neural Networks.

[79]  E. Capaldi,et al.  The organization of behavior. , 1992, Journal of applied behavior analysis.

[80]  T. Poggio,et al.  Hierarchical models of object recognition in cortex , 1999, Nature Neuroscience.

[81]  R Van Rullen,et al.  Face processing using one spike per neurone. , 1998, Bio Systems.

[82]  S. Hochstein,et al.  View from the Top Hierarchies and Reverse Hierarchies in the Visual System , 2002, Neuron.

[83]  Paul J. Werbos,et al.  The Roots of Backpropagation: From Ordered Derivatives to Neural Networks and Political Forecasting , 1994 .

[84]  D. Hubel,et al.  Receptive fields, binocular interaction and functional architecture in the cat's visual cortex , 1962, The Journal of physiology.

[85]  J. Maunsell,et al.  Sensory modality specificity of neural activity related to memory in visual cortex. , 1997, Journal of neurophysiology.

[86]  Stephen Grossberg,et al.  Art 2: Self-Organization Of Stable Category Recognition Codes For Analog Input Patterns , 1988, Other Conferences.

[87]  L. Trussell,et al.  Coactivation of Pre- and Postsynaptic Signaling Mechanisms Determines Cell-Specific Spike-Timing-Dependent Plasticity , 2007, Neuron.

[88]  Takeo Kanade,et al.  Neural Network-Based Face Detection , 1998, IEEE Trans. Pattern Anal. Mach. Intell..

[89]  F. Gage,et al.  Adult neurogenesis and neural stem cells of the central nervous system in mammals , 2002, Journal of neuroscience research.

[90]  Christoph D. Dahl,et al.  Individuation and holistic processing of faces in rhesus monkeys , 2007, Proceedings of the Royal Society B: Biological Sciences.

[91]  T. Shors,et al.  From stem cells to grandmother cells: how neurogenesis relates to learning and memory. , 2008, Cell stem cell.

[92]  Jilles Vreeken,et al.  Spiking neural networks, an introduction , 2003 .

[93]  Dennis Gabor,et al.  Theory of communication , 1946 .

[94]  Timothée Masquelier,et al.  Unsupervised Learning of Visual Features through Spike Timing Dependent Plasticity , 2007, PLoS Comput. Biol..

[95]  D C Van Essen,et al.  Functional properties of neurons in middle temporal visual area of the macaque monkey. I. Selectivity for stimulus direction, speed, and orientation. , 1983, Journal of neurophysiology.

[96]  Marc'Aurelio Ranzato,et al.  Efficient Learning of Sparse Representations with an Energy-Based Model , 2006, NIPS.

[97]  Thomas Serre,et al.  A feedforward architecture accounts for rapid categorization , 2007, Proceedings of the National Academy of Sciences.

[98]  J. Rauschecker,et al.  Maps and streams in the auditory cortex: nonhuman primates illuminate human speech processing , 2009, Nature Neuroscience.

[99]  Lyle N. Long,et al.  Review of Consciousness and the Possibility of Conscious Robots , 2010, J. Aerosp. Comput. Inf. Commun..

[100]  Lukasz A. Kurgan,et al.  A new synaptic plasticity rule for networks of spiking neurons , 2006, IEEE Transactions on Neural Networks.

[101]  V. Han,et al.  Synaptic plasticity in a cerebellum-like structure depends on temporal order , 1997, Nature.

[102]  Tomaso Poggio,et al.  Learning a dictionary of shape-components in visual cortex: comparison with neurons, humans and machines , 2006 .

[103]  Y. Dan,et al.  Spike-timing-dependent synaptic plasticity depends on dendritic location , 2005, Nature.

[104]  Paul Over,et al.  Evaluation campaigns and TRECVid , 2006, MIR '06.

[105]  Thomas Serre,et al.  A Component-based Framework for Face Detection and Identification , 2007, International Journal of Computer Vision.

[106]  Dezhe Z Jin,et al.  Fast computation with spikes in a recurrent neural network. , 2002, Physical review. E, Statistical, nonlinear, and soft matter physics.

[107]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[108]  David G. Lowe,et al.  Towards a Computational Model for Object Recognition in IT Cortex , 2000, Biologically Motivated Computer Vision.

[109]  Lyle N. Long,et al.  The critical need for software engineering education , 2008 .

[110]  Denis Fize,et al.  Speed of processing in the human visual system , 1996, Nature.

[111]  G. Kreiman,et al.  Timing, Timing, Timing: Fast Decoding of Object Information from Intracranial Field Potentials in Human Visual Cortex , 2009, Neuron.

[112]  Matthijs C. Dorst Distinctive Image Features from Scale-Invariant Keypoints , 2011 .

[113]  Patrice Y. Simard,et al.  Best practices for convolutional neural networks applied to visual document analysis , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..

[114]  J. Enns,et al.  What’s new in visual masking? , 2000, Trends in Cognitive Sciences.

[115]  Lyle N. Long,et al.  A Review of Biologically Plausible Neuron Models for Spiking Neural Networks , 2010 .

[116]  Hans P. Moravec Robot: Mere Machine to Transcendent Mind , 1998 .

[117]  E Harth,et al.  The inversion of sensory processing by feedback pathways: a model of visual cognitive functions. , 1987, Science.

[118]  Thomas Martinetz,et al.  Simple Method for High-Performance Digit Recognition Based on Sparse Coding , 2008, IEEE Transactions on Neural Networks.

[119]  L.N. Long,et al.  Scalable Biologically Inspired Neural Networks with Spike Time Based Learning , 2008, 2008 ECSIS Symposium on Learning and Adaptive Behaviors for Robotic Systems (LAB-RS).

[120]  A. Rauf Baig SPATIAL-TEMPORAL ARTIFICIAL NEURONS APPLIED TO ONLINE CURSIVE HANDWRITTEN CHARACTER RECOGNITION , 2004 .

[121]  H. Markram,et al.  Regulation of Synaptic Efficacy by Coincidence of Postsynaptic APs and EPSPs , 1997, Science.

[122]  Lyle N. Long,et al.  Character Recognition using Spiking Neural Networks , 2007, 2007 International Joint Conference on Neural Networks.

[123]  D. Perrett,et al.  Visual neurones responsive to faces in the monkey temporal cortex , 2004, Experimental Brain Research.

[124]  Ravi S. Menon,et al.  Haptic study of three-dimensional objects activates extrastriate visual areas , 2002, Neuropsychologia.

[125]  Richard Szeliski,et al.  Using character recognition and segmentation to tell computer from humans , 2003, Seventh International Conference on Document Analysis and Recognition, 2003. Proceedings..