Unsupervised learning of event-based image recordings using spike-timing-dependent plasticity

The remarkable versatility and efficiency of the brain makes it important to understand its principles in order to push the boundaries of modern computing. Many models in neuroscience effectively model the detailed biological properties of the brain — however, they do not exhibit good performance on any benchmark task. Other systems have good performance, but do not work in the manner that the brain does, thus they do not give any insight into the mechanisms of the brain. A recent model by Diehl and Cook [1] is a neuromorphic system modelled by spiking neural networks (SNN) and spike-timing-dependant plasticity (STDP) that is able to achieve very good results on the MNIST dataset, a benchmark dataset of handwritten numbers for classification algorithms. However, in this system the manner in which MNIST dataset is converted into spikes does not mimic the manner in which neuromorphic sensors record information. Therefore, in this paper, we test the method proposed by Diehl and Cook on the N-MNIST (neuromorphic MNIST) where the MNIST images were recorded using a silicon retina, a DVS sensor that was moved similar to the saccade-like movements of the retina. The output of the camera consists of events at pixels similar to the retinal spikes. In this paper we examined how the original system had to be changed to accommodate the new dataset and noted that with modified parameters that respond to characteristics of the new dataset the system had better accuracy. The 400 neuron system has an accuracy of 93.68% for a smaller 3-class training dataset, and entire 10-class N-MNIST dataset on an 800 neuron network has 80.63% accuracy. We noted that output neurons form clusters that respond to a particular class, making it suitable to stack the system in a hierarchical manner.

[1]  Wulfram Gerstner,et al.  Noise in Integrate-and-Fire Neurons: From Stochastic Input to Escape Rates , 2000, Neural Computation.

[2]  Tobi Delbruck,et al.  A 240 × 180 130 dB 3 µs Latency Global Shutter Spatiotemporal Vision Sensor , 2014, IEEE Journal of Solid-State Circuits.

[3]  Maurizio Mattia,et al.  Collective Behavior of Networks with Linear (VLSI) Integrate-and-Fire Neurons , 1999, Neural Computation.

[4]  Matthew Cook,et al.  Unsupervised learning of digit recognition using spike-timing-dependent plasticity , 2015, Front. Comput. Neurosci..

[5]  Tobi Delbruck,et al.  A Dynamic Vision Sensor With 1% Temporal Contrast Sensitivity and In-Pixel Asynchronous Delta Modulator for Event Encoding , 2015, IEEE Journal of Solid-State Circuits.

[6]  Jacques Gautrais,et al.  SpikeNET: A simulator for modeling large networks of integrate and fire neurons , 1999, Neurocomputing.

[7]  Gregory Cohen,et al.  Converting Static Image Datasets to Spiking Neuromorphic Datasets Using Saccades , 2015, Front. Neurosci..

[8]  Daniel Matolin,et al.  A QVGA 143 dB Dynamic Range Frame-Free PWM Image Sensor With Lossless Pixel-Level Video Compression and Time-Domain CDS , 2011, IEEE Journal of Solid-State Circuits.

[9]  Germán Mato,et al.  On Numerical Simulations of Integrate-and-Fire Neural Networks , 1998, Neural Computation.

[10]  Anthony N. Burkitt,et al.  A Review of the Integrate-and-fire Neuron Model: I. Homogeneous Synaptic Input , 2006, Biological Cybernetics.

[11]  S. Heshka,et al.  Brain and high metabolic rate organ mass: contributions to resting energy expenditure beyond fat-free mass. , 2010, The American journal of clinical nutrition.

[12]  W. Senn,et al.  Neocortical pyramidal cells respond as integrate-and-fire neurons to in vivo-like input currents. , 2003, Journal of neurophysiology.

[13]  Tobi Delbrück,et al.  A 128$\times$ 128 120 dB 15 $\mu$s Latency Asynchronous Temporal Contrast Vision Sensor , 2008, IEEE Journal of Solid-State Circuits.

[14]  Bernabé Linares-Barranco,et al.  A 128$\,\times$ 128 1.5% Contrast Sensitivity 0.9% FPN 3 µs Latency 4 mW Asynchronous Frame-Free Dynamic Vision Sensor Using Transimpedance Preamplifiers , 2013, IEEE Journal of Solid-State Circuits.

[15]  Bernabé Linares-Barranco,et al.  A 3.6 $\mu$ s Latency Asynchronous Frame-Free Event-Driven Dynamic-Vision-Sensor , 2011, IEEE Journal of Solid-State Circuits.

[16]  Tobi Delbruck,et al.  Real-time classification and sensor fusion with a spiking deep belief network , 2013, Front. Neurosci..

[17]  Garrick Orchard,et al.  Skimming Digits: Neuromorphic Classification of Spike-Encoded Images , 2016, Front. Neurosci..

[18]  Tobi Delbrück,et al.  A 128 X 128 120db 30mw asynchronous vision sensor that responds to relative intensity change , 2006, 2006 IEEE International Solid State Circuits Conference - Digest of Technical Papers.

[19]  Timothée Masquelier,et al.  Unsupervised Learning of Visual Features through Spike Timing Dependent Plasticity , 2007, PLoS Comput. Biol..

[20]  Nicolas Brunel,et al.  Fast Global Oscillations in Networks of Integrate-and-Fire Neurons with Low Firing Rates , 1999, Neural Computation.

[21]  Yoshua Bengio,et al.  Gradient-based learning applied to document recognition , 1998, Proc. IEEE.