Advancing Deep Residual Learning by Solving the Crux of Degradation in Spiking Neural Networks

Despite the rapid progress of neuromorphic computing, the inadequate depth and the resulting insufficient representation power of spiking neural networks (SNNs) severely restrict their application scope in practice. Residual learning and shortcuts have been evidenced as an important approach for training deep neural networks, but rarely did previous work assess their applicability to the characteristics of spike-based communication and spatiotemporal dynamics. This negligence leads to impeded information flow and the accompanying degradation problem. In this paper, we identify the crux and then propose a novel residual block for SNNs, which is able to significantly extend the depth of directly trained SNNs, e.g., up to 482 layers on CIFAR-10 and 104 layers on ImageNet, without observing any slight degradation problem. We validate the effectiveness of our methods on both framebased and neuromorphic datasets, and our SRM-ResNet104 achieves a superior result of 76.02% accuracy on ImageNet, the first time in the domain of directly trained SNNs. The great energy efficiency is estimated and the resulting networks need on average only one spike per neuron for classifying an input sample. We believe our powerful and scalable modeling will provide a strong support for further exploration of SNNs.

[1]  Kaushik Roy,et al.  Going Deeper in Spiking Neural Networks: VGG and Residual Architectures , 2018, Front. Neurosci..

[2]  Tiejun Huang,et al.  Deep Residual Learning in Spiking Neural Networks , 2021, NeurIPS.

[3]  Jian Sun,et al.  Identity Mappings in Deep Residual Networks , 2016, ECCV.

[4]  Arindam Sanyal,et al.  Temporal-Coded Deep Spiking Neural Network with Easy Training and Robust Performance , 2019, AAAI.

[5]  Lei Deng,et al.  H2Learn: High-Efficiency Learning Accelerator for High-Accuracy Spiking Neural Networks , 2021, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[6]  Jacek Tabor,et al.  Dynamical Isometry is Achieved in Residual Networks in a Universal Way for any Activation Function , 2018, AISTATS.

[7]  Steve Furber,et al.  SpiNNaker 2: A 10 Million Core Processor System for Brain Simulation and Machine Learning , 2019, ArXiv.

[8]  Malu Zhang,et al.  Spike-Timing-Dependent Back Propagation in Deep Spiking Neural Networks , 2020, ArXiv.

[9]  Gang Pan,et al.  Spiking Deep Residual Network , 2018, ArXiv.

[10]  Hong Yang,et al.  DART: Distribution Aware Retinal Transform for Event-Based Cameras , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.

[11]  L. Deng,et al.  The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.

[12]  Serge J. Belongie,et al.  Residual Networks Behave Like Ensembles of Relatively Shallow Networks , 2016, NIPS.

[13]  Kilian Q. Weinberger,et al.  Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[14]  Zhaofei Yu,et al.  Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[15]  Alex Krizhevsky,et al.  Learning Multiple Layers of Features from Tiny Images , 2009 .

[16]  Sungroh Yoon,et al.  Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object Detection , 2020, AAAI.

[17]  Wolfgang Maass,et al.  Optimized spiking neurons can classify images with high accuracy through temporal coding with two spikes , 2020, Nature Machine Intelligence.

[18]  Luping Shi,et al.  CIFAR10-DVS: An Event-Stream Dataset for Object Classification , 2017, Front. Neurosci..

[19]  Bernard Brezzo,et al.  TrueNorth: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip , 2015, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[20]  Kaushik Roy,et al.  Enabling Deep Spiking Neural Networks with Hybrid Conversion and Spike Timing Dependent Backpropagation , 2020, ICLR.

[21]  Hong Wang,et al.  Loihi: A Neuromorphic Manycore Processor with On-Chip Learning , 2018, IEEE Micro.

[22]  Sergey Ioffe,et al.  Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.

[23]  Kaushik Roy,et al.  RMP-SNN: Residual Membrane Potential Neuron for Enabling Deeper High-Accuracy and Low-Latency Spiking Neural Network , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[24]  Guoqi Li,et al.  Temporal-wise Attention Spiking Neural Networks for Event Streams Classification , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[25]  Jürgen Schmidhuber,et al.  Training Very Deep Networks , 2015, NIPS.

[26]  Dumitru Erhan,et al.  Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[27]  Eugene M. Izhikevich,et al.  Simple model of spiking neurons , 2003, IEEE Trans. Neural Networks.

[28]  Wolfgang Maass,et al.  Networks of Spiking Neurons: The Third Generation of Neural Network Models , 1996, Electron. Colloquium Comput. Complex..

[29]  Kostas Daniilidis,et al.  Spike-FlowNet: Event-based Optical Flow Estimation with Energy-Efficient Hybrid Neural Networks , 2020, ECCV.

[30]  Mingguo Zhao,et al.  Towards artificial general intelligence with hybrid Tianjic chip architecture , 2019, Nature.

[31]  Mark Horowitz,et al.  1.1 Computing's energy problem (and what we can do about it) , 2014, 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC).

[32]  Elisabetta Chicca,et al.  Efficient Processing of Spatio-Temporal Data Streams With Spiking Neural Networks , 2020, Frontiers in Neuroscience.

[33]  Lei Deng,et al.  Going Deeper With Directly-Trained Larger Spiking Neural Networks , 2020, AAAI.

[34]  Eero P. Simoncelli,et al.  Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.

[35]  Li Fei-Fei,et al.  ImageNet: A large-scale hierarchical image database , 2009, CVPR.

[36]  L. F Abbott,et al.  Lapicque’s introduction of the integrate-and-fire model neuron (1907) , 1999, Brain Research Bulletin.

[37]  Jian Sun,et al.  Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[38]  A. Hodgkin,et al.  A quantitative description of membrane current and its application to conduction and excitation in nerve , 1952, The Journal of physiology.

[39]  Shih-Chii Liu,et al.  Conversion of Continuous-Valued Deep Networks to Efficient Event-Driven Networks for Image Classification , 2017, Front. Neurosci..

[40]  Hesham Mostafa,et al.  Surrogate Gradient Learning in Spiking Neural Networks: Bringing the Power of Gradient-based optimization to spiking neural networks , 2019, IEEE Signal Processing Magazine.

[41]  Kay Chen Tan,et al.  Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition , 2019, Frontiers in Neuroscience.

[42]  Lei Deng,et al.  DashNet: A Hybrid Artificial and Spiking Neural Network for High-speed Object Tracking , 2019, ArXiv.

[43]  Lei Deng,et al.  Direct Training for Spiking Neural Networks: Faster, Larger, Better , 2018, AAAI.