Pruning of Deep Spiking Neural Networks through Gradient Rewiring

Spiking Neural Networks (SNNs) have been attached great importance due to their biological plausibility and high energy-efficiency on neuromorphic chips. As these chips are usually resource-constrained, the compression of SNNs is thus crucial along the road of practical use of SNNs. Most existing methods directly apply pruning approaches in artificial neural networks (ANNs) to SNNs, which ignore the difference between ANNs and SNNs, thus limiting the performance of the pruned SNNs. Besides, these methods are only suitable for shallow SNNs. In this paper, inspired by synaptogenesis and synapse elimination in the neural system, we propose gradient rewiring (Grad R), a joint learning algorithm of connectivity and weight for SNNs, that enables us to seamlessly optimize network structure without retraining. Our key innovation is to redefine the gradient to a new synaptic parameter, allowing better exploration of network structures by taking full advantage of the competition between pruning and regrowth of connections. The experimental results show that the proposed method achieves minimal loss of SNNs' performance on MNIST and CIFAR-10 datasets so far. Moreover, it reaches a ~3.5% accuracy loss under unprecedented 0.73% connectivity, which reveals remarkable structure refining capability in SNNs. Our work suggests that there exists extremely high redundancy in deep SNNs. Our codes are available at https://github.com/Yanqi-Chen/Gradient-Rewiring.

[1]  Ahmed M. Eltawil,et al.  Unsupervised Adaptive Weight Pruning for Energy-Efficient Neuromorphic Systems , 2020, Frontiers in Neuroscience.

[2]  Yonghong Tian,et al.  Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[3]  Friedemann Zenke,et al.  The remarkable robustness of surrogate gradient learning for instilling complex function in spiking neural networks , 2020, bioRxiv.

[4]  2019 15th International Conference on Computational Intelligence and Security (CIS) , 2019 .

[5]  Xianghong Lin,et al.  An Adaptive Structure Learning Algorithm for Multi-Layer Spiking Neural Networks , 2019, 2019 15th International Conference on Computational Intelligence and Security (CIS).

[6]  Kun Qian,et al.  Application of Deep Compression Technique in Spiking Neural Network Chip , 2019, IEEE Transactions on Biomedical Circuits and Systems.

[7]  Hesham Mostafa,et al.  Surrogate Gradient Learning in Spiking Neural Networks: Bringing the Power of Gradient-based optimization to spiking neural networks , 2019, IEEE Signal Processing Magazine.

[8]  Lei Deng,et al.  Comprehensive SNN Compression Using ADMM Optimization and Activity Regularization , 2019, IEEE Transactions on Neural Networks and Learning Systems.

[9]  M. Cernak,et al.  Spiking Neural Networks Trained With Backpropagation for Low Power Neuromorphic Implementation of Voice Activity Detection , 2019, ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).

[10]  Kaushik Roy,et al.  Towards spike-based machine intelligence with neuromorphic computing , 2019, Nature.

[11]  Sangheon Oh,et al.  A Soft-Pruning Method Applied During Training of Spiking Neural Networks for In-memory Computing Applications , 2019, Front. Neurosci..

[12]  Kaushik Roy,et al.  Enabling Spike-Based Backpropagation for Training Deep Neural Network Architectures , 2019, Frontiers in Neuroscience.

[13]  Steve B. Furber,et al.  Memory-Efficient Deep Learning on a SpiNNaker 2 Prototype , 2018, Front. Neurosci..

[14]  Lei Deng,et al.  Direct Training for Spiking Neural Networks: Faster, Larger, Better , 2018, AAAI.

[15]  Ieee Circuits,et al.  IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems information for authors , 2018, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[16]  Zhaohui Wu,et al.  Jointly Learning Network Connections and Link Weights in Spiking Neural Networks , 2018, IJCAI.

[17]  Robert A. Legenstein,et al.  Long short-term memory and Learning-to-learn in networks of spiking neurons , 2018, NeurIPS.

[18]  Michael Carbin,et al.  The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.

[19]  David Kappel,et al.  Deep Rewiring: Training very sparse deep networks , 2017, ICLR.

[20]  Kaushik Roy,et al.  STDP-Based Pruning of Connections and Weight Quantization in Spiking Neural Networks for Energy-Efficient Recognition , 2017, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.

[21]  Lei Deng,et al.  Spatio-Temporal Backpropagation for Training High-Performance Spiking Neural Networks , 2017, Front. Neurosci..

[22]  David Kappel,et al.  Network Plasticity as Bayesian Inference , 2015, PLoS Comput. Biol..

[23]  P. Huttenlocher Synaptogenesis, Synapse Elimination, and Neural Plasticity in Human Cerebral Cortex , 2013 .

[24]  Current Biology , 2012, Current Biology.

[25]  P. Cochat,et al.  Et al , 2008, Archives de pediatrie : organe officiel de la Societe francaise de pediatrie.

[26]  Shih-Chii Liu,et al.  Quantifying Input and Output Spike Statistics of a Winner-Take-All Network in a Vision System , 2007, 2007 IEEE International Symposium on Circuits and Systems.

[27]  H. Sompolinsky,et al.  The tempotron: a neuron that learns spike timing–based decisions , 2006, Nature Neuroscience.

[28]  Bartlett W. Mel,et al.  Cortical rewiring and information storage , 2004, Nature.

[29]  D. Koshland Frontiers in neuroscience. , 1988, Science.

[30]  Shoko Imaizumi,et al.  国際会議開催報告:IEEE International Symposium on Circuits and Systems , 2013 .

[31]  Wolfgang Maass,et al.  Networks of Spiking Neurons: The Third Generation of Neural Network Models , 1996, Electron. Colloquium Comput. Complex..