暂无分享,去创建一个
[1] Kaushik Roy,et al. Going Deeper in Spiking Neural Networks: VGG and Residual Architectures , 2018, Front. Neurosci..
[2] Tiejun Huang,et al. Deep Residual Learning in Spiking Neural Networks , 2021, NeurIPS.
[3] Jian Sun,et al. Identity Mappings in Deep Residual Networks , 2016, ECCV.
[4] Arindam Sanyal,et al. Temporal-Coded Deep Spiking Neural Network with Easy Training and Robust Performance , 2019, AAAI.
[5] Lei Deng,et al. H2Learn: High-Efficiency Learning Accelerator for High-Accuracy Spiking Neural Networks , 2021, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.
[6] Jacek Tabor,et al. Dynamical Isometry is Achieved in Residual Networks in a Universal Way for any Activation Function , 2018, AISTATS.
[7] Steve Furber,et al. SpiNNaker 2: A 10 Million Core Processor System for Brain Simulation and Machine Learning , 2019, ArXiv.
[8] Malu Zhang,et al. Spike-Timing-Dependent Back Propagation in Deep Spiking Neural Networks , 2020, ArXiv.
[9] Gang Pan,et al. Spiking Deep Residual Network , 2018, ArXiv.
[10] Hong Yang,et al. DART: Distribution Aware Retinal Transform for Event-Based Cameras , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[11] L. Deng,et al. The MNIST Database of Handwritten Digit Images for Machine Learning Research [Best of the Web] , 2012, IEEE Signal Processing Magazine.
[12] Serge J. Belongie,et al. Residual Networks Behave Like Ensembles of Relatively Shallow Networks , 2016, NIPS.
[13] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[14] Zhaofei Yu,et al. Incorporating Learnable Membrane Time Constant to Enhance Learning of Spiking Neural Networks , 2020, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[15] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[16] Sungroh Yoon,et al. Spiking-YOLO: Spiking Neural Network for Energy-Efficient Object Detection , 2020, AAAI.
[17] Wolfgang Maass,et al. Optimized spiking neurons can classify images with high accuracy through temporal coding with two spikes , 2020, Nature Machine Intelligence.
[18] Luping Shi,et al. CIFAR10-DVS: An Event-Stream Dataset for Object Classification , 2017, Front. Neurosci..
[19] Bernard Brezzo,et al. TrueNorth: Design and Tool Flow of a 65 mW 1 Million Neuron Programmable Neurosynaptic Chip , 2015, IEEE Transactions on Computer-Aided Design of Integrated Circuits and Systems.
[20] Kaushik Roy,et al. Enabling Deep Spiking Neural Networks with Hybrid Conversion and Spike Timing Dependent Backpropagation , 2020, ICLR.
[21] Hong Wang,et al. Loihi: A Neuromorphic Manycore Processor with On-Chip Learning , 2018, IEEE Micro.
[22] Sergey Ioffe,et al. Inception-v4, Inception-ResNet and the Impact of Residual Connections on Learning , 2016, AAAI.
[23] Kaushik Roy,et al. RMP-SNN: Residual Membrane Potential Neuron for Enabling Deeper High-Accuracy and Low-Latency Spiking Neural Network , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[24] Guoqi Li,et al. Temporal-wise Attention Spiking Neural Networks for Event Streams Classification , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).
[25] Jürgen Schmidhuber,et al. Training Very Deep Networks , 2015, NIPS.
[26] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Eugene M. Izhikevich,et al. Simple model of spiking neurons , 2003, IEEE Trans. Neural Networks.
[28] Wolfgang Maass,et al. Networks of Spiking Neurons: The Third Generation of Neural Network Models , 1996, Electron. Colloquium Comput. Complex..
[29] Kostas Daniilidis,et al. Spike-FlowNet: Event-based Optical Flow Estimation with Energy-Efficient Hybrid Neural Networks , 2020, ECCV.
[30] Mingguo Zhao,et al. Towards artificial general intelligence with hybrid Tianjic chip architecture , 2019, Nature.
[31] Mark Horowitz,et al. 1.1 Computing's energy problem (and what we can do about it) , 2014, 2014 IEEE International Solid-State Circuits Conference Digest of Technical Papers (ISSCC).
[32] Elisabetta Chicca,et al. Efficient Processing of Spatio-Temporal Data Streams With Spiking Neural Networks , 2020, Frontiers in Neuroscience.
[33] Lei Deng,et al. Going Deeper With Directly-Trained Larger Spiking Neural Networks , 2020, AAAI.
[34] Eero P. Simoncelli,et al. Image quality assessment: from error visibility to structural similarity , 2004, IEEE Transactions on Image Processing.
[35] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[36] L. F Abbott,et al. Lapicque’s introduction of the integrate-and-fire model neuron (1907) , 1999, Brain Research Bulletin.
[37] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] A. Hodgkin,et al. A quantitative description of membrane current and its application to conduction and excitation in nerve , 1952, The Journal of physiology.
[39] Shih-Chii Liu,et al. Conversion of Continuous-Valued Deep Networks to Efficient Event-Driven Networks for Image Classification , 2017, Front. Neurosci..
[40] Hesham Mostafa,et al. Surrogate Gradient Learning in Spiking Neural Networks: Bringing the Power of Gradient-based optimization to spiking neural networks , 2019, IEEE Signal Processing Magazine.
[41] Kay Chen Tan,et al. Deep Spiking Neural Networks for Large Vocabulary Automatic Speech Recognition , 2019, Frontiers in Neuroscience.
[42] Lei Deng,et al. DashNet: A Hybrid Artificial and Spiking Neural Network for High-speed Object Tracking , 2019, ArXiv.
[43] Lei Deng,et al. Direct Training for Spiking Neural Networks: Faster, Larger, Better , 2018, AAAI.