DropNet: Reducing Neural Network Complexity via Iterative Pruning
暂无分享,去创建一个
[1] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[2] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[3] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[4] Junfei Qiao,et al. A node pruning algorithm for feedforward neural network based on neural complexity , 2010, 2010 International Conference on Intelligent Control and Information Processing.
[5] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[6] Yiran Chen,et al. Learning Structured Sparsity in Deep Neural Networks , 2016, NIPS.
[7] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[8] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[9] Giovanna Castellano,et al. An iterative pruning algorithm for feedforward neural networks , 1997, IEEE Trans. Neural Networks.
[10] Gregory J. Wolff,et al. Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.
[11] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[12] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[13] Lukasz Kaiser,et al. Attention is All you Need , 2017, NIPS.
[14] Mathieu Salzmann,et al. Learning the Number of Neurons in Deep Networks , 2016, NIPS.
[15] Pritish Narayanan,et al. Deep Learning with Limited Numerical Precision , 2015, ICML.
[16] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[17] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.
[18] Zhiqiang Shen,et al. Learning Efficient Convolutional Networks through Network Slimming , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[19] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[21] Wonyong Sung,et al. Structured Pruning of Deep Convolutional Neural Networks , 2015, ACM J. Emerg. Technol. Comput. Syst..
[22] Rongrong Ji,et al. Accelerating Convolutional Networks via Global & Dynamic Filter Pruning , 2018, IJCAI.
[23] Xiangyu Zhang,et al. Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[24] Lior Wolf,et al. Channel-Level Acceleration of Deep Face Representations , 2015, IEEE Access.