The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks
暂无分享,去创建一个
[1] J. Rissanen. Stochastic Complexity and Modeling , 1986 .
[2] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[3] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[4] Gregory J. Wolff,et al. Optimal Brain Surgeon and general network pruning , 1993, IEEE International Conference on Neural Networks.
[5] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[6] Carl E. Rasmussen,et al. Occam's Razor , 2000, NIPS.
[7] Nicolas Le Roux,et al. Convex Neural Networks , 2005, NIPS.
[8] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[9] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[10] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[11] Yann LeCun,et al. Regularization of Neural Networks using DropConnect , 2013, ICML.
[12] Pierre Baldi,et al. Understanding Dropout , 2013, NIPS.
[13] Misha Denil,et al. Predicting Parameters in Deep Learning , 2014 .
[14] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[15] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[16] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[17] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[18] Song Han,et al. A Deep Neural Network Compression Pipeline: Pruning, Quantization, Huffman Encoding , 2015 .
[19] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[20] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[21] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[22] R. Venkatesh Babu,et al. Data-free Parameter Pruning for Deep Neural Networks , 2015, BMVC.
[23] R. Venkatesh Babu,et al. Learning the Architecture of Deep Neural Networks , 2015, ArXiv.
[24] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[25] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[26] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[27] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Transfer Learning , 2016, ArXiv.
[28] Joseph Paul Cohen,et al. RandomOut: Using a convolutional gradient norm to win The Filter Lottery , 2016, ArXiv.
[29] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] R. Venkatesh Babu,et al. Learning Neural Network Architectures using Backpropagation , 2015, BMVC.
[31] Suvrit Sra,et al. Diversity Networks , 2015, ICLR.
[32] Song Han,et al. DSD: Regularizing Deep Neural Networks with Dense-Sparse-Dense Training Flow , 2016, ArXiv.
[33] Rui Peng,et al. Network Trimming: A Data-Driven Neuron Pruning Approach towards Efficient Deep Architectures , 2016, ArXiv.
[34] Song Han,et al. Deep Compression: Compressing Deep Neural Network with Pruning, Trained Quantization and Huffman Coding , 2015, ICLR.
[35] Shuicheng Yan,et al. Training Skinny Deep Neural Networks with Iterative Hard Thresholding Methods , 2016, ArXiv.
[36] Ananthram Swami,et al. Distillation as a Defense to Adversarial Perturbations Against Deep Neural Networks , 2015, 2016 IEEE Symposium on Security and Privacy (SP).
[37] Forrest N. Iandola,et al. SqueezeNet: AlexNet-level accuracy with 50x fewer parameters and <1MB model size , 2016, ArXiv.
[38] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[39] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[40] Jianxin Wu,et al. ThiNet: A Filter Level Pruning Method for Deep Neural Network Compression , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[41] Erich Elsen,et al. Exploring Sparsity in Recurrent Neural Networks , 2017, ICLR.
[42] Alex Kendall,et al. Concrete Dropout , 2017, NIPS.
[43] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[44] Bo Chen,et al. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications , 2017, ArXiv.
[45] Vivienne Sze,et al. Designing Energy-Efficient Convolutional Neural Networks Using Energy-Aware Pruning , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[46] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.
[47] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[48] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[49] Xin Dong,et al. Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon , 2017, NIPS.
[50] Max Welling,et al. Bayesian Compression for Deep Learning , 2017, NIPS.
[51] Amnon Shashua,et al. Inductive Bias of Deep Convolutional Networks through Pooling Geometry , 2016, ICLR.
[52] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[53] R. Venkatesh Babu,et al. Training Sparse Neural Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[54] Dmitry P. Vetrov,et al. Structured Bayesian Pruning via Log-Normal Multiplicative Noise , 2017, NIPS.
[55] David A. Wagner,et al. Towards Evaluating the Robustness of Neural Networks , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[56] Xiangyu Zhang,et al. Channel Pruning for Accelerating Very Deep Neural Networks , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[57] Ryan P. Adams,et al. Compressibility and Generalization in Large-Scale Deep Learning , 2018, ArXiv.
[58] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[59] Yi Zhang,et al. Stronger generalization bounds for deep nets via a compression approach , 2018, ICML.
[60] David Kappel,et al. Deep Rewiring: Training very sparse deep networks , 2017, ICLR.
[61] Jason Yosinski,et al. Measuring the Intrinsic Dimension of Objective Landscapes , 2018, ICLR.
[62] Ryan P. Adams,et al. Non-vacuous Generalization Bounds at the ImageNet Scale: a PAC-Bayesian Compression Approach , 2018, ICLR.
[63] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[64] Barnabás Póczos,et al. Gradient Descent Provably Optimizes Over-parameterized Neural Networks , 2018, ICLR.