One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers
暂无分享,去创建一个
Yuandong Tian | Michela Paganini | Haonan Yu | Ari S. Morcos | Yuandong Tian | Michela Paganini | Haonan Yu
[1] Chenchen Liu,et al. Interpretable Convolutional Filter Pruning , 2018, ArXiv.
[2] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[3] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[4] Yoshua Bengio,et al. How transferable are features in deep neural networks? , 2014, NIPS.
[5] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[6] Andrew Zisserman,et al. Very Deep Convolutional Networks for Large-Scale Image Recognition , 2014, ICLR.
[7] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[8] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[9] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[10] Yuanzhi Li,et al. Learning and Generalization in Overparameterized Neural Networks, Going Beyond Two Layers , 2018, NeurIPS.
[11] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[12] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[13] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[14] Yuanzhi Li,et al. A Convergence Theory for Deep Learning via Over-Parameterization , 2018, ICML.
[15] Bolei Zhou,et al. Places: A 10 Million Image Database for Scene Recognition , 2018, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[16] Surya Ganguli,et al. Deep Information Propagation , 2016, ICLR.
[17] Gintare Karolina Dziugaite,et al. The Lottery Ticket Hypothesis at Scale , 2019, ArXiv.
[18] Tengyu Ma,et al. Fixup Initialization: Residual Learning Without Normalization , 2019, ICLR.
[19] Steve Kroon,et al. Critical initialisation for deep signal propagation in noisy rectifier neural networks , 2018, NeurIPS.
[20] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[21] Yann LeCun,et al. Towards Understanding the Role of Over-Parametrization in Generalization of Neural Networks , 2018, ArXiv.
[22] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[23] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[24] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[25] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[26] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[27] David Rolnick,et al. How to Start Training: The Effect of Initialization and Architecture , 2018, NeurIPS.
[28] Andrew Y. Ng,et al. Reading Digits in Natural Images with Unsupervised Feature Learning , 2011 .
[29] Fei-Fei Li,et al. ImageNet: A large-scale hierarchical image database , 2009, 2009 IEEE Conference on Computer Vision and Pattern Recognition.
[30] Jason D. Lee,et al. On the Power of Over-parametrization in Neural Networks with Quadratic Activation , 2018, ICML.
[31] Barnabás Póczos,et al. Gradient Descent Provably Optimizes Over-parameterized Neural Networks , 2018, ICLR.
[32] Samuel S. Schoenholz,et al. Deep Mean Field Theory: Layerwise Variance and Width Variation as Methods to Control Gradient Explosion , 2018, ICLR 2018.
[33] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[34] Vijay Vasudevan,et al. Learning Transferable Architectures for Scalable Image Recognition , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[35] Jacek M. Zurada,et al. Redundant feature pruning for accelerated inference in deep neural networks , 2019, Neural Networks.
[36] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.
[37] Chao Yang,et al. A Survey on Deep Transfer Learning , 2018, ICANN.
[38] Quoc V. Le,et al. Do Better ImageNet Models Transfer Better? , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).