暂无分享,去创建一个
[1] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[2] Michael C. Mozer,et al. Using Relevance to Reduce Network Size Automatically , 1989 .
[3] Jian Sun,et al. Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).
[4] Giovanna Castellano,et al. An iterative pruning algorithm for feedforward neural networks , 1997, IEEE Trans. Neural Networks.
[5] Ali Farhadi,et al. Discovering Neural Wirings , 2019, NeurIPS.
[6] Michael Carbin,et al. Comparing Rewinding and Fine-tuning in Neural Network Pruning , 2019, ICLR.
[7] Gintare Karolina Dziugaite,et al. Pruning Neural Networks at Initialization: Why are We Missing the Mark? , 2020, ArXiv.
[8] Timo Aila,et al. Pruning Convolutional Neural Networks for Resource Efficient Inference , 2016, ICLR.
[9] Erich Elsen,et al. The Difficulty of Training Sparse Neural Networks , 2019, ArXiv.
[10] Andrew Gordon Wilson,et al. Loss Surfaces, Mode Connectivity, and Fast Ensembling of DNNs , 2018, NeurIPS.
[11] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[12] Tengyu Ma,et al. Fixup Initialization: Residual Learning Without Normalization , 2019, ICLR.
[13] David J. Schwab,et al. The Early Phase of Neural Network Training , 2020, ICLR.
[14] Lawrence D. Jackel,et al. Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.
[15] Jascha Sohl-Dickstein,et al. Dynamical Isometry and a Mean Field Theory of CNNs: How to Train 10, 000-Layer Vanilla Convolutional Neural Networks , 2018, ICML.
[16] Pierre Geurts,et al. On the Transferability of Winning Tickets in Non-Natural Image Datasets , 2020, ArXiv.
[17] Lorien Y. Pratt,et al. Comparing Biases for Minimal Network Construction with Back-Propagation , 1988, NIPS.
[18] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[19] Yann Dauphin,et al. Empirical Analysis of the Hessian of Over-Parametrized Neural Networks , 2017, ICLR.
[20] Max Welling,et al. Bayesian Compression for Deep Learning , 2017, NIPS.
[21] Suman Jana,et al. Towards Practical Lottery Ticket Hypothesis for Adversarial Training , 2020, ArXiv.
[22] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[23] Ali Farhadi,et al. What’s Hidden in a Randomly Weighted Neural Network? , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[24] R.J.F. Dow,et al. Neural net pruning-why and how , 1988, IEEE 1988 International Conference on Neural Networks.
[25] Yuandong Tian,et al. One ticket to win them all: generalizing lottery ticket initializations across datasets and optimizers , 2019, NeurIPS.
[26] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[27] Surya Ganguli,et al. Pruning neural networks without any data by iteratively conserving synaptic flow , 2020, NeurIPS.
[28] Michael C. Mozer,et al. Skeletonization: A Technique for Trimming the Fat from a Network via Relevance Assessment , 1988, NIPS.
[29] Sergey Ioffe,et al. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift , 2015, ICML.
[30] Hermann Ney,et al. Successfully Applying the Stabilized Lottery Ticket Hypothesis to the Transformer Architecture , 2020, ACL.
[31] Ali Farhadi,et al. Soft Threshold Weight Reparameterization for Learnable Sparsity , 2020, ICML.
[32] Gintare Karolina Dziugaite,et al. Linear Mode Connectivity and the Lottery Ticket Hypothesis , 2019, ICML.
[33] Philip H. S. Torr,et al. SNIP: Single-shot Network Pruning based on Connection Sensitivity , 2018, ICLR.
[34] Luke Zettlemoyer,et al. Sparse Networks from Scratch: Faster Training without Losing Performance , 2019, ArXiv.
[35] Song Han,et al. EIE: Efficient Inference Engine on Compressed Deep Neural Network , 2016, 2016 ACM/IEEE 43rd Annual International Symposium on Computer Architecture (ISCA).
[36] Fred A. Hamprecht,et al. Essentially No Barriers in Neural Network Energy Landscape , 2018, ICML.
[37] Yoshua Bengio,et al. Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.
[38] Erich Elsen,et al. Rigging the Lottery: Making All Tickets Winners , 2020, ICML.
[39] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[40] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[41] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[42] Bindya Venkatesh,et al. Calibrate and Prune: Improving Reliability of Lottery Tickets Through Prediction Calibration , 2020, ArXiv.
[43] Peter Stone,et al. Scalable training of artificial neural networks with adaptive sparse connectivity inspired by network science , 2017, Nature Communications.
[44] Oriol Vinyals,et al. Qualitatively characterizing neural network optimization problems , 2014, ICLR.
[45] Gaël Varoquaux,et al. The NumPy Array: A Structure for Efficient Numerical Computation , 2011, Computing in Science & Engineering.
[46] Xin Wang,et al. Parameter Efficient Training of Deep Convolutional Neural Networks by Dynamic Sparse Reparameterization , 2019, ICML.
[47] Rudy Setiono,et al. A Penalty-Function Approach for Pruning Feedforward Neural Networks , 1997, Neural Computation.
[48] Balaji Lakshminarayanan,et al. Deep Ensembles: A Loss Landscape Perspective , 2019, ArXiv.
[49] Guodong Zhang,et al. Picking Winning Tickets Before Training by Preserving Gradient Flow , 2020, ICLR.
[50] Gilad Yehudai,et al. Proving the Lottery Ticket Hypothesis: Pruning is All You Need , 2020, ICML.
[51] Vardan Papyan,et al. Measurements of Three-Level Hierarchical Structure in the Outliers in the Spectrum of Deepnet Hessians , 2019, ICML.
[52] Gintare Karolina Dziugaite,et al. Stabilizing the Lottery Ticket Hypothesis , 2019 .
[53] Jascha Sohl-Dickstein,et al. The large learning rate phase of deep learning: the catapult mechanism , 2020, ArXiv.
[54] Yan Liu,et al. Deep residual learning for image steganalysis , 2018, Multimedia Tools and Applications.
[55] Shankar Krishnan,et al. An Investigation into Neural Net Optimization via Hessian Eigenvalue Density , 2019, ICML.
[56] J. Kruskal. Multidimensional scaling by optimizing goodness of fit to a nonmetric hypothesis , 1964 .
[57] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[58] Jason Yosinski,et al. Deconstructing Lottery Tickets: Zeros, Signs, and the Supermask , 2019, NeurIPS.