Reducing the number of neurons of Deep ReLU Networks based on the current theory of Regularization
暂无分享,去创建一个
[1] Hanan Samet,et al. Pruning Filters for Efficient ConvNets , 2016, ICLR.
[2] Nathan Srebro,et al. How do infinite width bounded norm networks look in function space? , 2019, COLT.
[3] Kai Yu,et al. Reshaping deep neural network for fast decoding by node-pruning , 2014, 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP).
[4] Josef Teichmann,et al. How implicit regularization of Neural Networks affects the learned function - Part I , 2019, ArXiv.
[5] Babak Hassibi,et al. Second Order Derivatives for Network Pruning: Optimal Brain Surgeon , 1992, NIPS.
[6] Nathan Srebro,et al. ` 1 Regularization in Infinite Dimensional Feature Spaces , 2007 .
[7] Emile Fiesler,et al. Pruning of Neural Networks , 1997 .
[8] Mingjie Sun,et al. Rethinking the Value of Network Pruning , 2018, ICLR.
[9] Michael Carbin,et al. The Lottery Ticket Hypothesis: Finding Sparse, Trainable Neural Networks , 2018, ICLR.
[10] Erich Elsen,et al. Sparse GPU Kernels for Deep Learning , 2020, SC20: International Conference for High Performance Computing, Networking, Storage and Analysis.
[11] Alexander M. Rush,et al. Movement Pruning: Adaptive Sparsity by Fine-Tuning , 2020, NeurIPS.
[12] Suvrit Sra,et al. Diversity Networks: Neural Network Compression Using Determinantal Point Processes , 2015, 1511.05077.
[13] Joan Bruna,et al. Gradient Dynamics of Shallow Univariate ReLU Networks , 2019, NeurIPS.
[14] Leslie Pack Kaelbling,et al. Generalization in Deep Learning , 2017, ArXiv.
[15] Sylvain Gelly,et al. Gradient Descent Quantizes ReLU Network Features , 2018, ArXiv.
[16] Ryota Tomioka,et al. In Search of the Real Inductive Bias: On the Role of Implicit Regularization in Deep Learning , 2014, ICLR.
[17] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[18] Yann LeCun,et al. Optimal Brain Damage , 1989, NIPS.
[19] R. Venkatesh Babu,et al. Data-free Parameter Pruning for Deep Neural Networks , 2015, BMVC.
[20] Rich Caruana,et al. Do Deep Nets Really Need to be Deep? , 2013, NIPS.
[21] Geoffrey E. Hinton,et al. Distilling the Knowledge in a Neural Network , 2015, ArXiv.
[22] Nathan Srebro,et al. A Function Space View of Bounded Norm Infinite Width ReLU Nets: The Multivariate Case , 2019, ICLR.