暂无分享,去创建一个
Richard G. Baraniuk | Daniel LeJeune | Hamid Javadi | Richard Baraniuk | Daniel LeJeune | Hamid Javadi
[1] Cun-Hui Zhang,et al. Scaled sparse linear regression , 2011, 1104.4595.
[2] P. Massart,et al. Gaussian model selection , 2001 .
[3] Stephen P. Boyd,et al. Enhancing Sparsity by Reweighted ℓ1 Minimization , 2007, 0711.1612.
[4] Colin Wei,et al. The Implicit and Explicit Regularization Effects of Dropout , 2020, ICML.
[5] Lianwen Jin,et al. Design of a Very Compact CNN Classifier for Online Handwritten Chinese Character Recognition Using DropWeight and Global Pooling , 2017, 2017 14th IAPR International Conference on Document Analysis and Recognition (ICDAR).
[6] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[7] Jianqing Fan,et al. Variable Selection via Nonconcave Penalized Likelihood and its Oracle Properties , 2001 .
[8] Emmanuel J. Candès,et al. SLOPE is Adaptive to Unknown Sparsity and Asymptotically Minimax , 2015, ArXiv.
[9] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[10] Erich Elsen,et al. The State of Sparsity in Deep Neural Networks , 2019, ArXiv.
[11] Jaehoon Lee,et al. Wide neural networks of any depth evolve as linear models under gradient descent , 2019, NeurIPS.
[12] Mikhail Belkin,et al. Classification vs regression in overparameterized regimes: Does the loss function matter? , 2020, J. Mach. Learn. Res..
[13] Suyog Gupta,et al. To prune, or not to prune: exploring the efficacy of pruning for model compression , 2017, ICLR.
[14] Nathan Srebro,et al. Dropout: Explicit Forms and Capacity Control , 2020, ICML.
[15] Julien Mairal,et al. Stochastic Majorization-Minimization Algorithms for Large-Scale Optimization , 2013, NIPS.
[16] Ji Xu,et al. On the proliferation of support vectors in high dimensions , 2020, ArXiv.
[17] Julien Mairal,et al. Optimization with Sparsity-Inducing Penalties , 2011, Found. Trends Mach. Learn..
[18] Xin Dong,et al. Learning to Prune Deep Neural Networks via Layer-wise Optimal Brain Surgeon , 2017, NIPS.
[19] Philip M. Long,et al. Surprising properties of dropout in deep networks , 2017, COLT.
[20] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[21] Wotao Yin,et al. Iteratively reweighted algorithms for compressive sensing , 2008, 2008 IEEE International Conference on Acoustics, Speech and Signal Processing.
[22] Adel Javanmard,et al. False Discovery Rate Control via Debiased Lasso , 2018, Electronic Journal of Statistics.
[23] Stephen P. Boyd,et al. Distributed Optimization and Statistical Learning via the Alternating Direction Method of Multipliers , 2011, Found. Trends Mach. Learn..
[24] David P. Wipf,et al. Iterative Reweighted 1 and 2 Methods for Finding Sparse Solutions , 2010, IEEE J. Sel. Top. Signal Process..
[25] Max Welling,et al. Learning Sparse Neural Networks through L0 Regularization , 2017, ICLR.
[26] Trevor Hastie,et al. Statistical Learning with Sparsity: The Lasso and Generalizations , 2015 .
[27] Adel Javanmard,et al. Debiasing the lasso: Optimal sample size for Gaussian designs , 2015, The Annals of Statistics.
[28] Christopher D. Manning,et al. Fast dropout training , 2013, ICML.
[29] Zoubin Ghahramani,et al. Variational Bayesian dropout: pitfalls and fixes , 2018, ICML.
[30] Dmitry P. Vetrov,et al. Variational Dropout Sparsifies Deep Neural Networks , 2017, ICML.
[31] H. Zou,et al. One-step Sparse Estimates in Nonconcave Penalized Likelihood Models. , 2008, Annals of statistics.
[32] Brendan J. Frey,et al. Adaptive dropout for training deep neural networks , 2013, NIPS.
[33] Raman Arora,et al. On Convergence and Generalization of Dropout Training , 2020, NeurIPS.
[34] I. Daubechies,et al. Iteratively reweighted least squares minimization for sparse recovery , 2008, 0807.0575.
[35] Ariel D. Procaccia,et al. Variational Dropout and the Local Reparameterization Trick , 2015, NIPS.
[36] Sida I. Wang,et al. Dropout Training as Adaptive Regularization , 2013, NIPS.
[37] Cun-Hui Zhang. Nearly unbiased variable selection under minimax concave penalty , 2010, 1002.4734.
[38] H. Zou,et al. Regularization and variable selection via the elastic net , 2005 .
[39] F. Bach,et al. Optimization with Sparsity-Inducing Penalties (Foundations and Trends(R) in Machine Learning) , 2011 .
[40] Francis R. Bach,et al. Structured Sparse Principal Component Analysis , 2009, AISTATS.
[41] Yurong Chen,et al. Dynamic Network Surgery for Efficient DNNs , 2016, NIPS.
[42] Zoubin Ghahramani,et al. Dropout as a Bayesian Approximation: Representing Model Uncertainty in Deep Learning , 2015, ICML.
[43] Song Han,et al. Learning both Weights and Connections for Efficient Neural Network , 2015, NIPS.
[44] Mike E. Davies,et al. Iterative Hard Thresholding for Compressed Sensing , 2008, ArXiv.
[45] Frederick R. Forst,et al. On robust estimation of the location parameter , 1980 .