暂无分享,去创建一个
Sylvain Sardy | Nicolas W Hengartner | Yen Ting Lin | Nikolai Bonenko | S. Sardy | N. Hengartner | Yen-Ting Lin | Nikolai Bobenko
[1] Nitish Srivastava,et al. Dropout: a simple way to prevent neural networks from overfitting , 2014, J. Mach. Learn. Res..
[2] I. Johnstone,et al. Ideal spatial adaptation by wavelet shrinkage , 1994 .
[3] I. Johnstone,et al. Wavelet Shrinkage: Asymptopia? , 1995 .
[4] Alexandros Kalousis,et al. Regularising Non-linear Models Using Feature Side-information , 2017, ICML.
[5] Andrew M. Saxe,et al. High-dimensional dynamics of generalization error in neural networks , 2017, Neural Networks.
[6] Andrea Montanari,et al. The Generalization Error of Random Features Regression: Precise Asymptotics and the Double Descent Curve , 2019, Communications on Pure and Applied Mathematics.
[7] Levent Sagun,et al. Scaling description of generalization with number of parameters in deep learning , 2019, Journal of Statistical Mechanics: Theory and Experiment.
[8] S. Sardy,et al. Quantile universal threshold , 2017 .
[9] R. Tibshirani. Regression Shrinkage and Selection via the Lasso , 1996 .
[10] Wyeth W. Wasserman,et al. Deep Feature Selection: Theory and Application to Identify Enhancers and Promoters , 2015, RECOMB.
[11] Peng Zhang,et al. Transformed 𝓁1 Regularization for Learning Sparse Deep Neural Networks , 2019, Neural Networks.
[12] S. Sardy,et al. Model Selection With Lasso-Zero: Adding Straw to the Haystack to Better Find Needles , 2018, J. Comput. Graph. Stat..
[13] Nitish Srivastava,et al. Improving neural networks by preventing co-adaptation of feature detectors , 2012, ArXiv.
[14] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[15] Julien Mairal,et al. Optimization with Sparsity-Inducing Penalties , 2011, Found. Trends Mach. Learn..
[16] Y. Benjamini,et al. Controlling the false discovery rate: a practical and powerful approach to multiple testing , 1995 .
[17] Sara van de Geer,et al. Statistics for High-Dimensional Data: Methods, Theory and Applications , 2011 .
[18] Marc Teboulle,et al. A Fast Iterative Shrinkage-Thresholding Algorithm for Linear Inverse Problems , 2009, SIAM J. Imaging Sci..
[19] Peter Bühlmann,et al. High-Dimensional Statistics with a View Toward Applications in Biology , 2014 .
[20] A. Belloni,et al. Square-Root Lasso: Pivotal Recovery of Sparse Signals via Conic Programming , 2010, 1009.5689.
[21] S. Frick,et al. Compressed Sensing , 2014, Computer Vision, A Reference Guide.
[22] Marc'Aurelio Ranzato,et al. Sparse Feature Learning for Deep Belief Networks , 2007, NIPS.
[23] Rajat Raina,et al. Efficient sparse coding algorithms , 2006, NIPS.
[24] I. Johnstone,et al. Needles and straw in haystacks: Empirical Bayes estimates of possibly sparse sequences , 2004, math/0410088.
[25] Arthur E. Hoerl,et al. Ridge Regression: Biased Estimation for Nonorthogonal Problems , 2000, Technometrics.
[26] Pushmeet Kohli,et al. Memory Bounded Deep Convolutional Networks , 2014, ArXiv.
[27] Andrea Montanari,et al. The Noise-Sensitivity Phase Transition in Compressed Sensing , 2010, IEEE Transactions on Information Theory.
[28] Michael A. Saunders,et al. Atomic Decomposition by Basis Pursuit , 1998, SIAM J. Sci. Comput..
[29] Emmanuel J. Candès,et al. Decoding by linear programming , 2005, IEEE Transactions on Information Theory.