暂无分享,去创建一个
[1] Mikhail Belkin,et al. Classification vs regression in overparameterized regimes: Does the loss function matter? , 2020, J. Mach. Learn. Res..
[2] Noureddine El Karoui,et al. The spectrum of kernel random matrices , 2010, 1001.0492.
[3] Tengyuan Liang,et al. A Precise High-Dimensional Asymptotic Theory for Boosting and Min-L1-Norm Interpolated Classifiers , 2020, ArXiv.
[4] Andrea Montanari,et al. The generalization error of max-margin linear classifiers: High-dimensional asymptotics in the overparametrized regime , 2019 .
[5] Peter L. Bartlett,et al. The importance of convexity in learning with squared loss , 1998, COLT '96.
[6] Reinhard Heckel,et al. Compressive sensing with un-trained neural networks: Gradient descent finds the smoothest approximation , 2020, ICML.
[7] Roman Vershynin,et al. Introduction to the non-asymptotic analysis of random matrices , 2010, Compressed Sensing.
[8] Philip M. Long,et al. Benign overfitting in linear regression , 2019, Proceedings of the National Academy of Sciences.
[9] Michael I. Jordan,et al. Advances in Neural Information Processing Systems 30 , 1995 .
[10] Ion Stoica,et al. Numpywren: Serverless Linear Algebra , 2018, ArXiv.
[11] Tengyuan Liang,et al. On the Multiple Descent of Minimum-Norm Interpolants and Restricted Lower Isometry of Kernels , 2019, COLT.
[12] V. Koltchinskii,et al. Empirical margin distributions and bounding the generalization error of combined classifiers , 2002, math/0405343.
[13] Lorenzo Rosasco,et al. FALKON: An Optimal Large Scale Kernel Method , 2017, NIPS.
[14] Christos Thrampoulidis,et al. A Model of Double Descent for High-dimensional Binary Linear Classification , 2019, ArXiv.
[15] Hans Ulrich Simon,et al. From noise-free to noise-tolerant and from on-line to batch learning , 1995, COLT '95.
[16] Stephen Tyree,et al. Exact Gaussian Processes on a Million Data Points , 2019, NeurIPS.
[17] David P. Woodruff,et al. Faster Kernel Ridge Regression Using Sketching and Preconditioning , 2016, SIAM J. Matrix Anal. Appl..
[18] Tengyuan Liang,et al. Just Interpolate: Kernel "Ridgeless" Regression Can Generalize , 2018, The Annals of Statistics.
[19] Albert B Novikoff,et al. ON CONVERGENCE PROOFS FOR PERCEPTRONS , 1963 .
[20] 俊一 甘利. 5分で分かる!? 有名論文ナナメ読み:Jacot, Arthor, Gabriel, Franck and Hongler, Clement : Neural Tangent Kernel : Convergence and Generalization in Neural Networks , 2020 .
[21] Jonathan Ragan-Kelley,et al. Neural Kernels Without Tangents , 2020, ICML.
[22] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[23] Ethan Bernstein. Absolute error bounds for learning linear functions online , 1992, COLT '92.
[24] Philip M. Long,et al. Finite-sample analysis of interpolating linear classifiers in the overparameterized regime , 2020, ArXiv.
[25] V. Vapnik,et al. Bounds on Error Expectation for Support Vector Machines , 2000, Neural Computation.
[26] Mikhail Belkin,et al. Diving into the shallows: a computational perspective on large-scale shallow learning , 2017, NIPS.