暂无分享,去创建一个
[1] Feller William,et al. An Introduction To Probability Theory And Its Applications , 1950 .
[2] M. W. Birch. Maximum Likelihood in Three-Way Contingency Tables , 1963 .
[3] E. Slud. Distribution Inequalities for the Binomial Law , 1977 .
[4] S. Addelman. Statistics for experimenters , 1978 .
[5] V. V. Buldygin,et al. Sub-Gaussian random variables , 1980 .
[6] David Haussler,et al. Decision Theoretic Generalizations of the PAC Model for Neural Net and Other Learning Applications , 1992, Inf. Comput..
[7] Peter L. Bartlett,et al. The Sample Complexity of Pattern Classification with Neural Networks: The Size of the Weights is More Important than the Size of the Network , 1998, IEEE Trans. Inf. Theory.
[8] Ronitt Rubinfeld,et al. Testing that distributions are close , 2000, Proceedings 41st Annual Symposium on Foundations of Computer Science.
[9] Peter L. Bartlett,et al. Rademacher and Gaussian Complexities: Risk Bounds and Structural Results , 2003, J. Mach. Learn. Res..
[10] Peter L. Bartlett,et al. Model Selection and Error Estimation , 2000, Machine Learning.
[11] Eli Upfal,et al. Probability and Computing: Randomized Algorithms and Probabilistic Analysis , 2005 .
[12] Max Buot. Probability and Computing: Randomized Algorithms and Probabilistic Analysis , 2006 .
[13] Santosh S. Vempala,et al. The geometry of logconcave functions and sampling algorithms , 2007, Random Struct. Algorithms.
[14] Omar Rivasplata,et al. Subgaussian random variables : An expository note , 2012 .
[15] Ronitt Rubinfeld,et al. Testing Closeness of Discrete Distributions , 2010, JACM.
[16] Ryota Tomioka,et al. Norm-Based Capacity Control in Neural Networks , 2015, COLT.
[17] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[18] Matus Telgarsky,et al. Spectrally-normalized margin bounds for neural networks , 2017, NIPS.
[19] Ohad Shamir,et al. Size-Independent Sample Complexity of Neural Networks , 2017, COLT.
[20] Tengyuan Liang,et al. Just Interpolate: Kernel "Ridgeless" Regression Can Generalize , 2018, The Annals of Statistics.
[21] Yuanzhi Li,et al. Learning Overparameterized Neural Networks via Stochastic Gradient Descent on Structured Data , 2018, NeurIPS.
[22] Ruosong Wang,et al. Fine-Grained Analysis of Optimization and Generalization for Overparameterized Two-Layer Neural Networks , 2019, ICML.
[23] J. Zico Kolter,et al. Uniform convergence may be unable to explain generalization in deep learning , 2019, NeurIPS.
[24] Yuan Cao,et al. Generalization Bounds of Stochastic Gradient Descent for Wide and Deep Neural Networks , 2019, NeurIPS.
[25] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[26] Mikhail Belkin,et al. Does data interpolation contradict statistical optimality? , 2018, AISTATS.
[27] Mikhail Belkin,et al. Reconciling modern machine-learning practice and the classical bias–variance trade-off , 2018, Proceedings of the National Academy of Sciences.
[28] Vitaly Feldman,et al. Does learning require memorization? a short tale about a long tail , 2019, STOC.
[29] Daniel M. Roy,et al. In Defense of Uniform Convergence: Generalization via derandomization with an application to interpolating predictors , 2019, ICML.
[30] Weijie J. Su,et al. Benign Overfitting and Noisy Features , 2020, arXiv.org.
[31] Michael W. Mahoney,et al. Exact expressions for double descent and implicit regularization via surrogate random design , 2019, NeurIPS.
[32] Philip M. Long,et al. Generalization bounds for deep convolutional neural networks , 2019, ICLR.
[33] A. Tsigler,et al. Benign overfitting in ridge regression , 2020 .
[34] Philip M. Long,et al. Benign overfitting in linear regression , 2019, Proceedings of the National Academy of Sciences.
[35] O. Papaspiliopoulos. High-Dimensional Probability: An Introduction with Applications in Data Science , 2020 .