Per-Example Gradient Regularization Improves Learning Signals from Noisy Data
暂无分享,去创建一个
[1] Quanquan Gu,et al. Benign Overfitting for Two-layer ReLU Networks , 2023, ArXiv.
[2] Zhiyuan Li,et al. How Does Sharpness-Aware Minimization Minimize Sharpness? , 2022, ArXiv.
[3] D. Barrett,et al. Why neural networks find simple solutions: the many regularizers of geometric complexity , 2022, NeurIPS.
[4] Nicolas Flammarion,et al. Towards Understanding Sharpness-Aware Minimization , 2022, ICML.
[5] Mikhail Belkin,et al. Benign Overfitting in Two-layer Convolutional Neural Networks , 2022, NeurIPS.
[6] Niladri S. Chatterji,et al. Benign Overfitting without Linearity: Neural Network Classifiers Trained by Gradient Descent for Noisy Linear Data , 2022, COLT.
[7] Yang Zhao,et al. Penalizing Gradient Norm for Efficiently Improving Generalization in Deep Learning , 2022, ICML.
[8] Jianfeng Yao,et al. Impact of classification difficulty on the weight matrices spectra in Deep Learning and application to early-stopping , 2021, 2111.13331.
[9] Mikhail Belkin,et al. Risk Bounds for Over-parameterized Maximum Margin Classification on Sub-Gaussian Mixtures , 2021, NeurIPS.
[10] Ariel Kleiner,et al. Sharpness-Aware Minimization for Efficiently Improving Generalization , 2020, ICLR.
[11] D. Barrett,et al. Implicit Gradient Regularization , 2020, ICLR.
[12] Philip M. Long,et al. Finite-sample analysis of interpolating linear classifiers in the overparameterized regime , 2020, J. Mach. Learn. Res..
[13] Andrea Montanari,et al. The Generalization Error of Random Features Regression: Precise Asymptotics and the Double Descent Curve , 2019, Communications on Pure and Applied Mathematics.
[14] Michael W. Mahoney,et al. Implicit Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory and Implications for Learning , 2018, J. Mach. Learn. Res..
[15] Yuanzhi Li,et al. Towards Understanding Ensemble, Knowledge Distillation and Self-Distillation in Deep Learning , 2020, ICLR.
[16] Philip M. Long,et al. Benign overfitting in linear regression , 2019, Proceedings of the National Academy of Sciences.
[17] Guy Blanc,et al. Implicit regularization for deep neural networks driven by an Ornstein-Uhlenbeck like process , 2019, COLT.
[18] Ekaba Bisong,et al. Regularization for Deep Learning , 2019, Building Machine Learning and Deep Learning Models on Google Cloud Platform.
[19] Yoshua Bengio,et al. Three Factors Influencing Minima in SGD , 2017, ArXiv.