What Neural Networks Memorize and Why: Discovering the Long Tail via Influence Estimation
暂无分享,去创建一个
[1] Joachim Denzler,et al. Do We Train on Test Data? Purging CIFAR of Near-Duplicates , 2019, J. Imaging.
[2] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[3] Pietro Perona,et al. The Devil is in the Tails: Fine-grained Classification in the Wild , 2017, ArXiv.
[4] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[5] Philip M. Long,et al. Benign overfitting in linear regression , 2019, Proceedings of the National Academy of Sciences.
[6] Dragomir Anguelov,et al. Capturing Long-Tail Distributions of Object Subcategories , 2014, 2014 IEEE Conference on Computer Vision and Pattern Recognition.
[7] S. Weisberg,et al. Residuals and Influence in Regression , 1982 .
[8] Dumitru Erhan,et al. Going deeper with convolutions , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] David Mease,et al. Explaining the Success of AdaBoost and Random Forests as Interpolating Classifiers , 2015, J. Mach. Learn. Res..
[10] Bernhard Schölkopf,et al. Data scarcity, robustness and extreme multi-label classification , 2019, Machine Learning.
[11] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[12] Mikhail Belkin,et al. Overfitting or perfect fitting? Risk bounds for classification and regression rules that interpolate , 2018, NeurIPS.
[13] Frederick Liu,et al. Estimating Training Data Influence by Tracking Gradient Descent , 2020, NeurIPS.
[14] Vitaly Feldman,et al. Does learning require memorization? a short tale about a long tail , 2019, STOC.
[15] Mikhail Belkin,et al. Does data interpolation contradict statistical optimality? , 2018, AISTATS.
[16] Ziheng Jiang,et al. Characterizing Structural Regularities of Labeled Data in Overparameterized Models , 2020 .
[17] Nathan Srebro,et al. Exploring Generalization in Deep Learning , 2017, NIPS.
[18] Vitaly Shmatikov,et al. Differential Privacy Has Disparate Impact on Model Accuracy , 2019, NeurIPS.
[19] Pradeep Ravikumar,et al. Representer Point Selection for Explaining Deep Neural Networks , 2018, NeurIPS.
[20] Mikhail Belkin,et al. To understand deep learning we need to understand kernel learning , 2018, ICML.
[21] Mikhail Belkin,et al. Two models of double descent for weak features , 2019, SIAM J. Math. Data Sci..
[22] Percy Liang,et al. Understanding Black-box Predictions via Influence Functions , 2017, ICML.
[23] J. Zico Kolter,et al. Uniform convergence may be unable to explain generalization in deep learning , 2019, NeurIPS.
[24] Anant Sahai,et al. Harmless interpolation of noisy data in regression , 2019, 2019 IEEE International Symposium on Information Theory (ISIT).
[25] Andrea Montanari,et al. Surprises in High-Dimensional Ridgeless Least Squares Interpolation , 2019, Annals of statistics.
[26] Vitaly Shmatikov,et al. Membership Inference Attacks Against Machine Learning Models , 2016, 2017 IEEE Symposium on Security and Privacy (SP).
[27] Hossein Mobahi,et al. Fantastic Generalization Measures and Where to Find Them , 2019, ICLR.
[28] Tengyuan Liang,et al. Just Interpolate: Kernel "Ridgeless" Regression Can Generalize , 2018, The Annals of Statistics.
[29] Qiyi Tang,et al. Additional material , 2022, The Works of Graham Greene, Volume 3.
[30] Alexander Rakhlin,et al. Consistency of Interpolation with Laplace Kernels is a High-Dimensional Phenomenon , 2018, COLT.
[31] Úlfar Erlingsson,et al. Distribution Density, Tails, and Outliers in Machine Learning: Metrics and Applications , 2019, ArXiv.
[32] Kilian Q. Weinberger,et al. Densely Connected Convolutional Networks , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[33] Yoshua Bengio,et al. An Empirical Study of Example Forgetting during Deep Neural Network Learning , 2018, ICLR.