暂无分享,去创建一个
[1] Matthew S. Nokleby,et al. Learning Deep Networks from Noisy Labels with Dropout Regularization , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).
[2] E. Mammen,et al. Smooth Discrimination Analysis , 1999 .
[3] J. L. Hodges,et al. Discriminatory Analysis - Nonparametric Discrimination: Consistency Properties , 1989 .
[4] A. Tsybakov,et al. Fast learning rates for plug-in classifiers , 2007, 0708.2321.
[5] Sanjoy Dasgupta,et al. Rates of Convergence for Nearest Neighbor Classification , 2014, NIPS.
[6] Bin Yang,et al. Learning to Reweight Examples for Robust Deep Learning , 2018, ICML.
[7] Joan Bruna,et al. Training Convolutional Networks with Noisy Labels , 2014, ICLR 2014.
[8] Xiaogang Wang,et al. Learning from massive noisy labeled data for image classification , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] J. Paul Brooks,et al. Support Vector Machines with the Ramp Loss and the Hard Margin Loss , 2011, Oper. Res..
[10] Ata Kabán,et al. Fast Rates for a kNN Classifier Robust to Unknown Asymmetric Label Noise , 2019, ICML.
[11] Manfred K. Warmuth,et al. Robust Bi-Tempered Logistic Loss Based on Bregman Divergences , 2019, NeurIPS.
[12] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[13] Nuno Vasconcelos,et al. On the Design of Loss Functions for Classification: theory, robustness to outliers, and SavageBoost , 2008, NIPS.
[14] Anima Anandkumar,et al. Learning From Noisy Singly-labeled Data , 2017, ICLR.
[15] Kevin Gimpel,et al. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise , 2018, NeurIPS.
[16] Dacheng Tao,et al. Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[17] Somesh Jha,et al. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples , 2017, ICML.
[18] Kibok Lee,et al. Robust Inference via Generative Classifiers for Handling Noisy Labels , 2019, ICML.
[19] Carla E. Brodley,et al. Identifying Mislabeled Training Data , 1999, J. Artif. Intell. Res..
[20] Aritra Ghosh,et al. Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.
[21] Li Fei-Fei,et al. MentorNet: Regularizing Very Deep Neural Networks on Corrupted Labels , 2017, ArXiv.
[22] Yale Song,et al. Learning from Noisy Labels with Distillation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[23] Jeff A. Bilmes,et al. Combating Label Noise in Deep Learning Using Abstention , 2019, ICML.
[24] Abhinav Gupta,et al. Learning from Noisy Large-Scale Datasets with Minimal Supervision , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[25] Arash Vahdat,et al. Toward Robustness against Label Noise in Training Deep Discriminative Neural Networks , 2017, NIPS.
[26] Robert D. Nowak,et al. Adaptive Hausdorff Estimation of Density Level Sets , 2009, COLT.
[27] A. Tsybakov,et al. Optimal aggregation of classifiers in statistical learning , 2003 .
[28] G. Lugosi,et al. On the Strong Universal Consistency of Nearest Neighbor Regression Function Estimates , 1994 .
[29] Nir Shavit,et al. Deep Learning is Robust to Massive Label Noise , 2017, ArXiv.
[30] Patrick D. McDaniel,et al. Deep k-Nearest Neighbors: Towards Confident, Interpretable and Robust Deep Learning , 2018, ArXiv.
[31] Heinrich Jiang,et al. Non-Asymptotic Uniform Rates of Consistency for k-NN Regression , 2017, AAAI.
[32] Ihab F. Ilyas,et al. Data Cleaning: Overview and Emerging Challenges , 2016, SIGMOD Conference.
[33] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[34] Dennis L. Wilson,et al. Asymptotic Properties of Nearest Neighbor Rules Using Edited Data , 1972, IEEE Trans. Syst. Man Cybern..
[35] C. J. Stone,et al. Consistent Nonparametric Regression , 1977 .
[36] Zhi-Hua Zhou,et al. On the Resistance of Nearest Neighbor to Random Noisy Labels , 2016 .
[37] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[38] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[39] Maya R. Gupta,et al. Completely Lazy Learning , 2010, IEEE Transactions on Knowledge and Data Engineering.
[40] Isabelle Guyon,et al. Discovering Informative Patterns and Data Cleaning , 1996, Advances in Knowledge Discovery and Data Mining.
[41] Ivor W. Tsang,et al. Masking: A New Perspective of Noisy Supervision , 2018, NeurIPS.
[42] Jacob Goldberger,et al. Training deep neural-networks using a noise adaptation layer , 2016, ICLR.
[43] Maya R. Gupta,et al. To Trust Or Not To Trust A Classifier , 2018, NeurIPS.
[44] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[45] Zhi-Hua Zhou,et al. On the Consistency of Exact and Approximate Nearest Neighbor with Noisy Data , 2016, ArXiv.
[46] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.