暂无分享,去创建一个
[1] Gilles Blanchard,et al. Decontamination of Mutually Contaminated Models , 2014, AISTATS.
[2] Yann LeCun,et al. The mnist database of handwritten digits , 2005 .
[3] Muhammad J. Amjad,et al. Robust Synthetic Control , 2017, J. Mach. Learn. Res..
[4] Somesh Jha,et al. Analyzing the Robustness of Nearest Neighbors to Adversarial Examples , 2017, ICML.
[5] Xiaogang Wang,et al. Deep Self-Learning From Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[6] Avrim Blum,et al. The Bottleneck , 2021, Monopsony Capitalism.
[7] David Sontag,et al. Electronic medical record phenotyping using the anchor and learn framework , 2016, J. Am. Medical Informatics Assoc..
[8] Praveen K. Paritosh,et al. “Everyone wants to do the model work, not the data work”: Data Cascades in High-Stakes AI , 2021, CHI.
[9] Kevin Gimpel,et al. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks , 2016, ICLR.
[10] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[11] Charles Elkan,et al. Learning classifiers from only positive and unlabeled data , 2008, KDD.
[12] Gilles Blanchard,et al. Classification with Asymmetric Label Noise: Consistency and Maximal Denoising , 2013, COLT.
[13] A. P. Dawid,et al. Maximum Likelihood Estimation of Observer Error‐Rates Using the EM Algorithm , 1979 .
[14] Tailin Wu,et al. Learning with Confident Examples: Rank Pruning for Robust Classification with Noisy Labels , 2017, UAI.
[15] Yanyao Shen,et al. Learning with Bad Training Data via Iterative Trimmed Loss Minimization , 2018, ICML.
[16] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[17] Jacob Goldberger,et al. Training deep neural-networks using a noise adaptation layer , 2016, ICLR.
[18] Seong Joon Oh,et al. Re-labeling ImageNet: from Single to Multi-Labels, from Global to Localized Labels , 2021, 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[19] Tomas Mikolov,et al. Bag of Tricks for Efficient Text Classification , 2016, EACL.
[20] Nagarajan Natarajan,et al. Cost-Sensitive Learning with Noisy Labels , 2017, J. Mach. Learn. Res..
[21] Marius Kloft,et al. Toward Supervised Anomaly Detection , 2014, J. Artif. Intell. Res..
[22] Junnan Li,et al. DivideMix: Learning with Noisy Labels as Semi-supervised Learning , 2020, ICLR.
[23] D. Angluin,et al. Learning From Noisy Examples , 1988, Machine Learning.
[24] Michael S. Bernstein,et al. ImageNet Large Scale Visual Recognition Challenge , 2014, International Journal of Computer Vision.
[25] Francisco Herrera,et al. Analyzing the presence of noise in multi-class problems: alleviating its influence with the One-vs-One decomposition , 2012, Knowledge and Information Systems.
[26] Binqiang Zhao,et al. O2U-Net: A Simple Noisy Label Detection Approach for Deep Neural Networks , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[27] Paul Dupuis,et al. Distinguishing and integrating aleatoric and epistemic variation in uncertainty quantification , 2011, 1103.1861.
[28] Pietro Perona,et al. Microsoft COCO: Common Objects in Context , 2014, ECCV.
[29] Noel E. O'Connor,et al. Unsupervised label noise modeling and loss correction , 2019, ICML.
[30] Yale Song,et al. Learning from Noisy Labels with Distillation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[31] Ata Kabán,et al. Multi-class classification in the presence of labelling errors , 2011, ESANN.
[32] Hongyi Zhang,et al. mixup: Beyond Empirical Risk Minimization , 2017, ICLR.
[33] Colin Wei,et al. Regularization Matters: Generalization and Optimization of Neural Nets v.s. their Induced Kernel , 2018, NeurIPS.
[34] Shengxin Zha,et al. EgoCom: A Multi-Person Multi-Modal Egocentric Communications Dataset , 2020, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[35] Arash Vahdat,et al. Toward Robustness against Label Noise in Training Deep Discriminative Neural Networks , 2017, NIPS.
[36] Gilles Blanchard,et al. Decontamination of Mutual Contamination Models , 2017, J. Mach. Learn. Res..
[37] Anima Anandkumar,et al. Learning From Noisy Singly-labeled Data , 2017, ICLR.
[38] Thore Graepel,et al. The Kernel Gibbs Sampler , 2000, NIPS.
[39] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[40] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[41] Clayton Scott,et al. A Rate of Convergence for Mixture Proportion Estimation, with Application to Learning from Noisy Labels , 2015, AISTATS.
[42] Antanas Verikas,et al. Agreeing to disagree: active learning with noisy labels without crowdsourcing , 2017, International Journal of Machine Learning and Cybernetics.
[43] Yizhou Wang,et al. L_DMI: A Novel Information-theoretic Loss Function for Training Deep Nets Robust to Label Noise , 2019, NeurIPS.
[44] Trevor Darrell,et al. Detector discovery in the wild: Joint multiple instance and representation learning , 2014, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[45] Carla E. Brodley,et al. Identifying Mislabeled Training Data , 1999, J. Artif. Intell. Res..
[46] Joan Bruna,et al. Training Convolutional Networks with Noisy Labels , 2014, ICLR 2014.
[47] Weilong Yang,et al. Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels , 2019, ICML.
[48] Julian J. McAuley,et al. Ups and Downs: Modeling the Visual Evolution of Fashion Trends with One-Class Collaborative Filtering , 2016, WWW.
[49] Dumitru Erhan,et al. Training Deep Neural Networks on Noisy Labels with Bootstrapping , 2014, ICLR.
[50] Qiang Liu,et al. On the Margin Theory of Feedforward Neural Networks , 2018, ArXiv.
[51] Swami Sankaranarayanan,et al. Learning From Noisy Labels by Regularized Estimation of Annotator Confusion , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[52] Bernhard Schölkopf,et al. Estimating a Kernel Fisher Discriminant in the Presence of Label Noise , 2001, ICML.
[53] Matthew S. Nokleby,et al. Learning Deep Networks from Noisy Labels with Dropout Regularization , 2016, 2016 IEEE 16th International Conference on Data Mining (ICDM).
[54] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[55] Wei Li,et al. WebVision Database: Visual Learning and Understanding from Web Data , 2017, ArXiv.
[56] Charles Elkan,et al. The Foundations of Cost-Sensitive Learning , 2001, IJCAI.
[57] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[58] Kevin Gimpel,et al. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise , 2018, NeurIPS.
[59] Frank Nielsen,et al. Loss factorization, weakly supervised learning and label noise robustness , 2016, ICML.
[60] Christopher De Sa,et al. Data Programming: Creating Large Training Sets, Quickly , 2016, NIPS.
[61] Richard Lippmann,et al. Neural Network Classifiers Estimate Bayesian a posteriori Probabilities , 1991, Neural Computation.
[62] Joaquin Vanschoren,et al. OpenML-Python: an extensible Python API for OpenML , 2019, ArXiv.
[63] Isaac L. Chuang,et al. Detecting and preventing "multiple-account" cheating in massive open online courses , 2015, Comput. Educ..
[64] Pengfei Chen,et al. Understanding and Utilizing Deep Neural Networks Trained with Noisy Labels , 2019, ICML.
[65] Qi Xie,et al. Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting , 2019, NeurIPS.
[66] Xingrui Yu,et al. Co-teaching: Robust training of deep neural networks with extremely noisy labels , 2018, NeurIPS.
[67] Xindong Wu,et al. Improving Crowdsourced Label Quality Using Noise Correction , 2018, IEEE Transactions on Neural Networks and Learning Systems.
[68] George Forman,et al. Counting Positives Accurately Despite Inaccurate Classification , 2005, ECML.
[69] Jonas Mueller,et al. Pervasive Label Errors in Test Sets Destabilize Machine Learning Benchmarks , 2021, NeurIPS Datasets and Benchmarks.
[70] Alexander J. Smola,et al. Detecting and Correcting for Label Shift with Black Box Predictors , 2018, ICML.
[71] Beata Beigman Klebanov,et al. Learning with Annotation Noise , 2009, ACL.
[72] George Forman,et al. Quantifying counts and costs via classification , 2008, Data Mining and Knowledge Discovery.
[73] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[74] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[75] Tomas Mikolov,et al. Enriching Word Vectors with Subword Information , 2016, TACL.
[76] Dacheng Tao,et al. Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.