Learning Noise Transition Matrix from Only Noisy Labels via Total Variation Regularization
暂无分享,去创建一个
[1] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[2] Clayton Scott,et al. A Rate of Convergence for Mixture Proportion Estimation, with Application to Learning from Noisy Labels , 2015, AISTATS.
[3] Fengmao Lv,et al. Can Cross Entropy Loss Be Robust to Label Noise? , 2020, IJCAI.
[4] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[5] Marius Kloft,et al. Efficient and Accurate ` p-Norm Multiple Kernel Learning , 2009 .
[6] Masashi Sugiyama,et al. On Symmetric Losses for Learning from Corrupted Labels , 2019, ICML.
[7] Shai Shalev-Shwartz,et al. Decoupling "when to update" from "how to update" , 2017, NIPS.
[8] Junnan Li,et al. DivideMix: Learning with Noisy Labels as Semi-supervised Learning , 2020, ICLR.
[9] D. Angluin,et al. Learning From Noisy Examples , 1988, Machine Learning.
[10] Yang Liu,et al. Peer Loss Functions: Learning from Noisy Labels without Knowing Noise Rates , 2019, ICML.
[11] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[12] Natalia Gimelshein,et al. PyTorch: An Imperative Style, High-Performance Deep Learning Library , 2019, NeurIPS.
[13] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[14] James Bailey,et al. Normalized Loss Functions for Deep Learning with Noisy Labels , 2020, ICML.
[15] Baharan Mirzasoleiman,et al. Coresets for Robust Training of Deep Neural Networks against Noisy Labels , 2020, NeurIPS.
[16] Gang Niu,et al. Provably End-to-end Label-Noise Learning without Anchor Points , 2021, ICML.
[17] Ambuj Tewari,et al. On the Consistency of Multiclass Classification Methods , 2007, J. Mach. Learn. Res..
[18] Guigang Zhang,et al. Deep Learning , 2016, Int. J. Semantic Comput..
[19] Xiaogang Wang,et al. Learning from massive noisy labeled data for image classification , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[20] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[21] Ambuj Tewari,et al. Mixture Proportion Estimation via Kernel Embeddings of Distributions , 2016, ICML.
[22] Kotagiri Ramamohanarao,et al. Learning with Bounded Instance- and Label-dependent Label Noise , 2017, ICML.
[23] Byron Boots,et al. Intra Order-preserving Functions for Calibration of Multi-Class Neural Networks , 2020, NeurIPS.
[24] Aritra Ghosh,et al. Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.
[25] P. Moral,et al. On contraction properties of Markov kernels , 2003 .
[26] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[27] Michael I. Jordan,et al. Convexity, Classification, and Risk Bounds , 2006 .
[28] Dimitris N. Metaxas,et al. A Topological Filter for Learning with Label Noise , 2020, NeurIPS.
[29] Gang Niu,et al. On the Minimal Supervision for Training Any Binary Classifier from Only Unlabeled Data , 2018, ICLR.
[30] Cheng Soon Ong,et al. Learning from Corrupted Binary Labels via Class-Probability Estimation , 2015, ICML.
[31] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[32] Kilian Q. Weinberger,et al. On Calibration of Modern Neural Networks , 2017, ICML.
[33] Dacheng Tao,et al. Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[34] Thomas Brox,et al. SELF: Learning to Filter Noisy Labels with Self-Ensembling , 2019, ICLR.
[35] Xingrui Yu,et al. How does Disagreement Help Generalization against Label Corruption? , 2019, ICML.
[36] Gilles Blanchard,et al. Classification with Asymmetric Label Noise: Consistency and Maximal Denoising , 2013, COLT.
[37] Gang Niu,et al. Are Anchor Points Really Indispensable in Label-Noise Learning? , 2019, NeurIPS.
[38] David Sontag,et al. Consistent Estimators for Learning to Defer to an Expert , 2020, ICML.
[39] Nagarajan Natarajan,et al. Learning from binary labels with instance-dependent noise , 2018, Machine Learning.
[40] Masashi Sugiyama,et al. Classification with Rejection Based on Cost-sensitive Classification , 2020, ICML.
[41] Le Song,et al. Iterative Learning with Open-set Noisy Labels , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[42] K. Schittkowski,et al. NONLINEAR PROGRAMMING , 2022 .
[43] Pheng-Ann Heng,et al. Robustness of Accuracy Metric and its Inspirations in Learning with Noisy Labels , 2020, AAAI.
[44] Jeff A. Bilmes,et al. Combating Label Noise in Deep Learning Using Abstention , 2019, ICML.
[45] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[46] Jacob Goldberger,et al. Training deep neural-networks using a noise adaptation layer , 2016, ICLR.
[47] Peter A. Flach,et al. Beyond temperature scaling: Obtaining well-calibrated multiclass probabilities with Dirichlet calibration , 2019, NeurIPS.
[48] Pietro Perona,et al. Learning object categories from Google's image search , 2005, Tenth IEEE International Conference on Computer Vision (ICCV'05) Volume 1.
[49] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[50] P. Diaconis,et al. Conjugate Priors for Exponential Families , 1979 .
[51] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[52] Matthias Hein,et al. Why ReLU Networks Yield High-Confidence Predictions Far Away From the Training Data and How to Mitigate the Problem , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[53] Bo An,et al. Combating Noisy Labels by Agreement: A Joint Training Method with Co-Regularization , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[54] Christopher Burgess,et al. beta-VAE: Learning Basic Visual Concepts with a Constrained Variational Framework , 2016, ICLR 2016.
[55] Gang Niu,et al. Learning from Complementary Labels , 2017, NIPS.
[56] Gang Niu,et al. Dual T: Reducing Estimation Error for Transition Matrix in Label-noise Learning , 2020, NeurIPS.
[57] Gang Niu,et al. Parts-dependent Label Noise: Towards Instance-dependent Label Noise , 2020, ArXiv.
[58] Jimmy Ba,et al. Adam: A Method for Stochastic Optimization , 2014, ICLR.
[59] Simon Haykin,et al. GradientBased Learning Applied to Document Recognition , 2001 .
[60] Ivor W. Tsang,et al. Masking: A New Perspective of Noisy Supervision , 2018, NeurIPS.
[61] Rocco A. Servedio,et al. Random classification noise defeats all convex potential boosters , 2008, ICML '08.
[62] Dacheng Tao,et al. Learning with Biased Complementary Labels , 2017, ECCV.
[63] Gilles Blanchard,et al. Decontamination of Mutually Contaminated Models , 2014, AISTATS.
[64] Gang Niu,et al. Confidence Scores Make Instance-dependent Label-noise Learning Possible , 2019, ICML.
[65] Geoffrey E. Hinton,et al. On the importance of initialization and momentum in deep learning , 2013, ICML.
[66] Mohan S. Kankanhalli,et al. Learning to Learn From Noisy Labeled Data , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[67] Qi Xie,et al. Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting , 2019, NeurIPS.
[68] Yoshua Bengio,et al. Gradient-based learning applied to document recognition , 1998, Proc. IEEE.
[69] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[70] Xingrui Yu,et al. Co-teaching: Robust training of deep neural networks with extremely noisy labels , 2018, NeurIPS.
[71] Gang Niu,et al. Analysis of Learning from Positive and Unlabeled Data , 2014, NIPS.
[72] Yueming Lyu,et al. Curriculum Loss: Robust Learning and Generalization against Label Corruption , 2019, ICLR.
[73] Weilong Yang,et al. Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels , 2019, ICML.