Wasserstein Adversarial Regularization for Learning With Label Noise
暂无分享,去创建一个
[1] Nicolas Courty,et al. Unbalanced minibatch Optimal Transport; applications to Domain Adaptation , 2021, ICML.
[2] Lei Feng,et al. Combating Noisy Labels by Agreement: A Joint Training Method with Co-Regularization , 2020, 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[3] Tom Goldstein,et al. Label Smoothing and Logit Squeezing: A Replacement for Adversarial Training? , 2019, ArXiv.
[4] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[5] Morgane Goibert,et al. Adversarial Robustness via Adversarial Label-Smoothing , 2019, ArXiv.
[6] Jae-Gil Lee,et al. SELFIE: Refurbishing Unclean Samples for Robust Deep Learning , 2019, ICML.
[7] Pengfei Chen,et al. Understanding and Utilizing Deep Neural Networks Trained with Noisy Labels , 2019, ICML.
[8] Kun Yi,et al. Probabilistic End-To-End Noise Correction for Learning With Noisy Labels , 2019, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Xingrui Yu,et al. How does Disagreement Help Generalization against Label Corruption? , 2019, ICML.
[10] N. Courty,et al. An Entropic Optimal Transport Loss for Learning Deep Neural Networks under Label Noise in Remote Sensing Images , 2018, Comput. Vis. Image Underst..
[11] Fei Wang,et al. The Devil of Face Recognition is in the Noise , 2018, ECCV.
[12] James Bailey,et al. Dimensionality-Driven Learning with Noisy Labels , 2018, ICML.
[13] Alessandro Rudi,et al. Differential Properties of Sinkhorn Approximation for Learning with Wasserstein Distance , 2018, NeurIPS.
[14] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[15] Masashi Sugiyama,et al. Co-teaching: Robust training of deep neural networks with extremely noisy labels , 2018, NeurIPS.
[16] Le Song,et al. Iterative Learning with Open-set Noisy Labels , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[17] Kiyoharu Aizawa,et al. Joint Optimization Framework for Learning with Noisy Labels , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[18] Nicolas Courty,et al. DeepJDOT: Deep Joint distribution optimal transport for unsupervised domain adaptation , 2018, ECCV.
[19] Bin Yang,et al. Learning to Reweight Examples for Robust Deep Learning , 2018, ICML.
[20] Gabriel Peyré,et al. Computational Optimal Transport , 2018, Found. Trends Mach. Learn..
[21] Kevin Gimpel,et al. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise , 2018, NeurIPS.
[22] Ruilin Li,et al. Learning to Match via Inverse Optimal Transport , 2018, J. Mach. Learn. Res..
[23] Yongyang Xu,et al. Building Extraction in Very High Resolution Remote Sensing Imagery Using Deep Learning and Guided Filters , 2018, Remote. Sens..
[24] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[25] Lei Zhang,et al. CleanNet: Transfer Learning for Scalable Image Classifier Training with Label Noise , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[26] Yu Cheng,et al. Sobolev GAN , 2017, ICLR.
[27] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[28] Roland Vollgraf,et al. Fashion-MNIST: a Novel Image Dataset for Benchmarking Machine Learning Algorithms , 2017, ArXiv.
[29] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[30] Mark Sandler,et al. The Effects of Noisy Labels on Deep Convolutional Neural Networks for Music Tagging , 2017, IEEE Transactions on Emerging Topics in Computational Intelligence.
[31] Gabriel Peyré,et al. Learning Generative Models with Sinkhorn Divergences , 2017, AISTATS.
[32] Arash Vahdat,et al. Toward Robustness against Label Noise in Training Deep Discriminative Neural Networks , 2017, NIPS.
[33] Jason Altschuler,et al. Near-linear time approximation algorithms for optimal transport via Sinkhorn iteration , 2017, NIPS.
[34] Ramesh Raskar,et al. Pairwise Confusion for Fine-Grained Visual Classification , 2017, ECCV.
[35] Shin Ishii,et al. Virtual Adversarial Training: A Regularization Method for Supervised and Semi-Supervised Learning , 2017, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[36] Yale Song,et al. Learning from Noisy Labels with Distillation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[37] Aritra Ghosh,et al. Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.
[38] Matt J. Kusner,et al. Supervised Word Mover's Distance , 2016, NIPS.
[39] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[40] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[41] Koray Kavukcuoglu,et al. Pixel Recurrent Neural Networks , 2016, ICML.
[42] Ross B. Girshick,et al. Seeing through the Human Reporting Bias: Visual Classifiers from Noisy Human-Centric Labels , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[43] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[44] Jonathan Krause,et al. The Unreasonable Effectiveness of Noisy Data for Fine-Grained Recognition , 2015, ECCV.
[45] Cheng Soon Ong,et al. Learning from Corrupted Binary Labels via Class-Probability Estimation , 2015, ICML.
[46] Hossein Mobahi,et al. Learning with a Wasserstein Loss , 2015, NIPS.
[47] Xiaogang Wang,et al. Learning from massive noisy labeled data for image classification , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[48] Michele Volpi,et al. Semantic segmentation of urban scenes by learning local class interactions , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition Workshops (CVPRW).
[49] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[50] Thomas Brox,et al. U-Net: Convolutional Networks for Biomedical Image Segmentation , 2015, MICCAI.
[51] Dumitru Erhan,et al. Training Deep Neural Networks on Noisy Labels with Bootstrapping , 2014, ICLR.
[52] Jonathon Shlens,et al. Explaining and Harnessing Adversarial Examples , 2014, ICLR.
[53] D. Tao,et al. Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[54] Rob Fergus,et al. Learning from Noisy Labels with Deep Neural Networks , 2014, ICLR.
[55] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[56] Marco Cuturi,et al. Sinkhorn Distances: Lightspeed Computation of Optimal Transport , 2013, NIPS.
[57] Jeffrey Dean,et al. Distributed Representations of Words and Phrases and their Compositionality , 2013, NIPS.
[58] Blaine Nelson,et al. Support Vector Machines Under Adversarial Label Noise , 2011, ACML.
[59] David Avis,et al. Ground metric learning , 2011, J. Mach. Learn. Res..
[60] Gaël Varoquaux,et al. Scikit-learn: Machine Learning in Python , 2011, J. Mach. Learn. Res..
[61] Ying Wu,et al. Context-aware clustering , 2008, 2008 IEEE Conference on Computer Vision and Pattern Recognition.
[62] Zhenguo Li,et al. Noise Robust Spectral Clustering , 2007, 2007 IEEE 11th International Conference on Computer Vision.
[63] Antonio Criminisi,et al. Harvesting Image Databases from the Web , 2007, 2007 IEEE 11th International Conference on Computer Vision.
[64] Nello Cristianini,et al. Spectral Kernel Methods for Clustering , 2001, NIPS.
[65] Carla E. Brodley,et al. Identifying Mislabeled Training Data , 1999, J. Artif. Intell. Res..
[66] Uri Shaham,et al. Understanding Adversarial Training: Increasing Local Stability of Neural Nets through Robust Optimization , 2015, ArXiv.
[67] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .