Generation and Analysis of Feature-Dependent Pseudo Noise for Training Deep Neural Networks
暂无分享,去创建一个
Venkat Devarajan | Ganesh Sankaranarayanan | Babak Namazi | Sree Ram Kamabattula | Kumudha Musini | G. Sankaranarayanan | V. Devarajan | Babak Namazi | Kumudha Musini
[1] James Bailey,et al. Dimensionality-Driven Learning with Noisy Labels , 2018, ICML.
[2] Ilkay Ulusoy,et al. Image Classification with Deep Learning in the Presence of Noisy Labels: A Survey , 2019, ArXiv.
[3] Kiyoharu Aizawa,et al. Joint Optimization Framework for Learning with Noisy Labels , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[4] Nir Shavit,et al. Deep Learning is Robust to Massive Label Noise , 2017, ArXiv.
[5] Simon K. Warfield,et al. Deep learning with noisy labels: exploring techniques and remedies in medical image analysis , 2020, Medical Image Anal..
[6] Yang Liu,et al. Learning with Instance-Dependent Label Noise: A Sample Sieve Approach , 2021, ICLR.
[7] Jacob Goldberger,et al. Training deep neural-networks using a noise adaptation layer , 2016, ICLR.
[8] Weilong Yang,et al. Beyond Synthetic Noise: Deep Learning on Controlled Noisy Labels , 2019, ICML.
[9] Li Fei-Fei,et al. ImageNet: A large-scale hierarchical image database , 2009, CVPR.
[10] Jae-Gil Lee,et al. SELFIE: Refurbishing Unclean Samples for Robust Deep Learning , 2019, ICML.
[11] Elyor Kodirov,et al. IMAE for Noise-Robust Learning: Mean Absolute Error Does Not Treat Examples Equally and Gradient Magnitude's Variance Matters , 2019 .
[12] Dumitru Erhan,et al. Training Deep Neural Networks on Noisy Labels with Bootstrapping , 2014, ICLR.
[13] Anima Anandkumar,et al. Learning From Noisy Singly-labeled Data , 2017, ICLR.
[14] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Shai Shalev-Shwartz,et al. Decoupling "when to update" from "how to update" , 2017, NIPS.
[16] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[17] Lei Zhang,et al. CleanNet: Transfer Learning for Scalable Image Classifier Training with Label Noise , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[18] Görkem Algan,et al. Label Noise Types and Their Effects on Deep Learning , 2020, ArXiv.
[19] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[20] Yi Sun,et al. Limited Gradient Descent: Learning With Noisy Labels , 2018, IEEE Access.
[21] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[22] Adam Tauman Kalai,et al. Noise-tolerant learning, the parity problem, and the statistical query model , 2000, STOC '00.
[23] Deliang Fan,et al. A Semi-Supervised Two-Stage Approach to Learning from Noisy Labels , 2018, 2018 IEEE Winter Conference on Applications of Computer Vision (WACV).
[24] Junnan Li,et al. DivideMix: Learning with Noisy Labels as Semi-supervised Learning , 2020, ICLR.
[25] Jae-Gil Lee,et al. Learning from Noisy Labels with Deep Neural Networks: A Survey , 2020, ArXiv.
[26] Sree Ram Kamabattula,et al. Identifying the Training Stop Point with Noisy Labeled Data , 2020, 2020 International Conference on Computational Science and Computational Intelligence (CSCI).
[27] Gang Niu,et al. Parts-dependent Label Noise: Towards Instance-dependent Label Noise , 2020, ArXiv.
[28] Jae-Gil Lee,et al. Prestopping: How Does Early Stopping Help Generalization against Label Noise? , 2019, ArXiv.
[29] Ali Farhadi,et al. You Only Look Once: Unified, Real-Time Object Detection , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[30] Bin Yang,et al. Learning to Reweight Examples for Robust Deep Learning , 2018, ICML.
[31] Xingrui Yu,et al. Co-teaching: Robust training of deep neural networks with extremely noisy labels , 2018, NeurIPS.
[32] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[33] Wei Li,et al. WebVision Database: Visual Learning and Understanding from Web Data , 2017, ArXiv.
[34] Subramanian Ramanathan,et al. Learning from multiple annotators with varying expertise , 2013, Machine Learning.
[35] Andrew McCallum,et al. Active Bias: Training More Accurate Neural Networks by Emphasizing High Variance Samples , 2017, NIPS.
[36] Pengfei Chen,et al. Understanding and Utilizing Deep Neural Networks Trained with Noisy Labels , 2019, ICML.
[37] Xiaogang Wang,et al. Learning from massive noisy labeled data for image classification , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[38] Yoshua Bengio,et al. A Closer Look at Memorization in Deep Networks , 2017, ICML.
[39] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[40] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[41] Kilian Q. Weinberger,et al. Identifying Mislabeled Data using the Area Under the Margin Ranking , 2020, NeurIPS.
[42] Takuhiro Kaneko,et al. Label-Noise Robust Generative Adversarial Networks , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).