暂无分享,去创建一个
Deyu Meng | Zongben Xu | Jun Shu | Keyu Chen | Qian Zhao | Deyu Meng | Zongben Xu | Jun Shu | Qian Zhao | Keyu Chen
[1] Lei Zhang,et al. CleanNet: Transfer Learning for Scalable Image Classifier Training with Label Noise , 2017, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[2] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[3] Naresh Manwani,et al. Noise Tolerance Under Risk Minimization , 2011, IEEE Transactions on Cybernetics.
[4] Deyu Meng,et al. Learning to Detect Concepts from Webly-Labeled Video Data , 2016, IJCAI.
[5] Jacob Goldberger,et al. Training deep neural-networks using a noise adaptation layer , 2016, ICLR.
[6] Geoffrey E. Hinton,et al. ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.
[7] Deyu Meng,et al. A theoretical understanding of self-paced learning , 2017, Inf. Sci..
[8] Richard Nock,et al. Making Deep Neural Networks Robust to Label Noise: A Loss Correction Approach , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[9] Manfred K. Warmuth,et al. Robust Bi-Tempered Logistic Loss Based on Bregman Divergences , 2019, NeurIPS.
[10] Aritra Ghosh,et al. Robust Loss Functions under Label Noise for Deep Neural Networks , 2017, AAAI.
[11] Samy Bengio,et al. Understanding deep learning requires rethinking generalization , 2016, ICLR.
[12] Qi Xie,et al. Self-Paced Learning for Matrix Factorization , 2015, AAAI.
[13] David M. Blei,et al. Robust Probabilistic Modeling with Bayesian Data Reweighting , 2016, ICML.
[14] Xiaogang Wang,et al. Learning from massive noisy labeled data for image classification , 2015, 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[15] Eric P. Xing,et al. AutoLoss: Learning Discrete Schedules for Alternate Optimization , 2018, ICLR 2018.
[16] Bin Yang,et al. Learning to Reweight Examples for Robust Deep Learning , 2018, ICML.
[17] Joan Bruna,et al. Training Convolutional Networks with Noisy Labels , 2014, ICLR 2014.
[18] Nuno Vasconcelos,et al. On the Design of Loss Functions for Classification: theory, robustness to outliers, and SavageBoost , 2008, NIPS.
[19] Brian D. Ziebart,et al. Robust Classification Under Sample Selection Bias , 2014, NIPS.
[20] Carlos Guestrin,et al. Addressing the Loss-Metric Mismatch with Adaptive Loss Alignment , 2019, ICML.
[21] Frank Nielsen,et al. On the Efficient Minimization of Classification Calibrated Surrogates , 2008, NIPS.
[22] Abhinav Gupta,et al. Learning from Noisy Large-Scale Datasets with Minimal Supervision , 2017, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[23] J. Paul Brooks,et al. Support Vector Machines with the Ramp Loss and the Hard Margin Loss , 2011, Oper. Res..
[24] Michael I. Jordan,et al. Convexity, Classification, and Risk Bounds , 2006 .
[25] Jürgen Schmidhuber,et al. Learning to Control Fast-Weight Memories: An Alternative to Dynamic Recurrent Networks , 1992, Neural Computation.
[26] Alex Krizhevsky,et al. Learning Multiple Layers of Features from Tiny Images , 2009 .
[27] Kevin Gimpel,et al. Using Trusted Data to Train Deep Networks on Labels Corrupted by Severe Noise , 2018, NeurIPS.
[28] Jasper Snoek,et al. Practical Bayesian Optimization of Machine Learning Algorithms , 2012, NIPS.
[29] Aritra Ghosh,et al. Making risk minimization tolerant to label noise , 2014, Neurocomputing.
[30] Jian Sun,et al. Deep Residual Learning for Image Recognition , 2015, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[31] Luca Antiga,et al. Automatic differentiation in PyTorch , 2017 .
[32] Yale Song,et al. Learning from Noisy Labels with Distillation , 2017, 2017 IEEE International Conference on Computer Vision (ICCV).
[33] Mert R. Sabuncu,et al. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels , 2018, NeurIPS.
[34] Deyu Meng,et al. Small Sample Learning in Big Data Era , 2018, ArXiv.
[35] Yoshua Bengio,et al. Random Search for Hyper-Parameter Optimization , 2012, J. Mach. Learn. Res..
[36] Sergey Levine,et al. Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.
[37] Lars Schmidt-Thieme,et al. Learning Surrogate Losses , 2019, ArXiv.
[38] Lijun Wu,et al. Learning to Teach with Dynamic Loss Functions , 2018, NeurIPS.
[39] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[40] Dumitru Erhan,et al. Training Deep Neural Networks on Noisy Labels with Bootstrapping , 2014, ICLR.
[41] Deyu Meng,et al. Easy Samples First: Self-paced Reranking for Zero-Example Multimedia Search , 2014, ACM Multimedia.
[42] Paolo Frasconi,et al. Bilevel Programming for Hyperparameter Optimization and Meta-Learning , 2018, ICML.
[43] Geoffrey E. Hinton,et al. Visualizing Data using t-SNE , 2008 .
[44] Ryan P. Adams,et al. Gradient-based Hyperparameter Optimization through Reversible Learning , 2015, ICML.
[45] Zhuowen Tu,et al. Learning to Predict from Crowdsourced Data , 2014, UAI.
[46] Andrew McCallum,et al. Active Bias: Training More Accurate Neural Networks by Emphasizing High Variance Samples , 2017, NIPS.
[47] Qi Xie,et al. Meta-Weight-Net: Learning an Explicit Mapping For Sample Weighting , 2019, NeurIPS.
[48] James Bailey,et al. Symmetric Cross Entropy for Robust Learning With Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[49] Kiyoharu Aizawa,et al. Joint Optimization Framework for Learning with Noisy Labels , 2018, 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition.
[50] Yao Li,et al. Attend in Groups: A Weakly-Supervised Deep Learning Framework for Learning from Web Data , 2016, 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).
[51] Yueming Lyu,et al. Curriculum Loss: Robust Learning and Generalization against Label Corruption , 2019, ICLR.
[52] Maoguo Gong,et al. Decomposition-Based Evolutionary Multiobjective Optimization to Self-Paced Learning , 2019, IEEE Transactions on Evolutionary Computation.
[53] Li Fei-Fei,et al. MentorNet: Learning Data-Driven Curriculum for Very Deep Neural Networks on Corrupted Labels , 2017, ICML.
[54] Paolo Frasconi,et al. Forward and Reverse Gradient-Based Hyperparameter Optimization , 2017, ICML.
[55] Fabian Pedregosa,et al. Hyperparameter optimization with approximate gradient , 2016, ICML.
[56] Arash Vahdat,et al. Toward Robustness against Label Noise in Training Deep Discriminative Neural Networks , 2017, NIPS.
[57] C A Nelson,et al. Learning to Learn , 2017, Encyclopedia of Machine Learning and Data Mining.