GMM discriminant analysis with noisy label for each class
暂无分享,去创建一个
Run-kun Lu | Jian-wei Liu | Zheng-ping Ren | Xiong-lin Luo | Xiong-lin Luo | Jian-wei Liu | Zheng-ping Ren | Run-kun Lu
[1] Clayton Scott,et al. A Rate of Convergence for Mixture Proportion Estimation, with Application to Learning from Noisy Labels , 2015, AISTATS.
[2] Michael Kearns,et al. Efficient noise-tolerant learning from statistical queries , 1993, STOC.
[3] Aditya Krishna Menon,et al. Learning with Symmetric Label Noise: The Importance of Being Unhinged , 2015, NIPS.
[4] Wensheng Zhang,et al. A method for predicting disease subtypes in presence of misclassification among training samples using gene expression: application to human breast cancer , 2006, Bioinform..
[5] Gilles Blanchard,et al. Classification with Asymmetric Label Noise: Consistency and Maximal Denoising , 2013, COLT.
[6] Jakramate Bootkrajang,et al. Supervised learning with random labelling errors , 2013 .
[7] Xiaogang Wang,et al. Deep Self-Learning From Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).
[8] Robert C. Williamson,et al. Learning in the Presence of Corruption , 2015, ArXiv.
[9] Nagarajan Natarajan,et al. Learning with Noisy Labels , 2013, NIPS.
[10] Sally A. Goldman,et al. Can PAC learning algorithms tolerate random attribute noise? , 1995, Algorithmica.
[11] Bernhard Schölkopf,et al. Estimating a Kernel Fisher Discriminant in the Presence of Label Noise , 2001, ICML.
[12] Andrés R. Masegosa,et al. Bagging Decision Trees on Data Sets with Classification Noise , 2010, FoIKS.
[13] Charles Bouveyron,et al. Robust supervised classification with mixture models: Learning from data with uncertain labels , 2009, Pattern Recognit..
[14] Eric R. Ziegel,et al. The Elements of Statistical Learning , 2003, Technometrics.
[15] Rocco A. Servedio,et al. Random classification noise defeats all convex potential boosters , 2008, ICML '08.
[16] Jieping Ye,et al. Generalized Linear Discriminant Analysis: A Unified Framework and Efficient Model Selection , 2008, IEEE Transactions on Neural Networks.
[17] Naresh Manwani,et al. Noise Tolerance Under Risk Minimization , 2011, IEEE Transactions on Cybernetics.
[18] Gunnar Rätsch,et al. Soft Margins for AdaBoost , 2001, Machine Learning.
[19] Charles Elkan,et al. Expectation Maximization Algorithm , 2010, Encyclopedia of Machine Learning.
[20] Ata Kabán,et al. Error bounds for Kernel Fisher Linear Discriminant in Gaussian Hilbert space , 2012, AISTATS.
[21] André Carlos Ponce de Leon Ferreira de Carvalho,et al. Pre-processing for noise detection in gene expression classification data , 2009, Journal of the Brazilian Computer Society.
[22] Cheng Soon Ong,et al. Learning from Corrupted Binary Labels via Class-Probability Estimation , 2015, ICML.
[23] Marcel J. T. Reinders,et al. Classification in the presence of class noise using a probabilistic Kernel Fisher method , 2007, Pattern Recognit..
[24] Sanjoy Dasgupta,et al. Learning mixtures of Gaussians , 1999, 40th Annual Symposium on Foundations of Computer Science (Cat. No.99CB37039).
[25] Jirí Grim,et al. On numerical evaluation of maximum-likelihood estimates for finite mixtures of distributions , 1982, Kybernetika.
[26] Tony R. Martinez,et al. Reduction Techniques for Instance-Based Learning Algorithms , 2000, Machine Learning.
[27] Heekuck Oh,et al. Neural Networks for Pattern Recognition , 1993, Adv. Comput..
[28] Taghi M. Khoshgoftaar,et al. Supervised Neural Network Modeling: An Empirical Investigation Into Learning From Imbalanced Data With Labeling Errors , 2010, IEEE Transactions on Neural Networks.
[29] Carla E. Brodley,et al. Identifying and Eliminating Mislabeled Training Instances , 1996, AAAI/IAAI, Vol. 1.
[30] Jun Du,et al. Modelling Class Noise with Symmetric and Asymmetric Distributions , 2015, AAAI.
[31] Volodymyr Melnykov,et al. Initializing the EM algorithm in Gaussian mixture models with an unknown number of components , 2012, Comput. Stat. Data Anal..
[32] Dacheng Tao,et al. Classification with Noisy Labels by Importance Reweighting , 2014, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[33] Bing Liu,et al. Learning with Positive and Unlabeled Examples Using Weighted Logistic Regression , 2003, ICML.
[34] Ata Kabán,et al. Label-Noise Robust Logistic Regression and Its Applications , 2012, ECML/PKDD.
[35] Aníbal R. Figueiras-Vidal,et al. Boosting by weighting critical and erroneous samples , 2006, Neurocomputing.
[36] Albert Fornells,et al. A study of the effect of different types of noise on the precision of supervised learning techniques , 2010, Artificial Intelligence Review.
[37] Radu Horaud,et al. DeepGUM: Learning Deep Robust Regression with a Gaussian-Uniform Mixture Model , 2018, ECCV.
[38] M. Verleysen,et al. Classification in the Presence of Label Noise: A Survey , 2014, IEEE Transactions on Neural Networks and Learning Systems.
[39] Taghi M. Khoshgoftaar,et al. Robustness of Threshold-Based Feature Rankers with Data Sampling on Noisy and Imbalanced Data , 2012, FLAIRS.
[40] Carla E. Brodley,et al. Identifying Mislabeled Training Data , 1999, J. Artif. Intell. Res..
[41] D. Rubin,et al. Maximum likelihood from incomplete data via the EM - algorithm plus discussions on the paper , 1977 .
[42] Galen Reeves,et al. Gaussian Mixture Models for Stochastic Block Models with Non-Vanishing Noise , 2019, 2019 IEEE 8th International Workshop on Computational Advances in Multi-Sensor Adaptive Processing (CAMSAP).
[43] Junnan Li,et al. DivideMix: Learning with Noisy Labels as Semi-supervised Learning , 2020, ICLR.
[44] D. Angluin,et al. Learning From Noisy Examples , 1988, Machine Learning.
[45] Junmo Kim,et al. NLNL: Negative Learning for Noisy Labels , 2019, 2019 IEEE/CVF International Conference on Computer Vision (ICCV).