Data Mining for Imbalanced Datasets: An Overview
暂无分享,去创建一个
[1] Robert C. Holte,et al. Explicitly representing expected cost: an alternative to ROC representation , 2000, KDD '00.
[2] James P. Egan,et al. Signal detection theory and ROC analysis , 1975 .
[3] Sauchi Stephen Lee. Noisy replication in skewed binary classification , 2000 .
[4] Jorma Laurikkala,et al. Improving Identification of Difficult Small Classes by Balancing Class Distribution , 2001, AIME.
[5] J. Ross Quinlan,et al. C4.5: Programs for Machine Learning , 1992 .
[6] Kai Ming Ting,et al. A Comparative Study of Cost-Sensitive Boosting Algorithms , 2000, ICML.
[7] David D. Lewis,et al. Heterogeneous Uncertainty Sampling for Supervised Learning , 1994, ICML.
[8] Yoav Freund,et al. Experiments with a New Boosting Algorithm , 1996, ICML.
[9] Nitesh V. Chawla,et al. Classification and knowledge discovery in protein databases , 2004, J. Biomed. Informatics.
[10] Nitesh V. Chawla,et al. SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..
[11] N. Japkowicz. Learning from Imbalanced Data Sets: A Comparison of Various Strategies * , 2000 .
[12] Carey E. Priebe,et al. COMPARATIVE EVALUATION OF PATTERN RECOGNITION TECHNIQUES FOR DETECTION OF MICROCALCIFICATIONS IN MAMMOGRAPHY , 1993 .
[13] Mark R. Wade,et al. Construction and Assessment of Classification Rules , 1999, Technometrics.
[14] Nathalie Japkowicz,et al. The Class Imbalance Problem: Significance and Strategies , 2000 .
[15] Steven Salzberg,et al. A Weighted Nearest Neighbor Algorithm for Learning with Symbolic Features , 2004, Machine Learning.
[16] Herna L. Viktor,et al. Learning from imbalanced data sets with boosting and data generation: the DataBoost-IM approach , 2004, SKDD.
[17] Stan Matwin,et al. Addressing the Curse of Imbalanced Training Sets: One-Sided Selection , 1997, ICML.
[18] Peter E. Hart,et al. The condensed nearest neighbor rule (Corresp.) , 1968, IEEE Trans. Inf. Theory.
[19] Gustavo E. A. P. A. Batista,et al. A study of the behavior of several methods for balancing machine learning training data , 2004, SKDD.
[20] Robert C. Holte,et al. C4.5, Class Imbalance, and Cost Sensitivity: Why Under-Sampling beats Over-Sampling , 2003 .
[21] Nitesh V. Chawla,et al. C4.5 and Imbalanced Data sets: Investigating the eect of sampling method, probabilistic estimate, and decision tree structure , 2003 .
[22] Susan T. Dumais,et al. Inductive learning algorithms and representations for text categorization , 1998, CIKM '98.
[23] Nathalie Japkowicz,et al. Concept-Learning in the Presence of Between-Class and Within-Class Imbalances , 2001, Canadian Conference on AI.
[24] Charles Elkan,et al. The Foundations of Cost-Sensitive Learning , 2001, IJCAI.
[25] David M. J. Tax,et al. One-class classification , 2001 .
[26] Peter D. Turney. Types of Cost in Inductive Concept Learning , 2002, ArXiv.
[27] Tom Fawcett,et al. Robust Classification for Imprecise Environments , 2000, Machine Learning.
[28] Charles X. Ling,et al. Data Mining for Direct Marketing: Problems and Solutions , 1998, KDD.
[29] Thomas G. Dietterich. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.
[30] J A Swets,et al. Measuring the accuracy of diagnostic systems. , 1988, Science.
[31] Robert P. W. Duin,et al. Uncertainty sampling methods for one-class classifiers , 2003 .
[32] Pedro M. Domingos. MetaCost: a general method for making classifiers cost-sensitive , 1999, KDD '99.
[33] Nitesh V. Chawla,et al. SMOTEBoost: Improving Prediction of the Minority Class in Boosting , 2003, PKDD.
[34] I. Tomek,et al. Two Modifications of CNN , 1976 .
[35] M. Maloof. Learning When Data Sets are Imbalanced and When Costs are Unequal and Unknown , 2003 .
[36] Damminda Alahakoon,et al. Minority report in fraud detection: classification of skewed data , 2004, SKDD.
[37] Salvatore J. Stolfo,et al. AdaCost: Misclassification Cost-Sensitive Boosting , 1999, ICML.
[38] C. G. Hilborn,et al. The Condensed Nearest Neighbor Rule , 1967 .
[39] A. S. Schistad Solberg,et al. A large-scale evaluation of features for automatic detection of oil spills in ERS SAR images , 1996, IGARSS '96. 1996 International Geoscience and Remote Sensing Symposium.
[40] Fredric C. Gey,et al. The Relationship between Recall and Precision , 1994, J. Am. Soc. Inf. Sci..
[41] Adam Kowalczyk,et al. Extreme re-balancing for SVMs: a case study , 2004, SKDD.
[42] Nathalie Japkowicz,et al. Supervised Versus Unsupervised Binary-Learning by Feedforward Neural Networks , 2004, Machine Learning.
[43] Foster J. Provost,et al. Learning When Training Data are Costly: The Effect of Class Distribution on Tree Induction , 2003, J. Artif. Intell. Res..
[44] Taeho Jo,et al. Class imbalances versus small disjuncts , 2004, SKDD.
[45] Eric Bauer,et al. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.
[46] Moninder Singh,et al. Learning Goal Oriented Bayesian Networks for Telecommunications Risk Management , 1996, ICML.
[47] William W. Cohen. Fast Effective Rule Induction , 1995, ICML.
[48] Nitesh V. Chawla,et al. Editorial: special issue on learning from imbalanced data sets , 2004, SKDD.
[49] Dunja Mladenic,et al. Feature Selection for Unbalanced Class Distribution and Naive Bayes , 1999, ICML.
[50] Stan Matwin,et al. Machine Learning for the Detection of Oil Spills in Satellite Radar Images , 1998, Machine Learning.
[51] Andreas Stolcke,et al. A study in machine learning from imbalanced data for sentence boundary detection in speech , 2006, Comput. Speech Lang..
[52] Andrew P. Bradley,et al. The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..
[53] Vipin Kumar,et al. Evaluating boosting algorithms to classify rare classes: comparison and improvements , 2001, Proceedings 2001 IEEE International Conference on Data Mining.