暂无分享,去创建一个
[1] Robert C. Holte,et al. Concept Learning and the Problem of Small Disjuncts , 1989, IJCAI.
[2] J. Ross Quinlan,et al. C4.5: Programs for Machine Learning , 1992 .
[3] David D. Lewis,et al. Heterogeneous Uncertainty Sampling for Supervised Learning , 1994, ICML.
[4] Pat Langley,et al. Static Versus Dynamic Sampling for Data Mining , 1996, KDD.
[5] Andrew P. Bradley,et al. The use of the area under the ROC curve in the evaluation of machine learning algorithms , 1997, Pattern Recognit..
[6] Antal van den Bosch,et al. When small disjuncts abound, try lazy learning: A case study , 1997 .
[7] Stan Matwin,et al. Addressing the Curse of Imbalanced Training Sets: One-Sided Selection , 1997, ICML.
[8] Carla E. Brodley,et al. Pruning Decision Trees with Misclassification Costs , 1998, ECML.
[9] Salvatore J. Stolfo,et al. Toward Scalable Learning with Non-Uniform Class and Cost Distributions: A Case Study in Credit Card Fraud Detection , 1998, KDD.
[10] Ron Kohavi,et al. The Case against Accuracy Estimation for Comparing Induction Algorithms , 1998, ICML.
[11] Catherine Blake,et al. UCI Repository of machine learning databases , 1998 .
[12] Mark R. Wade,et al. Construction and Assessment of Classification Rules , 1999, Technometrics.
[13] Tim Oates,et al. Efficient progressive sampling , 1999, KDD '99.
[14] Yoram Singer,et al. A simple, fast, and effective rule learner , 1999, AAAI 1999.
[15] Robert C. Holte,et al. Exploiting the Cost (In)sensitivity of Decision Tree Splitting Criteria , 2000, ICML.
[16] øöö Blockinøø. Well-Trained PETs : Improving Probability Estimation , 2000 .
[17] J A Swets,et al. Better decisions through science. , 2000, Scientific American.
[18] Haym Hirsh,et al. A Quantitative Study of Small Disjuncts , 2000, AAAI/IAAI.
[19] Charles Elkan,et al. The Foundations of Cost-Sensitive Learning , 2001, IJCAI.
[20] Gary M. Weiss,et al. The effect of class distribution on classifier learning: an empirical study , 2001 .
[21] Foster J. Provost,et al. Active Learning for Class Probability Estimation and Ranking , 2001, IJCAI.
[22] Nathalie Japkowicz,et al. A Mixture-of-Experts Framework for Learning from Imbalanced Data Sets , 2001, IDA.
[23] Bianca Zadrozny,et al. Learning and making decisions when costs and probabilities are both unknown , 2001, KDD '01.
[24] JapkowiczNathalie,et al. The class imbalance problem: A systematic study , 2002 .
[25] Nitesh V. Chawla,et al. SMOTE: Synthetic Minority Over-sampling Technique , 2002, J. Artif. Intell. Res..
[26] Marco Saerens,et al. Adjusting the Outputs of a Classifier to New a Priori Probabilities: A Simple Procedure , 2002, Neural Computation.
[27] Nathalie Japkowicz,et al. The class imbalance problem: A systematic study , 2002, Intell. Data Anal..
[28] Tom Fawcett,et al. Robust Classification for Imprecise Environments , 2000, Machine Learning.
[29] Foster J. Provost,et al. Active Sampling for Class Probability Estimation and Ranking , 2004, Machine Learning.
[30] Tom Fawcett,et al. Adaptive Fraud Detection , 1997, Data Mining and Knowledge Discovery.
[31] Paul R. Cohen,et al. Multiple Comparisons in Induction Algorithms , 2000, Machine Learning.
[32] David A. Cohn,et al. Improving generalization with active learning , 1994, Machine Learning.
[33] Eric Bauer,et al. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.
[34] Gary M. Weiss. A Quantitative Study of Small Disjuncts in Classifier Learning , 2022 .