Constructing Better Classifier Ensemble Based on Weighted Accuracy and Diversity Measure
暂无分享,去创建一个
[1] William Nick Street,et al. Ensemble Pruning Via Semi-definite Programming , 2006, J. Mach. Learn. Res..
[2] Rich Caruana,et al. Ensemble selection from libraries of models , 2004, ICML.
[3] Peter Norvig,et al. Artificial Intelligence: A Modern Approach , 1995 .
[4] J. Ross Quinlan,et al. C4.5: Programs for Machine Learning , 1992 .
[5] James P. Reilly,et al. Structures of α-type ions formed in the 157 nm photodissociation of singly-charged peptide ions , 2006 .
[6] Philip S. Yu,et al. Pruning and dynamic scheduling of cost-sensitive ensembles , 2002, AAAI/IAAI.
[7] John H. Holland,et al. Adaptation in Natural and Artificial Systems: An Introductory Analysis with Applications to Biology, Control, and Artificial Intelligence , 1992 .
[8] Tin Kam Ho,et al. The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..
[9] Grigorios Tsoumakas,et al. A Study on Greedy Algorithms for Ensemble Pruning , 2012 .
[10] Gonzalo Martínez-Muñoz,et al. Pruning in ordered bagging ensembles , 2006, ICML.
[11] Ludmila I. Kuncheva,et al. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.
[12] Bogdan Gabrys,et al. Classifier selection for majority voting , 2005, Inf. Fusion.
[13] Xindong Wu,et al. Ensemble pruning via individual contribution ordering , 2010, KDD.
[14] Thomas G. Dietterich,et al. Pruning Adaptive Boosting , 1997, ICML.
[15] Gavin Brown,et al. "Good" and "Bad" Diversity in Majority Vote Ensembles , 2010, MCS.
[16] Xin Yao,et al. An analysis of diversity measures , 2006, Machine Learning.
[17] D. Opitz,et al. Popular Ensemble Methods: An Empirical Study , 1999, J. Artif. Intell. Res..
[18] Sylvain Piechowiak,et al. On the Effectiveness of Diversity When Training Multiple Classifier Systems , 2009, ECSQARU.
[19] Thomas G. Dietterich. What is machine learning? , 2020, Archives of Disease in Childhood.
[20] Robert P. Sheridan,et al. Random Forest: A Classification and Regression Tool for Compound Classification and QSAR Modeling , 2003, J. Chem. Inf. Comput. Sci..
[21] Ian H. Witten,et al. The WEKA data mining software: an update , 2009, SKDD.
[22] Jon Atli Benediktsson,et al. Multiple Classifier Systems , 2015, Lecture Notes in Computer Science.
[23] Grigorios Tsoumakas,et al. An ensemble uncertainty aware measure for directed hill climbing ensemble pruning , 2010, Machine Learning.
[24] Fabio Roli,et al. Design of effective neural network ensembles for image classification purposes , 2001, Image Vis. Comput..
[25] Leo Breiman,et al. Bagging Predictors , 1996, Machine Learning.
[26] David B. Skalak,et al. The Sources of Increased Accuracy for Two Proposed Boosting Algorithms , 1996, AAAI 1996.
[27] Wei Tang,et al. Corrigendum to "Ensembling neural networks: Many could be better than all" [Artificial Intelligence 137 (1-2) (2002) 239-263] , 2010, Artif. Intell..
[28] Yoav Freund,et al. A decision-theoretic generalization of on-line learning and an application to boosting , 1995, EuroCOLT.
[29] Derek Partridge,et al. Software Diversity: Practical Statistics for Its Measurement and Exploitation | Draft Currently under Revision , 1996 .
[30] Thomas G. Dietterich. Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.
[31] Grigorios Tsoumakas,et al. An Ensemble Pruning Primer , 2009, Applications of Supervised and Unsupervised Ensemble Methods.
[32] Corinna Cortes,et al. Support-Vector Networks , 1995, Machine Learning.
[33] Robert E. Schapire,et al. The Boosting Approach to Machine Learning An Overview , 2003 .
[34] Wei Tang,et al. Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..
[35] Lawrence O. Hall,et al. Ensemble diversity measures and their application to thinning , 2004, Inf. Fusion.