Ensemble classification based on generalized additive models
暂无分享,去创建一个
[1] Agostino Di Ciaccio,et al. Improving nonparametric regression methods by bagging and boosting , 2002 .
[2] Michael G. Akritas,et al. Recent Advances and Trends in Nonparametric Statistics , 2003 .
[3] Francis K. H. Quek,et al. Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets , 2003, Pattern Recognit..
[4] Leo Breiman,et al. Bagging Predictors , 1996, Machine Learning.
[5] Pat Langley,et al. Crafting Papers on Machine Learning , 2000, ICML.
[6] Chun-Xia Zhang,et al. RotBoost: A technique for combining Rotation Forest and AdaBoost , 2008, Pattern Recognit. Lett..
[7] Ludmila I. Kuncheva,et al. Combining Pattern Classifiers: Methods and Algorithms , 2004 .
[8] Makoto Abe,et al. A Generalized Additive Model for Discrete-Choice Data , 1999 .
[9] Bruce K Armstrong,et al. Lung cancer rate predictions using generalized additive models. , 2005, Biostatistics.
[10] D. Ruppert. The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .
[11] Ludmila I. Kuncheva,et al. Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy , 2003, Machine Learning.
[12] Laurent Heutte,et al. Influence of Hyperparameters on Random Forest Accuracy , 2009, MCS.
[13] David W. Opitz,et al. Generating Accurate and Diverse Members of a Neural-Network Ensemble , 1995, NIPS.
[14] Subhash C. Bagui,et al. Combining Pattern Classifiers: Methods and Algorithms , 2005, Technometrics.
[15] O. J. Dunn. Multiple Comparisons among Means , 1961 .
[16] Paul H. C. Eilers,et al. Direct generalized additive modeling with penalized likelihood , 1998 .
[17] Juan José Rodríguez Diez,et al. Classifier Ensembles with a Random Linear Oracle , 2007, IEEE Transactions on Knowledge and Data Engineering.
[18] Dirk Van den Poel,et al. FACULTEIT ECONOMIE , 2007 .
[19] M. Friedman. The Use of Ranks to Avoid the Assumption of Normality Implicit in the Analysis of Variance , 1937 .
[20] Yoshua Bengio,et al. Boosting Neural Networks , 2000, Neural Computation.
[21] E. N. Zwane,et al. Semiparametric models for capture-recapture studies with covariates , 2004, Comput. Stat. Data Anal..
[22] Andy Liaw,et al. Classification and Regression by randomForest , 2007 .
[23] Torsten Hothorn,et al. Bundling Classifiers by Bagging Trees , 2002, Comput. Stat. Data Anal..
[24] Oleksandr Makeyev,et al. Neural network with ensembles , 2010, The 2010 International Joint Conference on Neural Networks (IJCNN).
[25] Kristof Coussement,et al. GAMens: Applies GAMens, GAMrsm and GAMbag ensemble classifiers. R Package version 1.11. , 2010 .
[26] Annibale Biggeri,et al. Parametric and semi-parametric approaches in the analysis of short-term effects of air pollution on health , 2007, Comput. Stat. Data Anal..
[27] Eric Bauer,et al. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.
[28] Robert P. W. Duin,et al. The Role of Combining Rules in Bagging and Boosting , 2000, SSPR/SPR.
[29] Juan José Rodríguez Diez,et al. Rotation Forest: A New Classifier Ensemble Method , 2006, IEEE Transactions on Pattern Analysis and Machine Intelligence.
[30] Tin Kam Ho,et al. The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..
[31] A. Prasad,et al. Newer Classification and Regression Tree Techniques: Bagging and Random Forests for Ecological Prediction , 2006, Ecosystems.
[32] M. Friedman. A Comparison of Alternative Tests of Significance for the Problem of $m$ Rankings , 1940 .
[33] Daniel Berg. Bankruptcy Prediction by Generalized Additive Models , 2006 .
[34] Thomas G. Dietterich. Multiple Classifier Systems , 2000, Lecture Notes in Computer Science.
[35] Jude W. Shavlik,et al. Combining the Predictions of Multiple Classifiers: Using Competitive Learning to Initialize Neural Networks , 1995, IJCAI.
[36] Pierre Geurts,et al. Extremely randomized trees , 2006, Machine Learning.
[37] Kellie J. Archer,et al. Empirical characterization of random forest variable importance measures , 2008, Comput. Stat. Data Anal..
[38] R. Tibshirani,et al. Generalized Additive Models: Some Applications , 1987 .
[39] Ron Kohavi,et al. The Case against Accuracy Estimation for Comparing Induction Algorithms , 1998, ICML.
[40] S. Eguchi,et al. An introduction to the predictive technique AdaBoost with a comparison to generalized additive models , 2005 .
[41] Ramón Díaz-Uriarte,et al. Gene selection and classification of microarray data using random forest , 2006, BMC Bioinformatics.
[42] J. Friedman. Special Invited Paper-Additive logistic regression: A statistical view of boosting , 2000 .
[43] P. Bühlmann. Bagging, subagging and bragging for improving some prediction algorithms , 2003 .
[44] Anne M. P. Canuto,et al. Investigating the influence of the choice of the ensemble members in accuracy and diversity of selection-based and fusion-based methods for ensembles , 2007, Pattern Recognit. Lett..
[45] Janez Demsar,et al. Statistical Comparisons of Classifiers over Multiple Data Sets , 2006, J. Mach. Learn. Res..
[46] Hyun-Chul Kim,et al. Constructing support vector machine ensemble , 2003, Pattern Recognit..
[47] Robert P. Sheridan,et al. Random Forest: A Classification and Regression Tool for Compound Classification and QSAR Modeling , 2003, J. Chem. Inf. Comput. Sci..
[48] Wei Tang,et al. Ensembling neural networks: Many could be better than all , 2002, Artif. Intell..
[49] Leo Breiman,et al. Random Forests , 2001, Machine Learning.
[50] Johannes R. Sveinsson,et al. Random Forests for land cover classification , 2006, Pattern Recognit. Lett..
[51] Hyun-Chul Kim,et al. Support Vector Machine Ensemble with Bagging , 2002, SVM.