Random Forests
暂无分享,去创建一个
[1] Yoav Freund,et al. Experiments with a New Boosting Algorithm , 1996, ICML.
[2] Robert Tibshirani,et al. Bias, Variance and Prediction Error for Classification Rules , 1996 .
[3] Yoav Freund,et al. Boosting the margin: A new explanation for the effectiveness of voting methods , 1997, ICML.
[4] Yali Amit,et al. Shape Quantization and Recognition with Randomized Trees , 1997, Neural Computation.
[5] Dale Schuurmans,et al. Boosting in the Limit: Maximizing the Margin of Learned Ensembles , 1998, AAAI/IAAI.
[6] Tin Kam Ho,et al. The Random Subspace Method for Constructing Decision Forests , 1998, IEEE Trans. Pattern Anal. Mach. Intell..
[7] L. Breiman. Arcing classifier (with discussion and a rejoinder by the author) , 1998 .
[8] L. Breiman. USING ADAPTIVE BAGGING TO DEBIAS REGRESSIONS , 1999 .
[9] L. Breiman. SOME INFINITY THEORY FOR PREDICTOR ENSEMBLES , 2000 .
[10] Eugene M. Kleinberg,et al. On the Algorithmic Implementation of Stochastic Discrimination , 2000, IEEE Trans. Pattern Anal. Mach. Intell..
[11] Leo Breiman,et al. Randomizing Outputs to Increase Prediction Accuracy , 2000, Machine Learning.
[12] David H. Wolpert,et al. An Efficient Method To Estimate Bagging's Generalization Error , 1999, Machine Learning.
[13] Thomas G. Dietterich. An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.
[14] Leo Breiman,et al. Bagging Predictors , 1996, Machine Learning.
[15] Eric Bauer,et al. An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.