Abstract The purpose of the ensemble methods is to increase the accuracy of predictionthrough combining many classi ers. According to recent studies, it is proved thatrandom forests and forward stagewise regression have good accuracies in classi cationproblems. However they have great prediction error in separation boundary pointsbecause they used decision tree as a base learner. In this study, we use the kernel ridgeregression instead of the decision trees in random forests and boosting. The usefulnessof our proposed ensemble methods was shown by the simulation results of the prostatecancer and the Boston housing data.Keywords: Boosting, ensemble method, forward stagewise regression, kernel ridge re-gression, random forest. 1. Introduction Ensemble methods are learning algorithms using a set of classi ers to predict a new data’sresponse value by taking a (weighted) vote or averaging of their predictions in the set ofclassi ers. The rst ensemble method is the one using Bayesian averaging. Also there areanother three famous ensemble methods such as bagging (Breiman, 1996), boosting (Freundand Schapire, 1997) and random forests (RF) (Breiman, 2001), which are based on thedecision trees. Especially, boosting and random forests are well known to be excellent in theaccuracy of prediction in classi cation problems.Bagging is a technique reducing the variance of an estimated prediction function and seemsto work well expecially for high-variance and low-bias procedures, such as in the decisiontree. For regression, we simply t the same regression trees many times to bootstrap sampledversions of the training data, to average the results. For classi cation, we cast a vote for thepredicted class in a set of trees.
[1]
J. Shim,et al.
Doubly penalized kernel method for heteroscedastic autoregressive datay
,
2010
.
[2]
Leo Breiman,et al.
Bagging Predictors
,
1996,
Machine Learning.
[3]
R. Tibshirani,et al.
Forward stagewise regression and the monotone lasso
,
2007,
0705.0269.
[4]
J. Shim.
Variable selection in the kernel Cox regression
,
2011
.
[5]
Leo Breiman,et al.
Random Forests
,
2001,
Machine Learning.
[6]
Yoav Freund,et al.
A decision-theoretic generalization of on-line learning and an application to boosting
,
1997,
EuroCOLT.
[7]
R. Tibshirani.
Regression Shrinkage and Selection via the Lasso
,
1996
.