Abstract This article proposes a two-step iterative procedure to improve the misclassification error rate of an initial classification rule. The first step involves an iterative method for generating a sequence of classifiers from the initial one; this is based on the augmentation of the feature vector with some new pseudo-predictors. Unlike other components of the feature vector, these new pseudo-predictors tend to provide information primarily on the performance or correctness of the classifier itself. The second step of the proposed procedure “pools together” the classifiers constructed in step one in order to produce a new classifier which is far more effective (in an asymptotic sense) than the initial classifier. In addition to these results, a data-splitting approach for selecting the number of iterations will also be discussed. Both the mechanics and the asymptotic validity of the proposed procedure are studied.
[1]
Luc Devroye,et al.
Automatic Pattern Recognition: A Study of the Probability of Error
,
1988,
IEEE Trans. Pattern Anal. Mach. Intell..
[2]
Yoav Freund,et al.
Boosting the margin: A new explanation for the effectiveness of voting methods
,
1997,
ICML.
[3]
Yoav Freund,et al.
A Short Introduction to Boosting
,
1999
.
[4]
G. Lugosi,et al.
Consistency of Data-driven Histogram Methods for Density Estimation and Classification
,
1996
.
[5]
M. Mojirsheibani.
Combining Classifiers via Discretization
,
1999
.
[6]
László Györfi,et al.
A Probabilistic Theory of Pattern Recognition
,
1996,
Stochastic Modelling and Applied Probability.