Ensemble methods and partial least squares regression

Recently, there has been increased attention in the literature on the use of ensemble methods in multivariate regression and classification. These methods have been shown to have interesting properties for both regression and classification. In particular, they can improve the accuracy of unstable predictors. Ensemble methods have so far been little studied in situations that are common for calibration and prediction in chemistry, i.e. situations with a large number of collinear x‐variables and few samples. These situations are often approached by data compression methods such as principal component regression (PCR) or partial least squares regression (PLSR). The present paper is an investigation of the properties of different types of ensemble methods used with PLSR in situations with highly collinear x‐data. Bagging and data augmentation by simulated noise are studied. The focus is on the robustness of the calibrations. Real and simulated data are used. The results show that ensembles trained on data with added noise can make PLSR robust against the type of noise added. In particular, the effects of sample temperature variations can be eliminated. Bagging does not seem to give any improvement over PLSR for small and intermediate numbers of components. It is, however, less sensitive to overfitting. Copyright © 2005 John Wiley & Sons, Ltd.

[1]  H. Martens,et al.  Extended multiplicative signal correction and spectral interference subtraction: new preprocessing methods for near infrared spectroscopy. , 1991, Journal of pharmaceutical and biomedical analysis.

[2]  Nathan Intrator,et al.  Bootstrapping with Noise: An Effective Regularization Technique , 1996, Connect. Sci..

[3]  L. Breiman Heuristics of instability and stabilization in model selection , 1996 .

[4]  Onno E. de Noord,et al.  Optimization of partial-least-squares calibration models by simulation of instrumental perturbations , 1997 .

[5]  Jianming Ye On Measuring and Correcting the Effects of Data Mining and Model Selection , 1998 .

[6]  S. Wold,et al.  Orthogonal signal correction of near-infrared spectra , 1998 .

[7]  L. Breiman Arcing Classifiers , 1998 .

[8]  A. J. Morris,et al.  Data augmentation: an alternative approach to the analysis of spectroscopic data , 1998 .

[9]  L. Breiman Arcing classifier (with discussion and a rejoinder by the author) , 1998 .

[10]  Tin Kam Ho,et al.  Complexity of Classification Problems and Comparative Advantages of Combined Classifiers , 2000, Multiple Classifier Systems.

[11]  T. Isaksson,et al.  Studies on the structure of water using two-dimensional near-infrared correlation spectroscopy and principal component analysis. , 2001, Analytical chemistry.

[12]  Jinbo Bi,et al.  Prediction of Protein Retention Times in Anion-Exchange Chromatography Systems Using Support Vector Regression , 2002, J. Chem. Inf. Comput. Sci..

[13]  Agostino Di Ciaccio,et al.  Improving nonparametric regression methods by bagging and boosting , 2002 .

[14]  Hyun-Chul Kim,et al.  Constructing support vector machine ensemble , 2003, Pattern Recognit..

[15]  J. Elder The Generalization Paradox of Ensembles , 2003 .

[16]  Jinbo Bi,et al.  Prediction of Protein Retention Times in Anion-Exchange Chromatography Systems Using Support Vector Regression. , 2003 .

[17]  Zafer Barutçuoglu A Comparison of Model Aggregation Methods for Regression , 2003, ICANN.

[18]  Sungzoon Cho,et al.  Smoothed Bagging with Kernel Bandwidth Selectors , 2004, Neural Processing Letters.

[19]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[20]  Leo Breiman,et al.  Using Iterated Bagging to Debias Regressions , 2001, Machine Learning.

[21]  Leo Breiman,et al.  Stacked regressions , 2004, Machine Learning.

[22]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[23]  Giorgio Valentini,et al.  Cancer recognition with bagged ensembles of support vector machines , 2004, Neurocomputing.

[24]  Eric Bauer,et al.  An Empirical Comparison of Voting Classification Algorithms: Bagging, Boosting, and Variants , 1999, Machine Learning.

[25]  R. Schapire The Strength of Weak Learnability , 1990, Machine Learning.

[26]  Subhabrata Chakraborti,et al.  Nonparametric Statistical Inference , 2011, International Encyclopedia of Statistical Science.