Self-Validated Ensemble Models for Design of Experiments

In the last twenty years, the prediction accuracy of machine learning models fit to observational data has improved dramatically. Many machine learning techniques require that the data be partitioned into at least two subsets; a training set for fitting models and a validation set for tuning models. Machine learning techniques requiring data partitioning have generally not been applied to designed experiments (DOEs), as the design structure and small run size limit the ability to withhold observations from the fitting algorithm. We introduce a new model-building algorithm, called self-validated ensemble models (SVEM), that emulates data partitioning by using the complete data simultaneously as both a training and a validation set. SVEM weights the two copies of the data differently under a weighting scheme based on the fractional-random-weight bootstrap (Xu et al., 2020). Similar to bagging (Breiman, 1994), this fractional-random-weight bootstrapping scheme is repeated many times and the final SVEM model is the sample average of the bootstrapped models. In this work, we investigate the performance of the SVEM algorithm with regression, Lasso, and the Dantzig Selector. However, the method is very general and can be applied in combination with most model selection and fitting algorithms. Through extensive simulations and a case study, we demonstrate that SVEM generates models with lower prediction error as compared to more traditional statistical approaches that are based on fitting a single model.

[1]  Yili Hong,et al.  Applications of the Fractional-Random-Weight Bootstrap , 2018, The American Statistician.

[2]  Geoffrey E. Hinton,et al.  Bayesian Learning for Neural Networks , 1995 .

[3]  D. Ruppert The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2004 .

[4]  Raul Cano On The Bayesian Bootstrap , 1992 .

[5]  Christopher J. Nachtsheim,et al.  A Class of Three-Level Designs for Definitive Screening in the Presence of Second-Order Effects , 2011 .

[6]  Terence Tao,et al.  The Dantzig selector: Statistical estimation when P is much larger than n , 2005, math/0506081.

[7]  G. Molenberghs,et al.  Past, Present, and Future of Statistical Science , 2014 .

[8]  G. Box,et al.  On the Experimental Attainment of Optimum Conditions , 1951 .

[9]  Hans C. van Houwelingen,et al.  The Elements of Statistical Learning, Data Mining, Inference, and Prediction. Trevor Hastie, Robert Tibshirani and Jerome Friedman, Springer, New York, 2001. No. of pages: xvi+533. ISBN 0‐387‐95284‐5 , 2004 .

[10]  M. Eissa,et al.  Application of quality by design approach for HPTLC simultaneous determination of amlodipine and celecoxib in presence of process-related impurity , 2020, Microchemical Journal.

[11]  Priyabrata Pattnaik,et al.  Production and purification of plasmid DNA vaccines: is there scope for further innovation? , 2014, Expert review of vaccines.

[12]  L. Breiman Heuristics of instability and stabilization in model selection , 1996 .

[13]  Mikhail Belkin,et al.  Reconciling modern machine-learning practice and the classical bias–variance trade-off , 2018, Proceedings of the National Academy of Sciences.

[14]  Galit Shmueli,et al.  To Explain or To Predict? , 2010, 1101.0891.

[15]  Peter Goos,et al.  Optimal Design of Experiments: A Case Study Approach , 2011 .

[16]  Trevor Hastie,et al.  An Introduction to Statistical Learning , 2013, Springer Texts in Statistics.

[17]  V. Roshan Joseph,et al.  Space-filling designs for computer experiments: A review , 2016 .

[18]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[19]  Geert Molenberghs,et al.  In praise of sparsity and convexity , 2014 .

[20]  J. Mackolil,et al.  Flow of nanoliquid past a vertical plate with novel quadratic thermal radiation and quadratic Boussinesq approximation: Sensitivity analysis , 2020 .

[21]  Snigdhansu Chatterjee,et al.  U-Statistics, Mm-Estimators and Resampling , 2018 .

[22]  Christopher J. Nachtsheim,et al.  Effective Design-Based Model Selection for Definitive Screening Designs , 2017, Technometrics.

[23]  N. Meinshausen,et al.  Discussion: A tale of three cousins: Lasso, L2Boosting and Dantzig , 2007, 0803.3134.

[24]  David R. Anderson,et al.  Model Selection and Multimodel Inference , 2003 .

[25]  Snigdhansu Chatterjee,et al.  © Institute of Mathematical Statistics, 2005 GENERALIZED BOOTSTRAP FOR ESTIMATING EQUATIONS , 2022 .

[26]  Robert W. Bondi,et al.  Effect of Experimental Design on the Prediction Performance of Calibration Models Based on Near-Infrared Spectroscopy for Pharmaceutical Applications , 2012, Applied spectroscopy.

[27]  Robert Tibshirani,et al.  In praise of sparsity and convexity , 2013 .

[28]  Douglas C. Montgomery,et al.  Analysis of definitive screening designs: Screening vs prediction , 2018 .

[29]  Robert Tibshirani,et al.  The Elements of Statistical Learning: Data Mining, Inference, and Prediction , 2001, Springer Series in Statistics.

[30]  Amin Karbasi,et al.  Multiple Descent: Design Your Own Generalization Curve , 2020, NeurIPS.

[31]  H. Zou,et al.  Structured variable selection and estimation , 2009, 1011.0610.

[32]  David Hinkley,et al.  Bootstrap Methods: Another Look at the Jackknife , 2008 .

[33]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[34]  Yi Lin,et al.  An Efficient Variable Selection Approach for Analyzing Designed Experiments , 2007, Technometrics.

[35]  Christopher J. Marley,et al.  A comparison of design and model selection methods for supersaturated experiments , 2010, Comput. Stat. Data Anal..

[36]  Robert Tibshirani,et al.  An Introduction to the Bootstrap , 1994 .

[37]  David J. Edwards,et al.  Response surface models: To reduce or not to reduce? , 2020 .

[38]  F. Vrečer,et al.  Application of experimental design methodology in development and optimization of drug release method. , 2005, International journal of pharmaceutics.