Confidence intervals for random forests: the jackknife and the infinitesimal jackknife

We study the variability of predictions made by bagged learners and random forests, and show how to estimate standard errors for these methods. Our work builds on variance estimates for bagging proposed by Efron (1992, 2013) that are based on the jackknife and the infinitesimal jackknife (IJ). In practice, bagged predictors are computed using a finite number B of bootstrap replicates, and working with a large B can be computationally expensive. Direct applications of jackknife and IJ estimators to bagging require B = Θ(n1.5) bootstrap replicates to converge, where n is the size of the training set. We propose improved versions that only require B = Θ(n) replicates. Moreover, we show that the IJ estimator requires 1.7 times less bootstrap replicates than the jackknife to achieve a given accuracy. Finally, we study the sampling distributions of the jackknife and IJ variance estimates themselves. We illustrate our findings with multiple experiments and simulation studies.

[1]  P. Hall,et al.  Effects of bagging and bias correction on estimators defined by estimating equations , 2003 .

[2]  Bradley Efron,et al.  Model selection , estimation , and bootstrap smoothing , 2012 .

[3]  C. Mallows Some Comments on Cp , 2000, Technometrics.

[4]  A. Buja,et al.  OBSERVATIONS ON BAGGING , 2006 .

[5]  Christina Gloeckner,et al.  Modern Applied Statistics With S , 2003 .

[6]  Joseph Sexton,et al.  Standard errors for bagged and random forest estimators , 2009, Comput. Stat. Data Anal..

[7]  B. Efron Jackknife‐After‐Bootstrap Standard Errors and Influence Functions , 1992 .

[8]  Leo Breiman,et al.  Bagging Predictors , 1996, Machine Learning.

[9]  Gérard Biau,et al.  Analysis of a Random Forests Model , 2010, J. Mach. Learn. Res..

[10]  Thomas G. Dietterich An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees: Bagging, Boosting, and Randomization , 2000, Machine Learning.

[11]  G DietterichThomas An Experimental Comparison of Three Methods for Constructing Ensembles of Decision Trees , 2000 .

[12]  Luc Devroye,et al.  Consistency of Random Forests and Other Averaging Classifiers , 2008, J. Mach. Learn. Res..

[13]  Achim Zeileis,et al.  Bias in random forest variable importance measures: Illustrations, sources and a solution , 2007, BMC Bioinformatics.

[14]  Andy Liaw,et al.  Classification and Regression by randomForest , 2007 .

[15]  B. Efron,et al.  The Jackknife Estimate of Variance , 1981 .

[16]  J. Friedman,et al.  On bagging and nonlinear estimation , 2007 .

[17]  Pierre Geurts,et al.  Extremely randomized trees , 2006, Machine Learning.

[18]  Jiangtao Duan Bootstrap-Based Variance Estimators for A Bagging Predictor. , 2011 .

[19]  Robert P. W. Duin,et al.  Bagging for linear classifiers , 1998, Pattern Recognit..

[20]  Guangjian Zhang,et al.  The Infinitesimal Jackknife with Exploratory Factor Analysis , 2012 .

[21]  Nicolai Meinshausen,et al.  Quantile Regression Forests , 2006, J. Mach. Learn. Res..

[22]  P. Bühlmann,et al.  Analyzing Bagging , 2001 .

[23]  B. Efron Estimation and Accuracy After Model Selection , 2014, Journal of the American Statistical Association.

[24]  J. Friedman Stochastic gradient boosting , 2002 .

[25]  Colin L. Mallows,et al.  Some Comments on Cp , 2000, Technometrics.

[26]  B. Efron,et al.  Compliance as an Explanatory Variable in Clinical Trials , 1991 .

[27]  Trevor Hastie,et al.  The Elements of Statistical Learning , 2001 .

[28]  Yi Lin,et al.  Random Forests and Adaptive Nearest Neighbors , 2006 .

[29]  T. Stamey,et al.  Prostate specific antigen in the diagnosis and treatment of adenocarcinoma of the prostate. II. Radical prostatectomy treated patients. , 1989, The Journal of urology.

[30]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.