Stability

Reproducibility is imperative for any scientific discovery. More often than not, modern scientific findings rely on statistical analysis of high-dimensional data. At a minimum, reproducibility manifests itself in stability of statistical results relative to"reasonable"perturbations to data and to the model used. Jacknife, bootstrap, and cross-validation are based on perturbations to data, while robust statistics methods deal with perturbations to models. In this article, a case is made for the importance of stability in statistics. Firstly, we motivate the necessity of stability for interpretable and reliable encoding models from brain fMRI signals. Secondly, we find strong evidence in the literature to demonstrate the central role of stability in statistical inference, such as sensitivity analysis and effect detection. Thirdly, a smoothing parameter selector based on estimation stability (ES), ES-CV, is proposed for Lasso, in order to bring stability to bear on cross-validation (CV). ES-CV is then utilized in the encoding models to reduce the number of predictors by 60% with almost no loss (1.3%) of prediction performance across over 2,000 voxels. Last, a novel"stability"argument is seen to drive new results that shed light on the intriguing interactions between sample to sample variability and heavier tail error distribution (e.g., double-exponential) in high-dimensional regression models with $p$ predictors and $n$ independent samples. In particular, when $p/n\rightarrow\kappa\in(0.3,1)$ and the error distribution is double-exponential, the Ordinary Least Squares (OLS) is a better estimator than the Least Absolute Deviation (LAD) estimator.

[1]  Arnold J Stromberg,et al.  Subsampling , 2001, Technometrics.

[2]  Peter Hall,et al.  Bootstrap methods in statistics , 2000 .

[3]  Arthur E. Hoerl,et al.  Ridge Regression: Biased Estimation for Nonorthogonal Problems , 2000, Technometrics.

[4]  Dana Ron,et al.  Algorithmic Stability and Sanity-Check Bounds for Leave-One-Out Cross-Validation , 1997, Neural Computation.

[5]  L. Breiman Heuristics of instability and stabilization in model selection , 1996 .

[6]  David J. Field,et al.  Emergence of simple-cell receptive field properties by learning a sparse code for natural images , 1996, Nature.

[7]  J. Shao Bootstrap Model Selection , 1996 .

[8]  S. I. Gelʹfand,et al.  Methods of Homological Algebra , 1996 .

[9]  C. Pace,et al.  How to measure and predict the molar absorption coefficient of a protein , 1995, Protein science : a publication of the Protein Society.

[10]  Joseph P. Romano,et al.  Large Sample Confidence Regions Based on Subsamples under Minimal Assumptions , 1994 .

[11]  S. Venyaminov,et al.  Circular dichroic analysis of denatured proteins: inclusion of denatured proteins in the reference set. , 1993, Analytical biochemistry.

[12]  Ping Zhang Model Selection Via Multifold Cross Validation , 1993 .

[13]  Jeffrey Adams,et al.  HARISH-CHANDRA'S METHOD OF DESCENT , 1992 .

[14]  K. Parker,et al.  Sequence motifs important for peptide binding to the human MHC class I molecule, HLA-A2. , 1992, Journal of immunology.

[15]  A. McMichael,et al.  A critical role for conserved residues in the cleft of HLA-A2 in presentation of a nonapeptide to T cells. , 1992, Science.

[16]  M. Goodale,et al.  Separate visual pathways for perception and action , 1992, Trends in Neurosciences.

[17]  David W. Allan,et al.  Frequency and time stability of GPS and GLONASS clocks , 1990, 44th Annual Symposium on Frequency Control.

[18]  H. Künsch The Jackknife and the Bootstrap for General Stationary Observations , 1989 .

[19]  K. Breslauer,et al.  Calculating thermodynamic data for transitions of any molecularity from equilibrium melting curves , 1987, Biopolymers.

[20]  Changbao Wu,et al.  Jackknife, Bootstrap and Other Resampling Methods in Regression Analysis , 1986 .

[21]  John Law,et al.  Robust Statistics—The Approach Based on Influence Functions , 1986 .

[22]  Ker-Chau Li,et al.  Asymptotic optimality of CL and generalized cross-validation in ridge regression with application to spline smoothing , 1986 .

[23]  E. Carlstein The Use of Subseries Values for Estimating the Variance of a General Statistic from a Stationary Sequence , 1986 .

[24]  Thomas Yunck,et al.  GPS-Based Satellite Tracking System for Precise Positioning , 1985, IEEE Transactions on Geoscience and Remote Sensing.

[25]  P. Hall Large Sample Optimality of Least Squares Cross-Validation in Density Estimation , 1983 .

[26]  D. Freedman,et al.  Some Asymptotic Theory for the Bootstrap , 1981 .

[27]  Jr. David A. Vogan Irreducible characters of semisimple Lie groups I , 1979 .

[28]  Luc Devroye,et al.  Distribution-free inequalities for the deleted and holdout error estimates , 1979, IEEE Trans. Inf. Theory.

[29]  M. Stone Cross‐Validatory Choice and Assessment of Statistical Predictions , 1976 .

[30]  P. Bickel One-Step Huber Estimates in the Linear Model , 1975 .

[31]  J. Hartigan NECESSARY AND SUFFICIENT CONDITIONS FOR ASYMPTOTIC JOINT NORMALITY OF A STATISTIC AND ITS SUBSAMPLE VALUES , 1975 .

[32]  C. J. Lawrence Robust estimates of location : survey and advances , 1975 .

[33]  F. Hampel The Influence Curve and Its Role in Robust Estimation , 1974 .

[34]  David M. Allen,et al.  The Relationship Between Variable Selection and Data Agumentation and a Method for Prediction , 1974 .

[35]  F. Hampel A General Qualitative Definition of Robustness , 1971 .

[36]  J. A. Hartigan,et al.  Using Subsample Values as Typical Values , 1969 .

[37]  Louis Essen,et al.  Accurate measurement of time , 1960 .

[38]  D. Hubel,et al.  Receptive fields of single neurones in the cat's striate cortex , 1959, The Journal of physiology.

[39]  M. H. Quenouille NOTES ON BIAS IN ESTIMATION , 1956 .

[40]  David A. Howe,et al.  Definitions of "total" estimators of common time-domain variances , 2001, Proceedings of the 2001 IEEE International Frequncy Control Symposium and PDA Exhibition (Cat. No.01CH37218).

[41]  Peter Dayan,et al.  Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems , 2001 .

[42]  R. Tibshirani Regression Shrinkage and Selection via the Lasso , 1996 .

[43]  M. Chavance [Jackknife and bootstrap]. , 1992, Revue d'epidemiologie et de sante publique.

[44]  B. Efron The jackknife, the bootstrap, and other resampling plans , 1987 .

[45]  P. Rousseeuw,et al.  Robust statistics: the approach based on influence functions , 1986 .

[46]  Frederick R. Forst,et al.  On robust estimation of the location parameter , 1980 .

[47]  Mario Bertero,et al.  The Stability of Inverse Problems , 1980 .

[48]  J. Engel,et al.  The triple helix in equilibrium with coil conversion of collagen-like polytripeptides in aqueous and nonaqueous solvents. Comparison of the thermodynamic parameters and the binding of water to (L-Pro-L-Pro-Gly)n and (L-Pro-L-Hyp-Gly)n. , 1977, Biopolymers.

[49]  S. Portnoy Robust Estimation in Dependent Situations , 1977 .

[50]  Rajesh P. N. Rao,et al.  The Jackknife – A review , 1974 .

[51]  F. Hampel Contributions to the theory of robust estimation , 1968 .

[52]  Arthur E. Hoerl,et al.  Application of ridge analysis to regression problems , 1962 .

[53]  M. H. Quenouille Approximate Tests of Correlation in Time‐Series , 1949 .

[54]  B. Chandrasekaran,et al.  The C957t Polymorphism in the Dopamine Receptor D2 (drd2) Gene Modulates Domain- General Category Learning (running Head: Drd2, Domain-general Category Learning) , 2022 .