Recent Editorials in this journal stressed the classicalparadigminclinicalepidemiologyofinsistingontesteretestevaluations for studies on diagnosis and prognosis [1] andspecifically prediction models [2]. Indeed, independentvalidationofpreviousresearchfindingsisanimportantscien-tific principle.Another recent debate was on the interpretation of thelack of external validation studies of published novel pre-diction models [3e5]. One issue is the role that validationshould have at the time of model development. Many re-searchers may be tempted to try to report some proof forexternal validity, that is, on discrimination and calibration,in independent samples with their publication that proposesa new prediction model. Major clinical journals currentlyseem to appreciate such reporting. Another issue is whetherexternal validation should be performed by different au-thors than those involved in the development of the predic-tion model [3,6]. We would like to comment on these andrelated key issues in the scientific basis of predictionmodeling.The recent review confirms that model developmentstudies are often relatively small for the complex chal-lenges posed by specifying the form of a prediction model(which predictors to include) and the estimation of predic-tor effects (overfit with standard estimation methods) [3].The median sample size was 445 subjects. The number ofevents is the limiting factor in this type of research andmay be far too low for reliable modeling [4]. In such smallsamples, internal validation is essential, and apparent per-formance estimates are severely optimistic (Fig. 1). Boot-strapping is the preferred approach for internal validationof prediction models [7e9]. A bootstrap procedure shouldinclude all modeling steps for an honest assessment ofmodel performance [10]. Specifically, any model selectionsteps, such as variable selection, need to be repeated perbootstrap sample if used.We recently confirmed that a split sample approach with50% held out leads to models with a suboptimal perfor-mance, that is, models with unstable and on average thesame performance as obtained with half the sample size[11]. We hence strongly advise against random split sampleapproaches in small development samples. Split sample ap-proaches can be used in very large samples, but again, weadvise against this practice because overfitting is no issueif sample size is so large that a split sample procedurecan be performed. Split sample approaches only work whennot needed.More relevant are attempts to obtain impressions ofexternal validity: do model predictions hold true indifferent settings, for example, in subjects from other cen-ters, or subjects seen more recently? Here, a nonrandomsplit can often be made in the development sample, forexample, by year of diagnosis. For example, we might vali-date a model on the most recent one-third of the sampleheld out from model development. Because the split is intime, this would qualify as a temporal external validation[6]. The disadvantages of a random split sample approachunfortunately equally hold here: a poorer model is devel-oped (on smaller sample size than the full developmentsample), and the validation findings are unstable (basedon a small sample size) [9].We make two propositions for validation at the time ofprediction model development (Fig. 2). First, we recom-mend an ‘‘internaleexternal’’ validation procedure. In thecontext of individual patient data meta-analysis (IPD-MA), internaleexternal cross-validation has been used toshow external validity of a prediction model [12,13].Inan MA context, the natural unit for splitting is by study.Every study is left out once, for validation of a model basedon the remaining studies. The final model is based on thepooled data set, which we label an ‘‘internallyeexternally
[1]
Ewout W Steyerberg,et al.
Internal and external validation of predictive models: a simulation study of bias and precision in small samples.
,
2003,
Journal of clinical epidemiology.
[2]
J. Habbema,et al.
Internal validation of predictive models: efficiency of some procedures for logistic regression analysis.
,
2001,
Journal of clinical epidemiology.
[3]
J. Knottnerus,et al.
Clinical prediction models are not being validated.
,
2015,
Journal of clinical epidemiology.
[4]
G. Bedogni,et al.
Clinical Prediction Models—a Practical Approach to Development, Validation and Updating
,
2009
.
[5]
Patrick Royston,et al.
Construction and validation of a prognostic model across several studies, with an application in superficial bladder cancer
,
2004,
Statistics in medicine.
[6]
Karel G M Moons,et al.
A new framework to enhance the interpretation of external validation studies of clinical prediction models.
,
2015,
Journal of clinical epidemiology.
[7]
John P A Ioannidis,et al.
Response to letter by Forike et al.: more rigorous, not less, external validation is needed.
,
2016,
Journal of clinical epidemiology.
[8]
J. Ioannidis,et al.
External validation of new risk prediction models is infrequent and reveals worse prognostic discrimination.
,
2015,
Journal of clinical epidemiology.
[9]
K. Covinsky,et al.
Assessing the Generalizability of Prognostic Information
,
1999,
Annals of Internal Medicine.
[10]
Sunil J Rao,et al.
Regression Modeling Strategies: With Applications to Linear Models, Logistic Regression, and Survival Analysis
,
2003
.
[11]
Gary S Collins,et al.
Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): Explanation and Elaboration
,
2015,
Annals of Internal Medicine.
[12]
G. Collins,et al.
Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPOD): The TRIPOD Statement
,
2015,
Annals of Internal Medicine.
[13]
A Cecile J W Janssens,et al.
External validation is only needed when prediction models are worth it (Letter commenting on: J Clin Epidemiol. 2015;68:25-34).
,
2016,
Journal of clinical epidemiology.
[14]
P. Austin,et al.
Events per variable (EPV) and the relative performance of different strategies for estimating the out-of-sample validity of logistic regression models
,
2014,
Statistical methods in medical research.
[15]
J. Knottnerus,et al.
Transferability/generalizability deserves more attention in 'retest' studies in Diagnosis and Prognosis.
,
2015,
Journal of clinical epidemiology.
[16]
C Legrand,et al.
Validation of prognostic indices using the frailty model
,
2009,
Lifetime data analysis.
[17]
Yvonne Vergouwe,et al.
Towards better clinical prediction models: seven steps for development and an ABCD for validation.
,
2014,
European heart journal.
[18]
J. Stockman,et al.
Prediction of MLH1 and MSH2 Mutations in Lynch Syndrome
,
2008
.
[19]
Juan Lu,et al.
Predicting Outcome after Traumatic Brain Injury: Development and International Validation of Prognostic Scores Based on Admission Characteristics
,
2008,
PLoS medicine.