Forecast evaluation with imperfect observations and imperfect models.

The field of statistics has become one of the mathematical foundations in forecast evaluations studies, especially in regard to computing scoring rules. The classical paradigm of proper scoring rules is to discriminate between two different forecasts by comparing them with observations. The probability density function of the observed record is assumed to be perfect as a verification benchmark. In practice, however, observations are almost always tainted by errors. These may be due to homogenization problems, instrumental deficiencies, the need for indirect reconstructions from other sources (e.g., radar data), model errors in gridded products like reanalysis, or any other data-recording issues. If the yardstick used to compare forecasts is imprecise, one can wonder whether such types of errors may or may not have a strong influence on decisions based on classical scoring rules. Building on the recent work of Ferro (2017), we propose a new scoring rule scheme in the context of models that incorporate errors of the verification data, we compare it to existing methods, and applied it to various setups, mainly a Gaussian additive noise model and a gamma multiplicative noise model. In addition, we frame the problem of error verification in datasets as scoring a model that jointly couples forecasts and observation distributions. This is strongly connected to the so-called error-in-variables models in statistics.

[1]  A. H. Murphy,et al.  A General Framework for Forecast Verification , 1987 .

[2]  A. Raftery,et al.  Strictly Proper Scoring Rules, Prediction, and Estimation , 2007 .

[3]  Philippe Naveau,et al.  Optimal fingerprinting under multiple sources of uncertainty , 2014 .

[4]  Manfred Dorninger,et al.  Quantifying verification uncertainty by reference data variation , 2012 .

[5]  Anton H. Westveld,et al.  Calibrated Probabilistic Forecasting Using Ensemble Model Output Statistics and Minimum CRPS Estimation , 2005 .

[6]  Tilmann Gneiting,et al.  Of quantiles and expectiles: consistent scoring functions, Choquet representations and forecast rankings , 2015, 1503.08195.

[7]  R. Buizza,et al.  The Skill of Probabilistic Precipitation Forecasts under Observational Uncertainties within the Generalized Likelihood Uncertainty Estimation Framework for Hydrological Applications , 2009 .

[8]  Neill E. Bowler Accounting for the effect of observation errors on verification of MOGREPS , 2008 .

[9]  T. Pham-Gia,et al.  Trace of the Wishart Matrix and Applications , 2015 .

[10]  C. Stein Estimation of the Mean of a Multivariate Normal Distribution , 1981 .

[11]  Pierre Pinson,et al.  Verification of the ECMWF ensemble forecasts of wind speed against analyses and observations , 2012 .

[12]  Olivier Mestre,et al.  Calibrated Ensemble Forecasts Using Quantile Regression Forests and Ensemble Model Output Statistics , 2016 .

[13]  I. Jolliffe,et al.  Forecast verification : a practitioner's guide in atmospheric science , 2011 .

[14]  Sarah L. Dance,et al.  Estimating correlated observation error statistics using an ensemble transform Kalman filter , 2014 .

[15]  S. Weijs,et al.  Accounting for Observational Uncertainty in Forecast Verification: An Information-Theoretical View on Forecasts, Observations, and Truth , 2011 .

[16]  Neill E. Bowler,et al.  Explicitly Accounting for Observation Error in Categorical Verification of Forecasts , 2006 .

[17]  A. M. Mathai,et al.  Further results on the trace of a noncentral wishart matrix , 1982 .

[18]  Alexander Kukush,et al.  Measurement Error Models , 2011, International Encyclopedia of Statistical Science.

[19]  R. Daley Estimating observation error statistics for atmospheric data assimilation , 1993 .

[20]  F. Zwiers,et al.  A new statistical approach to climate change detection and attribution , 2016, Climate Dynamics.

[21]  Retracted and replaced:Impact of observational error on the validation of ensemble prediction systems , 2008 .

[22]  Christopher A. T. Ferro Measuring forecast performance in the presence of observation error , 2017 .

[23]  Stefan Siegert,et al.  A Bayesian framework for verification and recalibration of ensemble forecasts: How uncertain is NAO predictability? , 2015, 1504.01933.

[24]  Michael Scheuerer,et al.  Probabilistic wind speed forecasting on a grid based on ensemble model output statistics , 2015, 1511.02001.

[25]  Nancy Nichols,et al.  On the representation error in data assimilation , 2018 .

[26]  T. Hamill Interpretation of Rank Histograms for Verifying Ensemble Forecasts , 2001 .

[27]  Jeffrey L. Anderson A Method for Producing and Evaluating Probabilistic Forecasts from Ensemble Model Integrations , 1996 .