Twin‐analysis verification: A new verification approach to alleviate pitfalls of own‐analysis verification

In operational numerical weather prediction (NWP), forecast verification against analysis from the same experiment is part of the standard evaluation practice. This “own‐analysis” verification is beneficial in providing complete and uniform spatial coverage but is known to suffer from overestimated skill scores when applied to short‐range forecasts due to the inevitable positive correlation between the forecast and analysis errors. To alleviate this problem, a new approach termed “twin‐analysis” verification is proposed. In this approach, the forecasts are verified against “twin analyses” that are produced by running an independent cycle using the same NWP system as the one used to produce the forecasts, but initializing the cycle using an independent first guess at the very beginning. In this way the error correlation between the forecasts and analyses is reduced, if not completely removed, while preserving the statistical properties of the analyses, allowing for a clearer interpretation of verification. This note reports the results of comparison between “twin‐analysis” and “own‐analysis” verification scores obtained for the Japan Meteorological Agency's global NWP system. It is reported that the two scores disagree up to 48 hr or even longer depending on the regions and elements, which highlights the reliability limit of “own‐analysis” verification. The proposed method does not completely remove the error correlation issue and also requires high computational costs; a practical use of the proposed approach in an operational environment is also discussed.

[1]  R. Todling,et al.  Evaluation of adjoint‐based observation impacts as a function of forecast length using an Observing System Simulation Experiment , 2020, Quarterly Journal of the Royal Meteorological Society.

[2]  Niels Bormann,et al.  Use and impact of Arctic observations in the ECMWF Numerical Weather Prediction system , 2019, Quarterly Journal of the Royal Meteorological Society.

[3]  Shunji Kotsuki,et al.  On the properties of ensemble forecast sensitivity to observations , 2019, Quarterly Journal of the Royal Meteorological Society.

[4]  M. Buehner,et al.  Using the hybrid gain algorithm to sample data assimilation uncertainty , 2018, Quarterly Journal of the Royal Meteorological Society.

[5]  Alan J. Geer,et al.  Significance of changes in medium-range forecast scores , 2016 .

[6]  Massimo Bonavita,et al.  On the use of EDA background error variances in the ECMWF 4D‐Var , 2012 .

[7]  P. Bauer,et al.  Direct 4D‐Var assimilation of all‐sky radiances. Part II: Assessment , 2010 .

[8]  Peter Bauer,et al.  Direct 4D‐Var assimilation of all‐sky radiances. Part I: Implementation , 2010 .

[9]  Barbara G. Brown,et al.  Forecast verification: current status and future directions , 2008 .

[10]  T. Palmer,et al.  Stochastic representation of model uncertainties in the ECMWF ensemble prediction system , 2007 .

[11]  F. Bouttier,et al.  Observing‐system experiments in the ECMWF 4D‐Var data assimilation system , 2001 .

[12]  M. Fisher,et al.  Background Error Covariance Modelling , 2003 .