A better measure of relative prediction accuracy for model selection and model estimation

Surveys show that the mean absolute percentage error (MAPE) is the most widely used measure of prediction accuracy in businesses and organizations. It is, however, biased: when used to select among competing prediction methods it systematically selects those whose predictions are too low. This has not been widely discussed and so is not generally known among practitioners. We explain why this happens. We investigate an alternative relative accuracy measure which avoids this bias: the log of the accuracy ratio, that is, log (prediction/actual). Relative accuracy is particularly relevant if the scatter in the data grows as the value of the variable grows (heteroscedasticity). We demonstrate using simulations that for heteroscedastic data (modelled by a multiplicative error factor) the proposed metric is far superior to MAPE for model selection. Another use for accuracy measures is in fitting parameters to prediction models. Minimum MAPE models do not predict a simple statistic and so theoretical analysis is limited. We prove that when the proposed metric is used instead, the resulting least squares regression model predicts the geometric mean. This important property allows its theoretical properties to be understood.

[1]  Barbara A. Kitchenham,et al.  A Further Empirical Investigation of the Relationship Between MRE and Project Size , 2003, Empirical Software Engineering.

[2]  Ingunn Myrtveit,et al.  Validity and reliability of evaluation procedures in comparative studies of effort prediction models , 2011, Empirical Software Engineering.

[3]  Len Tashman,et al.  Percentage Error: What Denominator? , 2009 .

[4]  Barbara A. Kitchenham,et al.  An empirical validation of the relationship between the magnitude of relative error and project size , 2002, Proceedings Eighth IEEE Symposium on Software Metrics.

[5]  Xiangzhu Gao,et al.  Assessing Software Cost Estimation Models: criteria for accuracy, consistency and regression , 1997, Australas. J. Inf. Syst..

[6]  T. Gneiting Making and Evaluating Point Forecasts , 2009, 0912.0902.

[7]  Heungsun Park,et al.  A Shrinked Forecast in Stationary Processes Favouring Percentage Error , 2006 .

[8]  Y. Miyazaki,et al.  Robust regression for developing software estimation models , 1994, J. Syst. Softw..

[9]  N. Nakagawa,et al.  Method to estimate parameter values in software prediction models , 1991 .

[10]  P. Goodwin,et al.  On the asymmetry of the symmetric MAPE , 1999 .

[11]  Barbara A. Kitchenham,et al.  A Simulation Study of the Model Evaluation Criterion MMRE , 2003, IEEE Trans. Software Eng..

[12]  Stephan Kolassa,et al.  Percentage Errors Can Ruin Your Day (and Rolling the Dice Shows How) , 2011 .

[13]  Stephen G. MacDonell,et al.  Evaluating prediction systems in software project estimation , 2012, Inf. Softw. Technol..

[14]  Christopher J. Lokan,et al.  What should you optimize when building an estimation model? , 2005, 11th IEEE International Software Metrics Symposium (METRICS'05).

[15]  Stephen G. MacDonell,et al.  What accuracy statistics really measure , 2001, IEE Proc. Softw..

[16]  Gautam Appa,et al.  On L1 and Chebyshev estimation , 1973, Math. Program..

[17]  Klaus Schrape,et al.  The business of forecasting , 2001 .

[18]  Spyros Makridakis,et al.  Accuracy measures: theoretical and practical concerns☆ , 1993 .

[19]  Stephen G. MacDonell,et al.  Assessing prediction systems , 1999 .

[20]  Fred L. Collopy,et al.  Error Measures for Generalizing About Forecasting Methods: Empirical Comparisons , 1992 .

[21]  E. Stensrud,et al.  MRE and Heteroscedasticity : An Empirical Validation of the Assumption of Homoscedasticity of the Magnitude of Relative Error , 2001 .

[22]  George Bernard Shaw,et al.  LONG-RANGE FORECASTING From Crystal Ball to Computer , 2010 .

[23]  Ingunn Myrtveit,et al.  Reliability and validity in comparative studies of software prediction models , 2005, IEEE Transactions on Software Engineering.