Evaluating Forecasting Methods by Considering Different Accuracy Measures

Abstract Choosing the appropriate forecasting technique to employ is a challenging issue and requires a comprehensive analysis of empirical results. Recent research findings reveal that the performance evaluation of forecasting models depends on the accuracy measures adopted. Some methods indicate superior performance when error based metrics are used, while others perform better when precision values are adopted as accuracy measures. As scholars tend to use a smaller subset of accuracy metrics to assess the performance of forecasting models, there is a need for a concept of multiple accuracy dimensions to assure the robustness of evaluation. Therefore, the main purpose of this paper is to propose a decision making model that allows researchers to identify the superiority of a forecasting technique over another by considering several accuracy metrics concurrently. A multi-criteria decision analysis approach, namely the preference ranking organization method for enrichment evaluation (PROMETHEE), was adopted to solve this problem. Bayesian Networks, Artificial Neural Networks, SVMs, Logistic Regression, and several Rule and Tree-based forecasting approaches were included in the analysis. After introducing a detailed description of accuracy measures, the performance of the prediction models are evaluated using a chosen dataset from the UCI Machine Learning Repository.

[1]  Seref Sagiroglu,et al.  The development of intuitive knowledge classifier and the modeling of domain dependent data , 2013, Knowl. Based Syst..

[2]  C. Cleverdon On the Inverse Relationship of Recall and Precision. , 1972 .

[3]  Yi Peng,et al.  FAMCDM: A fusion approach of MCDM methods to rank multiclass classification algorithms , 2011 .

[4]  Philip S. Yu,et al.  Top 10 algorithms in data mining , 2007, Knowledge and Information Systems.

[5]  Mehryar Mohri,et al.  Confidence Intervals for the Area Under the ROC Curve , 2004, NIPS.

[6]  David Enke,et al.  A New Hybrid Approach For Forecasting Interest Rates , 2012, Complex Adaptive Systems.

[7]  Alexander Dekhtyar,et al.  Information Retrieval , 2018, Lecture Notes in Computer Science.

[8]  Bing Xu,et al.  Performance evaluation of competing forecasting models: A multidimensional framework based on MCDA , 2012, Expert Syst. Appl..

[9]  Mandana Rezaeiahari,et al.  AHP based Classification Algorithm Selection for Clinical Decision Support System Development , 2014, Complex Adaptive Systems.

[10]  Peter Loos,et al.  Determination of Rule Patterns in Complex Event Processing Using Machine Learning Techniques , 2015, Complex Adaptive Systems.

[11]  อนิรุธ สืบสิงห์,et al.  Data Mining Practical Machine Learning Tools and Techniques , 2014 .

[12]  David J. Hand,et al.  Measuring classifier performance: a coherent alternative to the area under the ROC curve , 2009, Machine Learning.

[13]  Spyros Makridakis,et al.  Accuracy measures: theoretical and practical concerns☆ , 1993 .

[14]  Sotiris B. Kotsiantis,et al.  Supervised Machine Learning: A Review of Classification Techniques , 2007, Informatica.

[15]  Kaoru Tone,et al.  Forecasting Models Evaluation using A Slacks-based Context-dependent DEA Framework , 2014 .

[16]  Ian H. Witten,et al.  The WEKA data mining software: an update , 2009, SKDD.

[17]  M. Pencina,et al.  Evaluating the added predictive ability of a new marker: From area under the ROC curve to reclassification and beyond , 2008, Statistics in medicine.

[18]  Essam Mahmoud,et al.  Accuracy in forecasting: A survey , 1984 .

[19]  Fred L. Collopy,et al.  Error Measures for Generalizing About Forecasting Methods: Empirical Comparisons , 1992 .

[20]  Guy Lapalme,et al.  A systematic analysis of performance measures for classification tasks , 2009, Inf. Process. Manag..

[21]  Theodor J. Stewart,et al.  Multiple Criteria Decision Analysis , 2001 .

[22]  Mark Goadrich,et al.  The relationship between Precision-Recall and ROC curves , 2006, ICML.

[23]  Cesar H. Comin,et al.  A Systematic Comparison of Supervised Classifiers , 2013, PloS one.

[24]  Robert P. W. Duin,et al.  Approximating the multiclass ROC by pairwise analysis , 2007, Pattern Recognit. Lett..

[25]  Tom Fawcett,et al.  Analysis and Visualization of Classifier Performance: Comparison under Imprecise Class and Cost Distributions , 1997, KDD.

[26]  P. Vincke,et al.  Note-A Preference Ranking Organisation Method: The PROMETHEE Method for Multiple Criteria Decision-Making , 1985 .

[27]  Reza Baradaran Kazemzadeh,et al.  PROMETHEE: A comprehensive literature review on methodologies and applications , 2010, Eur. J. Oper. Res..

[28]  Rob J Hyndman,et al.  Another look at measures of forecast accuracy , 2006 .

[29]  Mary Felkin Comparing Classification Results between N-ary and Binary Problems , 2007, Quality Measures in Data Mining.

[30]  Hocine Cherifi,et al.  Accuracy Measures for the Comparison of Classifiers , 2012, ICIT 2012.

[31]  Jacob Cohen,et al.  Weighted kappa: Nominal scale agreement provision for scaled disagreement or partial credit. , 1968 .