Temporal Validated Meta-Learning for Long-Term Forecasting of Chaotic Time Series Using Monte Carlo Cross-Validation

Forecasting long-term values of chaotic time series is a difficult task, but it is required in several domains such as economy, medicine and astronomy. State-of-the-art works agree that the best accuracy can be obtained combining forecasting models. However, selecting the appropriate models and the best way to combine them is an open problem. Some researchers have been focusing on using prior knowledge of the performance of the models for combining them. A way to do so is by meta-learning, which is the process of automatically learning from tasks and models showing the best performances. Nevertheless, meta-learning in time series impose no trivial challenges; some requirements are to search the best model, to validate estimations, and even to develop new meta-learning methods. The new methods would consider performance variances of the models over time. This research addresses the meta-learning problem of how to select and combine models using different parts of the prediction horizon. Our strategy, called “Temporal Validated Combination” (TVC), consists of splitting the prediction horizon into three parts: short, medium and long-term windows. Next, for each window, we extract knowledge about what model has the best performance. This knowledge extraction uses a Monte Carlo cross-validation process. Using this, we are able to improve the long-term prediction using different models for each prediction window. The results reported in this chapter show that TVC obtained an average improvement of 1 % in the prediction of 56 points of the NN5 time series, when compared to a combination of best models based on a simple average. NARX neural networks and ARIMA models were used for building the predictors and the SMAPE metric was used for measuring performances.

[1]  Amir F. Atiya,et al.  A review and comparison of strategies for multi-step ahead time series forecasting based on the NN5 forecasting competition , 2011, Expert Syst. Appl..

[2]  Hava T. Siegelmann,et al.  Computational capabilities of recurrent NARX neural networks , 1997, IEEE Trans. Syst. Man Cybern. Part B.

[3]  Xiaozhe Wang,et al.  Rule induction for forecasting method selection: Meta-learning the characteristics of univariate time series , 2009, Neurocomputing.

[4]  Zheng Rong Yang,et al.  Evolving Stacked Time Series Predictors with Multiple Window Scales and Sampling Gaps , 2004, Neural Processing Letters.

[5]  R. Dennis Cook,et al.  Cross-Validation of Regression Models , 1984 .

[6]  Simon Haykin,et al.  Neural Networks: A Comprehensive Foundation , 1998 .

[7]  Henry D. I. Abarbanel,et al.  Analysis of Observed Chaotic Data , 1995 .

[8]  Rob J Hyndman,et al.  25 years of time series forecasting , 2006 .

[9]  Yi-Zeng Liang,et al.  Monte Carlo cross‐validation for selecting a model and estimating the prediction error in multivariate calibration , 2004 .

[10]  J. Nazuno Haykin, Simon. Neural networks: A comprehensive foundation, Prentice Hall, Inc. Segunda Edición, 1999 , 2000 .

[11]  David H. Wolpert,et al.  Stacked generalization , 1992, Neural Networks.

[12]  Bogdan Gabrys,et al.  Meta-learning for time series forecasting and forecast combination , 2010, Neurocomputing.

[13]  Rigoberto Fonseca,et al.  An assessment of ten-fold and Monte Carlo cross validations for time series forecasting , 2013, 2013 10th International Conference on Electrical Engineering, Computing Science and Automatic Control (CCE).

[14]  J. Shao Linear Model Selection by Cross-validation , 1993 .

[15]  I. J. Leontaritis,et al.  Input-output parametric models for non-linear systems Part II: stochastic non-linear systems , 1985 .

[16]  Robert J. Genetski,et al.  Long-Range Forecasting: From Crystal Ball to Computer , 1981 .

[17]  Sven F. Crone,et al.  Advances in forecasting with neural networks? Empirical evidence from the NN3 competition on time series prediction , 2011 .

[18]  Johan A. K. Suykens,et al.  Load forecasting using a multivariate meta-learning system , 2013, Expert Syst. Appl..

[19]  M. Rosenstein,et al.  A practical method for calculating largest Lyapunov exponents from small data sets , 1993 .

[20]  L. Glass,et al.  Oscillation and chaos in physiological control systems. , 1977, Science.

[21]  L. Cao Practical method for determining the minimum embedding dimension of a scalar time series , 1997 .

[22]  H. Kantz,et al.  Nonlinear time series analysis , 1997 .

[23]  J. Ross Macdonald Review and Reveal the Reviewers , 1996 .

[24]  Bernard Zenko,et al.  Is Combining Classifiers with Stacking Better than Selecting the Best One? , 2004, Machine Learning.

[25]  H. Abarbanel,et al.  Determining embedding dimension for phase-space reconstruction using a geometrical construction. , 1992, Physical review. A, Atomic, molecular, and optical physics.

[26]  Ron Kohavi,et al.  Wrappers for Feature Subset Selection , 1997, Artif. Intell..

[27]  Gwilym M. Jenkins,et al.  Time series analysis, forecasting and control , 1972 .

[28]  Eugen Diaconescu,et al.  The use of NARX neural networks to predict chaotic time series , 2008 .

[29]  Richard A. Davis,et al.  Time Series: Theory and Methods , 2013 .

[30]  Richard A. Davis,et al.  Time Series: Theory and Methods (2Nd Edn) , 1993 .

[31]  Holger Kantz,et al.  Practical implementation of nonlinear time series methods: The TISEAN package. , 1998, Chaos.

[32]  Spyros Makridakis,et al.  The M3-Competition: results, conclusions and implications , 2000 .

[33]  C. Dhanya,et al.  Nonlinear ensemble prediction of chaotic daily rainfall , 2010 .