Principles and Algorithms for Forecasting Groups of Time Series: Locality and Globality

Forecasting groups of time series is of increasing practical importance, e.g. forecasting the demand for multiple products offered by a retailer or server loads within a data center. The local approach to this problem considers each time series separately and fits a function or model to each series. The global approach fits a single function to all series. For groups of similar time series, global methods outperform the more established local methods. However, recent results show good performance of global models even in heterogeneous datasets. This suggests a more general applicability of global methods, potentially leading to more accurate tools and new scenarios to study. Formalizing the setting of forecasting a set of time series with local and global methods, we provide the following contributions: 1) Global methods are not more restrictive than local methods, both can produce the same forecasts without any assumptions about similarity of the series. Global models can succeed in a wider range of problems than previously thought. 2) Basic generalization bounds for local and global algorithms. The complexity of local methods grows with the size of the set while it remains constant for global methods. In large datasets, a global algorithm can afford to be quite complex and still benefit from better generalization. These bounds serve to clarify and support recent experimental results in the field, and guide the design of new algorithms. For the class of autoregressive models, this implies that global models can have much larger memory than local methods. 3) In an extensive empirical study, purposely naive algorithms derived from these principles, such as global linear models or deep networks result in superior accuracy. In particular, global linear models can provide competitive accuracy with two orders of magnitude fewer parameters than local methods.

[1]  Vitaly Kuznetsov Forecasting Non-Stationary Time Series : From Theory to Algorithms , 2014 .

[2]  Rob J. Hyndman,et al.  FFORMA: Feature-based forecast model averaging , 2020, International Journal of Forecasting.

[3]  Mehmed Kantardzic,et al.  Learning from Data , 2011 .

[4]  Robert Fildes,et al.  Forecasting third-party mobile payments with implications for customer flow prediction , 2020, International Journal of Forecasting.

[5]  Srihari Jaganathan,et al.  A combination-based forecasting method for the M4-competition , 2020 .

[6]  Tomasz Kornuta,et al.  Learning beyond simulated physics , 2018 .

[7]  Cosma Rohilla Shalizi,et al.  Nonparametric Risk Bounds for Time-Series Forecasting , 2012, J. Mach. Learn. Res..

[8]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[9]  Sander M. Bohte,et al.  Dilated Convolutional Neural Networks for Time Series Forecasting , 2018 .

[10]  Nicolas Chapados,et al.  N-BEATS: Neural basis expansion analysis for interpretable time series forecasting , 2019, ICLR.

[11]  Rob J Hyndman,et al.  Automatic Time Series Forecasting: The forecast Package for R , 2008 .

[12]  W. Hoeffding Probability Inequalities for sums of Bounded Random Variables , 1963 .

[13]  Fotios Petropoulos,et al.  forecast: Forecasting functions for time series and linear models , 2018 .

[14]  Serena Ng,et al.  Working Paper Series , 2019 .

[15]  Juan R. Trapero,et al.  On the identification of sales forecasting models in the presence of promotions , 2015, J. Oper. Res. Soc..

[16]  Evangelos Spiliotis,et al.  Statistical and Machine Learning forecasting methods: Concerns and ways forward , 2018, PloS one.

[17]  Valentin Flunkert,et al.  DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks , 2017, International Journal of Forecasting.

[18]  Mehryar Mohri,et al.  Generalization bounds for non-stationary mixing processes , 2016, Machine Learning.

[19]  J. Yosinski,et al.  Time-series Extreme Event Forecasting with Neural Networks at Uber , 2017 .

[20]  Evangelos Spiliotis,et al.  The M4 Competition: 100,000 time series and 61 forecasting methods , 2020 .

[21]  Slawek Smyl,et al.  A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting , 2020, International Journal of Forecasting.

[22]  Agata Chorowska,et al.  Weighted ensemble of statistical models , 2018, 1811.07761.

[23]  Heiga Zen,et al.  WaveNet: A Generative Model for Raw Audio , 2016, SSW.

[24]  Mark Landry,et al.  Probabilistic gradient boosting machines for GEFCom2014 wind forecasting , 2016 .

[25]  Christoph Bergmeir,et al.  Recurrent Neural Networks for Time Series Forecasting: Current Status and Future Directions , 2019, ArXiv.

[26]  Martin Stepnicka,et al.  On the results and observations of the time series forecasting competition CIF 2016 , 2017, 2017 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE).

[27]  Quoc V. Le,et al.  Learning Longer-term Dependencies in RNNs with Auxiliary Losses , 2018, ICML.

[28]  Amir F. Atiya,et al.  A review and comparison of strategies for multi-step ahead time series forecasting based on the NN5 forecasting competition , 2011, Expert Syst. Appl..

[29]  Christoph Bergmeir,et al.  LSTM-MSNet: Leveraging Forecasts on Sets of Related Time Series with Multiple Seasonal Patterns , 2020, IEEE transactions on neural networks and learning systems.

[30]  Christoph Bergmeir,et al.  Forecasting across time series databases using recurrent neural networks on groups of similar series: A clustering approach , 2017, Expert Syst. Appl..

[31]  Yoram Singer,et al.  Train faster, generalize better: Stability of stochastic gradient descent , 2015, ICML.

[32]  Syama Sundar Rangapuram,et al.  Probabilistic Forecasting with Spline Quantile Function RNNs , 2019, AISTATS.

[33]  Rob J. Hyndman,et al.  A note on the validity of cross-validation for evaluating autoregressive time series prediction , 2018, Comput. Stat. Data Anal..

[34]  Kim-Han Thung,et al.  A brief review on multi-task learning , 2018, Multimedia Tools and Applications.

[35]  George T. Duncan,et al.  Forecasting Analogous Time Series , 2001 .

[36]  Vitaly Kuznetsov,et al.  Foundations of Sequence-to-Sequence Modeling for Time Series , 2018, AISTATS.