Meta-learning framework with applications to zero-shot time-series forecasting

Can meta-learning discover generic ways of processing time-series (TS) from a diverse dataset so as to greatly improve generalization on new TS coming from different datasets? This work provides positive evidence to demonstrate this using a broad meta-learning framework which we show subsumes many existing meta-learning algorithms as specific cases. We further identify via theoretical analysis the meta-learning adaptation mechanisms within N-BEATS, a recent neural TS forecasting model. Our meta-learning theory predicts that N-BEATS iteratively generates a subset of its task-specific parameters based on a given TS input, thus gradually expanding the expressive power of the architecture on-the-fly. Our empirical results emphasize the importance of meta-learning for successful zero-shot forecasting to new sources of TS, supporting the claim that it is viable to train a neural network on a source TS dataset and deploy it on a different target TS dataset without retraining, resulting in performance that is at least as good as that of state-of-practice univariate forecasting models.

[1]  G. Yule On a Method of Investigating Periodicities in Disturbed Series, with Special Reference to Wolfer's Sunspot Numbers , 1927 .

[2]  Gilbert T. Walker,et al.  On Periodicity in Series of Related Terms , 1931 .

[3]  H. Harlow,et al.  The formation of learning sets. , 1949, Psychological review.

[4]  Peter R. Winters,et al.  Forecasting Sales by Exponentially Weighted Moving Averages , 1960 .

[5]  Robert L. Winkler,et al.  The accuracy of extrapolation (time series) methods: Results of a forecasting competition , 1982 .

[6]  R. Engle Autoregressive conditional heteroscedasticity with estimates of the variance of United Kingdom inflation , 1982 .

[7]  Neale WEATHER FORECASTING: MAGIC, ART, SCIENCE AND HYPNOSIS: Presidential Address delivered at the Fifth Annual Conference of the Meteorological Society of New Zealand, 10 October 1984 , 1985 .

[8]  Yoshua Bengio,et al.  Learning a synaptic learning rule , 1991, IJCNN-91-Seattle International Joint Conference on Neural Networks.

[9]  C. Chatfield,et al.  The M2-competition: A real-time judgmentally based forecasting study , 1993 .

[10]  H. C. Leung,et al.  Neural networks in supply chain management , 1995, Proceedings for Operating Research and the Management Sciences.

[11]  Michael R. Pearce,et al.  Retail sales force scheduling based on store traffic forecasting , 1998 .

[12]  K. Nikolopoulos,et al.  The theta model: a decomposition approach to forecasting , 2000 .

[13]  Spyros Makridakis,et al.  The M3-Competition: results, conclusions and implications , 2000 .

[14]  Kenneth B. Kahn How to Measure the Impact of a Forecast Error on an Enterprise , 2003 .

[15]  C. Holt Author's retrospective on ‘Forecasting seasonals and trends by exponentially weighted moving averages’ , 2004 .

[16]  Aris A. Syntetos,et al.  On the categorization of demand patterns , 2005, J. Oper. Res. Soc..

[17]  Rob J Hyndman,et al.  Another look at measures of forecast accuracy , 2006 .

[18]  Kaare Brandt Petersen,et al.  The Matrix Cookbook , 2006 .

[19]  Yoshua Bengio,et al.  On the Optimization of a Synaptic Learning Rule , 2007 .

[20]  Rob J Hyndman,et al.  Automatic Time Series Forecasting: The forecast Package for R , 2008 .

[21]  Jukka Korpela,et al.  Demand forecasting errors in industrial context: Measurement and impacts , 2009 .

[22]  Konstantinos Nikolopoulos,et al.  The Tourism Forecasting Competition , 2011 .

[23]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[24]  M. Rossetti,et al.  Exploring the Cost of Forecast Error in Inventory Systems , 2010 .

[25]  Yoshua Bengio,et al.  Deep Sparse Rectifier Neural Networks , 2011, AISTATS.

[26]  Rob J Hyndman,et al.  The value of feedback in forecasting competitions , 2011 .

[27]  Lee C. Baker,et al.  Winning methods for forecasting tourism time series , 2011 .

[28]  Jim Gao,et al.  Machine Learning Applications for Data Center Optimization , 2014 .

[29]  Nicolas Chapados,et al.  Retail store scheduling for profit , 2014, Eur. J. Oper. Res..

[30]  Carolo Friederico Gauss Theoria Motus Corporum Coelestium in Sectionibus Conicis Solem Ambientium , 2014 .

[31]  H. Mahoo,et al.  Integrating Indigenous Knowledge with Scientific Seasonal Forecasts for Climate Risk Management in Lushoto District in Tanzania , 2015 .

[32]  Pierre L'Ecuyer,et al.  Modeling and forecasting call center arrivals: A literature survey and a case study , 2015 .

[33]  Rob J Hyndman,et al.  Bagging exponential smoothing methods using STL decomposition and Box–Cox transformation , 2016 .

[34]  Joshua B. Tenenbaum,et al.  Building machines that learn and think like people , 2016, Behavioral and Brain Sciences.

[35]  Oriol Vinyals,et al.  Matching Networks for One Shot Learning , 2016, NIPS.

[36]  Inderjit S. Dhillon,et al.  Temporal Regularized Matrix Factorization for High-dimensional Time Series Prediction , 2016, NIPS.

[37]  A. Koehler,et al.  Models for optimising the theta method and their relationship to state space models , 2016 .

[38]  Hugo Larochelle,et al.  Optimization as a Model for Few-Shot Learning , 2016, ICLR.

[39]  Valentin Flunkert,et al.  DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks , 2017, International Journal of Forecasting.

[40]  Richard S. Zemel,et al.  Prototypical Networks for Few-shot Learning , 2017, NIPS.

[41]  Sergey Levine,et al.  Model-Agnostic Meta-Learning for Fast Adaptation of Deep Networks , 2017, ICML.

[42]  Hang Li,et al.  Meta-SGD: Learning to Learn Quickly for Few Shot Learning , 2017, ArXiv.

[43]  Germain Forestier,et al.  Transfer learning for time series classification , 2018, 2018 IEEE International Conference on Big Data (Big Data).

[44]  Aaron C. Courville,et al.  FiLM: Visual Reasoning with a General Conditioning Layer , 2017, AAAI.

[45]  Miriam A. M. Capretz,et al.  Transfer learning with seasonal and trend adjustment for cross-building energy forecasting , 2018 .

[46]  Evangelos Spiliotis,et al.  Statistical and Machine Learning forecasting methods: Concerns and ways forward , 2018, PloS one.

[47]  Matthias W. Seeger,et al.  Deep State Space Models for Time Series Forecasting , 2018, NeurIPS.

[48]  Evangelos Spiliotis,et al.  The M4 Competition: Results, findings, conclusion and way forward , 2018, International Journal of Forecasting.

[49]  Seungjin Choi,et al.  Gradient-Based Meta-Learning with Learned Layerwise Metric and Subspace , 2018, ICML.

[50]  Alexandre Lacoste,et al.  TADAM: Task dependent adaptive metric for improved few-shot learning , 2018, NeurIPS.

[51]  Razvan Pascanu,et al.  Meta-Learning with Latent Embedding Optimization , 2018, ICLR.

[52]  Ruslan Salakhutdinov,et al.  Multiple Futures Prediction , 2019, NeurIPS.

[53]  Syama Sundar Rangapuram,et al.  GluonTS: Probabilistic Time Series Models in Python , 2019, ArXiv.

[54]  Tim Januschowski,et al.  Deep Factors for Forecasting , 2019, ICML.

[55]  Evangelos Spiliotis,et al.  Forecasting with a hybrid method utilizing data smoothing, a variation of the Theta method and shrinkage of seasonal factors , 2018, International Journal of Production Economics.

[56]  Ibrahim Demir,et al.  Decentralized Flood Forecasting Using Deep Neural Networks , 2019, ArXiv.

[57]  Ratnesh Sharma,et al.  Energy Predictive Models with Limited Data using Transfer Learning , 2019, e-Energy.

[58]  Nicolas Chapados,et al.  N-BEATS: Neural basis expansion analysis for interpretable time series forecasting , 2019, ICLR.

[59]  Ahmet Murat Ozbayoglu,et al.  Financial Time Series Forecasting with Deep Learning : A Systematic Literature Review: 2005-2019 , 2019, Appl. Soft Comput..

[60]  Slawek Smyl,et al.  A hybrid method of exponential smoothing and recurrent neural networks for time series forecasting , 2020, International Journal of Forecasting.

[61]  Rob J. Hyndman,et al.  FFORMA: Feature-based forecast model averaging , 2020, International Journal of Forecasting.

[62]  Oriol Vinyals,et al.  Rapid Learning or Feature Reuse? Towards Understanding the Effectiveness of MAML , 2019, ICLR.