Time Series Forecasting for Self-Aware Systems

Modern distributed systems and Internet-of-Things applications are governed by fast living and changing requirements. Moreover, they have to struggle with huge amounts of data that they create or have to process. To improve the self-awareness of such systems and enable proactive and autonomous decisions, reliable time series forecasting methods are required. However, selecting a suitable forecasting method for a given scenario is a challenging task. According to the “No-Free-Lunch Theorem,” there is no general forecasting method that always performs best. Thus, manual feature engineering remains to be a mandatory expert task to avoid trial and error. Furthermore, determining the expected time-to-result of existing forecasting methods is a challenge. In this article, we extensively assess the state-of-the-art in time series forecasting. We compare existing methods and discuss the issues that have to be addressed to enable their use in a self-aware computing context. To address these issues, we present a step-by-step approach to fully automate the feature engineering and forecasting process. Then, following the principles from benchmarking, we establish a level-playing field for evaluating the accuracy and time-to-result of automated forecasting methods for a broad set of application scenarios. We provide results of a benchmarking competition to guide in selecting and appropriately using existing forecasting methods for a given self-aware computing context. Finally, we present a case study in the area of self-aware data-center resource management to exemplify the benefits of fully automated learning and reasoning processes on time series data.

[1]  Nikolas Herbst,et al.  Autonomic Forecasting Method Selection: Examination and Ways Ahead , 2019, 2019 IEEE International Conference on Autonomic Computing (ICAC).

[2]  Víctor M. Guerrero Time‐series analysis supported by power transformations , 1993 .

[3]  K. Nikolopoulos,et al.  The theta model: a decomposition approach to forecasting , 2000 .

[4]  P. Young,et al.  Time series analysis, forecasting and control , 1972, IEEE Transactions on Automatic Control.

[5]  WangXiaozhe,et al.  Rule induction for forecasting method selection , 2009 .

[6]  Estela Bee Dagum,et al.  Seasonal Adjustment Methods and Real Time Trend-Cycle Estimation , 2016 .

[7]  Motoaki Kawanabe,et al.  Machine Learning in Non-Stationary Environments - Introduction to Covariate Shift Adaptation , 2012, Adaptive computation and machine learning.

[8]  R. K. Agrawal,et al.  An Introductory Study on Time Series Modeling and Forecasting , 2013, ArXiv.

[9]  C. Holt Author's retrospective on ‘Forecasting seasonals and trends by exponentially weighted moving averages’ , 2004 .

[10]  Ada Diaconescu,et al.  Reference Scenarios for Self-aware Computing , 2017, Self-Aware Computing Systems.

[11]  D. Cox,et al.  An Analysis of Transformations , 1964 .

[12]  Xizhao Wang,et al.  Discretization of continuous-valued attributes in decision tree generation , 2010, 2010 International Conference on Machine Learning and Cybernetics.

[13]  Derek W. Bunn,et al.  Review of guidelines for the use of combined forecasts , 2000, Eur. J. Oper. Res..

[14]  Jörg Hähner,et al.  Local ensemble weighting in the context of time series forecasting using XCSF , 2016, 2016 IEEE Symposium Series on Computational Intelligence (SSCI).

[15]  Lars Grunske,et al.  Online Workload Forecasting , 2017, Self-Aware Computing Systems.

[16]  Irma J. Terpenning,et al.  STL : A Seasonal-Trend Decomposition Procedure Based on Loess , 1990 .

[17]  Marko Grobelnik,et al.  Feature Selection Using Support Vector Machines , 2002 .

[18]  Steven C. Hillmer,et al.  Issues Involved With the Seasonal Adjustment of Economic Time Series , 1984 .

[19]  Ron Kohavi,et al.  Wrappers for Feature Subset Selection , 1997, Artif. Intell..

[20]  Lawrence D. Jackel,et al.  Backpropagation Applied to Handwritten Zip Code Recognition , 1989, Neural Computation.

[21]  J. M. Bates,et al.  The Combination of Forecasts , 1969 .

[22]  Jeffrey O. Kephart,et al.  The Vision of Autonomic Computing , 2003, Computer.

[23]  Samuel Kounev,et al.  A Model-Based Approach to Designing Self-Aware IT Systems and Infrastructures , 2016, Computer.

[24]  Rami Bahsoon,et al.  A Survey of Self-Awareness and Its Application in Computing Systems , 2011, 2011 Fifth IEEE Conference on Self-Adaptive and Self-Organizing Systems Workshops.

[25]  Vladimir Vapnik,et al.  The Nature of Statistical Learning , 1995 .

[26]  J. Doran,et al.  Experiments with the Graph Traverser program , 1966, Proceedings of the Royal Society of London. Series A. Mathematical and Physical Sciences.

[27]  Guoqiang Peter Zhang,et al.  Time series forecasting using a hybrid ARIMA and neural network model , 2003, Neurocomputing.

[28]  C. D. Gelatt,et al.  Optimization by Simulated Annealing , 1983, Science.

[29]  Skander Soltani,et al.  On the use of the wavelet decomposition for time series prediction , 2002, ESANN.

[30]  L. Myers,et al.  Spearman Correlation Coefficients, Differences between , 2004 .

[31]  Samuel Kounev,et al.  Model-driven Algorithms and Architectures for Self-Aware Computing Systems (Dagstuhl Seminar 15041) , 2015, Dagstuhl Reports.

[32]  Pedro M. Domingos A few useful things to know about machine learning , 2012, Commun. ACM.

[33]  Jianhua Zhang,et al.  A hybrid forecasting model with parameter optimization for short-term load forecasting of micro-grids , 2014 .

[34]  竹安 数博,et al.  Time series analysis and its applications , 2007 .

[35]  Isabelle Guyon,et al.  An Introduction to Variable and Feature Selection , 2003, J. Mach. Learn. Res..

[36]  Benjamin Letham,et al.  Forecasting at Scale , 2018, PeerJ Prepr..

[37]  J. Yosinski,et al.  Time-series Extreme Event Forecasting with Neural Networks at Uber , 2017 .

[38]  Evgenia Smirni,et al.  Burstiness in Multi-tier Applications: Symptoms, Causes, and New Models , 2008, Middleware.

[39]  Chaker El Amrani,et al.  Sequence to Sequence Weather Forecasting with Long Short-Term Memory Recurrent Neural Networks , 2016 .

[40]  Rob J Hyndman,et al.  Forecasting Time Series With Complex Seasonal Patterns Using Exponential Smoothing , 2011 .

[41]  Rob J Hyndman,et al.  A state space framework for automatic forecasting using exponential smoothing methods , 2002 .

[42]  Ping-Feng Pai,et al.  A hybrid ARIMA and support vector machines model in stock price forecasting , 2005 .

[43]  Chris Chatfield,et al.  What is the ‘best’ method of forecasting? , 1988 .

[44]  Bogdan Gabrys,et al.  Meta-learning for time series forecasting and forecast combination , 2010, Neurocomputing.

[45]  Ada Diaconescu,et al.  Towards a Framework for the Levels and Aspects of Self-aware Computing Systems , 2017, Self-Aware Computing Systems.

[46]  A. Kraskov,et al.  Estimating mutual information. , 2003, Physical review. E, Statistical, nonlinear, and soft matter physics.

[47]  Stefano Nembrini,et al.  The revival of the Gini importance? , 2018, Bioinform..

[48]  Bernhard E. Boser,et al.  A training algorithm for optimal margin classifiers , 1992, COLT '92.

[49]  M. Shcherbakov,et al.  A Survey of Forecast Error Measures , 2013 .

[50]  Fotios Petropoulos,et al.  forecast: Forecasting functions for time series and linear models , 2018 .

[51]  M. Kendall Statistical Methods for Research Workers , 1937, Nature.

[52]  Xiaozhe Wang,et al.  Rule induction for forecasting method selection: Meta-learning the characteristics of univariate time series , 2009, Neurocomputing.

[53]  Samuel Kounev,et al.  Telescope: A Hybrid Forecast Method for Univariate Time Series , 2017 .

[54]  Peter R. Winters,et al.  Forecasting Sales by Exponentially Weighted Moving Averages , 1960 .

[55]  R. Forthofer,et al.  Rank Correlation Methods , 1981 .

[56]  Rob J Hyndman,et al.  Another look at measures of forecast accuracy , 2006 .

[57]  Heiga Zen,et al.  WaveNet: A Generative Model for Raw Audio , 2016, SSW.

[58]  Rami Bahsoon,et al.  Architectural Aspects of Self-Aware and Self-Expressive Computing Systems: From Psychology to Engineering , 2015, Computer.

[59]  Tianqi Chen,et al.  XGBoost: A Scalable Tree Boosting System , 2016, KDD.

[60]  Rob J Hyndman,et al.  Unmasking the Theta Method , 2003 .

[61]  VARUN CHANDOLA,et al.  Anomaly detection: A survey , 2009, CSUR.

[62]  Ada Diaconescu,et al.  The Notion of Self-aware Computing , 2017, Self-Aware Computing Systems.

[63]  Leo Breiman,et al.  Random Forests , 2001, Machine Learning.

[64]  Jonathan R. M. Hosking,et al.  Partitioning Nominal Attributes in Decision Trees , 1999, Data Mining and Knowledge Discovery.

[65]  Jeremy Pitt,et al.  The Computer After Me: Awareness and Self-Awareness in Autonomic Systems , 2014 .

[66]  Alexander J. Smola,et al.  Support Vector Regression Machines , 1996, NIPS.

[67]  Johan Tordsson,et al.  PEAS , 2016, ACM Trans. Model. Perform. Evaluation Comput. Syst..

[68]  George Athanasopoulos,et al.  Forecasting: principles and practice , 2013 .

[69]  Ajay Mohindra,et al.  Dynamic Scaling of Web Applications in a Virtualized Cloud Computing Environment , 2009, 2009 IEEE International Conference on e-Business Engineering.

[70]  M. G. Kendall,et al.  A Study in the Analysis of Stationary Time-Series. , 1955 .

[71]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[72]  Fred Collopy,et al.  Rule-Based Forecasting: Development and Validation of an Expert Systems Approach to Combining Time Series Extrapolations , 1992 .

[73]  Jóakim von Kistowski,et al.  How to Build a Benchmark , 2015, ICPE.

[74]  Wei-Yin Loh,et al.  Classification and regression trees , 2011, WIREs Data Mining Knowl. Discov..

[75]  David H. Wolpert,et al.  No free lunch theorems for optimization , 1997, IEEE Trans. Evol. Comput..

[76]  Wheeler Ruml,et al.  A Comparison of Greedy Search Algorithms , 2010, SOCS.