CDA-LSTM: an evolutionary convolution-based dual-attention LSTM for univariate time series prediction

Univariate time series forecasting is still an important but challenging task. Considering the wide application of temporal data, adaptive predictors are needed to study historical behavior and forecast future state in various scenarios. In this paper, inspired by human attention mechanism and decomposition and reconstruction framework, we proposed a convolution-based dual-stage attention (CDA) architecture combined with Long Short-Term Memory networks (LSTM) for univariate time series forecasting. Specifically, we first use the decomposition algorithm to generate derived variables from target series. Input variables are then fed into the CDA-LSTM machine for further forecasting. In the Encoder–Decoder phase, for the first stage, attention operation is combined with the LSTM acting as an encoder, which could adaptively learn the relevant derived series to the target. In the second stage, the temporal attention mechanism is integrated with decoder aiming to automatically select the relevant encoder hidden states across all time steps. A convolution phase is concatenated parallelly to the Encoder–Decoder phase to reuse the historical information of the target and extract the mutation features. The experimental results demonstrate the proposed method could be adopted as expert systems for forecasting in multiple scenarios, and the superiority is verified by comparing with twelve baseline models on ten datasets. The practicability of different decomposition algorithms and convolution architectures is also discussed by extensive experiment. Overall, our work carries a significant value not merely in adaptive modeling of deep learning in time series issues, but also in the field of univariate data processing and prediction.

[1]  Francisco C. Pereira,et al.  Combining time-series and textual data for taxi demand prediction in event areas: a deep learning approach , 2018, Inf. Fusion.

[2]  A. Murat Ozbayoglu,et al.  Algorithmic financial trading with deep convolutional neural networks: Time series to image conversion approach , 2018, Appl. Soft Comput..

[3]  Yi Liu,et al.  Hilbert-Huang Transform and the Application , 2020, 2020 IEEE International Conference on Artificial Intelligence and Information Systems (ICAIIS).

[4]  Qun Dai,et al.  Several novel evaluation measures for rank-based ensemble pruning with applications to time series prediction , 2015, Expert Syst. Appl..

[5]  Michael E. Fitzpatrick,et al.  Detecting anomalies in time series data via a deep learning algorithm combining wavelets, neural networks and Hilbert transform , 2017, Expert Syst. Appl..

[6]  Abolfazl Salami,et al.  A study of hybrid data selection method for a wavelet SVR mid-term load forecasting model , 2017, Neural Computing and Applications.

[7]  Yao Zhao,et al.  EA-LSTM: Evolutionary Attention-based LSTM for Time Series Prediction , 2018, Knowl. Based Syst..

[8]  Winston H. Hsu,et al.  Deep Multi-Kernel Convolutional LSTM Networks and an Attention-Based Mechanism for Videos , 2019, IEEE Transactions on Multimedia.

[9]  N. Huang,et al.  A study of the characteristics of white noise using the empirical mode decomposition method , 2004, Proceedings of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences.

[10]  Vladlen Koltun,et al.  An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.

[11]  Norden E. Huang,et al.  Ensemble Empirical Mode Decomposition: a Noise-Assisted Data Analysis Method , 2009, Adv. Data Sci. Adapt. Anal..

[12]  M. Steinhauser,et al.  A dual-stage two-phase model of selective attention. , 2010, Psychological review.

[13]  S. S. Shen,et al.  Applications of Hilbert–Huang transform to non‐stationary financial time series analysis , 2003 .

[14]  Garrison W. Cottrell,et al.  A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction , 2017, IJCAI.

[15]  G. Sudheer,et al.  A wavelet-nearest neighbor model for short-term load forecasting , 2015 .

[16]  Ponnuthurai Nagaratnam Suganthan,et al.  Empirical Mode Decomposition based ensemble deep learning for load demand time series forecasting , 2017, Appl. Soft Comput..

[17]  Alicia Troncoso Lora,et al.  A scalable approach based on deep learning for big data time series forecasting , 2018, Integr. Comput. Aided Eng..

[18]  Diego Cabrera,et al.  Bayesian approach and time series dimensionality reduction to LSTM-based model-building for fault diagnosis of a reciprocating compressor , 2020, Neurocomputing.

[19]  N. Huang,et al.  The empirical mode decomposition and the Hilbert spectrum for nonlinear and non-stationary time series analysis , 1998, Proceedings of the Royal Society of London. Series A: Mathematical, Physical and Engineering Sciences.

[20]  Yi-Ping Phoebe Chen,et al.  Hybrid deep learning and empirical mode decomposition model for time series applications , 2019, Expert Syst. Appl..

[21]  Ping-Huan Kuo,et al.  An Electricity Price Forecasting Model by Hybrid Structured Deep Neural Networks , 2018 .

[22]  Qian Zhang,et al.  Attention-based recurrent neural networks for accurate short-term and long-term dissolved oxygen prediction , 2019, Comput. Electron. Agric..

[23]  Qicai Wang,et al.  A Fusion Model-Based Label Embedding and Self-Interaction Attention for Text Classification , 2020, IEEE Access.

[24]  Tianrui Li,et al.  Multivariate time series forecasting via attention-based encoder-decoder framework , 2020, Neurocomputing.

[25]  Shanlin Yang,et al.  Electrical load forecasting based on self-adaptive chaotic neural network using Chebyshev map , 2016, Neural Computing and Applications.

[26]  Diyi Yang,et al.  Hierarchical Attention Networks for Document Classification , 2016, NAACL.

[27]  Liang Guo,et al.  A recurrent neural network based health indicator for remaining useful life prediction of bearings , 2017, Neurocomputing.

[28]  Jeremy S. Smith,et al.  Image captioning via hierarchical attention mechanism and policy gradient optimization , 2020, Signal Process..

[29]  Yi Yang,et al.  A hybrid application algorithm based on the support vector machine and artificial intelligence: An example of electric load forecasting , 2015 .

[30]  Zahra Hajirahimi,et al.  Sequence in Hybridization of Statistical and Intelligent Models in Time Series Forecasting , 2020, Neural Processing Letters.

[31]  Sven F. Crone,et al.  Advances in forecasting with neural networks? Empirical evidence from the NN3 competition on time series prediction , 2011 .

[32]  Ling Yang,et al.  DSTP-RNN: a dual-stage two-phase attention-based recurrent neural networks for long-term and multivariate time series prediction , 2019, Expert Syst. Appl..

[33]  Yimei Yang,et al.  Hybrid Method for Short-Term Time Series Forecasting Based on EEMD , 2020, IEEE Access.

[34]  Enhong Chen,et al.  Combining contextual neural networks for time series classification , 2020, Neurocomputing.

[35]  Simone Scardapane,et al.  A non-parametric softmax for improving neural attention in time-series forecasting , 2020, Neurocomputing.

[36]  Yoshua Bengio,et al.  On the Properties of Neural Machine Translation: Encoder–Decoder Approaches , 2014, SSST@EMNLP.

[37]  Bo Yu,et al.  Convolutional Neural Networks for human activity recognition using mobile sensors , 2014, 6th International Conference on Mobile Computing, Applications and Services.

[38]  Amy Loutfi,et al.  A review of unsupervised feature learning and deep learning for time-series modeling , 2014, Pattern Recognit. Lett..

[39]  Geoffrey E. Hinton,et al.  Deep Learning , 2015, Nature.

[40]  Dong Tian,et al.  An optimized hybrid model based on artificial intelligence for grape price forecasting , 2019, British Food Journal.

[41]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[42]  Gang Song,et al.  A novel double deep ELMs ensemble system for time series forecasting , 2017, Knowl. Based Syst..