Tensorized LSTM with Adaptive Shared Memory for Learning Trends in Multivariate Time Series

The problem of learning and forecasting underlying trends in time series data arises in a variety of applications, such as traffic management, energy optimization, etc. In literature, a trend in time series is characterized by the slope and duration, and its prediction is then to forecast the two values of the subsequent trend given historical data of the time series. For this problem, existing approaches mainly deal with the case in univariate time series. However, in many real-world applications, there are multiple variables at play, and handling all of them at the same time is crucial for an accurate prediction. A natural way is to employ multi-task learning (MTL) techniques in which the trend learning of each time series is treated as a task. The key point of MTL is to learn task relatedness to achieve better parameter sharing, which however is challenging in trend prediction task. First, effectively modeling the complex temporal patterns in different tasks is hard as the temporal and spatial dimensions are entangled. Second, the relatedness among tasks may change over time. In this paper, we propose a neural network, DeepTrends, for multivariate time series trend prediction. The core module of DeepTrends is a tensorized LSTM with adaptive shared memory (TLASM). TLASM employs the tensorized LSTM to model the temporal patterns of long-term trend sequences in an MTL setting. With an adaptive shared memory, TLASM is able to learn the relatedness among tasks adaptively, based upon which it can dynamically vary degrees of parameter sharing among tasks. To further consider short-term patterns, DeepTrends utilizes a multi-task 1dCNN to learn the local time series features, and employs a task-specific sub-network to learn a mixture of long-term and short-term patterns for trend prediction. Extensive experiments on real datasets demonstrate the effectiveness of the proposed model.

[1]  Rich Caruana,et al.  Multitask Learning , 1998, Encyclopedia of Machine Learning and Data Mining.

[2]  Andrew J. Davison,et al.  End-To-End Multi-Task Learning With Attention , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[3]  Larry S. Davis,et al.  Modeling Local Geometric Structure of 3D Point Clouds Using Geo-CNN , 2018, 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR).

[4]  Bo Zong,et al.  Deep Co-Clustering , 2019, SDM.

[5]  Ed H. Chi,et al.  SNR: Sub-Network Routing for Flexible Parameter Sharing in Multi-Task Learning , 2019, AAAI.

[6]  Ezio Bartocci,et al.  Temporal Logic Based Monitoring of Assisted Ventilation in Intensive Care Patients , 2014, ISoLA.

[7]  Xiang Zhang,et al.  Spatio-Temporal Attentive RNN for Node Classification in Temporal Attributed Graphs , 2019, IJCAI.

[8]  Garrison W. Cottrell,et al.  A Dual-Stage Attention-Based Recurrent Neural Network for Time Series Prediction , 2017, IJCAI.

[9]  Gerard de Melo,et al.  Cross-Lingual Propagation for Deep Sentiment Analysis , 2018, AAAI.

[10]  Zhiwei Wang,et al.  Deep Knowledge Tracing with Side Information , 2019, AIED.

[11]  Maosong Sun,et al.  Enhancing Stock Movement Prediction with Adversarial Training , 2018, IJCAI.

[12]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[13]  Nicolas Chapados,et al.  N-BEATS: Neural basis expansion analysis for interpretable time series forecasting , 2019, ICLR.

[14]  Shay B. Cohen,et al.  Stock Movement Prediction from Tweets and Historical Prices , 2018, ACL.

[15]  Gerard de Melo,et al.  A Robust Self-Learning Framework for Cross-Lingual Text Classification , 2019, EMNLP.

[16]  Jungwon Lee,et al.  Residual LSTM: Design of a Deep Recurrent Architecture for Distant Speech Recognition , 2017, INTERSPEECH.

[17]  Jackie Chi Kit Cheung,et al.  Learning Multi-Task Communication with Message Passing for Sequence Learning , 2019, AAAI.

[18]  Tao Lin,et al.  Hybrid Neural Networks for Learning the Trend in Time Series , 2017, IJCAI.

[19]  Bo Zong,et al.  Adaptive Neural Network for Node Classification in Dynamic Networks , 2019, 2019 IEEE International Conference on Data Mining (ICDM).

[20]  Martial Hebert,et al.  Cross-Stitch Networks for Multi-task Learning , 2016, 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR).

[21]  Christopher Joseph Pal,et al.  Delving Deeper into Convolutional Networks for Learning Video Representations , 2015, ICLR.

[22]  Nino Antulov-Fantulin,et al.  Exploring Interpretable LSTM Neural Networks over Multi-Variable Data , 2019, ICML.

[23]  Thomas S. Huang,et al.  Dilated Recurrent Neural Networks , 2017, NIPS.

[24]  Zhe Zhao,et al.  Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts , 2018, KDD.

[25]  Matthias W. Seeger,et al.  Deep State Space Models for Time Series Forecasting , 2018, NeurIPS.

[26]  Gerard de Melo,et al.  A Helping Hand: Transfer Learning for Deep Sentiment Analysis , 2018, ACL.

[27]  Vasant Honavar,et al.  LMLFM: Longitudinal Multi-Level Factorization Machines , 2019 .

[28]  Guokun Lai,et al.  Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks , 2017, SIGIR.

[29]  Liang Xiao,et al.  Wider and Deeper, Cheaper and Faster: Tensorized LSTMs for Sequence Learning , 2017, NIPS.

[30]  Xuanjing Huang,et al.  Deep Multi-Task Learning with Shared Memory , 2016, ArXiv.

[31]  Zhang Xiong,et al.  AlphaStock: A Buying-Winners-and-Selling-Losers Investment Strategy using Interpretable Deep Reinforcement Attention Networks , 2019, KDD.

[32]  Joachim Bingel,et al.  Latent Multi-Task Architecture Learning , 2017, AAAI.