DifFormer: Multi-Resolutional Differencing Transformer With Dynamic Ranging for Time Series Analysis

Time series analysis is essential to many far-reaching applications of data science and statistics including economic and financial forecasting, surveillance, and automated business processing. Though being greatly successful of Transformer in computer vision and natural language processing, the potential of employing it as the general backbone in analyzing the ubiquitous times series data has not been fully released yet. Prior Transformer variants on time series highly rely on task-dependent designs and pre-assumed “pattern biases”, revealing its insufficiency in representing nuanced seasonal, cyclic, and outlier patterns which are highly prevalent in time series. As a consequence, they can not generalize well to different time series analysis tasks. To tackle the challenges, we propose DifFormer, an effective and efficient Transformer architecture that can serve as a workhorse for a variety of time-series analysis tasks. DifFormer incorporates a novel multi-resolutional differencing mechanism, which is able to progressively and adaptively make nuanced yet meaningful changes prominent, meanwhile, the periodic or cyclic patterns can be dynamically captured with flexible lagging and dynamic ranging operations. Extensive experiments demonstrate DifFormer significantly outperforms state-of-the-art models on three essential time-series analysis tasks, including classification, regression, and forecasting. In addition to its superior performances, DifFormer also excels in efficiency – a linear time/memory complexity with empirically lower time consumption.

[1]  J. Kalagnanam,et al.  A Time Series is Worth 64 Words: Long-term Forecasting with Transformers , 2022, ICLR.

[2]  L. Zhang,et al.  Are Transformers Effective for Time Series Forecasting? , 2022, AAAI.

[3]  Junchi Yan,et al.  Transformers in Time Series: A Survey , 2022, IJCAI.

[4]  Jingbo Shang,et al.  UniTS: Short-Time Fourier Inspired Neural Networks for Sensory Time Series Classification , 2021, SenSys.

[5]  Jianmin Wang,et al.  Autoformer: Decomposition Transformers with Auto-Correlation for Long-Term Series Forecasting , 2021, NeurIPS.

[6]  Zhenghua Chen,et al.  Two-Stream Convolution Augmented Transformer for Human Activity Recognition , 2021, AAAI.

[7]  James Large,et al.  HIVE-COTE 2.0: a new meta ensemble for time series classification , 2021, Machine Learning.

[8]  Suhang Wang,et al.  Explainable Multivariate Time Series Classification: A Deep Neural Network Which Learns to Attend to Important Variables As Well As Time Intervals , 2021, WSDM.

[9]  Hui Xiong,et al.  Informer: Beyond Efficient Transformer for Long Sequence Time-Series Forecasting , 2020, AAAI.

[10]  Jun Luo,et al.  RF-net: a unified meta-learning framework for RF-enabled one-shot human activity recognition , 2020, SenSys.

[11]  James Large,et al.  The Canonical Interval Forest (CIF) Classifier for Time Series Classification , 2020, 2020 IEEE International Conference on Big Data (Big Data).

[12]  Gavin C. Cawley,et al.  The Temporal Dictionary Ensemble (TDE) Classifier for Time Series Classification , 2020, ECML/PKDD.

[13]  Zhenghua Chen,et al.  WiFi-Based Indoor Robot Positioning Using Deep Fuzzy Forests , 2020, IEEE Internet of Things Journal.

[14]  Jianfeng Gao,et al.  UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training , 2020, ICML.

[15]  Mani Srivastava,et al.  SenseHAR: a robust virtual activity sensor for smartphones and wearables , 2019, SenSys.

[16]  Peter J. Liu,et al.  Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer , 2019, J. Mach. Learn. Res..

[17]  Christian Bock,et al.  Set Functions for Time Series , 2019, ICML.

[18]  Junbo Zhang,et al.  Urban Traffic Prediction from Spatio-Temporal Data Using Deep Meta Learning , 2019, KDD.

[19]  Wenhu Chen,et al.  Enhancing the Locality and Breaking the Memory Bottleneck of Transformer on Time Series Forecasting , 2019, NeurIPS.

[20]  Jing Jiang,et al.  Graph WaveNet for Deep Spatial-Temporal Graph Modeling , 2019, IJCAI.

[21]  Hsiang-Fu Yu,et al.  Think Globally, Act Locally: A Deep Neural Network Approach to High-Dimensional Time Series Forecasting , 2019, NeurIPS.

[22]  Jiawei Han,et al.  STFNets: Learning Sensing Signals from the Time-Frequency Perspective with Short-Time Fourier Neural Networks , 2019, WWW.

[23]  VALENTIN RADU,et al.  Multimodal Deep Learning for Activity and Context Recognition , 2018, Proc. ACM Interact. Mob. Wearable Ubiquitous Technol..

[24]  Yunhao Liu,et al.  Design and Implementation of a CSI-Based Ubiquitous Smoking Detection System , 2017, IEEE/ACM Transactions on Networking.

[25]  Haoteng Yin,et al.  Spatio-temporal Graph Convolutional Neural Network: A Deep Learning Framework for Traffic Forecasting , 2017, IJCAI.

[26]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[27]  Valentin Flunkert,et al.  DeepAR: Probabilistic Forecasting with Autoregressive Recurrent Networks , 2017, International Journal of Forecasting.

[28]  Guokun Lai,et al.  Modeling Long- and Short-Term Temporal Patterns with Deep Neural Networks , 2017, SIGIR.

[29]  Tim Oates,et al.  Time series classification from scratch with deep neural networks: A strong baseline , 2016, 2017 International Joint Conference on Neural Networks (IJCNN).

[30]  Ivan Poupyrev,et al.  Soli , 2016, ACM Trans. Graph..

[31]  Jimmy Ba,et al.  Adam: A Method for Stochastic Optimization , 2014, ICLR.

[32]  Andrew Zisserman,et al.  Two-Stream Convolutional Networks for Action Recognition in Videos , 2014, NIPS.

[33]  Miguel A. Labrador,et al.  A Survey on Human Activity Recognition using Wearable Sensors , 2013, IEEE Communications Surveys & Tutorials.

[34]  Ali H. Shoeb,et al.  Application of Machine Learning To Epileptic Seizure Detection , 2010, ICML.

[35]  Xavier Glorot,et al.  Understanding the difficulty of training deep feedforward neural networks , 2010, AISTATS.

[36]  Abhinav Saxena,et al.  Damage propagation modeling for aircraft engine run-to-failure simulation , 2008, 2008 International Conference on Prognostics and Health Management.

[37]  Brij N. Singh,et al.  Optimal selection of wavelet basis function applied to ECG signal denoising , 2006, Digit. Signal Process..

[38]  R. E. Donatelli,et al.  Time Series Analysis , 2003, Statistics for Environmental Science and Management.

[39]  Chu‐Hua Kuei,et al.  Designing and Managing the Supply Chain Concepts, Strategies, and Case Studies , 2000 .

[40]  P. Young,et al.  Time series analysis, forecasting and control , 1972, IEEE Transactions on Automatic Control.

[41]  Alex X. Liu,et al.  Pyraformer: Low-Complexity Pyramidal Attention for Long-Range Time Series Modeling and Forecasting , 2022, ICLR.

[42]  Stephen Lin,et al.  Swin Transformer: Hierarchical Vision Transformer using Shifted Windows , 2021, 2021 IEEE/CVF International Conference on Computer Vision (ICCV).

[43]  Jaesik Choi,et al.  Conditional Temporal Neural Processes with Covariance Loss , 2021, ICML.

[44]  Xi Xiao,et al.  Adversarial Sparse Transformer for Time Series Forecasting , 2020, NeurIPS.

[45]  Siu Cheung Hui,et al.  Compositional De-Attention Networks , 2019, NeurIPS.

[46]  Matthias W. Seeger,et al.  Deep State Space Models for Time Series Forecasting , 2018, NeurIPS.