Repr2Seq: A Data-to-Text Generation Model for Time Series

Data-to-text generation takes structured data as input and produces text that sufficiently describes the data as output. Recently, it has received a lot of attention from both research field and industry. However, as a critical data form, time series is less discussed in this domain. This paper proposes Repr2Seq, a data-to-text generation model for time series. To better capture the structure and core information of time series, Repr2Seq obtains representation vectors using time series representation learning methods, which are then fed into a neural network-based model to generate text sequences. To demonstrate the effectiveness of Repr2Seq, a dataset consisting of stock price series and corresponding comments is proposed. Experiments show that Repr2Seq achieves significant improvement over standard approaches and leads to satisfactory results. We also conduct experiments to investigate the effect of hyperparameters on the model and detect performance improvements in various settings.

[1]  Zhenghua Chen,et al.  Time-Series Representation Learning via Temporal and Contextual Contrasting , 2021, IJCAI.

[2]  Yu Tong,et al.  TS2Vec: Towards Universal Representation of Time Series , 2021, AAAI.

[3]  Anna Goldenberg,et al.  Unsupervised Representation Learning for Time Series with Temporal Neighborhood Coding , 2021, ICLR.

[4]  Anuradha Bhamidipaty,et al.  A Transformer-based Framework for Multivariate Time Series Representation Learning , 2020, KDD.

[5]  Dimitris Vrakas,et al.  Signal2Vec: Time Series Embedding Representation , 2019, EANN.

[6]  Martin Jaggi,et al.  Unsupervised Scalable Representation Learning for Multivariate Time Series , 2019, NeurIPS.

[7]  Mirella Lapata,et al.  Data-to-Text Generation with Content Selection and Planning , 2018, AAAI.

[8]  Jinfeng Yi,et al.  Random Warping Series: A Random Features Method for Time-Series Embedding , 2018, AISTATS.

[9]  Lukasz Kaiser,et al.  Attention is All you Need , 2017, NIPS.

[10]  Emiel Krahmer,et al.  Survey of the State of the Art in Natural Language Generation: Core tasks, applications and evaluation , 2017, J. Artif. Intell. Res..

[11]  David Grangier,et al.  Neural Text Generation from Structured Data with Application to the Biography Domain , 2016, EMNLP.

[12]  Matthew R. Walter,et al.  What to talk about and how? Selective Generation using LSTMs with Coarse-to-Fine Alignment , 2015, NAACL.

[13]  Quoc V. Le,et al.  Sequence to Sequence Learning with Neural Networks , 2014, NIPS.

[14]  Yoshua Bengio,et al.  Neural Machine Translation by Jointly Learning to Align and Translate , 2014, ICLR.

[15]  Yoshua Bengio,et al.  Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation , 2014, EMNLP.

[16]  Li Wei,et al.  Experiencing SAX: a novel symbolic representation of time series , 2007, Data Mining and Knowledge Discovery.

[17]  Irfan A. Essa,et al.  Discovering Multivariate Motifs using Subsequence Density Estimation and Greedy Mixture Learning , 2007, AAAI.

[18]  Eamonn J. Keogh,et al.  A symbolic representation of time series, with implications for streaming algorithms , 2003, DMKD '03.

[19]  Salim Roukos,et al.  Bleu: a Method for Automatic Evaluation of Machine Translation , 2002, ACL.

[20]  Eamonn J. Keogh,et al.  Dimensionality Reduction for Fast Similarity Search in Large Time Series Databases , 2001, Knowledge and Information Systems.

[21]  Zbigniew R. Struzik,et al.  Wavelet Transform in Similarity Paradigm , 1998, PAKDD.

[22]  S. Hochreiter,et al.  Long Short-Term Memory , 1997, Neural Computation.

[23]  Christos Faloutsos,et al.  Efficiently supporting ad hoc queries in large datasets of time sequences , 1997, SIGMOD '97.

[24]  Yoshua Bengio,et al.  Learning long-term dependencies with gradient descent is difficult , 1994, IEEE Trans. Neural Networks.

[25]  Christos Faloutsos,et al.  Efficient Similarity Search In Sequence Databases , 1993, FODO.

[26]  Paul J. Werbos,et al.  Backpropagation Through Time: What It Does and How to Do It , 1990, Proc. IEEE.

[27]  Geoffrey E. Hinton,et al.  Learning representations by back-propagating errors , 1986, Nature.

[28]  Karen Kukich,et al.  Design of a Knowledge-Based Report Generator , 1983, ACL.

[29]  Marcella Corduas,et al.  Time series clustering and classification by the autoregressive metric , 2008, Comput. Stat. Data Anal..

[30]  Yoshua Bengio,et al.  Gradient Flow in Recurrent Nets: the Difficulty of Learning Long-Term Dependencies , 2001 .