Sig-wasserstein GANs for time series generation

Synthetic data is an emerging technology that can significantly accelerate the development and deployment of AI machine learning pipelines. In this work, we develop high-fidelity time-series generators, the SigWGAN, by combining continuous-time stochastic models with the newly proposed signature W1 metric. The former are the Logsig-RNN models based on the stochastic differential equations, whereas the latter originates from the universal and principled mathematical features to characterize the measure induced by time series. SigWGAN allows turning computationally challenging GAN min-max problem into supervised learning while generating high fidelity samples. We validate the proposed model on both synthetic data generated by popular quantitative risk models and empirical financial data. Codes are available at https://github.com/SigCGANs/Sig-WassersteinGANs.git

[1]  Magnus Wiese,et al.  Deep Hedging: Learning to Simulate Equity Option Markets , 2019, ArXiv.

[2]  Magnus Wiese,et al.  Conditional Sig-Wasserstein GANs for Time Series Generation , 2020, SSRN Electronic Journal.

[3]  Terry Lyons,et al.  Characteristic functions of measures on geometric rough paths , 2013, 1307.3580.

[4]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[5]  Cristopher Salvi,et al.  Sig-SDEs model for quantitative finance , 2020, ICAIF.

[6]  Terry Lyons,et al.  Logsig-RNN: a novel network for robust and efficient skeleton-based action recognition , 2021, BMVC.

[7]  Terry Lyons,et al.  A Data-Driven Market Simulator for Small Data Environments , 2020, SSRN Electronic Journal.

[8]  Mario Lucic,et al.  Are GANs Created Equal? A Large-Scale Study , 2017, NeurIPS.

[9]  Terry Lyons,et al.  Neural Controlled Differential Equations for Irregular Time Series , 2020, NeurIPS.

[10]  Terry Lyons,et al.  A Generalised Signature Method for Multivariate Time Series Feature Extraction , 2021 .

[11]  Christos H. Papadimitriou,et al.  Cycles in adversarial regularized learning , 2017, SODA.

[12]  S. Shankar Sastry,et al.  On Finding Local Nash Equilibria (and Only Local Nash Equilibria) in Zero-Sum Games , 2019, 1901.00838.

[13]  David Siska,et al.  Robust Pricing and Hedging via Neural SDEs , 2020, SSRN Electronic Journal.

[14]  Steven M. Bellovin,et al.  Privacy and Synthetic Datasets , 2018 .

[15]  Christa Cuchiero,et al.  A Generative Adversarial Network Approach to Calibration of Local Stochastic Volatility Models , 2020, Risks.

[16]  Terry Lyons,et al.  Uniqueness for the signature of a path of bounded variation and the reduced path group , 2005, math/0507536.

[17]  I. Chevyrev,et al.  Signature Moments to Characterize Laws of Stochastic Processes , 2018, J. Mach. Learn. Res..

[18]  Philip C. Treleaven,et al.  Generative adversarial networks for financial trading strategies fine-tuning and combination , 2019, Quantitative Finance.

[19]  H. Boedihardjo,et al.  The uniqueness of signature problem in the non-Markov setting , 2014, 1401.6165.

[20]  Constantinos Daskalakis,et al.  The Limit Points of (Optimistic) Gradient Descent in Min-Max Optimization , 2018, NeurIPS.

[21]  Magnus Wiese,et al.  Quant GANs: deep generation of financial time series , 2019, Quantitative Finance.

[22]  Terry Lyons,et al.  Learning stochastic differential equations using RNN with log signature features , 2019, ArXiv.

[23]  Michael I. Jordan,et al.  On Gradient Descent Ascent for Nonconvex-Concave Minimax Problems , 2019, ICML.

[24]  Samuel A. Assefa Generating synthetic data in finance: opportunities, challenges and pitfalls , 2020, ICAIF.