Quant GANs: deep generation of financial time series

Modeling financial time series by stochastic processes is a challenging task and a central area of research in financial mathematics. As an alternative, we introduce Quant GANs, a data-driven model which is inspired by the recent success of generative adversarial networks (GANs). Quant GANs consist of a generator and discriminator function, which utilize temporal convolutional networks (TCNs) and thereby achieve to capture long-range dependencies such as the presence of volatility clusters. The generator function is explicitly constructed such that the induced stochastic process allows a transition to its risk-neutral distribution. Our numerical results highlight that distributional properties for small and large lags are in an excellent agreement and dependence properties such as volatility clusters, leverage effects, and serial autocorrelations can be generated by the generator function of Quant GANs, demonstrably in high fidelity.

[1]  Yoshua Bengio,et al.  Generative Adversarial Nets , 2014, NIPS.

[2]  Sebastian Nowozin,et al.  Which Training Methods for GANs do actually Converge? , 2018, ICML.

[3]  Heiga Zen,et al.  WaveNet: A Generative Model for Raw Audio , 2016, SSW.

[4]  Georg M. Goerg The Lambert Way to Gaussianize Heavy-Tailed Data with the Inverse of Tukey's h Transformation as a Special Case , 2010, TheScientificWorldJournal.

[5]  Gunnar Rätsch,et al.  Real-valued (Medical) Time Series Generation with Recurrent Conditional GANs , 2017, ArXiv.

[6]  Yong Wang,et al.  Stock Market Prediction Based on Generative Adversarial Network , 2018, IIKI.

[7]  Sepp Hochreiter,et al.  Fast and Accurate Deep Network Learning by Exponential Linear Units (ELUs) , 2015, ICLR.

[8]  Leon A. Gatys,et al.  A Neural Algorithm of Artistic Style , 2015, ArXiv.

[9]  Andreas Dengel,et al.  Detection of Anomalies in Large Scale Accounting Data using Deep Autoencoder Networks , 2017, ArXiv.

[10]  Jian Sun,et al.  Delving Deep into Rectifiers: Surpassing Human-Level Performance on ImageNet Classification , 2015, 2015 IEEE International Conference on Computer Vision (ICCV).

[11]  V. Yakovenko,et al.  Probability distribution of returns in the Heston model with stochastic volatility , 2002, cond-mat/0203046.

[12]  R. Schilling Financial Modelling with Jump Processes , 2005 .

[13]  Jürgen Schmidhuber,et al.  Long Short-Term Memory , 1997, Neural Computation.

[14]  Luca Antiga,et al.  Automatic differentiation in PyTorch , 2017 .

[15]  F. Black,et al.  The Pricing of Options and Corporate Liabilities , 1973, Journal of Political Economy.

[16]  J. Teichmann,et al.  Deep hedging , 2018, Quantitative Finance.

[17]  Yoshua Bengio,et al.  Maxout Networks , 2013, ICML.

[18]  Geoffrey E. Hinton,et al.  Rectified Linear Units Improve Restricted Boltzmann Machines , 2010, ICML.

[19]  Sebastian Becker,et al.  Deep Optimal Stopping , 2018, J. Mach. Learn. Res..

[20]  Philip C. Treleaven,et al.  Generative adversarial networks for financial trading strategies fine-tuning and combination , 2019, Quantitative Finance.

[21]  Kurt Hornik,et al.  Approximation capabilities of multilayer feedforward networks , 1991, Neural Networks.

[22]  Razvan Pascanu,et al.  On the difficulty of training recurrent neural networks , 2012, ICML.

[23]  Damian Borth,et al.  Adversarial Learning of Deepfakes in Accounting , 2019, ArXiv.

[24]  Guigang Zhang,et al.  Deep Learning , 2016, Int. J. Semantic Comput..

[25]  Fernando De Meer Pardo,et al.  Enriching Financial Datasets with Generative Adversarial Networks , 2019 .

[26]  Yoshua Bengio,et al.  Empirical Evaluation of Gated Recurrent Neural Networks on Sequence Modeling , 2014, ArXiv.

[27]  Demis Hassabis,et al.  Mastering the game of Go with deep neural networks and tree search , 2016, Nature.

[28]  T. Munich,et al.  Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks , 2008, NIPS.

[29]  Geoffrey E. Hinton,et al.  ImageNet classification with deep convolutional neural networks , 2012, Commun. ACM.

[30]  Sepp Hochreiter,et al.  Self-Normalizing Neural Networks , 2017, NIPS.

[31]  Aaron C. Courville,et al.  Improved Training of Wasserstein GANs , 2017, NIPS.

[32]  Eric Jones,et al.  SciPy: Open Source Scientific Tools for Python , 2001 .

[33]  F. Abergel,et al.  Econophysics review: I. Empirical facts , 2011 .

[34]  S. Heston A Closed-Form Solution for Options with Stochastic Volatility with Applications to Bond and Currency Options , 1993 .

[35]  Vladlen Koltun,et al.  An Empirical Evaluation of Generic Convolutional and Recurrent Networks for Sequence Modeling , 2018, ArXiv.

[36]  Léon Bottou,et al.  Towards Principled Methods for Training Generative Adversarial Networks , 2017, ICLR.

[37]  T. Bollerslev,et al.  Generalized autoregressive conditional heteroskedasticity , 1986 .

[38]  C. Villani Optimal Transport: Old and New , 2008 .

[39]  R. Cont,et al.  Financial Modelling with Jump Processes , 2003 .

[40]  Damian Borth,et al.  Detection of Accounting Anomalies in the Latent Space using Adversarial Autoencoder Neural Networks , 2019, ArXiv.

[41]  Fernando De Meer Pardo,et al.  Mitigating Overfitting on Financial Datasets with Generative Adversarial Networks , 2020 .

[42]  Magnus Wiese,et al.  Deep Hedging: Learning to Simulate Equity Option Markets , 2019, ArXiv.

[43]  Zhisong Pan,et al.  Stock Market Prediction on High-Frequency Data Using Generative Adversarial Nets , 2018 .

[44]  Rizwan Mushtaq,et al.  Augmented Dickey Fuller Test , 2011 .

[45]  Kumiko Tanaka-Ishii,et al.  Modeling financial time-series with generative adversarial networks , 2019, Physica A: Statistical Mechanics and its Applications.

[46]  R. Cont Empirical properties of asset returns: stylized facts and statistical issues , 2001 .

[47]  Ira Kemelmacher-Shlizerman,et al.  Synthesizing Obama , 2017, ACM Trans. Graph..

[48]  Alexia Jolicoeur-Martineau,et al.  The relativistic discriminator: a key element missing from standard GAN , 2018, ICLR.

[49]  Jeff Donahue,et al.  Large Scale GAN Training for High Fidelity Natural Image Synthesis , 2018, ICLR.

[50]  Tianqi Chen,et al.  Empirical Evaluation of Rectified Activations in Convolutional Network , 2015, ArXiv.

[51]  Karen Simonyan,et al.  The challenge of realistic music generation: modelling raw audio at scale , 2018, NeurIPS.

[52]  Hans Buehler,et al.  Discrete Local Volatility for Large Time Steps (Extended Version) , 2015 .