Optimizing stock market execution costs using reinforcement learning

Stock market trading is a complex process where traders aim to maximize their expected return while minimizing associated risks. With the increasing availability of digital historical records, using automated agents for stock market trading becomes of a significant interest. Reinforcement learning is a machine learning branch which circumvents the problem of defining explicit targets and tackles problems which require sequential decisions. Reinforcement learning has been applied in finance problems, yet execution costs optimization problem among others still gets little attention. The optimization of execution costs in stock markets is a vital problem, where a trader wants to minimize the cost of buying a predefined amount of shares over a fixed time horizon. In this study, we propose a novel reinforcement learning Q-trade model to address the execution costs optimization problem. We tested the Q-trade model using historical data of the Egyptian stock market and Nasdaq stock market, and it showed in both markets a significant improvement (more than 60% for some securities) over the compared strategies.

[1]  D. Bertsimas,et al.  Optimal control of execution costs , 1998 .

[2]  Gur Huberman,et al.  Optimal Liquidity Trading , 2000 .

[3]  Matthew Saffell,et al.  Learning to trade via direct reinforcement , 2001, IEEE Trans. Neural Networks.

[4]  T. Downs,et al.  System for foreign exchange trading using genetic algorithms and reinforcement learning , 2004, Int. J. Syst. Sci..

[5]  Alexander Fadeev,et al.  Optimal execution for portfolio transactions , 2006 .

[6]  Richard S. Sutton,et al.  Reinforcement Learning: An Introduction , 1998, IEEE Trans. Neural Networks.

[7]  Stacy Williams,et al.  Limit order books , 2010, 1012.0349.

[8]  Alexander Schied,et al.  Risk aversion and the dynamics of optimal liquidation strategies in illiquid markets , 2009, Finance Stochastics.

[9]  Dale Schuurmans,et al.  Learning Exercise Policies for American Options , 2009, AISTATS.

[10]  Michael Kearns,et al.  Reinforcement learning for optimized trade execution , 2006, ICML.

[11]  Peter A. Forsyth,et al.  Optimal trade execution: A mean quadratic variation approach , 2012 .

[12]  Ben J. A. Kröse,et al.  Learning from delayed rewards , 1995, Robotics Auton. Syst..

[13]  Robert Almgren,et al.  Optimal Trading with Stochastic Liquidity and Volatility , 2012, SIAM J. Financial Math..

[14]  John Fearnley,et al.  Market Making via Reinforcement Learning , 2018, AAMAS.

[15]  Julian Lorenz,et al.  Mean–Variance Optimal Adaptive Execution , 2011 .

[16]  Dieter Hendricks,et al.  A reinforcement learning extension to the Almgren-Chriss framework for optimal trade execution , 2014, 2014 IEEE Conference on Computational Intelligence for Financial Engineering & Economics (CIFEr).

[17]  Marie-Josée Cros,et al.  MDPtoolbox: a multi-platform toolbox to solve stochastic dynamic programming problems , 2014 .

[18]  Olivier Gu 'eant Optimal execution and block trade pricing: a general framework , 2014 .

[19]  Robert Almgren,et al.  Optimal execution with nonlinear impact functions and trading-enhanced risk , 2003 .