Generalized Deep Reinforcement Learning for Trading

Authors

  • Junyoung Sim Ithaca High School
  • Benjamin Kirk Ithaca High School

DOI:

https://doi.org/10.47611/jsrhs.v12i1.4316

Keywords:

Artificial Intelligence, Deep Reinforcement Learning, Trading

Abstract

This paper proposes generalized deep reinforcement learning with multivariate state space, discrete rewards, and adaptive synchronization for trading any stock held in the S&P 500. Specifically, the proposed trading model observes the daily historical data of a stock held in the S&P 500 and multiple market-indicating securities (SPY, IEF, EUR=X, GSG), selects a trading action, and observes a discrete reward that is based on the correctness of the selected action and independent of the volatility of stocks. The proposed trading model’s reward-maximizing behavior is optimized by using a standard deep q-network (DQN) with adaptive synchronization that stabilizes and enables to track learning performance on generalizing new experiences from each stock. The proposed trading model was trained on the top 50 holdings of the S&P 500 and tested on the top 100 holdings of the S&P 500 starting from 2006 to 2022. Experimental results suggest that the proposed trading model significantly outperforms the 100% long-strategy benchmark in terms of annualized return, Sharpe ratio, and maximum drawdown.

Downloads

Download data is not yet available.

Author Biographies

Junyoung Sim, Ithaca High School

Junyoung Sim was born in Seoul, Korea, and moved to Ithaca, New York, in 2022. Junyoung attended Korea International School, Jeju, during his time in Korea and is currently a junior attending Ithaca High School. Junyoung is interested in computer science, finance, and operations research and engineering.

Benjamin Kirk, Ithaca High School

Benjamin Kirk is a mathematics teacher at Ithaca High School, where he has taught for 16 years. He is a New York State Master Teacher and regularly assists the state's education department as part of their Math Content Advisory Panel, where he advocates for increased attention to statistics and data science. 

References or Bibliography

A. W. Li and G. S. Bastos, “Stock Market Forecasting Using Deep Learning and Technical Analysis: A Systematic Review,” IEEE Access, vol. 8, pp. 185232–185242, 2020, doi: 10.1109/ACCESS.2020.3030226.

B. Lim, S. Zohren, and S. Roberts, “Enhancing Time-Series Momentum Strategies Using Deep Neural Networks,” The Journal of Financial Data Science, 2019, doi: 10.3905/jfds.2019.1.015.

S. Siami-Namini and A. S. Namin, “Forecasting Economics and Financial Time Series: ARIMA vs. LSTM.” arXiv, 2018. doi: 10.48550/ARXIV.1803.06386.

Q. Guo, S. Lei, Q. Ye, and Z. Fang, “MRC-LSTM: A Hybrid Approach of Multi-scale Residual CNN and LSTM to Predict Bitcoin Price,” in 2021 International Joint Conference on Neural Networks (IJCNN), 2021, pp. 1–8. doi: 10.1109/IJCNN52387.2021.9534453.

V. Mnih et al., “Human-level control through deep reinforcement learning,” Nature, vol. 518, no. 7540, pp. 529–533, Feb. 2015, doi: 10.1038/nature14236.

V. Mnih et al., “Playing Atari with Deep Reinforcement Learning.” arXiv, 2013. doi: 10.48550/ARXIV.1312.5602.

A. Millea, “Deep Reinforcement Learning for Trading—A Critical Survey,” Data, vol. 6, no. 11, 2021, doi: 10.3390/data6110119.

Z. Zhang, S. Zohren, and S. Roberts, “Deep Reinforcement Learning for Trading,” The Journal of Financial Data Science, vol. 2, no. 2, pp. 25–40, 2020, doi: 10.3905/jfds.2020.1.030.

Y. Li, W. Zheng, and Z. Zheng, “Deep Robust Reinforcement Learning for Practical Algorithmic Trading,” IEEE Access, vol. 7, pp. 108014–108022, 2019, doi: 10.1109/ACCESS.2019.2932789.

Y. Deng, F. Bao, Y. Kong, Z. Ren, and Q. Dai, “Deep Direct Reinforcement Learning for Financial Signal Representation and Trading,” IEEE Transactions on Neural Networks and Learning Systems, vol. 28, no. 3, pp. 653–664, 2017, doi: 10.1109/TNNLS.2016.2522401.

S. Carta, A. Corriga, A. Ferreira, A. S. Podda, and D. R. Recupero, “A multi-layer and multi-ensemble stock trader using deep learning and deep reinforcement learning,” Applied Intelligence, vol. 51, no. 2, pp. 889–905, Feb. 2021, doi: 10.1007/s10489-020-01839-5.

F. Rundo, “Deep LSTM with Reinforcement Learning Layer for Financial Trend Prediction in FX High Frequency Trading Systems,” Applied Sciences, vol. 9, no. 20, 2019, doi: 10.3390/app9204460.

X.-Y. Liu, Z. Xiong, S. Zhong, H. Yang, and A. Walid, “Practical Deep Reinforcement Learning Approach for Stock Trading.” arXiv, 2018. doi: 10.48550/ARXIV.1811.07522.

Y. Li, “Deep Reinforcement Learning: An Overview.” arXiv, 2017. doi: 10.48550/ARXIV.1701.07274.

L. P. Kaelbling, M. L. Littman, and A. W. Moore, “Reinforcement learning: A survey,” Journal of artificial intelligence research, vol. 4, pp. 237–285, 1996, doi: https://doi.org/10.1613/jair.301.

V. Mnih et al., “Asynchronous Methods for Deep Reinforcement Learning,” 2016, doi: 10.48550/ARXIV.1602.01783.

L. Menkhoff, “The use of technical analysis by fund managers: International evidence,” Journal of Banking & Finance, vol. 34, no. 11, pp. 2573–2586, 2010, doi: https://doi.org/10.1016/j.jbankfin.2010.04.014.

J. W. Lee, J. Park, J. O, J. Lee, and E. Hong, “A Multiagent Approach to $Q$-Learning for Daily Stock Trading,” IEEE Transactions on Systems, Man, and Cybernetics - Part A: Systems and Humans, vol. 37, no. 6, pp. 864–877, 2007, doi: 10.1109/TSMCA.2007.904825.

S. A. Badran and M. Rezghi, “An adaptive synchronization approach for weights of deep reinforcement learning.” arXiv, 2020. doi: 10.48550/ARXIV.2008.06973.

https://github.com/junyoung-sim/quant

Published

02-28-2023

How to Cite

Sim, J., & Kirk, B. (2023). Generalized Deep Reinforcement Learning for Trading. Journal of Student Research, 12(1). https://doi.org/10.47611/jsrhs.v12i1.4316

Issue

Section

HS Research Projects