Skip to main content
Log in

Application of deep reinforcement learning in stock trading strategies and stock forecasting

  • Published:
Computing Aims and scope Submit manuscript

Abstract

The role of the stock market across the overall financial market is indispensable. The way to acquire practical trading signals in the transaction process to maximize the benefits is a problem that has been studied for a long time. This paper put forward a theory of deep reinforcement learning in the stock trading decisions and stock price prediction, the reliability and availability of the model are proved by experimental data, and the model is compared with the traditional model to prove its advantages. From the point of view of stock market forecasting and intelligent decision-making mechanism, this paper proves the feasibility of deep reinforcement learning in financial markets and the credibility and advantages of strategic decision-making.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

References

  1. Abtahi F, Zhu Z, Burry AM (2015) A deep reinforcement learning approach to character segmentation of license plate images. In: 2015 14th IAPR international conference on machine vision applications (MVA), pp 539–542. IEEE

  2. Alimoradi MR, Kashan AH (2018) A league championship algorithm equipped with network structure and backward q-learning for extracting stock trading rules. Appl Soft Comput 68:478–493

    Article  Google Scholar 

  3. Almahdi S, Yang SY (2017) An adaptive portfolio trading system: a risk-return portfolio optimization using recurrent reinforcement learning with expected maximum drawdown. Expert Syst Appl 87:267–279

    Article  Google Scholar 

  4. Berutich JM, López F, Luna F, Quintana D (2016) Robust technical trading strategies using gp for algorithmic portfolio selection. Expert Syst Appl 46:307–315

    Article  Google Scholar 

  5. Chang PC, Liao TW, Lin JJ, Fan CY (2011) A dynamic threshold decision system for stock trading signal detection. Appl Soft Comput 11(5):3998–4010

    Article  Google Scholar 

  6. Chang V, Li T, Zeng Z (2019) Towards an improved adaboost algorithmic method for computational financial analysis. J Parallel Distrib Comput 134:219–232

    Article  Google Scholar 

  7. Cheng CH, Chen TL, Wei LY (2010) A hybrid model based on rough sets theory and genetic algorithms for stock price forecasting. Inf Sci 180(9):1610–1629

    Article  Google Scholar 

  8. Chien YWC, Chen YL (2010) Mining associative classification rules with stock trading data-a ga-based method. Knowl-Based Syst 23(6):605–614

    Article  Google Scholar 

  9. Dulac-Arnold G, Evans R, van Hasselt H, Sunehag P, Lillicrap T, Hunt J, Mann T, Weber T, Degris T, Coppin B (2015) Deep reinforcement learning in large discrete action spaces. arXiv preprint arXiv:1512.07679

  10. Foerster J, Assael IA, de Freitas N, Whiteson S (2016) Learning to communicate with deep multi-agent reinforcement learning. In: Advances in neural information processing systems, pp 2137–2145

  11. Foerster JN, Farquhar G, Afouras T, Nardelli N, Whiteson S (2018) Counterfactual multi-agent policy gradients. In: Thirty-second AAAI conference on artificial intelligence

  12. Guresen E, Kayakutlu G, Daim TU (2011) Using artificial neural network models in stock market index prediction. Expert Syst Appl 38(8):10389–10397

    Article  Google Scholar 

  13. Hastie T, Rosset S, Zhu J, Zou H (2009) Multi-class adaboost. Stat Interface 2(3):349–360

    Article  MathSciNet  Google Scholar 

  14. jpmorgan. https://www.businessinsider.com/jpmorgan-takes-ai-use-to-the-next-level-2017-8

  15. Koutník J, Schmidhuber J, Gomez F (2014) Online evolution of deep convolutional network for vision-based reinforcement learning. In: International conference on simulation of adaptive behavior, pp 260–269. Springer

  16. Krollner B, Vanstone BJ, Finnie GR (2010) Financial time series forecasting with machine learning techniques: a survey. ESANN 2010, 18th European Symposium on Artificial Neural Networks, Bruges, Belgium, April 28–30, 2010, Proceedings. https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2010-50.pdf

  17. Lange S, Riedmiller M (2010) Deep auto-encoder neural networks in reinforcement learning. In: The 2010 international joint conference on neural networks (IJCNN), pp 1–8. IEEE

  18. Lange S, Riedmiller M, Voigtlander A (2012) Autonomous reinforcement learning on raw visual input data in a real world application. In: The 2012 international joint conference on neural networks (IJCNN), pp 1–8. IEEE

  19. Liao Z, Wang J (2010) Forecasting model of global stock index by stochastic time effective neural network. Expert Syst Appl 37(1):834–841

    Article  Google Scholar 

  20. Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2015) Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971

  21. Mabu S, Obayashi M, Kuremoto T (2015) Ensemble learning of rule-based evolutionary algorithm using multi-layer perceptron for supporting decisions in stock trading problems. Appl Soft Comput 36:357–367

    Article  Google Scholar 

  22. Mnih V, Kavukcuoglu K, Silver D, Graves A, Antonoglou I, Wierstra D, Riedmiller M (2013) Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602

  23. Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529

    Article  Google Scholar 

  24. Schaul T, Quan J, Antonoglou I, Silver D (2015) Prioritized experience replay. arXiv preprint arXiv:1511.05952

  25. Schulman J, Levine S, Abbeel P, Jordan M, Moritz P (2015) Trust region policy optimization. In: International conference on machine learning, pp 1889–1897

  26. Shibata K, Iida M (2003) Acquisition of box pushing by direct-vision-based reinforcement learning. In: SICE 2003 annual conference, vol 3, pp 2322–2327. IEEE

  27. Shibata K, Okabe Y (1997) Reinforcement learning when visual sensory signals are directly given as inputs. In: International conference on neural networks, 1997. vol 3, pp 1716–1720. IEEE

  28. Silver D, Lever G, Heess N, Degris T, Wierstra D, Riedmiller M (2014) Deterministic policy gradient algorithms. In: ICML

  29. Van Hasselt H, Guez A, Silver D (2016) Deep reinforcement learning with double q-learning. In: AAAI, vol 2, p 5. Phoenix, AZ

  30. Vanstone B, Finnie G, Hahn T (2012) Creating trading systems with fundamental variables and neural networks: the aby case study. Math Comput Simul 86:78–91

    Article  MathSciNet  Google Scholar 

  31. Wang J, Hou R, Wang C, Shen L (2016) Improved v-support vector regression model based on variable selection and brain storm optimization for stock price forecasting. Appl Soft Comput 49:164–178

    Article  Google Scholar 

  32. Wang JZ, Wang JJ, Zhang ZG, Guo SP (2011) Forecasting stock indices with back propagation neural network. Expert Syst Appl 38(11):14346–14355

    Google Scholar 

  33. Wang Z, Schaul T, Hessel M, Van Hasselt H, Lanctot M, De Freitas N (2015) Dueling network architectures for deep reinforcement learning. arXiv preprint arXiv:1511.06581

  34. Wymann B, Espié E, Guionneau C, Dimitrakakis C, Coulom R, Sumner A (2000) Torcs, the open racing car simulator. Software available at http://torcs.sourceforge.net. Accessed 3 July 2019

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Victor Chang.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Li, Y., Ni, P. & Chang, V. Application of deep reinforcement learning in stock trading strategies and stock forecasting. Computing 102, 1305–1322 (2020). https://doi.org/10.1007/s00607-019-00773-w

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00607-019-00773-w

Keywords

Mathematics Subject Classification

Navigation