Abstract
The role of the stock market across the overall financial market is indispensable. The way to acquire practical trading signals in the transaction process to maximize the benefits is a problem that has been studied for a long time. This paper put forward a theory of deep reinforcement learning in the stock trading decisions and stock price prediction, the reliability and availability of the model are proved by experimental data, and the model is compared with the traditional model to prove its advantages. From the point of view of stock market forecasting and intelligent decision-making mechanism, this paper proves the feasibility of deep reinforcement learning in financial markets and the credibility and advantages of strategic decision-making.
Similar content being viewed by others
References
Abtahi F, Zhu Z, Burry AM (2015) A deep reinforcement learning approach to character segmentation of license plate images. In: 2015 14th IAPR international conference on machine vision applications (MVA), pp 539–542. IEEE
Alimoradi MR, Kashan AH (2018) A league championship algorithm equipped with network structure and backward q-learning for extracting stock trading rules. Appl Soft Comput 68:478–493
Almahdi S, Yang SY (2017) An adaptive portfolio trading system: a risk-return portfolio optimization using recurrent reinforcement learning with expected maximum drawdown. Expert Syst Appl 87:267–279
Berutich JM, López F, Luna F, Quintana D (2016) Robust technical trading strategies using gp for algorithmic portfolio selection. Expert Syst Appl 46:307–315
Chang PC, Liao TW, Lin JJ, Fan CY (2011) A dynamic threshold decision system for stock trading signal detection. Appl Soft Comput 11(5):3998–4010
Chang V, Li T, Zeng Z (2019) Towards an improved adaboost algorithmic method for computational financial analysis. J Parallel Distrib Comput 134:219–232
Cheng CH, Chen TL, Wei LY (2010) A hybrid model based on rough sets theory and genetic algorithms for stock price forecasting. Inf Sci 180(9):1610–1629
Chien YWC, Chen YL (2010) Mining associative classification rules with stock trading data-a ga-based method. Knowl-Based Syst 23(6):605–614
Dulac-Arnold G, Evans R, van Hasselt H, Sunehag P, Lillicrap T, Hunt J, Mann T, Weber T, Degris T, Coppin B (2015) Deep reinforcement learning in large discrete action spaces. arXiv preprint arXiv:1512.07679
Foerster J, Assael IA, de Freitas N, Whiteson S (2016) Learning to communicate with deep multi-agent reinforcement learning. In: Advances in neural information processing systems, pp 2137–2145
Foerster JN, Farquhar G, Afouras T, Nardelli N, Whiteson S (2018) Counterfactual multi-agent policy gradients. In: Thirty-second AAAI conference on artificial intelligence
Guresen E, Kayakutlu G, Daim TU (2011) Using artificial neural network models in stock market index prediction. Expert Syst Appl 38(8):10389–10397
Hastie T, Rosset S, Zhu J, Zou H (2009) Multi-class adaboost. Stat Interface 2(3):349–360
jpmorgan. https://www.businessinsider.com/jpmorgan-takes-ai-use-to-the-next-level-2017-8
Koutník J, Schmidhuber J, Gomez F (2014) Online evolution of deep convolutional network for vision-based reinforcement learning. In: International conference on simulation of adaptive behavior, pp 260–269. Springer
Krollner B, Vanstone BJ, Finnie GR (2010) Financial time series forecasting with machine learning techniques: a survey. ESANN 2010, 18th European Symposium on Artificial Neural Networks, Bruges, Belgium, April 28–30, 2010, Proceedings. https://www.elen.ucl.ac.be/Proceedings/esann/esannpdf/es2010-50.pdf
Lange S, Riedmiller M (2010) Deep auto-encoder neural networks in reinforcement learning. In: The 2010 international joint conference on neural networks (IJCNN), pp 1–8. IEEE
Lange S, Riedmiller M, Voigtlander A (2012) Autonomous reinforcement learning on raw visual input data in a real world application. In: The 2012 international joint conference on neural networks (IJCNN), pp 1–8. IEEE
Liao Z, Wang J (2010) Forecasting model of global stock index by stochastic time effective neural network. Expert Syst Appl 37(1):834–841
Lillicrap TP, Hunt JJ, Pritzel A, Heess N, Erez T, Tassa Y, Silver D, Wierstra D (2015) Continuous control with deep reinforcement learning. arXiv preprint arXiv:1509.02971
Mabu S, Obayashi M, Kuremoto T (2015) Ensemble learning of rule-based evolutionary algorithm using multi-layer perceptron for supporting decisions in stock trading problems. Appl Soft Comput 36:357–367
Mnih V, Kavukcuoglu K, Silver D, Graves A, Antonoglou I, Wierstra D, Riedmiller M (2013) Playing atari with deep reinforcement learning. arXiv preprint arXiv:1312.5602
Mnih V, Kavukcuoglu K, Silver D, Rusu AA, Veness J, Bellemare MG, Graves A, Riedmiller M, Fidjeland AK, Ostrovski G et al (2015) Human-level control through deep reinforcement learning. Nature 518(7540):529
Schaul T, Quan J, Antonoglou I, Silver D (2015) Prioritized experience replay. arXiv preprint arXiv:1511.05952
Schulman J, Levine S, Abbeel P, Jordan M, Moritz P (2015) Trust region policy optimization. In: International conference on machine learning, pp 1889–1897
Shibata K, Iida M (2003) Acquisition of box pushing by direct-vision-based reinforcement learning. In: SICE 2003 annual conference, vol 3, pp 2322–2327. IEEE
Shibata K, Okabe Y (1997) Reinforcement learning when visual sensory signals are directly given as inputs. In: International conference on neural networks, 1997. vol 3, pp 1716–1720. IEEE
Silver D, Lever G, Heess N, Degris T, Wierstra D, Riedmiller M (2014) Deterministic policy gradient algorithms. In: ICML
Van Hasselt H, Guez A, Silver D (2016) Deep reinforcement learning with double q-learning. In: AAAI, vol 2, p 5. Phoenix, AZ
Vanstone B, Finnie G, Hahn T (2012) Creating trading systems with fundamental variables and neural networks: the aby case study. Math Comput Simul 86:78–91
Wang J, Hou R, Wang C, Shen L (2016) Improved v-support vector regression model based on variable selection and brain storm optimization for stock price forecasting. Appl Soft Comput 49:164–178
Wang JZ, Wang JJ, Zhang ZG, Guo SP (2011) Forecasting stock indices with back propagation neural network. Expert Syst Appl 38(11):14346–14355
Wang Z, Schaul T, Hessel M, Van Hasselt H, Lanctot M, De Freitas N (2015) Dueling network architectures for deep reinforcement learning. arXiv preprint arXiv:1511.06581
Wymann B, Espié E, Guionneau C, Dimitrakakis C, Coulom R, Sumner A (2000) Torcs, the open racing car simulator. Software available at http://torcs.sourceforge.net. Accessed 3 July 2019
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Li, Y., Ni, P. & Chang, V. Application of deep reinforcement learning in stock trading strategies and stock forecasting. Computing 102, 1305–1322 (2020). https://doi.org/10.1007/s00607-019-00773-w
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00607-019-00773-w