Abstract
This article attempts to establish a trading strategy framework based on deep neural networks for the futures market, which consists of two parts: time series forecasting and trading strategies based on trading signals. In the time series forecasting task, we experimented with three types of methods with different entry points, namely recurrent neural networks with gate structure, networks combining time and frequency domain information, and network structures using attention mechanism. In the trading strategy part, the buying and selling signals and the corresponding trading volume are established according to the prediction results, and trading is conducted with the frequency of hours. In the empirical exploration part, we tested the prediction effect and strategic rate of return of various models on the copper contract. The data shows that in general, the best strategy can obtain a relatively stable income growth that has nothing to do with market fluctuations, but lacks countermeasures for rare external events with greater impact.
Peking University Grant 2020: “New Ideas for Teaching 2.0” Key Project; MSTC Grant 2019YFC1521203: research, development and demonstration of key technologies for knowledge organization and services for Antiques based on Knowledge Graph; NSFC Grant 61772044.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Sezer, O.B., Gudelek, M.U., Ozbayoglu, A.M.: Financial time series forecasting with deep learning: a systematic literature review: 2005–2019. arXiv (2019)
Rumelhart, David E., Hinton, Geoffrey E., Williams, Ronald J.: Learning representations by back propagating errors. Nature 323(6088), 533–536 (1986)
Bengio, Y.: Learning long-term dependencies with gradient descent is difficult. IEEE Trans Neural Netw. 5, 157–166 (1994)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Graves, A., Mohamed, A.R., Hinton, G.: Speech recognition with deep recurrent neural networks. In: 2013 IEEE International Conference on Acoustics, Speech and Signal Processing (2013)
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. CoRR abs/1409.3215 (2014). http://arxiv.org/abs/1409.3215
Schuster, M., Paliwal, K.K.: Bidirectional recurrent neural networks. IEEE Trans. Signal Process. 45(11), 2673–2681 (1997)
Graves, A., Schmidhuber, J.: Offline arabic handwriting recognition with multidimensional recurrent neural networks. In: Advances in Neural Information Processing Systems 21, Proceedings of the Twenty-Second Annual Conference on Neural Information Processing Systems, Vancouver, British Columbia, Canada, 8–11 December 2008 (2008)
Koutník, J., Greff, K., Gomez, F., et al.: A Clockwork RNN. Computer Science, pp. 1863–1871 (2014)
Chang, S., Zhang, Y., Han, W., et al.: Dilated recurrent neural networks (2017)
Cui, Z., Chen, W., Chen, Y.: Multi-scale convolutional neural networks for time series classification (2016)
Chung, J., Ahn, S., Bengio, Y.: Hierarchical multiscale recurrent neural networks (2016)
Zhang, L., Aggarwal, C., Qi, G.-J.: Stock price prediction via discovering multi-frequency trading patterns. In: Proceedings of the 23rd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining-KDD 2017, ACM Press the 23rd ACM SIGKDD International Conference - Halifax, NS, Canada, 13–17 August 2017, pp. 2141–2149 (2017)
Wei, B., Yue, J., Rao, Y., et al.: A deep learning framework for financial time series using stacked autoencoders and long-short term memory. PLoS ONE 12(7), e0180944 (2017)
Hui, L., Tian, H.Q., Pan, D.F., et al.: Forecasting models for wind speed using wavelet, wavelet packet, time series and Artificial Neural Networks. Appl. Energy 107, 191–208 (2013)
Wang, J., Wang, Z., Li, J., et al.: Multilevel wavelet decomposition network for interpretable time series analysis (2018)
Mallat, S.G.: A theory for multiresolution signal decomposition: the wavelet representation. IEEE Trans. Pattern Anal. Mach. Intell. 11(7), 674–693 (1989)
Bahdanau, D., Cho, K., Bengio, Y.: Neural machine translation by jointly learning to align and translate (2014)
Luong, T., Pham, H., Manning, C.D.: Effective approaches to attention-based neural machine translation. In: Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing. Lisbon, Portugal: Association for Computational Linguistics, September 2015, pp. 1412–1421. https://www.aclweb.org/anthology/D15-1166
Vaswani, A., Shazeer, N., Parmar, N., et al.: Attention is all you need (2017)
Li, S., Jin, X., Xuan, Y., et al.: Enhancing the locality and breaking the memory bottleneck of transformer on time series forecasting (2019)
Heaton, J., Goodfellow, I., Bengio, Y., Courville, A.: Deep learning. Genet. Program. Evolvable Mach. https://doi.org/10.1007/s10710-017-9314-z
Voita, E., Talbot, D., Moiseev, F., et al.: Analyzing multi-head self-attention: specialized heads do the heavy lifting, the rest can be pruned. In: Proceedings of the 57th Annual Meeting of the Association for Computational Linguistics (2019). http://dx.doi.org/10.18653/v1/p19-1580
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Yao, S., Hongfei, Y., Siping, Y., Chong, C., Qi, S. (2020). Empirical Research on Futures Trading Strategy Based on Time Series Algorithm. In: Dou, Z., Miao, Q., Lu, W., Mao, J., Jia, G. (eds) Information Retrieval. CCIR 2020. Lecture Notes in Computer Science(), vol 12285. Springer, Cham. https://doi.org/10.1007/978-3-030-56725-5_9
Download citation
DOI: https://doi.org/10.1007/978-3-030-56725-5_9
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-56724-8
Online ISBN: 978-3-030-56725-5
eBook Packages: Computer ScienceComputer Science (R0)