Abstract
Time series Forecasting has attracted attention over the last decade with the boost in processing power, the amount of data available and the development of more advanced algorithms. It is now widely used in a range of different fields including Medical Diagnostics, Weather Forecasting, Financial time series etc. In this paper, we propose a model of attention mechanism that allows for attended input to be fed to the model instead of the actual input. The motivation for the model is to show a new way to view the input so that the model can make more accurate predictions. The proposed LSTM model with the attention mechanism is then evaluated on common evaluation metrics and the results are compared with state of art models like CNN-LSTM and Stacked LSTM to show its benefits.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsReferences
Pal, A., Prakash, P.K.S.: Practical Time Series Analysis. Packt Publishing, Birmingham (2017)
Deshpande, B., Kotu, V.: Data Science. 2nd Ed. (2019)
Brockwell, P.J., Davis, R.A.: Introduction to Time Series and Forecasting. Springer, Heidelberg (2016). https://doi.org/10.1007/978-3-319-29854-2
Geron, A.: Hands-On Machine Learning with Scikit-Learn and TensorFlow Concepts, Tools, and Techniques to Build Intelligent Systems. O’Reilly Media, Sebastopol (2017)
Goodfellow, I., Bengio, Y., Courville, A.: Deep Learning. MIT Press, Cambridge (2017)
Bahdanau, D., Cho, K. , Bengio, Y.: Neural machine translation by jointly learning to align and translate. In: International Conference on Learning Representations (2015)
Tran, D.T., Iosifidis, A., Kanniainen, J., Gabbouj, M.: Temporal attention-augmented bilinear network for financial time-series data analysis. IEEE Trans. Neural Netw. Learn. Syst. 30(5), 1407–1418 (2018)
Qin, Y., et al.: A dual-stage attention-based recurrent neural network for time series prediction. arXiv preprint arXiv:1704.02971 (2017)
Wang, B., Huang, H., Wang, X.: A novel mining approach to financial time series forecasting. Neurocomputing 83, 136–145 (2012)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Mnih, V., Heess, N., Graves, A., Kavukcuoglu, K.: Recurrent models of visual attention. In: Proceedings of Advance Neural Information Processing System, pp. 2204–2212 (2014)
Xu, K., et al.: Show, attend and tell: Neural image caption generation with visual attention. In: Proceedings International Conference Machine Learning, pp. 2048–2057 (2015)
Chen, K., Wang, J., Chen, L.-C., Gao, H., Xu, W., Nevatia, R.: ABC-CNN: an attention based convolutional neural network for visual question answering (2015). https://arxiv.org/abs/1511.05960
Poterba, J.M., Summers, L.H.: Mean reversion in stock prices: evidence and implications. J. Finan. Econ. 22(1), 27–59 (1988)
Kohzadi, N., Boyd, M.S., Kermanshahi, B., Kaastra, I.: A comparison of artificial neural network and time series models for forecasting commodity prices. Neurocomputing 10(2), 169–181 (1996)
Kara, Y., Boyacioglu, M.A., Baykan, Ö.K.: Predicting direction of stock price index movement using artificial neural networks and support vector machines: the sample of the Istanbul stock exchange. Expert Syst. Appl. 38(5), 5311–5319 (2011)
Cao, L.J., Tay, F.E.H.: Support vector machine with adaptive parameters in financial time series forecasting. IEEE Trans. Neural Netw. 14(6), 1506–1518 (2003)
Graves, A., Mohamed, A.R., Hinton, G.: Speech recognition with deep recurrent neural networks. In: Proceedings of IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 6645–6649 (2013)
Hinton, G., et al.: Deep neural networks for acoustic modeling in speech recognition: the shared views of four research groups. IEEE Signal Process. Mag. 29(6), 82–97 (2012)
Gao, J., Pantel, P., Gamon, M., He, X., Deng, L.: Modeling interestingness with deep neural networks. In: Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pp. 2–13 (2014)
Shen, Y., He, X., Gao, J., Deng, L., Mesnil, G.: A latent semantic model with convolutional-pooling structure for information retrieval. In: Proceedings of the 23rd ACM International Conference on Conference on Information and Knowledge Management, pp. 101–110. ACM (2014)
Palangi, H., et al.: Deep sentence embedding using long short-term memory networks: analysis and application to information retrieval. In: IEEE/ACM Transactions on Audio, Speech and Language Processing (TASLP), vol. 24, no. 4, pp. 694–707 (2016)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2021 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Shali, Y., Brahma, B., Wadhvani, R., Gyanchandani, M. (2021). Attention LSTM for Time Series Forecasting of Financial Time Series Data. In: Misra, R., Kesswani, N., Rajarajan, M., Bharadwaj, V., Patel, A. (eds) Internet of Things and Connected Technologies. ICIoTCT 2020. Advances in Intelligent Systems and Computing, vol 1382. Springer, Cham. https://doi.org/10.1007/978-3-030-76736-5_8
Download citation
DOI: https://doi.org/10.1007/978-3-030-76736-5_8
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-76735-8
Online ISBN: 978-3-030-76736-5
eBook Packages: EngineeringEngineering (R0)