Skip to main content
Log in

Predicting the number of customer transactions using stacked LSTM recurrent neural networks

  • Original Article
  • Published:
Social Network Analysis and Mining Aims and scope Submit manuscript

Abstract

Time series forecasting is used to predict future values on a data sequence using time-based data. Time series of transactions or count data are difficult to predict because of their complex nonlinear patterns. The purpose of this study is to provide a solution for predicting the number of upcoming transactions in order to predict the anomaly of the system, with the purpose of checking for why the number of transactions in the next day to be lower than expected. This can be a kind of customer churn prediction and can help system administrators identify and prevent potential losses and can improve existing infrastructure. Required data were collected from 780 company, which contains 353 rows of transactions in 2018 and has two variables, date and number of transactions; the number of transactions has been set as a target to predict. The autoregressive integrated moving average (ARIMA) model cannot deal with nonlinear relationships, while the neural network model, conversely, is capable of processing nonlinear patterns. In this study, a deep learning approach is presented. The approach is a stacked long short-term memory (LSTM) structure, a model derived from recurrent neural networks. To find the most accurate prediction model, the performance measures of various recurrent neural network, PROPHET, and ARIMA models are compared using the same dataset. Experimental results show that for predicting the number of transactions, the stacked LSTM model is better than other approaches.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Notes

  1. Decline Curve Analysis.

  2. Auto regression integrated moving average.

  3. Gated Recurrent Unit.

References

  • Aczon M, et al., (2017) Dynamic mortality risk predictions in pediatric critical care using recurrent neural networks.

  • Agrawal K et al (2017) Fusion of statistical and machine learning approaches for time series prediction using earth observation data. Int J Comput Sci Eng 14(3):255–266

    Google Scholar 

  • Alkharif S, Lee K, Kim H (2018) Time-series analysis for price prediction of opportunistic cloud computing resources. in Proceedings of the 7th International Conference on Emerging Databases.

  • Ariyo AA, Adewumi AO, Ayo CK (2014) Stock price prediction using the ARIMA model. In 2014 UKSim-AMSS 16th International Conference on Computer Modelling and Simulation. IEEE.

  • Azoff EM (1994) Neural network time series forecasting of financial markets. John Wiley & Sons Inc

    Google Scholar 

  • Ballesteros M et al (2017) Greedy transition-based dependency parsing with stack lstms. Comput Linguist 43(2):311–347

    Article  MathSciNet  Google Scholar 

  • Bengio Y, Courville A, Vincent P (2013) Representation learning: A review and new perspectives. IEEE Trans Pattern Anal Mach Intell 35(8):1798–1828

    Article  Google Scholar 

  • Cao L-J, Tay FEH (2003) Support vector machine with adaptive parameters in financial time series forecasting. IEEE Trans Neural Networks 14(6):1506–1518

    Article  Google Scholar 

  • De Gooijer JG, Hyndman RJ (2006) 25 years of time series forecasting. Int J Forecast 22(3):443–473

    Article  Google Scholar 

  • Du S. et al. (2020) Multivariate Time Series Forecasting via Attention-based Encoder-Decoder Framework. Neurocomputing.

  • Fan Z-P, Che Y-J, Chen Z-Y (2017) Product sales forecasting using online reviews and historical sales data: A method combining the Bass model and sentiment analysis. J Bus Res 74:90–100

    Article  Google Scholar 

  • Faruk DÖ (2010) A hybrid neural network and ARIMA model for water quality time series prediction. Eng Appl Artif Intell 23(4):586–594

    Article  MathSciNet  Google Scholar 

  • Fu R, Zhang Z, Li L (2016) Using LSTM and GRU neural network methods for traffic flow prediction. in 2016 31st Youth Academic Annual Conference of Chinese Association of Automation (YAC). IEEE.

  • González-Caballero E, et al. (2021) Continuous linguistic variables and their applications to data mining and time series prediction. International Journal of Fuzzy Systems, 1–22.

  • Graves A, Mohamed A, Hinton G (2013) Speech recognition with deep recurrent neural networks. In 2013 IEEE international conference on acoustics, speech and signal processing. IEEE

  • Greff K et al (2016) LSTM: A search space odyssey. IEEE Trans Neural Networks Learn Syst 28(10):2222–2232

    Article  MathSciNet  Google Scholar 

  • Guo J et al (2021) Mechanical fault time series prediction by using EFMSAE-LSTM neural network. Measurement 173:108566

    Article  Google Scholar 

  • Guo T, et al. (2016) Robust online time series prediction with recurrent neural networks. In 2016 IEEE International Conference on Data Science and Advanced Analytics (DSAA). IEEE.

  • Hagenau M, Liebmann M, Neumann D (2013) Automated news reading: Stock price prediction based on financial news using context-capturing features. Decis Support Syst 55(3):685–697

    Article  Google Scholar 

  • Ho S, Xie M (1998) The use of ARIMA models for reliability forecasting and analysis. Comput Ind Eng 35(1–2):213–216

    Article  Google Scholar 

  • Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780

    Article  Google Scholar 

  • Hüsken M, Stagge P (2003) Recurrent neural networks for time series classification. Neurocomputing 50:223–235

    Article  Google Scholar 

  • Kamari A et al (2017) Decline curve based models for predicting natural gas well performance. Petroleum 3(2):242–248

    Article  Google Scholar 

  • Krause B, et al. (2016) Multiplicative LSTM for sequence modelling. arXiv preprint arXiv:1609.07959.

  • Längkvist M, Karlsson L, Loutfi A (2014) A review of unsupervised feature learning and deep learning for time-series modeling. Pattern Recogn Lett 42:11–24

    Article  Google Scholar 

  • Li J et al (2021) DTDR–ALSTM: Extracting dynamic time-delays to reconstruct multivariate data for improving attention-based LSTM industrial time series prediction models. Knowl Based Syst 211:106508

    Article  Google Scholar 

  • Mahalakshmi G, Sridevi S, Rajaram S (2016) A survey on forecasting of time series data. In 2016 International Conference on Computing Technologies and Intelligent Data Engineering (ICCTIDE'16). IEEE.

  • Majidpour M et al (2019) Fast univariate time series prediction of solar power for real-time control of energy storage system. Forecasting 1(1):107–120

    Article  Google Scholar 

  • Parmezan ARS, Souza VM, Batista GE (2019) Evaluation of statistical and machine learning models for time series prediction: Identifying the state-of-the-art and the best conditions for the use of each model. Inf Sci 484:302–337

    Article  Google Scholar 

  • Pascanu R, Mikolov T, Bengio Y (2013) On the difficulty of training recurrent neural networks. in International conference on machine learning.

  • Rabby MF et al (2021) Stacked LSTM based deep recurrent neural network with kalman smoothing for blood glucose prediction. BMC Med Inform Decis Mak 21(1):1–15

    Article  Google Scholar 

  • Ramos P, Santos N, Rebelo R (2015) Performance of state space and ARIMA models for consumer retail sales forecasting. Robotics Comput-Integr Manuf 34:151–163

    Article  Google Scholar 

  • Ruan L, et al. (2021) Workload time series prediction in storage systems: a deep learning based approach. Cluster Computing, p. 1–11.

  • Sapankevych NI, Sankar R (2009) Time series prediction using support vector machines: a survey. IEEE Comput Intell Mag 4(2):24–38

    Article  Google Scholar 

  • Wei X et al (2021) Machine learning for pore-water pressure time-series prediction: Application of recurrent neural networks. Geosci Front 12(1):453–467

    Article  Google Scholar 

  • Xiao Y et al (2021) A dual-stage attention-based Conv-LSTM network for spatio-temporal correlation and multivariate time series prediction. Int J Intell Syst 36(5):2036–2057

    Article  Google Scholar 

  • Yenidoğan I, et al. (2018) Bitcoin Forecasting Using ARIMA and PROPHET. In 2018 3rd International Conference on Computer Science and Engineering (UBMK). IEEE.

Download references

Acknowledgements

The authors express their sincere gratitude to 780 Co. for sharing their transaction database.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to M. V. Sebt.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Sebt, M.V., Ghasemi, S.H. & Mehrkian, S.S. Predicting the number of customer transactions using stacked LSTM recurrent neural networks. Soc. Netw. Anal. Min. 11, 86 (2021). https://doi.org/10.1007/s13278-021-00805-4

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s13278-021-00805-4

Keywords

Navigation