Abstract
Dynamic resource allocation and auto-scaling represent effective solutions for many cloud challenges, such as over-provisioning (i.e., energy-wasting, and Service level Agreement “SLA” violation) and under-provisioning (i.e., Quality of Service “QoS” dropping) of resources. Early workload prediction techniques play an important role in the success of these solutions. Unfortunately, no prediction technique is perfect and suitable enough for most workloads, particularly in cloud environments. Statistical and machine learning techniques may not be appropriate for predicting workloads, due to instability and dependency of cloud resources’ workloads. Although Recurrent Neural Network (RNN) deep learning technique considers these shortcomings, it provides poor results for long-term prediction. On the other hand, Sequence-to-Sequence neural machine translation technique (Seq2Seq) is effectively used for translating long texts. In this paper, workload sequence prediction is treated as a translation problem. Therefore, an Attention Seq2Seq-based technique is proposed for predicting cloud resources’ workloads. To validate the proposed technique, real-world dataset collected from a Google cluster of 11 k machines is used. For improving the performance of the proposed technique, a novel procedure called cumulative-validation is proposed as an alternative procedure to cross-validation. Results show the effectiveness of the proposed technique for predicting workloads of cloud resources in terms of accuracy by 98.1% compared to 91% and 85% for other sequence-based techniques, i.e. Continuous Time Markov Chain based models and Long short-term memory based models, respectively. Also, the proposed cumulative-validation procedure achieves a computational time superiority of 57% less compared to the cross-validation with a slight variation of 0.006 in prediction accuracy.
Article PDF
Similar content being viewed by others
References
Al-Sayed, M.M., Hassan, H.A., Omara, F.A.: Towards evaluation of cloud ontologies. J. Parallel Distrib. Comput. 126, 82–106 (2019)
Al-Sayed, M.M., Hassan, H.A., Omara, F.A.: CloudFNF: an ontology structure for functional and non-functional. J. Parallel Distrib. Comput. 14, 143–173 (2020)
Al-Sayed, M.M., Hassan, H.A., Omara, F.A.: Mapping lexical gaps in cloud ontology using BabelNet and FP-growth. Int. J. Comput. Sci. Secur. (IJCSS). 13(2), 36–52 (2019)
Zharikov, E., Telenyk, S., Bidyuk, P.: Adaptive workload forecasting in cloud data centers. J. Grid Comput. 18(1), 149–168 (2020)
Chen, Z., Hu, J., Min, G., Zomaya, A.Y., El-Ghazawi, T.: Towards accurate prediction for high-dimensional and highly-variable cloud workloads with deep learning. IEEE Trans. Parallel Distrib. Syst. 31(4), 923–934 (2020)
Amiri, M., Mohammad-Khanli, L.: Survey on prediction models of applications for resources provisioning in cloud. J. Netw. Comput. Appl. 82, 93–113 (2017)
Dabbagh, M., Hamdaoui, B., Guizani, M., Rayes, A.: Energy-efficient resource allocation and provisioning framework for cloud data centers. IEEE Trans. Netw. Serv. Manag. 12(3), 377–391 (2015)
Kumar, J., Goomer, R., Singh, A.K.: Long short term memory recurrent neural network (LSTM-RNN) based workload forecasting model for cloud datacenters. Procedia Comput. Sci. 125, 676–682 (2018)
Dabbagh, M., Hamdaoui, B., Guizani, M., Rayes, A.: An energy-efficient vm prediction and migration framework for overcommitted clouds. IEEE Trans. Cloud Comput. 6(4), 955–966 (2016)
Kumar, J., Singh, A.K., Buyya, R.: Self directed learning based workload forecasting model for cloud resource management. Inf. Sci. 543, 345–366 (2020)
GUREYA, D.D.: Resource Allocation for Data-Intensive Services in the Cloud. Universitetsservice US-AB, Stockholm (2021)
Schmidt, T., Marg, L.: How to Move to Neural Machine Translation for Enterprise-Scale Programs—An Early Adoption Case Study. In: Proceedings of the 21st Annual Conference of the European Association for Machine Translation, Alacant, Spain 2018
Al-Sayed, M.M., Khattab, S., Omarab, F.A.: Prediction mechanisms for monitoring state of cloud resources using Markov chain model. J. Parallel Distrib. Comput. 96, 163–171 (2016)
Amiri, M., Mohammad-Khanli, L., Mirandola, R.: An online learning model based on episode mining for workload prediction in cloud. Futur. Gener. Comput. Syst. 87, 83–101 (2018)
Saxena, D., Chauhan, R., Kait, R.: Dynamic fair priority optimization task scheduling algorithm in cloud computing: concepts and implementations. Int. J. Comput. Netw. Inf. Secur. 8(2), 41–48 (2016)
Saxena, D., Singh, A.K.: Communication cost aware resource efficient load balancing (care-lb) framework for cloud datacenter. Recent Adv. Comput. Sci. Commun. 12, 1–14 (2020)
Zhang, F., Liu, G., Fu, X., Yahyapour, R.: A survey on virtual machine migration: challenges techniques and open issues. IEEE Commun. Surv. Tutor. 20(2), 1206–1243 (2018)
Kumar, J., Singh, A.K.: Workload prediction in cloud using artificial neural network and adaptive differential evolution. Futur. Gener. Comput. Syst. 81, 41–52 (2018)
Kirchoff, D.F., Xavier, M., Mastella, J., De Rose, C.A.: A Preliminary Study of Machine Learning Workload Prediction Techniques for Cloud Applications. In: 27th Euromicro International Conference on Parallel, Distributed and Network-Based Processing (PDP), Pavia, Italy, Italy Feb. 2019
Tang, X.: Large-scale computing systems workload prediction using parallel improved LSTM neural network. IEEE Access. 7, 40525–40533 (2019)
Singh, P., Gupta, P., Jyoti, K.: TASM: technocrat ARIMA and SVR model for workload prediction of web applications in cloud. Clust. Comput. 22, 619–633 (2019)
Kim, I.K., Wang, W., Qi, Y., Humphrey, M.: Forecasting cloud application workloads with cloudinsight for predictive resource management. IEEE Trans. Cloud Comput (Early Access). 1–16 (2020)
Kumar, J., Singh, A.K., Buyya, R.: Self directed learning based workload forecasting model for cloud resource management. Inf. Sci. 543, 345–366 (2021)
Daraghmeh, M., Melhem, S.B., Agarwal A., Goel, N., Zaman, M.: Regression-Based Dynamic Provisioning and Monitoring for Responsive Resources in Cloud Infrastructure Networks. In: IEEE/ACS 15th International Conference on Computer Systems and Application (AICCSA), Jordan (2018)
Al-Dhuraibi, Y., Paraiso, F., Djarallah, N., Merle, P.: Elasticity in cloud computing: state of the art and research challenges. IEEE Trans. Serv. Comput. 11(2), 430–447 (2018)
Daraghmeh, M., Agarwal, A., Manzano, R., Zaman, M.: Time Series Forecasting Using Facebook Prophet for Cloud Resource Management. In: IEEE International Conference on Communications Workshops (ICC Workshops), Canada (2021)
de Nardin, I.F., da Righi, R.R., Lopes, T.R., Costa, C., YoungYeom, H.: On revisiting energy and performance in microservices applications: A cloud elasticity-driven approach. Parallel Comput. 108, 102858 (2021)
Dabbagh, M., Hamdaoui, B., Guizani, M., Rayes, A.: Exploiting task elasticity and price heterogeneity for maximizing cloud computing profits. IEEE Trans. Emerg. Top. Comput. 6(1), 85–96 (2015)
Malinowski, K., Mande, J.: Lionbridge Translation Technologies. [Online]. Available: https://www.lionbridge.com/blog/translation-localization/the-future-of-language-technology-the-future-of-machine-translation/ (23 12 2021). Accessed 12 2021
Dai, H.: Comparative Analysis of Machine Translation and Human Translation under the Background of Internet. In: International Conference on Cognitive Based Information Processing and Applications (CIPA), Singapore (2021)
Choudhary, H., Pathak, A.K., Shah, R.R., Kumaraguru, P.: Neural Machine Translation for English-Tamil. In: Proceedings of the Third Conference on Machine Translation (WMT), Brussels, Belgium (2018)
Yang, J., Yin, Y., Ma, S., Huang, H., Zhang, D., Li, Z., Wei, F.: Multilingual agreement for multilingual neural machine translation. In: Proceedings of the 59th annual meeting of the Association for Computational Linguistics and the 11th international joint conference on natural language processing (ACL-IJCNLP), Online (2021)
Cromieres, F., Nakazawa, T., Dabre, R.: Neural Machine Translation: Basics, Practical Aspects and Recent Trends. In: Proceedings of the the 8th International Joint Conference on Natural Language Processing (IJCNLP), Taiwan (2017)
Guo, J., Chang, Z., Wang, S., Ding, H., Feng, Y., Mao, L., Bao, Y.: Who Limits the Resource Efficiency of my Datacenter: an Analysis of Alibaba Datacenter Traces. In: IEEE/ACM 27th International Symposium on Quality of Service (IWQoS), Phoenix, AZ, USA, USA (2019)
Zhu, Y., Zhang, W., Chen, Y., Gao, H.: A novel approach to workload prediction using attention-based LSTM encoder-decoder network in cloud environment. EURASIP J. Wirel. Commun. Netw. 274, 1–18 (2019)
Hannan, E.J., Kavalieris, L.: REGRESSION, AUTOREGRESSION MODELS. J. Time Ser. Anal. 7(1), 27–49 (1986)
Robinson, P.M.: The estimation of a nonlinear moving average model. Stoch. Process. Appl. 5(1), 81–90 (1977)
Nelson, B.K.: Time series analysis using autoregressive integrated moving average (ARIMA) models. Acad. Emerg. Med. 5(7), 739–744 (1998)
Bernardo, J.M., Smith, A.F.: Bayesian theory, 1st edn. New York: John Wiley and Sons (2009)
Peterson, L.E.: K-nearest neighbor. Scholarpedia. 4(2), 1883 (2009)
Alpaydin, E.: Introduction to Machine Learning, 4th edn. MIT Press, United States (2020)
Bergmeir, C., Benítez, J.M.: On the use of cross-validation for time series predictor evaluation. Inf. Sci. 191, 192–213 (2012)
Tashman, L.J.: Out-of-sample tests of forecasting accuracy: an analysis and review. Int. J. Forecast. 16(4), 437–450 (2000)
Singh, G.B.: Fundamentals of Bioinformatics and Computational Biology, vol. 6, pp. 97–125. Springer International Publishing, Switzerland (2015)
Hochreiter, S., Schmidhuber, J.: Long short-term memory. Neural Comput. 9(8), 1735–1780 (1997)
Sutskever, I., Vinyals, O., Le, Q.V.: Sequence to sequence learning with neural networks. Adv. Neural Inf. Proces. Syst. 27, 3104–3112 (2014)
Goyal, P., Pandey, S., Jain, K.: Deep Learning for Natural Language Processing: Creating Neural Networks with Python. Apress, Berkeley (2018)
Schafer, R.W.: On the Frequency-Domain Properties of Savitzky-Golay Filters. In: Digital Signal Processing and Signal Processing Education Meeting (DSP/SPE), USA 2011
Farahnakian, F., Liljeberg, P., Plosila, J.: LiRCUP: Linear Regression Based CPU Usage Prediction Algorithm for Live Migration of Virtual Machines in Data Centers. In: 39th Euromicro Conference on Software Engineering and Advanced Applications, Santander, Spain (October 2013)
Islam, S., Keung, J., Lee, K., Liu, A.: Empirical prediction models for adaptive resource provisioning in the cloud. Futur. Gener. Comput. Syst. 28(1), 155–162 (2012)
Singh, S., Chana, I.: A survey on resource scheduling in cloud computing: issues and challenges. Journal of Grid Computing. 14(2), 217–264 (2016)
Chen, X., Wang, H., Ma, Y., Zheng, X., Guo, L.: Self-adaptive resource allocation for cloud-based software services based on iterative QoS prediction model. Futur. Gener. Comput. Syst. 105, 287–296 (2020)
Yoo, W., Sim, A.: Time-series forecast modeling on high-bandwidth network measurements. J. Grid Comput. 14, 463–476 (2016)
Naseera, S., Rajini, G.K., Prabha, N.A., Abhishek, G.: A comparative study on CPU load predictions in a computational grid using artificial neural network algorithms. Indian J. Sci. Technol. 8(35), 1–5 (2015)
Yadav, M.P.; Pal, N.; Yadav, D.K.: Workload Prediction over Cloud Server Using Time Series Data. In: 11th International Conference on Cloud Computing, Data Science & Engineering (Confluence), Noida, India (2021)
Désiré, K.K., Francis, K.A., Kouassi, K.H., Dhib, E., Tabbane, N., Asseu, O.: Fractional rider deep long short term memory network for workload prediction-based distributed resource allocation using spark in cloud gaming. Engineering. 13, 135–157 (2021)
Bi, J., Li, S., Yuan, H., Zhou, M.C.: Integrated deep learning method for workload and resource prediction in cloud systems. Neurocomputing. 424, 35–48 (2021)
Xue, J., Yan, F., Birke, R., Chen, L.Y., Scherer, T., Smirni, E.: PRACTISE: Robust Prediction of Data Center Time Series. In: 11th International Conference on Network and Service Management (CNSM), Barcelona, Spain (2015)
Box, G.E.P., Jenkins, G.M., Reinsel, G.C., Ljung, G.M.: Time Series Analysis: Forecasting and Control, 5th edn. Wiley, Hoboken (2015)
Hayashi, S., Tanimoto, A., Kashima, H.: Long-Term Prediction of Small Time-Series Data Using Generalized Distillation. In: International Joint Conference on Neural Networks (IJCNN), Budapest, Hungary, Hungary (July 2019)
Taieb, S.B.; Bontempi, G., Sorjamaa, A., Lendasse, A.: Long-Term Prediction of Time Series by Combining Direct and MIMO Strategies. In: International Joint Conference on Neural Networks, Atlanta, GA, USA (June 2009)
Mathworks documentation team, "Sliding Window Method and Exponential Weighting Method," The Mathworks, (1994-2022). [Online]. Available: https://www.mathworks.com/help/dsp/ug/sliding-window-method-and-exponential-weighting-method.html#:~:text=In%20the%20sliding%20window%20method,the%20Len%20%2D%201%20previous%20samples. Accessed 29 Apr 2022
Tang, X., Liao, X., Zheng, J.: Energy efficient job scheduling with workload prediction on cloud data center. Clust. Comput. 21(3), 1581–1593 (2018)
Tosun, E., Aydin, K., Bilgili, M.: Comparison of linear regression and artificial neural network model of a diesel engine fueled with biodiesel-alcohol mixtures. Alex. Eng. J. 55(4), 3081–3089 (2016)
Zuur, A.F., Ieno, E.N., Walker, N.J., Saveliev, A.A., Smith, G.M.: Limitations of Linear Regression Applied on Ecological Data. In: Mixed Effects Models and Extensions in Ecology with R. Statistics for Biology and Health, New York, Springer, pp. 11–33 (2009)
Amiri, M., Mohammad-Khanli, L., Mirandola, R.: A sequential pattern mining model for application workload prediction in cloud environment. J. Netw. Comput. Appl. 105, 21–62 (2018)
Amiri, M., Mohammad-Khanli, L., Mirandola, R.: A new efficient approach for extracting the closed episodes for workload prediction in cloud. Computing. 102, 141–200 (2020)
Amiri, M., Feizi-Derakhshi, M., Mohammad-Khanli, L.: IDS fitted Q improvement using fuzzy approach for resource provisioning in cloud. J. Intell. Fuzzy Syst. 32(1), 229–240 (2017)
Wang, Z., Hong, T.: Reinforcement learning for building controls: the opportunities and challenges. Appl. Energy. 269, 115036 (2020)
Bi, J., Yuan, H., Zhou, M.: Temporal prediction of multiapplication consolidated workloads in distributed clouds. IEEE Trans. Autom. Sci. Eng. 16(4), 1763–1773 (2019)
Sovic, A., Sersic, D.: Signal decomposition methods for reducing drawbacks of the DWT. Eng. Rev. 32(2), 70–77 (2012)
Kumar, J., Singh, A.K.: Performance evaluation of metaheuristics algorithms for workload prediction in cloud environment. Appl. Soft Comput. 113(Part A), 1–14 (2021)
Roodschild, M., Sardiñas, J.G., Will, A.: A new approach for the vanishing gradient problem on sigmoid. Prog. Artif. Intell. 9(4), 351–360 (2020)
Wang, X., Qin, Y., Wang, Y., Xiang, S., Chen, H.: ReLTanh: an activation function with vanishing gradient resistance for SAE-based DNNs and its application to rotating machinery fault diagnosis. Neurocomputing. 363, 88–98 (2019)
Bahdanau, D., Cho, K., Bengio, Y.: NEURAL MACHINE TRANSLATION BY JOINTLY LEARNING TO ALIGN AND TRANSLATE. In: arXiv preprint arXiv:1409.0473 (2015)
Chitrakar, R., Chuanhe, H.: Anomaly Detection Using Support Vector Machine Classification with K-Medoids Clustering. In: 3rd Asian Himalayas International Conference on Internet, Kathmundu, Nepal (2013)
Savitzky, A., Golay, M.J.: Smoothing and differentiation of data by simplified least squares procedures. Anal. Chem. 36(8), 1627–1639 (1964)
Luo, J., Ying, K., He, P., Bai, J.: Properties of Savitzky-Golay digital differentiators. Digital Signal Processing. 15(2), 122–136 (2005)
Sadeghi, M., Behnia, F., Amiri, R.: Window selection of the Savitzky–Golay filters for signal recovery from Noisy measurements. IEEE Trans. Instrum. Meas. 69(8), 5418–5427 (2020)
Brandl G. "pandas.cut," pandas development team (2008-2022). [Online]. Available: https://pandas.pydata.org/docs/reference/api/pandas.cut.html. Accessed 28 Aug 2021
Reiss, C., Tumanov, A.; Ganger, G.R., Katz, R.H.: Towards understanding heterogeneous clouds at scale: Google trace analysis. Intel science and Technology Center for Cloud Computing, Tech. Rep 84 (2012)
Chaves, S.A.D., Uriarte, R.B., Westphall, C.B.: Toward an architecture for monitoring private clouds. IEEE Commun. Mag. 49(12), 130–137 (2011)
Zou, F., Shen, L., Jie, Z., Zhang, W., Liu, W.: A Sufficient Condition for Convergences of Adam and RMSProp. In: IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA (2019)
Zhang, P., Zhang, H., Dai, J., Yi, Y., Zhang, H., Zhang, Y.: Deep learning on computational-resource-limited platforms: a survey. Mob. Inf. Syst. 2020, 1–19 (2020)
Al-Sayed, M.M.: Google Dataset (2022). [Online]. Available: https://drive.google.com/file/d/1p4Y3LitdBCIE0YZHvGFNJLG8ey72hXHH/view?usp=sharing. Accessed 29 Apr 2022
Data Availability Statement (DAS)
Experiments included in this paper have been conducted on the total CPU and memory usage, so the average CPU usages and memory usages of all tasks on every machine at every timestamp were calculated and stored in a text file with a name “the machine ID”. Therefore, 790 text files have been generated and organized in records (one for each 300 s) of the order “timestamp, CPU, and Memory”. These files are available in [85] repository.
Funding
Open access funding provided by The Science, Technology & Innovation Funding Authority (STDF) in cooperation with The Egyptian Knowledge Bank (EKB).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
There is no conflict of interest.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Al-Sayed, M.M. Workload Time Series Cumulative Prediction Mechanism for Cloud Resources Using Neural Machine Translation Technique. J Grid Computing 20, 16 (2022). https://doi.org/10.1007/s10723-022-09607-0
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s10723-022-09607-0