Abstract
Professional training involves the acquisition of knowledge, skills, and expertise required to perform specific job roles. It can take various forms, including classroom instruction, practical experience, and online training. While online training has the potential to reach a wider audience, a lack of interaction and support may lead to low engagement and high risk of failure. In this paper, we propose a solution to predict, as early as possible, the risk of failure in online professional training. To achieve this, we use a dataset of 13,719 observations, including 912 learners and 182 online courses delivered through an LMS from 2017 to 2022. Our data analysis led us to define the risk of failure based on three classes: low risk, medium risk, and high risk. The objective is to predict the class of each observation when the learner reaches the halfway point in each course. Initially, we tested nine predictive models, which revealed discrepancies in their results across the three classes. To address this gap, we propose a new solution that employs a weighted vote to improve classifications. This solution applies the Borda method to rank the nine models based on their predictive performance in each class. Then, the weight assigned to each model is calculated by considering both its rank and its F1 score in each risk of failure class. Finally, we apply a weighted majority vote involving the nine models. On average, and compared to the baseline models, our solution improves the F1 score by 1.2% for the three risk classes.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
References
Lambert, M., Vero, J., Zimmermann, B.: Vocational training and professional development: a capability perspective. Int. J. Train. Dev. 16(3), 164–182 (2012)
Mouaici, M., Vignollet, L., Galez, C., Etienne, M.: Learning analytics dashboards for professional training - challenges and proposal. In: CEUR Workshop Proceedings (2018)
Littlejohn, A.: Learning and work: professional learning analytics. In: Handbook of Learning Analytics, Society for Learning Analytics Research (SoLAR), pp. 269–277 (2017)
Ruiz-Calleja, A., Prieto, L.P., et al.: Learning analytics for professional and workplace learning: a literature review. In: Lecture Notes in Computer Science (2017)
Daumiller, M., Rinas, R., Olden, D., Dresel, M.: Academics’ motivations in professional training courses: effects on learning engagement and learning gains (2020)
Singh, V., Thurman, A.: How Many ways can we define online learning? a systematic literature review of definitions of online learning (1988-2018). Am. J. Distan. Educ. 33(4), 289–306 (2019). https://doi.org/10.1080/08923647.2019.1663082
Elias, T.: Learning analytics : definitions , processes and potential. Learning (2011)
Alhothali, A., et al.: Predicting student outcomes in online courses using machine learning techniques: a review. Sustain. 14, 6199 (2022)
Katarya, R., Gaba, J.: A review on machine learning based student’s academic performance prediction systems. In: Proceedings. - International Conference on Artificial Intelligence and Smart Systems, ICAIS (2021)
Adnan, M., et al.: Predicting at-risk students at different percentages of course length for early intervention using machine learning models. IEEE Access (2021)
Trakunphutthirak, R., Cheung, Y., Lee, V.: Detecting student at risk of failure: A case study of conceptualizing mining from internet access log files. In: ICDMW (2019)
Senthil, S., Lin, W.M.: Applying classification techniques to predict students’ academic results. In: ICCTAC 2017, vol. 2018, pp. 1–6 (2018)
You, J.W.: Identifying significant indicators using LMS data to predict course achievement in online learning. Internet High. Educ. 29, 23–30 (2016)
Souai, W., Mihoub, A., Tarhouni, M., Zidi, S., Krichen, M., Mahfoudhi, S.: Predicting at-risk students using the deep learning BLSTM approach. SMARTTECH (2022)
Mubarak, A.A., Cao, H., Hezam, I.M.: Deep analytic model for student dropout prediction in massive open online courses. Comput. Electr. Eng. (2021)
Waheed, H., et al.: Early prediction of learners at risk in self-paced education: a neural network approach. Expert Syst. Appl. 213, 118868 (2023)
Burgos, C., et al.: Data mining for modeling students’ performance: a tutoring action plan to prevent academic dropout. Comput. Electr. Eng. 66, 541–556 (2018)
Hong, B., Wei, Z., Yang, Y.: Discovering learning behavior patterns to predict dropout in MOOC. In: ICCSE 2017 - 12th International Conference on Computer Science and Education, pp. 700–704 (2017)
Kustitskaya, T.A., et al.: Early student-at-risk detection by current learning performance and learning behavior indicators. Bulg. Acad. Sci. Cybern. Inf. Technol. (2022)
What is SCORM and How it Works. https://scorm.com/. Accessed 05 Apr 2023
Zahid, M.A., De Swart, H.: The borda majority count. Inf. Sci. (Ny). (2015)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Mouaici, M. (2023). Early Prediction of Learners At-Risk of Failure in Online Professional Training Using a Weighted Vote. In: Viberg, O., Jivet, I., Muñoz-Merino, P., Perifanou, M., Papathoma, T. (eds) Responsive and Sustainable Educational Futures. EC-TEL 2023. Lecture Notes in Computer Science, vol 14200. Springer, Cham. https://doi.org/10.1007/978-3-031-42682-7_17
Download citation
DOI: https://doi.org/10.1007/978-3-031-42682-7_17
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-42681-0
Online ISBN: 978-3-031-42682-7
eBook Packages: Computer ScienceComputer Science (R0)