Abstract
Prediction of result of students in a particular subject based on their performance in continuous assessment during the semester can be accomplished by various available machine learning models. Every model has its own advantage and limitation due to the algorithm on which they work. Linear regression models have been used very popularly in the area of predictive analytics. Artificial neural networks have also proven their capabilities in prediction. Deep learning techniques are a trend nowadays in data analytics due to their accuracy and performance. This research paper will present a comparison of performance of five popular machine learning models used in predictive analytics—generalized linear model, multilayer perceptron, gradient boost model, Random Forest model, and deep neural network. We have used the result data of students at DIT University, Dehradun, which comprises of schooling marks, continuous assessment marks, and final marks in a subject. On the basis of schooling marks and continuous assessment marks, all these models will predict the final marks of a student in the subject. We will then compare and present the result and performance of these five models.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
O.C. Asogwa, A.V. Oladugba, Prediction of students academic performance rates using artificial neural networks. Am. J. Appl. Math. Stat. 3(4), 151–155 (2015)
B. Oancea, R. Dragoescu, S. Ciucu, Predicting students’ results in higher education using neural networks, in International Conference on Applied Information and Communication Technologies, Jelgava, Latvia (2013)
H. Agrawal, H. Mavani, Student performance prediction using machine learning. Int. J. Eng. Res. Technol. 4(03) (2015)
S.T. Karamouzis, A. Vrettos, An artificial neural network for predicting student graduation outcomes, in World Congress on Engineering and Computer Science, San Francisco, USA (2008)
S. Huang, N. Fang, Regression models of predicting student academic performance in an engineering dynamics course. Comput. Educ. 133–145 (2013)
P.M. Arsad, N. Buniyamin, J.-L. Ab Manan, Neural network model to predict electrical students’ academic performance, in 4th International Congress on Engineering Education (ICEED), Georgetown, Malaysia (2012)
Dr. M. Shakil, A multiple linear regression model to predict the student’s final grade in a mathematics class. Sam Houston State University
H. Hamsa, S. Indiradevi, J.J. Kizhakkethottam, Student academic performance prediction model using decision tree and fuzzy genetic algorithm, in Global Collegium in Recent Advancements and Effectual Researches in Engineering, Science and Technology (2016)
A. Polyzou, G. Karypis, Grade prediction with models specific to students and courses. Int. J. Data Sci. Anal. 2, 159–171 (2016)
A.M. Shahiria, W. Husaina, N.A. Rashida, A review on predicting student’s performance using data mining techniques. Proc. Comput. Sci. 72, 414–422 (2015)
S. Eguchi, Model Comparison for generalized linear models with dependent observations. Econom. Stat. 59 (2017)
J. Nelder, R. Wedderburn, Generalized linear models. J. R. Stat. Soc. (1972)
F. Rosenblatt, Principles of Neurodynamics: Perceptrons and the Theory of Brain Mechanisms. Spartan Books (1961)
D. Rumelhart, G. Hinton, R.J. Williams, Learning internal representations by error propagation, in Parallel Distribute Processing: Explorations in the Microstructure of Cognition, vol. 1 (1986)
S. Haykin, Neural Networks: A Comprehensive Foundation, 2nd edn. (Prentice Hall, 2012)
J. Kan, Evaluation of mining engineering technology innovation ability and application based on BP neural network, in International Conference on Industrial Technology and Management (ICITM) (2017)
Y. Zhang, A. Haghani, A gradient boosting method to improve travel time prediction. Trans. Res. Part C (2015)
Y. LeCun et al., Gradient based learning applied in document recognition, in Proceedings of the IEEE, vol. 86 (1998)
J.H. Friedman, Greedy Function Approximation: A Gradient Boosting Machine (1999)
T.K. Ho, Random decision forests, in 3rd International Conference on Document Analysis and Recognition, Montreal (1995)
T.K. Ho, The random subspace method for constructing decision forests. IEEE Trans. Pattern Anal. Mach. Intell. 20 (1998)
T.K. Ho, A data complexity analysis of comparative advantages of decision forest constructions. Pattern Anal. Appl. (2002)
J. Schmidhuber, Deep Learning in Neural Networks. Technical Report IDSIA-03-14 arXiv:1404.7828
J. Schmidhuber, LSTM recurrent networks learn simple context free and context sensitive languages. IEEE Trans. Neural Netw. 12 (2001)
V.K. Ojha, A. Abraham, V. Snasel, Metaheuristic design of feedforward neural networks: a review of two decades of research. Eng. Appl. Artif. Intell. 60 (2017)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2019 Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Kumar, V., Garg, M.L. (2019). Comparison of Machine Learning Models in Student Result Prediction. In: Kamal, R., Henshaw, M., Nair, P. (eds) International Conference on Advanced Computing Networking and Informatics. Advances in Intelligent Systems and Computing, vol 870. Springer, Singapore. https://doi.org/10.1007/978-981-13-2673-8_46
Download citation
DOI: https://doi.org/10.1007/978-981-13-2673-8_46
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-13-2672-1
Online ISBN: 978-981-13-2673-8
eBook Packages: EngineeringEngineering (R0)