Abstract
During the last decade, the use of AIs is being incorporated into the educational field whether to support the analysis of human behavior in teaching-learning contexts, as didactic resource combined with other technologies or as a tool for the assessment of the students.
This proposal presents a Systematic Literature Review and mapping study on the use of AIs for the assessment of students that aims to provide a general overview of the state of the art and identify the current areas of research by answering 6 research questions related with the evolution of the field, and the geographic and thematic distribution of the studies.
As a result of the selection process this study identified 20 papers focused on the research topic in the repositories SCOPUS and Web of Science from an initial amount of 129.
The analysis of the papers allowed the identification of three main thematic categories: assessment of student behaviors, assessment of student sentiments and assessment of student achievement as well as several gaps in the literature and future research lines addressed in the discussion.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Similar content being viewed by others
Notes
- 1.
The last two countries are not represented in the map in Fig. 3 due to their small size.
References
Makridakis, S.: The forthcoming Artificial Intelligence (AI) revolution: its impact on society and firms. Futures 90, 46–60 (2017)
Roll, I., Wylie, R.: Evolution and revolution in artificial intelligence in education. Int. J. Artif. Intell. Educ. 26, 582–599 (2016). https://doi.org/10.1007/s40593-016-0110-3
Gunning, D.: Explainable artificial intelligence (XAI). Defense Advanced Research Projects Agency (DARPA). http://www.darpa.mil/program/explainable-artificial-intelligence. Accessed 24 Feb 2020
Abdul, A., Vermeulen, J., Wang, D., Lim, B.Y., Kankanhalli, M.: Trends and trajectories for explainable, accountable and intelligible systems: an HCI research agenda. In: Proceedings of the 2018 CHI Conference on Human Factors in Computing Systems, pp. 1–18. ACM, April 2018
Amershi, S., Cakmak, M., Knox, W.B., Kulesza, T.: Power to the people: the role of humans in interactive machine learning. Ai Mag. 35, 105–120 (2014)
Cruz-Benito, J.: On data-driven systems analyzing, supporting and enhancing users’ interaction and experience. Doctoral dissertation, Universidad de Salamanca (2018)
O’neil, C.: Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Broadway Books, New York (2014)
Etzioni, A., Etzioni, O.: Incorporating ethics into artificial intelligence. J. Ethics 21, 403–418 (2017). https://doi.org/10.1007/s10892-017-9252-2
Russell, S., et al.: Letter to the editor: Research priorities for robust and beneficial artificial intelligence: an open letter. AI Mag. 36, 3–4 (2015)
Peirano, M.: El enemigo conoce el sistema: Manipulación de ideas, personas e influencias después de la economía de la atención. Debate (2019)
Miller, T.: Explanation in artificial intelligence: insights from the social sciences. Artif. Intell. 267, 1–38 (2019)
Aoun, J.E.: Robot-Proof: Higher Education in the Age of Artificial Intelligence. MIT Press, Cambridge (2017)
Henrie, C.R., Halverson, L.R., Graham, C.R.: Measuring student engagement in technology-mediated learning: a review. Comput. Educ. 90, 36–53 (2015)
Jonassen, D., Davidson, M., Collins, M., Campbell, J., Haag, B.B.: Constructivism and computer-mediated communication in distance education. Am. J. Distance Educ. 9, 7–26 (1995)
Perrotta, C., Williamson, B.: The social life of learning analytics: cluster analysis and the ‘performance’ of algorithmic education. Learn. Media Technol. 43, 3–16 (2018)
Papamitsiou, Z., Economides, A.A.: Learning analytics and educational data mining in practice: a systematic literature review of empirical evidence. J. Educ. Technol. Soc. 17, 49–64 (2014)
Roll, I., Winne, P.H.: Understanding, evaluating, and supporting self-regulated learning using learning analytics. J. Learn. Anal. 2, 7–12 (2015)
Rienties, B., Cross, S., Zdrahal, Z.: Implementing a learning analytics intervention and evaluation framework: what works? In: Kei Daniel, B. (ed.) Big Data and Learning Analytics in Higher Education, pp. 147–166. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-06520-5_10
Kluyver, T., et al.: Jupyter notebooks-a publishing format for reproducible computational workflows. In: Loizides, F., Schmidt, B. (eds.) Positioning and Power in Academic Publishing: Players, Agents and Agendas, pp. 87–90. IOS Press, Canada (2016)
Hamrick, J.B.: Creating and grading IPython/Jupyter notebook assignments with NbGrader. In: Alphonce, C., Tims, J. (eds.) Proceedings of the 47th ACM Technical Symposium on Computing Science Education, p. 242. ACM Press, New York (2016)
Blank, D. S., Bourgin, D., Brown, A., Bussonnier, M., Frederic, J., Granger, B.,… Page, L. nbgrader: A tool for creating and grading assignments in the Jupyter Notebook. The Journal of Open Source Education2, 32–34 (2019)
Kitchenham, B., Charters, S.: Guidelines for performing Systematic Literature Reviews in Software Engineering. Version 2.3 (EBSE-2007–01) (2007). https://www.elsevier.com/__data/promis_misc/525444systematicreviewsguide.pdf. Accessed 24 Feb 2020
Cruz-Benito, J., García-Peñalvo, F.J., Therón, R.: Analyzing the software architectures supporting HCI/HMI processes through a systematic review of the literature. Telematics and Inform. 38, 118–132 (2019)
Kitchenham, B.A., Budgen, D., Brereton, P.O.: Using mapping studies as the basis for further research – a participant-observer case study. Inf. Softw. Technol. 53, 638–651 (2011)
Kitchenham, B.: What’s up with software metrics? – a preliminary mapping study. J. Syst. Softw. 83, 37–51 (2010)
Neiva, F.W., David, J.M.N., Braga, R., Campos, F.: Towards pragmatic interoperability to support collaboration: a systematic review and mapping of the literature. Inf. Softw. Technol. 72, 137–150 (2016)
García Sánchez, F., Therón, R., Gómez-Isla, J.: Alfabetización visual en nuevos medios: revisión y mapeo sistemático de la literature. Educ. Knowl. Soc. 20, 1–35 (2019)
Moher, D., Liberati, A., Tetzlaff, J., Altman, D.G.: Preferred reporting items for systematic reviews and meta-analyses: the PRISMA statement. PLoS Med 6, 1–6 (2009)
Floryan, M., Dragon, T., Basit, N., Dragon, S., Woolf, B.: Who needs help? Automating student assessment within exploratory learning environments. In: Conati, C., Heffernan, N., Mitrovic, A., Verdejo, M.F. (eds.) AIED 2015. LNCS (LNAI), vol. 9112, pp. 125–134. Springer, Cham (2015). https://doi.org/10.1007/978-3-319-19773-9_13
Gurupur, V.P., Pankaj Jain, G., Rudraraju, R.: Evaluating student learning using concept maps and Markov chains. Expert Syst. Appl. 42, 3306–3314 (2015)
Newman, H., Joyner, D.: Sentiment analysis of student evaluations of teaching. In: Penstein Rosé, C., Martínez-Maldonado, R., Hoppe, H.U., Luckin, R., Mavrikis, M., Porayska-Pomsta, K., McLaren, B., du Boulay, B. (eds.) AIED 2018. LNCS (LNAI), vol. 10948, pp. 246–250. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-93846-2_45
Ma, J., Kang, J.-H., Shaw, E., Kim, J.: Workflow-based assessment of student online activities with topic and dialogue role classification. In: Biswas, G., Bull, S., Kay, J., Mitrovic, A. (eds.) AIED 2011. LNCS (LNAI), vol. 6738, pp. 187–195. Springer, Heidelberg (2011). https://doi.org/10.1007/978-3-642-21869-9_26
Tzacheva, A., Ranganathan, J., Jadi, R.: Multi-label emotion mining from student comments. In: Proceedings of the 2019 4th International Conference on Information and Education Innovations, pp. 120–124. ACM, New York (2019)
Lin, Q., Zhu, Y., Zhang, S., Shi, P., Guo, Q., Niu, Z.: Lexical based automated teaching evaluation via students’ short reviews. Comput. Appl. Eng. Educ. 27, 194–205 (2019)
Wang, M., Wang, C., Lee, C., Lin, S., Hung, P.: Type-2 fuzzy set construction and application for adaptive student assessment system. In: Proceedings of the 2014 IEEE International Conference on Fuzzy Systems (FUZZ-IEEE), pp. 888–894. IEEE (2014)
Wang, L., Hu, G., Zhou, T.: Semantic analysis of learners’ emotional tendencies on online MOOC education. Sustainability 10, 1–19 (2018)
Akhtar, J.: An interactive multi-agent reasoning model for sentiment analysis: a case for computational semiotics. Artif. Intell. Rev., 1–18 (2019). https://link.springer.com/article/10.1007/s10462-019-09785-6#citeas
Mahboob, T., Irfan, S., Karamat, A.: A machine learning approach for student assessment in E-learning using Quinlan’s C4.5, Naive Bayes and Random Forest algorithms. In: Proceedings of the 19th International Multi-Topic Conference (INMIC), pp. 1–8. IEEE (2017)
Livieris, I.E., Drakopoulou, K., Kotsilieris, T., Tampakas, V., Pintelas, P.: DSS-PSP - a decision support software for evaluating students’ performance. In: Boracchi, G., Iliadis, L., Jayne, C., Likas, A. (eds.) EANN 2017. CCIS, vol. 744, pp. 63–74. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-65172-9_6
Simjanoska, M., Gusev, M., Bogdanova, A.M.: Intelligent modelling for predicting students’ final grades. In: Proceedings of the 37th International Convention on Information and Communication Technology, Electronics and Microelectronics (MIPRO), pp. 1216–1221. IEEE (2014)
Hameed, I.: A fuzzy system to automatically evaluate and improve fairness of multiple-choice questions (MCQs) based exams. In: Proceedings of the 8th International Conference on Computer Supported Education - Volume 1: CSEDU, pp. 476–481. SciTePress (2016)
Dudek, D.: Survey analyser: effective processing of academic questionnaire data. In: Borzemski, L., Świątek, J., Wilimowska, Z. (eds.) ISAT 2018. AISC, vol. 852, pp. 245–257. Springer, Cham (2019). https://doi.org/10.1007/978-3-319-99981-4_23
Kuk, K., Milentijević, I.Z., Ranđelović, D., Popović, B.M., Čisar, P.: The design of the personal enemy - MIMLebot as an intelligent agent in a game-based learning environment. Acta Polytechnica Hungarica 14, 121–139 (2017)
Boongoen, T., Shen, Q., Price, C.: Fuzzy qualitative link analysis for academic performance evaluation. Int. J. Uncertainty Fuzziness and Knowl.-Based Syst. 19, 559–585 (2011)
Zatarain-Cabada, R., Barrón-Estrada, M.L., Ríos-Félix, J.M.: Affective learning system for algorithmic logic applying gamification. In: Pichardo-Lagunas, O., Miranda-Jiménez, S. (eds.) MICAI 2016. LNCS (LNAI), vol. 10062, pp. 536–547. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-62428-0_44
Caliskan, E., Tatar, U., Bahsi, H., Ottis, R., Vaarandi, R.: Capability detection and evaluation metrics for cyber security lab exercises. In: Bryant, A.R., Mills, R.F., Lopez, J. (eds.) Proceedings of the 2017 International Conference on Cyber Warfare and Security, pp. 407–414. Academic Conferences and Publishing International Ltd., UK (2017)
Luchoomun, T., Chumroo, M., Ramnarain-Seetohul, V.: A knowledge based system for automated assessment of short structured questions. In: Proceedings of the 2019 IEEE Global Engineering Education Conference (EDUCON), pp. 1349–1352. IEEE (2019)
Singh, S., Lal, S.P.: Educational courseware evaluation using Machine Learning techniques. In: Proceedings of the 2013 IEEE Conference on e-Learning, e-Management and e-Services, pp. 73–78. IEEE (2013)
Petrova, K., Li, C.: Focus and setting in mobile learning research: a review of the literature. Commun. IBIMA 10, 219–226 (2009)
Acknowledgement
This work has been partially funded by the Spanish Government Ministry of Economy and Competitiveness through the DEFINES project (Ref. TIN2016-80172-R).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Sánchez-Prieto, J.C., Gamazo, A., Cruz-Benito, J., Therón, R., García-Peñalvo, F.J. (2020). AI-Driven Assessment of Students: Current Uses and Research Trends. In: Zaphiris, P., Ioannou, A. (eds) Learning and Collaboration Technologies. Designing, Developing and Deploying Learning Experiences. HCII 2020. Lecture Notes in Computer Science(), vol 12205. Springer, Cham. https://doi.org/10.1007/978-3-030-50513-4_22
Download citation
DOI: https://doi.org/10.1007/978-3-030-50513-4_22
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-50512-7
Online ISBN: 978-3-030-50513-4
eBook Packages: Computer ScienceComputer Science (R0)