Abstract
The relevance of computational thinking as a skill for today's learners is no longer in question, but every skill needs an assessment system. In this study, we analyze two validated instruments for assessing computational thinking - the CTt (Computational Thinking Test) and the CTS (Computational Thinking Scale). The study involved 49 students in grades 8 and 9 (age 14–16). Prior to the study, students in both grades were taught computational thinking differently. One group learned computational thinking by completing tasks and creating projects in Scratch, the other group learned by completing tasks in “Minecraft: Education Edition”. The students were asked to take the CTt and CTS tests. The nature of these tests is different, one is computational thinking diagnostic tool, the other is a psychometric self-assessment test consisting of core abilities (subconstructs) important for computational thinking. The aim of this study was to determine how these tests related to each other and whether students’ gender and the different tools chosen to teach computational thinking had an impact on the level of computational thinking knowledge and abilities acquired based on the tests. The results have shown that the scores of the two tests correlated with each other only for male students’ subgroup. For a whole group CTt scores correlated only with CTS algorithmic thinking subconstruct. The results have also shown that teaching tools do have an impact on the acquisition of different computational thinking concepts skills: students taught with different tools had different test results. This study provides useful implications on computational thinking teaching improvement and its assessment better understanding.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Asbell-Clarke, J., et al.: The development of students’ computational thinking practices in elementary- and middle-school classes using the learning game. Zoombinis. Comput. Hum. Behav. 115, 106587 (2021)
Bocconi, S., et al.: Reviewing computational thinking in compulsory education. In: Inamorato Dos Santos, A., Cachia, R., Giannoutsou, N., Punie, Y. (eds.) Publications Office of the European Union, Luxembourg (2022). https://doi.org/10.2760/126955. ISBN 978-92-76-47208-7, JRC128347
CSTA: K12 computer science standards (2017). https://www.csteachers.org/page/about-csta-s-k-12-nbsp-standards
Djambong, T., Freiman, V., Gauvin, S., Paquet, M., Chiasson, M.: Measurement of computational thinking in K-12 education: the need for innovative practices. In: Sampson, D., Ifenthaler, D., Spector, J., Isaías, P. (eds.) Digital Technologies: Sustainable Innovations for Improving Teaching and Learning, pp. 193–222. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-73417-0_12
Dolgopolovas, V., Jevsikova, T., Dagiene, V., Savulioniene, L.: Exploration of computational thinking of software engineering novice students based on solving computer science tasks. Int. J. Eng. Educ. 32(3), 1–10 (2016)
Eloy, A., Achutti, C.F., Cassia, F., Deus Lopes, R.: A data-driven approach to assess computational thinking concepts based on learners’ artifacts. Inf. Educ. 21(1), 33–54 (2022)
European Commission: Digital Education Action Plan 2021–2027: Resetting education and training for the digital age (2020). https://eurlex.europa.eu/legal-content/EN/TXT/?uri=CELEX:52020DC0624
Groher, I, Sabitzer, B., Demarle-Meusel, H., Kuka, L., Hofer, A.: Work-in-progress: closing the gaps: diversity in programming education. In: 2021 IEEE Global Engineering Education Conference (EDUCON), pp. 1449–1453 (2021)
Guggemos, J.: On the predictors of computational thinking and its growth at the high-school level. Comput. Educ. 161, 104060 (2021)
Guggemos, J., Seufert, S. Román-González, M.:Computational thinking assessment – towards more vivid interpretations. Tech. Know. Learn.https://doi.org/10.1007/s10758-021-09587-2
Halpern, D.F.: Thoughts and Knowledge: An Introduction to Critical Thinking. Lawrence Erlbaum Associates, New Jersey-London (1996)
ISTE: Computational thinking: leadership toolkit (2015). https://www.iste.org/computational-thinking
Korkmaz, Ö., Çakir, R., Özden, M.Y.: A validity and reliability study of the computational thinking scales (CTS). Comput. Hum. Behav. 72, 558–569 (2017)
Labusch, A., Eickelmann, B.: Computational thinking competences in countries from three different continents in the mirror of students’ characteristics and school learning. In: Kong, S.C., et al., (eds.) Proceedings of International Conference on Computational Thinking Education 2020, pp. 2–7. The Education University of Hong Kong (2020)
Ma, H., Zhao, M., Wang, H., Wan, X., Cavanaugh, T.W., Liu, J.: Promoting pupils’ computational thinking skills and self-efficacy: a problem-solving instructional approach. Educ. Tech. Res. Dev. 69(3), 1599–1616 (2021). https://doi.org/10.1007/s11423-021-10016-5
Moreno-León, J., Robles, G., Román-González, M.: Dr. Scratch: automatic analysis of scratch projects to assess and foster computational thinking. RED-Rev. Educ. Distancia 46, 1–23 (2015)
Piatti, A., et al.: The CT-cube: a framework for the design and the assessment of computational thinking activities. Comput. Hum. Behav. Rep. 5, 100166 (2022)
Poulakis, E., Politis, P.: Computational thinking assessment: literature review. In: Tsiatsos, T., Demetriadis, S., Mikropoulos, A., Dagdilelis, V. (eds.) Research on e-Learning and ICT in Education, pp. 111–128. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-64363-8_7
Román-González, M., Pérez-González, J.-C., Jiménez-Fernández, C.: Which cognitive abilities underlie computational thinking? Criterion validity of the computational thinking test. Comput. Hum. Behav. 72, 678–691 (2017)
Román-González, M., Moreno-León, J., Robles, G.: Combining assessment tools for a comprehensive evaluation of computational thinking interventions. In: Kong, S.-C., Abelson, H. (eds.) Computational Thinking Education, pp. 79–98. Springer, Singapore (2019). https://doi.org/10.1007/978-981-13-6528-7_6
Rowe, E., et al.: Assessing implicit computational thinking in Zoombinis puzzle gameplay. Comput. Hum. Behav. 120, 106707 (2021)
Statter, D., Armoni, M.: Teaching abstraction in computer science to 7th grade students. ACM Trans. Comput. Educ. 20(1), 8–837 (2020)
Stupurienė, G., Jevsikova, T., Juškevičienė, A.: Solving ecological problems through physical computing to ensure gender balance in STEM education. Sustainability 14(9), 4924 (2022)
Sun, L., Hu, L., Zhou, D., Yang, W.: Evaluation and developmental suggestions on undergraduates’ computational thinking: a theoretical framework guided by Marzano’s new taxonomy. Interact. Learn. Environ. (2022). https://doi.org/10.1080/10494820.2022.2042311
Tang, X., Yin, Y., Lin, Q., Hadad, R., Zhai, X.: Assessing computational thinking: a systematic review of empirical studies. Comput. Educ. 148, 103798 (2020)
Tikva, C., Tambouris, E.: Mapping computational thinking through programming in K-12 education: a conceptual model based on a systematic literature Review. Comput. Educ. 162, 104083 (2021)
Troiano G., et al.: Is my game OK Dr. Scratch? Exploring programming and computational thinking development via metrics in student-designed serious games for STEM. In: Proceedings of the 18th ACM International Conference on Interaction Design and Children Association for Computing Machinery, New York, NY, USA, pp. 208–219 (2019)
Wei, X., Lin, L., Meng, N., Tan, W., Kong, S.-C., Kinshuk.: The effectiveness of partial pair programming on elementary school students’ computational thinking skills and self-efficacy. Comput. Educ. 160, 104023 (2021)
Author information
Authors and Affiliations
Corresponding authors
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Masiulionytė-Dagienė, V., Jevsikova, T. (2022). Assessing Computational Thinking: The Relation of Different Assessment Instruments and Learning Tools. In: Bollin, A., Futschek, G. (eds) Informatics in Schools. A Step Beyond Digital Education. ISSEP 2022. Lecture Notes in Computer Science, vol 13488. Springer, Cham. https://doi.org/10.1007/978-3-031-15851-3_6
Download citation
DOI: https://doi.org/10.1007/978-3-031-15851-3_6
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-15850-6
Online ISBN: 978-3-031-15851-3
eBook Packages: Computer ScienceComputer Science (R0)