Abstract
The development of computational thinking skills is attracting attention worldwide. The use of visual or block-based coding in primary schools has gained momentum. Yet, students' acceptance of such coding environments has been neglected in the literature. This study presents a measurement instrument that will allow pursuing such an endeavor. The Technology Acceptance Model was used as a guiding framework. A panel of eight experts and a purposive sample of 315 elementary school students (Mage = 10.18 years, SD = 1.13) were recruited. Inter-rater agreement between experts (75–100%) confirmed the content validity of the items. Exploratory factor analysis resulted in a three-factor solution of 12 items measuring perceived usefulness, perceived ease of use, and attitudes towards visual coding. Inspection of the correlation matrix provided evidence of convergent and discriminant validity. Cronbach’s alpha (.75–.82), McDonald’s omega (.68–.83), and Spearman-Brown’s split-half coefficients (.67–.80) provided evidence of reliability. Overall, findings support the validity and reliability of the proposed instrument. Its use can shed light on students' acceptance of different block-based coding environments.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Today's society is technology-driven, which increases the need for technological literacy. Educational policy advocates the development of students' computational thinking, defined as the ability to analyze, and solve problems using computational principles (Chalmers, 2018). It refers to a style of thinking and acting that is fundamental to everyone, not just computer scientists Indeed, it implies progressing from being passive consumers to understanding technology (Shute et al., 2017). As a result, learning to code has attracted interest in educational systems worldwide (Grover & Pea, 2013; Papavlasopoulou et al., 2018).
Coding is usually attributed to computer science studies subjects or majors. Yet, calls for the inclusion of coding from elementary levels onwards abounds (Berland & Wilensky, 2015; Bers et al., 2014; Chalmers, 2018). In this sense, several countries have proposed coding-related school subjects (Balanskat & Engelhardt, 2014). Similarly, more recent trends integrate coding tasks during science and mathematics instruction (Li et al., 2020; Sung et al., 2017). For example, the Next Generation Science Standards in the United States provides a basic curriculum for students’ early coding-literacy, in which mathematics and computational thinking are key scientific and engineering practices (NGSS Lead States, 2013). In Spain, current educational curricula include specific standards related to coding (LOMCE, 2013). A large number of EU countries, such as Denmark, Italy, or Portugal also incorporate coding in their school curricula (Balanskat & Engelhardt, 2014).
In summary, there is a renewed interest in introducing coding into elementary curricula. This is due in part to the availability of easy-to-use educational resources, such as board games, screen-based robots, or floor robots (Casey et al., 2020; Hamilton et al., 2020). Similarly, the growing emphasis on computational thinking has recently resulted in the development of user-friendly coding environments. Examples include Scratch, Code.org, Snap, or Micro:bit MakeCode, all of which use drag-and-drop strings (Fig. 1). Such resources, known as block-based coding or visual coding, enable the introduction of coding in elementary school (Sáez-López et al., 2016; Sung et al., 2017). Unlike text-based programming languages, visual coding uses pre-designed blocks. Each block represents actions, and their combination forms strings of instructions. Doing so, students code using pictographic language, which is then converted into error-free syntax.
2 Statement of the Problem
Much research has been conducted on the benefits of block-based coding at early ages (Sung et al., 2017). Likewise, research on how to introduce coding in primary and secondary school abounds. Yet, students’ acceptance of such resources remains unexplored, especially in the Spanish context. This lack of research is partly due to a lack of valid and reliable instruments for such a purpose, with few exceptions of ad-hoc measures. Hence, this study presents the development and psychometric evaluation of an instrument used to measure acceptance of block-based coding.
3 Theoretical Underpinnings
The proposed instrument is rooted in the Technology Acceptance Model (TAM; Davis, 1989). It was adapted from the Theory of Reasoned Action to study the acceptance of information system technologies. Specifically, TAM was proposed to explain the behavioral intentions when using technological innovations (Davis et al., 1989). This model describes the relationships between important determinants.
These include perceived usefulness and perceived ease of use of the system; attitudes toward use, behavioral intentions, and actual use of the technological innovation (Fig. 2). It represents one of the major models used to explore the acceptance of technology (for reviews, see Al-Maroof et al., 2021; Schepers & Wetzels, 2007). Its use in educational settings has provided insight into the acceptance of a broad array of information systems. Examples include cell phones (Zogheib & Daniela, 2021), statistics learning platforms (Song & Kong, 2017), floor robots (Casey et al., 2020), or role-play games (Suki & Suki, 2019).
The TAM posits three main aspects affecting individuals' acceptance behaviors. The first domain is perceived usefulness and relates to beliefs about whether the use of a certain system would improve performance. The second domain is ease of use. It refers to the notion that using a particular system would be simple, and it has a significant impact on perceived usefulness. Both domains are affected by external variables, such as gender, social influence, or support (Davis et al., 1989). The last major domain is attitudes, which are shaped by perceived usefulness and perceived, ease of use of the given system (Davis, 1989). It is postulated that greater levels of perceived ease of use and usefulness develop favorable attitudes; this increases the user's behavioral intention of use, which leads to acceptance of the information system technology (Al-Maroof et al., 2021).
4 Method
4.1 Study Design
This is an instrumental study design (Ato et al., 2013). This design involves research that examines the psychometric properties of measurement instruments.
4.2 Sample and Context
The sample was drawn using purposive sampling from 15 primary schools located in the province of Burgos (Spain). Participants attended a week-long intensive curriculum enrichment program, which included visual coding activities within science and math lessons. The BBC:MicroBit microcontroller board and its block-based environment MakeCode were used. Specifically, students programmed the BBC:MicroBit microcontroller boards into sensors that were used during hands-on laboratory investigations. For instance, a thermometer was programmed to measure the temperature of different water samples. Another example includes coding a timer to determine how long it takes to filter and purify the water (program detailed in Toma, 2020).
There were 337 elementary school students. After removing 22 questionnaires (6.5%) with missing data or more than one response per item, a valid sample of 315 students was formed. Almost half of the participants identified as girls (49.3%). Participants were enrolled in 3rd (n = 25), 4th (n = 77), 5th (n = 104), and 6th (n = 109) grades and had a mean age of 10.18 years (SD = 1.13).
A paper–pencil questionnaire was distributed on the last day of the program. To prevent social desirability, visiting school teachers were absent throughout data collection. Students were informed that their participation was voluntary and anonymous and that their responses would not affect school grades.
4.3 Instrument Development Framework
DeVellis (2017) guidelines for scale development were used. The first step was to determine the construct to be studied. As mentioned earlier, the TAM model was chosen. The second step was to generate a pool of items reflecting the latent variables of the TAM model. The recommended minimum number of items per construct is three. Therefore, six items measuring perceived usefulness, ease of use, and attitudes were developed. The items were worded to refer to the action of "coding" rather than to visual coding itself (Davis et al., 1989). The items on perceived usefulness related to different aspects (e.g., future study, school, life) to capture the phenomenon. The items on ease of use referred to students' perceived difficulty in using visual coding. Finally, the items measuring attitudes referred to positive dispositions toward visual coding. Therefore, simple terms (e.g., enjoyment, fun, or interest) were used. Some items were worded negatively to avoid acquiescence. In the third step, a five-point Likert scale (1: strongly disagree; 2: disagree; 3: Neither disagree or agree; 4: agree; and 5: strongly agree) was chosen as the measurement format. The fourth step was to assess the content validity of the items. Finally, in the fifth step, the revised items were administered to a large number of participants and tested for construct validity.
4.4 Data Analysis
4.4.1 Content Validity
A panel of eight experts (two university professors and six elementary school teachers) reviewed the original pool of items. The experts independently mapped the items to a TAM construct and rated their linguistic appropriateness and clarity (0: not appropriate; 1: appropriate).
4.4.2 Construct Validity
The revised items were subjected to an exploratory factor analysis following contemporary recommendations (Gaskin & Happell, 2014). The TAM model assumes that factors are correlated; therefore, the extraction method Principal Axis Factoring with Promax-oblique rotation was used. Retention of factors was determined using the results of Horn's parallel analysis (Hayton et al., 2004). Items with communalities > 0.30, loadings > 0.40, and no cross-loadings between factors were retained. Since there are gender differences in computer use and programming (Stoilescu & Egodawatte, 2010), an exploratory factor analysis by gender was also conducted. Finally, the correlation matrix was examined for evidence of convergent and discriminant validity. Items on one factor are expected to be more highly correlated than items on another factor. In addition, moderate correlations (r = 0.30 -0.50) between factors are preferable.
4.4.3 Reliability
Several indices were used to examine the internal consistency reliability of each retained factor of the proposed instrument (Hayes & Coutts, 2020). These include Cronbach’s alpha (α ≥ 0.70), McDonald’s Omega (ω), and Spearman-Brown’s split-half (≥ 0.60).
5 Results
5.1 Content Validity
Inter-rater agreement in linking the items to the TAM constructs was high, with at least seven of the eight experts reaching consensus (87.5–100%). Inter-rater agreement on the ratings of each item for linguistic appropriateness and clarity was also adequate, with at least six of the eight experts reaching consensus (75–100%). Few changes, however, were proposed. The word 'irrelevant' was changed to 'of little use to me'. Likewise, 'relevant' was replaced with 'useful', which are easier words for Spanish-speaking elementary students.
5.2 Construct Validity
The Kaiser–Meyer–Olkin measure of sampling adequacy was 0.858 and Bartlett's test for sphericity was statistically significant (χ2 = 1442.397, p < 0.01), supporting the factorability of the data matrix. After parallel analysis, only three latent factors exceeded the eigenvalues of a randomly generated sample with the same characteristics (Fig. 3).
The EFA was then repeated, retaining three factors. Three items had commonalities below 0.30. After removing these items, two items with loading less than.40 and one item with cross-loadings emerged. Removal of these items yielded a final three-factor solution of conceptual meaningfulness that explained 58.46% of the variance (Table 1). The final questionnaire consisted of 12 items distributed as follows: four items measuring attitudes toward visual coding (34.52% variance explained); two positively and two negatively worded items measuring perceived usefulness of visual coding (14.89% variance explained); and three positively and one negatively worded items measuring perceived ease of use of visual coding (9.05% variance explained). The Spanish version of the items is included in "Appendix 1".
Items in each factor were moderately and highly correlated with adequate ranges for the attitudes (0.42 ≤ r ≤ 0.61); usefulness (0.32 ≤ r ≤ 0.59); and ease of use (0.30 ≤ r ≤ 0.48). Hence, factors exhibit convergent validity. Similarly, a moderate correlation between the attitudes and usefulness (r = 0.55), and a low correlation between attitude and ease of use (r = 0.27), and usefulness and ease of use (r = 0.22) provides evidence of discriminant validity of the retained factors.
Exploratory factor analysis by gender resulted in the same three-factor structure explaining 60.49% and 62.49% of the variance for the girls and boys sample, respectively (Table 2). However, one of the items measuring ease of use had a commonality below the 0.30 cutoff for the sample of boys. Since its removal affected reliability and given its acceptable factor loading, it was retained in the final questionnaire. Taken together, these findings provide evidence for construct validity.
5.3 Reliability
Each factor exhibited adequate reliability. Cronbach’s alpha reliabilities were 0.82 for attitude, 0.75 for usefulness, and 0.76 for ease of use factors. McDonald’s omega coefficients were 0.83, 0.74, and 0.68, respectively. The Spearman-Brown split-half indices were 0.80, 0.67, and 0.73, respectively. Taken together, these findings provide evidence of good to excellent reliability for each retained factor.
6 Discussion
The current study presents the development and validation of an instrument for the measurement of Spanish-speaking elementary school students' acceptance of visual coding. Using the TAM model as a guiding framework, psychometric analysis suggests that the proposed instrument is robust regarding validity and reliability evidence. In particular, a panel of experts confirmed that the items developed were linguistically appropriate, and consistent with the theoretical framework adopted. Such results provide evidence of content validity (DeVellis, 2017).
These outcomes were supported by the findings of exploratory factor analysis. Specifically, factorial results indicated a parsimonious three-factor latent structure. Hence, in line with theoretical expectations, usefulness, ease of use, and attitudes towards visual coding are empirically distinct. Moreover, further analyses revealed adequate convergent and discriminant validity, thereby lending support to the construct validity of the instrument (DeVellis, 2017; Gaskin & Happell, 2014). Likewise, separate factor analyses by gender found that there were no differences in the latent structure. This suggests that boys' and girls' acceptance of visual coding can be assessed using the same items.
Additionally, each subscale of the proposed instrument exhibited good to excellent reliability, as indicated by several coefficients (Hayes & Coutts, 2020). Taken together, this investigation advances a promising, valid, and reliable instrument regarding the assessment of visual coding acceptance.
6.1 Implications
To the best of the author’s knowledge, this study represents the first effort to advance an acceptance of block-based or visual coding instruments for the Spanish context. Except for Cheng, (2019), whose study includes Chinese students and uses ad-hoc measures, research on the acceptance of visual coding environments such as Scratch, Blockly, Snap, or Micro:bit Makecode is scarce. Hence, this investigation has taken an important step forward to bridging the gap in prior literature on visual coding.
Elementary school students are introduced to coding through block-based resources (Chalmers, 2018; Sáez-López et al., 2016; Sung et al., 2017). Based on the TAM model, students' behavioral intention to engage in coding-related activities depends on their perception of such resources as useful and easy to use. Ultimately, this will lead to positive attitudes and acceptance of visual coding (Al-Maroof et al., 2021; Davis, 1989). The proposed instrument is therefore likely to be useful in determining whether the visual coding environments that are widely used in elementary schools are, indeed, accepted by students. For example, studies comparing existing block-based resources may be informative as to what coding resources are preferred by students of different ages and gender (Yildiz Durak, 2020). Furthermore, research into the features of such resources that can improve the perceived ease of use, usefulness, and students' attitudes may also be beneficial to the computational thinking research agenda. This could ultimately lead to the refinement and improvement of visual coding environments.
It is also beneficial to investigate the psychometric properties of the instrument in secondary school students. This would contribute to the development of cross-sectional and longitudinal studies that would provide a broader picture of students' acceptance of visual coding resources.
6.2 Limitations
The present investigation does exhibit some limitations. First, it should be noted that the psychometric properties of the instruments were tested with students engaging with one specific visual-coding resource (i.e., Micro:bit MakeCode). Therefore, the latent structure of the instrument should be confirmed in future studies using alternative visual coding environments (e.g., Scratch).
Second, a school-year difference in sample size was also noted. Children from the last four years of Spanish primary school (3rd to 6th grade) were included. Yet, third-graders made up fewer than 8% of participants. This has hampered additional analysis to establish if the factor structure is invariant across school grades. Future studies with a larger sample size are warranted (Gaskin & Happell, 2014).
References
Al-Maroof, R., Al-Qaysi, N., Salloum, S. A., & Al-Emran, M. (2021). Blended learning acceptance: A systematic review of information systems models. Technology, Knowledge and Learning. Netherlands: Springer. https://doi.org/10.1007/s10758-021-09519-0
Ato, M., López, J. J., & Benavente, A. (2013). A classification system for research designs in psychology. Anales De Psicologia, 29(3), 1038–1059. https://doi.org/10.6018/analesps.29.3.178511
Balanskat, A., & Engelhardt, K. (2014). Computing our future: Compuer programming and coding. Priorities, school curricula and initiatives across europe. Brussels: European Schoolnet
Berland, M., & Wilensky, U. (2015). Comparing virtual and physical robotics environments for supporting complex systems and computational thinking. Journal of Science Education and Technology, 24(5), 628–647. https://doi.org/10.1007/s10956-015-9552-x
Bers, M. U., Flannery, L., Kazakoff, E. R., & Sullivan, A. (2014). Computational thinking and tinkering: exploration of an early childhood robotics curriculum. Computers & Education, 72, 145–157. https://doi.org/10.1016/j.compedu.2013.10.020
Casey, J. E., Pennington, L. K., & Mireles, S. V. (2020). Technology acceptance model: assessing preservice teachers’ acceptance of floor-robots as a useful pedagogical tool. Technology, Knowledge and Learning. https://doi.org/10.1007/s10758-020-09452-8
Chalmers, C. (2018). Robotics and computational thinking in primary school. International Journal of Child-Computer Interaction, 17, 93–100. https://doi.org/10.1016/j.ijcci.2018.06.005
Cheng, G. (2019). Exploring factors influencing the acceptance of visual programming environment among boys and girls in primary schools. Computers in Human Behavior, 92, 361–372. https://doi.org/10.1016/j.chb.2018.11.043
Davis, F. D. (1989). Perceived usefulness, perceived ease of use, and user acceptance of information technology. MIS Quarterly, 13(3), 319–340. https://doi.org/10.2307/249008
Davis, F. D., Bagozzi, R. P., & Warshaw, P. R. (1989). User acceptance of computer technology: a comparison of two theoretical models. Management Science, 35(8), 982–1003. https://doi.org/10.1287/mnsc.35.8.982
DeVellis, R. F. (2017). Scale development. Theory and Applications. SAGE.
Gaskin, C. J., & Happell, B. (2014). On exploratory factor analysis: a review of recent evidence, an assessment of current practice, and recommendations for future use. International Journal of Nursing Studies, 51, 511–521. https://doi.org/10.1016/j.ijnurstu.2013.10.005
Grover, S., & Pea, R. (2013). Computational thinking in K-12: a review of the state of the field. Educational Researcher, 42(1), 38–43. https://doi.org/10.3102/0013189X12463051
Hamilton, M., Clarke-Midura, J., Shumway, J. F., & Lee, V. R. (2020). An emerging technology report on computational toys in early childhood. Technology, Knowledge and Learning, 25(1), 213–224. https://doi.org/10.1007/s10758-019-09423-8
Hayes, A. F., & Coutts, J. J. (2020). Use omega rather than cronbach’s alpha for estimating reliability. But Communication Methods and Measures, 14(1), 1–24. https://doi.org/10.1080/19312458.2020.1718629
Hayton, J. C., Allen, D. G., & Scarpello, V. (2004). Factor retention decisions in exploratory factor analysis: a tutorial on parallel analysis. Organizational Research Methods, 7(2), 191–205. https://doi.org/10.1177/1094428104263675
Li, Y., Schoenfeld, A. H., DiSessa, A. A., Graesser, A. C., Benson, L. C., English, L. D., & Duschl, R. A. (2020). On computational thinking and STEM education. Journal for STEM Education Research, 3(2), 147–166. https://doi.org/10.1007/s41979-020-00044-w
LOMCE. (2013). Ley Orgánica 8/2013, de 9 de diciembre, para la mejora de la calidad educativa [Organic Law 8/2013, of December 9, 2013, for the improvement of the quality of education]
NGSS Lead States. (2013). The next generation science standards: for states, by states. The National Academies Press. https://doi.org/10.1016/j.endm.2015.07.014
Papavlasopoulou, S., Sharma, K., & Giannakos, M. N. (2018). How do you feel about learning to code? Investigating the effect of children’s attitudes towards coding using eye-tracking. International Journal of Child-Computer Interaction, 17, 50–60. https://doi.org/10.1016/j.ijcci.2018.01.004
Sáez-López, J.-M., Román-González, M., & Vázquez-Cano, E. (2016). Visual programming languages integrated across the curriculum in elementary school: a two year case study using “Scratch” in five schools. Computers & Education, 97, 129–141. https://doi.org/10.1016/j.compedu.2016.03.003
Schepers, J., & Wetzels, M. (2007). A meta-analysis of the technology acceptance model: Investigating subjective norm and moderation effects. Information and Management, 44(1), 90–103. https://doi.org/10.1016/j.im.2006.10.007
Shute, V. J., Sun, C., & Asbell-Clarke, J. (2017). Demystifying computational thinking. Educational Research Review, 22, 142–158. https://doi.org/10.1016/j.edurev.2017.09.003
Song, Y., & Kong, S.-C. (2017). Investigating students’ acceptance of a statistics learning platform using technology acceptance model. Journal of Educational Computing Research, 55(6), 865–897. https://doi.org/10.1177/0735633116688320
Stoilescu, D., & Egodawatte, G. (2010). Gender differences in the use of computers, programming, and peer interactions in computer science classrooms. Computer Science Education, 20(4), 283–300. https://doi.org/10.1080/08993408.2010.527691
Suki, N. M., & Suki, N. M. (2019). Structural relationships in the embedding of role-play games in a class for Japanese language proficiency: towards a unified view. Technology, Knowledge and Learning, 24(1), 65–87. https://doi.org/10.1007/s10758-016-9294-8
Sung, W., Ahn, J., & Black, J. B. (2017). Introducing computational thinking to young learners: practicing computational perspectives through embodiment in mathematics education. Technology, Knowledge and Learning, 22(3), 443–463. https://doi.org/10.1007/s10758-017-9328-x
Toma, R. B. (2020). STEM education in elementary grades: Design of an effective framework for improving attitudes towards school science (Unpublished doctoral dissertation). Burgos, Spain: University of Burgos
Yildiz Durak, H. (2020). The effects of using different tools in programming teaching of secondary school students on engagement, computational thinking and reflective thinking skills for problem solving. Technology, Knowledge and Learning, 25(1), 179–195. https://doi.org/10.1007/s10758-018-9391-y
Zogheib, B., & Daniela, L. (2021). Students’ perception of cell phones effect on their academic performance: a latvian and a middle Eastern university cases. Technology, Knowledge and Learning. https://doi.org/10.1007/s10758-021-09515-4
Acknowledgements
The author is very grateful to the participating schools, the expert panel, and the teachers who delivered the coding sessions.
Funding
Open Access funding provided thanks to the CRUE-CSIC agreement with Springer Nature. This research did not receive any specific grant from funding agencies in the public, commercial, or not-for-profit sectors. The author of this manuscript has no relevant financial or non-financial interests to disclose, nor have affiliation with or involvement in any organization or entity with any financial or non-financial interests in the subject matter or materials discussed in this manuscript.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interest
The author of this manuscript declare that there is no conflict of interest.
Human and animal rights
All procedures performed in studies involving human participants were in accordance with the ethical standards of the Vicerectorate of Research at the University of Burgos and with the 1964 Helsinki Declaration and its later amendments or comparable ethical standards. Informed consent was obtained from the legal custodians or parents of the individual participants involved in the study.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix 1: The Spanish version of the items used in this study
Factor 1: Actitudes hacia la programación visual
-
1.
Me gusta aprender sobre programación.
-
4.
Creo que programar es divertido.
-
7.
Estoy interesado/a en actividades relacionadas con la programación.
-
10.
Creo que la programación es genial.
Factor 2: Utilidad percibida de la programación visual
-
2.
Para mis estudios futuros, lo que aprendo en las actividades de programación es útil.
-
5.
No espero usar mucho la programación cuando termine la escuela.
-
8.
Las lecciones de programación son de poca utilidad para mí.
-
11.
La programación es muy importante para mi vida.
Factor 3: Facilidad de uso percibida de la programación visual
-
3.
Soy muy bueno/a programando.
-
6.
Para mí es muy fácil terminar las actividades de programación.
-
9.
No tengo que invertir mucho tiempo para ser bueno/a en la programación.
-
12.
Creo que soy muy malo/a programando.
*Las respuestas a los ítems 5, 8, y 12 deben ser invertidas antes de proceder a su análisis.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Toma, R.B. Measuring Acceptance of Block-Based Coding Environments. Tech Know Learn 28, 241–251 (2023). https://doi.org/10.1007/s10758-021-09562-x
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10758-021-09562-x