Abstract
The main goal of this study was to evaluate the impact of an animation and visualization of data structures (AVDS) tool on both perceptions and objective test performance. The study involved a rigorous experiment that assessed the usability, acceptability, and effectiveness of the AVDS tool in solving exercises. A total of 78 participants responded to questionnaires and were exposed to the AVDS tool, after which they completed a performance test, half (39) with the AVDS tool (the experimental group) and half (39) without the tool (the control group). Findings showed that the usability of AVDS was good; the experimental group even perceived AVDS usability as excellent. The results show that perceived usefulness, perceived ease of use, and attitudes toward usage jointly led to positive intention to use the AVDS tool. Furthermore, perceived ease of use was a key factor influencing participants’ intention to use AVDS. In addition, the AVDS system improved test results and provided flexibility in use, enhancing learning experience and performance.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
The fields of computer science and software engineering require a strong understanding of data structures and algorithms. These concepts form the foundation for solving complex problems and are necessary for developing efficient and scalable software systems. Understanding algorithms and data structures pose a significant challenge for undergraduate computer science and software engineering students (Kitchin, 2017). Computer code, words, and individual pictures represent static descriptions, specific views, or instances of a dynamic process. This should be given special attention during the learning process because some students cannot translate static descriptions such as these into a dynamic process in their imagination (Kitzie, 2019). As a result, these students often have difficulty understanding the algorithm’s code. In addition to empirical and mathematical analysis, algorithm visualization (AV) is another technique used in teaching data structures and algorithms. Visualization is the process of data abstraction, operations, semantics of the algorithm software, and the creation of graphical content of the abstractions. Creating a visualization of the algorithm should make it easier to understand (Seaver, 2018; Rajala et al., 2007).
Research has demonstrated that visualization and animation can effectively enhance student learning outcomes in data structures courses. Osman and Elmusharaf (Osman & Elmusharaf, 2014) conducted an experiment to compare the performance of computer science and information technology students with and without a visualized learning environment. The study found that using the visualized learning environment positively impacted student understanding and performance, as reflected in their scores on the final exam. Nathasya et al. (2019) integrated algorithm visualization, presenting DS-PITON as a tool to help students understand data structure implementation. The study aimed to address the limitations of traditional algorithm visualization tools, which do not provide enough technical details. Their study showed that using the integrated DS-PITON tool improved students’ assessment scores and helped moderate-paced students complete assessments faster. Andreev et al. (2021) examined the use of visualization technologies in distance education and found that visualization improves the quality and effectiveness of training, and that visualization tools help to improve distance learning for students. However, they acknowledged the need for further research to explore other aspects of visualization in professional training. Perháč and Šimoňák (2022) developed a new algorithm visualization system designed to teach data structures and algorithms. Their study aimed to improve the usability, extensibility, availability, and functionality of existing systems. The newly-developed system provided most functionalities for data structure visualization and was evaluated as easy to understand and helpful in learning. The respondents requested the addition of new visualizations and functionalities, indicating the potential for further development in this area.
This paper explores the benefits of incorporating animations and visualizations in teaching and learning data structures and algorithms in computer science courses. We describe the case study performed during the fall of 2021 in a Data Structures course at SCEFootnote 1 College. For this study, we developed a dedicated interactive website tool tailored to support the Data Structures course at our institution. The tool encompasses all the data structures and algorithms taught in the course. We demonstrate the potential impact of using visual aids in the classroom, including willingness to use the tool to clarify complex concepts and ease of using the tool, and enhanced problem-solving skills.
Research objectives
The primary objective of this research was to design and implement a free online animation and visualization tool to complement the undergraduate Data Structures course and evaluate its impact on student comprehension and performance. The developed tool covers the data structures taught in the course, providing the visual capability to represent the algorithm execution process. We investigated the effect of this specialized tool on students’ comprehension and academic performance in the Data Structures course at the Sami Shamoon College of Engineering.
We formulated the following research questions:
-
RQ1. How do students evaluate the use of animation and imaging technology in learning data structures concepts?
-
RQ2. How do students accept the AVDS tool?
-
RQ3. How will the use of AVDS affect the understanding of algorithms and learning?
Based on the research questions, the following hypotheses were formulated:
-
H.1. Using animation and imaging technology in teaching data structures will be useful to students.
-
H.2.1. Perceived ease of use (PEOU) will have a positive impact on users’ attitudes toward using (ATU) the AVDS tool.
-
H.2.2. Perceived usefulness (PU) will positively influence users’ ATU the AVDS tool.
-
H.2.3. The PEOU will positively influence users’ PU towards the AVDS tool.
-
H.3.1. Using animation and imaging technology in teaching data structures will enhance students’ performance in solving exercises.
-
H.3.2. Students can complete more questions in a given amount of time (i.e., completion time will be shortened).
Methodology
The study was carried out using a performance test as an objective measurement designed by the research team (Appendix A, Table 12). In addition, we applied the common standardized questionnaires, a Technology Acceptance Model (TAM) and System Usability Scale (SUS) questionnaires (Appendix A, Tables 10 and 11), which are acceptable for testing new technologies.
The content of the questionnaires and the order of their presentation were precisely determined in advance. The SUS and TAM questionnaires were translated from English by three people and then validated by three others. The questionnaires included closed-ended questions in which students needed to choose an answer from a pre-determined set of answers. The performance test included two types of questions: open-ended and partially closed-ended; the latter required written justification of the answer. The answers were coded according to the percentage of correctness from 0 to 100.
Participants
A survey was taken among second-year students enrolled in the Data Structures course at the SCE College, Beer Sheva, Israel. Of 162 students in the course, 78 showed interest in participating in the research study: 23 females (mean age 21, SD = 2.56) and 55 males (mean age = 21, SD =1.97). Since all the students study in the same department and have completed all pre-course requirements, we assume they are similar to each other in terms of their educational background and academic level, and they can be classified as a homogeneous population.
Questionnaires
The SUS is a standardized questionnaire used to measure the perceived usability of a system (Suharsih et al., 2021). The SUS score is calculated by obtaining responses from participants on a 10-item questionnaire. Every item contribution score ranged between 1 (strongly disagree) and 5 (strongly agree) (Brooke et al., 1996).
The TAM (Al-Emran and Shaalan, 2021; Agbo et al., 2022; Dimitrijević and Devedžić, 2021) questionnaire used in the study measures three key factors related to the acceptance of the AVDS tool:
-
1.
Perceived usefulness (Davis et al., 1989): The extent to which individuals believe that using the AVDS tool will enhance their test performance and time efficiency.
-
2.
Perceived ease of use (Davis et al., 1989): The extent to which a student’s use of AVDS tools is free of effort.
-
3.
Attitude toward usage (Davis, 1993): The extent to which an individual’s attitude is favorably or unfavorably disposed toward using the AVDS tool and the influence of this attitude on the intention to use the tool.
These factors are measured to determine the level of acceptance and adoption of the AVDS tool and to provide insights into areas that need improvement.
Performance test
The performance test (see Appendix A, Table 12) helped researchers learn about the efficiency of the tool. For the test, we chose the AVL tree topic, a self-balancing binary search tree developed in 1963 (Adelson-Velskii & Landis, 1963). The AVL topic was taught in class a week before the experiment. The test consisted of six questions and lasted about 20 min. During the test, students in an experimental group used the developed Data Structures online tool, while the control group solved without that assistance. The AVDS tool helped the student solve the questions or verify the answers. At the time of the experiment, the system did not show the steps of the algorithm, but did provide a final answer only. To assess the students’ understanding of the solution, the assessment included more than just the final answer. It required students to explain the steps involved. For example, students needed to determine if a rotation was necessary during the insertion of a new node in the given AVL tree.
Data structures online tool
We found several publicly available AV tools: Data structure visualization, Visualgo, CS1332 Visualization Tool, Algorithm Visualizer, Vamonos, and Algorithm Wiki. However, the lack of several functionalities and features posed difficulties in utilizing them in our course and study. Some tools may not cover basic data structures, have restrictive licenses that prevent customization for education and research, lack required controls for navigating through data structures and algorithms, or do not cover some topics in enough detail. In addition, some tools do not provide the essential levels of freedom for students, such as zooming and panning to focus on specific data regions, the ability to easily export the visualized data, or to save the history for future reference. Another issue encountered is the unavailability of all necessary visualizations for our course within a single package. For these reasons, developing an alternative tool tailored to our needs seemed the best alternative.
We designed and implemented a free online animation and visualization tool that complements the undergraduate Data Structures course and provides an interactive learning experience on a deep level. The AVDS tool we developed is the one we utilized in our study. The AVDS tool covers a classic set of algorithms and data structures commonly studied in the undergraduate data structures course. The source code for the tool is available on the GitHub repository (https://github.com/genakogan/data-structures-21-sce.git), allowing others to build upon or modify it as they see fit. The AVDS tool can be especially helpful for students seeking to understand the code behind the visualization, as they can see precisely how the algorithms and data structures are implemented. While understanding the source code may require some technical skills, such as programming experience with JavaScript, HTML, and CSS, the GitHub repository can be a valuable resource for educators and students seeking to gain a deeper understanding of algorithms and data structures through visualization.
Experimental procedure
The experiment involved three stages: (1) Survey preparation, (2) Survey conduction, and (3) Data analysis. Below, we provide an overview of each, with Fig. 1 illustrating the experimental pipeline.
Stage 1: Survey preparation. The preparation consisted of the following steps:
-
1.1: Preparing the teaching staff. The teaching assistants received an explanation of the technology (AVDS) and its use, as well as an explanation of the experimental procedure.
-
1.2: Preparing the tool demonstration. A short video presenting the AVDS tool was recorded before the experiments. This video was available to the students before the beginning of the experiment (stage 2.1). The video demonstrated the AVDS tool described in Sect. 3.4, which students used during the survey.
-
1.3: Random assignment of students to experimental and control groups, with 39 participants in each.
-
1.4: AVL lecture, which was taught in class a week before the experiment.
Stage 2: Conducting the survey. The experiment took place during the practical session, the next week after the AVL tree topic was introduced in the lecture. The experiment was carried out as a sequence of the following steps (summarized in Table 1):
-
2.1: The students in both groups were provided with an explanation and clarification about how to use the AVDS online tool.
-
2.2: The students in each group completed a TAM questionnaire.
-
2.3: The students took the test: the experimental group used the AVDS tool, and the control group took the test without the AVDS tool. The test duration was approximately 20 min.
-
2.4: The students in the experimental group completed a TAM questionnaire a second time.
-
2.5: The students in both groups completed the SUS questionnaire.
Stage 3: Data analysis. The data analysis included cross-referencing data within the experimental group and between the groups, analyzing the results, and drawing conclusions.
Experimental design and variables
To analyze the results, we used integrated experimental sampling (De Datta, 1978). Integrated experimental sampling enables analysts to perform comparisons between groups as well as within the groups. There are three components in experimental design used to establish cause-and-effect relationships between variables by controlling for extraneous variables and reducing bias (Gribbons & Herman, 1996):
-
X—Activation of the manipulation, in our case, using the AVDS tool.
-
O—Indicates an observation or measurement of the dependent variables: \(O_1, O_2, \ldots , O_7\).
-
R—Random division into groups.
The study involved several observations that are defined in Table 2, including a pre-test and a post-test TAM questionnaire, a performance test, and a SUS questionnaire. Table 3 shows the steps of the experimental design used to evaluate the performance of the AVDS system. Using random division (R), the study included two groups: the experimental group (with access to the AVDS tool) and the control group (without access to the AVDS). The participants in both groups were given a TAM questionnaire (\(O_1\), \(O_2\), \(O_5\)). The experimental group used the AVDS tool (X) during the performance test, while the control group did not (\(O_3\), \(O_4\)). Additionally, both groups were given a SUS questionnaire (\(O_6\), \(O_7\)).
Independent Variables. The independent variables were: (1) Use of the AVDS tool; (2) Perceived usability of the AVDS tool; (3) Perceived usefulness of the AVDS tool; (4) Perceived ease of use of the AVDS tool; (5) Attitudes toward the use of the AVDS tool; (6) Experience using the AVDS; (7) Group assignment (experimental/control).
Dependent Variables. Dependent variables included: (1) Evaluation of the usability of the AVDS tool; (2) Assessment of the usefulness of the AVDS tool; (3) Acceptance of the AVDS tool; (4) Performance in solving exercises; (5) Response time to complete questions.
Analysis and findings
The results of the questionnaire surveys, based on 78 student responses, were analyzed using a combination of statistical tools, including the SPSS 28Footnote 2 analysis software, IBM SPSS AMOS 27 module for structural equation modeling (SEM) analysis (Bahcivan et al., 2019; Blunch, 2012; Guo et al., 2022; Yildiz Durak, 2019), and KaleidaGraphFootnote 3 for graph construction.
System usability scale score—(H.1.)
The SUS scores were converted to percentile ranks and adjective ratings (such as "very good" or "poor"). The combination of percentile ranks and adjective ratings can provide a more precise picture of the usability of a product or system (Bangor et al., 2009; Brooke, 2013).
The percentile ranks of both groups are presented in Fig. 2 (left-experimental group, right-control group). The mean score of the experimental group was 79.17, which is considered "Excellent," while the control group had a score of 69.23, which is considered "Good."
These results suggest that the AVDS tool was perceived as more usable by the experimental than the control group, with higher levels of ease of use and user perception of effectiveness and efficiency.
Technology acceptance model—(H2.1., H2.2. and H2.3.)
The study used a two-measurement model with three constructs: PU, PEOU, and ATU, depicted in Fig. 3. The TAM constructs are measured using multiple items, each of which measures a different aspect of the construct (see Appendix A, Table 10 for details).
The validity and reliability of these measures are evaluated to improve confidence in the results and quality of these models:
-
1.
First model with AVDS before (\(O_1\)) and after (\(O_5\)) the performance test (experimental group).
-
2.
Second model without AVDS before (\(O_2\)) performance test (control group) and with AVDS after (\(O_5\)) performance test (experimental group).
This improves confidence in the results and quality of the models, ensuring accurate reflection of intended constructs and consistent results, thereby enhancing credibility and providing overall confidence in the model.
Improving and validating the model
The results of a Cronbach’s alpha reliability coefficient, shown in Table 4, indicate the internal consistency of the scale by measuring the degree of interrelatedness among the questionnaire items. The table displays the corrected item-total correlation for each item, and the Cronbach’s alpha result if the item is deleted from the scale. It suggests that the PEOU2 and PEOU4 items measure a similar construct and explain the high R-squared value. Removing PEOU2 from the model shown in Fig. 3 decreased the R-squared value of the attitude toward usage variable below one. Overall, the results demonstrate high internal consistency and reliability of the TAM questionnaire items.
Adding a double-curved arrow between two error terms can be a potential solution for a Heywood case (Wang et al., 2021). A Heywood case refers to a scenario in the SEM model where the estimated covariance matrix of latent variables produces unfeasible or unreasonable results, such as negative variances or correlations greater than 1.0. In Fig. 4, the solid double-curved arrow is used to represent a correlated error in the first model, and the dashed double-curved arrow is used for the second model. This indicates that the error terms share a common source of variance, which could be due to similar conceptual content. This approach can help improve the model’s fit and address issues with model identification.
The results of model validation demonstrate a moderate to good model fit to the data, with reasonable fit indices (see Appendix B, Table 13), including RMSEA and RMR, which indicate a close model fit and a good fit, respectively. The factor loadings of the indicators (Appendix B, Table 14 ) were found to be greater than 0.4, demonstrating the construct validity of the instrument. The convergent validity was good, with AVE values greater than 0.5 (Appendix B, Table 15 ). The discriminant validity was satisfactory, with the exception of one case where the construct PEOU had poor discriminant validity to ATU (see Appendix B, Table 16 ). Finally, the constructs had acceptable consistency, as shown by the composite reliability values, which were greater than or equal to 0.6 (Appendix B, Table 17).
Hypotheses testing
The chi-square for the structural model was calculated before and after applying the weight constraints to the research hypotheses (Al-Aulamie, 2013). Table 5 shows the chi-square, degrees of freedom, and p-value for two structural models, before and after applying weight constraints, and the difference between them. The results indicate that the constrained models have a higher chi-square and lower p-value than the unconstrained models, suggesting that the relationships among the variables differ between cases at the model level.
After establishing a significant difference between the cases presented in Table 5, the next step is identifying the hypotheses that may account for these differences. Identification is accomplished by performing the weight constraints method separately on each hypothesis in two models.
For the first model, the first hypothesis (H2.1) was verified using the relationship between PEOU and ATU (see Table 6). The t-values calculated for PEOU and ATU are equal to 3.480 and 5.502, respectively, for before (\(O_1\)) and after (\(O_2\)) using AVDS. Both t-values are greater than the Z-score which is equal to 1.977. Thus, the relationship between variables related to PEOU of technology with an ATU is acceptable because it gives a positive and significant influence, meaning that hypothesis H2.1 can be proven acceptable. The standardized coefficient showed that to be the case after using AVDS, resulting in a stronger relationship (\(\beta\) = 0.755, p < .01) than before using AVDS (\(\beta\) = 0.752, p < .01).
In the second hypothesis (H2.2), by examining the relationship between PU and ATU, the resulting t-value of 2.567, which represents the result before (\(O_1\)) using AVDS, and 2.224 after (\(O_2\)) using AVDS, is stated to be greater than 1.977. It can be concluded that the relationship between variables on the construct of PU of AVDS technology is proven to have a positive and significant influence on ATU, which means that H2.2 is proven to be accepted. The standardized coefficient showed that the case before using AVDS has a stronger relationship (\(\beta\) = 0.345, p = .010) than after using AVDS (\(\beta\) = 0.274, p = .026).
The third hypothesis (H2.3) was tested using the relationship between PEOU and PU variables and produced a t-value of 4.665 > 1.977 after using AVDS. At the same time, the case before using AVDS showed that the t-value is less than the Z-score (t = 1.829, Z-score = 1.977; p = .067). Thus, the relationship between variables regarding the PEOU of technology on the PU has a positive and significant influence, meaning that H2.3 is proven to be accepted only after using AVDS (\(\beta\) = 0.726, p < .01).
The structural model test revealed that the variance in the dependent variable PU that was explained by the independent variable PEOU, differed between \(O_1\) (12%) and \(O_5\) (53%). Additionally, the predictors PU and PEOU accounted for 98% of the variation in ATU for \(O_1\) and 95% for \(O_5\), according to the R square value, which means that this model is rational, although other unknown factors may impact dependent variables.
For the second model (see Table 7) hypothesis H2.1 was verified using the relationship between PEOU and ATU. The t-values calculated for PEOU and ATU equal 3.790 and 5.586, respectively, for the control group (case \(O_2\)) and experimental group (case \(O_5\)). Both t-values are greater than the Z-score, which is equal to 1.977. Thus, the relationship between variables related to the PEOU of technology with an ATU is acceptable because it gives a positive and significant influence, meaning that hypothesis H2.1 can be proven acceptable. A standardized coefficient showed that the control group (case \(O_2\)) had stronger relationship (\(\beta\) = 0.753, p < .01) than the experimental group (\(\beta\) = 0.740, p < .01) (case \(O_5\)).
Hypothesis H2.2 was checked out using the relationship between PU and ATU variables. In this case, it was found that the t-value for the experimental group is equal to 2.500, which is greater than the Z-score. At the same time, a comparison of the t-value with the Z-score carried out for the control group showed that, in that case, the t-value was less than the Z-score (t = 1.471, Z-score = 1.977; p = .141). Therefore, the relationship between variables on the construct of PU is proven to have a positive and significant influence on ATU, which means that H2.2 is proven to be accepted only for the experimental group (\(\beta\) = 0.302, p = .012) (case \(O_5\)).
Hypothesis H2.3 was tested using the relationship between PEOU and PU variables. The t-value of 5.618 > 1.977 (Z-score) for the control group (case \(O_2\)) and 4.415 > 1.977 (Z-score) for the experimental group (case \(O_5\)). Thus, the relationship between variables has a positive and significant influence meaning that H2.3 from both relations can be accepted. A standardized coefficient showed that the control group (case \(O_2\)) had a stronger relationship (\(\beta\) = 0.831, p < .01) than the experimental group (case \(O_5\)) (\(\beta\) = 0.696, p < .01).
Additionally, the results of the structural model test showed that the impact of the independent variable PEOU on the dependent variable PU differed between cases \(O_2\) (12%) and \(O_5\) (53%). The predictors PU and PEOU were found to explain 95% of the variation in ATU for both cases, indicating that the model has a high level of explanatory power. However, it is important to note that the model may not account for other unknown factors that could also impact the dependent variables.
Multiple linear regression for performance—(H3.1. and H3.2.)
A significant difference existed in the test scores of the experimental and control groups (see Fig. 5). For the experimental group, the test scores were considerably higher: more than \(75\%\) of students who got grades from 80 to 100 belonged to the group that used the AVDS tool. The findings also indicate uniform distribution among the students that used the AVDS tool: 37 students out of 39 received grades above 80. Furthermore, students in the experimental group demonstrated significantly shorter response times.
Additionally, we used multiple linear regression to predict the performance test (Table 8). The model is developed for the following case: control group \(O_2\) without AVDS (i.e., before the test) and experimental group \(O_5\) with AVDS (i.e., after the test ), which answers the TAM questionnaire after using the tool. The performance test is based on four significant main effects: PU1 (p \(\le\) .027), ATU3 (p \(\le\) .004), PEOU3 (p \(\le\) .030), and group with (after) or without AVDS (before) (p \(\le\) .001). A significant regression equation was found (F(4.73) = 13.728, p <.001), with an adjusted \(R^2\) of 0.398. The predicted test performance of participants is equal to 0.424 + 9.431 (ATU3) − 4.415 (PU1) − 5.602 (PEOU3) + 32.428 (before and after). For the first main effect, higher perceived flexible use leads to a higher performance test (\(\beta\) = 9.431, SE = 3.137). In addition, it can be concluded that for the experimental group that used AVDS, the higher perceived improvement performance (\(\beta\) = \(-\)4.415, SE = 1.959) resulted in lower test performance. Similar to the previous findings, higher perceived easiness to the question PEOU3 (\(\beta\) = \(-\) 5.602, SE = 2.527) led to lower test performance. Generally, the test performance was higher for the experimental group (\(\beta\) = 32.428, SE = 4.9).
The second multiple linear regression was calculated to predict performance time (Table 9). The model was developed for the groups described previously in the first regression. The performance time is based on three significant effects: (a) PEOU2 (p \(\le\) .008), (b) PEOU3 (p \(\le\) .035), and (c) group with (after) AVDS or without (before) AVDS (p \(\le\) .001). A significant regression equation was found: (F(3,74) = 15.700, p <.001), with an adjusted \(R^2\) of 0.364. The predicted performance test of participants is equal to 15.700–1.066 (PEOU2) + 0.756 (PEOU3) − 4.707 (before and after). For the first main effect, for the experimental group that used AVDS, the PEOU2 (\(\beta\) = \(-\) 1.066, SE = 0.388) led to lower performance time results. Moreover, the higher PEOU3 (\(\beta\) = 0.76, SE = 0.353) resulted in higher time performance. Mostly, the performance time decreased for the experimental group (ES = \(-\) 4.71, SE = 0.8).
Limitations of the study
This study has limitations. Firstly, the small sample size may be unreliable, as RMR, NFI, GFI, RMSEA, and regression coefficients among variables are often done by Maximum Likelihood (Rao et al., 2011), which assumes normal distribution among the indicator variables. As a general rule, sample size requirements for SEM are often much larger than for other types of statistical analyses, such as multiple regression. A minimum sample size of 200 to 300 observations is often recommended for SEM, but the specific requirements will depend on the specific model and the data being analyzed. (Wolf et al., 2013). Secondly, this study may not fully capture the periodicity of AVDS usage. Therefore, the results of this study should be viewed as preliminary evidence for examining the relationship between students and their intention to use the AVDS system. Thirdly, the development of the site was based mainly on the course staff requirements, and was evaluated by the students’ course. Although, results show that most of the participants found the system with high usability, ∼ 1% of the participants in both groups rated usability as ’worst imaginable’. Therefore, the next phase of the development should take into account the requirements of the pilot outcomes and the students’ inputs.
Discussion
This research aimed to examine the efficacy of the AVDS tool in improving students’ learning and understanding of data structures and algorithms. The study found that the AVDS tool had a positive impact on students’ perception of usability, usefulness, and acceptance of the tool. The results showed that the AVDS tool was acceptable and effective for students’ learning.
The SUS score of the AVDS tool for the control group was 69.23 (in the "Good" category); for the experimental group, the score was 79.17 (in the "Excellent" category). This indicates that the AVDS tool is good and acceptable.
Using the TAM theory, the study explored the factors influencing the adoption of the AVDS tool by students. The findings suggested that students’ perception of the AVDS tool’s ease of use significantly affected their attitudes toward using the tool. If students found the tool easy to use, they were more likely to consider it useful and have a positive attitude toward integrating technology into learning. The study also showed that the AVDS tool improved students’ performance in terms of grades and completion time, such that \(\sim 95\%\) of them achieved grades above 80. These findings suggest that the developed AVDS tool was an effective learning medium that was easy for users to learn, understand, and operate. Overall, the study’s findings can be used to enhance understanding of user acceptance of the AVDS tool and lead to more successful adoption. The results of our study align with previous research (Andreev et al., 2021; Nathasya et al., 2019; Osman & Elmusharaf, 2014; Perháč & Šimoňák, 2022) which demonstrated that using visualized learning environments positively impacts student understanding, performance, and assessment completion times.
The application can help promote the learning of the Data Structures and Algorithms topics in the undergraduate computer science and software engineering courses. Beyond the common functionalities, the proposed tool provides valuable analytics for the teaching staff, e.g., usage reports for each topic. These reports can be utilized to prioritize more challenging topics during the course.
Conclusions and future work
This study investigated the effects of the AVDS tool, using a wide range of variables including TAM, SUS, and the students’ performance. According to the SUS results, we found that the experimental group perceived the usability of AVDS to be "Excellent", while the control group estimated it only as "Good".
Similarly, TAM findings showed that the experience with AVDS led to positive intentions to use the tool and higher levels of ease of use, effectiveness, and efficiency. Moreover, there is evidence that the use of AVDS elevated the overall learning experience and achievement levels. The students in the experimental group obtained higher and more uniform scores than those in the control group, and it took them less time to solve tasks than the control group.
At present, our team is actively enhancing the application by improving its functionalities and addressing usability issues identified in the current study. Recently we conducted a second additional pilot to evaluate the effectiveness of the additional developments and now we are analyzing its outcomes. The current version of the website may be accessed at https://ds-sce.sce-fpm.com/.
Data availability
Not applicable.
Code availability
References
Adelson-Velskii, M., & Landis, E. M. (1963). An algorithm for the organization of information. Tech. rep., Joint Publications Research Service Washington DC. https://zhjwpku.com/assets/pdf/AED2-10-avl-paper.pdf
Agbo, F. J., Oyelere, S. S., Suhonen, J., & Tukiainen, M. (2022). Design, development, and evaluation of a virtual reality game-based application to support computational thinking. Educational Technology Research and Development, 71, 1–33. https://doi.org/10.1007/s11423-022-10161-5
Al-Aulamie, A. (2013). Enhanced technology acceptance model to explain and predict learners’ behavioural intentions in learning management systems. PhD Thesis, University of Bedfordshire, http://hdl.handle.net/10547/323773
Al-Emran, M., & Shaalan, K. (2021). Recent advances in technology acceptance models and theories. Springe. https://doi.org/10.1007/978-3-030-64987-6
Andreev, A. N., Emtseva, O. V., Babkina, E. S., Dolzhenkov, V. N., & Kudryashova, N. V. (2021). The effectiveness of using visualization tools and forms in distance learning. Revista Tempos E Espaços Em Educação, 14(33), 2. https://doi.org/10.20952/revtee.v14i33.16053
Bahcivan, E., Gurer, M. D., Yavuzalp, N., & Akayoglu, S. (2019). Investigating the relations among pre-service teachers’ teaching/learning beliefs and educational technology integration competencies: A structural equation modeling study. Journal of Science Education and Technology, 28, 579–588. https://doi.org/10.1007/s10956-019-09788-6
Bangor, A., Kortum, P., & Miller, J. (2009). Determining what individual sus scores mean: Adding an adjective rating scale. Journal of usability studies, 4(3), 114–123.
Blunch, N. (2012). Introduction to structural equation modeling using IBM SPSS statistics and AMOS. Sage.
Brooke, J., (1996). Sus-a quick and dirty usability scale. Usability Evaluation in Industry, 189(194), 4. https://doi.org/10.1201/9781498710411
Brooke, J. (2013). Sus: a retrospective. Journal of Usability Studies, 8(2), 29–40.
Davis, F. D. (1993). User acceptance of information technology: system characteristics, user perceptions and behavioral impacts. International Journal of Man-Machine Studies, 38(3), 475–48. https://doi.org/10.1006/imms.1993.1022
Davis, F. D., Bagozzi, R. P., & Warshaw, P. R. (1989). User acceptance of computer technology: A comparison of two theoretical models. Management Science, 35(8), 982–1003. https://doi.org/10.1287/mnsc.35.8.982
De Datta, S. K. (1978). A handbook on the methodology for an integrated experiment: survey on rice yield constraints. Int. Rice Res. Inst.
Dimitrijević, S., & Devedžić, V. (2021). Utilitarian and experiential aspects in acceptance models for learning technology. Educational Technology Research and Development, 69, 627–65. https://doi.org/10.1007/s11423-021-09970-x
Gribbons, B., & Herman, J. (1996). True and quasi-experimental designs. Practical Assessment, Research, and Evaluation, 5(1), 1. https://doi.org/10.7275/fs4z-nb61
Guo, Q., Qiao, C., & Ibrahim, B. (2022). The mechanism of influence between ict and students’science literacy: a hierarchical and structural equation modelling study. Journal of Science Education and Technology, 31(2), 272–288. https://doi.org/10.1007/s10956-021-09954-9
Joseph, F., Black, W. C., Babin, B. J., & Anderson, R. E. (2010). Multivariate data analysis: A global perspective. Pearson Educatio, 10, 2141158.
Kenny, D. A., & McCoach, D. B. (2003). Effect of the number of variables on measures of fit in structural equation modeling. Structural Equation Modeling, 10(3), 333–35. https://doi.org/10.1207/S15328007SEM1003_1
Kenny, D. A., Kaniskan, B., & McCoach, D. B. (2015). The performance of rmsea in models with small degrees of freedom. Sociological Methods & Research, 44(3), 486–507.
Kitchin, R. (2017). Thinking critically about and researching algorithms. Information, Communication & Society, 20(1), 14–29. https://doi.org/10.1080/1369118X.2016.1154087
Kitzie, V. (2019). “that looks like me or something i can do”: Affordances and constraints in the online identity work of us lgbtq+ millennials. Journal of the Association for Information Science and Technology, 70(12), 1340–1351. https://doi.org/10.1002/asi.24217
Kline, R. B. (1998). Structural equation modeling. Guilford.
Nathasya, R. A., Karnalim, O., & Ayub, M. (2019). Integrating program and algorithm visualisation for learning data structure implementation. Egyptian Informatics Journal, 20(3), 193–20. https://doi.org/10.1016/j.eij.2019.05.001
Osman, W. I., & Elmusharaf, M. M. (2014). Effectiveness of combining algorithm and program animation: A case study with data structure course. Issues in Informing Science and Information Technology, 11, 155–168.
Perháč, P., & Šimoňák, S. (2022). Interactive system for algorithm and data structure visualization. Computer Science Journal of Moldova, 30(1), 28–48.
Rajala, T., Laakso, M. J., Kaila, E., & Salakoski, T. (2007). Ville: A language-independent program visualization tool. Proceedings of the Seventh Baltic Sea Conference on Computing Education Research-Citeseer, 88, 151–159.
Rao, C., Miller, J., & Rao, D. (2011). Essential statistical methods for medical statistics. Elsevier.
Schumacker, R. E., & Lomax, R. G. (2004). A beginner’s guide to structural equation modeling. Psychology Pres. https://doi.org/10.4324/9780203851319
Seaver, N. (2018). What should an anthropology of algorithms do? Cultural Anthropology, 33(3), 375–38. https://doi.org/10.14506/ca33.3.04
Steiger, J. H. (1980). Statistically based tests for the number of common factors. In: Paper presented at the Annual Meeting of the Psychometric Society, Iowa Cyty.
Suharsih, R., Febriani, R., & Triputra, S. (2021). Usability of jawara sains mobile learning application using system usability scale (sus). Jurnal Online Informatika, 6(1), 41–5. https://doi.org/10.15575/join.v6i1.700
Wang, S., De Boeck, P., & Yotebieng, M. (2021). Heywood cases in unidimensional factor models and item response models for binary data. Applied Psychological Measurement, 47(2), 141–15. https://doi.org/10.1177/01466216231151701
Wolf, E. J., Harrington, K. M., Clark, S. L., & Miller, M. W. (2013). Sample size requirements for structural equation models: An evaluation of power, bias, and solution propriety. Educational and psychological measurement, 73(6), 913–934. https://doi.org/10.1177/0013164413495237
Yildiz Durak, H. (2019). Examining the acceptance and use of online social networks by preservice teachers within the context of unified theory of acceptance and use of technology model. Journal of Computing in Higher Education, 31(1), 173–20. https://doi.org/10.1007/s12528-018-9200-6
Funding
Open access funding provided by Sami Shamoon College of Engineering. No funding was received for conducting this study.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors have no competing interests to declare that are relevant to the content of this article.
Ethical approval
The study was conducted in compliance with ethical standards and regulations and was approved by the college ethical committee.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kogan, G., Chassidim, H. & Rabaev, I. The efficacy of animation and visualization in teaching data structures: a case study. Education Tech Research Dev (2024). https://doi.org/10.1007/s11423-024-10382-w
Accepted:
Published:
DOI: https://doi.org/10.1007/s11423-024-10382-w