Introduction

One of the most significant changes to take place in the approach to teaching the first 2 years of medical education (pre-clerkship) in the last several decades is the conversion of discipline-based to integrated curricula. Teaching basic science content in conjunction with clinical disciplines provides medical students motivation to learn important basic science content and encourages faculty to organize material in a way that is beneficial to future physicians [1]. Although curricula at individual institutions vary, in many cases, curricular integration includes integrated examinations. Assessing content that traditionally involves a laboratory component, such as gross anatomy and histology, is often an exception and retains its own stand-alone, practical examination (e.g., [2]).

In parallel with curricular changes, technological advances have made a significant impact on medical education. In particular, histology (microscopic anatomy) instruction has transformed dramatically over the past 20 years due to the near universal implementation of virtual microscopy to supplement or replace study of histology using glass slides viewed through microscopes [3,4,5,6,7,8,9,10,11,12,13]. Although the implementation was gradual, by 2017, most medical schools in the USA had incorporated virtual microscopy into the teaching of histology [14].

Evolution of Histology Education at the University of Cincinnati College of Medicine

Histology instruction at the University of Cincinnati College of Medicine (UCCOM) underwent a substantial shift following the transition from stand-alone courses to an integrated, organ-systems-based curriculum in 2011. By that point, the implementation of virtual microscopy had already begun, starting with the use of virtual slides as a supplement to glass slides and microscopes. Alongside this change was a shift from histology being taught in a traditional, lecture/laboratory format to histology instruction being almost entirely virtual via self-study modules. Previous research has assessed the impact of this alteration, which was largely positive in terms of student outcomes and preference [15]. When this change occurred around 2016, histology was taught within the integrated courses, but still tested independently on stand-alone, practical examinations. As the curriculum continued to evolve, there were several factors that led to a push for histology content to be included on the integrated, computer-based lecture examinations. First, a declining emphasis on microscope-related skills resulted in digital slides being used more frequently on histology examinations. This brought up the question of whether histology needed a separate examination when digital images of slides could easily be included on the computer-based, integrated examinations. Second, having histology tested separately allowed students to study a large volume of material over a short period of time leading up to the examination. Pedagogically, this approach is not favorable for long-term retention [16, 17]. Including histology on integrated exams would, in theory, require students to consistently review the material and integrate it with related lecture content. Lastly, including histology on integrated examinations would eliminate the former stand-alone practical examination, providing both students and course directors with greater flexibility in scheduling.

When making a substantial change in assessment strategy, it is critical to evaluate how student learning outcomes are influenced. The purpose of this study is to determine how student performance, study approach, and student preference were impacted when histology was assessed as part of integrated versus stand-alone examinations. In addition to overall trends, this study also investigates whether students at different levels of academic achievement were affected equally and if any temporal trends in student performance exist across the transition in assessment strategy.

Materials and Methods

To investigate the transition from stand-alone to integrated histology examinations, several different cohorts and two different courses at the University of Cincinnati College of Medicine were considered (Table 1). The academic years investigated ranged from 2018 to 2022. Over this period, the examination format of histology changed from stand-alone to integrated examinations. The first course to incorporate this transition in 2020 was the Gastrointestinal, Endocrine, and Reproduction (GER) course, which is part of the second-year medical school curriculum. In 2022, the Musculoskeletal-Integumentary (MSK) course, which is part of the first-year medical school curriculum, also shifted from stand-alone to integrated histology examinations. To evaluate whether student performance was impacted by this change in assessment format, data from stand-alone examinations that occurred 2 years prior to these transition points were included.

Table 1 Overview of years and courses investigated in this study

While similar changes occurred in other courses in the curriculum, the GER and MSK courses were selected for this study because the assessment items over this transitional period were the most comparable in these courses. Only questions that were exactly the same before and after the assessment format change were included since that provides the most direct comparison. In total, 18 questions in the GER course and 16 questions in the MSK course met this criterion. When histology was included on integrated examinations, both courses employed assessment strategies that spread the histology content out over three examinations that were each separated by 2-week time periods.

Study Design

Examination data were investigated using several different approaches. The first was to compare overall student performance on histology questions when delivered as a stand-alone versus part of an integrated examination within each course. Next, class rank was used to group students in each cohort into an upper and lower 50%. This was done to determine if the change in examination format disproportionally affected higher- or lower-performing students. For this study, class rank reflects a calculation that took place at the end of the academic year a given class was enrolled in either the MSK or GER course. This calculation includes all course grades leading up to that point in time. Lastly, student performance was investigated by year to determine if there were any temporal trends that might be obscured with pooled data.

In addition to examination data, students enrolled in the 2022 MSK course were surveyed to elicit their feedback on the change in assessment strategy. This group of students was selected because the course that occurs just prior to MSK, Fundamentals of Cellular Medicine, still tests histology as a stand-alone examination. At the end of the MSK course, students were given an open-ended question asking them to comment on whether their approach to studying histology changed when it was tested on an integrated versus stand-alone examination.

Statistical Methods

All examination data were analyzed using analysis of covariance (ANCOVA), with scores on the Medical College Admissions Test (MCAT) used as a covariate. This was done to account for the possibility that differences in student aptitude could impact the results. When applicable, a Bonferroni post hoc test was used to evaluate pairwise comparisons. The effect size for ANCOVA results are reported as eta squared n2, which was calculated by dividing the sum of squares between by the sum of squares total [18]. Cohen [19] suggests that values should be interpreted as follows: 0.01 = small, 0.06 = medium, and 0.14 = large. All quantitative data were analyzed using SPSS version 27 (IBM Cort., Armonk, NY). The student survey results were analyzed and reported using thematic analysis [20]. This study was approved by the Institutional Review Board at the University of Cincinnati College of Medicine (IRB#2021–1061).

Results

Overall Comparison

The mean score for the two cohorts that took the stand-alone histology examination in the GER course was 90.8 ± 8.8%. In contrast, when the same questions were asked as part of integrated examinations, the average was 86.4 ± 11.8%. ANCOVA results indicated this difference is statistically significant (p < 0.01) with a small to medium effect size n2=0.043 (Table 2). In the MSK course, similar results were observed. The class average for students who took stand-alone histology examinations was 92.2 ± 8.2% whereas students who were assessed on histology as part of the integrated examinations averaged 81.8 ± 11.1%. This difference was statistically significant (p < 0.01) with a large effect size n2=0.201 (Table 2).

Table 2 Comparison of examination performance by delivery format

Comparison by Class Standing

To further investigate the difference in performance, students within each course and cohort were divided into an upper and lower 50% according to class rank. In the GER course, this revealed that students in the upper 50% of the class who were tested on histology as a stand-alone examination averaged 94.7 ± 5.1% whereas students in the upper 50% who were tested as part of an integrated examination averaged 91.0 ± 7.4%. This difference was statistically significant (p < 0.01) with a medium effect size n2=0.068. A similar pattern was also present among students who ranked in the lower 50% of the class, where the stand-alone examination average was 87.1 ± 10.1% compared to an integrated examination average of 81.8 ± 13.4%. This difference was also statistically significant (p < 0.01) with small-medium effect size n2=0.044 (Table 3).

Table 3 Comparison of examination performance by class standing

In the MSK course, students in the upper 50% who took stand-alone histology examinations averaged 95.2 ± 5.5% whereas students in the upper 50% who took an integrated histology examination averaged 88.5 ± 9.2%. This difference was statistically significant (p < 0.01) with a large effect size n2=0.160.Students in the lower 50% who took stand-alone examinations averaged 89.2 ± 9.3%, while students in the lower 50% who took integrated histology examinations averaged 74.7 ± 13.2%. This difference was statistically significant (p < 0.01) with a large effect size n2=0.284 (Table 3).

Comparison by Cohort

Examination averages by cohort were considered to determine if there were any patterns in the data over time. Within the GER course, ANCOVA revealed a statistically significant difference in examination scores between the five cohorts (p < 0.01) with a medium effect size (n2 = 0.074). While there were significant differences between several post hoc comparisons, the most notable finding was that the examination averages the initial year histology was integrated were significantly different than every other year that was analyzed (Table 4). The longitudinal trend of scores in the GER course is also visualized in Fig. 1. This illustrates how the initial decline in performance during the first year of integrated examinations is followed by results in the two most recent years that are comparable to when the examinations were delivered in a stand-alone format.

Table 4 p value post hoc comparison of examination performance by cohort
Fig. 1
figure 1

Temporal trends in examination data in the GER course. Error bars represent one standard deviation from the mean

In the MSK course, there was a statistically significant difference between the three cohorts that were compared (p < 0.01) with a large effect size n2=0.202 Post hoc tests revealed that both groups who took the stand-alone histology examination performed significantly better than those who took the integrated examination (Table 4).

Student Survey

The response rate for the survey was 70% (40/57). When students were asked for feedback on histology being assessed as part of a stand-alone versus integrated examination, two themes emerged. The first had to do with study habits. Of the twenty-nine students who specifically commented on their study routine, 69% indicated that having histology as part of an integrated examination forced them to review the content more frequently, whereas the remaining 31% indicated that their study habits did not change. The second theme that emerged was related to overall preference. In this case, 86% of the twenty-one students who commented on this theme indicated they preferred when histology was tested on integrated examinations. Those who preferred the integrated format often commented that it forced them to continuously review the material, and while causing more work initially, it was less stressful than “cramming” neglected material shortly before a stand-alone examination. Those who preferred stand-alone examinations commented that this format allowed them to better organize their time and focus on histology content when preparing for an examination that assessed the material all at once.

Discussion

This study explored outcomes related to testing histology on integrated versus stand-alone examination formats. Results indicate that overall, student performance on histology questions was higher when they were delivered on a stand-alone examination. Investigation by cohort revealed temporal trends in the GER course that show an improvement over time in the average scores of histology questions on integrated examinations. Another aspect considered in this study was whether the change in testing format of histology impacted students differently based on their academic class rank. While significant differences were identified among both the upper and lower 50% of the class, the decrease in performance was larger among the lower 50% after histology was included on integrated examinations. This finding was particularly evident in the MSK course, where students in the lower 50% of the class performed 14.5% lower on integrated examinations compared to 6.7% lower among those in the upper 50% of the class.

On the surface, these results appear to suggest that testing histology in a stand-alone format produces better learning outcomes. However, academic performance is only one aspect of a much larger picture. Survey data collected from this study show that students overwhelmingly prefer when histology is included on integrated examinations. Numerous students commented how testing histology on integrated examinations forced them to revisit the material on a regular basis, which, pedagogically, has benefits from the standpoint of long-term retention [16, 17]. Assessment integration may also influence student preparation strategies, subsequently impacting performance [21]. Beyond material retention, an integrated approach has also been shown to facilitate the transfer of conceptual knowledge to clinical applications when assimilated in a longitudinal manner [22, 23].

When contextualizing the performance outcomes presented in this study, it is important to consider the larger picture of whether the observed differences translate to large knowledge gaps. To keep the comparisons as consistent as possible, only examination questions that were identical and asked every year were included in this study. This resulted in a relatively small number of overall questions that were investigated. This point is critical for interpreting the results because it means that a statistically significant difference of a 4.4% lower average performance on integrated examinations in GER translates to less than a single examination question. While the difference is slightly more notable in MSK, it begs the question of whether the slight drop in immediately observable performance might be outweighed by potential long-term benefits. Most agree that student study habits are driven by assessment [24,25,26,27,28]. Thus, including histology as part of an integrated examination encourages students to integrate material, which has been shown to have a positive influence on learning [29].

Comparing the results of this study to previous findings is challenging since there are few examples in the literature that evaluate the impact that shifts from stand-alone to integrated testing have on student performance. Thompson et al. [30] found that student performance decreased when discipline-integrated assessments were introduced among second-year medical students. However, further inquiry revealed that the integrated examinations tested a greater proportion of higher-order concepts, which likely explained the difference in performance. The current study avoided confounding factors such as this by only including identical questions. Hudson and Tonkin [31] investigated the impact of a move from discipline-specific to integrated examinations and found overall positive outcomes. However, the authors focused on examination validity rather than student performance. In a recent paper, a decrease in the amount of time dedicated to histology and a reduction in the number of histology questions included on assessments was shown to negatively impact student performance, particularly among students in the lower quartile of the class [32]. The authors offer two possible explanations for their overall findings: students spent less time studying histology, and students were no longer informed of their histology-specific performance on integrated examinations, which may have made it difficult to self-identify areas in need of improvement. The latter point is relevant to this study since students went from receiving a separate histology practical examination score to a histology score that was a relatively small component of a larger examination. While students were provided with a score report that included a breakdown of histology question performance, it is unclear whether they use this information to modify their study strategies.

This study focused on two different courses, one of which takes place in the first year of the curriculum (MSK) whereas the other course (GER) is in the second year. While the results from these courses were not distinctly different, the decrease in performance was more pronounced in the MSK course. When comparing the performance of first-year to second-year medical students, it is important to note that educational experience may impact student preparation and outcomes on the integrated examinations. For example, the level of a student’s previous exposure to coursework can have a substantial influence on the understanding of the material and assessment outcomes [33, 34]. Therefore, second-year students, having significant experience with medical school course materials and a greater appreciation for integrated assessments, may have an increased ability to perform well on integrated examinations when compared to their first-year counterparts. Another possibility that could explain the improvement in scores over time in the GER course is updated study materials. To improve the subsequent cohorts’ academic performance, students often create and “pass down” study resources. It is possible that changes in curricular content or assessment strategies might not be immediately reflected in these materials. As a result, students who rely heavily on these resources may not allocate enough time studying content that is different from the previous year. Additional years of data from the MSK course could provide further insight into whether this could explain the temporal trends observed in the GER course.

A finding from the present study that is difficult to explain is why students in the lower 50% of the class were differentially impacted by the histology assessment strategy change. Previous studies have suggested several potential factors that differentiate lower- and higher-performing learners, including differences in anxiety, test-taking strategies, study strategies, and motivation [35,36,37,38]. However, these only serve as general explanations for potential sources of the difference. Why the gap between upper- and lower-performing students would increase with integrated histology examinations requires further inquiry. At present, the most logical explanation is that students who are struggling to keep up with course material choose to place less emphasis on studying histology when it is on integrated examinations because it represents only a small portion of a larger examination. Student survey comments lend some support to this possibility, as two different students specifically stated that studying histology was not a priority when tested on an integrated examination because there was so much additional material they needed to learn. A more targeted study is needed to further explore this theory.

Limitations

Despite attempts to create a controlled study design, there are a number of factors that limit the interpretation of the findings presented in this study. First, this study focused on two courses that are part of a much larger curriculum. While this was done because these courses provided the most direct comparison, any other curricular changes that occurred over the study period could not be accounted for in the results. This includes external factors, such as the COVID-19 pandemic and its influence on the delivery of the curriculum. Since histology was assigned as self-study modules over the entire study period, COVID-19 likely had less of an effect than other areas. Nonetheless, there are implications related to this that extend beyond the delivery of material that could have impacted the results. Another issue is the relatively small number of questions included in this study. Previous research suggests that 25–30 examination items are needed to reach adequately reliability [39]. Thus, it is not certain how well the identical items selected for use in this study represent overall histology knowledge. In addition, the study period for the MSK course was short, with only a single year of integrated examination data that were available. Additional data points will help further evaluate the trends observed in this study. Lastly, survey data was retrospective in the sense that students were asked to compare experiences that occurred months apart. It is possible that students did not accurately remember, for example, the way they studied months prior for a stand-alone compared to an integrated examination.

Conclusion

Evaluating the impact that changes in assessment strategy have on student performance is a critical step in quality control and improvement. Results from the present study indicate that student performance on histology questions declined when they were included as part of integrated examination. Trends observed in the longitudinal data in the GER course showed that the initial drop in student performance was followed by improvement that resulted in averages that reached levels similar to previous standards within 2 years. While the reason for the initial decline in scores is not entirely clear, several factors are suggested, including shifts in the amount of time devoted to studying histology as well as a transitional period where student-generated study resources do not align with curricular/assessment changes. Although the shift to integrated histology examinations did not produce evidence of benefits from a student performance standpoint, the change was viewed as largely positive from a student perspective and eliminating the stand-alone examination provided both students and course directors more flexibility in scheduling. It is hoped that this study will provide insight for those at other institutions who are considering a similar change in assessment strategies.