Medical Science Educator

, Volume 26, Issue 1, pp 111–116 | Cite as

Enhancing Knowledge Retention of Cardiovascular Physiology Using Simulation

  • Maureen Hall
  • Maria Sheakley
  • Diana Callender
  • David Pederson
  • Gregory E. Gilbert
  • Kim Leighton
Open Access
Original Research

Abstract

Background

Basic science education plays an integral role in preparing medical students to be competent physicians and lifelong learners. Faculty in the preclinical years of undergraduate medical education, formerly focused on the transmission of biomedical principles and factual information, are now presenting concepts using clinical activities, with emphasis on clinical relevance, while advancing active learning and critical thinking.

Purpose

The purpose of this study was to evaluate the outcomes of a simulation intervention on short-term student knowledge gain. We investigated whether integrating simulation using the Harvey after completing lectures on cardiovascular physiology in a basic science course led to significant knowledge increases in first year medical students.

Methods

Wilcoxon rank sum (Mann-Whitney U) tests were used to test for significant differences in students receiving a lecture-only curriculum and students receiving a lecture-plus-simulation curriculum. Since this is an educational intervention, an a priori alpha level of 0.10 was specified. A two-sided test was used to analyze for differences between the curricula. All analyses were done using R software.

Results

There were statistically significant differences between outcomes associated with teaching modalities for the summative course exam, comprehensive final exam, and the cumulative average of these two exams (p value = .0006, <.0001, and .0980, respectively). Students exposed to simulation plus lecture performed better on the summative exam, but not the final exam.

Conclusions

The use of simulation and cardiac physiology lectures for first year medical students was found to have a significant impact on students’ cardiac physiology exam scores in the short-term. A longitudinal study is needed to see if there is long-term knowledge retention and improvement in clinical skills.

Keywords

Manikins High-fidelity simulation Education, medical, undergraduate Teaching and Learning Cardiovascular physiology 

Introduction

Basic science education has been a staple in medical education as students learn how to provide patient care. Retention of knowledge learned in these foundational courses declines over time [1]; however, it is necessary for knowledge to be retained beyond course assessments and throughout the preclinical and clinical years. Educational technologies may be able to aid in enhancing knowledge retention. Simulated clinical activities offer an endless variety of learning experiences for undergraduate medical students, which can aid in improving knowledge retention to bridge the gap between basic science education and clinical skills. The purpose of this study was to evaluate the outcomes of a simulation intervention on short-term student knowledge gain in first year medical students.

Literature Review

Retention of basic science knowledge during and beyond medical school has been a known challenge for well over a century, with the historical literature documented in a review by Custers [2]. He reports that if knowledge is to be retained, educators should review studies that show how to enhance long-term retention and include those strategies, such as creating repetitious practice sessions. Another strategy is to ensure students understand the relevance of what they are learning.

Chamberlain et al. [3] studied the impact of using case presentations to underscore the relevance of learning content from two basic science courses that would be needed later in clinical practice. Exam scores indicating knowledge retention were significantly higher for the students in the case-based group than those in the traditional lecture-laboratory setting. In a study to determine if medical students’ perceived relevance of biomedical sciences, demonstrated in clinical situations, would impact retention of basic science knowledge, Malau-Aduli et al. [4] found that higher perception of subject relevance correlated positively with increased retention of knowledge. They also suggest that using teaching strategies that increase awareness of clinical relevance will improve retention of knowledge.

Many preclinical medical educators, who formerly focused on verbal transmission of biomedical principles and factual information, are now presenting specific concepts with emphasis on clinical relevance [5]. Organ systems-based curricula increase integration and long-term retention of knowledge by presenting information organized around specific organ systems [2]. Following a longitudinal retrospective review after a curriculum revision, one study of medical students found that performance on the preclinical subject board examination, United States Medical Licensing Exam (USMLE) Step 1, and United States Medical Licensing Exam Step 2 was significantly better for those taught with the integrated organ systems curriculum than those taught with the traditional curriculum [6].

High-fidelity patient simulation is a teaching strategy that can combine case-based learning, developed within organ system-based curriculum, with a simulated clinical environment, allowing students to learn in context, thereby increasing the potential for students to see the relevancy of the material learned in the classroom that is now being applied to patient care experiences, and in a laboratory that can allow for repetition. The use of simulation, as a teaching strategy, offers an interactive means for organizing basic science knowledge around systems and disease processes, which may increase the perceived relevance of basic science education and thus increase knowledge retention.

The faculty developed a unique approach to teaching cardiovascular physiology using the Harvey® cardiovascular simulator in a basic science course [7]. We investigated whether the inclusion of a simulation event designed to allow first year medical students to apply basic science information learned in didactic lectures led to significant knowledge increases as assessed by course exam scores. With the growth of simulation use in medical education, it is important to investigate the impact of using simulation in undergraduate medical education and its role in a student’s ability to integrate, understand, and retain knowledge.

Methods and Materials

This investigation was approved by the Institutional Review Board of a large offshore United States (US) medical school where this study was conducted and subscribed to the tenets of the Declaration of Helsinki. Voluntary completion and return of the survey served as the informed consent to participate in the medical simulation research study. This study used historical controls for a comparison as curricular change had taken place. The medical school program admits three new classes of students per year (January, May, and September), which allows for the large sample sizes in this study. The control group that was chosen was the most recent group of students receiving the traditional curriculum (lecture only).

Sample

Two groups of students participated in this investigation. The first group was a historical control of 515 students from the 2009 matriculating class at the offshore US medical school. The group receiving the intervention was a group of 1066 students from the 2011 matriculating class at the medical school program at which the study took place. Twelve hours of cardiovascular physiology lectures were presented over 2 weeks, four of which were directly related to the simulation experience in this study. The day after the lectures ended, the students were divided into groups of eight to participate in a 1-hour (h) supplemental simulation exercise.

Instruction

Usual Instruction Method

Four hours of didactic physiology lectures directly related to the simulation were delivered during the first year cardiovascular course. The topics covered were the cardiac cycle (1 h), pressure-volume loops (1 h), and the physiological basis of heart sounds and murmurs (2 h). The content delivered was standard for any medical physiology course, based on the learning objectives provided by the American Physiological Society (APS) for these topics. The September 2009 and 2011 cohorts received the same lectures, delivered by the same faculty member, covering the same content.

Intervention

In addition to the didactic physiology lectures, the intervention group participated in a simulation exercise designed to reinforce the learning objectives of the didactic lectures. The Harvey Cardiovascular Simulator was used for simulation activities [7], and students completed clinical tasks including (1) stethoscope placement in the proper anatomical positions on the chest; (2) auscultation of first, second, third, and fourth heart sounds; and (3) auscultation of split heart sounds, with and without respiratory changes.

Clinically trained simulation facilitators attended the cardiovascular physiology lectures or reviewed lectures via video streaming to ensure a common understanding of what the students learned in lecture. Facilitators were given handouts that outlined the simulation objectives, content, and the desired instructional technique. A simulation training session was developed where facilitators discussed the objectives and content and a standardized approach to facilitation was demonstrated. After each cohort completed the simulation, there was a facilitator debrief where some videos of the simulations were reviewed and facilitators given feedback. This iterative process was repeated after each cohort completed the simulation.

Independent Variables

The independent variable assessed in this investigation was curriculum type—lecture-only curriculum versus lecture plus simulation.

Dependent Variables

Assessment of the cardiovascular course learning outcomes occurred via two examinations. The assessment questions were written in USMLE style and were identical for both cohorts. The first exam was a summative course exam at the end of the cardiovascular course; the second was a comprehensive final exam at the end of the semester that covered three consecutive courses. There were four assessment questions on the summative course exam and six on the comprehensive final exam that were used in this study, all directly mapped to the shared learning objectives for the physiology lectures and simulation exercise. The questions on the summative course exam and comprehensive final exam were different, but similar in content and difficulty. Scores were entered as percent correct. The cumulative average is the arithmetic average of the summative course exam score and comprehensive final score. The assessment was used to track differences in performance between the intervention group and the historical control group. The validity and reliability of the questions (and instruments) has not been assessed. Tests were administered through and data collected using Medical Education Technology, Incorporated’s, LearningSpace (METI LearningSpace®, now known as Canadian Aviation Electronics (CAE), [Limited] Healthcare’s LearningSpace) [8] a web-based clinical learning management system.

Analysis

Data were entered into a Microsoft Excel® [9] spreadsheet and saved into a comma-separated value (CSV) file. A frequency distribution of the teaching modality was calculated. Summative course exam grades, comprehensive final exam grades, and cumulative averages were tested for normality using normal probability plots and the Anderson-Darling, Shapiro-Francia, and the Shapiro-Wilk normality tests [10, 11, 12]. The Anderson-Darling test is the recommended empirical distribution function test by Stephens compared to other tests of normality giving more weight to the tails of the distribution than the Cramer-von Mises test [13]. The Shapiro-Francia test was chosen because of its known performance and the Shapiro-Wilk test was chosen because it is one of the best-known tests for normality [12]. An a priori alpha level for the goodness-of-fit tests was specified to be .10 [14]. Data were not normally distributed and thus the Wilcoxon rank sum (Mann-Whitney U) tests were used for significance testing. Since this is an educational intervention, an a priori alpha level of .10 was specified [15]. A two-sided test was used to test for differences in exam scores between curricula. All analyses were done using R software [16, 17].

Results

Five hundred fifteen (33 %) students participated in a lecture-only curriculum and 1066 (67 %) students participated in a curriculum augmented with simulation.

All exam scores (three) demonstrated marked departure from normality graphically and by all three tests of normality; therefore, hypothesis testing was done using the Wilcoxon rank sum (Mann-Whitney U) test. The central tendency and measures of dispersion, as well as p values, and associated confidence intervals can be seen in Table 1.
Table 1

Central tendency, p values, and 90 % confidence intervals for the means of three exam scores

 

Mean (standard error of mean)

Median (interquartile range)

90 % confidence interval

Construct

Lecture

Simulation plus lecture

Lecture

Simulation plus lecture

p value

Lecture

Simulation plus lecture

Summative course exam

66.2 (1.19)

71.5 (.75)

75 (50)

75 (50)

.0006

64.2, 68.2

70.3, 73.9

Comprehensive final exam

92.4 (.79)

85.2 (.72)

100 (0)

100 (33)

<.0001

91.1, 93.7

84.0, 87.6

Cumulative average of exams

79.4 (.76)

78.4 (.54)

88 (25)

83 (25)

.0980

78.1, 80.7

77.5, 80.6

There were statistically significant differences between the knowledge outcomes for each teaching modalities for the summative course exam and the comprehensive final exam (p value = .0006 and <.0001, respectively). Not surprisingly, the average of these two exams, labeled “cumulative average of exam,” were also significantly different (p value = .0980). Mean scores were also substantially different in the cases of summative course exam scores and comprehensive final exam scores differing by over five points. However, these means were not consistently higher or lower. Summative course exam scores were higher for the simulation-plus-lecture group and comprehensive final exam scores were higher for the lecture-only group. Median scores for both exams were the same for both groups, 75 and 100 for the summative course exam and comprehensive final exam, respectively. There was no practical difference in averages of the two exams between the groups (one percentage point); however, there was a statistical difference owing to the large sample size.

In conclusion, mean summative course exam scores and mean comprehensive final exam scores were significantly and meaningfully different, but median scores were identical. In the case of the average of the two scores, mean scores were almost identical, but there was a five percentage point difference in median exam scores favoring the lecture curriculum.

Discussions

Results of this investigation demonstrated that first year medical students in a basic science curriculum augmented with simulation scored significantly better than historical controls on the summative course exam. According to Bowe et al. [5], preclinical educators now present concepts within a clinical relevance framework, which advances active learning, critical thinking, and other professional competencies rather than using the historical method of transmitting biomedical principles and facts via lecture. This investigation tested and reported the results of using simulation as a teaching strategy; our data analyzes the hypothesis that first year medical students learned cardiovascular physiology concepts better when lectures were paired with simulation when tested early in the term. Students also gained clinical skills earlier in the curriculum, as traditionally, the Harvey simulation was not introduced until the fourth semester midway through the second preclinical year at this medical school. One important benefit of using the Harvey with first semester students is that students were introduced to clinical skills in the first semester and provided clinical context of basic science education immediately.

Pairing simulation with cardiovascular physiology lectures may allow students to apply basic science knowledge earlier and within clinical context, and the corresponding relationship may narrow the retention gap traditionally seen in undergraduate medical education. A systemic review by Issenberg et al. [18] concluded that simulation-based education was effective and complements medical education. Increasing short-term retention, improving the retention interval, and transferring knowledge to practice is challenging because it takes more time to offer hands-on learning opportunities than it does to lecture. McGaghie et al. [19] call for integration throughout the curriculum to solve this challenge.

Students participated in the Harvey simulation sessions as first year students and again as second year students, doubling their simulation exposure, thus having had more opportunity to enhance their basic science knowledge and gain additional clinical skills. Simulation coupled with lectures in cardiovascular physiology allows students to learn through experience, and according to the course evaluations, it is valued highly by students. Using this paradigm, new knowledge is acquired in a clinical context, aiding subsequent retrieval and reinforcing current knowledge [1].

It was surprising to discover that there was not a significant difference in the comprehensive final exam grades between the two study groups. There may be several reasons for this lack of difference. One explanation could be attributed to the differences in intervals between time of exposure to the intervention and taking the summative course exam and the comprehensive final exam. The summative course exam took place approximately 2 weeks following the intervention while the comprehensive final took place a month later. During these periods, students did not have scheduled compulsory small group practice sessions with the Harvey that would allow for integrated review of the materials at least 1 week before their comprehensive final exam. Furthermore, the final exam was comprehensive; students are more likely to be overwhelmed when taking the comprehensive final since they are expected to study a larger volume of material involving multiple disciplines. In addition, they become anxious since the comprehensive final is a high-stakes exam accounting for a greater percentage weighting of their cumulative average. Students who fail the comprehensive final exam are at risk of having to remediate the entire program which may impose emotional distress due to adding four additional months to the time of the anticipated conferral of their doctorate of medicine degree, further contributing to their financial burden.

As with all studies, this study has its limitations. We did not assess the learner’s retention of the material in a longitudinal fashion. Additionally, no a priori power calculations were done; one could argue an adequate sample size was achieved. However, because of the larger sample size, a much smaller effect size was detectable, perhaps one not of importance. However, there were also meaningful differences between cardiovascular physiology scores of the historical cohort (78 % average) and the lecture-plus-simulation cohort (85 % average), which would result in the difference of a full letter grade for the student. Additionally, only facilitators and lecturers directly affiliated with the medical school were included in the study, and therefore, generalization must be made with caution. Finally, the simulation experience was facilitated by several different faculty members, which introduces variability in facilitation. To limit this variability, the facilitators underwent training of the simulation exercise together.

Future directions for investigation into knowledge gain would be to enhance participation. One possible approach would be to decrease the group size to better actively engage students in the learning activity. Another logical extension of this investigation would be to study longer term retention employing a longitudinal study within the term and in students’ second year of medical school using an experimental design approach. Finally, the number of questions should be increased on the assessment.

The findings in this study correlate to the historical literature [2] that established retention of content to be a challenge. While it was hypothesized that enhancing lecture with simulation would improve retention, in this study it did not. Placing the learning within context of a simulated environment using high-fidelity mannequins was found to improve scores initially similar to the findings of Chamberlain et al. [3], but not over time which further supports the need for repetition to lead to mastery. The findings of this study promote the need to ensure that opportunities to repeat active learning opportunities, such as those conducted in simulation environments, are made available and promoted for student practice.

Conclusions

The combination of simulation with the cardiac physiology lectures for first year medical students was found to have a significant impact on students’ cardiac physiology exam scores in the short-term. Offering additional practice sessions with the Harvey, throughout the term, could enhance knowledge retention of the information as repetition has been shown to positively impact retention. A longitudinal study is needed to see if there is long-term knowledge retention and improvement in clinical skills. Further studies are needed to replicate these findings in our student population regarding the magnitude of knowledge retention. Overall, the intervention was successful other than the limitations based on the small number of exam questions used to track students’ knowledge and the inherent variations among the trained facilitators. Thus, the quality of the outcomes depends primarily on faculty development and consistencies among facilitators.

Notes

Compliance with Ethical Standards

This investigation was approved by the Institutional Review Board of a large offshore US medical school where this study was conducted and subscribed to the tenets of the Declaration of Helsinki. Voluntary completion and return of the survey served as the informed consent to participate in the medical simulation research study.

References

  1. 1.
    Lindsey RV, Shroyer JD, Pashler H, Mozer MC. Improving students’ long-term knowledge retention through personalized review. Psychol Sci. 2014;25:639. doi:10.1177/0956797613504302.CrossRefGoogle Scholar
  2. 2.
    Custers EJFM. Long-term retention of basic science knowledge: a review study. Adv Health Sci Educ. 2010;15(1):109–28. doi:10.1007/s10459-008-9101-y.CrossRefGoogle Scholar
  3. 3.
    Chamberlain NR, Stuart MK, Singh VK, Sargentini NJ. Utilization of case presentations in medical microbiology to enhance relevance of basic science for medical students. Med Educ Online. 2012;17:15943. doi:10.3402/meo.v17lQ15943.CrossRefGoogle Scholar
  4. 4.
    Malau-Aduli BS, Lee AYS, Cooling N, Catchpole M, Jose M, Turner R. Retention of knowledge and perceived relevance of basic sciences in an integrated case-based learning (CBL) curriculum. BMC Med Educ. 2013;13:139. doi:10.1186/1472-6920-13-139.CrossRefGoogle Scholar
  5. 5.
    Bowe CM, Voss J, Thomas Aretz H. Case method teaching: an effective approach to integrate the basic and clinical sciences in the preclinical medical curriculum. Med Teach. 2009;31(9):834–41.CrossRefGoogle Scholar
  6. 6.
    Williams PC, Epps AC, McCammon S. The strategic impact of a changing curriculum and learning environment on medical students’ academic performance. J Natl Med Assoc. 2011;103(9-10):802–10.CrossRefGoogle Scholar
  7. 7.
    The “Harvey®” The Cardiopulmonary Patient Simulator. Covered by U.S. Patent No. 7,316,568.Google Scholar
  8. 8.
    Microsoft Corporation. Microsoft excel® for Mac® 2011. 2010.Google Scholar
  9. 9.
    Medical Education Technologies, Incorporated, Learning Space ®, version 11.2010–2011.Google Scholar
  10. 10.
    Anderson T, Darling DA. A test of goodness-of-fit. J Am Stat Assoc. 1954;49(268):765–9. doi:10.1080/01621459.1954.10501232.CrossRefGoogle Scholar
  11. 11.
    Shapiro SS, Francia RS. An approximate analysis of variance test for normality. J Am Stat Assoc. 1972;67(337):215–6. doi:10.1080/01621459.1972.10481232.CrossRefGoogle Scholar
  12. 12.
    Shapiro SS, Wilk MB. An analysis of variance test for normality (complete samples). Biometrika. 1965;52(3-4):591–611. doi:10.1093/biomet/52.3-4.591.CrossRefGoogle Scholar
  13. 13.
    Stephens M. Tests based on EDF statistics. In: D’Agostino RB, Stephens M, editors. Goodness-of-fit techniques. New York: Marcel Dekker; 1986.Google Scholar
  14. 14.
    Yap BW, Sim CH. Comparisons of various types of normality tests. J Stat Comput Simul. 2011;81(12):2141–55. doi:10.1080/00949655.2010.520163.CrossRefGoogle Scholar
  15. 15.
    Welke TM, LeBlanc VR, Savoldelli GL, et al. Personalized oral debriefing versus standardized multimedia instruction after patient crisis simulation. Anesth Analg. 2009;109(1):183–9. doi:10.1213/ane.0b013e3181a324ab.CrossRefGoogle Scholar
  16. 16.
    R Core Team. R: a language and environment for statistical computing. 2014.Google Scholar
  17. 17.
    Gross J, Ligges U. Nortest: tests for normality. 2012.Google Scholar
  18. 18.
    Issenberg SB, McGaghie W, Petrusa E, Lee Gordon D, Scalese R. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005;27(1):10–28. doi:10.1080/01421590500046924.CrossRefGoogle Scholar
  19. 19.
    McGaghie W, Issenberg S, Petrusa E, Scalese R. A critical review of simulation-based medical education research: 2003–2009. Med Educ. 2010;44(1):50–63. doi:10.1111/j.1365-2923.2009.03547.x.CrossRefGoogle Scholar

Copyright information

© The Author(s) 2015

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.

Authors and Affiliations

  • Maureen Hall
    • 1
  • Maria Sheakley
    • 2
  • Diana Callender
    • 3
  • David Pederson
    • 4
  • Gregory E. Gilbert
    • 1
    • 4
  • Kim Leighton
    • 4
  1. 1.Center for Teaching and LearningRoss University School of MedicineRoseauCommonwealth of Dominica, West Indies
  2. 2.Western Michigan University Homer Stryker M.D. School of MedicineKalamazooUSA
  3. 3.The Commonwealth Medical CollegeScrantonUSA
  4. 4.DeVry Medical InternationalIselinUSA

Personalised recommendations