Abstract
Purpose
Debriefing is necessary for effective simulation education. The PEARLS (Promoting Excellence and Reflective Learning in Simulations) is a scripted debriefing model that incorporates debriefing best practices. It was hypothesized that student simulation performance might impact facilitator adherence to the PEARLS debriefing model. There are no published findings on the effect of student performance on debriefer behavior.
Methods
Third-year medical students participated in a video-recorded, formative simulation to treat a high-fidelity mannequin for an asthma exacerbation. A faculty debriefer trained in the PEARLS model evaluated student performance with a standardized rubric and conducted a recorded debriefing. Debriefing recordings were analyzed for debriefer adherence to the PEARLS model. Debriefers were assigned a debriefing score (DS) from 0 to 13; 13 was perfect adherence to the model. Definitive intervention (DI) for asthma exacerbation was defined as bronchodilator therapy. Critical actions were as follows: a focused history, heart/lung exam, giving oxygen, and giving a bronchodilator.
Results
Mean DS for the debriefers of students who provided DI was 8.57; 9.14 for those students who did not (P = 0.25). Mean DS for debriefers of students who completed all critical actions was 8.68; 8.52 for those students who did not (P = 0.62). Analysis of elapsed time to DI showed no relationship between the time DI was provided and DS.
Conclusions
Student performance had no impact on debriefer performance, suggesting the PEARLS model is an effective aid for debriefers, regardless of learner performance. These findings suggest student performance may not bias facilitators’ ability to conduct quality debriefings.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
Introduction
Medical simulation is an application of experiential learning [1, 2]. Reflective practices, such as debriefing, are essential to the success of the experiential learning cycle [3]. In contrast to feedback, debriefing is bidirectional, interactive, specific, and focuses on reflection-on-action [4,5,6,7].
Existing literature suggests debriefing is a necessary component for learning in medical simulation. One randomized control trial of simulated airway management found that only simulation in combination with debriefing, maintained the skills learned during the course [8]. A cardiology review course using simulation with debriefing showed an improvement in clinical skills when compared to a historical comparison group who did not receive debriefing [9]. A meta-analysis of debriefing showed a 25% improvement in performance with debriefing [10]. A systematic review of medical simulation showed that debriefing was the single most important component to effective learning [11].
Debriefing should focus on learner development as opposed to assessment. A focus on assessment may impede learner development by making the learner feel defensive and less open to change, while also biasing debriefer behaviors [12]. In fact, studies of the impression-perseverance effect show that even when an observer knows that good or bad performance is a false condition, this biases the observer’s perception of competence [13, 14]. This presents a conundrum for a would-be debriefer. In order to provide effective debriefing, the debriefer must assess how well the learner performed the task(s), so as to focus the debriefing on ways for the learner to improve. Yet, too much focus on assessment may impede the learner’s receptiveness, as well as bias the debriefer. It was hypothesized that good or poor learner performance may exacerbate this problem by influencing debriefer behavior. At the time of this study, there was no known literature specifically addressing the impact of learner performance on debriefer behavior. Thus, it was hypothesized that good or poor student performance might improve or worsen debriefing behavior, or vice versa (i.e., no directional relationship was hypothesized due to the exploratory nature of the study).
Since debriefing is a skill that must be taught [15], the answer to this question has implications at both the individual facilitator and programmatic level. If good and/or poor student performance worsens debriefing quality, this would negatively impact learning outcomes [8,9,10]. Though this begs the questions: what constitutes a “good” debriefing, how does an individual debriefer achieve this, and how can it be scaled across an organization? While there are faculty development programs available [16, 17], there is no national standard for faculty training in medical simulation or debriefing, despite ample literature supporting specific debriefing best practices [8,9,10,11, 18,19,20,21,22,23,24,25,26,27,28,29,30,31,32].
The Promoting Excellence and Reflective Learning in Simulations (PEARLS) debriefing model is a scripted approach to simulation debriefing which offers sample phrases for each stage of the debriefing to assist facilitators in incorporating debriefing best practices [18,19,20,21,22]. In a randomized trial of scripted vs non-scripted debriefing in Pediatric Advanced Life Support (PALS), scripting increased student knowledge acquisition and team leader performance [21]. PEARLS scripting improves model accessibility for those debriefers still developing their debriefing skills [18]. PEARLS circumvents some limitations of other debriefing models, including the prioritization of expediency and a requirement for high levels of experience for successful application [20, 23,24,25]. The PEARLS model presents a sequence of four stages: (1) reactions, (2) description, (3) analysis, and (4) summary [18]. These stages and their associated sample phrases assist debriefers in setting the stage for the debriefing, organizing the debriefing, and helping debriefers pose questions [18]. The PEARLS approach combines the debriefing strategies of learner self-assessment [20, 24, 26], facilitated reflection and understanding [27,28,29], directed performance feedback [30, 31], and focused teaching [28, 32]. PEARLS contains these blended strategies within a consolidated model [18]. This gives debriefers several clear and actionable strategies to employ as they facilitate learning. The PEARLS framework also assists the debriefer in choosing the best suited educational approach(es) for their learners during the analysis phase [18].
The Medical College of Georgia (MCG) Educational Simulation Program provides support to MCG faculty in the design, development, and implementation of high-quality formative and summative educational simulation activities. The program also runs an annual mid-clerkship simulation in which approximately 200 medical students participate in an individual, formative simulation activity. In order to facilitate the training and development of debriefing best practices for the facilitators, the PEARLS model was selected after a needs assessment of the Educational Simulation Program at the medical school and an analysis of available debriefing models. MCG faculty had very little or no debriefing experience with the PEARLS model prior to its adoption. To prepare for the event, faculty facilitators were required to participate in simulation and debriefing training. Facilitators were first provided with asynchronous training materials, which included video-recorded examples of the asthma exacerbation case and PEARLS debriefing. Following the completion of the asynchronous training, facilitators attended individual and small-group in-person training in the simulation center, where they could walk through the case and debriefing model with the director of the simulation program and get hands-on experience with the simulation technology.
Given the variable simulation and debriefing experiences of the facilitators, the PEARLS model was selected by the program for the goal of improving consistency in facilitator debriefing practices, thereby providing more consistent learner experiences. This study evaluated the effects of student performance on debriefer behavior during the mid-clerkship simulation activity. For this study, PEARLS best practices were conceptualized as a proxy for debriefing quality. This study was conducted in order to better understand the relationship between student simulation performance and debriefer behaviors, as this relationship will have direct implications for the continued development of faculty debriefers [8,9,10].
Materials and Methods
Simulation Scenario Overview
The simulation case was a formative, non-graded, individual activity that was designed as an opportunity for student learning and reflection. This study was considered expedited by the institutional review board of Augusta University. All students were collectively oriented to the activity, briefed on the scenario parameters, and given an orientation to the simulation room and its capabilities. Each student then rotated through the same case: a patient exposed to outdoor smoke that triggered an asthma exacerbation. The patient had a history of asthma, presented in obvious respiratory distress with diffuse wheezes on exam in both lungs, and vital signs showing tachypnea and hypoxia. The student was instructed to play the role of the physician, evaluating the patient, ordering diagnostic tests, managing the asthma exacerbation, and then calling the patient’s primary care physician to arrange a disposition, after which the case was ended.
Simulation Room Overview
The simulations were conducted in one of five dedicated simulation rooms designed to simulate an emergency department room. The patient was a high-fidelity mannequin with voice capability under manual control by the scenario facilitator. All students used the same kind of mannequin. Real-time vital signs were on display at all times. Results of tests were given as handouts to the student by a trained actor who played the role of the nurse in the room. Equipment and supplies available to the student included intravenous (IV) catheters and fluids, nasal cannula and oxygen mask, nebulizer to mimic albuterol, medications for IV or oral administration (i.e., prednisone). Consultations were made via a phone located in the simulation room.
Simulation Scenario Facilitator and Actor
The scenario was facilitated from a control room by a trained faculty member who acted as the patient and the consultant. The control room was divided by a one-way glass or was a room with a monitor that projected the simulation room. A computer controlled the mannequin’s vital signs and lung sounds regardless of location. A push to talk microphone allowed the facilitator to voice the patient. A phone allowed the facilitator to act as the family physician who calls for a case presentation via phone. The nurse actor performed the student’s orders and provided the student with test results in the simulation room. Both the facilitators and nurse actors were trained on the scenario. The facilitators were also trained in the PEARLS debriefing model and given a PEARLS handout to follow during the debriefing. Each facilitator and nurse actor managed a single simulation encounter with a single student. After the scenario ended, the nurse actor stayed in the room for the next encounter, while the facilitator and student went to another room for debriefing.
A total of 32 faculty facilitators participated. Of these facilitators, five were non-physicians. These five facilitators were Doctors of Philosophy, one each in pharmacology, anesthesia, physiology, biochemistry, anatomy. Of the 27 physicians, the breakdown by specialty was: seven internal medicine, six pediatric emergency medicine, five emergency medicine, two surgery, two pediatrics, two family medicine, and three OB/GYN. The facilitators were of varied simulation experience. We assessed experience level of the facilitators using the Dreyfus model of skill acquisition [33]. The breakdown by competency was: 11 novice, nine advanced beginner, eight competent, two proficient, two expert.
Simulation Scenario Assessment
Although this was a formative experience without any formal evaluation, the facilitator completed a case evaluation form for each student as the scenario unfolded. Key components of the form were a start time, end time, and time at which the bronchodilator was first given. The specific medication(s) and dose(s) were also recorded. The form was also used to record the order in which each action was taken in order to determine how students prioritized their decision-making (e.g., if the student’s first action was to gather an HPI, that would be marked “1,” and if they then ordered a nebulizer treatment that would be marked “2”). If an action was not performed, it was not assigned a number.
Debriefing Assessment
A 15-min debriefing followed each 15-min simulation scenario. We chose a 1:1 time ratio for simulation to debriefing based on the importance of debriefing in the experiential learning cycle and the time allotted for the event [3, 8,9,10,11]. Both the scenario and debriefing were recorded. After developing a new, structured rubric to assess facilitator adherence to the PEARLS model, evidence of construct and content validity for the instrument was collected through a pilot test and expert review. Recordings of a prior simulation case debriefing were scored with the instrument as a pilot test, which also allowed for an estimate of reliability. Problematic items in the pilot test were revised for clarity. The revised instrument was then evaluated by experts using cognitive interviewing. Inter-rater reliability was established between all 32 facilitators with a correlation coefficient of 0.913. A primary researcher then observed each video and assessed how many PEARLS practices were performed during each debriefing. The instrument to assess adherence to the PEARLS model contained 13 observable facilitator behaviors. The debriefing scored one point for when the debriefer initiated the behavior and one-half point if the student initiated the behavior instead of the debriefer. A score of zero was assigned for no observable behavior. Thus, this analysis generated a debriefing score (DS) from 0 to 13 for each individual debriefing. Debriefing scores were then compared to the student evaluation forms in order to generate the results. The researcher that analyzed all of the videos was a medical student trained in the PEARLS debriefing model and use of the new PEARLS assessment rubric. The student was supervised by the last author as part of a student research program funded by the Medical College of Georgia.
Results
Debriefing Score
As noted above, each debriefer was assigned a DS for each debriefing they facilitated. Given that this score consisted of observed PEARLS best practices, the debriefing score was considered a proxy for both the debriefing quality and as an assessment of the debriefer’s behaviors; a higher score indicated both a higher quality debriefing and a greater adherence to the PEARLS debriefing model. This was based on the idea that each PEARLS component contributes to a better debriefing, and therefore, improves learning [18,19,20,21,22,23,24,25,26,27,28,29,30,31,32]. Table 1, PEARLS debriefing items, lists each item in the PEARLS debriefing model along with the respective number of times Facilitators Initiated (FI) and Students Initiated (SI) each debriefing item (out of a total of 187 debriefings). Table 1 also has a combined (C) column. This column adds together the FI and SI instances of each item, to give a total number of times each item was performed. Debriefers performed each item the majority of the time with the exception of item 4 (“Asked the student to summarize the case”), item 10 (“Used preview statements to introduce new topics”), and item 13 (“Clearly ended the debriefing”).
Definitive Intervention
For the purpose of this analysis, bronchodilator administration was defined as the definitive intervention (DI) for the scenario [34]. The DS for the debriefer for those students who gave a bronchodilator during the scenario was compared to the DS for the debriefer of those students who failed to give a bronchodilator. Table 2, comparison of learner performance to debriefing score, shows mean DS for the debriefer of those students who gave a bronchodilator compared to those that did not. The mean DS for the debriefer of those students who gave a bronchodilator was 8.57 (2.22) with n = 180 (96%). The mean DS for the debriefer of those students who did not give a bronchodilator was 9.14 (2.17) with n = 7 (4%). The p value between means was 0.25.
Critical Actions
For the purpose of this analysis, critical actions were defined as follows: (a) conduct HPI, (b) conduct heart exam, (c) conduct lung exam, (d) order bronchodilator treatment, or (e) order O2 [34]. The critical actions were not defined prior to the simulation exercise and were not communicated as such to students during debriefing. Making a disposition for the patient was not considered a critical action because a disposition discussion with the patient’s primary physician was scripted into the scenario in order to provide every student an opportunity to present the patient as part of the simulation. Table 2 shows mean facilitator DS for debriefers of those students who completed all five critical actions versus those students who failed to complete all critical actions. The mean DS for debriefers of those students who completed all critical actions was 8.68 (2.16) with n = 88 (47%). The mean DS for debriefers of those students who did not complete all critical actions was 8.52 (2.28) with n = 99 (53%). The p value between means was 0.62.
Time to Definitive Intervention
The time at which the student gave a bronchodilator was defined as the Time to Definitive Intervention (TTDI). A desired TTDI was not determined prior to the simulation and was not communicated to students during debriefing. The time intervals were recorded by debriefers at 1-min intervals. A time score of 0 indicates the student administered a bronchodilator at a time less than one minute after case start. Plotting TTDI and DS produces the scatter plot shown in Fig. 1—TTDI vs DS. There was no discernable relationship identified between the two variables. DS was further analyzed as a function of the time difference between the student’s TTDI and the mean. The mean TTDI was 4.11. The hypothesis was that it might not be absolute performance that influenced debriefer behavior, but deviation from mean performance that influenced debriefer behavior. When the time distance between mean TTDI was plotted against DS, it produced Fig. 2—time distance from mean TTDI vs DS. As with TTDI vs DS, this analysis showed no discernable relationship between the metric and DS.
Discussion
This study was conducted in order to better understand the relationship between student simulation performance and debriefer behaviors. To assess debriefer behaviors, a new, reliable instrument was created to measure debriefer adherence to the PEARLS debriefing model. Analysis was conducted on the number of PEARLS observable best practices that occurred during the debriefing, for which a DS was assigned. The DS was then compared to measures of student performance, including DI by the student, which was defined as giving a bronchodilator medication; how quickly the student provided DI, termed TTDI; and if the student completed all critical actions (i.e., conduct HPI, conduct heart exam, conduct lung exam, give oxygen, give DI). Overall, the data showed that student performance has no effect on DS, and by proxy, no impact on debriefing quality or debriefer adherence to the PEARLS model.
Bronchodilator administration was used as the DI because bronchodilator administration is the sole intervention that could reverse or ameliorate the simulated patient’s underlying pathology of an asthma exacerbation [34]. Given how crucial bronchodilator administration is for an asthma exacerbation, it was hypothesized that students failing to provide a bronchodilator would influence DS. Of all the available performance metrics, this provided the starkest contrast between ideal and suboptimal performance. Suboptimal student performance might concern the debriefer, leading them to deviate from the PEARLS model and engage in more didactic teaching behaviors. However, poor student performance could have also motivated the debriefer to rely more heavily on the sample phrases. No directional relationship was hypothesized due to the exploratory nature of the study. However, despite the initial hypothesis, no relationship was identified between DS and providing, or failing to provide, DI.
While bronchodilator administration was defined as the DI, four other critical actions were identified as necessary for appropriate diagnosis and treatment of an asthma exacerbation [34]. The data was split approximately evenly between those students who performed all critical actions (n = 88, 47%) and those who did not (n = 99, 53%). Comparison of these cohorts showed even less evidence of a difference in DS and also reversed the trend seen in the DI comparison, with a slightly higher mean DS for debriefers of the students who completed all critical actions.
It was also hypothesized that TTDI might influence debriefer model adherence. A small TTDI was conceptualized as a better performance, as a debriefer might feel concerned when having to wait for a student to figure out what to do (i.e., provide DI) and this reaction might influence debriefer performance. As with the analysis above, it was hypothesized that a delay in TTDI might improve or degrade debriefer adherence. However, comparison of TTDI to DS indicated no relationship. It was considered that it might not be better student performance, per se, that influenced the debriefer adherence to the PEARLS model; it might be deviation from average performance (i.e., either exceptionally high performance or particularly poor performance) that influenced debriefer behavior. This prompted analysis comparing time distance from mean TTDI to DS. However, once again, this comparison showed no relationship.
This study demonstrates that student performance does not influence debriefer performance when the debriefer is using the PEARLS debriefing model. These results can be generalized to other simulation programs. The student body at the Medical College of Georgia is diverse in race, gender, nationality, and socioeconomic status. The faculty debriefers were similarly diverse. They included members of several medical specialties, included practicing clinicians, and basic science researchers. Debriefers also had a broad range of simulation and debriefing experience. See Methods for a detailed breakdown.
Facilitators performed the majority of the PEARLS items the majority of the time. However, facilitators performed items 4, 10 and 13 (Table 1) at the lowest frequency. The action performed least frequently was item 4, “Ask the student to summarize the case.” The reason for this is unclear. The preceding item asked the student to discuss how they felt about the case. Depending on how the student responded to this question, the facilitator may have perceived the student as summarizing the case in their response to this request (item 3 was performed in 182 of 187 debriefings, 97%). Also, items 5–7 refer to student’s areas of strengths and areas for improvement. The student’s discussion of their feelings about the case, may have led naturally to a discussion of strengths and weaknesses, thereby causing the debriefing conversation to naturally flow past the opportunity for a case summary. Similarly, item 10, “Used preview statements to introduce new topics,” was rarely performed (34 of 187 debriefings, 18%). Again, the reason is unclear. It may be that preview statements are more formal than is typical of normal conversational flow. It may be that natural transitions often occurred during debriefings, seeming to obviate the need for preview statements, or facilitators were not practiced at utilizing this technique for changing subjects. Facilitators performed item 13, “Clearly ended the debriefing,” a little less than half the time (92 out of 187, 49%). The reason for this is unclear. It may be that the facilitator was rushed by the end the debriefing, and/or felt the debriefing had clearly ended based on non-verbal cues.
A limitation of this study is the lack of a control group. Future research could include data collection on the behaviors of debriefers trained in and using a different debriefing model, trained debriefers using no identified debriefing model, and untrained debriefers without an identified model. However, this study does indicate that when the PEARLS debriefing model is used, students at all performance levels received debriefings of similar quality. Because prior research showed that debriefing models improve student learning [8,9,10,11, 18,19,20,21,22,23,24,25,26,27,28,29,30,31,32], and that debriefing training is necessary to achieve these benefits [15,16,17], it is reasonable to conclude that using the PEARLS model may help debriefers provide quality debriefings to students at all performance levels. Another limitation in the data is that only seven of the students failed to provide DI. There was a non-significant trend towards improved DS for those students who failed to provide DI. A larger data set may identify a difference in DS for debriefings of students with suboptimal performance. However, the analysis of critical actions suggests this is unlikely, given that when there are similar n values between cohorts, the difference between cohorts narrowed and the trend towards improved DS for worse performance was reversed. A further limitation in this study is it analyzes only specific, objective performance measures. The measurement included binary actions recorded or not observed, the order of actions performed, and time when each action occurred. What this study did not include was a measurement of students’ interpersonal interactions with the patient, nurse, and consultant. These behaviors could affect debriefers overall gestalt of “good” vs “poor” performance. For example, a student could have a low TTDI and complete all critical actions while being brusque and uncaring towards the patient and rude towards the nursing staff and consultant. This would likely be perceived by the debriefer as “poor” performance despite “good” metrics. Another confounder is the student’s reaction to the debriefing, as this could significantly impact debriefer performance. Future studies should assess debriefer perceptions of the student after conducting the debriefing to assess whether students’ interactions with the patient and staff were appropriate, and/or, student reactions and openness to the debriefing.
Conclusions
In conclusion, this study demonstrated that student performance does not affect debriefer adherence to the PEARLS debriefing model. The analysis examined student performance through a conceptualization of ideal performance as definitive intervention and completion of critical actions. This study also explored any relationship between debriefer adherence and the time it took students to provide definitive intervention. This suggests use of the PEARLS model may help debriefers provide quality debriefing regardless of student performance.
References
Kolb DA. Experiential learning experience as the source of learning and development. Second edition. ed. Upper Saddle River, New Jersey: Pearson Education, Inc.,; 2015.
Kolb AY, Kolb DA. The learning way: Meta-cognitive aspects of experiential learning. Simul Gaming. 2009;40(3):297–327. https://doi.org/10.1177/1046878108325713.
Schon DA. The reflective practitioner: how professionals think in action. New York: Basic Books; 1983.
Eppich WJ, Hunt EA, Duval-Arnould JM, Siddall VJ, Cheng A. Structuring feedback and debriefing to achieve mastery learning goals. Acad Med. 2015;90(11):1501–8. https://doi.org/10.1097/ACM.0000000000000934.
Voyer S, Hatala R. Debriefing and feedback: two sides of the same coin? Simul Healthc. 2015;10(2):67–8. https://doi.org/10.1097/SIH.0000000000000075.
Ellis S, Davidi I. After-event reviews: drawing lessons from successful and failed experience. J Appl Psychol. 2005;90(5):857–71. https://doi.org/10.1037/0021-9010.90.5.857.
Locke EA, Latham GP. Building a practically useful theory of goal setting and task motivation: A 35-year odyssey. Amer Psych. 2002;57(9):705.
Kovacs G, Bullock G, Ackroyd-Stolarz S, Cain E, Petrie D. A randomized controlled trial on the effect of educational interventions in promoting airway management skill maintenance. Ann Emerg Med. 2000;36(4):301–9. https://doi.org/10.1067/mem.2000.109339.
Issenberg SB, McGaghie WC, Gordon DL, Symes S, Petrusa ER, Hart IR, et al. Effectiveness of a cardiology review course for internal medicine residents using simulation technology and deliberate practice. Teach Learn Med. 2002;14(4):223–8. https://doi.org/10.1207/S15328015TLM1404_4.
Tannenbaum SI, Cerasoli CP. Do team and individual debriefs enhance performance? A meta-analysis Hum Factors. 2013;55(1):231–45. https://doi.org/10.1177/0018720812448394.
Issenberg SB, McGaghie WC, Petrusa ER, Lee Gordon D, Scalese RJ. Features and uses of high-fidelity medical simulations that lead to effective learning: a BEME systematic review. Med Teach. 2005;27(1):10–28. https://doi.org/10.1080/01421590500046924.
Murphy KR, Cleveland J. Understanding performance appraisal: social, organizational, and goal-based perspectives. Thousand Oaks, Calif.: Sage Publications; 1995.
Ross L, Lepper MR, Hubbard M. Perseverance in self-perception and social perception: biased attributional processes in the debriefing paradigm. J Pers Soc Psychol. 1975;32(5):880.
Wegner DM, Coulton GF, Wenzlaff R. The transparency of denial: briefing in the debriefing paradigm. J Pers Soc Psychol. 1985;49(2):338.
Roze des Ordons AL, Cheng A, Gaudet JE, Downar J, Lockyer JM, Cumming School of Medicine F et al. Exploring faculty approaches to feedback in the simulated setting: are they evidence informed? Simul Healthc. 2018;13(3):195–200. doi:https://doi.org/10.1097/SIH.0000000000000289.
Peterson DT, Watts PI, Epps CA, White ML. Simulation faculty development: a tiered approach. Simul Healthc. 2017;12(4):254–9. https://doi.org/10.1097/SIH.0000000000000225.
Hallmark BF. Faculty development in simulation education. Nurs Clin N Am. 2015;50(2):389–97. https://doi.org/10.1016/j.cnur.2015.03.002.
Eppich W, Cheng A. Promoting excellence and reflective learning in simulation (PEARLS): development and rationale for a blended approach to health care simulation debriefing. Simul Healthc. 2015;10(2):106–15. https://doi.org/10.1097/SIH.0000000000000072.
Bajaj K, Meguerdichian M, Thoma B, Huang S, Eppich W, Cheng A. The PEARLS healthcare debriefing tool. Acad Med. 2018;93(2):336. https://doi.org/10.1097/ACM.0000000000002035.
Ahmed M, Arora S, Russ S, Darzi A, Vincent C, Sevdalis N. Operation debrief: a SHARP improvement in performance feedback in the operating room. Ann Surgery. 2013;258(6):958–63.
Cheng A, Hunt EA, Donoghue A, Nelson-McMillan K, Nishisaki A, Leflore J, et al. Examining pediatric resuscitation education using simulation and scripted debriefing: a multicenter randomized trial. JAMA Pediatr. 2013;167(6):528–36. https://doi.org/10.1001/jamapediatrics.2013.1389.
Cheng A, Rodgers DL, Van Der Jagt É, Eppich W, O’Donnell J. Evolution of the pediatric advanced life support course: enhanced learning with a new debriefing tool and web-based module for pediatric advanced life support instructors. Pediatr Crit Care Me. 2012;13(5):589–95.
Sawyer TL, Deering S. Adaptation of the US Army’s after-action review for simulation debriefing in healthcare. Simul Healthc. 2013;8(6):388–97.
Mullan PC, Wuestner E, Kerr TD, Christopher DP, Patel B. Implementation of an in situ qualitative debriefing tool for resuscitations. Resus. 2013;84(7):946–51. https://doi.org/10.1016/j.resuscitation.2012.12.005.
Kolbe M, Weiss M, Grote G, Knauth A, Dambach M, Spahn DR, et al. TeamGAINS: a tool for structured debriefings for simulation-based team trainings. BMJ Qual Saf. 2013;22(7):541–53.
Gaba DM, Howard SK, Fish KJ, Smith BE, Sowb YA. Simulation-based training in anesthesia crisis resource management (ACRM): a decade of experience. Simul Gaming. 2001;32(2):175–93.
Diekmann P. Videoassisted debriefing in simulation-based training of crisis resource management. Clin Simul: Oper Eng Manag. 2008. p. 667–76.
Dismukes RK, Gaba DM, Howard SK. So many roads: facilitated debriefing in healthcare. LWW; 2006.
Rudolph JW, Simon R, Dufresne RL, Raemer DB. There’s no such thing as “nonjudgmental” debriefing: a theory and method for debriefing with good judgment. Simul Healthc. 2006;1(1):49–55.
Hewson MG, Little ML. Giving feedback in medical education: verification of recommended techniques. J Gen Intern Med. 1998;13(2):111–6.
Archer JC. State of the science in health professional education: effective feedback. Med Educ. 2010;44(1):101–8. https://doi.org/10.1111/j.1365-2923.2009.03546.x.
Dieckmann P, Molin Friis S, Lippert A, Østergaard D. The art and science of debriefing in simulation: ideal and practice. Med Teach. 2009;31(7):e287–94.
Dreyfus S DH. A five stage model of the mental activities involved in directed skill acquisition. California University Berkeley Operations Research Center. 1980. http://www.dtic.mil/dtic/index.html. Accessed 14 Sept 2018.
Camargo CA Jr, Rachelefsky G, Schatz M. Managing asthma exacerbations in the emergency department: summary of the National Asthma Education and Prevention Program Expert Panel Report 3 guidelines for the management of asthma exacerbations. P Am Thorac Soc. 2009;6(4):357–66.
Acknowledgements
The authors wish to thank Lucas Smith, medical student, for conducting the video reviews necessary for the paper.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Ethical Approval
This study was approved and considered expedited by the institutional review board of Augusta University. The procedures used in this study adhere to the tenets of the Declaration of Helsinki.
Informed Consent
All necessary consent was obtained from participants prior to participation. Consent included for participation in the study activities, analysis of the results, and publication of those results. All participants were analyzed in a de-identified manner.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
McNutt, R., Tews, M. & Kleinheksel, A.J. Student Performance During a Simulated Patient Encounter Has No Impact on Debriefer Adherence to PEARLS Debriefing Model. Med.Sci.Educ. 31, 1141–1148 (2021). https://doi.org/10.1007/s40670-021-01290-2
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40670-021-01290-2