Background

Longstanding health inequities based on race, gender, socioeconomic status, and other social influencers of health have been the subject of renewed attention in light of current events such as the COVID-19 pandemic and our national reckoning with systemic racism [1, 2]. The Agency for Healthcare Research and Quality reports that patients of Black, LatinX, or indigenous race receive worse care in relation to 40% of quality measures assessed, and the annual National Healthcare Disparities Report consistently demonstrates that white patients receive better quality of care than other racial groups [3]. This differential in care remains after controlling for economic status, educational level, and healthcare access, suggesting discrimination on the part of the medical system [4], where covert prejudice remains present at the individual and institutional levels.

There are numerous factors which contribute to health inequities, but mounting research suggests that implicit bias toward patients may have measurable impacts on healthcare [5]. Implicit bias is an unconscious and unintentional association between a category of people and some attribute [6]. While explicit attitudes are deliberate and conscious, implicitattitudes can affect behavior without conscious volition [3].

Post-graduate physicians may be an attractive target audience for educational interventions about implicit bias because they are responsible both for making clinical decisions and training future generations of physicians. Despite this, the availability of opportunities for physicians to explore their biases in a formal setting after medical school is unclear, and no literature review has been conducted on post-graduate physician implicit bias interventions [7,8,9].

We conducted a comprehensive scoping review to present the content and outcomes of educational interventions which address post-graduate physicians’ implicit bias toward patients, to potentially inform decision-making of medical educators seeking similar interventions.

Methods

We employed a rigorous scoping review methodology, using the JBI Manual for Evidence Synthesis framework [10]. Using this strategy, our scoping review was split into the stages below:

Developing the research question

We asked, how can implicit bias toward patients be addressed through physician educational programs?

Inclusion criteria

The population of focus was post-graduate physicians, such as resident physicians, fellows, and attending physicians of all specialties, including populations in which physicians were a subgroup of a larger group of learners. We focused on curricula addressing implicit bias toward patients and defined implicit bias as stigma, prejudice, stereotype, and other forms of unconscious bias based on race, socioeconomic status, sexual orientation, weight, substance use, and any other personal identifying trait. We defined curricula as any planned educational experiences, including clinical rotations, didactics, training programs, and conferences. Primary research, systematic reviews, books, editorials, guidelines, videos, and conference abstracts were included, while non-English language studies were excluded. We included literature describing implemented curricula as well as literature which provided recommendations and theoretical background for potential interventions. We did not limit studies by publication date.

Search strategy

With the aid of an experienced research librarian (P. Bain, Countway), we conducted a search of MEDLINE (Ovid), Embase, Web of Science, ERIC, CINAHL, and PsycINFO in February 2020 using the search strategy and keywords in Additional file 1: Appendix 1. Because we found relevant articles from the database MedEdPORTAL which were not identified in this initial search, we conducted a manual search of MedEdPORTAL in November 2020 using the terms “implicit bias,” “unconscious bias,” “prejudice,” and “stigma.”

Sources of evidence selection

We used Covidence systematic review management software (Melbourne, Australia) for each step of screening and data extraction. First, all reviewers applied inclusion criteria to 10% of the papers to ensure that we were uniform in our screening. We conducted the remaining screening in two stages: titles and abstracts were screened initially, then the full texts of included articles were screened to determine final eligibility (Fig. 1). All coauthors (S.G., M.C., B.A., R.J., N.K., K.S., R.S., J.T., C.V., and J.K.) participated in both rounds of screening. Each article was independently reviewed by two coauthors using predefined selection criteria and we resolved disagreements with reviewer discussion until consensus was reached.

Fig. 1
figure 1

Flowchart of the screening process using PRISMA (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) guidelines [11]

† We assigned reviewers for full text review such that the screeners for each article’s full text were different from the screeners for its title/abstract review. This ensured that each article was screened in total by 4 different reviewers, in order to minimize effects of individual biases or subjective interpretations of criteria

‡ We imported sources cited in the bibliographies of included studies into Covidence and repeated the two-phase screening process

Data extraction

Coauthors (S.G., M.C., B.A., R.J., N.K., K.S., R.S., J.T., C.V., and J.K.) collected data from the included studies using a data extraction form (Additional file 2: Appendix 2). The form’s data fields were guided by educational principles deemed most relevant by the coauthors as well as the Guideline for Reporting Evidence-Based Practice Education Interventions and Teaching (GREET) checklist [12].

Analysis of the evidence

Three coauthors (S.G., M.C., and J.K.) analyzed data qualitatively and quantitatively, using frequency counts for key characteristics identified. Interventions were categorized into 4 distinct educational models developed iteratively via inductive coding by the authors. We analyzed outcomes using Kirkpatrick’s 4 levels of program evaluation, an analytic model for curricular outcome measurement [13].

Results

Curriculum characteristics

Our review identified 90 articles on implicit bias interventions for post-graduate physicians. Table 1 presents the aggregated data from these articles, and Additional file 3: Appendix 3 summarizes characteristics of all 90 articles.

Table 1 Aggregated data from 90 studies included in scoping review on post-graduate physician implicit bias curricula

Educational models

Inductive data analysis revealed 4 educational models used in implicit bias curricula: Competence, Skills-Based, Social Contact, and Critical Models. Their different theoretical foundations and pedagogical approaches are summarized in Table 2.

Table 2 Educational models identified in curricula addressing post-graduate physicians’ implicit bias toward patients

Outcomes reported

Eighty percent of the educational interventions reported outcomes. Outcome assessments most frequently relied on learners to self-report the perceived effects of the curriculum through pre and post surveys (62%). Figure 2 depicts the interventions’ approaches to outcome measurement through the lens of Kirkpatrick’s model for program evaluation [13].

Fig. 2
figure 2

Number (%) falling into each of the 4 levels of Kirkpatrick’s Triangle for Program Evaluation [13], of 53 articles identified

Curriculum analysis

Most of the educational methods employed were interactive (67%), and this was the most commonly identified curricular strength (26%). The most common weaknesses identified were related to resource availability, such as schedule and timing, funding, and institutional investment (53%).

Discussion

Our review identified several elements and challenges of effective physician implicit bias curricula. Below we highlight a spectrum of educational approaches to these curricula, as well as areas for improvement in implementation and outcome assessment.

Educational models

The 4 educational models (Table 2) identified in our analysis present various strengths and weaknesses. Competence Models have been critiqued for presenting implicit bias as a problem to be understood and resolved at the level of the individual [15,16,17], often by increasing learners’ awareness of their bias. Although evidence does not support the premise that increased awareness alone will allow clinicians to manage their own implicit bias [18, 19], self-reflection may trigger cognitive dissonance and increase learner motivation to change. In our review, 20% of interventions identified self-reflection on personal bias as a strength. On the other hand, when Competence Models are used to improve learners’ understanding of cultural groups by focusing on categorical traits rather than individuation, they may have the counterproductive effect of actually increasing reliance on stereotypes [20,21,22]. It is critical that interventions demonstrate heterogeneity rather than homogeneity within stereotyped groups, a strength which was recognized in 20% of curricula published in this review.

Skills-Based Models draw upon evidence-based strategies in Social Cognitive Psychology that aim to reduce stereotyping outside of healthcare settings [18, 23,24,25]. These skills may include “perspective-taking,” which fosters empathy by asking learners to imagine themselves in a patient’s position. Another practice, called individuation, consciously focuses on “specific information about an individual,” [18] which may “increase [learners’] capacity to see others as members of a common ingroup” instead of an outgroup [23]. Such models sometimes employ mindfulness, which encourages “attention to one’s own thought processes…and how they affect decisions so that one pays attention to the details of clinical care rather than falling back on habits…such as stereotypes” [20].

Social Contact Models facilitate direct interaction with diverse patients to foster empathy and enhance learners’ comfort, confidence, and positive emotions in interactions with people they perceive to be outgroup members [23, 24, 26, 27]. Evidence suggests that social contact only leads to these positive outcomes in specific conditions, namely, the presence of shared goals and equal status between both parties [20, 27]. Otherwise, such interactions have the potential to strengthen previously held stereotypes [20, 27]. To address this risk, novel approaches incorporate standardized patient encounters with debriefing [20]. One downside to Social Contact Models is that lessons learned with specific populations may not be easily applied to other contexts, in contrast to Skills-Based Models, which provide tools meant to be universally applicable.

Critical Models seek to profoundly transform the paradigms through which learners think about equity and justice in the medical system. In contrast to other models, which seek to avoid provoking discomfort or defensiveness among learners [16, 20], Critical Models intentionally present learners with experiences designed to arouse emotions, destabilize assumptions, and trigger cognitive dissonance. According to transformative learning [19, 28, 29], an educational theory which focuses on adult learning, such an exposure to a “disorienting dilemma” [30] prompts learners to “engage in a process of self-examination,” leading to paradigm shift [31].

Curriculum implementation

Each educational model encountered challenges in its implementation. Our review revealed barriers related to institutional investment and culture, availability of experienced facilitators, and learner-related factors.

Institutional attitudes can support or impede learning by impacting the time and funding available for implicit bias programs [29]. Given the multiple competing demands for medical staff time [32], it is unsurprising that over half of the interventions held only a single session, despite concern that “the lessons of a onetime workshop…tend to fade as the volume of work increases, and old practices reassert themselves” [33]. When institutional investment is lacking, the burden is carried by a handful of sometimes overtaxed individuals, as one author recalls, “we had momentum. What we didn’t have was money…which was a recipe for a lot of talk and no action…it seemed pretty clear I was going to have to find the funding for it myself” [34]. We also observed an uneven distribution of implicit bias programs between various specialties, illustrating how departmental subcultures may affect the accessibility of such trainings.

Another barrier identified was the availability of facilitators who were comfortable and well-versed in the subject matter [20, 26, 29, 33, 35, 36]. Only half of the interventions discussed the training of facilitators. A deficiency of experienced facilitators could detract from curriculum feasibility and quality while compounding variability in learner experiences. Facilitators may be wary of teaching implicit bias because of the sensitivity of the subject matter, inadequate preparation and training, or institutional cultures of silence with relation to bias [29]. Some questioned the evidence behind implicit bias, or felt antagonized when confronted with inequities in their establishment [34]. In response, several articles investigated best practices for facilitator training and identified this as a crucial area for future research [15, 29, 37].

Implicit bias programs were also impacted by factors related to learners. Multiple studies relayed concerns that the voluntary nature of these curricula meant that attendees were “self-selected,” [38] such that the program may have been “preaching to the choir.” Interventions can reach a greater array of learners if their institutions value implicit bias training and support learners in making time for it [26]. Changing institutional culture may also address another learner-related factor: the defensiveness and feelings of shame, fear [29] or denial [39] that may be experienced when confronting one’s own bias. Although such discomfort can be part of the process, as in the case of Critical Models [30, 31], too much discomfort can be counterproductive. Educators should provide a supportive environment to intentionally channel learner discomfort into behavioral change [20, 31].

Environments which support vulnerability and are free of criticism are optimal if learners are to experience transformative change [16]. One study suggested that “self-reflection, self-awareness, discovering…of often shameful past experiences of bias—could only be accomplished through…a non-judgmental environment in which everyone feels comfortable expressing their views with little fear of mockery or embarrassment” [16]. It is also crucial to avoid taxing learners who are underrepresented minorities by treating them as token representatives of their group or expecting them to educate other learners [40]. Educators must strive to “create a learning environment that fosters safety, trust, and respect,” “vet speakers, content, and materials carefully,” and “employ andragogical versus pedagogical methods of learning” which treat learners as active agents in their own learning [41]. Striking this balance may be especially difficult when power differentials exist between facilitators and trainees, which reinforces the need for robust faculty development [29].

Outcomes reported

Program evaluation is an essential component of curriculum development [42, 43]. Seventeen percent of studies in this review labeled evidence supporting interventions as a strength. This suggests that educators are seeking data to guide curricula, yet 20% of interventions did not report results. Faculty development initiatives should explicitly encourage educators to create a prospective evaluation plan to measure and disseminate outcomes, so that others may benefit from the lessons learned.

Kirkpatrick’s model for program evaluation (Fig. 2) is a well-known paradigm to categorize approaches to outcome measurement. The reported outcomes of included publications most commonly mapped to Level 2: Learning, which relates to learners’ knowledge, attitudes, or skills, as well as confidence or commitment to change [13]. Noteworthy shortcomings exist within this subset of data. While optimal measurement at Level 2 would involve an external evaluator [42], many studies reported outcomes via self-assessments, raising concerns about their validity [44]. As an alternative, several authors measured IAT scores, often in a pre/post intervention format. The advantages of such an approach are the rigor with which IAT instruments are developed and evidence that the IAT has greater predictive validity than other self-report measures [45], but some publications question the validity and precision of IAT-based data [46,47,48].

Few included studies attempted to measure outcomes at Kirkpatrick Levels 3–4. Level 3 assesses the degree to which learners apply what they learned, and Level 4 assesses targeted outcomes and organizational benefits [13]. Although measurement at these higher levels is challenging due to the time, money, and methodologic expertise required [49, 50], investing in such outcome evaluation presents the best opportunity to demonstrate meaningful impact on physician implicit bias and patient care [13]. Many of the interventions described in this review do not measure efficacy at these higher level outcomes, a limitation which has been recognized in prior implicit bias research [51]. Educators wishing to adopt similar curricula should understand that evidence directly supporting these interventions’ reduction of implicit bias in the clinical or learning environments is lacking. It is our hope that with higher level outcome assessment, more longitudinal interventions employing engaging teaching modalities, increased faculty training, and organizational culture eager to address implicit bias, our field will refine implicit bias curricula and benefit from more compelling data supporting them.

Conclusions

Our analysis of the literature on post-graduate physician implicit bias curricula highlights opportunities for next steps in the field:

  1. 1.

    Educators seeking implicit bias curricula can consider the educational models, teaching modalities, and challenges identified in this review to critically apply and improve aspects of these interventions in their own programs.

  2. 2.

    Institutional investment and faculty development were commonly identified challenges in this review of implicit bias curricula. Educators should examine whether their organizational culture, leaders, and teaching faculty will support implicit bias curricula and commit needed resources.

  3. 3.

    Implicit bias curricula should be evidence-based. This requires more widespread program evaluation using well-validated instruments, and especially assessing changes in physician behavior and impacts on patients.

Limitations

This scoping review presents an extensive yet incomplete snapshot of implicit bias curricula for physicians. It is limited to the databases we searched, although we identified additional papers through the iterative process of screening included studies’ bibliographies. In addition, many articles provided only brief information in the form of an abstract. Each stage of screening, data extraction, and coding likely introduced a degree of bias from the reviewers, which we mitigated by having 2 reviewers reach consensus at each step. Finally, per scoping review methodology, we did not consider the quality of the studies we included. This lack of discrimination should be considered when extrapolating results.