INTRODUCTION

The Veterans Health Administration (VHA), part of the US Department of Veterans Affairs (VA), is the largest integrated health care delivery system in the USA, caring for 6.3 million patients in fiscal year 2021.1 VHA delivers care through its own facilities, supplemented through purchasing care for enrolled veterans from community (non-VHA) providers. It has faced difficulties with access to care, specifically wait times, over the past decade, prompting congressional and administrative action—through legislation, regulation, and appropriations.2,3,4,5 The VA Maintaining Internal Systems and Strengthening Integrated Outside Networks (MISSION) Act was enacted by the US Congress in 2018.6 In order to find ways to improve access and patient experience and to reduce provider burnout, Sect. 507 of the MISSION Act mandated a 2-year pilot program to study medical scribe introduction in VA Medical Centers (VAMCs), focused on specialty care clinics and emergency departments.

Medical scribes are trained but clinically unlicensed professionals who can assist providers in navigating the electronic health record (EHR) and entering patient visit information.7 Scribes are employed in clinical settings to increase provider productivity and satisfaction by minimizing physicians’ documentation burden and burnout, and to improve patient experience by increasing the time spent with patients instead of entering patient information into the EHR. Evidence from past studies suggests that the use of medical scribes may be associated with increased productivity in a variety of clinical settings, although not uniformly. Studies found that scribes may be associated with increased provider productivity measured by the number of patients seen per hour or relative value units (RVUs) per hour in primary care,8,9,10,11,12 specialty care,13,14,15,16 and emergency department settings.17,18,19 Only a few studies have examined the impact of scribes on provider satisfaction, time spent on documentation, time spent face-to-face with patients, and patient satisfaction. Scribes have been associated with decreases in time spent on documentation and increases in provider satisfaction.12,14,20 While studies found evidence for increases in time spent face-to-face with patients associated with the introduction of scribes,9,21 there is not strong evidence for an association between scribes and overall patient satisfaction.12,14,22,23 Moreover, outside VA, evidence suggests that scribes may be associated with increased productivity, and reduced time spent on clinical documentation without affecting patient satisfaction. However, this may be dependent on which providers are assigned or choose to use scribes.24,25,26 Notably, none of this evidence on the impact of scribes on outcomes was generated within VA, was generally on a smaller scale, and does not support causal conclusions.19

As mandated by the MISSION Act, our objective was to evaluate the impact of medical scribes on provider productivity, wait times, and patient satisfaction in specialty care clinics. The focus of the MISSION Act scribe pilot on specialty care is partly due to the high wait times in specialty care that have necessitated greater access to non-VHA care in the community. Furthermore, with over a third of VA physicians reporting burnout, albeit lower than the private sector,27 and the challenges that VA faces in recruiting and retaining physicians,28 any strategy that can improve productivity and/or decrease provider burnout could have a considerable impact on access and quality of care. Our study provides insights into whether scribes can help improve productivity, decrease wait times, and enhance patient experience in VA specialty clinics.

METHODS

Study Design and Setting

We conducted a cluster randomized trial to determine the impact of medical scribes on provider productivity, wait times, and patient satisfaction in VHA. We compared the outcomes from the intervention sites with baseline (pre-intervention) data (January–December 2019) and pilot period data (June 2020–June 2022) from the comparison sites. The months in between (January–May 2020) were excluded as some sites had begun to hire in anticipation of the start of the pilot, which was delayed by the onset of the COVID-19 pandemic and also caused sizable disruptions in patterns of care. Section 507 of the MISSION Act specified that at least four VAMCs included in the pilot must be located in rural and urban areas, with at least two in underserved areas, which we identified using the highest quartile of cardiology and orthopedics wait times from January 2018 to June 2019. Furthermore, 30% of the medical scribes were to be employed in ED and 70% in specialty care clinics. The Office of Veterans Access to Care (OVAC), now known as the Office for Integrated Veteran Care (IVC), developed a list of 32 VAMCs interested in participating in the pilot. We stratified the VAMCs into the relevant subcategories based on the requirements of the law and randomly selected VAMCs within each group for assignment to the intervention, with the remaining VAMCs included as the comparison group (Table 1).29

Table 1 VA Medical Centers Included, by Randomization Status and Specialty

Intervention

Section 507 of the MISSION Act specified that four scribes were to be assigned to each of the 10 VAMCs, which increased to 12 upon implementation, with VA hiring two of the scribes as VA employees and the other two as contractors at each site. The hiring of VA and contract scribes began before the official start of the pilot (June 30, 2020), which was pushed back several times due to the onset of the COVID-19 pandemic and continued throughout the duration of the pilot (ended July 1, 2022). Two scribes were to be assigned to two providers each at each VAMC, with providers volunteering to participate. Scribes assisted their assigned provider in documenting relevant patient information from each visit in the VA EHR. The notes made by scribes were tagged with their name as well as the date and time and were approved by the provider before becoming viewable in the EHR—also allowing for identification of visits for research purposes in the VA Clinical Data Warehouse (CDW) that involved scribes.

Outcomes

Our study includes provider productivity, wait times, and patient satisfaction outcomes as mandated by the MISSION Act. We included three clinic-pay period level measures of provider productivity—(1) relative value units (RVUs) per full-time equivalent (FTE) per pay period, (2) visits per FTE per pay period, and (3) patients per day per pay period, which included the workload of both physicians and other advanced practice providers (e.g., physician assistants, nurse practitioners). For each provider in our included specialty clinics, we obtained patient visits and associated RVUs from the CDW and their clinical workload on an FTE basis for each pay period from the Specialty Productivity Report and Quadrant Tool (SPARQ) maintained by the Office of Productivity, Efficiency, and Staffing (OPES). Providers with scribe and non-scribe visits in the intervention sites were included to account for the potential site-wide spillover effect from scribe intervention.

We also included three clinic-pay period wait times measures—(1) request to appointment day (total wait time), (2) request to appointment made (time between request for an appointment and appointment creation, the administrative time to make the appointment), and (3) appointment made to appointment day (time between appointment creation and appointment day). Appointments that were cancelled, a no-show, or created on the same day as the visit (e-referrals, and text/email messages between providers30) were not included. These represent the wait times for new consultations only, a validated measure of wait times in specialty care, with lower wait times having been previously associated with higher patient satisfaction.31

Finally, we included six patient satisfaction outcomes measured using survey data from Veterans Signals (V-Signals), a nationally administered email-based survey used to understand the veteran experience and satisfaction with the care received at VAMCs (approximately 18% response rate).32 We used two V-Signals surveys (Outpatient–Scheduling an Appointment and Outpatient–Health Care Visit) to capture these six measures—(1) “It was easy to get my appointment”; (2) “I got my appointment on a date/time that worked for me”; (3) “After I checked in for my appointment, I knew what to expect”; (4) “My provider listened carefully to me”; (5) “My provider explained things in a way that I could understand”; and (6) “I trust this clinic for my health care needs.” Veterans responded to each on a 5-point Likert scale (1 is strongly disagree and 5 is strongly agree), which we dichotomized for analysis (1 for strongly agree or agree, 0 for neutral, disagree, or strongly disagree), and these were averaged to the clinic-pay period level for cardiology and orthopedics.

Statistical Analysis

We used difference-in-differences (DID) regression models to compare the changes in productivity, wait times, and patient satisfaction outcomes at the clinic-pay period level with an intent-to-treat approach, identifying differences between intervention and comparison sites after the scribe intervention began accounting for baseline differences and numerous covariates. All clinical activity was included in our productivity and wait time models because the presence of scribes can have a spillover effect, thus impacting outcomes for providers and visits involving scribes and those without. The models include a randomization status indicator (Intervention, 1 if randomized to have scribes), pilot period indicator (Pilot, 1 if after June 2020), and their interaction (Intervention × Pilot) plus the covariates described below. Our unadjusted models only include the DID terms as well as clinic and pay period fixed effects while the fully adjusted models also include the proportion of enrollees over age 65, the proportion of enrollees aged between 35 and 65, the proportion of high-priority status enrollees (7 and 8), Zillow Home Value Index, Medicare Advantage (MA) penetration rate (Centers for Medicare and Medicaid Services), average patient risk scores (NOSOS score from VHA Support Service Center), and proportion of enrollees by race (African American, American Indian, Asian, and Native Hawaiian). All enrollee proportions were obtained from the VA Planning Systems Support Group (PSSG) or Survey of Enrollees. This study was Congressionally mandated and therefore considered part of VA operations rather than human subjects research by the VA Boston Health Care System Internal Review Board and was registered on ClinicalTrials.gov (NCT04154462) in November 2019. The datasets generated during and/or analyzed during the current study are not publicly available and cannot be shared upon request as they include confidential VHA patient and workforce data.

Deviations from Protocol

The implementation of the MISSION 507 scribe pilot experienced a few deviations from the protocol registered on ClinicalTrials.gov. Due to challenges in hiring and retaining scribes, the planned number of scribes in both hiring modalities was not reached. Also, one of the sites randomized to implement scribes in its emergency department dropped out of the pilot; however, it was not an intervention site for specialty care and therefore is not a relevant deviation for this analysis. We also made slight variations to the outcomes and covariates included in our regression models. Our protocol specified using scribe FTEs per 1000 patients and quartiles of scribe FTEs per physician FTE; however, given the incomplete fidelity to the dosage of scribe hires that the randomized clinics were able to achieve, we felt that an intent-to-treat approach was better than introducing potential endogeneity (selection bias) by using a measure of intervention dose that would likely be correlated with our outcomes and other factors about each facility (e.g., those that did a better job of hiring and keeping scribes also likely are better in other unobserved ways). We were unable to include several of the prespecified covariates, including age 18–64 male insurance rate, median household income, veteran unemployment rate, and average enrollee driving distance because data for 2021 and 2022 were not yet available at the time of analysis. However, we expect that Medicare Advantage penetration can partially account for variation in insurance coverage, and Zillow Home Value Index and VA priority status can control for variation in the socioeconomic profile of communities and the relevant enrollee population for each clinic. We decided not to include community care wait times for specialty care because of the potential simultaneity with VA wait times.

RESULTS

Despite randomization, stratified within specialties and meeting other requirements of the MISSION Act, there were meaningful differences between our intervention and comparison sites (Table 2). The proportion of enrollees between ages 35 and 65 and the proportion of white enrollees were higher in the intervention sites while the proportion of enrollees over age 65, the proportions of African American, Asian, and Native Hawaiian enrollees, Medicare Advantage penetration rate, and NOSOS risk scores were lower in the intervention sites. There were no significant differences between intervention and comparison sites in the proportion of enrollees with VA priority status and Zillow Home Value Index.

Table 2 Baseline Patient and Station Enrollee Characteristics, by Randomization Status

Baseline provider productivity, as measured by RVUs or visits per FTE, was lower in cardiology intervention sites relative to comparison sites and higher for patients per day per provider (Table 3). The only difference in baseline provider productivity between orthopedics intervention and comparison sites was for patients per day per provider, where it was significantly higher for intervention sites. Baseline wait times for new consults in both cardiology and orthopedics were significantly higher for intervention sites relative to comparison sites. A comparison of patient satisfaction across randomization status and specialty showed similar ratings, with mostly no significant differences between intervention and comparison sites. Trend graphs for all outcomes, used to assess parallel trends during the pre-intervention period, are shown in Appendix Figs. 1 through 24.

Table 3 Baseline Provider Productivity, Wait Times, and Patient Satisfaction, by Randomization Status and Specialty

Provider Productivity

Provider productivity declined from the baseline to pilot period along all three measures in both specialties, with greater declines in the comparison group in all cases (Table 4, Appendix Tables 1 and 2). In our fully adjusted DID models, randomization to scribes was associated with an increase of 8.48 visits per FTE (95% CI: 3.13, 13.81; p = 0.002) and 25.17 RVUs per FTE (95% CI: 17.05, 33.29; p < 0.001), but no change in patients per day per provider. Similarly, randomization to scribes was associated with an increase of 12.50 visits per FTE (95% CI 5.17, 19.83; p = 0.001) and 17.26 RVUs per FTE (95% CI: 7.43, 27.10; p = 0.001), but no change in patients per day per provider. The unadjusted DID results are also shown for comparison, which yields similar conclusions but larger coefficient estimates (Table 4).

Table 4 Difference-in-Differences Results for Provider Productivity in Cardiology and Orthopedics

Wait Times

Request to appointment day (total) wait times and request to appointment made, the administrative component of wait times, significantly increased in cardiology from the baseline to pilot period in both intervention and comparison clinics (Table 5, Appendix Tables 3 and 4), with no change in appointment made to appointment day wait times. In our fully adjusted DID models, we found no association between randomization to scribes and changes in wait times in cardiology. However, in orthopedics, we observed significant increases in all three wait times measures pre-post in the comparison sites, with decreases (request to appointment day, appointment made to appointment day) or no change (request to appointment made) for intervention sites. In our fully adjusted DID models, we found an 8.54-day decrease (95% CI: − 12.63, − 4.46; p < 0.001) in request to appointment day wait times and a 5.74-day decrease (95% CI: − 8.38, − 3.09; p < 0.001) in appointment made to appointment day wait times in orthopedics clinics that had the scribe intervention relative to the comparison sites, and no significant change in request to appointment made wait times.

Table 5 Difference-in-Differences Results for Wait Times in Cardiology and Orthopedics

Patient Satisfaction

We found no evidence for meaningful changes in patient satisfaction related to the scribe pilot (Table 6, Appendix Tables 5 and 6); only one of 12 fully adjusted models showed a significant difference associated with randomization to scribes—an improvement of 0.07 (p < 0.05) for “I trust this clinic for my healthcare needs” in cardiology.

Table 6 Difference-in-Differences Results for Patient Satisfaction in Cardiology and Orthopedics

DISCUSSION

Based on a 2-year trial of medical scribes involving 18 VA Medical Center specialty clinics, we find suggestive evidence of randomization to scribes being associated with increased productivity in cardiology and orthopedics, decreased wait times in orthopedics, and no change in patient satisfaction. Our findings in the VA setting are consistent with previous non-VA studies that found associations between scribe intervention and increased productivity.8,13,14,15 In our study, randomization to the scribe pilot was associated with improved productivity when measured as visits per FTE and RVUs per FTE, but not patients per day per provider, which may be a result of the lack of FTE adjustment in that measure. Although scribes’ assistance may increase providers’ availability in the clinic and reduce wait times,22,33,34 providers’ scheduling patterns and administrative staff support are also key to reducing wait times. Providers’ scheduling patterns or other support staff availability were not adjusted as part of the pilot; further examination of the impact of scribes on wait times in future studies would be beneficial. Although our quantitative evaluation did not suggest any meaningful change in patient satisfaction, the findings from a concurrent qualitative evaluation performed by the VA Collaborative Evaluation Center (VACE) suggest a potential association between scribes and improvements in providers’ attention to patient care, accuracy of visit documentation, patient experience, and provider satisfaction, depending on the scribe expertise.35

Our study has several limitations. First, sites were randomly selected from a list of sites that exhibited interest in participating in the scribe pilot rather than being randomly selected from all VA facilities, potentially limiting generalizability if this intervention were to be scaled across VHA. Second, provider participation was voluntary, which again could have implications for scalability and what effects could be expected if scribes were introduced to the care process without buy-in from providers. Third, we do not account for the difference in hiring modality of scribes (VA or contractor) in our models because of insufficient statistical power. There may be differences in training between VA and contract scribe training and/or prior experience that are relevant to these outcomes. Finally, cost analysis that can help inform adoption decisions of scribe programs in the VA is not included in this article.

To our knowledge, this is the first study to examine the impact of medical scribes in a multi-center randomized trial within a large integrated health care system, like the VHA. Moreover, medical scribes may increase provider satisfaction, improve the quality of provider and patient interactions, and decrease physician burnout due to the documentation burden related to electronic health records, none of which we are able to capture in this study.12,20,36 The tradeoff between cost and benefits should be considered in any assessment of the potential scalability of scribes across VHA particularly given the difficulties in hiring observed during this pilot. Multiple non-VA studies have found associations between the use of scribes and financial benefits for practices,10,14 although the budgetary and revenue considerations within the VA context are markedly different than in the community. Given our findings that scribes have the potential to increase productivity and decrease wait times for VA specialty care and continued policy and management focus on improving access to care for veterans, further evaluation and consideration for implementation across VHA and other access-related interventions are needed.