Background and Objective

The patient-centered medical home (PCMH) is a widely accepted model for transforming the organization and delivery of primary care.1 PCMHs strive to provide comprehensive care, build partnerships with patients, and emphasize care coordination and communication. Primary care teams within integrated health systems require training and other resources to support their transformation into successful PCMHs and promote high-quality care.

One such resource is the online toolkit, where PCMH staff can easily access materials and strategies. Several free and fee-for-use toolkits are available, usually aiming to support initial PCMH implementation24 or PCMH certification.5,6 However, few toolkits focus on quality improvement (QI) once basic PCMH implementation has been completed, and published evaluations of toolkits are limited.

In this article, we present findings from a multimethod evaluation of a PCMH toolkit developed by the Veterans Health Administration (VA). The Consolidated Framework for Implementation Research (CFIR) was used to guide this evaluation, which synthesized data from web analytics, online surveys, and interviews with end users.

Innovation

In 2010, the VA began the Patient-Aligned Care Team (PACT) Initiative to transform more than 800 VA primary care practices into PCMHs.7 As the nation’s largest integrated health care system, the VA established a range of PACT performance measures and defined three “pillars” of PACT success: access to care, care coordination, and practice redesign. The PACT Toolkit (“Toolkit”) was developed to share “tools”—that is, technical, clinical, or organizational innovations—linked to one or more performance measures or pillars.

The Toolkit is a VA intranet site that now contains more than 60 tools, which can be accessed and downloaded by any VA staff. It grew out of in-person regional learning collaboratives, where PACT teams gathered to learn process improvement strategies.8,9 Target Toolkit users are PACT coaches and team members, including providers, nurses, and clinic managers. Tool adoption is voluntary, and most tools can be customized to local needs.

Clinic staff created the component tools, which were vetted by physicians and nurse program leaders. Each tool’s webpage identifies the relevant performance pillar, briefly describes the tool, and provides download links (Figure available online). Examples of tools include EHR note templates, training materials, patient brochures, and care process flowmaps.

To promote use of the Toolkit, a two-pronged social marketing strategy employed direct contact and facilitation to stimulate website visits and the use of individual tools.1012 Toolkit staff created a brand logo, sent e-brochures to distribution lists of collaborative participants, and e-mailed periodic updates to a Toolkit users’ listserv. They also gave presentations to VA opinion leaders and demonstrated the Toolkit at each regional collaborative. Collaborative attendees were given promotional brochures and encouraged to share information about the Toolkit with colleagues.

Design and Participants

Our evaluation integrated web analytic data, online survey responses, and interview responses. Because the Toolkit was developed as an enhanced Microsoft SharePoint® website, we used SharePoint’s web analytics function to measure the number of website visits by unique users and facilities. We present data from the Toolkit launch in September 2011 through March 2013.

The CFIR guided survey design and interview questions.13 This framework comprises five domains: intervention (i.e., Toolkit) characteristics, outer setting (e.g., health care system policies), inner setting (e.g., facility characteristics, local culture), characteristics of individuals involved, and implementation processes (e.g., dissemination). We focused on the inner setting and individual characteristics domains. The intervention characteristics domain was measured using respondent ratings of the Toolkit’s quality and the implementation process domain using a survey item asking how the respondent heard of the Toolkit. We assumed the outer setting remained constant across facilities because of the VA’s simultaneous, system-wide PACT rollout.

Staff at 142 VA medical centers and subsidiary outpatient clinics were asked to participate in an online survey and telephone interviews. To identify a survey sample frame, we compiled a list of collaborative attendees and other PACT team members. E-mail invitations to complete the anonymous survey using SurveyMonkey™ were sent to 2,899 people. From June-August 2012, we received 893 responses (a 30.8 % response rate). As shown in Table 1, 65 % of respondents were primary care providers, RNs, or care managers, which are roles central to the PACT model. Respondents represent 136 facilities, from all geographic regions and VA facility complexity levels, for a facility response rate of 95.8 %. We excluded 305 respondents who were not PACT team members or were unsure, and 44 who were clerical associates; they were asked only about personal characteristics.

Table 1 Respondent and Facility Characteristics

The final survey sample of 544 respondents received up to 23 questions focused on: Toolkit awareness and use, PACT involvement, personal characteristics, and organizational readiness for change. Respondents who had not heard of the Toolkit, had not visited the website, or whose facility had not implemented tools were skipped out of some questions. We included a final open-text field for comments; 124 respondents offered remarks.

Interview subjects were recruited from among the 1,147 PACT collaborative participants and 615 Toolkit super-users (defined as those who had browsed more than ten Toolkit webpages in the first 30 days after their initial visit). Because we suspected there might be regional and facility-level variation in PACT implementation, we purposively recruited individuals from among each of five different regions and three different levels of facility complexity. We completed 35 interviews from August-October 2012; interviews with three staff who were not PACT team members were excluded from the analysis. Super-users comprised approximately 25 % of interviewees. As shown in Table 1, approximately half of interviewees were primary care providers or RNs/care managers.

Semistructured telephone interviews elicited detailed descriptions of facility- and individual-level Toolkit use, enriching survey findings and grounding our understanding of the data. Each 30-min interview was recorded with the interviewee’s consent. The interview guide touched on several CFIR domains, including: (1) demographic information (individual characteristics), (2) awareness and adoption of the Toolkit (implementation process), and (3) regional/facility characteristics and culture (inner setting) contributing to Toolkit use. A note-taker produced notes for review; quotations were confirmed from recordings. We followed a directed content analysis strategy14 using a stepped procedure in which three analysts coded the interviews to ensure reliability. Coded interviews were then used to identify common themes.

This evaluation was determined to be non-human subject research by the Stanford University Institutional Review Board.

Main Measures

We used specific survey items to define three Toolkit use outcomes: awareness of the Toolkit, use of the website, and tool implementation at the facility. For tool implementation, we excluded respondents who worked only at subsidiary outpatient sites and collapsed individual responses to the facility (medical center) level; we inferred that affirmative responses were correct when other responses from the same facility were negative. Facility-level values for primary care patient population and facility complexity were attributed to each respondent.

We constructed multivariable logistic regression models for the Toolkit awareness and website use outcomes, with predictors drawn from CFIR domains as described above. Because respondents were nested within facilities, cluster robust standard errors15 were calculated. A block bootstrap (10,000 re-samplings) of each model was created to address small sample sizes relative to covariate sets16 and nesting of respondents within facilities.17 This procedure repeatedly took a subsample of our survey data, estimated a statistic, and then aggregated all these subsample statistics to reflect likely activity beyond our survey responses. Bias-corrected 95 % confidence intervals are reported for all models. Analyses were conducted using Stata® v11.1.

Key Results

Awareness and Use of the Toolkit

Web analytic data showed that 6,745 unique users visited the Toolkit website in the first 19 months, with an average of 1,108 new and return users per month. Usage was dispersed across all 21 VA regional networks, with a median of 271 users per network. Unique users represented 32.8 % of the 20,300 VA primary care providers and staff implementing PACT.

Survey responses indicated that 80.3 % of respondents had heard of the Toolkit from a variety of sources, including Toolkit promotional emails (46 %), emails from colleagues (14 %), word-of-mouth (8 %), and collaborative sessions (42 %; multiple responses total greater than 100 %). As shown in Table 2a, no significant differences in Toolkit awareness were observed across individual characteristics or inner setting features we measured, including team roles, medical center or outpatient location, facility complexity, or facility primary care enrollment.

Table 2 Predictors of Toolkit Awareness and Website Access by Survey Respondents

Among survey respondents who had heard of the Toolkit, 70.3 % had browsed the website. Of those who had not, 73 % cited not having time as the main reason; 16 % cited having no need to adopt tools and 5 % lack of interest. Different categories of users were equally likely to have visited the website, with no significant differences observed across team roles, practice location, facility complexity, or facility primary care enrollment (Table 2b).

Survey respondents who browsed the Toolkit gave it an average quality rating of 3.77/5. Of respondents who visited the website, 34 % spent an hour or more there; 62 % reported having downloaded a tool, and 7 % requested a tool that requires clinical application support for implementation. Ten percent reported having uploaded or suggested a tool, but only 4 % had contributed to an online discussion forum.

Aggregating multiple responses per facility showed that 65.1 % of facilities had implemented one or more PACT tools. Figure 1 shows how frequently individual tools were adopted.

Figure 1
figure 1

Frequency of tool implementation among facilities (n = 69).

The majority of interviewees expressed a favorable view of the Toolkit. Interviewees reported visiting the Toolkit for a number of reasons, including to address a specific quality deficit (approximately one quarter of interviewees) and learn how other clinics solved implementation problems (approximately one third of interviewees). One interviewee expressed support for the Toolkit saying, “I wish…we could have seen it early on in [PACT] development so that we didn’t have to reinvent the wheel…One of the things we had done before the Toolkit was…email around to see who had a best practice” [IntM]. Another explained how they use the Toolkit saying, “When I have to implement something, I go to the tool pages, like the Huddle Worksheet, which is good for new providers and teams…working together as a team, and communication…” [IntR]. Interviewees were most interested in point-of-care support tools, patient materials, and PACT training tools.

Another frequently cited use of the Toolkit was to support PACT training efforts. For example, one interviewee reported “When we do local PACT training…I go in and access some of the tools…I actually use it as a training tool” [IntC]. This finding from interview analysis was supported by survey comments. For example, one survey respondent commented “I work in PACT Learning Center education and include the PACT Toolkit in every session” [SR730].

Impact of Training on Toolkit Use

Survey respondents who reported receiving PACT training at the regional level were significantly more likely to have heard of the Toolkit and to have visited the website. However, receipt of facility-level training was not associated with Toolkit awareness or use (Table 2a, b).

Interviews revealed variation in content and depth of local PACT training sessions. Some interviewees described a top-down training approach where coaches trained at a PACT Center of Excellence, then returned to the facility to deliver standardized content. Some interviewees indicated their clinics were closed so staff could attend training sessions, yet others only mentioned attending monthly in-service trainings. Finally, some interviewees indicated they received almost no formal training. One subject said, “[I]…just had PACT virtual training…PACT info not provided in orientation; what I learned, I found on my own…” [IntL].

Impact of Facility Environment on Toolkit Use

Almost half (46.2 %) of survey respondents who indicated that their facility had implemented one or more tools reported that it was very or somewhat easy to implement new tools to improve primary care there, while 25.8 % reported it was very or somewhat difficult. Similarly, more respondents (44.2 %) reported that no special approval would be needed to implement a tool than reported permission would be required (31.2 %) or were unsure (24.6 %).

To better understand the local culture within which the Toolkit was used, we asked interviewees about their familiarity with QI and motivations to engage in it. Interviewees were, for the most part, familiar with local performance measurement. One interviewee reported that their facility has “monthly performance meetings where we go over that data…then spread out the information to the separate teamlets…” [IntM]. Another indicated that facility leadership plays a role in disseminating information about performance, “We know all our performance measures; we are updated regularly by…leadership…” [IntD].

Dissemination of performance data, by quality teams or facility leadership, was another recurring theme. Almost half of interviewees reported regularly seeking or receiving local performance feedback. They were also aware of the consequences for meeting or failing to meet specific goals. One respondent described these positive and negative consequences, saying, “The penalty is having to give action plans every month…The reward is that you get lots of verbal praise when a facility goes into the green. We are recognized at a monthly ceremony…the [PACT] metrics are part of performance pay for [certain staff positions]” [IntAC].

Survey comments provided additional information for explaining Toolkit use within a facility. Comments reflected three themes (Table 3). Some praised the Toolkit’s usefulness. Another group thought the Toolkit would be helpful if they had time to use it. The last group indicated that using the Toolkit would represent yet another burden within the already stressful environment of PACT implementation. Comments also identified a number of specific barriers to Toolkit use, including: (1) insufficient time to implement required PACT elements, let alone optional elements like the Toolkit; (2) inadequate staff; (3) insufficient PACT training; (4) out-of-touch or uncaring leadership; and (5) insufficient space or equipment.

Table 3 Select Comments to Open-Ended Survey Items, by Theme

Discussion

We evaluated use of an online interactive Toolkit enabling PCMH teams to share QI tools across a national health care system. Website data show that a large number and diverse set of VA staff used the Toolkit. Survey data and interviews show that users rated the Toolkit’s quality highly, and many spent substantial time on the website. Interviews and survey comments provided insights about individual and organizational factors that facilitated use of the Toolkit, but also identified barriers to use.

Several toolkits have been developed by professional societies,2,18 state medical organizations,3,19 or health plans4 to provide primary care practices with the basic building blocks for PCMH implementation. Other toolkits, often fee-based, contain extensive detail to help practices achieve PCMH certification from the National Committee for Quality Assurance or other organizations.5,6 However, these toolkits provide limited support for QI activities, and we are unaware of published evaluations of them.

Outside primary care, Speroff found that an online toolkit was less likely to change care processes than an online collaborative, although neither intervention improved long-term outcomes.20 Supporters view toolkits as empowering staff to customize solutions,21 while others view toolkits as passive resources requiring staff motivation to adopt change.20 DeWalt found clinic staff were willing to adopt “concise and actionable” tools but required some knowledge of QI methods to succeed.22 Our study did not measure clinical outcomes, but is the first to examine the combination of an online toolkit containing concise, actionable tools with in-person collaboratives.

Dissemination of quality improvement techniques presents serious challenges. Statistical analyses of CFIR’s individual characteristic and inner setting domains showed a broad reach of Toolkit awareness and use across medical centers and outpatient settings and from facilities of all sizes, complexities, and regions. This suggests that social marketing is one key to successful dissemination.

Another key is social interaction. Participation in regional PACT training sessions promoted awareness and use of the Toolkit; facility-level training, however, did not. We hypothesize that in-person interaction outside the PCMH team helps explain this difference. Moreover, qualitative results suggest that participation in external and/or internal networks (e.g., learning collaboratives or listservs) promoted Toolkit use and that recommendations from trusted colleagues were especially influential.23 Combined with interview findings about variations in training across facilities, these results suggest that future toolkit efforts may benefit from standardizing local PCMH training content and providing opportunities for out-of-team social interaction.

QI culture also plays a role. Aspects of a facility’s inner setting, including awareness about performance, a learning culture, a supportive climate, and readiness for change were associated with Toolkit use.13 Interviewees highlighted a perceived need for improving local deficits as a driver of performance improvement, and several indicated that a local performance focus was conducive to Toolkit use. Survey respondents in facilities that had implemented tools were more likely than not to report that it was easy to implement new tools there. Survey comments also describe aspects of local environments that are less favorable for Toolkit use, including lack of leadership support and inadequate time to identify and adopt tools.

The barriers to Toolkit use highlighted by our qualitative results may also reflect the inner-setting complexity of launching the Toolkit during the VA’s larger PCMH transformation. Staff whose time is most occupied with patient care seem to have benefitted less from the Toolkit than those whose roles (e.g., care managers) enabled them to review it in depth. But interviewees also report using the Toolkit to learn how other facilities addressed PACT implementation challenges. Thus, in the midst of system-wide primary care transformation, an online toolkit is an accessible knowledge-sharing resource that may allow facilities struggling with PCMH implementation to learn from more successful sites.

Some limitations should be considered when interpreting these findings. Although our survey response rate was good, non-response bias cannot be ruled out. Interview responses may not reflect the range of persons who declined to be interviewed. Data collection occurred during PACT implementation; use and perceptions of the Toolkit may change once PACT becomes routine system-wide. Lastly, we could not evaluate the Toolkit’s effect on provider behavior, a subject for future research.

The VA successfully developed an online peer-to-peer Toolkit to share QI tools across PCMHs nationwide. Other health systems may benefit from developing such online resources. Our findings suggest that they can enhance toolkit use by using systematic social marketing, providing standardized training and opportunities for social interaction, and emphasizing performance measurement. System-level and local leadership, as well as adequate time and resources for clinicians to implement PCMH, can also help overcome barriers to toolkit use.