Advocates for primary care reform have promoted patient-centered medical home (PCMH) models as a way to strengthen primary care. PCMH structures and processes, which are intended to help practices deliver patient-centered, proactive, coordinated care can be coupled with financial support and accountability for a defined population of patients.1 PCMHs are intended to transform the business model of primary care while improving quality of care and reducing costs, largely by reducing emergency department and hospital utilization. Medicare, Medicaid and commercial payers have launched dozens of PCMH pilots across the nation.2,3

Two recent systematic reviews found mixed evidence that PCMH initiatives affect utilization, quality, and costs within two years.4,5 Many studies have found evidence of improved quality based on selected process measures and reduction in emergency department visits, but the evidence for cost savings is limited. A study of Geisinger Health System’s PCMH pilot found an 18 % reduction in hospital admissions after two years of a PCMH initiative.6 Similarly, Group Health of Puget Sound’s pilot medical home yielded 6 % fewer hospitalizations compared with other Group Health clinics in the first year and savings of $10.30 per patient per month in the second year.7

While some recent evaluations have not identified significant effects of PCMH initiatives on health care utilization and cost,810 the heterogeneity of PCMH interventions and contexts may lead to differential effectiveness, the patterns of which could illuminate facilitators and barriers to effective transformation.11 In addition, few studies of PCMH pilots have evaluated outcomes beyond two years, even though the literature suggests that practice transformation takes time to unfold.8 We examined the cumulative effects of a PCMH intervention after two and three years, and endeavored to document details of both the initiative and context to enable comparison with similar studies.

METHODS

The institutional review board at the Harvard School of Public Health approved the study.

Study Setting

We examined changes in patient care following initiation of a PCMH pilot in the Front Range area of Colorado, convened and supported by HealthTeamWorks, a non-profit organization that supports practices in implementing continuous quality improvement using a systems approach. The HealthTeamWorks pilot was launched in April 2009 with support from five commercial insurers (Aetna, Anthem, Cigna, Humana, and United HealthCare), one high-risk plan, which predominantly includes patients with preexisting conditions (CoverColorado), and the state Medicaid agency.

The HealthTeamWorks pilot included 15 practices involving 51 physicians, 35 allied health professionals (NPs, PAs) and 205 staff collectively serving approximately 98,000 patients. All 15 pilot sites were small to medium-sized practices (nine or fewer physicians). The National Committee for Quality Assurance (NCQA) certifies PCMHs at three accreditation levels, depending on to what degree that practice meets PCMH benchmarks and capabilities, with Level 3 being the highest level of accreditation.12 At baseline, NCQA certified three pilot practices as Level-2 PCMH practices and 12 as Level-3 PCMHs.

Participating health plans paid the PCMH practices $5.3 million over three years in per-member per-month fees based on the level of NCQA accreditation that each practice attained and $720,000 from a related pay-for-performance program, which awarded bonuses to practices based on meeting both quality and utilization benchmarks. The average payment per practice per year was approximately $118,000, or $34,700 per primary care physician. For pay for performance, quality was measured using medical record and registry data for diabetes, depression screening, and tobacco counseling. Utilization targets included reductions in emergency department visits, hospital admissions and increases in generic medication use. The practices received technical assistance, including monthly in-office coaching, learning collaborative sessions three times per year, monthly webinars and innovative technology interventions from HealthTeamWorks outlined in Appendix Figure 1.

One distinguishing feature of the HealthTeamWorks pilot was its deliberate focus on strategies to build “medical neighborhoods” and improve coordination of care. These include efforts such as systems for test and referral tracking to improve information flow, new staff positions to coordinate care, and care compacts with specialists to define shared responsibilities and accountability for coordinated patient care.13

Data Collection

To evaluate changes in patient care after the HealthTeamWorks pilot, we used a difference-in-difference approach. The pre-intervention period began 1 April 2007 and ended 31 March 2009. We examined the post-intervention period at two intervals: two years after initiation of the pilot (1 April 2009 through 31 March 2011) and three years after initiation of the pilot (1 April 2009 through 31 March 2012).

The primary data source for our analysis was administrative claims data from participating health plans. Two health plans (Humana and Cigna) were unable to provide usable claims data. We excluded Colorado Medicaid because of differences in eligibility, and patient population. We hypothesized that the intervention would reduce rates of hospital and emergency department admissions and improve quality of care. We expected that better identification and management of chronically ill patients would result in increased use of prescription drugs. We did not have a fixed hypothesis about primary care and specialist visits because we expected that greater access overall and to more non-visit-based care could yield either positive or negative effects on face-to-face visits with either type of physician. Hospital admissions, emergency department visits, primary care and specialist visits, and prescriptions were measured using standard definitions.14 To examine ambulatory care sensitive hospital admissions and ER visits, we used previously developed algorithms and used the Elixhauser index to stratify the population on health status.1517 We examined changes in quality on the basis of an adaptation of NCQA's Health Effectiveness Data and Information Set (HEDIS) measures for colon, breast, and cervical cancer, as well as three measures of diabetes care quality.

To quantify the change in total costs of care following the HealthTeamWorks PCMH intervention, we took the payer perspective and generated a standardized measure of cost by re-pricing the study claims according to an estimate of the average commercial fees for each service. The pricing data and algorithm were provided under license by OptumHealth.

We replicated all of our claims-data analyses for the subgroup of enrollees with two or more comorbidities identified using the Elixhauser algorithm. These patients represent approximately the top quintile of our overall sample in terms of comorbidities, have more frequent acute care utilizations, and we hypothesized that they could have more to gain from a PCMH intervention.

Quantitative Analysis

We attributed each patient to the practice that provided the plurality of their primary care and repeated this attribution annually. If a patient was enrolled in a health plan but had no primary care visits in a given year, we used their previous year’s primary care practice for attribution. We excluded patients older than 65 years since Medicare coverage, out of concern that we would not observe complete data where Medicare was a secondary payer. To be included in any year of the analysis, patients were required to have a minimum of six months of insurance eligibility in the year.

For each practice in the PCMH pilot, we identified comparison practices in the same geographic region through propensity score matching using the claims data from the baseline period (1 April 1 2007 through 31 March 2009).17 In the propensity score model, we included the Elixhauser comorbidity index, the number of patients from the high-risk CoverColorado plan, and practice-level average rates of inpatient admissions, emergency department visits, and primary care visits. Matching was based on a caliper width of 0.6 standard deviations of the logit transformation of the propensity score, with a maximum of ten matches for each pilot practice. We selected the caliper width to balance bias reduction through selecting closer matches against the desirability of increasing the number of matches to increase power.18

To validate our comparison cohort, we examined differences in practice-level patient characteristics and utilization patterns using Wilcoxon-Mann–Whitney statistics. Our difference-in-difference approach relies on the assumption that comparison practices’ performance is on a common trend with the pilot practices; hence, any post-intervention differences in pilot practices’ performance relative to that pre-intervention trend and concurrent comparison practice performance are attributable to the pilot program. A fundamental assumption is that pilot and comparison practices’ performance would exhibit a similar trend absent the pilot program. Therefore, we further tested for differences in the quarterly trends in our utilization and quality measures before the pilot program.

To adjust for residual confounding after matching, we estimated the effect of the intervention on utilization using generalized estimating equations (GEE) models assuming negative binomial or logistic distributions with a log link, and then adjusted for patient level clustering. The dependent variables in these models were constructed as counts per quarter (per year for quality measures), and the independent variables were patient age, sex, the Elixhauser comorbidity index, an indicator for whether the patient’s primary care physician was in the pilot program, number of months of eligibility, a linear time trend to account for secular changes in utilization, an indicator for the post-intervention period, and an interaction between the pilot and post-intervention practices. Using model predictions based on simulating alternative scenarios, we calculated difference-in-difference estimates. We then bootstrapped the difference-in-difference to obtain standard errors. Bootstrapping was required due to the non-linearity of the model and our interest in the magnitude of the transformed interaction effect.19 We used a standard two-part GEE model for cost analyses.20 The two-part model breaks down the estimation into the probability of having any cost and, for those with costs greater than zero, a model of the level of costs. In the first part where the dependent variable is whether there is any cost greater than zero, we used a binomial distribution with a logit link. For the second part, where the dependent variable is the level of cost for those with non-zero values, we used a gamma distribution with log link. For the subgroup of patients with two or more comorbidities, one-part linear models were estimated because the vast majority of observations had costs greater than zero and two-part models failed to converge. All estimates are reported both in terms of the absolute and relative (i.e., percentage) changes in the original units of measurement from pilot baseline levels for all patients and for those with two or more comorbidities separately. p < 0.05 was considered statistically significant.

RESULTS

Practice Characteristics

There were no differences between the pilot and the 66 comparison practices on most baseline characteristics with the exception of percent female, emergency department costs, and Hemoglobin A1C (Hba1c) testing for patients with diabetes (Tables 1 & 2). There were no differences in quarterly trends on utilization measures prior to the intervention (data not shown). For lipid testing and breast cancer screening, however, pilot practices were improving less rapidly than comparison practices (data not shown).

Table 1 Sample Characteristics of Pilot and Comparison Practices in the Baseline Period for the Full Analytic Sample (1 April 2007 – 31 March 2009)
Table 2 Sample Characteristics of Pilot and Comparison Practices in the Baseline Period for Patients with Two or More Comorbidities (1 April 2007 – 31 March 2009)

Utilization

Table 3 provides difference-in-difference estimates for utilization and costs for each time period relative to the pre-intervention levels, first for all patients and then for the sub-sample of patients with two or more comorbidities. All estimates should be interpreted as the change in pilot practices’ performance relative to the change in comparison practices’ performance. For ease of exposition, we describe these differential changes as reductions or increases, but they should be interpreted as changes relative to the trend in comparison practices over the same period.

Table 3 Changes in Health Care Utilization and Cost in Pilot Practices Relative to Comparison Practices

At the end of two years, HealthTeamWorks pilot practices reduced their patients’ use of the emergency department by 1.4 visits per 1000 member months, or approximately 7.9 % (p = 0.02). At the end of three years, pilot practices sustained this difference with 1.6 fewer emergency department visits per 1000 member months, or a 9.3 % reduction from baseline (p = 0.01). At three years, there was a reduction of 4.2 primary care visits per 1000 member months in the pilot practices, or approximately 1.5 % (p = 0.02). Among patients with two or more comorbidities, there was a reduction from baseline in primary care visits at both two years (a net effect of 12 visits per 1000 member months or a 2.7 % reduction, p = 0.006) and at three years (a net effect of eight visits per 1000 member months or a 1.8 % reduction, p = 0.04). For patients with two or more comorbidities, at three years there was a reduction of 0.9 ambulatory care sensitive inpatient admissions per 1000 member months (10.3 %; p = 0.05).

Cost

Among all patients, after two years we found a reduction in emergency department costs of $4.11 per member per month (a 13.9 % reduction, p < 0.0001). After three years, the reduction from baseline in emergency department costs was sustained at $3.50 per member per month (a 11.8 % reduction, p = 0.001) (Table 3). For patients with two or more comorbidities, we found a reduction in emergency department costs of $11.54 per member per month after two years and $6.61 per member per month after three years (25.2 % and 14.5 % reduction from baseline, respectively, after two and three years, p < 0.0001 and p = 0.003).

Clinical Quality

Table 4 presents the results for clinical quality indicators for both the full sample and for patients with two or more comorbidities. Participation in the HealthTeamWorks pilot was associated with improved cervical cancer screening after two years by 4.7 % (a 12.5 % relative improvement; p < 0.001) and after three years by 3.3 % (a 9.0 % relative improvement, p < 0.001). Conversely, colon cancer screening declined in intervention practices (a 21.1 % decrease, p = 0.01). For patients with two or more comorbidities, at the end of two years, there were two improvements: a 6.2 % increase in cervical cancer screening (a 16.2 % relative improvement, p < 0.001) and a 3.5 % increase in breast cancer screening (a 6.8 % relative improvement, p = 0.01). After three years, the significant improvements in both areas were smaller but sustained (cervical cancer screening: an increase of 4.4 %, an 11.5 % relative improvement, p = 0.001; breast cancer: an increase of 2.6 % , a 5.1 % relative improvement, p = 0.03). In contrast, colon cancer screening in patients with two or more comorbidities decreased after two years (a 23.0 % reduction, p = 0.005) and three years (a 20.3 % reduction, p = 0.02).

Table 4 Changes in the Quality of Care in Pilot Practices Relative to Comparison Practices

DISCUSSION

The HealthTeamWorks PCMH pilot was one of the first, and at the time largest in terms of participating plans and practices, multi-payer medical home initiatives in the United States.. Our analysis suggests that the HealthTeamWorks PCMH pilot was associated with meaningful reductions in emergency department utilization that were sustained into the third year of the pilot. Holding all other costs equal, the reductions in emergency department costs we identified translate into nearly $5 million per year in savings for the approximately 100,000 patients touched by the pilot. Notably, however, we did not find overall cost savings, perhaps because of offsetting increases in other categories of spending. We also found a reduction in ambulatory care sensitive inpatient hospital admissions for patients with two or more comorbidities, suggesting that some PCMH interventions may be able to deliver on the promise to reduce hospital use by patients with chronic illness.

Alongside acute care reductions, we found small but significant reductions in primary care visits. The mechanism that produced these reductions cannot be determined from our data. One possibility that is consistent with our logic model for the PCMH is that telephone consultations, email communication and group visits were provided as substitute for primary care visits.21 Another element of the PCMH involves more systematic care planning, which could improve the effectiveness of visits, reducing the need for return visits. Future studies should be designed to examine these substitution possibilities, which might offer some hope for more efficient use of an already strained workforce.

This PCMH intervention produced variable effects on measures of clinical quality; however, our study cannot definitively detect what mechanism may have led to this change. It is possible that the effort practices invested to make the changes promoted by the PCMH intervention distracted them from some screening programs. The reductions in the number of face-to-face primary care visits we observed may also have led to fewer opportunities to counsel patients and order screening tests. If a PCMH intervention reduces visits, practices may need to develop other methods for prompting screening.

It is notable that reductions in utilization and costs were observed for the attributed patient population as a whole, as well as for patients with two or more comorbidities. The fact that these benefits were sustained into the third year of the initiative and a new reduction in ambulatory care sensitive admissions accrued at this stage bodes well for the long-term prospects of PCMH interventions in Colorado and elsewhere. With a sample of one PCMH pilot in this study, we cannot tease out why the HealthTeamWorks was able to produce reductions in acute care use and costs when other PCMH pilots have failed to do so. As experience with PCMH initiatives accumulates researchers can contribute to policy and practice by identifying which design features (e.g., on-site coaching) or market characteristics are associated with greater improvements in patient care.

Our analysis is subject to several limitations. First, the HealthTeamWorks intervention was introduced in a small number of practices, limiting our statistical power. We were unable to obtain complete data from all payers involved, which reduced our sample size and thus power. Second, practices volunteered and were subsequently selected for the pilot, and despite statistical matching, we cannot exclude the possibility of selection bias in our results. Third, our analysis of quality effects is limited in that we did not have clinical data or measures of health outcomes. Fourth, we rely on a quasi-experimental approach to identifying the potential effects of the HealthTeamWorks pilot, and our findings may be sensitive to untestable assumptions about the comparability of our control practices to the HealthTeamWorks practices. Comparisons of baseline demographic and utilization means and pre-intervention trends revealed few significant differences, but our results should be interpreted in light of those differences. Finally, the HealthTeamWorks PCMH pilot was a complex intervention involving new payments, quality incentives, technical assistance and a collaborative approach to quality improvement. Our analysis is unable to identify the extent to which those components or other factors contributed the outcomes we observed.

Interpretation of our results should take into account differences in pilot and matched comparison practices. In the pre-intervention period, pilot practices had significantly lower percentages of female patients in both samples, and for the overall sample, higher per-member per-month emergency department costs. Moreover, prior to the intervention trend tests indicated that lipid testing and breast cancer testing were not improving as rapidly in the pilot practices as in the comparison practices. While levels at baseline were not different, these trend differences could have biased our findings towards the null.

Evaluations of PCMH interventions have produced mixed results on cost and quality measures. Possible explanations for these results include the lack of comparison groups and small sample sizes in some studies, and differences between the design features of PCMH interventions may have contributed to both positive and negative findings in this literature. Our study suggests that a PCMH intervention can produce reductions in acute care utilization and enhance some aspects of the quality of care, while potentially reducing others.

Both in Colorado and nationally, PCMH recognition continues to grow, and a wide variety of payers use PCMH status as a criterion for payment.22 In addition, some Accountable Care Organizations consider PCMH-model primary care as a foundation for their efforts to manage population health under a global budget.23 Given the momentum to enhance and extend PCMH models and interventions, evaluating whether populations are also healthier as a result will be an important future endeavor.