Community behavioral health clinics provide critical services and function as safety-net providers for people with mental and substance use disorders. States have increasingly relied on these clinics to help people live independently in the community and avoid institutional care (
1,
2). There are nearly 2,600 community mental health centers and 5,600 specialty outpatient mental health clinics across the United States (
3). In the absence of federal licensing or accreditation standards, these clinics vary in the types of services they provide and populations they serve, which may also reflect the local workforce and the ability of these clinics to leverage different funding streams (
4–
6). Medicaid is an important funding source for these clinics, but Medicaid payment rates have not historically covered the full costs of the services that these clinics provide (
7,
8). As a result, clinics have turned to a patchwork of state and local funding and philanthropy to supplement costs for Medicaid beneficiaries and people without insurance (
9). Such variation in funding and the services available from these clinics could contribute to disparate access to care across communities (
10,
11).
Section 223 of the Protecting Access to Medicare Act authorized the Certified Community Behavioral Health Clinic (CCBHC) demonstration, which allows states to test a new strategy to deliver and reimburse services in behavioral health clinics (
12). Participating states certify that clinics provide a standard set of comprehensive ambulatory mental health and substance use services, crisis services, primary care screening and monitoring, and care coordination to adults and children (
12). Participation in the demonstration also requires CCBHCs to maintain relationships with hospitals and a wide range of other providers to coordinate care. They must also implement other activities to increase access to care, such as offering same-day appointments and conducting outreach to underserved populations.
The CCBHC demonstration established a new Medicaid prospective payment system designed to cover the full costs of CCBHC services. In this system, state Medicaid programs elect to reimburse all CCBHCs in the state by using either a fixed daily or a monthly rate for each day or month, respectively, that a Medicaid beneficiary receives care from a CCBHC. The CCBHC receives the same daily or monthly payment regardless of the number or type of services provided during a visit or month. This reimbursement mechanism gives clinics some flexibility to tailor services to the beneficiary without being concerned about the financial impact of every encounter or procedure. CCBHCs also report a common set of quality measures, and states can award quality bonus payments based on measure performance. Congress initially authorized the demonstration for 2 years, and eight states began implementing the CCBHC model in mid-2017. Congress has extended the demonstration and allowed additional states to participate.
In this study, we examined how the demonstration affected emergency department (ED) visits and hospitalizations during the first 2 years in Missouri, Oklahoma, and Pennsylvania. The CCBHC model could increase access to comprehensive services, thereby helping people avoid EDs and hospitalizations. We hypothesized that Medicaid beneficiaries who received care from CCBHCs would have fewer ED visits and hospitalizations, compared with Medicaid beneficiaries who received care from other community behavioral health clinics.
Methods
Data
We selected three of the original demonstration states to reflect different geographic areas and CCBHC payment models. Missouri and Pennsylvania used a daily rate to reimburse CCBHCs, whereas Oklahoma used a monthly rate (
13). We limited the analysis to these states because of resource constraints and concerns about the usability of Medicaid data from some other states. At the beginning of the study period, there were 15 CCBHCs in Missouri, which served 78% of the counties; three CCBHCs in Oklahoma, serving 22% of the counties; and seven CCBHCs in Pennsylvania, serving 10% of the counties. We established a data use agreement with each state to obtain Medicaid fee-for-service claims and managed care encounter data covering calendar years 2015 through 2019. The data included inpatient, ED, and ambulatory claims for adults and children-adolescents who received care from CCBHCs and other community behavioral health clinics for a 2-year period before the demonstration (mid-2015–mid-2017) and for the first 2 years of the demonstration (mid-2017–mid-2019). The analysis was exempt from institutional review board approval.
Treatment and Comparison Groups
We conducted analyses separately for each state rather than pool data across states because of state differences in CCBHC payment models and Medicaid programs. In each state, the sample included beneficiaries who received care from a community behavioral health clinic (including those that became CCBHCs) at any time in the 24 months before the demonstration. Beneficiaries were assigned to the treatment group if their last visit to a behavioral health clinic before the demonstration was to a clinic that became a CCBHC. Beneficiaries were assigned to the comparison group if their last visit to a behavioral health clinic was a clinic that did not become a CCBHC, thus representing usual care. We defined the samples on the basis of where they received care before the demonstration to minimize bias that could be introduced by changes in CCBHC case mix during the demonstration. We excluded beneficiaries if they died before the demonstration or were dually enrolled in Medicare (because we did not have Medicare data) and beneficiaries who were not eligible for full Medicaid benefits or were not continuously enrolled in Medicaid for at least 6 months during the demonstration (a table in the online supplement to this article provides additional details). The compositions of the treatment and comparison groups did not differ whether we made the assignment on the basis of the clinic where the beneficiary received most behavioral health services in the 2 years before the demonstration or whether we made the assignment on the basis of the last visit. The treatment and comparison groups had <5% crossover during the demonstration, likely because these clinics typically serve different catchment areas.
Outcomes Measures and Variables Describing Baseline Characteristics
Outcome measures included ED visits that did not result in a hospitalization and hospitalizations (to avoid double counting hospitalizations and ED visits) per 1,000 beneficiaries. We categorized hospitalizations and ED visits as related to behavioral health if the claim included a principal behavioral health diagnosis (
ICD-9 and
ICD-10 codes for mental health and substance use diagnoses, excluding developmental and intellectual disabilities, autism, dementia, and Alzheimer’s disease). Claims without a principal behavioral health diagnosis were categorized as general medical health. The analysis included variables for demographic characteristics, Medicaid eligibility category, residence (urban, suburban, or rural), and health status in the 2-year predemonstration period (
14,
15). For Missouri and Pennsylvania, we included variables for enrollment into specific managed care plans; Oklahoma did not enroll beneficiaries in managed care.
Propensity Score Methods
The sample for propensity score analyses included 21,453 beneficiaries in Missouri, 36,866 in Oklahoma, and 186,414 in Pennsylvania. For Missouri and Oklahoma, the comparison group was smaller than the treatment group. As a result, we used the variables in
Table 1 to reweight the comparison group to resemble the treatment group on the basis of a beneficiary’s predicted probability of being assigned to treatment given the person’s observable predemonstration characteristics. In Missouri, the treatment and comparison groups generally had similar levels (e.g., rates of hospitalizations per 1,000 beneficiaries) and quarterly trends in outcomes in the baseline period before and after propensity score weighting. In Oklahoma, the groups had similar trends but levels differed for some outcomes. We prioritized parallel trends during the eight baseline quarters, because the parallel-trend assumption is critical for the difference-in-differences design and helps protect against regression to the mean (
16,
17), which we examined visually in graphs after weighting.
For Pennsylvania, we used an optimal-matching algorithm to develop matched sets of treatment and comparison group beneficiaries, because the potential comparison group was much larger than the treatment group. We exact-matched beneficiaries who were enrolled in the same managed care plan at the beginning of the demonstration to ensure that the treatment and comparison groups were drawn from the same regions and had access to the same provider networks. We gave each treatment group beneficiary a weight of 1 and each comparison beneficiary a weight equal to the ratio of treatment beneficiaries to comparison beneficiaries in the matched set. We assessed the quality of the weighted or matched samples on the basis of standardized differences in means (calculated as the ratio of the treatment-comparison difference and the treatment group SD), percentage difference in means, equivalence tests, and trend plots (
18).
Impact Analyses
We fit ordinary least-squares (OLS), difference-in-differences regression models with beneficiary fixed effects to estimate the impact of the demonstration on the number of hospitalizations and ED visits over the full 24-month demonstration period and for each of the two 12-month periods after the start of the demonstration. In our main analyses, we limited the baseline period to the 12 months preceding the demonstration. Standard errors (SEs) were adjusted for multiple observations of the same beneficiary to allow for serial correlation of the outcomes within individual beneficiaries over time. We weighted each observation by using the weights from the propensity score models and an eligibility weight that accounted for the number of months the beneficiary was enrolled in Medicaid in each observation period. A key assumption of this design is that the change in outcomes observed among those in the comparison group is what would have been observed in the treatment group in the absence of the demonstration. Consistent with other studies that have used claims data to examine the impacts of new service delivery models (
19,
20), we defined p≤0.10 as statistically significant at the outset of the study to avoid falsely concluding that the demonstration did not have effects. We also interpreted the findings in the context of the SEs and effect sizes.
We conducted two sensitivity tests. First, to determine whether the results were sensitive to outliers (that is, to a small number of beneficiaries with high service use), we truncated outcomes at the 98th percentile. Second, we implemented the models by using 24 months (instead of 12 months) of baseline data to examine whether the impact estimates changed when we accounted for longer predemonstration trends.
We considered using zero-inflated negative binomial (ZINB) models, given that hospitalizations and ED visits are not normally distributed. However, ZINB models did not accommodate beneficiary fixed effects to adjust for time-invariant beneficiary characteristics. Previous studies have found that ZINB models produced point estimates similar to those of OLS models but with less conservative SEs (
20). Our application of OLS is consistent with previous studies that measured impacts on hospitalizations and ED visits (
19–
22). However, we cannot rule out that ZINB models could have produced different findings.
Results
The treatment and comparison groups were well balanced after propensity score adjustment, with some exceptions. In Oklahoma, the treatment group included a larger share of beneficiaries with a mental health condition during the baseline period, relative to the comparison group (80% vs. 74%, standardized difference=0.16), mostly driven by a larger share of beneficiaries with depression (
Table 1). The treatment group in Oklahoma also had a larger share of beneficiaries with an ED visit during the baseline period, relative to the comparison group (69% vs. 63%, standardized difference=0.14).
We noted some differences across states in the characteristics of beneficiaries included in the final analytic samples after propensity score adjustment. For example, 55% of the overall sample in Missouri qualified for Medicaid on the basis of disability, compared with less than one-quarter of the samples in Oklahoma and Pennsylvania. The sample in Missouri was also, on average, slightly older (mean age=30 years), compared with the Oklahoma and Pennsylvania samples (mean age=24 years for both states). The racial composition of the samples also varied by state; approximately 80% of beneficiaries in Missouri were White, compared with 63% in Oklahoma and almost 50% in Pennsylvania. About 25% of beneficiaries in Oklahoma were in the “other” race category, compared with 4% in Missouri and 12% in Pennsylvania. Approximately 15% of beneficiaries in Pennsylvania had an opioid use disorder, compared with about 5% in the other two states. Across states, >80% of treatment and comparison group beneficiaries remained in the sample by month 19 of the 24-month demonstration period.
The demonstration had a statistically significant (p≤0.10) impact on the average number of behavioral health ED visits in Oklahoma and Pennsylvania but not in Missouri (
Table 2). Over the 24-month demonstration period, reductions of 11% and 13% were observed in Oklahoma and Pennsylvania, respectively, in the average number of behavioral health ED visits among those who received care from CCBHCs, relative to the comparison group (p=0.08 and p=0.09, respectively). No impact was observed on all-cause ED visits, likely because behavioral health ED visits represented only about 10% of ED visits in each state.
The demonstration did not have a statistically significant impact on the average number of all-cause, general medical health, or behavioral health hospitalizations in any state in our main analysis (
Table 3). However, findings from sensitivity analyses suggested that CCBHCs could have reduced hospitalizations. In Oklahoma, the demonstration was associated with an approximately 22% decrease in all-cause hospitalizations, behavioral health hospitalizations, and general medical health hospitalizations when we used a 2-year baseline period to account for longer predemonstration trends (p<0.05 for all analyses) (see table in
online supplement). In Pennsylvania, beneficiaries who received care from CCBHCs had a 10% decrease in all-cause hospitalizations, relative to the comparison group (p=0.06), but only when we truncated the number of all-cause hospitalizations at the 98th percentile.
Discussion
In this study, we examined the impacts of the CCBHC demonstration on ED visits and hospitalizations in Missouri, Oklahoma, and Pennsylvania. In Oklahoma and Pennsylvania, Medicaid beneficiaries who received care from CCBHCs had a greater reduction in behavioral health ED visits, compared with those who received care from other behavioral health clinics in these two states. Findings from sensitivity analyses suggested that CCBHCs could have reduced hospitalizations in the two states. Several features of the CCBHC model could have contributed to these findings, including requirements for CCBHCs to provide access to comprehensive behavioral health and crisis services, peer support, and care coordination. CCBHCs also undertake other activities to increase access to care, such as offering same-day appointments and delivering care in locations beyond the clinic. The demonstration payment system does not require CCBHCs to bill for every procedure or type of service, which could have allowed them to tailor services to clients and help these individuals avoid seeking care from EDs and hospitals. However, daily and monthly billing processes impeded use of the claims data to determine whether the delivery of specific services contributed to these impacts. It is notable that the CCBHC demonstration in these two states achieved these impacts during the first demonstration year and despite early implementation challenges related to workforce shortages and changes in billing processes (
23). The CCBHCs were able to provide the required scope of services at the launch of the demonstration, which could have had an immediate impact on ED visits and hospitalizations. Detecting impacts on hospitalizations for the full CCBHC population could require a longer evaluation period, given that psychiatric hospitalizations are less common than ED visits (
24,
25).
Previous interventions to increase access to comprehensive care in community behavioral health clinics have typically been narrower in scope than the CCBHC demonstration and have yielded mixed findings. For example, some state Medicaid programs have implemented behavioral health home models that share some features of CCBHCs (
26); some studies of these models have reported positive impacts on all-cause (but not behavioral health) ED visits and no impacts on hospitalizations (
27), whereas other studies found positive impacts on hospitalizations (
28), likely reflecting differences in interventions, implementation contexts, and study periods. Likewise, studies have produced mixed findings on whether integrating primary care services into behavioral health clinics has an impact on ED visits and hospitalizations (
29,
30). These past efforts have generally focused on changing a limited number of care processes often for specific populations, supported by grant funding or modest changes in Medicaid payment rates (
31,
32). In contrast, the CCBHC demonstration requires clinic-wide implementation of a new model supported by a full redesign of the Medicaid payment system. This study provides the first insights into how this new model and payment system function together to affect ED visits and hospitalizations.
Readers should interpret the findings within the context of each state’s population and service delivery system characteristics, which we could not fully account for by using the available data. We did not design the study to directly compare states or draw conclusions about the best approach to implementing CCBHCs. As noted, there were some differences in each state’s implementation approach (for example, Missouri implemented CCBHCs to serve most counties, whereas the other two states implemented within regions) and populations, which could explain some differences in the findings across states. However, there were also evaluation design constraints (for example, fewer comparison group beneficiaries in Missouri relative to other states) that could explain these different findings across states.
This study had limitations. Although the study design allowed us to attribute impacts to the demonstration, it required limiting the analytic population to Medicaid beneficiaries who received care from these clinics before the demonstration. Therefore, the findings reflect the impacts among beneficiaries who were already receiving care from these clinics. Findings could differ among those who newly entered services during the demonstration. Future studies could use alternative designs to compare their results with these findings. In addition, although the treatment and comparison groups were comparable on observable characteristics, they could have differed on characteristics that are not measurable with Medicaid data. This may have been particularly relevant in Missouri, where the areas not affected by the demonstration were more limited than in other states. Small residual imbalances in observed characteristics, such as the higher prevalence of depression and higher baseline rate of ED visits in the treatment group in Oklahoma relative to the comparison group, could also have affected impact estimates. Finally, the findings reflect the first 2 years of the demonstration because this was the period initially authorized by Congress, and the analysis did not include all demonstration states.
Conclusions
The CCBHC demonstration reduced behavioral health ED visits among Medicaid beneficiaries, and there was some evidence of reductions in hospitalizations. Because the demonstration has continued beyond the initial 2 years and expanded to additional states, future research could examine impacts over a longer period and in other implementation contexts.
Acknowledgments
The authors thank Crystal Blyler, Ph.D., for providing feedback on earlier drafts of the manuscript; Rachel Hildrich Gross, B.S., Mark Lee, M.P.H., Sybil Pan, B.S., and McCayla Sica, B.S., for providing Medicaid data acquisition and management support; and Harold Alan Pincus, M.D., for consultation on the study design and interim findings. They also appreciate the assistance of the state Medicaid staff, who provided data for the study and guidance on the use of the data.