Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Effects of quality-based procedure hospital funding reform in Ontario, Canada: An interrupted time series study

  • Alvin Ho-ting Li ,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Project administration, Software, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    ali66@uwo.ca

    Affiliations Ottawa Hospital Research Institute, Ottawa, Ontario, Canada, Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada

  • Karen S. Palmer,

    Roles Conceptualization, Formal analysis, Investigation, Project administration, Supervision, Writing – original draft, Writing – review & editing

    Affiliations Women’s College Research Institute, Women’s College Hospital, Toronto, Ontario, Canada, Faculty of Health Sciences, Simon Fraser University, Burnaby, British Columbia, Canada

  • Monica Taljaard,

    Roles Formal analysis, Methodology, Writing – review & editing

    Affiliations Ottawa Hospital Research Institute, Ottawa, Ontario, Canada, School of Epidemiology and Public Health, University of Ottawa, Ottawa, Ontario, Canada

  • J. Michael Paterson,

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Software, Validation, Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada

  • Adalsteinn Brown,

    Roles Conceptualization, Funding acquisition, Methodology, Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada, Dalla Lana School of Public Health, University of Toronto, Toronto, Ontario, Canada, Department of Obstetrics and Gynecology, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada

  • Anjie Huang,

    Roles Data curation, Formal analysis, Investigation, Software, Writing – review & editing

    Affiliation Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada

  • Husayn Marani,

    Roles Writing – review & editing

    Affiliations Women’s College Research Institute, Women’s College Hospital, Toronto, Ontario, Canada, Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada

  • Lauren Lapointe-Shaw,

    Roles Formal analysis, Writing – review & editing

    Affiliation Department of Medicine, University of Toronto, Toronto, Ontario, Canada

  • Daniel Pincus,

    Roles Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada, Division of Orthopaedic Surgery, Department of Surgery, University of Toronto, Toronto, Canada

  • Marian S. Wettstein,

    Roles Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada, Department of Surgical Oncology (Urology), Princess Margaret Cancer Centre, University Health Network, University of Toronto, Toronto, Ontario, Canada

  • Girish S. Kulkarni,

    Roles Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Department of Surgical Oncology (Urology), Princess Margaret Cancer Centre, University Health Network, University of Toronto, Toronto, Ontario, Canada, Division of Urology, Department of Surgery, Faculty of Medicine, University of Toronto, Toronto, Ontario, Canada

  • David Wasserstein,

    Roles Writing – review & editing

    Affiliations Division of Orthopaedic Surgery, Department of Surgery, University of Toronto, Toronto, Canada, Sunnybrook Research Institute, Sunnybrook Health Sciences Centre, Toronto, Ontario, Canada

  • Noah Ivers

    Roles Conceptualization, Formal analysis, Funding acquisition, Investigation, Methodology, Project administration, Supervision, Validation, Writing – review & editing

    Affiliations Institute for Clinical Evaluative Sciences, Toronto, Ontario, Canada, Women’s College Research Institute, Women’s College Hospital, Toronto, Ontario, Canada, Department of Family and Community Medicine and Institute of Health Policy, Management and Evaluation, University of Toronto, Toronto, Ontario, Canada

Abstract

Background

The Government of Ontario, Canada, announced hospital funding reforms in 2011, including Quality-based Procedures (QBPs) involving pre-set funds for managing patients with specific diagnoses/procedures. A key goal was to improve quality of care across the jurisdiction.

Methods

Interrupted time series evaluated the policy change, focusing on four QBPs (congestive heart failure, hip fracture surgery, pneumonia, prostate cancer surgery), on patients hospitalized 2010–2017. Outcomes included return to hospital or death within 30 days, acute length of stay (LOS), volume of admissions, and patient characteristics.

Results

At 2 years post-QBPs, the percentage of hip fracture patients who returned to hospital or died was 3.13% higher in absolute terms (95% CI: 0.37% to 5.89%) than if QBPs had not been introduced. There were no other statistically significant changes for return to hospital or death. For LOS, the only statistically significant change was an increase for prostate cancer surgery of 0.33 days (95% CI: 0.07 to 0.59). Volume increased for congestive heart failure admissions by 80 patients (95% CI: 2 to 159) and decreased for hip fracture surgery by 138 patients (95% CI: -183 to -93) but did not change for pneumonia or prostate cancer surgery. The percentage of patients who lived in the lowest neighborhood income quintile increased slightly for those diagnosed with congestive heart failure (1.89%; 95% CI: 0.51% to 3.27%) and decreased for those who underwent prostate cancer surgery (-2.08%; 95% CI: -3.74% to -0.43%).

Interpretation

This policy initiative involving a change to hospital funding for certain conditions was not associated with substantial, jurisdictional-level changes in access or quality.

Introduction

Policymakers worldwide are experimenting with hospital funding models to improve system performance [13]. Although such reforms may contribute to improvements in resource allocation and patient outcomes, they may also invoke unintended consequences [46].

In April 2011, the Government of Ontario, Canada, announced a multi-year phased-in implementation of “patient-based” hospital funding [7]. These hospital funding reforms reduced reliance on global hospital budgets (i.e., fixed annual amount based largely on historical spending) by introducing two new components to hospital funding: Health Based Allocation Model (HBAM), organizational-level funding based on service and patient characteristics; and Quality-Based Procedures (QBPs), a novel approach to hospital funding sharing some characteristics with activity-based funding (ABF) [8]. QBPs consist of pre-set reimbursement rates for managing patients with specific diagnoses or those undergoing specific procedures, coupled with best-practice clinical handbooks for each QBP [7]. Between April 2012-April 2016, 19 priority QBPs were implemented for a range of medical and surgical conditions [9].

The stated goal of QBPs was to “facilitate adoption of best clinical evidence-informed practices” and appropriately reduce “variation in costs and practice across the province while improving outcomes” [10]. The provincial government’s rationale and assumed mechanism of action for QBPs was as follows: "QBPs are specific clusters of patient services that offer opportunities for health care providers to share best practices and will allow the system to provide even better quality care, while increasing system efficiencies. By promoting the adoption of clinical evidence-informed practices, clinical practice variation should be reduced across the province while improving patient outcomes to ensure that patients receive the right care, in the right place, at the right time" [11].

Like all “patient focused” activity-based funding systems, QBPs established a prospective payment rate based on service type and volume. Funding was carved out of hospitals’ global budgets and then reallocated to hospitals at the start of the relevant fiscal year as a fixed fee and fixed volume, for each QBP procedure or diagnosis. The fixed volume of QBP-funded cases per hospital is based on historical volume levels at each hospital. The fixed fee is adjusted for each hospital based on its unique case-mix index (CMI) to account for the complexity in its overall patient population.

However, QBPs differ from most ABF reforms in that funding applies only to a very limited set of diagnoses and procedures, and they rely on the use of handbooks to encourage incorporation of best practices [1215]. To create these handbooks for each QBP, the Ministry of Health and Long Term Care, in collaboration with partners such as Health Quality Ontario, Cancer Care Ontario, and the Cardiac Care Network, established expert advisory panels with leading clinicians, scientists, and patients. They defined episodes of care for selected diagnoses or procedures, developed best practice recommendations for patient care, and suggested indicators to monitor for ongoing quality improvement. The resulting QBP Clinical Handbooks serve as a compendium of evidence and clinical consensus [11].

There is no mechanism in place to enforce adherence to the clinical pathways in the handbooks or to measure adherence; hospitals are paid via QBPs whether they follow the pathways or not, but the intent was that following the pathways would enable hospitals to deliver care for the amount paid by QBPs [7, 16].

To date, there has been no peer-reviewed evaluation of the overall effects of QBPs on key indicators of patient care. We took advantage of Ontario’s data infrastructure to evaluate this new hospital payment model, focusing on system-level changes in measures of quality of care, access to care, and hospital coding behaviour for four QBPs including planned and unplanned surgical procedures, and medical diagnoses, selected a priori by our research team: (1) congestive heart failure, (2) hip fracture, (3) pneumonia, and (4) prostate cancer surgery.

Methods

Setting, context, and design

Hospital-based care in Ontario, Canada is publicly-funded. Ontario’s 141 publicly-funded hospital corporations comprise 262 hospital sites [17], of which a majority receive QBP funding. Small hospitals (n = 55 with typically fewer than 2700 inpatients or day surgery cases per year in two of the last three years) and specialty hospitals—such as for mental health, children, chronic care, and rehabilitation—primarily receive funding through global budgets and have only implemented select QBPs (e.g. tonsillectomy) depending on their specific patient population (e.g. children). These hospitals are excluded from our analyses because they only perform a very small number, if any, of the diagnoses and/or procedures performed by QBP-funded hospitals [7].

Using population-based interrupted time series (ITS) analyses, and based on a dated pre-specified protocol and dataset creation plans held at ICES, we evaluated patients admitted to Ontario hospitals for four pre-specified QBPs. We selected these QBPs with input from health system decision makers and hospital leaders to represent a range of acute versus elective and surgical versus medical issues. This was further informed by our prior qualitative work which identified sources of potential variation in the extent to which, and the ways in which, hospitals responded to QBPs [7, 16]. We chose a priori to incorporate a 3 month transition period to allow time for any clinical changes in response to the funding model change to be implemented. We used an ITS design which is a robust quasi-experimental design that can be used to evaluate policy changes at the whole system- and population-level when randomization is infeasible. [1822] The study interval depended on data availability and varied by QBP: congestive heart failure (April 2010-February 2017) launched April 2013; hip fracture (April 2012-February 2017) launched April 2014; pneumonia (April 2012-February 2017) launched April 2014; and prostate cancer surgery (April 2010-February 2017) launched April 2015. We chose these intervals to ensure at least 24 monthly data points pre-policy.

Ethics approval

The use of data in this project was authorized under section 45 of Ontario’s Personal Health Information Protection Act, which does not require review by a research ethics board.

Study patients

We separately identified patients for each QBP cohort using inclusion and exclusion criteria detailed in each clinical handbook [1215]. In short, cohorts for congestive heart failure and pneumonia were defined using specific qualifying hospital discharge diagnoses; hip fracture, using a combination of discharge diagnoses and procedures; and prostate cancer surgery, using specific procedure codes. We considered only admissions to hospitals that received funding for one of the QBPs under evaluation. Episodes of care had to be separated by at least 30 days (to exclude 30-day hospital readmissions). We excluded patients without a valid Ontario Health Insurance Plan (OHIP) number who could not be accurately followed in our data sets, and patients with missing demographic information (<0.1%).

Data sources and quality measures

We used multiple linked health administrative databases to describe study patients and ascertain outcomes. Patient demographic information and vital status were obtained from the Registered Persons Database. Hospital diagnoses and procedures were obtained from the Canadian Institute for Health Information (CIHI) Discharge Abstract Database (CIHI-DAD). Emergency department admissions were captured using CIHI’s National Ambulatory Care Reporting System (NACRS).

We described patients according to age at hospital admission, sex, neighbourhood income quintile, rurality of residence, Deyo-Charlson Comorbidity Index [23], and number of emergency department visits and hospitalization days in the year preceding qualifying admission.

Outcomes were assessed for each QBP in three domains:

  1. Quality of care: i) death or return to hospital (i.e. unplanned presentation to emergency department or hospital admission within thirty days, among patients discharged alive and not transferred); ii) mean acute hospital length of stay (LOS); and iii) mean total LOS for entire episode of care including transfers;
  2. Access to care: i) total volume of admissions; ii) proportion of patients aged 65 years or older; and iii) proportion of patients living in lowest neighborhood income quintile;
  3. Coding behaviour: hospital discharge coding behavior as assessed by mean HBAM Inpatient Group (HIG) resource intensity weight. HIG weight is the Ontario-specific acute inpatient grouping methodology used to account for patients’ clinical- and resource-utilization characteristics [24].

We selected these measures because policymakers hoped that QBPs would reduce length of stay in settings where it was longer than optimal without decreasing quality (i.e., outcomes such as deaths, return to hospital, or inappropriate coding). The expectation was also that shorter lengths of stay, as typically seen in other countries implementing ABF-like reforms, would facilitate greater throughput to increase total patient volume across the system, and that access to care would not be compromised by inequity across age and income [8, 25]. Socioeconomic status (SES) may contribute to inequalities in access to care, so we used neighborhood income quintile is an indicator of SES [2628]. Prior research has shown that financial incentives associated with hospital funding reforms, such as QBPs, may alter coding behaviour to maximize reimbursement [2932]. HIG weight is a measure of coding behaviour because it incorporates both case mix and the resource intensity of each patient care episode adjusted for patient characteristics. If upcoding is occurring we would expect to see changes in HIG weight. Thus, to the extent changes in HIG weight do not represent true abrupt changes in patient case mix, the HIG weight is one potential measure by which to evaluate effects on coding.

Statistical analysis

For each outcome, we calculated monthly summaries, aggregated across hospitals (percent, mean, or raw count) and plotted them over time. For each QBP, we excluded three months of data following start of the funding change to account for a policy “transition” period [33]. We chose a three month transition a priori, postulating that it would take a fiscal quarter for any policy effects to occur. We accounted for seasonality by decomposing the data in trend, seasonal, and random components, and then removing the seasonal component [34].

We used segmented linear regression analysis of the seasonally-adjusted data. We used the forecast library in the R statistical software package to fit the model and used an automated stepwise selection procedure based on the Akaike Information Criterion (AIC) to include autoregressive terms accounting for the serial correlation [35, 36]. We used visual inspection of the observed and fitted data as well as residual plots to verify goodness of fit. Our model included fixed terms for pre-policy intercept and slope, intercept change at the time of policy (immediate difference in level following implementation of QBPs accounting for the postulated three-month transition period), and post-policy trend change (difference in the slope following implementation of QBPs). For the main results, we expressed the effect of QBP on each outcome as the counterfactual difference after 2 years, that is, the difference between the observed rate and the rate that would have occurred had QBPs been not implemented. This was estimated as the difference at 2 years post-implementation between the fitted post-implementation rates and the projected rates estimated from the pre-intervention intercept and slope. All analyses were performed at ICES (www.ices.on.ca) using linked, coded data. We used SAS v. 9.3 to prepare the monthly time series data for each outcome measure, and R version 3.4.4 to perform the regression analyses (nlme, car) and plot and compute the 95% confidence intervals for the counterfactual [37].

Results

Patient characteristics

S1S4 Tables describe the overall characteristics for each cohort. The patient characteristics remained largely unchanged throughout the study period.

Results from segmented regression analysis

The counterfactual estimates from the segmented regression analyses, representing the effect of QBPs on outcomes at 2 years post-implementation, are presented in Table 1. Figs 13 and S1S4 Figs present the observed data and fitted values from the segmented regression analyses. The full results from the segmented regression analyses are presented in S5 and S6 Tables.

thumbnail
Fig 1. Percent of patients returned to hospital or died.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.g001

thumbnail
Fig 2. Mean acute length of stay for the episode of care (days).

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.g002

thumbnail
Fig 3. Total volume of admissions.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.g003

thumbnail
Table 1. Estimated effect of implementation of QBPs on outcomes after 2 years, calculated as counterfactual difference from the segmented regression analysis (absolute difference, 95% confidence intervals).

https://doi.org/10.1371/journal.pone.0236480.t001

Quality of care.

At 2 years post-implementation, the estimated percentage of hip fracture patients who returned to hospital or died within 30 days was higher by an absolute 3.13% (95% CI: 0.37% to 5.89%) than if QBPs had not been introduced. There was no change in LOS for hip fracture patients in comparison to the counterfactual (Table 1). For prostate cancer surgery patients, the increase in mean acute LOS over the counterfactual was 0.33 days (95% CI: 0.07 to 0.59) and mean total LOS 0.34 days (95% CI: 0.06 to 0.61). There were no other statistically significant changes observed for the LOS outcome (Fig 2 and S1 Fig).

Access to care.

At 2 years post-implementation, the volume of patients admitted with congestive heart failure was higher by 80 patients (95% CI: 2 to 159) than if QBPs had not been introduced (Table 1). The volume of hip fracture admissions was lower by 138 patients (95% CI: -183 to -93). Percentage of admitted patients living in the lowest income quintile was higher (1.89%; 95% CI: 0.51% to 3.27%) for those diagnosed with congestive heart failure and decreased for those with prostate cancer surgery (-2.08%; 95% CI: -3.74% to -0.43%).

Hospital coding behaviour.

We observed no statistically significant changes in mean HIG weight for any of the four cohorts (S4 Fig).

Interpretation

Summary of findings

For the seven outcomes across the four diagnoses and/or procedures we studied, we compared the observed against expected counterfactual findings across outcomes in the domains of quality, access, and coding behaviour and found an inconsistent and generally weak response to the QBP funding reform at the system-level. In general, QBPs did not appear to result in changes to prevailing trends in return to hospital or death. Contrary to expectations, LOS increased slightly for prostate cancer surgery. Counterintuitively, despite no change in LOS, QBP funding was associated with a decrease in overall volume of admissions for hip fracture surgeries and a small absolute increase (of 3%) in the monthly percentage of hip fracture patients who returned to hospital or died within 30 days of discharge. We observed small increases in the percentage of patients admitted with congestive heart failure and decreases for prostate cancer surgery residing in the lowest neighborhood income quintile (<2%).

Explanation of findings

To our knowledge, there is no published quantitative research on the broad effects of the implementation of the QBP funding reform policy. Prior qualitative analyses of QBPs has revealed challenges associated with implementation of this complex hospital funding reform policy [7, 16, 38].

Previous studies found mixed reactions after other types of hospital funding reform [25, 39]. For example, evaluation of a limited experiment with activity-based funding in British Columbia, Canada, showed small decreases in volume, small increases in patients’ length of stay, and no changes in measures of quality (i.e., unplanned readmissions and in-hospital mortality) [25]. Conversely, a systematic review of ABF affirmed that transition to activity-based funding initially decreased length of stay in the US and internationally, and also found important policy- and clinically-relevant changes, including substantial increases in admissions to post-acute care following hospitalization [8].

Unintended consequences typically associated with ABF-like reforms include, for example, patients being discharged “sicker and quicker” to post-acute care facilities or home, and “upcoding”, which may be appropriate if it represents more accurate coding, or inappropriate [8]. We did not see the decrease in LOS that might have been precipitated by accelerated discharge. Nor did we observe changes in coding behaviour for the QBPs studied.

The slight variation we observed in effects—some positive, others negative—may be partly explained by our prior qualitative work, in which we observed variation in response to the reform related to complexity of changes required, internal capacity for organizational change, and availability and appropriateness of supports to manage change [16]. It may also simply represent noise rather than signal. Our goal in this paper was not to understand quantitatively the variation in responsiveness by hospital, but to evaluate the system-level effects of the jurisdictional policy change, as this is the level at which ‘success’ of the policy reform must ultimately be judged.

The lack of large-scale meaningful changes in association with Ontario’s shift to QBP funding is perhaps not surprising. Funding reforms may not be necessary or effective when desirable changes are already occurring. For example, hospitals were already under long-standing pressures to reduce length of stay and may have reached a floor, which may explain why further financial pressure from QBPs had little effect. Similarly, hospitals may have also lacked effective incentives or supports to address readmissions, since, unlike activity-based funding reforms in other countries, Ontario’s QBP funding reform did not financially disincentivize return to hospital nor link funding to care outcomes.

Limitations

Our study has several important limitations that are common to observational studies of policy changes. Teasing apart the effect of QBPs in the presence of multiple system-level changes is challenging. First, other initiatives to improve patient care and/or control costs may have overlapped with the timing of QBP implementation. Specific initiatives that we are aware of included passage in 2010 of the Excellent Care for All Act (ECFA) [40], the introduction of Health Based Allocation Model hospital funding reforms in April 2012 [24], and the introduction of Community Health Links in December 2012 [41] (S5S7 Figs, S7 Table). Visual inspection of data points around the timing of introduction of these initiatives however suggests that they are unlikely to have had a major impact on the outcomes studied in our analyses. A possible exception is the introduction of Community Health Links in December 2012: due to its timing close to that of the QBPs for congestive heart failure (CHF) in April 2013, it is difficult to independently assess the effect of the QBPs for this condition. Undetected confounding is always possible in any uncontrolled study. Policies aimed at improving health care are constantly being tinkered with, which may influence any particular intervention, such as QBPs, in ways not easily detected. Second, given the nearly ubiquitous implementation of QBP in Ontario, we did not identify suitable contemporaneous comparators in this study, which could have strengthened the inferences drawn. Although one of the ways to optimize an ITS is to add negative controls, we did not add these because the goal of the reform was to effect broad change across the entire system, resulting in only a small number of unique and, therefore, non-comparable hospitals being exempted from implementing QBPs (i.e., very small hospitals with few beds and/or those with unique targeted populations). Third, because our analyses were restricted to QBP-funded hospitals rather than all hospitals, we cannot be certain that our results are generalizable to the whole system; however, the proportion of QBP procedures occurring outside of QBP-funded hospitals is low (<11%). Fourth, examining a broader range of outcomes (e.g. extent to which patient care is aligned with evidence-based care processes described in QBP clinical pathways; reduction in inter-hospital variation in care, cost, and wait-times) may be more sensitive to, or reveal different effects of QBPs on, patient care and outcomes, care providers, and the health care system as a whole. Fifth, it is noteworthy that QBPs are unique to Ontario, making generalizability to other jurisdictions (both within and outside Canada) difficult to assess. Although somewhat similar in design to ABF reforms elsewhere, critical differences include the absence of financial disincentives for readmission with QBPs; a smaller and less ubiquitous funding scope limited to fewer priority diagnoses and procedures than ABF reforms elsewhere. However, this study is relevant to health system funding reforms that attempt to improve quality while also cutting costs, though contextual factors that influence the linkage between quality and cost are difficult to capture in relatively simple funding reforms [42]. Sixth, we did not assess how QBPs impacted hospitals’ finances, so we cannot make any inferences about whether increases or decreases in hospitals’ budgets affected patient care and/or outcomes for the QBPs we evaluated. Seventh, there may be benefits or harms of QBPs that we did not measure, or other policy objectives that may have been met, such as those related to total cost per episode-of-care or cost to the system overall. We were careful to limit our conclusions to only the QBPs and outcomes we evaluated to avoid being overbroad.

Conclusion

We found mixed and generally very small effects on quality of care, access to care, and coding behaviour, across the four QBPs we studied. We speculate that challenges with implementing the best practice pathways featured in the QBP handbooks, together with progressive controls on hospital expenditures, and a worsening overall fiscal picture in Ontario coincident with QBP implementation, may have led to inconsistent and weak signals. Further experimentation with funding reform as a potential mechanism to improve outcomes might yield greater impact if focused on specific diagnoses and procedures in which suboptimal process or outcome measures are well-established and for which efforts to improve outcomes by other means have been inadequate.

Supporting information

S1 Fig. Mean length of stay for entire episode (days).

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s001

(TIFF)

S2 Fig. Percent change in patients over 65 (%).

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s002

(TIFF)

S3 Fig. Percent change in patients living in the lowest neighborhood income quintile.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s003

(TIFF)

S4 Fig. Mean HIG weight.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s004

(TIFF)

S5 Fig. Percent of patients returned to hospital or died with competing initiatives.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Competing initiatives are outlined in the legend. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s005

(TIFF)

S6 Fig. Mean acute length of stay for the episode of care (days) with competing initiatives.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. Competing Interventions are outlined in the legend. The grey shaded area represents the three months of “transition” period. Competing Initiatives are outlined in the legend. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s006

(TIFF)

S7 Fig. Total volume of diagnoses/procedures with competing initiatives.

Red solid line represents the fitted model. The red dashed line represents the counterfactual (i.e. if no policy change occurred). The vertical dashed line represents the date of policy change. The grey shaded area represents the three months of “transition” period. Competing Initiatives are outlined in the legend. Data are seasonally adjusted.

https://doi.org/10.1371/journal.pone.0236480.s007

(TIFF)

S1 Table. Cohort characteristics for congestive heart failure patients included in the analysis.

https://doi.org/10.1371/journal.pone.0236480.s008

(DOCX)

S2 Table. Cohort characteristics for hip fracture patients included in the analysis.

https://doi.org/10.1371/journal.pone.0236480.s009

(DOCX)

S3 Table. Cohort characteristics for pneumonia patients included in the analysis.

https://doi.org/10.1371/journal.pone.0236480.s010

(DOCX)

S4 Table. Cohort characteristics for prostate cancer patients included in the analysis.

https://doi.org/10.1371/journal.pone.0236480.s011

(DOCX)

S5 Table. Estimated coefficients from the interrupted time series analysis on the impact of quality-based procedure policy (% patients returned to hospital or died, mean acute length of stay, total volume).

https://doi.org/10.1371/journal.pone.0236480.s012

(DOCX)

S6 Table. Estimated coefficients from the interrupted time series analysis on the impact of quality-based procedure policy (% change in patients over 65, % change in patients living in lowest neighborhood income quintile, mean HIG weight.

https://doi.org/10.1371/journal.pone.0236480.s013

(DOCX)

Acknowledgments

We thank Michael Law, University of British Columbia, for insightful comments on an earlier draft of the manuscript.

References

  1. 1. Damberg CL, Sorbero ME, Lovejoy SL, Martsolf G, Raaen L, Mandel D. Measuring Success in Health Care Value-Based Purchasing Programs: Findings from an Environmental Scan, Literature Review, and Expert Panel Discussions.: 242.
  2. 2. Mattison CA, Wilson MG. Rapid synthesis: Examining the effects of value-based physician payment models. Hamilton, Canada: McMaster Health Forum, 10 October 2017.
  3. 3. Sutherland J, Crump RT, Repin N, Hellsten E. Paying for Hospital Services: A Hard Look at the Options. SSRN Electronic Journal. 2013 [cited 13 Jun 2018].
  4. 4. Gu Q, Koenig L, Faerberg J, Steinberg CR, Vaz C, Wheatley MP. The Medicare Hospital Readmissions Reduction Program: Potential Unintended Consequences for Hospitals Serving Vulnerable Populations. Health Services Research. 2014;49: 818–837. pmid:24417309
  5. 5. Weeks WB, Rauh SS, Wadsworth EB, Weinstein JN. The Unintended Consequences of Bundled Payments. Annals of Internal Medicine. 2013;158: 62. pmid:23277901
  6. 6. Lipsitz LA. Understanding Health Care as a Complex System: The Foundation for Unintended Consequences. JAMA. 2012;308: 243–244. pmid:22797640
  7. 7. Palmer KS, Brown AD, Evans JM, Marani H, Russell KK, Martin D, et al. Qualitative analysis of the dynamics of policy design and implementation in hospital funding reform. PLOS ONE. 2018;13: e0191996. pmid:29373587
  8. 8. Palmer KS, Agoritsas T, Martin D, Scott T, Mulla SM, Miller AP, et al. Activity-Based Funding of Hospitals and Its Impact on Mortality, Readmission, Discharge Destination, Severity of Illness, and Volume of Care: A Systematic Review and Meta-Analysis. Jimenez-Soto E, editor. PLoS ONE. 2014;9: e109975. pmid:25347697
  9. 9. Government of Ontario, Ministry of Health and Long Term Care. Health System Funding Reform—Health Care Professionals—MOHLTC. [cited 16 Jul 2018]. http://www.health.gov.on.ca/en/pro/programs/ecfa/funding/hs_funding_qbp.aspx
  10. 10. Ontario Ministry of Health and Long-Term Care, Quality-Based Procedures Indicators, An Implementation Guidance Document, 2014. http://health.gov.on.ca/en/pro/programs/ecfa/docs/qbp_indicator_guidance_en.pdf
  11. 11. Ontario Hospital Association, Toolkit to Support Implementation of Quality-Based Procedures, ISBN # 978-0-88621-353-4, https://www.oha.com/Documents/QBP%20Toolkit.pdf
  12. 12. Health Quality Ontario, Ministry of Health and Long-Term Care. Quality-based procedures: Clinical handbook for community-acquired pneumonia. Toronto, ON: Health Quality Ontario; 2014 Feb p. 67. www.hqontario.ca/evidence/evidence-process/episodes-of-care#community-acquired-pneumonia
  13. 13. Health Quality Ontario, Ministry of Health and Long-Term Care. Quality-Based Procedures: Clinical Handbook for Hip Fracture. Toronto, ON: Health Quality Ontario; 2013 May p. 97. http://www.health.gov.on.ca/en/pro/programs/ecfa/docs/qbp_hipfracture.pdf
  14. 14. Health Quality Ontario, Ministry of Health and Long-Term Care. Quality-based procedures: clinical handbook for heart failure (acute and postacute). Toronto, ON: Health Quality Ontario; 2015 Feb p. 78. http://www.health.gov.on.ca/en/pro/programs/ecfa/docs/qbp_heart.pdf
  15. 15. Health Quality Ontario, Ministry of Health and Long-Term Care. Quality-Based Procedures Clinical Handbook for Cancer Surgery. Toronto, ON: Health Quality Ontario; 2016 Jan p. 92. http://www.health.gov.on.ca/en/pro/programs/ecfa/docs/qbp_cancer_surgery.pdf
  16. 16. Palmer KS, Brown AD, Evans JM, Marani H, Russell KK, Martin D, et al. Standardising costs or standardising care? Qualitative evaluation of the implementation and impact of a hospital funding reform in Ontario, Canada. Health Research Policy and Systems. 2018;16: 74. pmid:30075735
  17. 17. Ontario Hospital Association—Your Hospitals. [cited 23 Jun 2018]. https://www.oha.com/your-hospitals
  18. 18. Fine B, Schultz SE, White L, Henry D. Impact of restricting diagnostic imaging reimbursement for uncomplicated low back pain in Ontario: a population-based interrupted time series analysis. cmajo. 2017;5: E760–E767. pmid:29042408
  19. 19. Hardy G, Colas JA, Weiss D, Millar D, Forster A, Walker M, et al. Effect of an innovative community-based care model, the Monarch Centre, on postpartum length of stay: an interrupted time-series study. CMAJ Open. 2018;6: E261–E268. pmid:30012645
  20. 20. Jandoc R, Burden AM, Mamdani M, Lévesque LE, Cadarette SM. Interrupted time series analysis in drug utilization research is increasing: systematic review and recommendations. Journal of Clinical Epidemiology. 2015;68: 950–956. pmid:25890805
  21. 21. Kontopantelis E, Doran T, Springate DA, Buchan I, Reeves D. Regression based quasi-experimental approach when randomisation is not an option: interrupted time series analysis. BMJ. 2015;350: h2750. pmid:26058820
  22. 22. Rudoler D, de Oliveira C, Cheng J, Kurdyak P. Payment incentives for community-based psychiatric care in Ontario, Canada. CMAJ. 2017;189: E1509–E1516. pmid:29229712
  23. 23. Charlson M, Szatrowski TP, Peterson J, Gold J. Validation of a combined comorbidity index. J Clin Epidemiol. 1994;47: 1245–1251. pmid:7722560
  24. 24. Health Based Allocation Model: HBAM Inpatient Group Methodology and Reports at CIHI. https://www.oha.com/Documents/HBAM-What%20You%20Need%20To%20Know.pdf
  25. 25. Sutherland JM, Liu G, Crump RT, Law M. Paying for volume: British Columbia’s experiment with funding hospitals based on activity. Health Policy. 2016;120: 1322–1328.
  26. 26. Moscelli G, Siciliani L, Gutacker N, Cookson R. Socioeconomic inequality of access to healthcare: Does choice explain the gradient? Journal of Health Economics. 2018;57: 290–314. pmid:28935158
  27. 27. Health Quality Ontario. Income and health: opportunities to achieve health equity in Ontario. 2016. https://www.hqontario.ca/Portals/0/documents/system-performance/health-equity-report-en.pdf
  28. 28. Canadian Institute for Health Information, Trends in Income-Related Health Inequalitites in Canada, Technical Report, Revised July 2016. Ottawa, ON: CIHI 2016. https://secure.cihi.ca/free_products/trends_in_income_related_inequalities_in_canada_2015_en.pdf
  29. 29. Seiber EE. Physician Code Creep: Evidence in Medicaid and State Employee Health Insurance Billing. Health Care Financ Rev. 2007;28: 83–93. pmid:17722753
  30. 30. Abler S, Verde P, Stannigel H, Mayatepek E, Hoehn T. Effect of the introduction of diagnosis related group systems on the distribution of admission weights in very low birthweight infants. Arch Dis Child Fetal Neonatal Ed. 2011;96: F186–189. pmid:21037282
  31. 31. Helms CM. A pseudo-epidemic of septicemia among Medicare patients in Iowa. Am J Public Health. 1987;77: 1331–1332. pmid:3115124
  32. 32. Medicare Fraud Strike Force Charges 89 Individuals for Approximately $223 Million in False Billing. [cited 23 Jun 2018]. https://www.justice.gov/opa/pr/medicare-fraud-strike-force-charges-89-individuals-approximately-223-million-false-billing
  33. 33. Taljaard M, McKenzie JE, Ramsay CR, Grimshaw JM. The use of segmented regression in analysing interrupted time series studies: an example in pre-hospital ambulance care. Implementation Science. 2014;9: 77. pmid:24943919
  34. 34. Jebb AT, Tay L, Wang W, Huang Q. Time series analysis for psychological research: examining and forecasting change. Front Psychol. 2015;6. pmid:26106341
  35. 35. Hyndman RJ, Khandakar Y. Automatic Time Series Forecasting: The forecast Package for R. Journal of Statistical Software. 2008;27.
  36. 36. auto.arima function | R Documentation. [cited 3 Jul 2019]. https://www.rdocumentation.org/packages/forecast/versions/8.7/topics/auto.arima
  37. 37. Zhang F, Wagner AK, Soumerai SB, Ross-Degnan D. Methods for estimating confidence intervals in interrupted time series analyses of health interventions. J Clin Epidemiol. 2009;62. pmid:19010644
  38. 38. Baxter P, Cleghorn L, Alvarado K, Cummings G, Kennedy D, McKey C, et al. Quality-based procedures in Ontario: exploring health-care leaders’ responses. J Nurs Manag. 2016;24: 50–58.
  39. 39. Moreno-Serra R, Wagstaff A. System-wide impacts of hospital payment reforms: evidence from Central and Eastern Europe and Central Asia. J Health Econ. 2010;29: 585–602. pmid:20566226
  40. 40. Government of Ontario, Ministry of Health and Long Term Care. Excellent Care For All—Health Care Professionals—MOHLTC. [cited 26 Jan 2019]. http://www.health.gov.on.ca/en/pro/programs/ecfa/legislation/act.aspx
  41. 41. Government of Ontario, Ministry of Health and Long Term Care. Healthy Change—Ontario’s Action Plan for Health Care—Public Information—MOHLTC. [cited 26 Jan 2019]. http://www.health.gov.on.ca/en/ms/ecfa/healthy_change/healthlinks.aspx
  42. 42. Øvretveit J, Health Foundation (Great Britain). Does improving quality save money?: a review of evidence of which improvements to quality reduce costs to health service providers. London: The Health Foundation; 2009.