Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Sustainability planning in the US response to the opioid crisis: An examination using expert and text mining approaches

  • Carlos Gallo ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Conceptualization, Data curation, Formal analysis, Methodology, Software, Visualization, Writing – original draft, Writing – review & editing

    carlos.gallo@northwestern.edu

    Affiliation Department of Psychiatry and Behavioral Sciences, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, United States of America

  • Karen Abram ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Conceptualization, Data curation, Methodology, Project administration, Writing – review & editing

    Affiliation Department of Psychiatry and Behavioral Sciences, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, United States of America

  • Nanette Hannah ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Data curation, Methodology, Writing – original draft, Writing – review & editing

    Affiliation Department of Psychiatry and Behavioral Sciences, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, United States of America

  • Lauren Caton ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Data curation, Methodology, Writing – original draft, Writing – review & editing

    Affiliation Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Palo Alto, California, United States of America

  • Barbara Cimaglio ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Methodology, Writing – review & editing

    Affiliation Illinois Department of Human Services, Division of Substance Use Prevention and Recovery, Chicago, Illinois, United States of America

  • Mark McGovern ,

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Conceptualization, Data curation, Funding acquisition, Methodology, Writing – original draft, Writing – review & editing

    Affiliations Department of Psychiatry and Behavioral Sciences, Stanford University School of Medicine, Palo Alto, California, United States of America, Department of Medicine, Stanford University School of Medicine, Palo Alto, California, United States of America

  • C. Hendricks Brown

    Contributed equally to this work with: Carlos Gallo, Karen Abram, Nanette Hannah, Lauren Caton, Barbara Cimaglio, Mark McGovern, C. Hendricks Brown

    Roles Conceptualization, Formal analysis, Funding acquisition, Methodology, Writing – original draft, Writing – review & editing

    Affiliation Department of Psychiatry and Behavioral Sciences, Feinberg School of Medicine, Northwestern University, Chicago, Illinois, United States of America

Abstract

Between January 2016 and June 2020, the Substance Abuse and Mental Health Services Administration rapidly distributed $7.5 billion in response to the U.S. opioid crisis. These funds are designed to increase access to medications for addiction treatment, reduce unmet treatment need, reduce overdose death rates, and provide and sustain effective prevention, treatment and recovery activities. It is unclear whether or not the services developed using these funds will be sustained beyond the start-up period. Based on 34 (64%) State Opioid Response (SOR) applications, we assessed the states’ sustainability plans focusing on potential funding sources, policies, and quality monitoring. We found variable commitment to sustainability across response plans with less than half the states adequately describing sustainability plans. States with higher proportions of opioid prescribing, opioid misuse, and poverty had somewhat higher scores on sustainment. A text mining/machine learning approach automatically rated sustainability in SOR applications with an 82% accuracy compared to human ratings. Because life saving evidence-based programs and services may be lost, intentional commitment to sustainment beyond the bolus of start-up funding is essential.

Introduction

The opioid crisis in the United States has continued to escalate. There were 46,802 total opioid related deaths in 2018, a historical high that continues to rise, with a provisional number of opioid related deaths exceeding 50,000 in 2019 [1]. An increasing number of individuals who die from opioids are also using stimulants such as methamphetamine and cocaine [2]. From 2015 to 2019, the death toll per 100,000 in the US population increased from 16.3 deaths to 595 deaths attributable to opioid overdose [3]. Despite some states’ recent reductions in opioid related mortality [4], there are signs that these rates will continue to increase at least in some areas, especially with the influx of synthetic opioids [5] and rising mixtures of opioids with methamphetamines and cocaine [2].

In December, 2016, the 21st Century Cures Act provided $1 billion in new funding to address the exponentially increasing opioid crisis [6]. Via the State Targeted Response (STR) grants, the Substance Abuse and Mental Health Services Administration (SAMHSA) began distributing these funds to the states through two allocations in April 2017 and March 2018. SAMHSA announced a third $1.5 billion funding opportunity called the State Opioid Response Grant (SOR) due in August 2018, and money was allocated in September 2019 to continue progress made by states and territories through the STR grants. For the purpose of this paper we refer to these as the 2018 SOR applications. The recently announced FY 2020 SOR budget has an equal amount of support [6]. The SOR grants aim to address the opioid crises by increasing access and infrastructure support for medication for treatment of opioid use disorder (MOUD, i.e., using the three FDA-approved medications for the treatment of opioid use disorder of methadone, buprenorphine, and naloxone); expanding services and reducing unmet prevention, treatment and recovery need; and reducing opioid overdose-related deaths from prescription opioids, heroin and synthetic opioids. SOR applications required a section dedicated to explaining sustainability plans. In particular, states were required to “Describe your plan to ensure these [Required] activities are sustained after grant funding ends” [7].

While the large influx of recent funding offers the potential for expanding states’ and territories’ ability to provide services, the question remains whether such expansions are likely to be sustained when the federal funding streams for this crisis start to decline. The goal of sustainability originally was defined as the extent to which an Evidence Based Intervention (EBI) can continue to deliver its intended benefit over an extended period of time after external support is finished [8, 9]. Recent empirical work on sustainability has led to an expansion of this definition. Thus sustainability can also refer to conditions that create lasting health outcomes by creating infrastructures during the time of funding support. For instance, Palinkas and colleagues wrote “the continued use of program components and activities for the continued achievement of desirable program and population outcomes… some definitions of sustainment focus on the continued existence of the coalition itself while others focus on the activities and impacts of the coalition” [8] after external support is finished. Sustainability was originally defined as the ability to maintain EBI components but has evolved to include a broader definition, including elements of ongoing collaborations/partnerships, infrastructure support, community need, ongoing evaluation of performance and outcomes, and availability of funding.

Recent research has contributed to conceptualizing and evaluating sustainability [1015], but still lags far behind that of the other core phases of implementation science, namely exploration, preparation and implementation with fidelity [14]. This is not to say that sustainability planning is being ignored; sustainability planning was in fact a requirement of the State Opioid Response proposals. However, there is a general tendency during a period marked by financial expansion for governments to concentrate on initiating new implementation strategies at the expense of retaining these capacities beyond grant periods.

The focus of this paper is to measure how well states’ SOR grants included planning for sustainment, and whether state characteristics of population, severity of opioid use disorder, and availability of existing services influenced plans for sustainment. With a view towards the future in which automated procedures for classifying sustainability in similar documents could be used in place of labor-intensive human coding, we also used text mining and machine learning to examine the degree to which such automated procedures could replicate human coding. Similar procedures have been proposed [16] and developed [17] in implementation research to monitor implementation process, but this is the first time we know of that specifically addresses the use of these procedures for sustainability or for treatment of substance abuse.

Methods

Collection of State Opioid Response reports

This study was carried out through the support of a National Institute on Drug Abuse (NIDA) supplement to one of the authors (MM); it was classified as an exempt, non-human subject study by both the Northwestern University and Stanford IRBs. In seeking to determine the degree to which states have described their sustainability programs, we requested the first SOR applications from each of the 50 states. We chose to exclude territories and the District of Columbia from this study as their administrations were quite different from the states. Additionally, territories and the District of Columbia received a set, equal amount of support regardless of population characteristics or need. We used SAMHSA’s website [18] to find the email addresses of the single state agencies (SSAs) for substance abuse services in each state, and wrote directly to each state’s director [18].

We contacted SAMHSA to inform them of our plan to contact current state SSAs requesting full SOR applications and collect additional personnel contacts through former SSAs. We were guided by the National Association for State Alcohol and Drug Abuse Directors (NASADAD) in how to communicate these requests so as to minimize burden to busy state directors. All SSAs, as well as SAMHSA and NASADAD, were provided a summary of the research protocol. We informed all parties that all documents would be held in confidence, and the analyses would be reported in aggregate across states with no identification. All states’ provision of their SORs were completely voluntary. Neither SAMHSA nor NASADAD took an active part nor directly encouraged states to participate, but could answer questions if states contacted them. The funder, NIDA, also did not take an active part. A total of 34 states (68%) provided their SORs.

Data abstraction from the SOR applications: Human coding of sustainability

For this paper we borrow from the recent empirical work of Palinkas and colleagues that identifies common elements of sustainment across multiple SAMHSA programs [8]. In this work, sustainability including elements of ongoing collaborations/partnerships, infrastructure support, community need, ongoing evaluation of performance and outcomes, and availability of funding. This work guided our selection of general categories organized into three major elements that we were able to code reliably from the SOR abstracts. Specifically, our category of “funding” was described as “availability of funding,” in that paper. Our category of “policies, partnerships, or regulations” included “collaborations/partnerships” and “infrastructure support” listed in that paper. Our category of “quality monitoring” mirrored “ongoing evaluation of performance and outcomes” in the recently published paper. We did not include the paper’s other category of future “community need” as some communities’ need was expected to remain high in every state, and projections of such need were absent from these SOR applications.

A team of five doctoral level coders were assembled to review the SOR applications using a standard protocol described below to assess the degree of sustainability across three critical categories: 1) Funding, 2) Policies, and 3) Quality Monitoring. These components were specified as follows:

  1. Funding. Funding was considered sustainable if it included one or more of the following sources that was judged reasonable beyond the grant period: a) Medicaid 1115 waiver was described in the SOR, regardless of whether it is was approved or pending; b) State mentioned sources of funding from private insurance; c) The SOR described new CPT/Billing codes that allowed billing of insurance for opioid treatment services; d) SOR described a partnership with offices that already provide funded services for Drug Control Policy or Alliance of Recovery Residences; or e) existing funding streams supporting the Hub and Spoke model, which required high startup costs.
  2. Policies. The SOR was considered to have sustainable policies when it presented new policies, partnerships or regulations of the state that supports the proposed program beyond the funding period. Ongoing partnership included different state agencies that were committed to continue providing services. These include criminal justice agencies, public health departments, and police departments. Other sustainable activities under this category include mandated trainings for all providers in opioid related services within private specialty clinics, federally qualified health centers (FQHC) and emergency rooms.
  3. Quality monitoring. When quality adherence programs or ongoing program evaluation are explicitly referenced after funding ends, then the application is deemed to meet the quality monitoring criteria. This ongoing quality monitoring includes tracking patients on medication, tracking care coordination, linkage, retention, monitoring and feedback to ongoing learning collaboratives, and finally a description of a data-based evaluation system that has ongoing support.

As an example, a positive score of one (1) is given on funding when an SOR application met the criteria described above for funding, otherwise a score of zero (0) is given when it lacked an indication of sustainable funding or simply mentioned there would be future funding but gave no clear plans. We developed a new protocol for sustainability coding based on a full qualitative assessments of sustainability plans by interviewing previous and current single state authorities (SSAs) on recently published work [19]. The highest score of 5 was assigned to SOR proposals that met all three criteria for its clear discussions of sustainability. A 4 was assigned when only two sustainability criteria were met, and a 3 when only one criterion was met. If an SOR proposal had some discussion of sustainability but no definite plans were identified, a value of 2 was assigned. In other words, unless there was a clear indication of how a program would be sustained, that component of sustainability planning was scored as absent, with a value of 2. Finally, a value of 1 was given when there was no mention of sustainability at all. Three coders rated each SOR independently on this five-point scale. Once all SORs were individually scored, any disagreements were resolved by consensus [20]. Nearly all the initial individual scores agreed within 1 unit, and consensus was reached on all 34 reviews.

When an SOR application was assessed to have any sustainability discussion (a score of 2 to 5), then we coded which of five specific activities were being sustained. These five activities are: a) Medications for Opioid Use Disorder (MOUD) including Suboxone, Buprenorphine, Naltrexone, and Methadone. b) Overdose Services: Naloxone, syringe exchange programs. c) Prevention Services including harm reduction, screening for misuse and risk factors (e.g., SBIRT), primary prevention with drug abuse prevention programs, prescription drug monitoring program (PDMP), prescription disposal services such as drop-off and take back. d) Maintenance of Recovery involving behavioral interventions targeting social determinants of health, job skills, and physical health. e) Implementation Support includes strategies such as ECHO training, Hub and Spoke models, mentoring, coaching, learning collaboratives, and technical assistance.

State measures tested as predictors of sustainability ratings

To investigate how states’ sustainability relates to general population characteristics, need for opioid services, and the service delivery capacity, each state was characterized along the following dimensions: state’s population characteristics, opioid metrics including prescription and mortality, and treatment service availability. Wherever possible, measures from these three dimensions were based on 2017 data, as the SOR applications were written and due in 2018, funded in 2019, but used, at the time, the most updated data, namely 2017’s data. Hence, 2017 information is used to predict sustainability ratings (Table 3), as well as to identify whether there was any sampling bias between the states that provided their SOR application and the states that did provide their SOR application (Table 1).

thumbnail
Table 1. Comparison of states characteristics depending on whether an SOR application.

https://doi.org/10.1371/journal.pone.0245920.t001

State Population Characteristics were compiled from recent public records for 2017 [21]. The US Census datasets were used to examine state’s characteristics in population [22], and percentages of urban density percentage [22], minority population [23], people living in poverty [24] and uninsured individuals [25]. State’s opioid use, misuse, and mortality rates were collected in 2017 for number of opioid prescriptions per 100 persons, opioid misuse percentage for 12 years and older, along with specific age brackets of 12 years to 17 years, 18 years to 25 years, and 26 years and older [26], opioid hospital admissions per 100 thousand people [27], and opioid deaths per 100 thousand people [4, 2830]. Other potentially useful variables such as synthetic opioid deaths [5, 31] were only available for a portion of states, leaving too much missing data for a careful analysis. Hence, such variables were excluded. To the extent possible, we verified data consistency. For instance, we treated a datum as missing when we could not resolve state’s reporting accuracy in numbers. Likewise, when hospital admissions were far greater than discharges rates we treated that state’s admissions values as missing (N = 1). State’s treatment capacity and utilization per 100 thousand people were measured by the SOR budget amount, number of methadone maintenance treatment centers per 100 thousand people [3234], number of prescribing physicians [35], state Medicaid expansion status [36], and whether or not MOUD services were covered by insurance or Medicaid [28]. Additional variables such as copayment and limits on MOUD services [37] were not included in analysis due to significant missing data and high variation in eligibility for reduced rates.

MOUD services were collected from SAMHSA websites [38]. Data on the different types of drug types and deaths were collected from the CDC [29]. Other data specific to opioid and other prescription drugs were collected from KFF.org [33, 39]. Iowa Community Indicator Program was used to collect rural area data for the states [40]. Medicaid expansion and waiver information on the states was obtained from the National Conference of State Legislatures website and the KFF organization [4143]. Medicaid Expansion is a provision in the Affordable Care Act (ACA) that calls for expansion of Medicaid eligibility to cover more low-income Americans [44]. Under the expansion, Medicaid eligibility would be extended to adults up to age 64 with incomes up to 138 percent of the federal poverty level. Each state determines whether to participate in the Medicaid expansion program. In keeping with our agreement not to identify states, we excluded variables where less than 5 states were coded in one category (e.g. 1115 Waiver Exception).

Analytical procedures

We first examined (Table 1) whether those states providing their SOR applications differed on the full set of predictors mentioned above. Continuous data were compared using t-tests after suitable transformation (e.g., log for population); binary proportions were compared using χ2 tests, and count data (e.g., number of opioid treatment centers) were compared using Poisson regression, after adjusting for the relevant population (e.g., those with opioid use disorder). Because of the small number of states and corresponding modest power—a traditional 0.5 two-sided t-test has 70% power to detect a 0.8 effect size and similarly a χ2 test has this same power to detect a true difference of 0.2 and 0.55 –we also relied on diagnostic p-p plots to compare the observed p-values to those derived from a corresponding null distribution [45].

To examine how sustainability rating can be explained by state level predictors, we used ordinal regression (Table 3), which assumes that the log-odds based on all the splits into lower and higher scores on the 1–5 scale is the same (i.e., a cumulative logistic model) [46]. We report this log-odds estimate and its standard error; positive values indicate higher ratings of sustainability with higher values of the predictor variable. Tests of relationships with the sustainability score were based on univariate Wald-type tests of the log-odds estimate divided by its standard error, and we designate tests with p-values less than 0.05 as significant. In addition to the univariate tests, we conducted a backward stepwise multiple regression model that incorporated all predictor variables that reached 0.20 significance, and dropping variables that had substantial multicollinearity. In analyzing binary outcomes (e.g., whether MOUD was sustainable), we used standard logistic regression.

Assessing sustainability using automatic methods: Text mining and machine learning

Text mining.

As the human coding of sustainability requires substantial effort, we explored whether accurate automated classifications of sustainability could be obtained (Table 4). Recognizing that 34 documents is a small number to use text mining to classify states into 5 categories to extract key words and phrases, we chose to dichotomize the sustainability scale into two categories: 4–5 versus 1–3. Text mining searches identified specific linguistic content pertaining to words, chosen by the authors and implementation and opioid experts, in the following categories: education, prevention, rescue, interventions, implementation and sustainability. For instance, the sustainability category contained the terms: sustaining, sustainment, sustainability, and beyond funding, among others (see S1 Appendix-Content for lists of content-related words). The implementation category contained: acceptability, adoption, feasibility, fidelity, among others. Evaluation terms included: monitor, assessment, and feedback. We included 8 categories of implementation strategies from a recent classification system [47]. We also tested to see whether automatically extracted linguistic features related to content and style can predict sustainability as assessed by human raters. The list of keywords used in these categories is located in S1 Appendix-Content. Linguistic features in both content and style have been used to predict success of funding in crowdsourcing campaigns [48] and academic success [49]. We tested two linguistic dimensions as predictors of sustainability using multiple regression modeling: 1) function words (articles, prepositions, auxiliary verbs), and 2) verb tenses in future form. To control for document length, each count of linguistic features was divided by the text length. These stylistic indicators were examined the machine learning algorithm described below.

Machine learning.

Machine classifier algorithms were also tested to explore if sustainability could be predicted in an automatic fashion. Support vector machines (SVM) [50] were trained to identify sustainability ratings based on the language patterns found in the SOR application documents. The SVMs were trained and tested using cross validation with 1-hold-out [51] where one SOR application selected at random was used for testing the algorithm prediction. The rest of the SOR applications were used as training data. The accuracy results were averaged over the cross-validation samples. This setup aimed to make most of the small sample size while reducing potential bias from training and testing document selection. The SVM algorithm has several parameters that can be determined empirically based on the nature of the dataset. The parameters explored in this paper are the kernel used to transform the data into a separable hyperplane (radial, linear, and linear with cost choice), and the cost of misclassification (C). The kernel function transforms the input data into a different space so that classification can be accomplished by a hyperplane [52]. The optimal hyperplane separates the training data using a maximum margin. We can allow misclassifications by controlling the size of the margin. A large margin allows more misclassifications. A narrow margin allows fewer misclassifications. This is called the “cost” function. The optimal cost of misclassification can be found by comparing classification performance and trial and error of the cost value [53].

SVM classifiers do not provide interpretable coefficients; however, they are useful in recognizing categories in unseen data. We also tested additional algorithms such as the Multinomial Naive Bayes classifier implemented in the Weka machine learning toolkit [54]. The independent variables included a more comprehensive set of indicators for the structure and form of the language, including words related to articles, propositions, words longer than 6 letters, number of words per sentence, verb tenses to measure a focus on past, present or future, propositions, articles, and auxiliary verbs.

Results

Table 1 compares two groups: 1) states with SOR documents (N = 34) and 2) states without an SOR document available (N = 16). The comparison was done using a t-test and chi-square significant tests across 15 characteristics of the state’s population, opioid use and mortality, or treatment capacity. Overall we found very limited differences between the two groups (i.e. states with SORs applications versus states without available SORs applications). The only dimension for which a significant difference was found was the percentage of opioid misuse among youth ages 12 to 17 years old (p = 0.015), with 3.66% of states with available SORs versus 4.49% for those without. This significant comparison is not surprising given the 15 multiple tests conducted at the 0.05 level. We examined this further with a p-p plot (Fig 1), which compares the observed ordered p-values against what one would obtain by simulation under a null distribution of 15 independent uniform random variables. In this plot, the observed 15 p-values all fell within or adjacent to the 95% confidence interval from simulation, and there was no indication that there were smaller than expected p-values at the left end of the plot. Thus we found minimal indication of bias selection in the available set of 34 states with SORs documents.

thumbnail
Fig 1. P-P plot comparing states with and without SOR reports.

https://doi.org/10.1371/journal.pone.0245920.g001

Table 2 describes the sustainability distribution of the 34 SOR applications that were available. The scale used for rating sustainability ranged from a lowest of 1 to a highest of 5 with the following frequencies: Grade 1 was assigned to three states (8.8%), grade 2 to eight states (23.5%), grade 3 to eleven states (32.3%), grade 4 to ten states (29.4%), and the highest grade 5 to two states (5.9%). Thus one-third had a rating of 4 or 5, and a similar proportion provided no evidence that any of the three criteria had plans to be sustained. The three criteria necessary for sustainability had the following distribution. Funding criterion was described by 20 states (58.8%), Policies criterion by 16 states (47.1%), and Quality Monitoring criterion by 3 states (8.8%). The main five activities sustained in the SORs were services related to MOUD service activities in 25 states (73.5%), Overdose service activities in 16 states (47.1%), Prevention service activities in 23 states (67.6%), Maintenance service activities in 23 states (67.6%), and Implementation support service activities in 16 states (47.1%).

thumbnail
Table 2. Distribution of sustainability ratings in 34 SOR documents.

https://doi.org/10.1371/journal.pone.0245920.t002

Table 3 summarizes the coefficient and standard error (SE) estimated by univariate prediction of the dependent variables; specifically the sustainability scores, criteria and activities. Significant predictors are shown in the rows with stars. The predictor variables are the state population characteristics, opioid use, misuse and mortality, and the opioid treatment capacity and utilization. The Sustainability rating was significantly likely to be higher when there was a larger number of opioid prescriptions per 100 people (Coef = 0.052, SE = 0.024, p = 0.03), higher percentage of opioid misuse in people 12 years and older (Coef = 1.281, SE = 0.476, p = 0.007), and higher percentage of the opioid misuse in people ages 18 to 25 years old (Coef = 0.845, SE = 0.261, p = 0.001). The funding criteria was more likely to be described in the SOR when there was a higher percentage of poverty in the states (Coef = 0.364, SE = 0.176, p = 0.034), higher percentage of opioid misuse in people 12 years and older (Coef = 1.463, SE = 0.621, p = 0.018), as well as those ages 26 years and older (Coef = 1.158, SE = 0.559, p = 0.038). Both the policies and quality monitoring criteria were more likely to be described in the SOR when there was a higher percentage of opioid misuse among those ages 18 years and 25 years old (Coef = 0.631, SE = 0.277, p = 0.023) and (Coef = 1.225, SE = 0.586, p = 0.037), respectively.

thumbnail
Table 3. Predicting sustainability, three criteria for sustainability, and five activities based on states characteristics, opioid use, misuse, mortality, and opioid service capacity and utilization.

https://doi.org/10.1371/journal.pone.0245920.t003

With respect to the five activities that were sustained, we found the following: MOUD service activities were more likely to be described in the SOR application when there was a higher percentage of opioid misuse by those ages 18 to 28 years (Coef = 0.736, SE = 0.351, p = 0.036). MOUD service activities and Overdose service activities were both more likely to be present when the following four (4) predictors were higher, namely: 1) with higher percentages of populations under the poverty line (Coef = 0.492, SE = 0.226, p = 0.037) and (Coef = 0.292, SE = 0.148, p = 0.049), for MOUD and Overdose outcomes respectively, 2) with higher percentages of uninsured populations (Coef = 0.658, SE = 0.285, p = 0.021) and (Coef = 0.377, SE = 0.171, p = 0.028), for MOUD and Overdose outcomes respectively, 3) with higher opioid prescriptions per 100 people (Coef = 0.083, SE = 0.039, p = 0.03) and (Coef = 0.074, SE = 0.032, p = 0.019), for MOUD and Overdose outcomes respectively, and, finally, 4) with higher percentages of opioid misuse among those ages 12 years and older (Coef = 1.496, SE = 0.695, p = 0.031) and (Coef = 1.081, SE = 0.552, p = 0.05), for MOUD and Overdose outcomes respectively. Prevention service activities were more likely to be described when there was a higher percentage of opioid misuse among those ages 12 years and older (Coef = 2.061, SE = 0.766, p = 0.007), and from those ages 18 to 25 years old (Coef = 2.089, SE = 0.809, p = 0.01). Maintenance of recovery service activities were more likely among those with a higher percentage of opioid misuse among those ages 12 years and older (Coef = 1.104, SE = 0.558, p = 0.048). Implementation support activities were also more likely among states with a higher percentage of opioid misuse among those ages 12 years and older (Coef = 1.7, SE = 0.664, p = 0.01), as well as those with ages 26 years and older (Coef = 1.586, SE = 0.631, p = 0.012).

To extend the univariate tests presented in Table 3, we conducted stepwise multiple regression models that began with all predictor variables that reached significance of up to 0.20 (p< = 0.2), then removed if non-significant when included with this set of predictor variables. Predictors that were colinear above .9 were removed from the multiple regression model. Only one such case occurred, namely, opioid misuse among ages 12 and older and misuse among ages 26 and older had a correlation of 0.94. We kept the former variable, and removed the latter. In the multiple regression model treating Sustainability Ratings as the ordinal dependent variable, two predictors remained. Sustainability score was more likely to be higher when state’s population is poorer (Coef = 0.444, SE = 0.18, p<0.014), and with a higher percentage of opioid misuse among those 12 years and older (Coef = 0.695, SE = 0.305, p = 0.023). Of the three criteria of sustainability, again we removed the opioid misuse among ages 26 and older due to collinearity with those ages 12 and older. Funding criteria was more likely to be described in states with higher percentage of population living under poverty line (Coef = 0.809, SE = 0.381, p = 0.034). No other variables reached 0.05 significance regarding funding. In the multiple regression for Policies criteria and Quality Monitoring criteria, no variable reached significance. Of the five activities being sustained, Overdose activities were more likely to be described in states with smaller populations (Coef = -1.93, SE = 0.86, p = 0.025), and with a higher percentage of opioid misuse among those 12 years and older (Coef = 3.436, SE = 1.488, p = 0.021). Here again we removed collinear predictor opioid misuse among 26 years and older. No other variables reached significance in predicting Overdose activity. Namely, Implementation Support activities were more likely among states with higher percentage of uninsured population (Coef = 0.477, SE = 0.218, p = 0.036), and with higher percentage of opioid misuse among those 12 years and older (Coef = 2.477, SE = 1.006, p<0.014). No other variables reached significance.

To examine sustainability related activities more closely, we also ran a multivariable regression with the five activities as predictors of Sustainability score. While significance effects were not found, the univariate models showed that MOUD, Prevention, Implementation Sustainment Support activities were most closely related to Sustainability Score.

We now turn to the machine learning models. Table 4 summarizes the coefficient and standard error estimated by predicting the dependent variables of sustainability scores, criteria and activities using key words pertinent to broad classes of implementation strategies. In particular, the independent variables are counts of in-house selected words related to education, prevention, rescue, intervention, implementation, sustainability. In addition, we borrowed words in 8 categories from a recent classification system [47], and two linguistic features related to style and future verb tense. Sustainability ratings were more likely to be higher with lower word counts of Category 1 Evaluation (Coef = -4.037, SE = 1.506, p = 0.007). Funding criteria was more likely to be described in the SOR application when there were higher word counts in Category 7 Financial Strategies (Coef = 12.676, SE = 6.396, p = 0.047), and when there were lower words counts in Prevention (Coef = -15.17, SE = 6.777, p = 0.025), Category 1 Evaluation (Coef = -6.076, SE = 2.339, p = 0.009), and Category 6 Consumer Engagement (Coef = -11.273, SE = 5.16, p = 0.029).

thumbnail
Table 4. Predicting sustainability, three criteria for sustainability, and five activities based on linguistic features extracted from 34 SOR documents.

https://doi.org/10.1371/journal.pone.0245920.t004

No variable reached significance when predicting Policy criteria. Quality Monitoring criteria was more likely to be present with higher number of words describing Category 5 Education (Coef = 5.079, SE = 2.579, p = 0.049). Implementation Support was more likely to be described in the SORs that had fewer words pertinent to Category 1 Evaluation (Coef = -3.872, SE = 1.857, p = 0.037). We found no significant predictors for Policies, MOUD, Overdose, Prevention, and Maintenance of recovery activities.

In addition to the univariate regressions, we conducted multiple regression models with all predictor variables that with significance below p< = 0.20 in the univariate model in Table 4. The correlation among these variables was under 0.4, and hence no variable with p value less than 0.2 was removed from the analysis due to collinearity. Sustainability was modestly related to Evaluation Category 1 words. Sustainability was higher with lower number of words within Evaluation Category 1 (Coef = -3.536, SE = 1.836, p = 0.054). For the five activities to be sustained, Implementation support activity was significantly more likely to be described when there were less words in Evaluation Category 1 (Coef = -5.595, SE = 2.657, p = 0.035). Prevention services activity was modestly related to the number of words describing Rescue activities. Prevention was less likely to appear when more words of Rescue activities were mentioned (Coef = -32.885, SE = 17.148, p = 0.0551). We found no significant predictors for Funding, Policies, and Quality Monitoring, MOUD, Overdose, and Maintenance of recovery activities.

Using machine learning to predict sustainability

Three Support Vector Machines (SVM) implemented in R with kernels radial, linear, and linear with cost choice achieved an accuracy of 17%, 52%, and 78% respectively. We tested the optimal cost choice number from 0 to 1, and found that .63 leads to the lowest number of misclassifications. This cost choice leads to an accuracy of 78%. Additionally, a Naive Bayes Multinomial implemented in WEKA, reached an 82% accuracy in predicting Sustainability ratings.

Discussion

Grant applications were requested from all 50 states in order to assess their sustainability commitments beyond federal funding; 34 states shared their application documents. There was no measurable bias between SOR and non-SOR states (Table 1). A protocol for assessing commitment to sustainability was developed to understand state’s response to the opioid crisis in their respective populations. We identified three criteria for sustainability beyond the funding period: identification of funding sources, development of policies, and assurance of quality monitoring.

Despite being a requirement of the SOR application, most states reported limited details on sustainability. Specifically, one third had no clear or no description of how they would sustain their programs. Another third met at least two of the criteria for sustainability, and only 2 states (5.9%) met all 3 criteria (Table 2).

A statement requiring a sustainability section in an SOR application is insufficient for understanding how states plan to sustain their strategies. While new research instruments are just beginning to examine sustainment [55], this area has been under-researched and under-practiced. This suggests that leaders are not fully equipped to enact policies and practices that will promote uninterrupted services and are unlikely to be coordinated with statewide efforts to scale up. Additionally, a state application that said state leaders intend to implement a policy does not guarantee that the policy will in fact be implemented. Our coding system was more rigorous than simply saying that a state intended to sustain programs. Unless there was a clear indication of how a program would be sustained, that component of sustainability was scored as absent. A full qualitative assessments of sustainability plans by interviewing previous and current single state authorities (SSAs) has also analyzed the context, intentions and difficulties on implementing sustainability plans on a recently published paper [19].

States with higher sustainability ratings generally had a higher proportion of the population receiving opioid prescriptions, as well as higher proportions with misuse particularly among young adults. Regarding the three sustainability criteria, funding was more likely to be described in the SOR applications for states having higher poverty rates and opioid misuse, especially among those 26 and older. Policies as well as quality monitoring were more likely to be present among states with higher rates of opioid misuse among young adults. Regarding the five activities, MOUD services were predicted by higher rates of poverty, uninsured citizens, opioid prescriptions and opioid misuse. Prevention services were more likely to be described in states that had higher misuse among youth and young adults. Maintenance of recovery and implementation support were more likely described when opioid misuse was high. In sum, states with the most opioid prescriptions, opioid misuse, and poverty generally had higher sustainability ratings, criteria, and activities.

Text mining can be used to extract linguistic features automatically without human actions, thus providing an opportunity to classify SOR and related documents rapidly and cheaply. Findings from standard text mining showed minimal or sometimes counter-intuitive directional relationships with sustainability measures. This may be due to the inability for single words have to reflect complex concepts such as sustainability. However, results from more complex machine learning algorithms were remarkably successful in predicting the sustainability ratings obtained by expert human raters. We found that 82% of the ratings were accurately predicted by the machine learning method, which provides a more sophisticated decision rule than that using individual word counts by themselves. This represents a useful first step towards measuring sustainability using automatic methods which can facilitate rapid assessments, and hence overcome common bottlenecks from using expert human raters. Other machine learning applications include the recognition of applied prevention research in the NIH portfolio [56]. In this work, Villani et al (2018) developed two machine learning methods with the goal of recognizing prevention studies in the NIH portfolio. Their accuracy ranged from 81% to 88%. Thus our performance of 82% is comparable. The NIH portfolio of prevention studies reached a sample size of N = 3814 which is several order of magnitude bigger than our data set. Machine learning algorithms do learn and improve as more data is available to train and test on.

It is important to interpret these findings through the lens of the state agencies responsible for these time-limited grants and ultimately the addiction prevention and treatment services in their jurisdiction. With the 21st Century Cures Act bolus of funding, states were under pressure to spend funds rapidly. Thus, systems leaders may not have had adequate time or opportunity to plan beyond start-up. The mechanisms to achieve lasting policy and financing changes may be elusive to state leaders who often encounter changing policies at executive and legislative levels of government. These state leaders are situated in varying levels of government, from departments to divisions to bureaus that often have limited authority or capacity to communicate and align long-term strategies. We suggest that state authorities may benefit from sustainability strategies that establish and maintain inter-governmental alliances (e.g., leadership and data sharing across substance abuse care, mental health care, indigent care, child welfare, criminal justice, and emergency medical services). Factors of poverty, opioid treatment deserts, stigma, and discrimination all have pervasive influences on substance abuse and its treatment, so broader systems and policies related to housing, transportation, and education need to be coordinated in a system of systems approach. Drivers of sustainable policies and financing of evidence-based treatments are likely to require sophistication in politics, communication platforms, and continued engaging with advocacy, faith, and other community groups.

To date, implementation research has primarily accumulated knowledge on how to overcome barriers at the organizational or individual health care provider-level within specific macro level contexts [57]. In this paper we did examine widely different contexts based on population and service characteristics. However, it is also likely that higher state-level organizational structure can influence sustainment of opioid treatment and prevention programs. Further, the functioning of governmental as well as service delivery systems can change dramatically under stress, such as economic downturns and the recent COVID-19 pandemic. Thus an important limitation of this paper is that potentially relevant governmental structure and functioning dimensions are not included. This is because we were not able to locate accurate administrative or other data to reflect states’ recent organizational structures and functioning similar to that used to predict local government information sharing [58]. For future research in discriminating successful versus unsuccessful sustainment, it would be useful to include measures of inter-governmental organizational data as well as bridging data reflecting the relationships between providers and their outer contexts [59, 60]. We expect such information would be useful not only for research but have practical value as well.

There are other limitations relevant to this paper. We did not obtain one-third of the states’ SOR applications, so even though there were no important characteristics we could discern from states that did and did not provide their applications, there may still be some differences that we did not detect. Also, the modest sample size, the ordinal and dichotomous measures, along with the multiple univariate tests could have identified some spurious relationships. Indeed, a few relationships that were significant in univariate analysis were no longer significant in multiple regressions. Accounting somewhat more for the possibility of spurious findings, the machine learning analyses have built-in cross-validation procedures that minimize effects of overfitting.

From an implementation practice perspective, state addiction prevention and treatment system authorities need and want guidance on how to leverage change in policy, financing and regulations so that the best available programs are, and remain, accessible to their citizens across the entire system. Practical information for systems leaders about effective and efficient sustainment strategies is therefore sorely needed [61].

Supporting information

S1 Appendix. Quantitative analysis keywords opiods 2020.07.31.

This file contains the keywords used in the text mining and the categories of each list of keywords.

https://doi.org/10.1371/journal.pone.0245920.s001

(XLSX)

References

  1. 1. Statistics NCfH. 12 Month-ending Provisional Counts of Drug Overdose Deaths [Internet]. Center for Disease Control and Prevention2019 [updated 5 Jul 2020. Available from: https://www.cdc.gov/nchs/nvss/vsrr/drug-overdose-data.htm#data-tables.
  2. 2. Ellis MS, Kasper ZA, Cicero TJ. Twin epidemics: The surging rise of methamphetamine use in chronic opioid users. Drug Alcohol Depend. 2018;193:14–20. pmid:30326396
  3. 3. National Institute on Drug Abuse. Opioid Summaries by State [Internet]. National Intstitute of Health [updated 2019 May. Available from: https://www.drugabuse.gov/drugs-abuse/opioids/opioid-summaries-by-state].
  4. 4. Ahmad F, Rossen L, Sutton P. Provisional Drug Overdose Death Counts [Internet]. Center for Disease Control and Prevention 2019 [Available from: https://www.cdc.gov/nchs/nvss/vsrr/drug-overdose-data.htm].
  5. 5. Center for Disease Control and Prevention. Synthetic Overdose Data [Internet]. 2020 [Available from: https://www.cdc.gov/drugoverdose/data/fentanyl.html].
  6. 6. National Association of State Alcohol and Drug Abuse Directors. Timeline of Funding for the State Targeted Response (STR) and State Opioid Response (SOR) Grant Programs. 2019.
  7. 7. Substance Abuse and Mental Health Services Administration. State Opioid Response Grants. Substance Abuse and Mental Health Administration: US Department of Health and Public Services; 2020 8/12/2020.
  8. 8. Palinkas LA, Spear SE, Mendon SJ, Villamar J, Reynolds C, Green CD, et al. Conceptualizing and measuring sustainability of prevention programs, policies, and practices. Translational Behavioral Medicine. 2020;10(1):136–45. pmid:31764968
  9. 9. Rabin BA, Brownson RC. Terminology for dissemination and implementation research. Dissemination and implementation research in health: Translating science to practice 2017. p. 19–45.
  10. 10. Moore JE, Mascarenhas A, Bain J, Straus SE. Developing a comprehensive definition of sustainability. Implementation Science. 2017;12(1):110. pmid:28865479
  11. 11. Wiltsey Stirman S, Kimberly J, Cook N, Calloway A, Castro F, Charns M. The sustainability of new programs and innovations: a review of the empirical literature and recommendations for future research. Implemention Science. 2012;7:17. pmid:22417162
  12. 12. Hailemariam M, Bustos T, Montgomery B, Barajas R, Evans LB, Drahota A. Evidence-based intervention sustainability strategies: a systematic review. Implementation Science. 2019;14(1):57. pmid:31171004
  13. 13. Palinkas L, Mendon S. Conceptualizing and measuring sustainability of prevention programs and initiatives. PSMG Virtual Grand Rounds 2018.
  14. 14. Scheirer MA, Dearing JW. An agenda for research on the sustainability of public health programs. American Journal of Public Health. 2011;101(11):2059–67. pmid:21940916
  15. 15. Palinkas LA, Chou CP, Spear SE, Mendon SJ, Villamar JA, Brown CH. Measurement of sustainment of prevention programs and initiatives: the sustainment measurement system scale. Implementation Science. 2020;15(1):71. pmid:32883352
  16. 16. Gallo CG, Pantin H, Villamar J, Prado GJ, Tapia M, Ogihara M, et al. Blending Qualitative and Computational Linguistics Methods for Fidelity Assessment: Experience with the Familias Unidas Preventive Intervention. Administration and Policy in Mental Health and Mental Health Services Research. 2015;42(5):574–85. pmid:24500022
  17. 17. Wang D, Ogihara M, Gallo CG, Villamar J, Smith JD, Vermeer W, et al. Automatic classification of communication logs into implementation stages via text analysis. Implementation Science. 2016;11(1):119. pmid:27600612
  18. 18. Substance Abuse and Mental Health Services Administration. Directory of Single State Agencies (SSA) for Substance Abuse Services [Internet]. samhsa.gov 2016 [30 Mar 2020]. Available from: https://www.samhsa.gov/sites/default/files/ssadirectory.pdf].
  19. 19. Caton L, Yuan M, Louie D, Gallo C, Abram K, Palinkas L, et al. The Prospects for Sustaining Evidence-based Responses to the US Opioid Epidemic State Leadership Perspectives. To Appear in Substance Abuse Treatment, Prevention, and Policy. 2020. pmid:32727559
  20. 20. Willms DG, Best JA, Taylor DW, Gilbert JR, Wilson DMC, Lindsay EA, et al. A Systematic Approach for Using Qualitative Methods in Primary Prevention Research. Medical Anthropology Quarterly. 1990;4(4):391–409.
  21. 21. United States Census Bureau. Annual Estimates of the Resident Population for the United States, Regions, States, and Puerto Rico: April 1, 2010 to July 1, 2019.
  22. 22. United States Census Bureau. US Census QuickFacts [Internet]. census.gov [Available from: https://www.census.gov/quickfacts/US].
  23. 23. Governing the future of states and localities. State Population By Race, Ethnicity Data [Internet]. 2017 [Available from: https://www.governing.com/gov-data/census/state-minority-population-data-estimates.html].
  24. 24. List of U.S. states and territoris by poverty rate US Census Bureau2020 [Available from: https://www.census.gov/quickfacts/fact/table/US/PST045219.
  25. 25. Kaiser Family Foundation. Health Insurance Coverage of the Total Population [Internet]. KFF.org 2018 [Available from: https://www.kff.org/other/state-indicator/total-population/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  26. 26. Kaiser Family Foundation. Individuals Reporting Past Year Opioid Use Disorder [Internet]. KFF.org 2017–2018 [Available from: https://www.kff.org/other/state-indicator/past-year-opioid-use-disorder/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  27. 27. Substance Abuse and Mental Health Services Administration CfBHSaQ. Treatment Episode Data Set (TEDS): 2017. Admissions to and Discharges from Publicly-Funded Substance Use Treatment. Substance Abuse and Mental Health Services Administration. 2017.
  28. 28. 2018 N-SSATS State Profiles. Substance Abuse and Mental Health Services Administration [Internet]. 2018 [Available from: https://www.samhsa.gov/data/sites/default/files/reports/rpt23232/2018_NSSATS_StPro_combined.pdf].
  29. 29. Center for Disease Control and Prevention. Nonfatal Overdoses: All Opioids [Internet]. 2020 [updated 2020 Apr 21. Available from: https://www.cdc.gov/drugoverdose/data/nonfatal/nonfatal-opioids.html].
  30. 30. Kaiser Family Foundation. Opioid Overdose Death Rates and All Drug Overdose Death Rates per 100,000 Population (Age-Adjusted) [Internet]. KFF.org 2020 [Available from: https://www.kff.org/other/state-indicator/opioid-overdose-death-rates/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  31. 31. Prekupec MP, Mansky PA, Baumann MH. Misuse of Novel Synthetic Opioids: A Deadly New Trend. Journal of Addiction Medicine. 2017;11(4):256–65. pmid:28590391
  32. 32. Kaiser Family Foundation. Individuals Reporting Needing but Not Receiving Treatment for Illicit Drug Use in the Past Year [Internet]. KFF.org 2017–2018 [Available from: https://www.kff.org/other/state-indicator/individuals-reporting-needing-but-not-receiving-treatment-for-illicit-drug-use-in-the-past-year/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  33. 33. Kaiser Family Foundation. Medicaid Behavioral Health Services: Naloxone Coverage Provided for Family Members or Friends Obtaining a Naloxone Prescription on Enrollee’s Behalf [Internet]. KFF.org 2018 [Available from: https://www.kff.org/other/state-indicator/medicaid-behavioral-health-services-naloxone-coverage-provided-for-family-members-or-friends-obtaining-naloxone-prescription-on-enrollees-behalf/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  34. 34. Substance Abuse and Mental Health Services Administration. Opioid Treatment Program Directory [Internet]. samhsa.gov [Available from: https://dpt2.samhsa.gov/treatment/directory.aspx].
  35. 35. Substance Abuse and Mental Health Services Administration. Buprenorphine Practitioner Locator [Internet]. samhsa.gov [Available from: https://www.samhsa.gov/medication-assisted-treatment/practitioner-program-data/treatment-practitioner-locator].
  36. 36. Healthinsurance.org. What is Medicaid expansion? [Internet]. Healthinsurance.org [updated 21 Feb 2020. Available from: https://www.healthinsurance.org/glossary/medicaid-expansion/].
  37. 37. Kaiser Family Foundation. Medicaid Behavioral Health Services: Methadone for Medication Assisted Treatment (MAT) [Available from: https://www.kff.org/other/state-indicator/medicaid-behavioral-health-services-methadone-for-medication-assisted-treatment-mat/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D.
  38. 38. Substance Abuse and Mental Health Services Administration. Supplemental NSDUH Opioid Tables. 2018.
  39. 39. Kaiser Family Foundation. Medicaid Behavioral Health Services: Methadone for Medication Assisted Treatment (MAT) [Internet]. KFF.org 2018 [Available from: https://www.kff.org/other/state-indicator/medicaid-behavioral-health-services-methadone-for-medication-assisted-treatment-mat/?currentTimeframe=0&sortModel=%7B%22colId%22:%22Location%22,%22sort%22:%22asc%22%7D].
  40. 40. Iowa State University Iowa Community Indicators Program. Urban Percentage of the Population for States, Historical [Internet]. [Available from: https://www.icip.iastate.edu/tables/population/urban-pct-states].
  41. 41. National Conference of State Legislatures. Medicaid 1115 Waivers by State [Internet]. ncsl.org 2019 [updated 2019 Aug 30. Available from: https://www.ncsl.org/research/health/medicaid-1115-waivers-by-state.aspx].
  42. 42. KFF. Medicaid Waiver Tracker: Approved and Pending Section 1115 Waivers by State 2020 [Available from: https://www.kff.org/medicaid/issue-brief/medicaid-waiver-tracker-approved-and-pending-section-1115-waivers-by-state/.
  43. 43. Kaiser Family Foundation. Medicaid Waiver Tracker: Approved and Pending Section 1115 Waivers by State [Internet]. KFF.org 2020 [Available from: https://www.kff.org/medicaid/issue-brief/medicaid-waiver-tracker-approved-and-pending-section-1115-waivers-by-state/].
  44. 44. Mitchell AB, Sara. Overview of the ACA Medicaid Expansion [Internet]. 2018 [updated 2018 Dec 3. Available from: https://fas.org/sgp/crs/misc/IF10399.pdf].
  45. 45. Wilk MB, Gnanadesikan R. Probability Plotting Methods for the Analysis of Data. Biometrika. 1968;55(1):1–17. pmid:5661047
  46. 46. Agresti A. Modelling ordered categorical data: recent advances and future challenges. Statistics in Medicine. 1999;18(17–18):2191–207. pmid:10474133
  47. 47. Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implementation Science. 2015;10(1):21. pmid:25889199
  48. 48. Parhankangas A, Renko M. Linguistic style and crowdfunding success among social and commercial entrepreneurs. Journal of Business Venturing. 2017;32(2):215–36.
  49. 49. Pennebaker JW, Chung CK, Frazee J, Lavergne GM, Beaver DI. When small words foretell academic success: the case of college admissions essays. PLoS One. 2014;9(12):e115844. pmid:25551217
  50. 50. Karatzoglou A, Meyer D, Hornik K. Support Vector Machines in R. 2006. 2006;15(9):28. pmid:16567719
  51. 51. Gu B, Sheng VS, Tay KY, Romano W, Li S. Cross Validation Through Two-Dimensional Solution Surface for Cost-Sensitive SVM. IEEE Transactions on Pattern Analysis and Machine Intelligence. 2017;39(6):1103–21. pmid:27295653
  52. 52. Yekkehkhany B, Safari A, Homayouni S, Hasanlou M. A Comparison Study of Different Kernel Functions for SVM-Based Classification of Multi-Temporal Polarimetry SAR Data. International Archives of the Photogrammetry Remote Sensing and Spatial Information Science. 2014;XL-2/W3:281–5.
  53. 53. Cao P, Zhao D, Zaiane O, editors. An Optimized Cost-Sensitive SVM for Imbalanced Data Learning 2013; Berlin, Heidelberg: Springer Berlin Heidelberg.
  54. 54. Kotthoff L, Thornton C, Hoos HH, Hutter F, Leyton-Brown K. Auto-WEKA 2.0: Automatic model selection and hyperparameter optimization in WEKA. Journal of Machine Learning Research. 2017;18:826–30.
  55. 55. Palinkas LA, Spear S, Mendon S, Villamar JA, Reynolds C, Green CD, et al. Conceptualizing and measuring sustainability of prevention programs and initiatives. Translational and Behavioral Medicine. 2019;10(1):136–45.
  56. 56. Villani J, Schully SD, Meyer P, Myles RL, Lee JA, Murray DM, et al. A Machine Learning Approach to Identify NIH-Funded Applied Prevention Research. Am J Prev Med. 2018;55(6):926–31. pmid:30458951
  57. 57. Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to Improve the Selection and Tailoring of Implementation Strategies. The Journal of Behavioral Health Services & Research. 2017;44(2):177–94. pmid:26289563
  58. 58. Gil-Garcia JR, Sayogo DS. Government inter-organizational information sharing initiatives: Understanding the main determinants of success. Elsevier. 2016; 33:572–82.
  59. 59. Moullin JC, Dickson KS, Stadnick NA, Rabin B, Aarons GA. Systematic review of the Exploration, Preparation, Implementation, Sustainment (EPIS) framework. Implementation Science. 2019;14(1):1. pmid:30611302
  60. 60. Lengnick-Hall R, Willging C, Hurlburt M, Fenwick K, Aarons GA. Contracting as a bridging factor linking outer and inner contexts during EBP implementation and sustainment: a prospective study across multiple U.S. public sector service systems. Implementation Science. 2020;15(1):43. pmid:32527274
  61. 61. Caton L, Yuan M, Louie D, Gallo C, Abram K, Palinkas L, et al. The prospects for sustaining eidence-based responses to the US opioid epidemic: state leadership perspectives. Under Review. 2020. pmid:33148283