- Systematic review
- Open Access
Cost analysis in implementation studies of evidence-based practices for mental health and substance use disorders: a systematic review
Implementation Science volume 16, Article number: 26 (2021)
This study is a systematic literature review of cost analyses conducted within implementation studies on behavioral health services. Cost analysis of implementing evidence-based practices (EBP) has become important within implementation science and is critical for bridging the research to practice gap to improve access to quality healthcare services. Costing studies in this area are rare but necessary since cost can be a barrier to implementation and sustainment of EBP.
We followed Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) methodology and applied the Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist. Key search terms included: (1) economics, (2) implementation, (3) EBP, and (4) behavioral health. Terms were searched within article title and abstracts in: EconLit, SocINDEX, Medline, and PsychINFO. A total of 464 abstracts were screened independently by two authors and reduced to 37 articles using inclusion and exclusion criteria. After a full-text review, 18 articles were included.
Findings were used to classify costs into direct implementation, direct services, and indirect implementation. While all studies included phases of implementation as part of their design, only five studies examined resources across multiple phases of an implementation framework. Most studies reported direct service costs associated with adopting a new practice, usually summarized as total EBP cost, cost per client, cost per clinician, and/or cost per agency. For studies with detailed analysis, there were eleven direct cost categories represented. For five studies that reported costs per child served, direct implementation costs varied from $886 to $9470 per child, while indirect implementation costs ranged from $897 to $3805 per child.
This is the first systematic literature review to examine costs of implementing EBP in behavioral healthcare settings. Since 2000, 18 studies were identified that included a cost analysis. Given a wide variation in the study designs and economic methods, comparison across studies was challenging, which is a major limitation in the field, as it becomes difficult to replicate studies or to estimate future costs to inform policy decisions related to budgeting. We recommend future economic implementation studies to consider standard economic costing methods capturing costs across implementation framework phases to support comparisons and replicability.
Cost analysis is broadly applied through economic evaluations of treatment interventions and related programs for mental health/substance use disorders (SUD), often as part of an effectiveness/cost-effectiveness study. Recently, cost analysis has become an important area within the implementation science field, which focuses on the translation of research into practice and disseminating methods and applications of evidence-based practices (EBP) on a broader scale . Implementation science has developed to understand the factors that facilitate the adoption of EBP to assist in bridging the research to practice gap, and to improve the accessibility and quality of health services . Health services focused on mental health and SUD, across the lifespan, fall under the broader area of behavioral health. Research and evaluation of the implementation of EBP in behavioral healthcare is a rapidly expanding area of importance, particularly since EBP adoption in this area has lagged behind that in traditional healthcare settings. Economic studies in this area are especially important since costs can be a barrier to implementation of EBPs in behavioral healthcare agencies , and funding/payment for behavioral health services are often separate from other health services.
In their seminal article, Proctor et al. specifically called for implementation research to include costs as an outcome . The inclusion of costs in implementation research is important because the costs of implementing EBP can drive decision-making by service providers. Such decision-making impacts whether an EBP will be adopted or sustained, which ultimately impacts service quality and patient outcomes . However, despite the importance of including costs, a systematic review of implementation research on practice guidelines found that only about a quarter of implementation studies include costs . This is relevant because the bulk of costing studies in implementation research focuses on practice guidelines .
Implementation studies within the broader health service delivery field that included economic evaluations were summarized in a recent systematic literature review by Roberts et al. . This review identified 30 studies, most of which were hospital-based. More than half of the included studies used cost-effectiveness analyses, while smaller proportions used cost utility analyses or cost-consequence analyses. Measured costs included staff training, development of new care processes/pathways, and patient/caregiver-related costs. While this study provided a useful overview of the economic tools used in implementation research, it did not specify how these tools have been used in the area of nonhospital-based behavioral health services or how costs are collected across implementation phases.
The extent to which researchers have included costs in their implementation studies of EBP in behavioral health services is not known. Additionally, the costing methods applied in these studies have not been systematically described. We address this gap by conducting a systematic review of the literature to understand the types of costs and costing methods which have been used to date in implementation research of evidence-based behavioral health interventions in the USA and Canada. We focus on these two countries due to their similarities in the epidemiology of behavioral health disorders, overlapping related health policies , and flow of both patients and programs across the border . We exclude other countries due to the unique nature of the US healthcare financing . Addressing these issues can elucidate areas of potential growth in how implementation researchers incorporate cost analysis as a core measure. While costs are an important driver, they certainly do not overshadow the importance of patient outcomes, which is why we focus only on the implementation of EBP, which have already been linked to quality services.
Identification of studies
We conducted an initial systematic review (February 2019) and updated the search prior to submission (April 2020). We investigated costing studies in behavioral health implementation research using Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) methodology  and checklist . We developed key search terms through reviewing search terms in MeSH® . Key search terms included four areas: (1) economics, (2) implementation, (3) EBP, and (4) behavioral health. Included articles contained at least one key term from each area. Specific key terms were economic, cost, implementation, evidence-based practice, behavioral health, mental health, mental illness, mental disorder, psychiatric, addiction, drug abuse, substance use, and drug use. Key words related to specific types of substances (e.g., opioids) were not included to reduce complexity.
We based our search strategy on suggestions from psychology and economics published best practices for systematic literature reviews [14, 15]. Terms were searched within article title and abstracts in the following databases: EconLit, SocINDEX, Medline, and PsychINFO. We did not include Embase, as we were not interested in biomedically or pharmaceutically related research, and excluded studies focused only on drugs. Included articles contained at least one key term from each area. Abstracts were downloaded into Excel and independently screened by two of the authors to determine if they met inclusion and exclusion criteria. Full-text screening was conducted for all articles meeting inclusion criteria. Full-text screening was also conducted for articles where insufficient information was provided in the abstract to determine if criteria were met. The Boolean search and selection criteria are listed below. There is no published review protocol.
((TI economic) OR (TI cost) OR (AB economic) OR (AB cost)) AND ((TI implementation) OR (AB implementation)) AND ((TI evidence based practice) OR (AB evidence based practice)) AND ((TI behavioral health) OR (AB behavioral health) OR (TI mental health) OR (AB mental health) OR (TI mental illness) OR (AB mental illness) OR (TI mental disorder) OR (AB mental disorder) OR (TI psychiatric) OR (AB psychiatric) OR (TI addiction) OR (AB addiction) OR (TI drug abuse) OR (AB drug abuse) OR (TI substance use) OR (AB substance use) OR (TI drug use) OR (AB drug use))
The following are the inclusion criteria:
Articles that address implementation studies of behavioral health services which incorporate costing (formal economic analysis was not required for inclusion)
Studies published after 2000
Studies only if original research is presented (quantitative or qualitative data)
Community-based studies (outpatient level of care or services provided to people residing in the community, rather than a hospital)
Published in English
Research was conducted in the USA or Canada
Studies providing services to people of any age
The following are excluded in the study:
Published before 2000
Editorials, newspapers, and other pieces of popular media
Dissertations or theses
Book chapters, book reviews, proceedings, government and/or organization reports, and other publications from the gray literature
Protocols without preliminary data
Published in a language other than English.
Studying a system outside the USA or Canada
Studies based in hospitals
Did not include costing data
Focusing solely on comparing medications
Data extraction and analysis
As shown in Fig. 1, the initial abstract search returned 636 articles (four from EconLit, 33 from SocIndex, 335 from Medline, and 264 from PsychINFO). Seven articles were added based on review of references during full text review. After excluding articles based on date of publication and those that were duplicates, a total of 464 unique abstracts were screened independently by two authors. Authors also screened articles individually for risk of bias, but did not exclude any articles based on this issue. This was reduced to 37 articles after applying inclusion/exclusion criteria to abstracts. The final number of included articles after full-text reviews was 18.
The following characteristics were extracted from all included studies: authors, year published, journal, area of behavioral health, type of EBP, sample size, implementation framework, economic methodology, costing approach, costing perspective(s), cost categories and specific resources, cost analysis results, and economic study design. Thematic analysis was conducted to summarize similarities and differences across studies. Content analysis was used to identify how these characteristics overlapped within included studies, such as identifying costing perspectives and resource categories that overlapped across studies. Content analysis classified three categories of costs for the purpose of the study: direct implementation, direct service, and indirect implementation. For a subset of studies, direct and indirect implementation cost results were extracted and reported. In studies where summary cost estimates were not provided, we used the aggregated cost results (as presented) combined with patient/family/client sample sizes provided within the articles to approximate summary costs (i.e., cost per child served) to support comparisons across studies.
The Consolidated Health Economic Evaluation Reporting Standards (CHEERS) checklist  was used to assess the quality of the articles. This checklist was designed to provide a standardized reporting system for economic evaluations of health interventions. Since we included other implementation studies with costing components, in addition to economic evaluations, we reported the overall percent of included items on the checklist (removing not applicable items from the denominator, rather than a score). We eliminated several checklist items that were not relevant for costing studies that did not also include an economic evaluation. We converted CHEERS checklist quality scores used in previous research [17, 18] to a percentage. Previous studies applied the following scoring categories: excellent (24 out of 24 points or 100%), good (at least 18 out of 24 points or 75-99%), average (at least 12 out of 24 points or 50-74%), and poor (11 or fewer out of 24 points or 0-49%).
Table 1 provides the list of included studies with details on study design, type of EBP, target population, implementation framework, and economic methodology. Only one of the included articles was published before 2010 . Most articles (12 articles) were published in specialty journals focusing on behavioral health. The remainder were published in specialty journals related to public health (four articles [20,21,22,23]), children’s services (one article ), or implementation science (one article ). Several studies focused on specific demographic groups [children with substance use disorders, mental health disorders, or trauma (six articles [19, 23,24,25,26,27]), adolescents/youth with mental health disorders (two articles [21, 28]), justice-involved youth (two articles [29, 30]), geriatric mental health (one article ), and veteran’s mental health (one article ). Studies focusing on specific diagnoses included: alcohol use (one article ), addiction (one article ), co-occurring disorders (one article ), mental health disorders (two articles [22, 35]), and serious mental illness (one article ).
Most articles (15 articles) investigated the implementation of a single EBP. However, some studies (three articles) considered implementation of multiple practices [21, 33, 36]. EBPs delivered to patients included multisystemic therapy, screening and brief intervention for alcohol use, trauma-focused cognitive behavioral therapy, gender-responsive trauma-informed care, assertive community treatment, integrated treatment, family psychoeducation, integrated dual disorders treatment, illness management/recovery, and supported employment. EBPs that were delivered to providers focused on remote learning and mentoring, improved practice guidelines, quality measures, and enhanced interagency collaboration.
Studies varied in sample size and unit of analysis. Four studies measured EBP adoption on individual patient outcomes. The sample size in these studies ranged from 72-1,869 people; however, three of the studies had a sample size between 72-155 [23, 25, 34]. One study measured the impact on families (N=45) . Another study measured outcomes for patients (N=1,410) and clinicians (N=574) . Yet another study measured outcomes for clinicians only (N=154) . Three studies focused on management staff and/or teams, with sample sizes ranging from 32 (teams) to 75 (staff) [21, 31, 35]. Eight studies measured EBP impact on the organization/agency or system and had sample sizes ranging from 3-307 organizations [22, 24, 26, 27, 30, 32, 33, 36].
Implementation frameworks and costing approaches
Most articles (11 articles [21, 24, 26,27,28,29, 31,32,33,34,35],) reported using an implementation framework, although there were no studies that reported using the same framework. This is indicative of the broad set of implementation frameworks that exist within the field, and the complexity of internal and external factors that must be considered in evaluating an implementation process based on type of EBP, setting, and target population. The Consolidated Framework for Implementation Research (CFIR) provides a point of reference as it includes a comprehensive list of constructs (intervention characteristics, agency/staff characteristics, outer and inner contexts within which the agency operates, and the implementation process itself) . In this framework, cost is mentioned as a key measure of the implementation process, but no details are provided as to recommended approaches, measures, or data sources for estimating implementation costs. In this review, about half (five articles) of the included studies were developing a framework, or specific tools, for understanding implementation processes and/or phases. These tools/frameworks focused on costs (two articles [26, 33]), EBP sustainability (two articles [27, 32]), and implementation stages and fidelity (one article ). The remaining studies adopted existing implementation frameworks to conceptually guide their research.
Most implementation frameworks follow a phased structure, generally including pre-implementation, implementation, and sustainment activities and corresponding evaluation metrics. Cost analyses must define an analytic perspective, and in implementation research, two key perspectives are the agency/provider/organization that would be deciding to adopt an EBP and the payer, sometimes the agency itself, health insurance plan, or Medicaid/Medicare. In cost-effectiveness research, the Second Panel on Cost-Effectiveness in Health and Medicine recommends adopting the healthcare sector and societal perspectives, both of which are broader than a provider or payer perspective. The societal perspective includes costs to the healthcare system, other systems (e.g., justice, education, social services), and patient/caregivers. The cost analysis perspective defines the relevant resources and monetary conversion factors for estimating costs across implementation phases. General categories of implementation costs include direct labor costs, indirect labor costs, and nonlabor costs . Direct labor costs reflect personnel time spent implementing an EBP, including time spent training and delivering services to clients. Indirect labor costs refer to personnel time outside of clinical services, such as time spent recording case notes or administrative duties. Nonlabor costs essentially cover all other resources required for implementation such as contracted services, equipment, supplies, licensing, building/space costs, and administrative overhead. Studies use other many terms to characterize costs such as direct implementation costs, direct service costs, opportunity costs, accounting costs, as well as fixed costs and variable costs. All implementation phases (pre-implementation, implementation, and sustainment) have some combination of these costs. Pre-implementation costs would typically include labor costs associated with meetings and trainings during preparation and planning phases, as well as one-time expenditures on equipment or supplies to prepare for the implementation phase. Implementation costs would include labor costs for those delivering services or performing administrative roles, and recurring expenditures on supplies, equipment, space, communications. Sustainment costs would include the cost of implementation resources that are required to continuously support the EBP.
For the purposes of this review, we classify three categories of costs guided by the presentation of costs in the included studies: direct implementation costs, direct service costs, and indirect implementation costs (Table 2). As shown in Table 2, direct implementation costs capture the actual expenditures incurred by agencies/providers implementing the EBP for pre-implementation and implementation activities such as meetings, trainings, purchasing training manuals, and travel. Direct implementation costs were calculated mainly through activity-based costing, tallying the time spent on implementation activities and applying reported or imputed salary data to estimate the direct labor costs for implementation activities. Direct service costs are those costs associated with billable healthcare and other services resulting from the implementation process or specific EBP activities. About half of the included studies calculated direct service costs using Medicaid or other claims data [19, 20, 22,23,24,25,26, 29, 30, 34].
Indirect implementation costs capture the opportunity cost for agencies due to lost revenues and/or time spent on implementation activities rather than standard (pre-implementation) clinical activities that could be billed and reimbursed. Indirect costs were estimated using the amount of time that a clinician reported spending on implementation activities, multiplied by the reimbursement rate for their billable activities [27, 28]. Two of the 18 studies measured the indirect costs associated with implementation of EBPs [19, 24]. Seven studies do not report cost figures, but rather used primary data collection and qualitative methods to summarize specific examples of both direct and indirect implementation costs incurred at the agency during the implementation process [21, 22, 24, 32, 33, 35, 36]. Almost all studies (sixteen) applied direct costing, although one study used cost-benefit analysis , and another used cost-effectiveness analysis . Different costing perspectives were represented across the 18 studies. Eleven studies described using a provider (i.e., organizational) perspective, with one other study also including costs incurred by organization staff (but not reimbursed directly)  and another including costs incurred by staff as well as consumers . One study applied a facility perspective (one location of a multi-location organization) , one reported an interagency perspective (e.g., representing costs incurred across collaborating agencies) , and one reported a taxpayer perspective . Two studies did not fully describe their perspective but provided enough detail to infer that they used a societal perspective [22, 23].
Using the cost categories defined in Table 2, we noted five studies examining costs within all phases of implementation (including sustainment) based on the adopted framework [19, 21, 24, 33, 35]. Three studies examined only pre-implementation and implementation phases, all of which focused on trauma-focused cognitive behavioral therapy. These studies measured either direct implementation costs, direct services costs, and/or indirect implementation costs [26,27,28]. Eight studies included costs aligned only with the implementation phase [20, 22, 23, 25, 29,30,31, 34]. All of these studies measured direct services costs, except one study, which measured direct implementation costs . The remaining two studies were focused on the sustainability phase, and provided qualitative measurement of direct implementation costs [32, 36].
As shown in Table 3, there was a wide variation in the types of direct cost categories in the implementation costing studies. We identified eleven distinct categories of direct implementation costs captured across these five studies. Eight of these categories captured direct costs attributable to staff time spent in implementation activities such as trainings, meetings, on-site consultations, follow-up calls, supervision, project management, and preparation time. While the actual name of the cost categories differed slightly across the studies, the objective of including these eight direct cost categories was similar. For example, on-site meetings are called “learning sessions”  in the study of community-based implementation of trauma-focused cognitive behavioral therapy, and “on-site assessor reviews” in another study . The purpose of including this cost category is to understand how much time is devoted to learning how to appropriately implement the EBP. The three final cost categories captured time devoted to data activities required of implementation studies or the actual purchase of data equipment as well as transportation expenses and training materials purchased. No single study included all 11 cost categories. One study of the implementation of a trauma-focused cognitive behavioral therapy program for youth included nine of the eleven cost categories . Another study examining three specific methods for assessing implementation activities (on-site, phone, and self-report), included four cost categories only . One article focused implementation costing on training and travel expenses only .
Studies reporting quantitative cost results across the highlighted costing categories are summarized in Table 4. Since it was difficult to compare cost results between studies given the variation in the type of EBP implemented and the number of cost categories included, the results were summarized by type of cost (direct versus indirect), type of EBP, and average costs (per child, per clinician, per collaborative). The results show that even within similar cost categories and types of costs, there was still considerable variation. For example, among the five studies that reported direct implementation costs [19, 26,27,28, 31], four reported direct implementation costs per child, which ranged from $886 per child to $9470 per child. For the three studies that implemented the same EBP (trauma-focused cognitive behavioral therapy for children), direct implementation cost per child varied between $886 and $2742 [26,27,28]. Two studies reported indirect implementation costs ranging from $897 per child to $3805 per child [27, 28].
Three studies examined costs across phases of implementation: pre-implementation, implementation, and sustainment. All three of these studies also reported average per-child costs. One study combined pre-implementation and implementation direct costs totaling $886 per child . A second study reported average direct implementation costs per child of $1,000 and indirect implementation costs per child of $897 per year during the sustainment period . In another study, while the authors reported staff costs separately for pre-implementation and implementation, they report an aggregate summary of the average direct implementation costs ($2,742 per child) for both phases together .
Results from the CHEERS checklist (Table 5) indicate that, on average, studies included 87% of applicable items from the CHEERS checklist, which corresponds to good quality. The highest quality (excellent) studies include all applicable items [32, 33, 35, 36]. All studies, except one, included at least 75% of applicable items (good quality), with the one exception including only 65% (average quality) . The most common item that was not included was the discount rate, with only 8% of studies including that item. Several other items from the CHEERS checklist related to cost-effectiveness analysis and were largely not applicable. These items included measurement of effectiveness, measurement of valuation, and characterizing uncertainty/heterogeneity.
Overview of the field
This is the first systematic literature review to examine cost analyses within implementation studies of EBP for behavioral health services. The literature review identified 18 studies published since 2000 that included a formal cost analysis. Most of this work has been published in recent years (16 articles since 2013). Given the increase in cost analyses within implementation research of behavioral health EBPs, this study provides important context for the current state of the field, including a summary of findings, and we offer suggestions for how the field might standardize economic concepts and measures to increase translatability and support data harmonization going forward. In particular, the large number of not applicable categories from the CHEERS checklist may indicate a need for a new checklist of costing as part of implementation research.
Review of results
We found several studies proposing the development of an implementation research framework, but no studies using a previously published framework. This finding resonates with a recent study by Wensing and Grol (2019) that identified a large number of implementation frameworks within the field, but very little consistency across contextual determinants . Our findings describe and categorize existing methodologies for calculating direct costs of EBP implementation, direct costs of healthcare services, and indirect costs of EBP implementation. These three categories of cost most closely align with how cost results were presented in the included studies, but they diverge slightly from standard cost categories in the economic evaluation literature. For instance, economic analyses of treatment interventions typically include start-up costs (aligned with pre-implementation costs) and intervention management/operational costs (aligned with implementation and sustainment costs). Within these categories, resources are typically characterized as sunk costs (e.g., start-up), variable costs (dependent on number of patients or clients), time-dependent (recurring costs during the year to support implementation), and societal costs (e.g., opportunity costs of subsidized/donated resources, staff and participant’s time and travel costs).
We identified a wide variation in study type and costing methods, thus limiting the comparability across cost analysis results. Further complicating comparability is the possibility that there could be some double counting in the direct implementation and direct service cost categories if staff time is being reported as a provider expense as well as a billable service cost. Only one study included both of these categories , and the methods and results were not detailed enough to determine the magnitude of double counting. However, this limitation is important for future comparisons of implementation costs. Despite these challenges, we were able to highlight 11 distinct categories of direct implementation costs that are captured in sufficient detail across five of the studies. This review also highlights the importance of estimating the indirect implementation costs—or opportunity costs—in both seeking and providing services. Opportunity cost in this context refers to the lost revenue from billable clinical services due to implementation activities. Several of the implementation studies calculated opportunity costs. These calculations are extremely important from the provider perspective as they represent the large monetary investment that is made in implementing EBPs and other programs at the expense of direct billing for providing healthcare services.
Regarding the presentation of summary cost estimates, three studies examined the same EBP (trauma-focused cognitive behavioral therapy), which facilitated comparison. However, even among these studies there was wide variation in reported costs, which may suggest that costs are substantially impacted by factors outside of the implementation process. Some of the reported costs were also relatively high (as much as $3,805 per child), and may be inflated due to the inclusion of fixed startup costs (i.e., pre-implementation costs that would not vary with the number of children served) and smaller caseloads. In fact, one of the included studies highlights this in reporting that per child and per session incremental implementation costs are highest in the start-up phase and decrease over time as more children are served . However, as is known, the cost of healthcare in the United States is highest in the world, and this includes the costs of behavioral healthcare. As pharmacotherapy is a large component to behavioral healthcare, relatively high drug prices in the United States could be an important driver. For example, one study reported the annual per capita cost of treating an adult for a behavioral health disorder to be $2,745 . Other research has identified that factors outside implementation can impact costs. For example, behavioral health patient caseloads and state level Medicaid and Child Health Insurance Program (CHIP) funding may be associated with lower implementation costs and better retention in behavioral health services programs for justice-involved youth .
Implications for future research
Given the wide variation in the types of factors included in implementation research it is difficult to make comparisons across studies. This is a major limitation of implementation research focusing on behavioral health services, as it becomes difficult to replicate studies or use studies to estimate future costs to inform policy decisions related to budgeting. With increasing emphasis on the economics of implementation science , the field could benefit from adopting standardized guidelines, especially for costing perspectives, approaches (e.g., activity-based costing), instrumentation, and categorization of cost components. In the broader economic evaluation literature, standard methods for cost analysis are established [41, 42]. Recently, standardized approaches for economic analysis in behavioral health were also described . As implementation science grows, similar guidelines need to be developed. Without standardization and harmonization of data across studies, it is difficult to fully assess how the evidence of effectiveness and economic impact generalizes from one study to a broad area of practice or research.
Based on our findings, we recommend that future studies include more details about the specific activities and resources associated with an implementation process, so that other researchers and policy makers can anticipate what costs will be incurred in changing existing or adopting new practices. We also recommend that studies include measures of environmental context such as treatment capacity, available funding (i.e.: block grants) and urbanicity, all available from public sources, to better capture how outside factors may impact the costs of implementing EBPs . As shown in two studies, urbanicity was related to behavioral health EBP delivery [45, 46]. Beyond impacting the effectiveness of the EBP, contextual factors impact costs directly. For example, salaries may be impacted by geography due to higher costs of living in certain areas.
Limitations and strengths
While this study adds substantially to the literature by describing the state of the field to date, there are still several limitations. For example, since there was very little overlap between studies in the types of EBP being implemented or outcome measures, we were unable to conduct a meta-analysis. Our findings were also focused on implementation studies of evidence-based behavioral health services, and therefore do not extend to other health-related services. Further, only one included study reported both direct implementation costs and direct service costs , which limited our ability to estimate the ratio of implementation to service costs. Concerns about double counting staff time or other resources across direct implementation and direct service cost categories could not be explored in this study given the level of detail provided. Additionally, we also focused only on nonhospital-based services within the United States and Canada, therefore excluding hospital-based services. Since healthcare systems and behavioral health epidemiology in countries outside the United States and Canada are quite different, this decision likely improved the specificity of our findings. However, we are unable to generalize outside of this geographical area. Additional research focused on single-payer healthcare systems, like in the United Kingdom, that would illuminate the role of costing in both policy decisions and implementation of services. While these choices likely improve selection process, there are likely other important areas of research conducted outside the scope of this work. Future studies should aim to describe findings in these areas. Had we included hospital-based services and health services more broadly, it is likely that there would have been greater variation in costing approaches and results.
Implementation research on EBPs in behavioral health has grown significantly in the last several years. However, the field has not yet standardized the use of economic methods or measures. Recommendations based on our findings include moving toward standard cost measures to facilitate cross study comparisons and the potential for results to drive policy decisions.
Availability of data and materials
Data can be obtained by replicating the key search terms in the listed academic databases.
Preferred Reporting Items for Systematic Reviews and Meta-Analyses
Consolidated Health Economic Evaluation Reporting Standards
Substance use disorders
Consolidated Framework for Implementation Research
Child Health Insurance Program
Bhattacharyya O, Reeves S, Zwarenstein M. What is implementation research? Rationale, concepts, and practices. Res Soc Work Pract. 2009;19(5):491–502.
Bauer MS, Damschroder L, Hagedorn H, Smith J, Kilbourne AM. An introduction to implementation science for the non-specialist. BMC Psychol. 2015;3(1):32.
Gaudiano BA, Dalrymple KL, D’Avanzato C, Bunaciu L. The need for quality improvement in behavioral health. In: Quality improvement in behavioral health. Switzerland: Springer; 2016. p. 33–54.
Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Admin Policy Ment Health. 2011;38(2):65–76.
Raghavan R. In: Brownson RC, GAC, Proctor EK, editors. The role of economic evaluation in dissemination and implementation research. Oxford: Oxford University Press; 2018. p. 113.
Grimshaw J, Thomas R, MacLennan G, Fraser C, Ramsay C, Vale L, et al. Effectiveness and efficiency of guideline dissemination and implementation strategies. 2004.
Roberts SLE, Healey A, Sevdalis N. Use of health economic evaluation in the implementation and improvement science fields—a systematic literature review. Implement Sci. 2019;14(1):72.
Donroe JH, Socias ME, Marshall BD. The deepening opioid crisis in North America: historical context and current solutions. Curr Addict Rep. 2018;5(4):454–63.
Rochefort DA. Making Single-Payer Reform Work for Behavioral Health Care: Lessons From Canada and the United States. Int J Health Services.. 2020;50(3):334-49. https://doi.org/10.1177/0020731420906746.
Böhm K, Schmid A, Götze R, Landwehr C, Rothgang H. Five types of OECD healthcare systems: empirical results of a deductive classification. Health Policy. 2013;113(3):258–69.
Moher D, Shamseer L, Clarke M, Ghersi D, Liberati A, Petticrew M, et al. Preferred reporting items for systematic review and meta-analysis protocols (PRISMA-P) 2015 statement. Syst Rev. 2015;4(1):1.
Moher D, Liberati A, Tetzlaff J, Altman DG, The PRISMA Group. Preferred Reporting Items for Systematic Reviews and Meta-Analyses: the PRISMA statement. PLoS Med. 2009;6(6):e1000097.
Lu Z, Kim W, Wilbur WJ. Evaluation of query expansion using MeSH in PubMed. Inform Retrl. 2009;12(1):69–80.
Siddaway AP, Wood AM, Hedges LV. How to do a systematic review: a best practice guide for conducting and reporting narrative reviews, meta-analyses, and meta-syntheses. Ann Rev Psychol. 2019;70:747–70.
Thielen F, Van Mastrigt G, Burgers L, Bramer W, Majoie H, Evers S, et al. How to prepare a systematic review of economic evaluations for clinical practice guidelines: database selection and search strategy development (part 2/3). Expert Rev Pharmacoecon Outcomes Res. 2016;16(6):705–21.
Husereau D, Drummond M, Petrou S, Carswell C, Moher D, Greenberg D, et al. Consolidated health economic evaluation reporting standards (CHEERS) statement. Cost Effectiveness and Resource Allocation. 2013;11(1):6.
Dopp AR, Coen AS, Smith AB, Reno J, Bernstein DH, Kerns SE, et al. Economic impact of the statewide implementation of an evidence-based treatment: multisystemic therapy in New Mexico. Behav Ther. 2018;49(4):551–66.
Dopp AR, Hanson RF, Saunders BE, Dismuke CE, Moreland AD. Community-based implementation of trauma-focused interventions for youth: economic impact of the learning collaborative model. Psychol Serv. 2017;14(1):57.
Fisher E, Hasselberg M, Conwell Y, Weiss L, Padrón NA, Tiernan E, et al. Telementoring primary care clinicians to improve geriatric mental health care. Popul Health Manage. 2017;20(5):342–7.
Greer D, Grasso DJ, Cohen A, Webb C. Trauma-focused treatment in a state system of care: Is it worth the cost? Administration and Policy in Mental Health and Mental Health Services Research. 2014;41(3):317–23.
King DK, Gonzalez SJ, Hartje JA, Hanson BL, Edney C, Snell H, et al. Examining the sustainability potential of a multisite pilot to integrate alcohol screening and brief intervention within three primary care systems. Transl Behav Med. 2018;8(5):776–84.
Knusden HK, Roman PM. Innovation attributes and adoption decisions: perspectives from leaders of a national sample of addiction treatment organizations. J Substance Abuse Treat. 2015;49:1–7.
Lang JM, Connell CM. Measuring costs to community-based agencies for implementation of an evidence-based practice. J Behav Health Serv Res. 2017;44(1):122–34.
Lee TG, Walker SC, Bishop AS. The impact of psychiatric practice guidelines on medication costs and youth aggression in a juvenile justice residential treatment program. Psychiatric Serv. 2015;67(2):214–20.
McKee SA, Harris GT, Cormier CA. Implementing residential integrated treatment for co-occurring disorders. J Dual Diagnosis. 2013;9(3):249–59.
Palinkas LA, Um MY, Jeong CH, Chor KHB, Olin S, Horwitz SM, et al. Adoption of innovative and evidence-based practices for children and adolescents in state-supported mental health clinics: a qualitative study. Health Res Policy Syst. 2017;15(1):27.
Rollins AL, Kukla M, Salyers MP, McGrew JH, Flanagan ME, Leslie DL, et al. Comparing the costs and acceptability of three fidelity assessment methods for Assertive Community Treatment. Administration and Policy in Mental Health and Mental Health Services Research. 2017;44(5):810–6.
Roundfield KD, Lang JM. Costs to community mental health agencies to sustain an evidence-based practice. Psychiatric Serv. 2017;68(9):876–82.
Saldana L. The stages of implementation completion for evidence-based practice: protocol for a mixed methods study. Implement Sci. 2014;9(1):43.
Shi Y, Scanlon DP, Kang R, McHugh M, Greene J, Christianson JB, et al. The longitudinal impact of Aligning Forces for Quality on measures of population health, quality and experience of care, and cost of care. Am J Manag Care. 2016;22(suppl 12):S373–S81.
Stewart RE, Adams DR, Mandell DS, Hadley TR, Evans AC, Rubin R, et al. The perfect storm: collision of the business of mental health and the implementation of evidence-based practices. Psychiatric Serv. 2016;67(2):159–61.
Suarez E, Jackson DS, Slavin LA, Michels MS, McGeehan KM. Project Kealahou: improving Hawai ‘i's system of care for at-risk girls and young women through gender-responsive, trauma-informed care. Hawai'i J Med Public Health. 2014;73(12):387.
Swain K, Whitley R, McHugo GJ, Drake RE. The sustainability of evidence-based practices in routine mental health agencies. Commun Ment Health J. 2010;46(2):119–29.
Swenson CC, Randall J, Henggeler SW, Ward D. The outcomes and costs of an interagency partnership to serve maltreated children in state custody. Childrens Serv. 2000;3(4):191–209.
Mihalopoulos C, Chatterton ML. Economic evaluations of interventions designed to prevent mental disorders: a systematic review. Early Interv Psychiatry. 2015;9(2):85–92.
Mihalopoulos C, Vos T, Pirkis J, Carter R. The economic analysis of prevention in mental health programs. Ann Rev Clin Psychol. 2011;7:169–201.
Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4(1):50.
Wensing M, Grol R. Knowledge translation in health: how implementation science could contribute more. BMC Med. 2019;17(1):88.
Thorpe K, Jain S, Joski P. Prevalence and spending associated with patients who have a behavioral health disorder and other conditions. Health Affairs. 2017;36(1):124–32.
Bowser D, Henry BF, McCollister KE. An overlapping systems conceptual framework to evaluate implementation of a behavioral health intervention for justice–involved youth. Health Serv Insights. 2019;12:1178632919855037.
Drummond MF, Sculpher MJ, Claxton K, Stoddart GL, Torrance GW. Methods for the economic evaluation of health care programmes. Oxford: University Press; 2015.
Neumann PJ, Sanders GD, Russell LB, Siegel JE, Ganiats TG. Cost-effectiveness in health and medicine: Oxford University Press; 2016.
Kaplan RM, Gold M, Duffy SQ, Miller N, Glassman JR, Chambers DA, et al. Economic analysis in behavioral health: toward application of standardized methodologies. Health Psychol. 2019;38(8):672.
Bowser D, Henry BF, Wasserman GA, Knight D, Gardner S, Krupka K, et al. Comparison of the overlap between juvenile justice processing and behavioral health screening, assessment and referral. J Appl Juv Justice Serv. 2018;2018:97–125.
Robertson AA, Hiller M, Dembo R, Dennis M, Scott C, Henry BF, et al. National survey of juvenile community supervision agency practices and caregiver involvement in behavioral health treatment. J Child Fam Stud. 2019;28(11):3110–20.
Marks KR, Leukefeld CG, Dennis ML, Scott CK, Funk R, Cooperative J-T. Geographic differences in substance use screening for justice-involved youth. J Substance Abuse Treat. 2019;102:40–6.
This study was funded by the National Institute on Drug Abuse (NIDA), and the National Institute on Alcohol Abuse and Alcoholism (NIAAA) both at the National Institutes of Health (NIH) (T32AA007567, T32DA37801, R21DA044378, P30 DA040500, R01DA035808). The contents of this publication are solely the responsibility of the authors and do not necessarily represent the official views of the NIAAA, NIDA, NIH, or the participating universities.
Ethics approval and consent to participate
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Bowser, D.M., Henry, B.F. & McCollister, K.E. Cost analysis in implementation studies of evidence-based practices for mental health and substance use disorders: a systematic review. Implementation Sci 16, 26 (2021). https://doi.org/10.1186/s13012-021-01094-3
- Evidence-based practices
- Behavioral health