Implementation strategies are necessary to ensure that evidence-based practices are successfully incorporated into routine clinical practice. Such strategies, however, are frequently modified to fit local populations, settings, and contexts. While such modifications can be crucial to implementation success, the literature on documenting and evaluating them is virtually nonexistent. In this paper, we therefore describe the development of a new framework for documenting modifications to implementation strategies.
We employed a multifaceted approach to developing the Framework for Reporting Adaptations and Modifications to Evidence-based Implementation Strategies (FRAME-IS), incorporating multiple stakeholder perspectives. Development steps included presentations of initial versions of the FRAME-IS to solicit structured feedback from individual implementation scientists (“think-aloud” exercises) and larger, international groups of researchers. The FRAME-IS includes core and supplementary modules to document modifications to implementation strategies: what is modified, the nature of the modification (including the relationship to core elements or functions), the primary goal and rationale for the modification, timing of the modification, participants in the modification decision-making process, and how widespread the modification is. We provide an example of application of the FRAME-IS to an implementation project and provide guidance on how it may be used in future work.
Increasing attention is being given to modifications to evidence-based practices, but little work has investigated modifications to the implementation strategies used to implement such practices. To fill this gap, the FRAME-IS is meant to be a flexible, practical tool for documenting modifications to implementation strategies. Its use may help illuminate the pivotal processes and mechanisms by which implementation strategies exert their effects.
Implementation strategies are typically required to ensure that evidence-based practices (EBPs) are adopted in real-world clinical settings.
These implementation strategies, like EBPs themselves, are often modified based on the clinical setting or patient population being served.
To our knowledge, the field lacks a systematic way of documenting and tracking modifications to implementation strategies, complicating efforts to understand the mechanisms by which those implementation strategies exert their effects.
To fill this gap, in this paper, we describe the development of the Framework for Reporting Adaptations and Modifications to Evidence-based Implementation Strategies (FRAME-IS).
Systematic use of the FRAME-IS will allow healthcare system leaders, administrators, and implementation scientists to better understand the ways that implementation strategies may be modified to improve health outcomes across settings and populations.
Implementation science aims to maximize the adoption, use, and sustainability of evidence-based practices (EBPs) in real-world healthcare settings. These EBPs can include specific medications, treatment algorithms, manualized therapies, and ways of structuring care. Adaptation of an EBP has been defined as the deliberate alteration of its design or delivery to improve its fit in a given context [1, 2]. For example, an evidence-based psychotherapy that is traditionally completed in 12 sessions could be shortened for use in a new clinical setting if 12 sessions is shown to be burdensome, too costly, or unnecessary for that patient population. Modifications to EBPs may be made in either an ad hoc or planful manner; the latter category of planned modifications may be labeled as adaptations [3, 4]. Despite the importance of modifications for improving EBP effectiveness  and decreasing healthcare disparities , they remain understudied in implementation science. Recent work has aimed to fill this gap by fleshing out ways to design, document, and evaluate modifications to EBPs (e.g., [7,8,9,10]).
It is increasingly evident that similar processes of tailoring and modification occur for implementation strategies as they are used to support implementation of EBPs in different contexts. Implementation strategies are methods or techniques used to adopt, implement, and sustain EBPs in routine practice . Implementation strategies range from relatively “light touches” (e.g., audit and feedback ) to more intensive strategies that include multiple components and may act on more than one level of a health system (e.g., implementation facilitation ). Recent work suggests little consensus in the field regarding the selection of specific implementation strategies for a given implementation project, or the mechanisms of action for those strategies . As such, scholars have suggested that implementation strategies should be theoretically or empirically driven, and their components (e.g., actions, actors, goals) described so as to promote replicability and testing of their mechanisms of action [15,16,17].
Concurrently, scholars acknowledge the importance of modifying implementation strategies to fit with the unique characteristics of the EBP, setting, and stakeholders involved in an implementation effort . For example, consider provider training as an implementation strategy. A traditional training may involve in-person workshops, but due to the COVID-19 pandemic, many trainings have been adapted to be conducted online. In addition to tracking the specific implementation strategies being used in a given project [19, 20], we advocate for the need to document and evaluate modifications to implementation strategies that have been well-defined and characterized. Although implementation strategies should always be operationalized to fit each context, documentation and reporting of these adaptations has to date not been consistently undertaken. It is crucially important to develop a more nuanced understanding of modifications to implementation strategies because of the inherent dynamism and complexity of implementation itself . Without such an understanding, it is difficult to determine the processes or mechanisms by which implementation strategies exert their effects on implementation outcomes. Furthermore, because many implementation strategies may consist of several components delivered in dynamic contexts (e.g., ), changes are typically required to one or more of those components to improve fit or address constraints. Indeed, the modification of implementation strategies fits within co-creation models, in which researchers and stakeholders collaborate to exchange, generate, and utilize knowledge [21,22,23]. Local modifications are key to co-creation models , which provide opportunities to generate collaborative knowledge about how modifications to implementation strategies impact clinical and implementation outcomes.
A first step for understanding modifications to implementation strategies is to develop a framework for characterizing those modifications across studies and settings. This can allow implementation scientists, healthcare leaders, and quality improvement specialists to track when, why, and how implementation strategies have been modified. Akin to developing a database of modifications to EBPs , we propose that systematic documentation is needed to determine what modifications to implementation strategies are associated with successful versus unsuccessful implementation.
Based on our previous work on developing a framework for tracking adaptations of EBPs, we have used the Framework for Reporting Adaptations and Modifications – Expanded (FRAME)  as a foundation from which to build a new framework, focusing on modifications that may be made to implementation strategies rather than EBPs. As described in more detail below, we found that a new framework was necessary because our pilot work highlighted key differences in tracking modifications to implementation strategies versus EBPs. Thus, the Framework for Reporting Adaptations and Modifications to Evidence-based Implementation Strategies (FRAME-IS) aims to meet the need in the implementation science field for a tool to document modifications to implementation strategies. Below, we describe the development of the FRAME-IS and include an example of its application to a recent implementation trial.
Development of the FRAME-IS
The FRAME-IS is based on the FRAME, the development of which is described in detail elsewhere [10, 26] Briefly, the FRAME is a framework for tracking modifications to EBPs focused on what to adapt, alongside a related adaptation framework blending the content and processes of adaptation [27, 28]. The FRAME was developed based on earlier research , combined with results from a literature review, focus groups, and coding process rooted in grounded theory [29, 30].
Our process for in turn developing the FRAME-IS can be found in Table 1. We began with the FRAME  and used it to develop a first draft of the FRAME-IS by changing the wording to refer to modifications to implementation strategies instead of EBPs. We also changed the language regarding the personnel involved in the modification (e.g., emphasizing the role of implementers themselves in the modification process). The authors then used the preliminary version of the FRAME-IS to describe modifications to implementation strategies in their own ongoing and completed projects, which informed further refinement (e.g., ). While the FRAME-IS is meant to be applicable to implementation efforts in traditional healthcare settings, it also includes language related to educational settings where some healthcare interventions may be implemented.
We then presented the FRAME-IS to several groups of stakeholders, including implementation experts from our networks, as well as international groups of implementation scientists and practitioners. These groups included the Washington University Network for Dissemination and Implementation Researchers (WUNDIR ;) and the VA Quality Enhancement Research Initiative Implementation Research Group (QUERI IRG ;). During these presentations, stakeholders (total n > 50) noted the importance of streamlining the FRAME-IS (e.g., eliminating some options, emphasizing a modular approach). We also used those presentations to identify volunteers to complete one-on-one “think-aloud” sessions . Thinking aloud requires stakeholders to talk aloud while performing a task, allowing developers to understand their experiences and perspectives to shape the product to better meet their needs and constraints. This method has been applied in research on cognitive processes and in human-centered design approaches to product development. In our case, think-aloud participants (n = 6) were asked to review the FRAME-IS, apply it to implementation strategies while verbalizing their thought process as they did so, and provide general feedback afterward. The authors facilitated this process and took notes, which were reviewed by the full author team and used to refine the FRAME-IS. These sessions took about 30 min to complete and focused on a variety of implementation strategies including implementation facilitation, provider training, and audit/feedback. Aggregate results suggested the need for additional streamlining and clarification of the answer options to be included in the FRAME-IS. Stakeholders also suggested that the module documenting the rationale for the modification (Module 4 below) should be tied more closely to existing theories and frameworks (e.g., [16, 35]). Incorporating these suggested revisions led to the final version, which consists of a series of modules meant to capture different aspects of the modification process. The resulting FRAME-IS is multifaceted and is intended to be applicable to a variety of implementation strategies and contexts. Below, we describe each component of the FRAME-IS, as well as literature relevant to its conceptualization and challenges that may emerge in its use.
Components of the FRAME-IS
Based on stakeholder feedback, the FRAME-IS is modular and includes both core (i.e., required) and optional modules to maximize its practicality across implementation projects with a variety of goals, priorities, and available resources. Core modules can be found in Fig. 1, and optional modules can be found in Fig. 2. The decision regarding which modules to designate as core versus optional was made by the authors based on consensus discussions, incorporating stakeholder feedback. Additional descriptive text can be found in the supplement. Completing core modules requires specifying: a brief description of the EBP, implementation strategy, and modifications (Module 1); what is modified (Module 2); the nature of the modification (content, evaluation, or training modifications only; Module 3); and the rationale for the modification (Module 4). Optional modules, which can be completed at the discretion of the study investigators or project team, involve specifying when the modification occurred, and whether it was planned (Module 5); who participated in the decision to modify (Module 6); and how widespread the modification is (Module 7). Module 3, while itself considered a core module, includes the option of documenting the extent to which the modification was considered fidelity-consistent with the original implementation strategy.
Module 1: Brief description of the EBP, implementation strategy, and modification(s)
To facilitate tracking modifications and to complete the remainder of the FRAME-IS, we recommend briefly describing the EBP in question, and the initially defined implementation strategy. The Expert Recommendations for Implementing Change (ERIC) compilation  may be useful in describing the implementation strategy. If it is feasible, more comprehensive guidance  may be used to describe the implementation strategy in more detail to delineate the core elements, processes, or functions of the strategy. We also suggest identifying potential initial modifications to the strategy. We note that many modifications may be “bundled”—i.e., may involve changes to multiple aspects of the implementation strategy. For example, the content and the length of a provider training may be modified simultaneously. In those cases, the research or implementation team can determine whether to complete the FRAME-IS separately for each modification or to document all of the separate modifications at once. Completing it separately for each individual modification may allow for finer-grained analysis of what was most helpful—but may also represent an undue documentation burden in cases where separate components of a modification cannot be disentangled. Similarly, depending on the goal of the project, users should determine whether it is best to use the FRAME-IS separately for each component of a multifaceted implementation strategy or to define the entire strategy and modification in Module 1. For example, Leadership and Organizational Change for Implementation (LOCI ) is a multi-component implementation strategy that includes coaching calls. If the modification in question involved changing the frequency of LOCI coaching calls, the research team would need to decide whether the change in coaching call frequency should be documented alongside modifications that may be made simultaneously to other LOCI components.
Module 2: What is modified?
The FRAME-IS includes four broad categories of modifications to implementation strategies: Content, Context, Evaluation, and Training. These four categories mirror those included in the FRAME, but with some key distinctions. A Context modification refers to changes to the setting or the way the implementation strategy is delivered. For example, if the implementation strategy being modified was implementation facilitation, the context could change if facilitation was provided virtually as opposed to in person. We note that there are some distinctions in Content and Training modifications to implementation strategies. For example, if the implementation strategy in question is a clinician training workshop, then modifications to the content of the workshop itself would qualify as a Content modification (because it is part of the package of implementation strategies). In contrast, changes to how implementers are trained—e.g., by modifying the ways that external facilitators are trained in the context of implementation facilitation—would qualify as a Training modification. Modifications to evaluation refer to changes in the way that an implementation strategy is evaluated. For example, recent work has aimed to uncover the “core activities” of implementation facilitation ; ongoing studies are incorporating assessments of these core activities into their implementation evaluation plans (e.g., ).
Module 3: What is the nature of the content, evaluation, or training modification?
For Content, Evaluation, and Training modifications, it is important to track the nature of the modification itself. These modifications can range from relatively small tweaks or more sweeping changes to the implementation strategy—or even abandonment of an implementation strategy or component altogether. The elements of the FRAME-IS in this domain closely mirror those of the FRAME, but with language specifying modifications made to an implementation strategy rather than an EBP. Adding or removing/skipping elements refer to specific aspects of a discrete implementation strategy (e.g., removing role plays from a training workshop) or package of strategies (e.g., removing feedback, adding incentives in a leadership program). Pacing may refer to pacing of training or frequency of feedback. Integration refers to incorporation of other implementation strategies or approaches (e.g., adding feedback or incentives to training and supervision when the first round of implementation suggests that results are not as robust as desired). Loosening structure might include coaching “on the fly” or immediately after a clinical interaction rather than at scheduled times. Substituting may include adaptations such as changing fidelity assessment from observer-rated to self-report.
As an optional portion of this module, practitioners or researchers may also be interested in documenting the extent to which the modification was conducted with fidelity . In this context, fidelity may be defined as “the adherence to the intervention components, competence with which the intervention is delivered, and differentiation from other treatments” . Fidelity-consistent modifications may be expected to have better outcomes than fidelity-inconsistent ones, which may represent “drift” as they remove core elements of the strategy. Core elements or functions, in this context, refer to components or topics considered essential to that implementation strategy . For example, removing the feedback component from an audit and feedback implementation strategy would likely be fidelity-inconsistent. We acknowledge, though, that the relationships among fidelity, adaptation, and outcomes are complex and not fully understood . Fundamental questions regarding balancing adaptation and fidelity have yet to be answered—especially for the substantial proportion of implementation strategies that are meant to be inherently adaptable, flexible, or modular (e.g., ). Thus, completion of this portion of the module may not be warranted in all cases.
Module 4: What is the rationale for the modification?
The goal of this module is to document why a given modification was made to a given implementation strategy. This may allow interested parties to determine what rationales are associated with more or less successful modifications. We break the rationale into two components. First, regarding the goal of the modification, we have derived answer options in part from the Reach – Effectiveness – Adoption – Implementation – Maintenance ((RE-AIM) Framework  and Proctor’s implementation outcomes , plus one option related to health equity. We note that some goals may be related to the implementation effort itself, while others may be more directly related to the EBP being implemented. We also note that many modifications may aim to achieve multiple goals. For example, shortening an EBP training may simultaneously reduce costs, increase adoption, and increase reach. In completing this Module, we recommend selecting the box corresponding to the primary goal of the modification or selecting multiple boxes in the context of several co-equal goals.
Second, we recommend documenting the level of the organization that most directly informed the modification. For example, modifications made to accommodate available staffing at a clinic would qualify as the organizational level, while modifications made to fit with the professional or cultural values of frontline staff delivering the EBP would qualify as the practitioner level.
Module 5: When the modification is initiated, and whether it is planned
The timing of modifications is crucial in implementation science, with the pre-implementation, implementation, and sustainment phases featuring distinct pressures, challenges, and goals [37, 41, 42]. Modifications made early in the implementation process may leave more time for implementers and practitioners to adjust to the change. Modifications made later (during the implementation or sustainment phases) may nonetheless be required to accommodate shifting priorities or resources (e.g., shifting initial training to a web-based format based on travel restrictions). Documenting the timing of modifications to implementation strategies will allow the field to develop a better understanding of how such timing affects implementation processes and outcomes.
Note that for our purposes here the primary goal is to document when a modification is initiated, rather than when it occurs, as many modifications to implementation strategies may happen over large portions of the implementation period. For example, consider an implementation strategy of provider training to increase uptake of an evidence-based psychotherapy. If the timing of those trainings is modified, the important question for this section is when the decision was made to change the timing rather than when the (modified) trainings were offered. Documenting when decisions regarding modifications are made may ultimately help shed light on whether modifications made early versus late in the implementation period are differentially successful.
Regardless of the phase in which modifications to implementation strategies are made, there is a conceptual distinction between those that are made in a planful versus reactive manner [43, 44]. For example, consider a hypothetical implementation project featuring a learning collaborative that was originally designed to meet face-to-face four times during the implementation year. Let us further imagine that, due to budget constraints, only two learning collaborative meetings could be funded. If the budget constraints were known early in the implementation year, then the implementer and/or content experts would likely be able to modify the curriculum or format to ensure that all core content of the learning collaborative is covered (i.e., an adaptation or planned modification). Such proactive changes could include, for example, establishing or expanding a virtual component for the learning collaborative, or lengthening the two face-to-face sessions that could be funded. In contrast, having two face-to-face learning collaborative meetings abruptly canceled midway through the implementation year (i.e., a reactive modification) might necessitate more substantial changes to the curriculum that leaves some core content unaddressed. Differentiating unplanned, reactive modifications from proactive and planful adaptations will allow implementation scientists to better understand the circumstances under which impromptu modifications to implementation strategies may be more or less helpful.
Module 6: Who participates in the decision to modify?
In some cases, modifications to implementation strategies may be made in a collaborative manner, with multiple stakeholders or “actors” (e.g., administrators, frontline clinicians, implementation specialists) agreeing that a given modification is needed in a given setting . In other cases, the decision to modify an implementation strategy may be unilateral (as when a health system leader requires that a given implementation strategy be scaled back based on personnel changes or competing priorities). Documenting this distinction can inform future decisions regarding when broader consensus on certain types of modifications to implementation strategies is required for implementation success, consistent with the principles of stakeholder engagement . As researchers start to grapple with the intersection of implementation science and health equity , carefully identifying who is suggesting the modifications may be an important aspect of tracking the co-creation of implementation strategies [21,22,23,24]. Identifying sources of power in the implementation process , and incorporating the voices of those in the community, will be important for the field as we move toward equitable practice in our science.
Module 7: How widespread is the modification?
For modifications documented in Module 2 as Content, Training, and Evaluation modifications, it may be important to document the breadth or scope of the modification to the implementation strategy. This can range from relatively narrow modifications (e.g., in the context of an individual consultation call for a clinician who missed a day of group consultation) to broad-based ones (e.g., modifications made by an entire health system that is using an implementation strategy to roll out an EBP).
Note that some of the answer options refer to the individuals receiving the EBP, while others refer to the practitioners delivering the EBP, and yet others refer to those tasked with supporting the use of the EBP. It is possible that boxes within all three of these categories could be checked. If a single implementation facilitator adds an audit and feedback component to an implementation facilitation strategy within one clinic (and no other facilitators are using the unmodified strategy within the clinics they oversee), then that would qualify as a modification at one clinic/unit and one specific implementer/facilitator.
Case example: Application of FRAME-IS in a recent implementation trial
Here, and in Table 2, we illustrate application of the FRAME-IS in an ongoing trial that seeks to train lay health workers to task-share with mental health professionals to improve engagement for Spanish-speaking Latinx families receiving Parent-Child Interaction Therapy (PCIT ;). PCIT is an evidence-based parenting program for young children with disruptive behavior disorders . In this task-sharing model, professional clinicians provide PCIT and lay health workers conduct outreach and promote treatment adherence to improve access, adherence, and skill acquisition for families . Lay health workers have been identified as an important workforce to decrease disparities in access to care for marginalized communities . However, limited research has identified what implementation supports lay health workers need to successfully engage in EBPs.
Module 1: Description of the implementation strategy
In this case example, we will apply the FRAME-IS to the training model for lay health workers that was developed to support parental engagement in PCIT within a university-based clinic in Miami, Florida. Initial training aimed to prepare lay health workers to identify and refer appropriate cases to PCIT, teach them the parenting skills targeted in the program (e.g., giving specific praises of positive behaviors), and prepare them to promote adherence to home practice of these skills. Lay health workers were taught how to use a behavioral coding system in PCIT; they were also taught to provide feedback on how to improve weaker skills and reinforce skills that parents were using well. Each lay health worker was provided with an e-book with videos and scripts to help them describe PCIT and the parenting skills to caregivers. Lay health workers in the initial training demonstrated improved knowledge of PCIT over the course of training. They also increased their ability to model the parenting skills taught in PCIT, in order to help parents use them in the home. However, it was challenging for lay health workers to conduct behavioral observations of parent’s skill use and provide accurate feedback on which skills to improve .
Module 2: What was modified?
The original training model was adapted for community-based mental health settings within California (context), with changes to the training program content to meet this context and address challenges identified in the initial training program evaluation.
Module 3: The nature of the content modification
The majority of modifications involved tailoring (i.e., making minor changes to) the training materials to fit with the local context. Furthermore, the focus on training the lay health workers to conduct behavioral coding was removed from the training, as this was not considered a feasible or necessary skill. Instead, lay health workers were taught how to provide general feedback regarding parenting skills.
Module 4: The rationale for the modification
The goal of the adaptations made was to increase the acceptability, appropriateness, and feasibility of the training to fit with cultural differences for providers and clients across the two contexts. Trainings in Miami and California were conducted in Spanish; however, regional language differences needed to be addressed. The Latinx population served in Miami was predominantly from the Caribbean (e.g., Cuba), South America (e.g., Colombia), and Central America (e.g., Honduras), whereas the population served in California was predominantly from Mexico. Therefore, certain vocabulary and idioms were tailored in training materials. New videos were created for the e-book, featuring children and parents of Mexican descent, to increase cultural and linguistic similarities with the families being served. Behavioral coding was removed from the training to increase the feasibility of training and to ensure adequate fidelity (skill at providing accurate feedback to the parents).
Module 5: When the modification was initiated and planned
Modifications of the training protocol occurred during the pre-implementation/planning phase of the current trial and were planned as part of a research-community partnership (i.e., were proactively made prior to beginning training).
Module 6: Who participated in decisions to modify?
Researchers, who developed and evaluated the initial training in Miami, led decisions surrounding the modifications to the training program based on data from the Miami project. They incorporated feedback from local program leaders and lay health workers, who participated in qualitative interviews and a community-advisory board. Ultimately, this feedback led to modifications that the research team finalized.
Module 7: How widespread is the modification?
Modifications were made for the entire network system/community, specifically for lay health workers living in California and working within community-mental health settings.
Recent implementation science work has emphasized the importance of documenting modifications to EBPs [8, 10]. Implementation strategies are used to support the implementation of EBPs, and scholars have recommended specifying components of implementation strategies  to support their reproducibility and further elucidate mechanisms of change. As implementation strategies are deployed, however, they can be modified. Comprehensively documenting modifications to implementation strategies will allow the field to study the relationship between those modifications and implementation outcomes across settings. While previous frameworks (e.g., ) are meant to track modifications to EBPs, the FRAME-IS is to our knowledge the first tracking tool developed specifically for modifications to implementation strategies, incorporating modular coding and a novel focus on those implementing the EBP in question (see Supplemental File). Systematic application of the FRAME-IS to implementation projects—alongside careful assessment of implementation outcomes using RE-AIM  or similar evaluation-oriented implementation frameworks—will ideally help to answer fundamental questions about whether, when, how, and why an implementation strategy has been effective in a given setting. This will allow for a co-creation of collaborative knowledge between implementation researchers and stakeholders [21,22,23,24], as the FRAME-IS may guide systematic adaptations to implementation strategies when planning a project, and a way to track ad hoc modifications made throughout the project. The FRAME-IS may also be useful to test hypotheses related to core components or functions of implementation strategies: if a component of an implementation strategy is modified in some way, and that modified implementation strategy’s effectiveness rivals or surpasses that of its unmodified version, then it raises the possibility that the modified component (in its original form) was not in fact central to the implementation strategy’s success and that there are alternative, adaptive forms of the component that can be deployed successfully.
The FRAME-IS is not without limitations, of course. First, completing the FRAME-IS may be difficult in the context of subtle modifications that emerge longitudinally as implementation progresses (e.g., ongoing tailoring of written training materials), depending on who completes the reporting and how frequently it occurs. It may also be difficult to apply the FRAME-IS to multi-component implementation strategies (e.g., implementation facilitation ) that are inherently adaptable or intended to be tailored without guidance from the developers of the implementation strategy. In these circumstances, it may still represent a useful tool for tracking the specific ways that the implementation strategy was applied—albeit one that may require time-intensive multi-method assessment to achieve acceptable validity. It is also unclear whether local implementers are able to accurately report on modifications to implementation strategies as they occur in real time, or whether supplemental personnel are required. Local implementer reports may be necessary when expert observation is not feasible or scalable; in those situations, adequate training in the use of the FRAME-IS will be pivotal. One thing that remains to be determined is how frequently reporting should occur to capture the full extent of modifications. We also note that the FRAME-IS is meant to capture modifications to one a priori identified implementation strategy. Clear operationalization of what the original strategy entails is essential to accurately track adaptations. For projects featuring multiple implementation strategies rolled out over time (e.g., ), additional tracking (e.g., informed by the ERIC compilation  and with well-specified descriptions of the strategies as originally designed) may also be required to ensure a robust understanding not just of the modification process, but of how implementation proceeded more broadly.
Implementation strategies often undergo modification to improve their fit with the EBPs, target populations, and clinical contexts in which they are applied. Without systematically tracking such modifications, it will be difficult for the implementation science field to determine how best to maximize their effectiveness or fully address health disparities . To meet this need, the FRAME-IS is meant to be a first step toward better assessing the ways that implementation strategies may be modified as EBPs are put into practice. Fundamental questions, such as what role the FRAME-IS could play in designing, analyzing, and publishing implementation trials, will need to be explored in future work. Currently, the FRAME-IS is being piloted in a multinational ten-site implementation study; this and other applications of the FRAME-IS may help answer such questions regarding best practices for formatting, administering, and tracking its completion. It is our hope that application of the FRAME-IS more broadly may shed light on the processes and mechanisms by which implementation strategies exert their effects, ultimately improving the uptake of EBPs across settings.
Availability of data and materials
Framework for Reporting Adaptations and Modifications – Expanded
Framework for Reporting Adaptations and Modifications to Evidence-based Implementation Strategies
Washington University Network for Dissemination and Implementation Researchers
Stirman SW, Gutner CA, Crits-Christoph P, Edmunds J, Evans AC, Beidas RS. Relationships between clinician-level attributes and fidelity-consistent and fidelity-inconsistent modifications to an evidence-based psychotherapy. Implement Sci. 2015;10(1):1–10.
Wiltsey Stirman S, Gamarra JM, Bartlett BA, Calloway A, Gutner CA. Empirical examinations of modifications and adaptations to evidence-based psychotherapies: Methodologies, impact, and future directions. Clin Psychol Sci Pract. 2017;24(4):396–420.
Barrera M, Berkel C, Castro FG. Directions for the advancement of culturally adapted preventive interventions: local adaptations, engagement, and sustainability. Prev Sci. 2017;18(6):640–8. https://doi.org/10.1007/s11121-016-0705-9.
Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implement Sci. 2013;8(1):117. https://doi.org/10.1186/1748-5908-8-117.
Bernal GE, Domenech Rodríguez MM. Cultural adaptations: tools for evidence-based practice with diverse populations. Washington, DC: American Psychological Association; 2012.
Escoffery C, Lebow-Skelley E, Haardoerfer R, Boing E, Udelson H, Wood R, et al. A systematic review of adaptations of evidence-based public health interventions globally. Implement Sci. 2018;13(1):125. https://doi.org/10.1186/s13012-018-0815-9.
Escoffery C, Lebow-Skelley E, Udelson H, Böing EA, Wood R, Fernandez ME, et al. A scoping study of frameworks for adapting public health evidence-based interventions. Transl Behav Med. 2019;9(1):1–10. https://doi.org/10.1093/tbm/ibx067.
Miller CJ, Wiltsey-Stirman S, Baumann AA. Iterative Decision-making for Evaluation of Adaptations (IDEA): a decision tree for balancing adaptation, fidelity, and intervention impact. J Community Psychol. 2020;48(4):1163–77. https://doi.org/10.1002/jcop.22279.
Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):21. https://doi.org/10.1186/s13012-015-0209-1.
Kirchner JE, Ritchie MJ, Pitcock JA, Parker LE, Curran GM, Fortney JC. Outcomes of a partnered facilitation strategy to implement primary care–mental health. J Gen Intern Med. 2014;29(4):904–12. https://doi.org/10.1007/s11606-014-3027-2.
Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011;38(2):65–76. https://doi.org/10.1007/s10488-010-0319-7.
Rogal SS, Yakovchenko V, Morgan T, Bajaj JS, Gonzalez R, Park A, et al. Getting to implementation: a protocol for a Hybrid III stepped wedge cluster randomized evaluation of using data-driven implementation strategies to improve cirrhosis care for Veterans. Implement Sci. 2020;15(1):1–10.
Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Serv Res. 2017;44(2):177–94. https://doi.org/10.1007/s11414-015-9475-6.
Boyd MR, Powell BJ, Endicott D, Lewis CC. A method for tracking implementation strategies: an exemplar implementing measurement-based care in community behavioral health clinics. Behav Ther. 2018;49(4):525–37. https://doi.org/10.1016/j.beth.2017.11.012.
Bunger AC, Powell BJ, Robertson HA, MacDowell H, Birken SA, Shea C. Tracking implementation strategies: a description of a practical approach and early findings. Health Res Policy Syst. 2017;15(1):15. https://doi.org/10.1186/s12961-017-0175-y.
Engell T, Follestad IB, Andersen A, Hagen KA. Knowledge translation in child welfare—improving educational outcomes for children at risk: study protocol for a hybrid randomized controlled pragmatic trial. Trials. 2018;19(1):1–17.
Greenhalgh T, Jackson C, Shaw S, Janamian T. Achieving research impact through co-creation in community-based health services: literature review and case study. Milbank Q. 2016;94(2):392–429. https://doi.org/10.1111/1468-0009.12197.
Stirman SW, Miller CJ, Toder K, Calloway A. Development of a framework and coding system for modifications and adaptations of evidence-based interventions. Implement Sci. 2013;8(1):65. https://doi.org/10.1186/1748-5908-8-65.
Baumann AA, Domenech Rodríguez MM, Amador NG, Forgatch MS, Parra-Cardona JR. Parent management training-oregon model (PMTO™) in Mexico City: integrating cultural adaptation activities in an implementation model. Clin Psychol Sci Pract. 2014;21(1):32–47. https://doi.org/10.1111/cpsp.12059.
Rodriguez MMD, Baumann AA, Schwartz AL. Cultural adaptation of an evidence based intervention: from theory to practice in a Latino/a community context. Am J Community Psychol. 2011;47(1-2):170–86. https://doi.org/10.1007/s10464-010-9371-4.
Corbin J, Strauss A. Basics of qualitative research: techniques and procedures for developing grounded theory. Boston: Sage Publications; 2014.
Glaser, B.G. and Strauss, A.L., Discovery of grounded theory: strategies for qualitative research. 2017: Routledge.
Barnett M, Miranda J, Kia-Keating M, Saldana L, Landsverk J, Lau AS. Developing and evaluating a lay health worker delivered implementation intervention to decrease engagement disparities in behavioural parent training: a mixed methods study protocol. BMJ Open. 2019;9(7):e028988. https://doi.org/10.1136/bmjopen-2019-028988.
Brownson RC, Proctor EK, Luke DA, Baumann AA, Staub M, Brown MT, et al. Building capacity for dissemination and implementation research: one university’s experience. Implement Sci. 2017;12(1):104. https://doi.org/10.1186/s13012-017-0634-4.
Glasgow RE, Vogt TM, Boles SM. Evaluating the public health impact of health promotion interventions: the RE-AIM framework. Am J Public Health. 1999;89(9):1322–7. https://doi.org/10.2105/AJPH.89.9.1322.
Aarons GA, Ehrhart MG, Farahnak LR, Hurlburt MS. Leadership and organizational change for implementation (LOCI): a randomized mixed method pilot study of a leadership and organization development intervention for evidence-based practice implementation. Implement Sci. 2015;10(1):11. https://doi.org/10.1186/s13012-014-0192-y.
Smith JL, R.M., Miller CJ, Chinman MJ, Landes S, Kelly PA, Kirchner JE. Getting to fidelity: identifying core components of implementation facilitation strategies. In Innovation to Impact: Research to Advance VA’s Learning Healthcare Community (VA HSR&D / QUERI National Conference). 2019 (October). Washington, DC.
Iverson KM, Dichter ME, Stolzmann K, Adjognon OL, Lew RA, Bruce LE, et al. Assessing the Veterans Health Administration’s response to intimate partner violence among women: protocol for a randomized hybrid type 2 implementation-effectiveness trial. Implement Sci. 2020;15:1–10.
Jolles MP, Lengnick-Hall R, Mittman BS. Core functions and forms of complex health interventions: a patient-centered medical home illustration. J Gen Intern Med. 2019;34(6):1032–8. https://doi.org/10.1007/s11606-018-4818-7.
Durlak JA, DuPre EP. Implementation matters: a review of research on the influence of implementation on program outcomes and the factors affecting implementation. Am J Community Psychol. 2008;41(3-4):327–50. https://doi.org/10.1007/s10464-008-9165-0.
Kilbourne AM, Neumann MS, Pincus HA, Bauer MS, Stall R. Implementing evidence-based interventions in health care: application of the replicating effective programs framework. Implement Sci. 2007;2(1):42. https://doi.org/10.1186/1748-5908-2-42.
Salloum RG, Shenkman EA, Louviere JJ, Chambers DA. Application of discrete choice experiments to enhance stakeholder engagement as a strategy for advancing implementation: a systematic review. Implement Sci. 2017;12(1):140. https://doi.org/10.1186/s13012-017-0675-8.
Rogers L, De Brún A, Birken SA, Davies C, McAuliffe. The micropolitics of implementation; a qualitative study exploring the impact of power, authority, and influence when implementing change in healthcare teams. BMC Health Serv Res. 2020;20(1):1–13.
Barnett ML, Davis EM, Callejas LM, White JV, Acevedo-Polakovich ID, Niec LN, et al. The development and evaluation of a natural helpers' training program to increase the engagement of urban, Latina/o families in parent-child interaction therapy. Children Youth Serv Review. 2016;65:17–25. https://doi.org/10.1016/j.childyouth.2016.03.016.
Smith S, Almirall D, Bauer M, Liebrecht C, Kilbourne A. (When) Is more better? comparative effectiveness of external vs external+ internal facilitation on site-level uptake of a collaborative care model in community-based practices that are slow to adopt. Health Serv Res. 2020;55(S1):61–2. https://doi.org/10.1111/1475-6773.13413.
The authors would like to thank the implementation scientists and practitioners who provided feedback on the FRAME-IS through think-aloud exercises. We also thank members of the Washington University Network for Dissemination and Implementation Researchers (WUNDIR) and Implementation Research Group (IRG) for providing feedback on the FRAME-IS in a group call format.
Funding sources include NIMH R01 MH 109506 (Stirman), QUERI QUE 20-026 (Miller), K01MH110608 (Barnett), UL1 TR002345 (Baumann), P50 CA-19-006 (Baumann), 3U01HL13399403S1 (Baumann), and 5U24HL136790-02 (Baumann). We would also like to acknowledge the Implementation Research Institute (IRI; NIMH 5R25MH08091607), of which Drs. Barnett, Gutner, Miller, and Wiltsey-Stirman are past fellows, and Dr. Baumann is current faculty.
Authors and Affiliations
VA Boston Healthcare System, Center for Healthcare Organization and Implementation Research (CHOIR), Boston, MA, USA
Christopher J. Miller
Department of Psychiatry, Harvard Medical School, Boston, MA, USA
Christopher J. Miller
Department of Counseling, Clinical, & School Psychology, University of California, Santa Barbara, Santa Barbara, CA, USA
Miya L. Barnett
Washington University at St. Louis, St. Louis, MO, USA
Ana A. Baumann
ViiV Healthcare, Innovation & Implementation Science, Research Triangle, NC, USA
Cassidy A. Gutner
Department of Psychiatry, Boston University School of Medicine, Boston, MA, USA
Cassidy A. Gutner
National Center for PTSD Dissemination and Training Division, Palo Alto, CA, USA
Department of Psychiatry and Behavioral Sciences, Stanford University, Stanford, CA, USA
CJM developed the initial version of the FRAME-S, and wrote the initial version of the manuscript. MLB applied the FRAME-IS to the example project included in the text. All authors (CJM, MLB, AAB, CAG, and SWS) contributed in the following ways: (a) provided substantive refinement of the FRAME-IS, (b) got feedback on initial versions of the FRAME-IS from other stakeholders, and (c) read, edited, and approved the final manuscript.
This .ppt file contains supplemental descriptive text on the core and optional modules of the FRAME-IS.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
Miller, C., Barnett, M.L., Baumann, A.A. et al. The FRAME-IS: a framework for documenting modifications to implementation strategies in healthcare.
Implementation Sci16, 36 (2021). https://doi.org/10.1186/s13012-021-01105-3