- Open Access
- Open Peer Review
The FRAME: an expanded framework for reporting adaptations and modifications to evidence-based interventions
Implementation Science volume 14, Article number: 58 (2019)
This paper describes the process and results of a refinement of a framework to characterize modifications to interventions. The original version did not fully capture several aspects of modification and adaptation that may be important to document and report. Additionally, the earlier framework did not include a way to differentiate cultural adaptation from adaptations made for other reasons. Reporting additional elements will allow for a more precise understanding of modifications, the process of modifying or adapting, and the relationship between different forms of modification and subsequent health and implementation outcomes.
We employed a multifaceted approach to develop the updated FRAME involving coding documents identified through a literature review, rapid coding of qualitative interviews, and a refinement process informed by multiple stakeholders. The updated FRAME expands upon Stirman et al.’s original framework by adding components of modification to report: (1) when and how in the implementation process the modification was made, (2) whether the modification was planned/proactive (i.e., an adaptation) or unplanned/reactive, (3) who determined that the modification should be made, (4) what is modified, (5) at what level of delivery the modification is made, (6) type or nature of context or content-level modifications, (7) the extent to which the modification is fidelity-consistent, and (8) the reasons for the modification, including (a) the intent or goal of the modification (e.g., to reduce costs) and (b) contextual factors that influenced the decision. Methods of using the framework to assess modifications are outlined, along with their strengths and weaknesses, and considerations for research to validate these measurement strategies.
The updated FRAME includes consideration of when and how modifications occurred, whether it was planned or unplanned, relationship to fidelity, and reasons and goals for modification. This tool that can be used to support research on the timing, nature, goals and reasons for, and impact of modifications to evidence-based interventions.
Adaptation, a key concept in implementation, has been defined as a process of thoughtful and deliberate alteration to the design or delivery of an intervention, with the goal of improving its fit or effectiveness in a given context [1, 2]. It is a form of modification, which is a broader concept that encompasses any changes made to interventions, whether deliberately and proactively (adaptation), or in reaction to unanticipated challenges that arise in a given session or context [3, 4]. The process, nature, and outcomes of modifications to evidence-based programs/practices (EBPs) have often not been well documented, despite considerable recent interest in the field of implementation science [5, 6]. Consequently, modification has historically not been fully evaluated or understood.
Understanding what, how, and when modifications occur is a vital aspect of implementation science because the process of implementing EBPs is dynamic . Modifications may occur for a variety of purposes and with differing implications. Some may enhance outcomes, particularly if they more closely align the intervention with the needs of the specific population in a particular system or context. In fact, modifications that focus on increasing the fit of the EBPs with the target population can lead to improved engagement, acceptability, and clinical outcomes, particularly when working with minority populations [8,9,10]. However, modifications that remove key elements of an intervention, or fail to align with population needs, may be less effective [7, 11,12,13,14]. Inconsistent reporting has resulted in uncertainty regarding modification’s impact on health and the types of modifications that can maximize implementation success [15, 16].
Without understanding forms of modification that occur, the systematic evaluation of processes and strategies that lead to more and less successful implementation may be hindered [2, 17,18,19]. To facilitate a more nuanced consideration of modifications and to work toward identifying forms of modifications that may enhance specific interventions vs. those that may reduce effectiveness , Stirman and colleagues previously developed a Framework for Modification and Adaptations . This framework characterized different forms of modifications to interventions, and later work differentiated fidelity-consistent from fidelity-inconsistent modification [1, 2, 16]. More recently, other research groups have used the framework to characterize modifications to various healthcare interventions and prevention programs [20,21,22,23]. Some key aspects of the 2013 framework were shown to have high clarity (rater agreement) and acceptable to high coverage (percentage of identified adaptations that could be classified using the taxonomy; ).
Despite its value in distinguishing and categorizing different forms of modifications, the original framework did not capture other considerations that may be important to document. Because it was originally developed largely to identify forms of modification rather than to fully document the process itself, it did not include potential reasons for modifications, which can range from improving individual or contextual fit (e.g., [20, 21, 25, 26]) to addressing systemic constraints. In 2017, Baumann et al. consulted the implementation literature and the literature on social determinants of health and engaged in a consensus process to expand the framework to include possible reasons for adaptation , a process that differed from the original process of framework development. They recommended additional work to refine the resulting framework.
As we planned a process to refine the framework further, we identified additional limitations and opportunities for expansion. For example, while the 2013 and 2017 frameworks [15, 16] laid out distinct forms of modification that could be made to the content or mode of delivery, they did not specify when those modifications were designed to address important differences between the original population and the stakeholders in the current implementation effort (i.e., cultural adaptations). The frameworks also did not specify whether modifications were planned (i.e., adaptations) or unplanned (e.g., [2, 4, 27]), or allow investigators to consider modifications in conjunction with fidelity, a related but distinct implementation outcome . Further, we recognized that other aspects of reporting that may be important for improving understanding of modification and its impact, such as when in the implementation process the modification was made, were also not included [28,29,30,31]. To address the aforementioned issues, our goal was to develop a refined framework that expanded the original framework to facilitate documentation of additional aspects of the implementation process.
Process for refining the framework
We employed a pragmatic, multifaceted approach, detailed in Table 1, to develop a more comprehensive strategy for characterizing adaptation design and process. This approach included multiple sources of data that better aligned with our goals than approaches such as scoping or systematic reviews . We chose this pragmatic approach because several systematic reviews of adaptation have been published in recent years, and we recognized that such reviews might not capture aspects of modification and adaptation that had not been adequately documented.
Review of the literature
We first searched the literature and identified existing frameworks, systematic reviews, and discussions of adaptations of public health and behavioral health interventions that had been published since 2013 [1, 6, 29, 33]. We also drew on systematic reviews of intervention modifications and adaptations, existing adaptation frameworks, and reviews on cultural adaptation [3,4,5, 15, 17, 20, 21, 33, 34]. We employed a snowballing approach to examine a total of 170 individual articles in the literature that described adaptations to interventions. After reviewing all articles from three systematic reviews [1, 17, 29], we implemented a “stop rule” such that if no new aspects of adaptation were identified after examining 10 original sources from each additional systematic review, we ceased reviewing individual articles. Additionally, we examined two widely cited frameworks of implementation that listed potential determinants [30, 35] and a framework of social determinants of health [36, 37] to further refine the reasons for adaptation. After steps 1–8 (Table 1) were complete, we piloted the resulting framework by coding a sample of articles and interview responses to ensure that no additional elements of adaptation were identified, once again implementing the 10-article “stop rule”.
Reviewing qualitative data
To complement our literature review, we conducted a rapid coding process with a set of responses to questions about adaptation. We generated the responses from interviews with 55 mental health providers and administrators that detailed processes and reasons for adaptation of a psychosocial intervention (see Table 1). These interviews were conducted through studies on sustainability and adaptation [2, 38].
Using the information generated through the two data sources, we categorized both reasons for adaptation and aspects of the adaptation process not captured in the previous framework. We then collapsed similar subcategories and organized them by consensus among the three authors.
Finally, to increase the likelihood that the updated framework would document aspects of modification and adaptation that were important to stakeholders, we presented a draft of the framework to several groups of stakeholders and solicited suggestions regarding additions, refinements, and clarifications.
Overview of the resulting framework and reporting recommendations
Our approach yielded several additions and refinements to the original framework, which are indicated in bold in Fig. 1. Just as Proctor and colleagues  advocate for a multifaceted approach to reporting implementation strategies, our framework is intended to facilitate comprehensive documentation of modifications. Our Framework for Reporting Adaptations and Modifications-Enhanced (FRAME) includes the following eight aspects: (1) when and how in the implementation process the modification was made, (2) whether the modification was planned/proactive or unplanned/reactive, (3) who determined that the modification should be made, (4) what is modified, (5) at what level of delivery the modification is made, (6) type or nature of context or content-level modifications, (7) the extent to which the modification is fidelity-consistent, and (8) the reasons for the modification, including (a) the intent or goal of the modification (e.g., improve fit, adapt to a different culture, reduce costs, etc.) and (b) contextual factors that influenced the decision. Below, we describe in further detail aspects of the FRAME, with attention to aspects not included in the original framework.
When and how in the implementation process the modification was made
Timing, not included in the original framework, is important to measure as modification can occur in any phase of the long-term implementation process: pre-implementation activities, an implementation phase, and scale-up and sustainment [40,41,42]. During the planning or pre-implementation phase, there are opportunities to anticipate changes and discover adaptations that need to be made through a pilot phase. Despite these efforts, new challenges, constraints, or potential enhancements may come to light during the implementation phase. During scale-up or scale-out, other changes may be necessary, particularly for contexts where the intervention is likely to reach populations that may differ from the population that received the intervention during implementation . Additionally, during sustainment, changes in the system or population served by that system may also necessitate changes [7, 44, 45].
Whether the modification was planned/proactive or unplanned/reactive
Reporting on when and how decisions to modify interventions are made will allow investigation of whether planned adaptations are different in nature or in outcomes than those that are improvised during implementation. To avoid unplanned or reactive modifications that are inconsistent with an intervention’s goals, research base, or theory, investigators have advocated a process of planned adaptation , ideally as early as possible in the implementation process. However, in practice, many modifications are made less systematically . Reactive modifications have been defined as those that “occur during the course of program implementation, often due to unanticipated obstacles” . These modifications often occur in an impromptu manner, in reaction to constraints or challenges that are encountered , and may or may not be aligned with the elements of the intervention that make it effective [2, 4].
Adaptations are typically made proactively through a planning process that identifies ways to maximize fit and implementation success while minimizing disruption of the intervention . However, due to the reactive nature of other modifications, we added a specifier to the framework to capture whether modifications were planned prior to delivery (i.e., an adaptation), or unplanned and made in response to an unanticipated challenge. It is important to note that iterative changes are not necessarily reactive—iteration can accommodate unanticipated challenges. For example, a modification made during the “Act” portion of a “Plan-do-study-act” cycle would not be considered reactive, because it was determined through a systematic process rather than through improvisation.
Who determined the modification
The driver of change, and how participatory the decision was, may be important predictors of whether the changes have the desired impact. The FRAME can be used to list all who play a role in the decision, but in some contexts, it may be important to specify who makes the ultimate decision, as this detail might affect whether and how widespread the modification may occur. It may also be closely linked to reasons for modification. For example, administrators’ decisions to modify an EBP may be related to restrictions in sourcing or contracting while policymakers’ decisions may be in response to political factors or funding availability. Modifications made by individual providers may respond to recipient-level needs or local constraints that may not be visible to policymakers.
What is modified
The original 2013 framework  focused largely on characterizing what types of changes were made to facilitate understanding of which changes are associated with implementation success and recipient-level outcomes. Much of this component of the framework remains unchanged, although we added implementation and scale-up activities to reflect that these processes may differ across contexts. If little about the intervention were changed, but the implementation strategies differed significantly across otherwise similar contexts, then differences in outcomes may be attributable to differences in how the intervention was implemented.
At what level of delivery the modification is made
Codes for the level of delivery also remain mostly unchanged from the 2013 framework. Reporting the level at which modifications occur has implications for understanding whether and under what circumstances implication success or effectiveness are associated with making individual-level modifications, and when modifications may need to be applied more broadly. The FRAME now differentiates the entire target group (e.g., women who are at risk of developing diabetes) and individual sub-groups (e.g., new mothers who are at risk of developing diabetes), as modifications may be made for broad or specific groups depending on the circumstances.
Type or nature content-level modifications
Although many items in this section are unchanged from the original framework, items were added to reflect a larger variety of modifications identified through the literature, ongoing observational work (c.f., [1, 3, 20, 21, 23, 48]), and stakeholder interviews. For example, it may be important to understand whether drift occurs for a relatively brief period of time before returning to a protocol or whether it occurs for the duration of a meeting or session, without a return to planned content. Additionally, spreading out psychosocial or educational content intended for a single meeting or session over multiple sessions was added, as this may occur when an individual requires more time to understand content or when unforeseen or emergent issues need to be addressed during a given session.
The relationship to fidelity
In previous work, we have made a distinction between fidelity-consistent and fidelity-inconsistent modifications. Fidelity-consistent modifications are defined as those that preserve core elements of a treatment that are needed for the intervention to be effective . In contrast, fidelity-inconsistent modifications are those that alter the intervention in a manner that fails to preserve its core elements. The identification of fidelity-consistent and fidelity-inconsistent modifications can be made in consultation with the existing literature, input from the treatment developer, and any available evaluation data. Others have suggested that the function of intervention elements be prioritized over the form, such that what is core to the intervention is conceptualized as one of several possible activities or materials that accomplish a core intervention function (e.g., education, skill building, connecting to supports or resources) .
When implementing an intervention with a population that differs in important ways from the populations with which the intervention has been tested and implemented, new understanding of which elements are core vs. peripheral may emerge. For example, a component intended to improve basic health literacy may be essential for one population, but less necessary for a population that has a solid foundation of specific health information. Moreover, it may not be known whether some modifications, such as preserving the original spacing of sessions or activities, would have an impact on outcomes, and there may not be theory to inform decisions about such modifications either way. We have therefore added an “unknown” code that can be used when there is no theory or evidence to inform a decision about whether an element is core vs. peripheral.
The rationale for the modifications made
The additions related to the rationale for modifications represent the most significant enhancements to the original framework. Capturing the rationale for a given modification may be crucial in determining links to key implementation or health outcomes . For example, modifications made to cut costs may have a very different impact than those made to improve fit or engagement. Thus, we specify the goal for modification—to improve feasibility or acceptability, to increase reach or engagement, to improve fit (note that cultural modifications intended to improve fit are assigned a subcategory under this goal), to reduce costs, to improve clinical outcomes, or to align the intervention with cultural values, norms, or priorities.
To identify reasons to modify EBPs for FRAME, we referred to existing reviews and original literature that enumerated reasons for specific projects and implementation frameworks that specified potential determinants at different levels. We specifically added cultural modifications within the FRAME due to the importance of cultural modifications in implementation science. Attention to cultural factors is important in investigating modifications to EBPs because of its potential consequences in terms of healthcare disparities . Cultural adaptation can be defined as the systematic modification of an intervention to “consider language, culture, and context in such a way that it is compatible with the client’s cultural patterns, meanings, and values” . Focusing on culture allows us to expand the characterization of content modifications (e.g., whether content is added, removed, or tailored) and context-level modifications (e.g., whether personnel, training, or education are modified) to explicitly address cultural patterns or values at the client, provider, or sociopolitical level. Designation of cultural or religious norms at the organizational level distinguish adaptations made to distinguish this form of culture from the construct of organizational culture used in the implementation and organization literature to signify “the way things are done in an organization” . At times, cultural or religious beliefs may contribute to organizational culture or policies, but the constructs do not fully overlap.
By explicitly defining cultural aspects and determinants that affect inequities in care delivery, we hope to identify the types of modifications made to address cultural aspects of the populations included in our studies. Thus, we also examine, more broadly, factors at the recipient, provider, organizational, and socio-political levels that are examples of important components that affect the modifications of the EBPs.
Socio-political factors may also be important determinants of modifications to EBPs captured in the FRAME. For example, socio-historical factors such as stigma attached to receiving mental health treatment may require modifications so that interventions are instead delivered by spiritual leaders or peers, or in settings that are more comfortable for members of a community. Existing laws and policies can impact whether aspects of an intervention are removed due to constraints (e.g., telephone check-ins may be removed if they cannot be reimbursed; licensure restrictions for telehealth across state lines) and may also result in adaptations to personnel or setting.
At the organizational level, we found that factors that may lead to modifications overlap somewhat with those found in existing determinant frameworks (e.g., [41, 52]). For example, staffing shortages may suggest a variety of context modifications. This may include delivery by different personnel (e.g., providers from a different discipline) or changes to the format or timing of delivery. Such staffing shortages may also affect training or evaluation of the intervention. For example, implementation may require streamlining training sessions, adapting them to fit with providers’ previous training, or spreading them over several weeks to accommodate busy clinic schedules.
Other practical constraints play a role in decisions to adapt or modify interventions [2, 15, 19, 23]. Space shortages may indicate the need for context modifications (e.g., changing from group to individual delivery). In contexts without easily accessed health centers, community-, home-, or telehealth-based delivery formats, there may be contextual adaptations to address demand or need. Limitations to available technology can have far-reaching implications as well as it may require removal or adaptation of aspects of the original intervention that can be delivered in under-resourced settings. Time constraints may lead to removing elements or compressing the intervention.
Aspects of organizational/setting (including local community if not delivered in a healthcare setting) culture may also impact how interventions are delivered. Competing demands, de-prioritization of an intervention, or high rates of turnover may lead to changes in who delivers the intervention, how many sessions or elements are provided, and whether and how training is provided. Regulatory or compliance issues or legal concerns may lead to certain aspects of an intervention not being delivered (e.g., limits to the types of physical activity or activities that may occur off-premises). An organizational culture that has long espoused a different theory or intervention may leave providers wary of new practices and lead to integration of elements of interventions into existing practices rather than de-implementation of preferred practices . In sum, organizational/setting factors—including organizational culture and available resources—may necessitate a variety of modifications even to interventions with a strong evidence base in specific contexts.
At the provider level, there has been discussion around the positive or negative consequences of modifying interventions to fit with provider preferences or to improve the interaction between providers and their clients . Providers of psychosocial interventions frequently modify interventions for a variety of reasons, including perceived client preferences, providers’ preferences or self-efficacy , and efforts to maintain a good therapeutic alliance [55,56,57,58]. Factors such as provider gender and cultural beliefs may also impact decisions about delivery of the intervention. Additionally, some provider factors, such as previous training and experience, may lead to changes to training and evaluation.
Recipient level factors are also identified in Fig. 1. Each identified factor may contribute to a need for modification to promote optimal levels of engagement and outcomes at different levels. For example, limited transportation might make face to face meetings in a clinic difficult, which could lead to modifications in the format (e.g., telephone or internet-based) or setting (home-based). On the other hand, low literacy might make full engagement in an intervention that relies on written materials difficult and require tailoring to deliver content to clients in other formats. Other aspects, such as cultural norms, legal status, or physical capacity, may necessitate the removal or alteration of some elements of the EBP.
Although our approach to refining the framework is unique in its use of multiple data sources and stakeholder input, some limitations are important to describe. We did not employ a systematic review or a traditional thematic analysis because these approaches did not fully align with the current project goals. While it is possible that additional items would have been identified through these processes, our use of a “stop rule”, stakeholder feedback, and coding of subsequent articles using the FRAME after consolidating information from all of our data sources suggested that the framework was sufficiently comprehensive.
Furthermore, stakeholders identified a need to balance comprehensiveness with feasibility and pragmatism in documentation and reporting. The feasibility of using a comprehensive framework is likely to differ across research and applied settings. We must also acknowledge that comprehensively cataloging modifications to EBPs may be difficult in some contexts even with a well-developed framework like the FRAME. For example, if an intervention has not been exhaustively described and tested—as is frequently the case in healthcare—then it may be impossible to reliably detect adaptations and their associated consequences. In these cases, the FRAME will only be as useful as the data informing its application. Evaluation of different approaches to using the FRAME for documentation is needed to inform efforts to achieve the appropriate balance.
Recommendations and future directions in reporting
Several strategies for reporting adaptations and modifications that may be applicable to the FRAME have been developed and described in the literature. In this section, we discuss the advantages and drawbacks to these strategies.
To identify modifications that are made during routine treatment delivery, Stirman and colleagues developed the Modification and Adaptation Checklist (MAC; ), an observational coding system intended to be used in conjunction with fidelity assessments. Observation, the “gold standard” for fidelity coding, may be useful when providers may not realize they are making content modifications or when they have difficulty recalling, identifying, or describing which modifications they made. It may be particularly useful in contexts in which providers may be reluctant to report modifications such as drift or removing key intervention elements. However, observation is not feasible in many contexts as it is labor intensive and would require familiarity with the FRAME and the intervention. Intermittent observation, although more feasible, might lead investigators to miss certain forms of modification, such as extending a protocol or repeating material in a session or encounter that is not observed . Furthermore, without additional information from stakeholders, the rationale for making a modification cannot be confirmed.
Provider or key informant self-report
A self-report version of the MAC includes both content-level modifications and brief questions about format, level of delivery, and reasons for modification. The reasons for modification are less detailed than those included in the FRAME. However, space for details about the specific contextual factors that were considered or that led to modification can allow for free responses that can be coded. Self-report may be more feasible when frequent assessment is required, although it may entail greater burden on providers than participating in a one-time interview or having encounters observed. Providers may over-report some forms of modification (c.f., ), while underreporting others. In fact, at times, providers may not recognize that they are modifying the interventions or whether adaptations are fidelity-consistent or fidelity-inconsistent. Core elements and functions of the intervention may not be fully established, making it challenging to report these aspects. These challenges may be heightened when modifications are not tracked in real-time, and reporting may be subject to recall bias, or when training has not been sufficient to promote awareness of fidelity. Additionally, incentives and contingencies may be present that impact reporting of modification and fidelity to an intervention.
Despite these limitations, self-report measures may be the most feasible strategy for real-time reporting of modifications that occur during implementation and sustainment phases. However, many of the context modifications that are reported can be validated through other forms of documentation. It remains to be determined whether self-report and observer ratings agree, and the optimal frequency for self-report. Additionally, it is unclear whether self-reports are more informative and accurate as global self-assessments (e.g., inquiring which interventions a provider has made over a given period of time) or for a single encounter or time point (e.g., focusing assessments on which intervention a provider used for a single encounter). Recent and ongoing research are attempting to address these types of questions [22, 38].
Interviews may provide richer data than provider checklists or observation of single encounters, although they may be subject to similar biases as self-report. They allow an understanding of who made the ultimate decision to modify, the level of delivery, reasons for adaptation, and contextual factors that were considered. We developed a codebook for our original framework  to allow investigators to operationalize and identify modifications that were made to interventions during the implementation process. It was originally used for coding articles in the literature  and interviews with community-based clinicians  and has since been used in a variety of contexts [20,21,22,23]. Interview guides based on an expanded codebook can be used to facilitate understanding of the eight FRAME elements. For example, Rabin and colleagues recently described a measurement system that expanded the Stirman et al. 2013 framework to include RE-AIM concepts, framed as Who, How, When, What, and Why? . A potential drawback to interviews is that they may not be feasible to administer frequently due to the time required for interviews and coding.
Differing forms of assessment will need to be compared to assess accuracy and reliability, and factors such as burden on stakeholders and research participants will need to be considered in determining the best assessment strategy for a given project. For some interventions, modifications may be most reliably identified through self-report checklists (with sufficient descriptions of each adaptation to facilitate reliable reporting) that are completed soon after the intervention is delivered, while others may be best identified through detailed interviews with stakeholders. Triangulation of strategies may be necessary when modifications are not easily observed and to better assess reasons for modification. For example, Rabin et al.  used observational data in addition to interviews to construct intervention process maps and identify additional contextual factors that may be relevant to adaptation.
Future directions and research agenda
Measurement and reporting
While attention to modification has greatly increased in the past decade, the science of measurement and reporting remains nascent. Strategies for reporting and measurement have not yet been empirically compared, nor have psychometric properties of self-reports been examined. When used for research, detailed versions of a FRAME measure can facilitate comprehensive reporting and analyses. For the most precise coding, the FRAME figure itself could be used for each separate modification that was identified, with the reporter or interviewer circling the appropriate selection in each of the sections. However, elements that do not apply in a given context could be removed to streamline reporting and evaluation in routine care contexts. As with sustainability , it is unlikely that assessment of modification at a single time point will reflect the dynamic process of implementation . Assessment at multiple time points will provide a richer understanding of why, how, and with what impact modification occurs in complex systems.
Linking and understanding modifications and outcomes
Ultimately, the FRAME is intended to facilitate understanding of associations between the process, types, and reasons that interventions are modified and key outcomes. Over time, such research may identify which aspects of the FRAME are particularly important to attend to when planning and documenting adaptations. Key outcomes to consider include increased, decreased, or unchanged levels of reach, diagnostic outcomes, engagement, or acceptability. However, in designing analyses to examine whether modifications may lead to differences in outcomes, it is important not to confound the impact of potential moderating factors that inspired the modification (e.g., comorbidity) with the impact of the modification itself.
We developed and reviewed some study methodologies, summarized elsewhere [1, 15], that can facilitate prospective research on modifications. We also reviewed experimental, prospective investigations of adaptations , but many of the adaptations and combinations thereof that occur in settings in which interventions routinely occur have not been represented. Methods to tease apart the impact of individual modifications when they frequently occur in conjunction with multiple others remain to be developed and may require large samples. Chambers and Norton  suggest the development of a database in which data from multiple projects can be pooled, using a common taxonomy, to facilitate more rapid understanding of what adaptations are necessary or effective for similar interventions when delivered to different populations or in different contexts. Within efforts to implement or scale-up across multiple sites, strategies such as qualitative comparative analysis may identify combinations of contextual factors and adaptations that associated with outcomes.
Much work remains to be done to develop generalized knowledge about the process, nature, and outcomes of modifications made to different types of interventions in vastly different contexts. The FRAME is intended to capture information that reflects the complex and dynamic settings in which implementation occurs. Documenting with the FRAME can facilitate more rigorous study that includes efforts not only to characterize adaptations themselves, but also to clarify the timing, context, and process of modifying interventions to facilitate their implementation, scale-up, spread, and sustainment.
Framework for Reporting Adaptations and Modifications-Enhanced
Modification and Adaptation Checklist
National Institute of Health
Qualitative Comparative Analysis
Reach, Effectiveness, Adoption, Implementation and Maintenance
Stirman SW, Gamarra JM, Bartlett BA, Calloway A, Gutner CA. Empirical examinations of modifications and adaptations to evidence-based psychotherapies: methodologies, impact, and future directions. Clin Psychol Sci Pract. 2017;24(4):396–420.
Stirman SW, Gutner C, Edmunds J, Evans AC, Beidas R. Relationships between clinician-level attributes and fidelity-consistent and fidelity-inconsistent modifications to an evidence-based psychotherapy. Implementation Sci. 2015;10(1):115.
Barrera M, Berkel C, Castro FG. Directions for the advancement of culturally adapted preventive interventions: local adaptations, engagement, and sustainability. Prev Sci. 2017;18(6):640–8.
Cooper BR, Shrestha G, Hyman L, Hill L. Adaptations in a community-based family intervention: replication of two coding schemes. J Primary Prevent. 2016;37(1):33–52.
von Thiele Schwarz U, Förberg U, Sundell K, Hasson H. Colliding ideals–an interview study of how intervention researchers address adherence and adaptations in replication studies. BMC Med Res Methodol. 2018;18(1):36.
Sundell K, Beelmann A, Hasson H, von Thiele Schwarz U. Novel programs, international adoptions, or contextual adaptations? Meta-analytical results from German and Swedish intervention research. J Clin Child Adolesc Psychol. 2016;45(6):784–96.
Chambers DA, Glasgow RE, Stange KC. The dynamic sustainability framework: addressing the paradox of sustainment amid ongoing change. Implementation Sci. 2013;8(1):117.
Steinka-Fry KT, Tanner-Smith EE, Dakof GA, Henderson C. Culturally sensitive substance use treatment for racial/ethnic minority youth: a meta-analytic review. J Subst Abus Treat. 2017;75:22–37.
Rathod S, Gega L, Degnan A, Pikard J, Khan T, Husain N, et al. The current status of culturally adapted mental health interventions: a practice-focused review of meta-analyses. Neuropsychiatr Dis Treat. 2018;14:165–78.
Bernal GE, Domenech Rodríguez MM. Cultural adaptations: tools for evidence-based practice with diverse populations: American Psychological Association; 2012.
Pérez D, Van der Stuyft P, del Carmen Zabala M, Castro M, Lefèvre P. A modified theoretical framework to assess implementation fidelity of adaptive public health interventions. Implementation Sci. 2015;11(1):91.
Castro FG, Barrera M Jr, Martinez CR Jr. The cultural adaptation of prevention interventions: resolving tensions between fidelity and fit. Prev Sci. 2004;5(1):41–5.
Kumpfer KL, Scheier LM, Brown J. Strategies to avoid replication failure with evidence-based prevention interventions: case examples from the strengthening families program. Eval Health Prof. 2018:1–34.
Barrera M, Castro FG. A heuristic framework for the cultural adaptation of interventions. Clin Psychol: Sci and Pract. 2006;13(4):311–6.
Baumann A, Cabassa LJ, Stirman SW. Adaptation in dissemination and implementation science. In: Brownson RC, Colditz GA, Proctor EK, editors. Dissemination and implementation research in health: translating science to practice, vol. 2. Oxford: Oxford University Press; 2017. p. 286–300.
Stirman SW, Miller CJ, Toder K, Calloway A. Development of a framework and coding system for modifications and adaptations of evidence-based interventions. Implementation Sci. 2013;8:65.
Baumann AA, Powell BJ, Kohl PL, Tabak RG, Penalba V, Proctor EK, et al. Cultural adaptation and implementation of evidence-based parent-training: a systematic review and critique of guiding evidence. Child Youth Serv Rev. 2015;53:113–20.
Glasziou P, Altman DG, Bossuyt P, Boutron I, Clarke M, Julious S, et al. Reducing waste from incomplete or unusable reports of biomedical research. Lancet. 2014;383(9913):267–76.
Baumann AA, Domenech Rodríguez MM, Amador NG, Forgatch MS, Parra-Cardona JR. Parent management training-Oregon model (PMTO™) in Mexico City: integrating cultural adaptation activities in an implementation model. Clin Psychol: Sci and Pract. 2014;21(1):32–47.
Rabin BA, McCreight M, Battaglia C, Ayele R, Burke RE, Hess PL, et al. Systematic, multimethod assessment of adaptations across four diverse health systems interventions. Front Health Policy Res. 2018;6:102.
Escoffery C, Lebow-Skelley E, Haardoerfer R, Boing E, Udelson H, Wood R, et al. A systematic review of adaptations of evidence-based public health interventions globally. Implement Sci. 2018;13(1):125.
Lau A, Barnett M, Stadnick N, Saifan D, Regan J, Wiltsey Stirman S, et al. Therapist report of adaptations to delivery of evidence-based practices within a system-driven reform of publicly funded children’s mental health services. J Consult Clin Psychol. 2017;85(7):664.
Cook JM, Dinnen S, Thompson R, Simiola V, Schnurr PP. Changes in implementation of two evidence-based psychotherapies for PTSD in VA residential treatment programs: a national investigation. J Trauma Stress. 2014;27(2):137–43.
Roscoe JN, Shapiro VB, Whitaker K, Kim BE. Classifying changes to preventive interventions: applying adaptation taxonomies. J Prim Prev. 2019:1–21.
Parra-Cardona R, López-Zerón G, Leija SG, Maas MK, Villa M, Zamudio E, et al. A culturally adapted intervention for Mexican-origin parents of adolescents: the need to overtly address culture and discrimination in evidence-based practice. Fam Process. 2018.
Rodríguez MMD, Baumann AA, Schwartz AL. Cultural adaptation of an evidence based intervention: from theory to practice in a Latino/a community context. Am J Community Psychol. 2011;47(1–2):170–86.
Ivanich JD, Mousseau AC, Walls M, Whitbeck L, Whitesell NR. Pathways of adaptation: two case studies with one evidence-based substance use prevention program tailored for indigenous youth. Prev Sci. 2018:1–11.
Rabin B, McCreight M, Glasgow R, editors. Systematic multi-method documentation of adaptations in five health system interventions. Implementation Sci 2018;13.
Escoffery C, Lebow-Skelley E, Udelson H, Böing EA, Wood R, Fernandez ME, et al. A scoping study of frameworks for adapting public health evidence-based interventions. Transl Behav Med. 2018;9(1):1–10.
Brown CH, Curran G, Palinkas LA, Aarons GA, Wells KB, Jones L, et al. An overview of research and evaluation designs for dissemination and implementation. Annu Rev Public Health. 2017;38:1–22.
Ramsay JE, Janevic MR, Hogan CK, Edwards DL, Connell CM. Using the replicating effective programs framework to adapt a heart health intervention. Health Promot Pract. 2018.
Kennedy-Martin T, Bae JP, Paczkowski R, Freeman E. Health-related quality of life burden of nonalcoholic steatohepatitis: a robust pragmatic literature review. Journal of Patient-Reported Outcomes. 2018;2(1):28–42.
Chu J, Leino A. Advancement in the maturing science of cultural adaptations of evidence-based interventions. J Consult Clin Psychol. 2017;85(1):45.
Miller-Day M, Pettigrew J, Hecht ML, Shin Y, Graham J, Krieger J. How prevention curricula are taught under real-world conditions: types of and reasons for teacher curriculum adaptations. Health Educ. 2013;113(4):324–44.
Bernal G, Bonilla J, Bellido C. Ecological validity and cultural sensitivity for outcome research: issues for the cultural adaptation and development of psychosocial treatments with Hispanics. J Abnorm Child Psychol. 1995;23(1):67–82.
World Health Organization. Social determinants of mental health: World Health Organization; 2014.
Braveman P, Gottlieb L. The social determinants of health: it's time to consider the causes of the causes. Public Health Rep. 2014;129(1_suppl2):19–31.
Stirman SW, Finley E, Shields N, Cook J, Suvak M, Dimeff L, Koerner K, Haine-Schlagel R, Gutner C, Burgess J, Gagnon D, Beristianos M, Mallard K, Ramirez V, Monson, C. M. Improving and sustaining delivery of CPT for PTSD in mental health systems: a cluster randomized trial. Implement Sci. 2017;12:3260.
Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implementation Sci. 2013;8(1):139.
Kilbourne AM, Neumann MS, Pincus H, Bauer MS, Stall R. Implementing evidence-based interventions in health care: application of the replicating effective programs framework. Implementation Sci. 2007;2(1):42.
Aarons G, Hurlburt M, Horwitz SM. Advancing a conceptual model of evidence-based practice implementation in public service sectors. Admin Pol Ment Health. 2011;38(1):4–23.
Rogers EM. Diffusion of innovations: New York The Free Press, A Division of Simon & Schuster, Inc; 2010.
Aarons GA, Sklar M, Mustanski B, Benbow N, Brown CH. “Scaling-out” evidence-based interventions to new populations or new health care delivery systems. Implementation Sci. 2017;12(1):111.
Stirman SW, Kimberly J, Cook N, Calloway A, Castro F, Charns M. The sustainability of new programs and innovations: a review of the empirical literature and recommendations for future research. Implement Sci. 2012;7(1):17.
Gruen RL, Elliott JH, Nolan ML, Lawton PD, Parkhill A, McLaren CJ, et al. Sustainability science: an integrated approach for health-programme planning. Lancet. 2008;372(9649):1579–89.
Lee SJ, Altschul I, Mowbray CT. Using planned adaptation to implement evidence-based programs with new populations. Am J Community Psychol. 2008;41(3–4):290–303.
Moore J, Bumbarger B, Cooper B. Examining adaptations of evidence-based programs in natural contexts. J Primary Prevent. 2013;34(3):147–61.
Park AL, Chorpita BF, Regan J, Weisz JR. Research network on youth mental health. Integrity of evidence-based practice: are providers modifying practice content or practice sequencing? Admin Pol Ment Health. 2015;42(2):186–96.
Shelton RC, Cooper BR, Stirman SW. The sustainability of evidence-based interventions and practices in public health and health care. Annu Rev Public Health. 2018;39:55–76.
Mittman B. Evaluating complex interventions: confronting and guiding (versus ignoring and suppressing) heterogeneity and adaptation. In: Brown H, Smith JD, editors. Prevention Science Methods Group Webinar; 2018.
Glisson C. Assessing and changing organizational culture and climate for effective services. Res Soc Work Pract. 2007;17(6):736–47.
Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implementation Sci. 2009;4(1):50.
Helfrich CD, Rose AJ, Hartmann CW, van Bodegom-Vos L, Graham ID, Wood SJ, et al. How the dual process model of human cognition can inform efforts to de-implement ineffective and harmful clinical practices: a preliminary model of unlearning and substitution. J Eval Clin Pract. 2018;24(1):198–205.
Meza RD, Jungbluth N, Sedlar G, Martin P, Berliner L, Wiltsey-Stirman S, Dorsey S. Clinician-Reported Modification to a CBT Approach in Children’s Mental Health. J Emotion Behav Disord. 2019;1-10.
Aarons GA, Miller EA, Green AE, Perrott JA, Bradway R. Adaptation happens: a qualitative case study of implementation of the incredible years evidence-based parent training program in a residential substance abuse treatment program. Journal of Children's Services. 2012;7(4):233–45.
Stirman SW, Calloway A, Toder K, Miller CJ, DeVito AK, Meisel SN, et al. Community mental health provider modifications to cognitive therapy: implications for sustainability. Psychiatr Serv. 2013;64(10):1056–9.
Cook JM, Simiola V, McCarthy E, Ellis A, Stirman SW. Use of reflective journaling to understand decision making regarding two evidence-based psychotherapies for PTSD: practice implications. Pract Innov. 2018;3(3):153–67.
Naeem F, Habib N, Gul M, Khalid M, Saeed S, Farooq S, et al. A qualitative study to explore patients’, carers’ and health professionals’ views to culturally adapt CBT for psychosis (CBTp) in Pakistan. Behav Cogn Psychother. 2016;44(1):43–55.
Marques L, Valentine SE, Kaysen D, Mackintosh MA, De Silva D, Louise E, Ahles EM, Youn SJ, Shtasel DL, Simon NM, Wiltsey Stirman S. Provider fidelity and modifications to cognitive processing therapy in a diverse community health clinic: Associations with clinical change. J Consult Clin Psychol. 2019;87(4):357.
Chambers DA, Norton WE. The adaptome: advancing the science of intervention adaptation. J of Prev Med. 2016;51(4):S124–S31.
The authors wish to express their gratitude to Clara Johnson for her assistance with manuscript preparation.
The development of new elements of the FRAME was informed in part by ongoing data collection in a study funded by the National Institute of Mental Health (R01 MH 106506 PIs: Stirman and Monson). AB is also funded by U01HL133994-03S; UL1TR00234; R01HG009351; U24HL136790-02. CM’s work on this project was also supported by VA HSR&D QUERI QUE 15-289. The authors gratefully acknowledge the Implementation Research Institute (IRI), at the George Warren Brown School of Social Work, Washington University in St. Louis, through an award from the National Institute of Mental Health (5R25MH08091607).
Availability of data and materials
Ethics approval and consent to participate
The study that generated the interview data was IRB approved and individuals who were interviewed provided informed consent.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.