This article has Open Peer Review reports available.
Implementing evidence-based interventions in health care: application of the replicating effective programs framework
- Amy M Kilbourne†1, 2Email author,
- Mary S Neumann3,
- Harold A Pincus4,
- Mark S Bauer5 and
- Ronald Stall6
© Kilbourne et al; licensee BioMed Central Ltd. 2007
Received: 18 July 2006
Accepted: 09 December 2007
Published: 09 December 2007
We describe the use of a conceptual framework and implementation protocol to prepare effective health services interventions for implementation in community-based (i.e., non-academic-affiliated) settings.
The framework is based on the experiences of the U.S. Centers for Disease Control and Prevention (CDC) Replicating Effective Programs (REP) project, which has been at the forefront of developing systematic and effective strategies to prepare HIV interventions for dissemination. This article describes the REP framework, and how it can be applied to implement clinical and health services interventions in community-based organizations.
REP consists of four phases: pre-conditions (e.g., identifying need, target population, and suitable intervention), pre-implementation (e.g., intervention packaging and community input), implementation (e.g., package dissemination, training, technical assistance, and evaluation), and maintenance and evolution (e.g., preparing the intervention for sustainability). Key components of REP, including intervention packaging, training, technical assistance, and fidelity assessment are crucial to the implementation of effective interventions in health care.
REP is a well-suited framework for implementing health care interventions, as it specifies steps needed to maximize fidelity while allowing opportunities for flexibility (i.e., local customizing) to maximize transferability. Strategies that foster the sustainability of REP as a tool to implement effective health care interventions need to be developed and tested.
Closing the gap between research and practice has been a priority for many agencies, including the U.S. National Institutes of Health, Veterans Health Administration (VHA) and the Agency for Healthcare Research and Quality . Despite the development of effective interventions to improve health care quality, most of these interventions have only been implemented in the academic settings in which they were developed, and few have been successfully disseminated into non-academic-affiliated (i.e., community-based) organizations . A primary reason for this research-to-practice gap is the lack of a framework for implementing effective interventions in community-based organizations that maintains fidelity while maximizing transferability when the interventions are adopted across different settings [2–4].
An effective strategy for implementing clinical and health services (i.e., health care) interventions is needed. Many interventions fail to achieve the outcomes observed when tested in research settings once they are implemented in community-based organizations. This "voltage drop" in effectiveness maybe due to reduced fidelity to the intervention when disseminated outside the academic realm to community-based organizations, as well as lack of guidance in customizing interventions to community-based populations [5–8]. Moreover, stakeholders (e.g., health care purchasers, plans, and providers) increasingly have to choose which interventions to implement for their populations, and need guidance in implementing them in a cost-efficient manner without diminishing the intervention's effectiveness. At the same time, these stakeholders need to ensure that the intervention will be accepted and adopted across different organizations.
A number of strategies for guiding implementation efforts have been proposed [9–14], many of which primarily focus on adapting interventions by (a) determining when an organization is "ready" to adapt or adopt an intervention, and (b) working with senior leaders and frontline providers to overcome barriers to adaptation. For example, the VA's Quality Enhancement Research Initiative (QUERI) framework outlines a process for choosing, implementing, and marketing evidence-based interventions in health care organizations . Simpson describes the underlying organizational characteristics that need to be considered (readiness to change, resources, and culture) when planning an implementation of evidence-based interventions . Bartholomew et al. describe Intervention Mapping, which involves planning an implementation of a program based on information from the target population .
However, these frameworks do not address a fundamental issue in implementation: achieving a balance between adequate fidelity to the intervention and accommodating differences across organizations in order to maximize the effectiveness of the intervention. To date, no implementation frameworks have outlined strategies for maintaining treatment fidelity while providing opportunities to adapt interventions to fit local needs. For example, previous frameworks do not provide specific guidance in replicating interventions for use in community-based organizations, and do not specify training or technical assistance programs for facilitating implementation across different settings. Without the appropriate tools and materials available in a user-friendly format, community-based organizations are less likely to implement an evidence-based intervention; or if implemented, treatment fidelity may be suboptimal.
In this article, we describe an innovative implementation framework: Replicating Effective Programs (REP). In contrast to previously published implementation frameworks, REP provides a roadmap for implementing evidence-based interventions into community-based settings through a combination of intervention "packaging," training, technical assistance, and other strategies to maximize the chances for sustaining the interventions. REP has been empirically evaluated through a randomized controlled trial of its effectiveness in achieving intervention uptake and fidelity across different organizations [15, 16]. To date, other implementation frameworks have included some but not all of the components of REP [9–14], and there have been no evaluations of implementation frameworks that combine strategies to maximize both fidelity and flexibility in implementing interventions.
History and underlying theories of REP
REP was developed and applied by the U.S. Centers for Disease Control and Prevention (CDC) to package and disseminate HIV behavioral and treatment interventions for implementation in community-based service settings, notably AIDS service organizations . CDC's REP project was initiated in 1996 to address a critical link in bridging research and practice; namely, packaging interventions so that they can be easily implemented in non-academic, community-based settings. Extensive research has focused on developing and evaluating interventions designed to change risk behaviors related to HIV transmission since shortly after the first cases of AIDS were reported in the United States . The CDC's approach involved: applying rigorous criteria to identify effective interventions ; packaging interventions' complete scientific protocols into non-technical language ; supporting the implementation of the interventions through training and technical assistance ; and providing ongoing financial and technical support to sustain the interventions' implementation [22–24].
Outline of REP process for health services-based interventions
--Identify high-burden condition
--Identify barriers to implementation
Identify effective intervention
--Identify intervention tested in a completed, randomized controlled study
--Organizational needs assessment, usual care
Researchers, representatives from practices, providers of community-based organizations (target population)
--Write package into everyday language
--Distinguish core elements, menu options
Community Working Group
--Select Community Working Group (CWG)
Researchers, CWG (i.e., health plans, practices, providers, patients, purchasers)
--Refine package, Core elements and menu options refined based on CWG input, adjudicated by intervention developers
--Refine training, technical assistance approach strategies per CWG input
Pilot test package
--Further refinement of package
--Identify eligible organizations
Researchers, CWG, organizations participating in implementation
--Logistics of dissemination
--Kick-off meeting, package dissemination
--Organization staff training
Researchers and staff
--Follow-up with organizations
Technical assistance expert
--Return on investment
--Continue CWG, site visits
Feedback and refinement
--Analyze data, inform sustainability
Maintenance and evolution
Organizational, financial changes
--CWG advises on sustainability strategies
Researchers, CWG, sites
--Develop business case for intervention and REP process
--Refine business case: return on investment
Re-customize delivery as need arises
--Continue to refine package (e.g., menu options)
Pre-conditions for intervention adoption are: the identification of the need for a new intervention for a target population; identification of an effective intervention that fits local settings (e.g., mission of organization and benefits to the organization) and the intended target population (e.g., behavioral risks and culture); identifying implementation barriers; and drafting a user-friendly manual (i.e., "package") of the intervention.
The first step of the pre-conditions phase is to identify appropriate at-risk populations and suitable evidence-based interventions.
Identifying effective interventions
Interventions are suitable for REP if they have been rigorously evaluated (e.g., via randomized controlled trials or other quantitative studies) and have produced statistically significant positive effects on the health outcomes of interest.
Ensuring intervention fits local settings
In reviewing effective interventions, researchers should be cognizant of whether the intervention has been shown to be effective in patient populations similar to those in the organizations taking part in the implementation process. Many community-based organizations serve a disproportionate number of low income, minority, and co-morbidly ill patients when compared to patients from academic settings. Another consideration is determining what the "usual care" conditions were in the effectiveness studies. Often academic practices have resources that facilitate the conduct of research studies (e.g., staffing, electronic medical records, space), which many community-based organizations lack.
Identifying implementation barriers
To ensure that the intervention is feasible in local settings, researchers should meet with the staff members from participating organizations, introduce the intervention, and conduct an assessment of potential barriers to its implementation. Such meetings also foster buy-in, especially if the researchers develop a list of benefits of participation, from training opportunities to potential cost savings (i.e., the "business case"). During these meetings, researchers should gather information in the form of a needs assessment to benchmark usual care, including resources and functioning characteristics that may directly affect the mode by which the intervention will be implemented and the types of technical assistance that might be needed (e.g., staffing, patient volume, management characteristics, information technology capability). This information can inform the content of the intervention package.
Drafting the intervention package
An intervention package is then drafted that is suitable for implementation in community-based organizations. Training and technical assistance plans are also drafted during this phase. Packaging involves the conversion of intervention protocols into non-academic language and into a user-friendly manual that can be readily disseminated . The package should be drafted by the developers of the intervention, but edited by non-technical writers to ensure that the intervention package materials are free of technical jargon. Supporting documents, such as a staff training curriculum, a technical assistance guide, and promotional materials are also included. The package is refined and finalized during the pre-implementation phase of REP based on input from the pilot-testing organizations.
The REP intervention package conveys the intervention's theoretical foundation (i.e., core elements), components, and methods. Core elements – the critical features of the design and intent of the intervention that are thought to be responsible for the intervention's effectiveness – are essential for maintaining fidelity. Still, while core elements are standardized, the mechanisms by which they can be operationalized may vary across organizations. The package can therefore articulate menu options for adapting intervention delivery across different organizations to allow flexibility in implementation. For example, menu options for organizations with a substantial number of minorities may include culturally appropriate options for delivering the intervention, such as outreach to families or community groups. Prior evidence from the management literature suggests that articulating a priori the core elements and adaptation options of a new product to potential users of the product is necessary for its successful adaptation . Thus, having the core elements detailed, while also providing options for implementing these core elements, is vital for optimizing both fidelity to the intervention and flexibility in its implementation.
The REP packaging process has distinct advantages over current strategies for disseminating interventions. For example, a number of behavioral health intervention toolkits (i.e., descriptions of the intervention components) have been published or posted on websites, such as the U.S. Substance Abuse and Mental Health Services Administration's National Health Information Center Evidence-Based Practices website . However, unlike REP packages, these toolkits often lack a detailed description of the intervention's set-up procedures, underlying theories and logic flow, scripts, and other specific materials, as well as instructions for implementing these materials across different settings. In contrast, REP packages provide specific details regarding the intervention as well as operationalized options for adapting delivery of intervention core elements to local organizations in a way that does not compromise the intervention's core elements. The REP packaging process has already been widely used by the CDC to successfully implement and disseminate HIV interventions. For example, between 1996 and 2005, the REP project packaged ten interventions for use by health departments, clinics, and community-based organizations; and more packages are now being prepared. CDC and state health departments fund the implementation of REP-packaged interventions by over 500 prevention organizations nationwide .
Pre-implementation involves 1) input from a Community Working Group (CWG) on developing the package, training, and technical assistance programs, 2) package pilot test, and 3) orientation and logistics in preparing the intervention for implementation.
Community working group – develop package
To accomplish the aims of the pre-implementation phase, the intervention developer convenes a group of stakeholders from organizations serving the target population to participate in a Community Working Group (CWG). We use a comprehensive definition of stakeholders based on the Pincus multi-level 6-P framework that describes the different levels of health care: populations (i.e., communities), purchasers, plans, practices, providers, and patients (consumers) . Given that these different levels play an important role in the use of REP to implement interventions, representatives from all levels should be considered for active participation in the CWG. For example, providers would be knowledgeable of the day-to-day barriers to implementing the intervention, consumers can comment on participant recruitment and burden, and health plan leaders can provide input on how to sustain the intervention beyond the REP implementation phase (e.g., reimbursement of intervention services). CWG participants should be persons who have influence over their peers (i.e., opinion leaders), yet at the same time provide input on the real-world experiences and have the clout to leverage any necessary changes to minimize barriers to intervention implementation.
The CWG members meet regularly throughout the pre-implementation phase to review the intervention materials, finalize the prototype package, advise on the staff training and technical assistance plans, and plan the logistics of the implementation. The goal of these meetings is to review and advise on the package content and develop options suitable for customizing the intervention to specific institutions (e.g., different options for organizing group sessions, care manager communication protocols with providers, guideline dissemination techniques). The intervention developers or researchers familiar with the science and core elements of the intervention should also participate in the CWG in order to provide expertise in the intervention. Notably, core elements and menu options are refined based on CWG input, but the final decisions to include menu options are made by the intervention developers/researchers. Whenever possible, notes should be taken at the CWG meetings to preserve insight and feedback on the package and implementation process garnered from the CWG members.
The final package should include the following components: the intervention technical manual, training curriculum, and guidance for the technical assistance. The technical manual contains a general overview (operationalized core elements, recruitment tips, and selling points for organization staff and leaders on the return on investment); intervention materials (verbatim scripts, and recommendations such as selecting an appropriate venue to conduct the intervention), as well as session workbooks and exhibits; a document describing staff members' roles, time and resource requirements, job qualifications for staff involved, and supervision guidelines; electronic files of printable materials pertinent to the intervention; and pocket cards, flowsheets, and other materials to facilitate the intervention implementation. All materials should be able to be photocopied and easy to read.
Pilot testing the package
Once the package has been reviewed by the CWG, it can be tested for clarity and functionality within a few intervention sites. During this process, the organizations can identify five to eight patients to participate in a full intervention cycle, and study investigators should collect information on feasibility, acceptance, and any problems with the overall package, so that the package can be refined based on their input.
Orientation and logistics
Based on input from the site visits, needs assessments, pilot test, and CWG, the prototype intervention package, training, and technical assistance plans should be finalized for distribution. The package can be distributed through orientation meetings, which can be helpful in bringing recipient organizations together as well. Prior to package distribution, a program champion should be identified at each organization. Program champions can help identify appropriate staff members who can implement the intervention and mobilize support for the intervention within their organization. Program champions, along with key staff members who will be implementing the intervention, should be asked to attend the orientation meeting.
The implementation phase begins upon the dissemination of the REP package to organizations, and continues with intervention training, technical assistance, and evaluation (e.g., fidelity and outcomes monitoring). The CWG continues to be involved in this phase to assist researchers in interpreting feedback from the evaluation process.
Training of selected staff members at participating organizations is essential in implementing and sustaining the intervention. The training should begin soon after package dissemination and should review the manuals and include role playing to showcase specific communication skills with participants. Booster training sessions should also be implemented later in the implementation process.
The technical assistance (TA) component usually occurs after training and involves regular phone calls with the organization representative within a month after training is complete. Regular calls are necessary in order for the TA experts to be proactive rather than reactive when problems regarding implementation need to be solved. The trained TA specialist takes notes during the call and documents implementation progress at each organization. He or she also advises on how to maintain fidelity (e.g., discerning core elements from menu options), integration with existing services, and troubleshooting the implementation process. One of the biggest challenges is ensuring that the core elements are maintained (fidelity) while its implementation is customized and adapted to local design specifications (flexibility). The TA specialist should be encouraged to discuss the tension between fidelity and adaptability with organizations, stressing that the essential core elements can be implemented using different methods while remaining consistent with the intervention's intent, and to provide positive feedback on the organization's efforts. The TA specialist should also be trained to address fidelity issues with both more- and less-experienced employees, as the former may be reluctant to change their traditional ways of patient care, while the latter may require more intensive training.
Four types of evaluation strategies ought to be considered: interpretative evaluation of the intervention implementation process; measurement of intervention fidelity at the organization and patient level; patient-level outcomes; and return on investment (e.g., costs). Interpretative evaluation involves collection of data via qualitative interviews of providers and consumers to determine how the intervention was actually implemented, and notes from training and TA calls to determine to what extent these programs were useful for the organizations. Intervention fidelity measures should be developed to determine whether core elements were successfully implemented. For example, medical charts and staff logs can be used to count the frequency and intensity of services provided, as well as patient attendance and adherence, and staff interviews can be used to gather additional information on whether core elements were implemented correctly. Patient-level outcomes, including processes of care and clinical and functional outcomes, should be chosen based on which outcomes the intervention was designed to impact. The return on investment (e.g., cost-benefit, or how does the cost of implementing the intervention compare to the savings in patient care) is one of the most important evaluations to be conducted, and represents a key argument in making the business case for the intervention to stakeholders (e.g., health care purchasers, plans, and consumers).
Feedback and refinement
Once the evaluation is complete, results should be vetted to the CWG for input and suggestions on how to refine the package for further dissemination. In addition, stakeholders on the CWG with knowledge and leverage over the financing of care should consider how the intervention could be sustained beyond the implementation.
Maintenance and evolution
Maintenance and evolution involve organizational or financial changes to sustain the intervention (e.g., incorporation into job duties, securing funding, and training new personnel), re-customizing intervention delivery as circumstances change, and preparing the refined package, training, and TA program for national dissemination. This phase is often considered the most challenging and least studied, in part because its key component (sustaining interventions) involves concerted multilevel efforts to change the current practice and the organizational and financial incentives to make possible long-term national adoption. For example, one of the most promising strategies to sustain health services interventions is to create billable services based on the intervention's core elements. Among health care purchasers and health plans, sustainability may involve making the business case for initiating financial incentives to provide services that are crucial to the intervention. Sustaining these changes may require strategies beyond financial incentives, such as publicizing results on improved outcomes. Ultimately, good evaluation data, especially on quality performance measures and the return on investment, are essential for making the business case for these changes.
Applying the REP process to health care settings
REP is a promising approach to implement effective clinical and health services interventions in community-based settings because many of these interventions involve complex behavioral components akin to HIV prevention interventions (e.g., group psychoeducation and individual self-efficacy training). There are several lessons learned from CDC's experience of REP that can be applied to implementing interventions in routine health care settings: time and resources are needed to develop prototype packages of effective interventions so that they have fidelity to the original intervention trial procedures and yet are presented in language that is friendly to community-based organizations; staff orientation and training are essential for the faithful replication of effective interventions; and when materials and training are given in ways that are accessible to community-based organizations, science-based intervention procedures are enthusiastically adopted.
How REP addresses multilevel barriers to implementing effective treatments in health care settings
REP components addressing barriers
Health care purchasers
Lack of awareness of evidence-based interventions
Lack of a "business case" for technology transfer models
Business case for REP process (added value of training/TA) via evaluation
Different provider organizations fragment implementation efforts
Lack of financial incentives to implement intervention
Package can be disseminated to several sites
Community Working Group (CWG) to include plan/purchaser and minority representatives- match model to program/state mandates
Lack of time, resources to train staff in intervention
Interventions not adapted to practice organization
Lack of engagement if intervention is imposed on them
Manual and guidance on intervention application through TA to facilitate customization to local sites based on input from site representatives from CWG, enhance group scheduling, phone self-management sessions, and culturally appropriate options including community and family engagement
Lack of time, information
Lack of training opportunities in intervention
REP packaging- manual on how to supplement provider services
Lack of access to customized interventions
Lack of ongoing support, interest
Identification and packaging of interventions tailored to different populations; Identification of intervention core elements
In addition, many front-line providers lack the time and tools to implement interventions in everyday practice, and often feel that these interventions are being imposed on them without their input . Finally, at the patient level, the HIV intervention literature has taught us that one size does not fit all, and that intervention and dissemination efforts need to be customized to particular patient populations [18, 19]. This means that separate intervention packages ought to be considered for different groups if there is substantial evidence that the intervention has a differential effect on outcomes across group differences (i.e., moderator effects).
Discussion: future directions
Closing the gap between research and practice is a priority, especially for community-based practices outside the academic realm. The implementation of effective interventions in community-based organizations has been impeded by the multilevel barriers across the health care system. Practical frameworks to facilitate the implementation of interventions in routine care settings are needed to facilitate a more rapid implementation of research into practice. REP is a promising tool that can be generalized to implement and disseminate clinical and health services interventions .
Nonetheless, REP has not been evaluated for its effectiveness beyond disseminating HIV prevention interventions, nor has it been evaluated for its effect on patient outcome or costs (return on investment). Kelly and colleagues  conducted the first randomized controlled trial of REP in over 70 AIDS service organizations from across the U.S. Organizations were randomized to receive the REP package, or the package, training, and TA. Those receiving the package, training, and TA were more likely to use the intervention, and intervention fidelity was greater than the package-only group. Moreover, REP was equally effective in implementing different types of HIV interventions (e.g., group-level and individual-level) for different populations (e.g., at-risk women, injection drug users, etc.) However, it was not known whether TA and training improve outcomes at the patient level above and beyond packaging alone, and whether the added time and costs of training/TA result in a more successful implementation of the intervention and improved patient outcomes.
Furthermore, a more thorough assessment of what constitutes an intervention core element is needed. Many health services interventions are limited in their ability to be replicated because they often involve multiple components that are idiosyncratic to the particular setting. Hence, a more sophisticated analysis of intervention mediators, or active ingredients, is warranted in order to develop a more precise logic flow between the intervention core element and desired outcome.
Moreover, there have been no comprehensive studies of the long-term effects of REP beyond the implementation phase (i.e., sustainability). For example, implementation fidelity may change in the months and years following training, and as the intervention package is disseminated to organizations that are at different stages of adoption (e.g., late majority, laggards, per Rogers' diffusion model). There is also little research on the degree to which this final phase could or should be tailored for different types of adopters. For example, are the late majorities and laggards more likely to respond to financial incentives, or should a technical assistance program be adapted specifically for them? Further research is needed on whether the added investment of training and technical assistance can mitigate this voltage drop in fidelity and patient-level effect, especially across the different stages of adoption. Moreover, incentives used in health services to promote long-term sustainability of an intervention, such as billing codes, performance measures, or administrative changes (e.g., job function changes) show promise but have not been fully studied on a large scale.
Applying the REP framework to health services interventions can potentially address these knowledge gaps, notably through studies focused on whether interventions packaged through the REP process improve patient outcomes, and which of the intervention's key components are likely to foster its sustainability beyond the implementation phase. For example, we are evaluating the effectiveness of REP in implementing a bipolar disorder care program within community-based practices using a randomized controlled trial approach (e.g., randomizing organizations to receive training and technical assistance versus dissemination of the bipolar disorder program package alone). Nonetheless, quasi-experimental designs with non-equivalent comparison groups are promising alternatives, especially if health care organizations are resistant to randomization, or if there are insufficient numbers of sites. Such designs can also be conducted on a larger scale, especially if patient outcomes are measurable at the population level (i.e., using claims data), avoiding the expense of primary data collection.
Overall, REP is an valuable framework for implementing health care interventions, as it specifies steps needed to maximize fidelity to effective interventions while allowing opportunities for flexibility (i.e., community input) to maximize transferability. REP should be further applied to implement health services interventions and evaluated for its effectiveness in implementing different interventions in health care settings. Strategies that foster the long-term outcomes and sustainability of REP as a tool to implement effective health care interventions need to be developed and tested in health care settings.
The findings and conclusions in this report are those of the authors and do not necessarily represent the views of the Centers for Disease Control and Prevention or the Department of Veterans Affairs.
This research was supported by the U.S. Department of Veterans Affairs, Veterans Health Administration, Health Services Research and Development Service (IIR 02-283-2; A. Kilbourne, PI). Dr. Kilbourne is funded by the Career Development Award Merit Review Entry Program from the VA Health Services Research and Development Service.
- National Institutes of Health: Roadmap for Medical Research. 2005, [http://nihroadmap.nih.gov/initiatives.asp]Google Scholar
- Roy-Byrne PP, Sherbourne CD, Craske MG, Stein MB, Katon W, Sullivan G, Means-Christianson A, Bystritsky A: Moving treatment research from clinical trials to the real world. Psychiatr Serv. 2003, 54: 327-32. 10.1176/appi.ps.54.3.327.View ArticlePubMedGoogle Scholar
- Bradley EH, Webster TR, Baker D, Schlesinger M, Inouye SK, Barth C, Lapane KL, Lipson D, Stone R, Koren MJ: Translating research into practice: speeding the adoption of innovative health care programs. The Commonwealth Fund Issue Brief. 2004, 1-12. 724Google Scholar
- Kelly JA, Heckman TG, Stevenson LY, Williams PN, Ertl T, Hays RB, Leonard NR, O'Donnell L, Terry MA, Sogolow ED, Neumann MS: Transfer of research-based HIV prevention interventions to community service providers: fidelity and adaptation. AIDS Educ Prev. 2000, 12 (5 Suppl): 87-98.PubMedGoogle Scholar
- Rotheram-Borus MJ, Rebchook GM, Kelly JA, Adams J, Neumann MS: Bridging research and practice: community-researcher partnerships for replicating effective interventions. AIDS Educ Prev. 2000, 12 (5 Suppl): 49-61.PubMedGoogle Scholar
- Duan N, Braslow JT, Weisz JR, Wells KB: Fidelity, adherence, and robustness of interventions. Psychiatr Serv. 2001, 52: 413-10.1176/appi.ps.52.4.413.View ArticlePubMedGoogle Scholar
- Landon BE, Wilson IB, McInnes K, Landrum MB, Hirschhorn L, Marsden PV, Gustafson D, Cleary PD: Effects of a quality improvement collaborative on the outcome of care of patients with HIV infection: the EQHIV study. Ann Intern Med. 2004, 140: 887-96.View ArticlePubMedGoogle Scholar
- Mittman BS: Creating the evidence base for quality improvement collaboratives. Ann Intern Med. 2004, 140: 897-901.View ArticlePubMedGoogle Scholar
- Green LW, Glasgow RE: Evaluating the relevance, generalization, and applicability of research: issues in external validation and translation methodology. Eval Health Prof. 2006, 29: 126-53. 10.1177/0163278705284445.View ArticlePubMedGoogle Scholar
- Cohen D, McDaniel RR, Crabtree BF, Ruhe MC, Weyer SM, Tallia A, Miller WL, Goodwin MA, Nutting P, Solberg LI, Zyzanski SJ, Jaen CR, Gilchrist V, Stange KC: A practice change model for quality improvement in primary care practice. J Healthc Manag. 2004, 49: 155-68.PubMedGoogle Scholar
- Sales A, Smith J, Curran G, Kochevar L: Models, Strategies, and Tools: Theory in implementing evidence-based findings into health care practice. J Gen Intern Med. 2006, 21 (Suppl 2): S43-S59.PubMedPubMed CentralGoogle Scholar
- Rubenstein LV, Mittman BS, Yano EM, Mulrow CD: From understanding health care provider behavior to improving health care: the QUERI framework for quality improvement. Quality Enhancement Research Initiative. Med Care. 2000, 38 (6 Suppl 1): I129-I141.PubMedGoogle Scholar
- Simpson DD: A conceptual framework for transferring research to practice. J Subst Abuse Treat. 2002, 22 (4): 171-82. 10.1016/S0740-5472(02)00231-3.View ArticlePubMedGoogle Scholar
- Bartholomew LK, Parcel GS, Kok G: Intervention mapping: a process for developing theory- and evidence-based health education programs. Health Educ Behav. 1998, 25 (5): 545-63. 10.1177/109019819802500502.View ArticlePubMedGoogle Scholar
- Richardson JL, Milam J, McCutchan A, Stoyanoff S, Bolan R, Weiss J, Kemper C, Larsen RA, Hollander H, Weismuller P, Chou CP, Marks G: Effect of brief safer-sex counseling by medical providers to HIV-1 seropositive patients: a multi-clinic assessment. AIDS. 2004, 18: 1179-1186. 10.1097/00002030-200405210-00011.View ArticlePubMedGoogle Scholar
- Kelly JA, Somlai AM, DiFranceisco WJ, Otto-Salaj LL, McAuliffe TL, Hackl KL, Heckman TG, Holtgrave DR, Rompa D: Bridging the gap between the science and service of HIV prevention: transferring effective research-based HIV prevention interventions to community AIDS service providers. Am J Public Health. 2000, 90: 1082-8.View ArticlePubMedPubMed CentralGoogle Scholar
- Centers for Disease Control and Prevention: Replicating Effective Programs Plus website. 2006, [http://www.cdc.gov/hiv/projects/rep/default.htm]Google Scholar
- Semaan S, Kay L, Strouse D, Sogolow E, Mullen PD, Neumann MS, Flores SA, Peersman G, Johnson WD, Lipman PD, Eke A, Des Jarlais DC: A profile of U.S.-based trials of behavioral and social interventions for HIV risk reduction. J Acquir Immune Defic Syndr. 2002, 30 (Suppl 1): 30-50.View ArticleGoogle Scholar
- Sogolow E, Peersman G, Semaan S, Strouse D, Lyles CM: The HIV/AIDS Prevention Research Synthesis Project: Scope, methods, and study classification results. J Acquir Immune Defic Syndr. 2002, 30 (Suppl 1): 15-29.View ArticleGoogle Scholar
- Kegeles SM, Rebchook GM, Hays RB, Terry MA, O'Donnell L, Leonard NR, Kelly JA, Neumann MS: From science to application: the development of an intervention package. AIDS Educ Prev. 2000, 12 (Suppl 5): 62-74.PubMedGoogle Scholar
- Adams J, Terry MA, Rebchook GM, O'Donnell L, Kelly JA, Leonard NR, Neumann MS: Orientation and training: preparing agency administrators and staff to replicate an HIV prevention intervention. AIDS Educ Prev. 2000, 12 (5 Suppl): 75-86.PubMedGoogle Scholar
- Kraft JM, Mezoff JS, Sogolow ED, Neumann MS, Thomas PA: A technology transfer model for effective HIV/AIDS interventions: science and practice. AIDS Educ Prev. 2000, 12 (5 Suppl): 7-20.PubMedGoogle Scholar
- Sogolow ED, Kay LS, Doll LS, Neumann MS, Mezoff JS, Eke AN, Semaan S, Anderson JR: Strengthening HIV prevention: application of a research-to-practice framework. AIDS Educ Prev. 2000, 12 (5 Suppl): 21-32.PubMedGoogle Scholar
- Rogers EM: Diffusion of Innovations. 1962, New York: The Free Press, 1Google Scholar
- Tax S: The Fox Project. Human Organization. 1958, 17: 17-19.View ArticleGoogle Scholar
- Green LW, Kreuter MW: Health Promotion Planning: An Educational and Environmental Approach. 1991, Mountain View, CA: Mayfield Publishing Company, 2Google Scholar
- Bandura A: Self-efficacy: towards unifying theory of behavioral change. Psychol Rev. 1977, 84: 191-215. 10.1037/0033-295X.84.2.191.View ArticlePubMedGoogle Scholar
- Henderson MG, Souder BA, Bergman A, Collard AF: Private sector initiatives in case management. Health Care Financ Rev. 1988, Spec: 89-95.Google Scholar
- SAMHSA National Health Information Center Evidence-Based Practices. [http://www.mentalhealth.samhsa.gov/cmhs/communitysupport/toolkits/community/]
- Pincus HA, Hough L, Houtsinger JK, Rollman BL, Frank RG: Emerging models of depression care: multi-level ('6 P') strategies. Int J Methods Psychiatr Res. 2003, 12: 54-63. 10.1002/mpr.142.View ArticlePubMedGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.