- Open Access
- Open Peer Review
Realities of replication: implementation of evidence-based interventions for HIV prevention in real-world settings
© Cunningham and Card; licensee BioMed Central Ltd. 2014
- Received: 7 June 2013
- Accepted: 2 January 2014
- Published: 6 January 2014
To have public health impact, evidence-based interventions (EBIs) must be implemented appropriately at meaningful scale. The Center for Disease Control and Prevention’s Replicating Effective Programs and Diffusion of Effective Behavioral Interventions programs disseminate select EBIs by providing program materials and training health providers on their appropriate use and implementation. Sociometrics’ HIV/AIDS Prevention Program Archive (HAPPA) and Program Archive for Sexuality, Health, and Adolescents (PASHA) are likewise the largest EBI collections targeting sexual risk behaviors in the private sector. This study examined the extent to which organizations that obtain EBIs from HAPPA and PASHA implement, adapt and evaluate them and factors associated with program implementation.
Survey data were collected from 123 organizations that acquired, and had been in possession for a minimum of six months, at least one EBI from HAPPA or PASHA between January 2009 and June 2011. Data regarding program characteristics and date of acquisition were obtained from Sociometrics’ sales and marketing databases. Logistic regression was used to assess barriers to program implementation.
Among organizations that obtained an EBI from Sociometrics intending to implement it, 53% had implemented the program at least once or were in the process of implementing the program for the first time; another 22% were preparing for implementation. Over the three-year time period assessed, over 11,381 individuals participated in these interventions. Almost two-thirds (65%) of implementers made changes to the original program. Common adaptations included: editing content to be more current and of local relevance (81%); adding, deleting or modifying incentives for participation (50%); changing the location in which the program takes place (44%); and/or changing the number, length and/or frequency of program sessions (42%). In total, 80% of implementers monitored program delivery. Participant outcomes were tracked by 78%; 28% of which used evaluation designs that included a control or comparison group. Lack of adequate resources was significantly associated with decreased likelihood of program implementation (odds ratio = 0.180, p <0.05).
Findings provide greater understanding of implementation processes, barriers and facilitators that may be used to develop strategies to increase the appropriate use of EBIs.
- Evidence-based intervention
- Behavioral intervention
Since the beginning of the epidemic, considerable resources have been spent developing and rigorously evaluating HIV prevention programs. A wide variety of evidence-based interventions (EBIs) have been shown to reduce HIV-related sex and drug risk behaviors among a range of high risk populations. Consequently, over the past several years, government and other funding agencies have expressed a preference for making use of existing, well-tested HIV prevention interventions and approaches ‘that work,’ rather than spending resources on innovation [1–3]. The Centers for Disease Control and Prevention’s Division of HIV/AIDS Prevention (DHAP) includes the implementation of EBIs as part of its strategic plan . EBIs originally developed for HIV prevention are also included on the approved program lists for other health initiatives such as the Office of Adolescent Health (OAH)-administered Teenage Pregnancy Prevention Program, which, in 2010, awarded $75 million in competitive grants to a broad range of organizations and agencies to implement evidence-based programs to prevent teen pregnancy . By replicating effective programs, practitioners operating on limited budgets have a promising blueprint to work with, thereby improving their chances of success.
Several strategies have been developed to promote the translation of efficacious HIV prevention programs into real-world settings. For example, the CDC’s Replicating Effective Programs (REP) and Diffusion of Effective Behavioral Interventions (DEBI) projects provide materials, training, and technical assistance to support implementation of EBIs [6, 7]. Although this effort has been proven effective at encouraging organizations to adopt evidence-based programming, the trainings and technical assistance provided are extremely resource-intensive. Sociometrics Corporation has likewise developed the HIV/AIDS Prevention Program Archive (HAPPA) and Program Archive for Sexuality, Health, and Adolescents (PASHA), collections of HIV prevention program packages shown to be efficacious in preventing HIV infection or its risk-related behavioral antecedents among adults and adolescents, respectively [8–10]. Each program package contains everything that a new site would need to implement an EBI, such as a user guide that gives an overview of the program and the evidence of its effectiveness; a facilitator’s manual that gives step-by-step implementation protocols for each session; and session implementation materials referenced in the facilitator’s manual, such as slides, video clips, participant handouts, activity masters, checklists, and homework assignments for the next session. The program packages also contain evaluation materials such as surveys and questionnaires that were used in the original demonstration of effectiveness and that may be used to reevaluate the program as implemented in a new setting. Programs are selected by an independent Scientific Expert Panel based on information available in peer-reviewed journals regarding quality of program implementation; scientific rigor of evaluation; and positive effects on HIV risk behavior, antiretroviral therapy adherence, or biological markers (STI/HIV rates or viral load). Training is available, but not required, to implement the programs. Free technical assistance is provided to all users upon request as needed. Over the past 17 years, Sociometrics has disseminated thousands of HAPPA and PASHA products, generating more than $1.1 million in sales revenues. ETR Associates and other program developers also independently publish and disseminate EBIs that they created.
The availability of replication kits for effective HIV prevention programs facilitates, but does not guarantee, their appropriate use by practitioners . Harshbarger, Simmons, Coelho, Sloop and Collins, for example, found that of the 162 agencies that had obtained and even sent a staff member to be trained to implement a brief, 45-minute, single-session, HIV prevention intervention, only 38% implemented the program within the expected time frame . Moreover, despite recommendations against making substantial changes to pre-packaged EBIs, organizations frequently intentionally modify the content, scope, focus and/or delivery method [13, 14]. Such changes may be necessary to improve the fit between the intervention and new target population or context and promote maintenance and sustainability of the program over time, but may also compromise effectiveness.
A growing body of literature has identified a wide variety of reasons why organizations may have difficulty implementing EBIs [15–19]. Among the most prominent of these are factors associated with the organizational contexts into which interventions are introduced, such as organizational capacity, congruence with other agency programs or goals, availability of technical and resource assistance, and ‘buy-in’ among staff and other local stakeholders. Despite their perceived importance, however, few empirical studies have examined how such factors affect implementation and maintenance over time .
The overall objective of this study was to examine how EBIs for HIV prevention disseminated via Sociometrics’ HAPPA and PASHA are used in real world settings. Specifically, we assessed: the extent to which these EBIs are implemented; the extent to which the EBIs that are implemented are modified, evaluated and/or sustained over time; and organizational characteristics associated with implementation. This is the first implementation study of EBIs for HIV prevention to assess a private-sector alternative to the CDC’s REP and DEBI dissemination programs.
We employed a universal sampling scheme to reach all organizations that had obtained at least one EBI via HAPPA or PASHA in the recent past (between January 2009 and June 2011). During this time, a total of 187 organizations obtained 381 program packages for 50 unique programs. Reasons an agency may order multiple program packages include, but are not limited to, that they plan to implement a particular program in multiple sites or that they would like to closely examine materials for a variety of EBIs before selecting which, if any, to implement. A total of 16 (9%) organizations were unable to be contacted, 48 (26%) were successfully contacted but did not participate in the survey because an individual responsible for the program(s) obtained could not be identified or declined to participate, and 123 (66%) completed the survey. Individuals who completed the survey had worked an average of 9.8 years (standard deviation [SD] = 7.5 years) in HIV prevention and included executive directors, program directors, managers, coordinators, and facilitators, health educators, and other prevention specialists. In total, 107 (87%) were female. A total of 86 (70%) self-identified as being White, 25 (20%) as Black, 10 (8%) as Hispanic, 1 (1%) as Asian, and 1 (1%) as mixed race/ethnicity.
A minimum of five phone call attempts were made and two emails sent to reach an eligible representative from each organization. Eligibility requirements included being 18 years of age or older and responsibility for the selection and/or implementation the HIV prevention program package(s) obtained from Sociometrics. All organizations had been in possession of their selected EBI(s) for six months to almost three years prior to being contacted.
Survey data were collected from September to December 2011 via SurveyMonkey. Items assessed included: organizational characteristics such as type and main activities of the organization, type of area served, extent to which HIV prevention is a priority, number of paid full-time equivalent staff, number of paid full-time equivalent staff who work specifically in HIV prevention, and implementation of other EBIs; reasons for EBI acquisition; extent to which program implementation was achieved; and level of agreement (1 = strongly disagree to 5 = strongly agree) regarding six factors posited to influence program implementation, including lack of program alignment with the needs of local population and/or setting, poor program fit with other agency programs or goals, lack of availability of funding and other resources, absence of a ‘champion’ committed to program implementation, lack of organizational support for the program, and lack of support from local stakeholders. For each factor, responses were dichotomized as ‘yes’ or ‘no’ that it was a barrier to program implementation. Participants representing organizations that had implemented or were currently implementing the program were additionally asked a series of yes or no questions regarding: characteristics of the target population reached (i.e., number, age, sex, race/ethnicity, sexual orientation, HIV status, and substance use behaviors of persons enrolled in the program); type of adaptations made (e.g., translated materials into another language, edited content to be more current and of local relevance, added activities, removed activities); reasons for program adaptations made (e.g., to make the program more acceptable to the target population, to increase the relevance and compatibility of the program with the implementing agency’s mission, to increase ‘buy-in’ for the program among other local stakeholders, to make it more feasible to implement the program given resource constraints); process evaluation activities conducted (e.g., tracked program participation, assessed participant satisfaction with the program, assessed staff attitudes toward program, assessed fidelity of program implementation); outcome evaluation activities conducted (e.g., types of surveys conducted, whether comparison or control groups were included); and training and technical assistance received (e.g., general training on program implementation, training on how to implement or help implementing a specific program, general training on how to adapt evidence-based programs, training on how to adapt or help adapting a specific program, general training on program evaluation, training on how to evaluate or help evaluate a specific program). Data regarding characteristics of the EBIs such as type (e.g., one-on-one, small group, or community-wide), duration (e.g., single versus multi-session), original implementation setting (e.g., clinic-, community-, or school-based), program package format (e.g., download versus hard-copy), and number of months organizations were in possession of the program prior to completing the survey, were obtained from Sociometrics’ sales and marketing databases.
Respondents completed the survey for only one program obtained per organization. For organizations that had obtained multiple programs (as opposed to multiple copies of the same program), study staff randomly selected the program about which the participant completed the survey. The surveys lasted approximately 10 minutes to 30 minutes, depending on whether the organization had ever implemented the program obtained. Survey participants were compensated $25 for their time, either as a personal check or donation to their organization. All of the procedures were approved by Sociometrics’ Institutional Review Board.
All analyses were conducted using SPSS 17.0 (SPSS Inc, Chicago, IL). Initial descriptive analyses were performed to examine the extent to which program implementation was achieved as well as reasons for not implementing the EBI obtained. Statistical analyses were performed in a step-wise fashion using logistic regression with 95% confidence intervals for each odds ratio. First, we tested whether various organizational and program characteristics were associated with whether program implementation had occurred. All of the variables that were significantly associated with program implementation were then entered in a single block. All analyses controlled for length of time an organization had been in possession of the program prior to completing the survey. Statistical significance was defined as p <0.05.
Organizational and intervention characteristics
Organizational and program characteristics (N = 123)
Type of organizationa
Network alliance/umbrella organization
Other (e.g., health department, detention center, school health program )
Both rural and urban
Priority of HIV prevention
Not a priority
A low priority
A medium priority
A high priority
Don’t know/No response
Number of paid full-time staff
Don’t know/No response
Number of full-time staff devoted to HIV prevention
Don’t know/No response
Implementing other EBI for HIV prevention
Don’t know/No response
Original implementation setting
Program package format
The majority of EBIs obtained by organizations included in the study were small-group (76%), multi-session (93%) programs, originally implemented in community-based (40%) or clinic-based (39%) settings. A total of 57% of the program packages for these EBIs were obtained from Sociometrics in a downloadable, versus hard-copy, format. On average, organizations had been in possession of their EBIs for 16.4 (SD = 8.1) months.
A total of 45 (37%) organizations had neither implemented their selected EBI nor were planning to do so. Of these, 19 (42%) had not ever intended to implement the program. Rather, commonly cited reasons for EBI acquisition included: to obtain information about the program for a grant application or research proposal (11%); to use activities from the intervention as part of another program offered (5%); to help inform the development of a new program (11%); and/or to provide training or technical assistance to others implementing the program (42%).
Among organizations that obtained their selected EBI with the intent to implement it (N = 104), 55 (53%) had implemented the program at least once or were implementing it for the first time (30% and 23%, respectively), and another 23 (22%) were in the process of preparing for implementation. Among those that had implemented their selected EBI at least once (N = 31), 68% had done so more than once, and 48% were still offering the program.
EBI implementation setting and participant characteristics (N = 55)
Black or African American
Native Hawaiian or other Pacific Islander
American Indian/Alaska Native
Gay/Men who have sex with men (MSM)
Both negative and positive
Types of and reasons for changes made to EBIs a (N = 36)
Types of Adaptations
Translated program materials into another language
Edited content to be more current and of local relevance
Changing the number, length, and/or frequency of program sessions
Adding activities without removing any of the original ones
Deleting activities without adding any new ones
Adding, deleting, or modifying incentives for participation
Changing the location in which the program takes place (e.g., from a clinic to a community-based organization)
Reasons for adaptations
To make the program more relevant, acceptable, or appealing to the target population
To increase the relevance and compatibility of the program with the implementing agency’s mission
To increase ‘buy-in’ for the program among other local stakeholders
To make it more feasible to implement the program given resource constraints
Among organizations that implemented the EBI they obtained (N = 55), 44 (80%) had conducted or were conducting one or more of the following process evaluation activities: tracking program participation (84%); assessing participant satisfaction with the program (86%); assessing staff attitudes toward the program (66%); and assessing fidelity of program implementation (64%). In total, 43 (78%) had assessed and/or were tracking participant outcomes such as changes in participant knowledge, attitudes or behaviors. Study designs for these assessments included: post-test survey only (5%); pre-test and post-test surveys (53%); and pre-test, post-test, and follow-up surveys (42%). A total of 28% included a control or comparison group. Among organizations that reported having made adaptations to the EBI they implemented or were currently implementing (N = 36), 89% and 75% assessed or were monitoring program delivery and/or assessing participant outcomes, respectively.
Factors associated with program implementation
Odds ratios for potential factors associated with program implementation
Organization implemented or is currently implementing program, N (%)
Yes (N = 55)
Noa(N = 49)
Program did not meet (and could not be easily adapted to meet) the needs of the local population and/or setting.
Program is not a good fit (and cannot be easily adapted to be a good fit) with other agency programs or goals.
Lack of adequate funding, staff, and other resources are available to implement the program.
Lack of a ‘champion’ committed to implementing the program.
Lack of organizational support for the program.
Lack of support from local stakeholders.
In bivariate analysis, program implementation having not occurred was associated with a lack program alignment with the needs of local population and/or setting; lack of adequate funding, staff, and other resources; and lack of the presence of a ‘champion’ committed to program implementation. In multivariable analysis, only lack of adequate funding, staff, and other resources remained independently associated with decreased likelihood of organizations having implemented the EBI they obtained (odds ratio = 0.180, p <0.05).
Among program implementers (N = 55), 85% reported having received a grant or donation to fund the program. Only 5% reported that no additional funds were needed for program implementation; rather, the EBI had been integrated into other ongoing initiatives, programs or activities. The remainder did not specify how program implementation had been or was being supported.
Training and technical assistance received
Among organizations that implemented the EBI they obtained (N = 55), 76% reported having received training or technical assistance with program implementation and/or evaluation activities including: training on program implementation in general (45%); training on how to implement or help implement a specific program (47%); training on how to adapt EBIs in general (47%); training on how to adapt or help adapt a specific program (33%); training on program evaluation in general (45%); and/or training on how to evaluate or help evaluate a specific program (31%). Likewise, among organizations that were planning for but had not yet begun program implementation (N = 23), 65% reported that staff had received or would receive training and/or technical assistance in one or more of the areas described above.
In order to maximize their public health impact and successfully change behavior among diverse populations, EBIs must be widely disseminated, appropriately implemented, and sustained over time. Our findings indicate that organizations may obtain materials for evidence-based HIV prevention programs for a variety of reasons, including: to implement the program, to obtain information for a grant application or research proposal, and/or to provide training or technical assistance to others implementing the program. Among those who obtained EBIs with the intent to implement them, a majority (53%) had implemented the program at least once or were in the middle of implementing it for the first time. Almost another quarter (22%) were in the preimplementation phase (i.e., preparing for implementation). Thus, a total of 75% of organizations were actively working with their respective programs within a minimum of six months of having obtained them. Moreover, many organizations that obtained EBIs via HAPPA and PASHA have successfully been able to sustain program implementation over time.
These rates of adoption are consistent with, and in some cases higher than, those reported for DEBI programs [12, 20], despite attendance at program-specific training not being a requirement to obtain intervention materials from Sociometrics. Having received training is associated with an increased likelihood of AIDS services organizations offering EBIs to their clients . Nearly half of organizations that had implemented or were currently implementing the EBI they obtained via HAPPA or PASHA reported that they had received training on or technical assistance with program implementation in general (45%) and/or for their specific EBI (47%). Training is not mandated to acquire programs from these collections; thus, this finding indicates that many sites will take the initiative to seek such assistance on their own.
Within the three-year study time frame, EBIs obtained from Sociometrics were implemented with a diverse pool of over 11,300 individuals. The target populations for the majority of these programs included blacks and/or Hispanics, groups disproportionately affected by HIV and who have been a priority for new EBI development . Adolescents were the most common age group targeted, particularly adolescent females, likely due to several current funding streams focused on teen pregnancy prevention that include EBIs disseminated via PASHA on their lists of approved programs such as the OAH-sponsored Teen Pregnancy Prevention Program . Another high priority population exposed to over a quarter of the EBIs implemented by organizations assessed in this study was gay men, including men who have sex with men (MSM) who may or may not self-identify as gay. Few programs that were implemented or being implemented specifically targeted PLH. An increased number of evidence-based secondary HIV prevention programs have become available in recent years. Coupled with increasing efforts to include behavioral interventions for PLH as part of more comprehensive approaches to HIV prevention [4, 21], the proportion of organizations implementing such programs is expected to grow.
Our finding that almost two-thirds (65%) of program implementers had made changes to the original content and/or design is consistent with previous implementation studies of DEBI programs, where adaptations were found to be commonplace [12, 20, 22, 23]. The rationale for these changes was likewise similar, with the most commonly cited reasons being to make the program more relevant, acceptable or appealing to the target population and/or more feasible to implement given resources constraints. The opportunity to make adaptations may also enable agency staff to develop a sense of ownership of a program .
The most common type of changes reported by study participants were editing content to be more current and of local relevance, and adding, deleting or modifying incentives for participation. These are relatively minor, and generally considered allowable, adaptations. Conversely, changing the number, length and/or frequency of program sessions, substituting activities, and/or deleting activities (adaptations reported by 42%, 28%, and 11% of participating organizations that had implemented or were implementing the EBIs they obtained, respectively) are much more likely to threaten program fidelity and, in turn, potentially negatively affect desired outcomes. The ADAPT and ADAPT-ITT frameworks are logic models that detail the steps involved in organizations’ adoption, adaptation and implementation of EBIs [13, 24]. There is a recognized need, however, for more explicit guidance regarding how specific programs can and cannot be adapted during translation .
In recent decades, the nonprofit sector has faced increasing pressures of accountability for program success, with an emphasis on measuring outcomes and impacts rather than just activities and inputs . Many funders, for example, now require detailed evaluation plans as part of proposal applications. Most organizations (80%) that implemented the EBI they obtained actively monitored program delivery, including tracking program participation, assessing participant satisfaction with and staff attitudes toward the program, and/or, to a slightly lesser extent, assessing fidelity of program implementation. A similarly large proportion (78%) had assessed and/or was tracking participant outcomes such as changes in participant knowledge, attitudes and behaviors. Only a small minority of these (28%) included a control or comparison group. HAPPA and PASHA are marketed as evidence-based products for service providers. Service providers who implement EBIs should conduct outcome monitoring to determine whether a reduction in risk behaviors occurs among their clients that participate. This is especially true if substantial changes were made to the program. Determining whether risk behavior change is fully attributable to the adapted intervention would require a more complex evaluation design using some comparison condition, but this is often beyond the capacity of many direct HIV prevention service providers. Other research has also shown that agencies seldom pre-test or pilot EBIs to which they have made changes prior to implementation .
Although use of EBIs is considered to be a more cost-effective approach to HIV prevention than developing new programs, of the organizational characteristics assessed in this study, availability of funding, staff, and other resources was nonetheless the only factor found to be independently significantly associated with whether program implementation took place. For most (85%) of the organizations that had achieved program implementation, this work was supported by an outside grant or donation. Given funding to implement an EBI, two other important factors associated with program implementation are the goodness of fit between the intervention and needs of the local context, and the presence of a program champion. The relationship between other factors associated with organizational context not assessed in this study (e.g., culture, climate, work attitudes) and program implementation should be the focus of future research . In particular, it would be beneficial for future work to investigate how these organizational contextual factors might facilitate successful implementation and maintenance of EBIs as part of their services, absent additional funding.
There were several limitations to this study. Participants represented only 66% of eligible organizations; thus, caution should be exercised when generalizing the findings. Selection bias among the respondents may have led to a skewed picture of EBI implementation. This study also relied on self-reported, rather than objective, measures of program implementation, adaptation, and evaluation activities, thus may be subject to recall and/or social desirability biases. We attempted to mitigate the former by limiting our sample to include only organizations that had obtained their respective EBIs in the past three years. Likewise, the selection criteria for respondents included that they have been involved in the review, selection, implementation, and evaluation of evidence-based programs at their respective organizations and were familiar with the program package that was obtained from Sociometrics. It is possible, however, that staff turnover may have affected the reliability of some respondents’ answers (e.g., their knowledge of reasons for EBI acquisition if they were not on staff at that time). Although the survey asked about attempts at evaluation, we do not know what measures were assessed or how many of these efforts were able to document significant changes. Owing to the cross-sectional nature of this research, we also cannot make causal interpretations regarding factors that influence program implementation. The study procedures controlled for the influence of agencies that purchased more than one type of program package. It is a good strategy for organizations to consider multiple options and choose that which is the best fit for their contexts. By randomly selecting the program about which representative from these agencies completed the survey, in some cases, it might appear that an agency had failed to implement an EBI when in fact stakeholders may have made a thoughtful determination to implement an alternative one.
The decision to acquire a program is distinct from the decision to implement it and how it gets implemented. When translating EBIs to real world settings, challenges are to be expected. Despite its limitations, this study demonstrates the success of Sociometrics’ program archives to promote the adoption, implementation and evaluation of a wide variety of evidence-based HIV prevention programs among a diverse pool of practitioners nationwide. Moreover, its findings provide greater understanding of implementation processes, barriers and facilitators that may be used to develop strategies to increase the use of EBIs.
This study was funded by the National Center for Minority Health and Health Disparities (Grant R44 MD005177-02). The Program Archive on Sexuality, Health and Adolescence (PASHA) is funded by the National Institute on Child Health and Human Development and the U.S. Office of Adolescent Health. The HIV/AIDS Prevention Program Archive (HAPPA) is funded by the National Institute of Allergy and Infectious Diseases and the National Center for Minority Health and Health Disparities. We thank Dr. Rachel Golden for reviewing an earlier draft of this manuscript.
- National Institutes of Health (NIH): Dissemination, Implementation, and Operational Research for HIV Prevention Interventions (RO1).http://grants.nih.gov/grants/guide/pa-files/PA-08-166.html,
- Norton WE, Amico KR, Cornman DH, Fisher WA, Fisher JD: An agenda for advancing the science of implementation of evidence-based HIV prevention interventions. AIDS Behav. 2009, 13: 424-429. 10.1007/s10461-009-9556-8.View ArticlePubMedPubMed CentralGoogle Scholar
- Rotheram-Borus MJ, Swendeman D, Flannery D, Rice E, Adamson DM, Ingram B: Common factors in effective HIV prevention programs. AIDS Behav. 2009, 13: 399-408. 10.1007/s10461-008-9464-3.View ArticlePubMedGoogle Scholar
- Centers for Disease Control and Prevention (CDC): Division of HIV/AIDS Prevention Strategic Plan. 2011,http://www.cdc.gov/hiv/pdf/policies_DHAP-strategic-plan.pdf, -2015,Google Scholar
- Office of Adolescent Health (OAH): Teenage Pregnancy Prevention: Replication of Evidence-based Programs.http://familyplanning.uchicago.edu/policy/publications-resources/TPPI%20Fact%20Sheet%20FINAL.pdf,
- Collins C, Harshbarger C, Sawyer S, Hamdallah M: The diffusion of effective behavioral interventions project: development, implementation, and lessons learned. AIDS Educ Prev. 2006, 18 (Suppl A): 5-20.View ArticlePubMedGoogle Scholar
- Neumann MS, Sogolow ED: Replicating effective programs: HIV/AIDS prevention technology transfer. AIDS Educ Prev. 2000, 12 (Suppl A): 35-48.PubMedGoogle Scholar
- Card JJ: The sociometrics program archives: promoting the dissemination of evidence-based practices through replication kits. Res Soc Work Pract. 2001, 11: 521-526. 10.1177/104973150101100407.View ArticleGoogle Scholar
- Card JJ, Benner T, Shields JP, Feinstein N: The HIV/AIDS Prevention Program Archive (HAPPA): a collection of promising prevention programs in a box. AIDS Educ Prev. 2001, 13: 1-28. 10.1521/aeap.22.214.171.12426.View ArticlePubMedGoogle Scholar
- Card JJ, Lessard L, Benner T: PASHA: Facilitating the replication and use of effective adolescent pregnancy and STI/HIV prevention programs. J Adolesc Health. 2007, 40: 275-e1-275.e14View ArticlePubMedGoogle Scholar
- Rotheram-Borus MJ, Duan N, Flannery D: Interventions that are CURRES: Cost-effective, useful, realistic, robust, evolving, and sustainable. Facilitating Pathways: Care, Treatment, and Prevention in Child and Adolescent Mental Health. Edited by: Remschmidt H, Belfer ML, Goodyer I. 2004, Heidelberg, Germany: Springer-Verlag Telos, 235-244.View ArticleGoogle Scholar
- Harshbarger C, Simmons G, Coelho H, Sloop K, Collins C: An empirical assessment of implementation, adaptation, and tailoring: The evaluation of CDC’s National Diffusion of VOICES/VOCES. AIDS Educ Prev. 2006, 18 (Suppl A): 184-197.View ArticlePubMedGoogle Scholar
- Dolcini M, Gandelman AA, Vogan SA, Kong C, Leak TN, Leak TN, King AJ, Desantis L, O'Leary A: Translating HIV interventions into practice: community-based organizations experience with the diffusion of effective behavioral interventions (DEBIs). Soc Sci Med. 2010, 71 (10): 1839-1846. 10.1016/j.socscimed.2010.08.011.View ArticlePubMed CentralGoogle Scholar
- Veniegas RC, Kao UH, Rosales R: Adapting HIV prevention evidence-based interventions in practice settings: an interview study. Implement Sci. 2009, 4: 76-10.1186/1748-5908-4-76.View ArticlePubMedPubMed CentralGoogle Scholar
- Durlak JA, Dupree EP: Implementation matters: a review of the research on the influence of implementation on program outcomes and the factors affecting implementation. Am J Com Psychol. 2008, 41: 327-350. 10.1007/s10464-008-9165-0.View ArticleGoogle Scholar
- Glasgow RE, Lichtenstein E, Marcus A: Why don’t we see more translation of health promotion research to practice? Rethinking the efficacy-to-effectiveness transition. Am J Public Health. 2003, 93: 1261-1267. 10.2105/AJPH.93.8.1261.View ArticlePubMedPubMed CentralGoogle Scholar
- Kelly JA, Sogolow ED, Neumann MS: Future directions and emerging issues in technology transfer between HIV prevention researchers and community-based service providers. AIDS Educ Prev. 2000, 12 (Suppl A): 126-141.PubMedGoogle Scholar
- Kelly JA, Somlai AM, DiFranceisco WJ, Otto-Salaj LL, McAuliffe TL, Hackl KL, Heckman TG, Holtgrave DR, Rompa D: Bridging the gap between science and service in HIV prevention: Transferring effective research-based HIV prevention interventions to community AIDS service providers. Am J Public Health. 2009, 90: 1082-1088.Google Scholar
- Veniegas RC, Kao UH, Rosales R, Arellanes M: HIV prevention technology transfer: challenges and strategies in the real world. Am J Public Health. 2009, 99 (Suppl 1): S124-S130.View ArticlePubMedPubMed CentralGoogle Scholar
- Kalichman SC, Hudd K, DiBerto G: Operational fidelity to an evidence-based HIV prevention intervention for people living with HIV/AIDS. J Prim Prev. 2010, 31: 235-245. 10.1007/s10935-010-0217-5.View ArticlePubMedPubMed CentralGoogle Scholar
- Centers for Disease Control and Prevention (CDC): Evolution of HIV/AIDS prevention programs – United States, 1981-2006. MMWR. 2006, 55: 597-603.Google Scholar
- Galbraith JS, Stanton B, Boekeloo B, Kind W, Desmond S, Howard D, Black MM, Carey JW: Exploring implementation and fidelity of evidence-based behavioral interventions for HIV prevention: Lessons learned from the focus on kids diffusion case study. Health Educ Behav. 2009, 36: 532-549. 10.1177/1090198108315366.View ArticlePubMedGoogle Scholar
- Iverson EF, Balusuriya D, Garcia GP, Sheng M, Richardson JL, Stoyahoff S, King JB: The challenges of assessing fidelity to physician-driven HIV prevention interventions: Lessons learned implementing partnership for health in a Los Angeles HIV clinic. AIDS Behav. 2008, 12: 978-988. 10.1007/s10461-008-9392-2.View ArticlePubMedGoogle Scholar
- Wingood GM, DiClemente RJ: The ADAPT-ITT model: A novel method of adapting evidence-based HIV Interventions. J Acquir Immune Defic Syndr. 2008, 47 (Suppl 1): S40-S46.View ArticlePubMedGoogle Scholar
- Moxham C, Boaden R: The impact of performance measurement in the voluntary sector: Identification of contextual and processual factors. Int J Oper Prod Man. 2007, 27: 826-845. 10.1108/01443570710763796.View ArticleGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an open access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.