Testing implementation support for evidence-based programs in community settings: a replication cluster-randomized trial of Getting To Outcomes®
Implementation Science volume 13, Article number: 131 (2018)
Community organizations can have difficulty implementing evidence-based prevention programs. More research is needed on implementation support interventions designed to help these organizations implement programs with quality.
Preparing to Run Effective Programs (PREP) is a randomized controlled trial testing Getting To Outcomes (GTO), a 2-year implementation support intervention. It compares 15 Boys and Girls Club sites implementing CHOICE (control group), a five-session evidence-based alcohol and drug prevention program, with 14 similar sites implementing CHOICE supported by GTO (intervention group). PREP replicates a previous GTO study that had the same design, but featured a teen pregnancy prevention program instead. All sites received typical CHOICE training. Fourteen intervention sites received GTO manuals, training, and onsite technical assistance to help practitioners complete implementation best practices specified by GTO (i.e., GTO steps). During the first year, technical assistance providers helped the intervention group adopt, plan, and deliver CHOICE. Then, this group was trained on evaluation and quality improvement steps of GTO using feedback reports summarizing their own data, which yielded revised plans for subsequent implementation of CHOICE. This paper presents results regarding GTO’s impact on CHOICE fidelity (adherence, quality of delivery, dosage) and the proximal outcomes of the youth participants (aged 10–14)—attitudes and intentions regarding cigarettes, alcohol, and marijuana use. Fidelity was assessed at all sites by observer ratings and attendance logs. Proximal outcomes were assessed via survey at baseline, 3, and 6 months.
After 1 year, fidelity and proximal outcomes were similar between Intervention and control groups. After 2 years (which included GTO quality improvement activities that took place between years 1 and 2), intervention sites had higher ratings of CHOICE adherence and quality of delivery (dosage remained similar). Proximal outcomes did not differ between groups in either year, although there was universally high endorsement of prosocial responses to those outcomes from the start.
Findings suggest that systematic implementation support provided by GTO can help community organizations achieve better fidelity. Findings replicate the implementation results from a previous GTO study using the same design, but with a different evidence-based program and different fidelity measures. Although proximal outcomes did not change, in large part due to ceiling effects, the implementation findings suggest GTO can support a variety of programs.
This project is registered at ClinicalTrials.gov with number NCT02135991. The trial was first registered on May 12, 2014.
Problematic rates of alcohol, marijuana, and other drug use among US adolescents highlight the need for good implementation of prevention evidence-based programs. In 2015, over half of high school seniors reported alcohol use in the past year, with one third reporting being drunk in this time frame. One third of high school seniors report past month drinking, and over 20% report using marijuana monthly . Also, the use of opioids has reached epidemic proportions, and electronic cigarette use has skyrocketed in the past 2 years, outpacing the use of regular cigarettes among youth . The estimated costs of alcohol misuse, illicit drug use, and substance use disorders are more than $400 billion . Despite the need and availability of scores of alcohol and drug prevention evidence-based programs (see the Penn State Clearinghouse, https://militaryfamilies.psu.edu/programs-review/), communities often face difficulty implementing evidence-based programs with the quality needed to achieve outcomes [3,4,5,6,7,8,9]. This poor implementation often results from limited resources and a lack of capacity—the knowledge, attitudes, and skills—individual practitioners need to implement “off the shelf” evidence-based programs.
Strong implementation includes best practices, such as setting realistic goals, thoughtful planning, evaluation, quality improvement, and program sustainability. Many youth-serving organizations require help with these practices. Preparing to Run Effective Prevention (PREP) is a 2-year, randomized controlled trial of an implementation support intervention called Getting To Outcomes® or GTO , which is designed to build capacity for these practices. The aim of the PREP study was to test GTO’s impact on fidelity and youth outcomes of an evidence-based, substance use prevention program called CHOICE , carried out by community-based, youth-serving organizations. PREP is a replication of an earlier GTO study, Enhancing Quality Interventions Promoting Healthy Sexuality (EQUIPS, ) in which the evidence-based program was a teen pregnancy prevention program called Making Proud Choices .
Getting to outcomes—an implementation support intervention
GTO builds capacity for implementing evidence-based programs by strengthening the knowledge, attitudes, and skills needed to carry out implementation best practices for running any program —i.e., goal setting, planning, evaluation, quality improvement, and sustaining. Rooted in social cognitive theories of behavioral change [15,16,17,18] and implementation science theories such as the Consolidated Framework for Implementation Research (see [19, 20]), GTO’s logic model (see Fig. 1) states that GTO training and technical assistance builds practitioner capacity to perform multiple implementation best practices needed for an evidence-based program (see Table 1) . Improved performance of these implementation best practices when delivering a specific evidence-based program can improve program fidelity, which results in more positive outcomes .
GTO was developed by Chinman, Imm, and Wandersman as a written guide and published by the RAND Corporation in 2004 to help individuals conduct drug and alcohol prevention programs , https://www.rand.org/pubs/technical_reports/TR101.html. It was developed by reviewing multiple literatures on planning, implementation, and program evaluation and then distilling down key points that could be more easily understood by community-based practitioners . Also, tools—or worksheets—were added to the guide to prompt users to make and record key decisions. As part of the first GTO study, a quasi-experimental trial from 2002 to 2005 , RAND added face to face training and ongoing technical assistance to the existing written guide to increase GTO’s impact. From then on, in all subsequent studies, the GTO approach provides three supports: (1) the GTO manual (tailored to a variety of content domains including drug and alcohol prevention, which was used in PREP, ), (2) face-to-face training, and (3) ongoing, onsite, and proactive technical assistance. GTO has been applied to multiple content areas including teen pregnancy prevention , underage drinking prevention , and positive youth development .
Key to GTO’s capacity-building is asking practitioners to be active learners. GTO establishes expectations and gives opportunities and guidance for practitioners to carry out for themselves the implementation best practices that GTO specifies.
In previous quasi-experimental  and randomized controlled trials [19, 28], GTO has been found to improve capacity of individual practitioners and performance of alcohol and drug prevention programs. However, those studies involved mostly non-evidence-based programs of widely varying type and quality, and thus were not able to assess common outcomes across program participants. The EQUIPS study had the same design as PREP and showed that community-based organizations (Boys and Girls Clubs) using GTO demonstrated better capacity, performance, fidelity, and youth outcomes from a teen pregnancy prevention evidence-based program (Making Proud Choices) than clubs not using GTO [12, 29]. Other work has also demonstrated that implementation support can improve fidelity and outcomes of substance use prevention evidence-based programs, but those trials were not able to track programming being implemented, or its fidelity, in the control communities [30, 31] or track technical assistance usage and blind fidelity observers .
Contributions of the PREP study
PREP builds upon past studies of implementation support in general, and of GTO in particular. The PREP design replicates the EQUIPS study (cluster randomized controlled trial comparing evidence-based program vs evidence-based program+GTO), but does so using a different evidence-based program (i.e., CHOICE) in a different content domain (i.e., substance use), with some similar and some different measures of fidelity and outcomes (access to the CHOICE developer, ED, was helpful to ensure we used the same measures as in past CHOICE trials). Replicating findings in a different content domain represents a strong test of GTO’s robustness. In particular, testing the fidelity of CHOICE included examining fidelity to motivational interviewing, a non-judgmental, non-confrontational counseling approach typically used with a variety of health risk behaviors [33, 34]. Given that CHOICE program delivery utilizes motivational interviewing , delivering the program involves a more complicated set of skills than many universal prevention programs require, and thus presents a greater challenge to achieve program fidelity. Finally, PREP, like EQUIPS, has rigorous design features that past GTO and other implementation support studies have not been able to incorporate including use of a single evidence-based program to reduce variation between intervention and control groups, measures of implementation (fidelity) in the intervention and control groups, fidelity observers blinded to group condition, and measures of both fidelity and individual outcomes [19, 24, 28, 36,37,38,39].
PREP is a 2-year randomized controlled trial (RCT) comparing 15 sites within 8 Boys and Girls Clubs (BGCs) who received typical training to implement the CHOICE program  (control group) with 14 sites within seven BGCs who received the same CHOICE training, plus GTO manuals, training, and technical assistance (intervention group). As in EQUIPS, GTO was provided over a 2-year period, allowing all sites to deliver CHOICE twice. The trial assessed fidelity (e.g., adherence, quality of CHOICE delivery, dosage) and the alcohol and drug outcomes of participating middle school youth. Based on results from the EQUIPS trial [40, 41], it was hypothesized that the intervention sites would be higher on fidelity than the control sites in the second year, and that the youth in the intervention sites would show more improvement in alcohol and drug outcomes than youth in control sites in the second year.
The 29 sites are in the greater Los Angeles, California area, covering Los Angeles (23 sites) and Orange (six sites) counties. BGCs provide youth programming ranging from recreation in gyms to leadership, character education, health and wellness, and academic programs. A BGC often has several sites (i.e., geographic locations). Despite some variability, each site typically has its own facility and a small number of full- and part-time staff (n = 7–10). A sub-set of staff (between 1 and 10; mean = 2.2, median = 3) at each site participated in the study. CHOICE was initially developed in Los Angeles with a diverse population of youth [35, 42] and was therefore appropriate for the mostly Latino and African-American sample found in these BGC sites. Invitations were made to all BGCs in the area (n = 38) via meetings of a BGC alliance. The study team stopped recruiting when the above sample was reached. The site level sample size was justified at 80% power by taking into account the estimated correlation between baseline and follow-up assessments of the site level measures (.5 to .6) and the moderate to large effect sizes expected based on previous GTO  and CHOICE  studies.
The youth level sample size was justified using data from previous CHOICE trials , including the expected correlation between baseline and follow-up assessments (r = 0.4), the intraclass correlation measure of clustering (ICC = 0.3), and the small to medium effect sizes achieved on various outcome measures of interest.
In study year 1, 356 youth in self-reported grades 7–9 participated in the youth survey. These youth ranged in age from 10 to 14 (M = 11.9, SD = 1.0); 48% were in grade 7, 37% in grade 8, and 15% in grade 9. Gender was 50% girls. The survey requested binary responses to ethnicity (Hispanic or Latino/Latina vs. not) and six separate racial identifications; youth were permitted to choose all that applied. Sixty-four percent reported Latinx ethnicity; 17% reported being Black or African American; 13% White or Caucasian; 9% Asian or Asian American; 6% American Indian or Alaska Native, and less than 5% Native Hawaiian or Pacific Islander. “Other race” was selected by 58% of respondents; 86% of these youth indicated Hispanic or Latinx ethnicity. Multiple racial identifications were indicated by 7% of youth.
In year 2, n = 253, ranging in age from 10 to 15 (M = 11.9, SD = 1.0); 55% were in grade 7, 38% in grade 8, and 8% in grade 9. Approximately half (51%) were girls. Sixty-six percent reported Latinx ethnicity; 14% were Black or African American; 12% Asian or Asian American; 11% White or Caucasian; 8% American Indian or Alaska Native, and less than 5% Native Hawaiian or Pacific Islander. “Other race” was selected by 59%; 90% of these youth were Hispanic or Latinx. Multiple races were selected by 7% of youth.
Using a random number generator, we randomized the 15 BGC clubs to intervention (8 BGCs—15 sites) or control (7 BGCs—14 sites). We randomized at the BGC level (versus site) to minimize potential contamination due to communication between sites within the same club system. The principal investigator informed each club about their assignment.
At baseline (after randomization), we conducted a web-based survey of BGC staff involved in CHOICE to assess for potential differences in demographic variables and attitudes toward evidence-based programs. All staff who were contacted, responded (control = 29/29; intervention = 34/34). Staff in the control and intervention groups had largely similar demographic makeup (no significant differences based on bivariate models accounting for clustering within BGC and county). Half (49%) of the staff were female; most (59%) were over 25 years old; half (50%) had a 4-year college degree or more; and 56% were Hispanic or Latinx, 22% were non-Hispanic African-American and 22% were non-Hispanic White, multiracial, or of other races. Over a third (38%) were full-time employees at their respective BGC site.
The web survey included the Evidence-Based Practice Attitude Scale . Its four scales, and their McDonald’s ω coefficients in this study (with 95% confidence intervals [CIs]), assess the degree to which practitioners would adopt an evidence-based program based on the following factors: appeal (how much the evidence-based program was intuitively attractive), ω = .65, CI [.47, .83]; requirements (degree to which it was considered mandatory), ω = .88 [.7, .94]; openness (willingness to trying new interventions), ω = .77 [.66, .85]; and divergence (belief that experience is more important than research), ω = .81 [.66, .89]). Coefficient ω  is a measure of internal consistency on the same scale as coefficient alpha, but makes less biased estimations, has fewer problems with inflation due to number of items, and has CIs to more accurately evaluate reliability . The ω values found here are considered acceptable to good . To evaluate baseline group differences on each scale, we fit a linear mixed-effects regression model with fixed treatment effect (intervention vs. control) and a random club (BGC) and county (Los Angeles vs. Orange) intercepts. There were no significant differences between the two groups on any scale at baseline, ps > .2, with or without staff-level demographic covariates.
CHOICE—an evidence-based alcohol and drug prevention program
CHOICE involves five, half-hour sessions based on social learning theory , decision-making theory , and self-efficacy theory . Delivered using motivational interviewing , CHOICE uses role plays to teach resistance skills and discuss pros and cons of cutting down and stopping use. The five sessions cover (1) normative feedback on alcohol and drug use among middle-school youth; (2) how beliefs about substances can affect behavior and how substances affect people; (3) how substances can be used to deal with stress or other negative emotions and how social substance use can become problematic; (4) how to identify certain pressures to use substances and how to resist them; and (5) how to use problem solving skills to avoid using substances when they are present . CHOICE training consists of two, mostly full-day sessions on content and motivational interviewing role playing. The evidence for CHOICE comes from two randomized trials that found the program was associated with reductions in alcohol and marijuana use [11, 51].
CHOICE implementation supported by GTO
Using existing staff, each BGC site was asked to implement CHOICE once a year for 2 years with a different group of adolescents each year between May 2014 and April 2016. Two half-time, masters level technical assistance providers delivered standard CHOICE manuals and training to all sites. For intervention sites, technical assistance providers also delivered GTO manuals, face-to-face training in GTO, and onsite technical assistance with phone and email follow-up to support implementation during the two rounds of CHOICE delivery (per site in year 1: M = 11.17 h of TA, SD = 3.4; year 2: M = 14.7 h of TA, SD = 3.9). The GTO manual contains written guidance about how to complete multiple implementation best practices important for evidence-based programs—i.e., GTO steps. Most GTO steps contain tools that prompt practitioners to make, and then record, decisions about various practices. For example, the GTO Goals Tool has prompts that assist practitioners to write goal and desired outcome statements. Table 1 shows how BGC staff assigned to the GTO condition performed the various implementation practices in each of the GTO steps to implement CHOICE.
Before the first CHOICE implementation, technical assistance providers delivered two GTO trainings to participating staff at each intervention site. The first focused on GTO steps 1–3 (needs assessment, setting goals and concrete objectives, and best practices). A few weeks later, each site received training on GTO steps 4–6, focusing on program fit, capacity, and planning. Concurrently, technical assistance providers met periodically with BGC staff to help them complete each GTO step (i.e., complete the tools) and guide the planning of CHOICE. Then, BGC staff at all sites implemented CHOICE and facilitated the collection of fidelity and youth outcome data (described below). Intervention sites then received training on evaluation and quality improvement (GTO steps 7–9), along with feedback reports summarizing fidelity and youth outcome data from their sites, which were used in a TA-facilitated quality improvement process that resulted in a revised plan for the second implementation of CHOICE. The year 2 implementation followed the same process and collected the same data, supplemented by training on sustainability (GTO step 10). All 29 BGC sites received $2000 a year to defray some costs of study participation—e.g., supplies and snacks used during CHOICE sessions.
Measures and data collection
PREP was approved by RAND’s Institutional Review Board. Data collectors and technical assistance staff watched for harms of GTO and CHOICE while GTO was active. None were reported.
All sites were rated on three fidelity dimensions—adherence to CHOICE, quality of CHOICE delivery, and dosage . Adherence and quality of CHOICE delivery ratings were made by a pool of eight research data collectors (blind to condition). Rather than calculate reliability by doubling up on observations (which were objectionable to the sites), the entire data collector team rated 16 videotaped sessions developed by the CHOICE trainer, spread over the intervention period. Krippendorff’s α was calculated comparing observers’ ratings for each video to the “master ratings” by the CHOICE trainer.
Each year, data collectors observed and rated two CHOICE sessions per site (randomly selected) on how closely BGC staff implemented activities as designed (not at all, partially, fully) using a CHOICE fidelity tool . A similar tool was used in the EQUIPS study to rate activities  and was found to be reliable (Cohen’s weighted Kappa = 0.92 in year 1; 0.96 in year 2) and valid (showed improvement in the hypothesized direction once GTO was fully implemented in the second year). In each year, a total of 1344 activities were conducted across all 29 sites (a full CHOICE program contains 46 discrete activities). In year 1, we observed and rated 489 activities (36%), distributed across all 29 sites (n = 235 for the control group, 254 for the intervention group). In year 2, we observed and rated 515 activities (38%), distributed across all 29 sites (n = 255 for the control group, 260 for the intervention group). Ordinal α comparing ratings from each of the eight coders to the master key ranged from .50 to .91, median = .70, acceptable to good by common standards .
Quality of CHOICE delivery
Motivational interviewing is key to CHOICE delivery. We measured its fidelity with the Motivational Interviewing Treatment Integrity scale (MITI; ), the standard measure in clinical trials of motivational interviewing-based interventions [54, 55]. The MITI has five specific behaviors that are counted during the session and five ‘global’ ratings in which the entire session is scored on a scale from 1 = low to 5 = high. The MITI has shown acceptable psychometric characteristics across multiple research settings [56,57,58,59], and its scores have correlated with outcomes as expected, suggesting its validity [60,61,62]. However, because the five global ratings had low inter-rater reliability in this study, we omitted them from analyses and relied on the behavioral data as our operationalization of delivery quality.
The behaviors counted during the session are the number of open- and closed-ended questions, statements that are MI-adherent (e.g., “If it’s ok, I’d like to hear what you thought about that.”) or non-adherent (e.g., “You need to stop using drugs”), and reflections that are simple (e.g., “some of you are ready to make changes”) or complex (e.g., “some of you are hoping that by making changes, things will improve in your lives”). Data collectors administered the MITI at the same sessions they rated adherence. From this count data, we derived four indicators used in the analyses, per the MITI scoring instructions: percent complex reflections (complex reflection/total number of reflections), percent open questions (open questions/total number of questions), reflection to question ratio (total number of reflections/total number of questions), and percent MI adherent (number of MI adherent statements/number of MI adherent + MI non-adherent statements). Across all variables derived from behavioral counts, α for each data collector were in a high range from .88 to .93 , median = .90.
This measure was operationalized as the attendance rates at each site. BGC staff at intervention sites sent RAND the recorded attendance of enrolled youth at each CHOICE session. We calculated attendance (in percent) by dividing the number of youth who completed the baseline survey by the number of youth attending each session, averaged across each session, each year. Control site rates were calculated using attendance data gathered by data collectors during their two visits at each site (averaged together), each year.
For three substances—cigarettes, alcohol, and marijuana—all youth participants were asked questions about proximal outcomes, which have been shown to predict actual use. Proximal outcome items come from large national surveys, such as monitoring the future , and have been used in many different randomized controlled trials with youth [11, 65,66,67]. All of these measures were used in the previous trials of CHOICE [11, 51]. These measures include intentions to use cigarettes/alcohol/marijuana. Separate single items assessed student intentions to use in the next 6 months (1 = definitely yes to 4 = definitely no) . Resistance self-efficacy (RSE) measured the likelihood one would refuse an offer of cigarettes/alcohol/marijuana in three specific situations (e.g., best friend using, bored at a party, all friends using at a party). Items are rated on a scale from 1 = I would definitely use to 4 = I would definitely not use, and averaged. Higher scores indicated greater RSE (alcohol α = 0.92; marijuana α = 0.96) . Perception of peer use. Youth were asked three separate questions to estimate the percentage of peers in their grade who use cigarettes, alcohol, or marijuana . They were also asked three separate items for whether they think their best friend uses these substances . Youth reported on three separate items asking them how often they spend time around teens who use cigarettes, alcohol, or marijuana (1 = never to 4 = often) . Positive and negative beliefs were assessed using well-established measures with adolescents [11, 67, 72]. For each substance, youth were asked how much they agreed (1 = strongly agree to 4 = strongly disagree) with various positive consequences of using cigarettes (three items; α = 0.78), alcohol (two items; α = 0.81), and marijuana (three items; α = 0.81). Youth were asked the same number of items asking about negative consequences for each substance (cigarette α = 0.82; alcohol α = 0.82; marijuana α = 0.83). Because most measures were ordinal and substantially skewed, we dichotomized each variable into lower-risk vs. higher-risk responses for each substance to improve estimation in logistic regression models. We coded intentions: “definitely no” vs. any other; for RSE: “would definitely not use” across all scenarios vs. any other combination; for percentage of peers perceived using in their grade, “0” or “10%” vs. 20% or more; for best friend use, “no” vs. “yes”; for spending time with teens, “never” vs. any other; and for positive and negative consequences, the lowest-risk response across all respective consequences vs. any other combination.
We also asked youth three separate items about their lifetime use and three separate items for their 30-day use of cigarettes/alcohol/marijuana. However, the rates of use were very low. Across the entire sample, the rate of 30-day cigarette use was 1.5% (baseline), 1.4% (3 months), and 1.0% (6 months). For alcohol, it was 3.7%, 4.2%, and 6.7%, respectively. For marijuana, it was 3.3%, 5.9%, and 9.7%, respectively. Because of the low use, the statistical models showed convergence difficulties; therefore, we decided to only analyze the proximal outcomes.
Data collection and response rates
In each of the 2 years, the BGC staff in both groups recruited participants for CHOICE from their eligible membership. Staff sent information flyers to parents, approached parents when they were present at the site, and/or held CHOICE information sessions at the site. Parents provided written consent and youth provided assent. RAND (blind to study condition) and BGC staff facilitated paper and pencil youth survey sessions with each site before the launch of CHOICE (Baseline) and at a 3 and 6-month follow-up. We used several methods to boost response rates including make-up sessions, mailed surveys, and phone reminders.
The overall retention rate was 88% at both the 3-month post-test and at the 6-month follow-up. We evaluated differential attrition by condition using discrete-time survival analysis, accounting for clustering within site and BGC, using PROC SURVEYLOGISTIC in SAS. There was a significant effect of assignment to GTO, hazard ratio = 1.09, 95% CI [1.02, 1.16], indicating a greater likelihood of attrition at each follow-up for youth in the GTO group. We also screened for differential attrition by study year, race/ethnicity, gender, age, and lifetime use of different substances at baseline. After applying the Benjamini-Hochberg false discovery rate (FDR) correction, there was a significant difference by gender, with girls more likely to leave the study, HR = 1.06, 95% CI [1.02, 1.10].
We compared intervention and control sites on CHOICE fidelity (adherence, quality of CHOICE delivery, and dosage) and the seven proximal outcomes for cigarettes, alcohol, and marijuana. For each measure, we conducted a mixed model spanning condition and study year, using weighted comparisons to (1) compare intervention and control groups in years 1 and 2 separately; (2) examine year 1-to-year 2 change within intervention and control groups, separately; and (3) assess the interaction term between group and year, to test whether the change from year 1 to 2 differed between groups. All analyses were conducted in SAS v9.4, predominantly with PROC MIXED and PROC GLIMMIX. Effect size estimates were based on spreadsheets provided by Lakens [73, 74]. Confidence intervals for estimated Cohen’s d effect sizes were calculated using the SAS macro Effect_CI .
Type I error control
To control the FDR, we adjusted p values using the Benjamini-Hochberg procedure  such that, across significant findings after adjustment, a proportion of no more than α (.05 herein) reflect type I error. We made this correction within two sets of multiple tests addressing the same conceptual result: the four MI quality indicators and the proximal outcomes analyses for the three substances. Attendance and CHOICE Adherence measures were treated as individual outcomes. We made corrections separately within analyses for year 1, year 2, and the condition by year interaction because these analyses address different conceptual questions. The within-condition change analyses have not been adjusted because they are not part of the study hypotheses and are shown only for illustrative purposes.
We compared control and intervention sites across all three dimensions of fidelity (i.e., CHOICE adherence, MI quality and dosage). For adherence, we fit a mixed-effects, proportional-odds, logistic regression model where the observational unit was one rated CHOICE activity, nested within session, site, BGC, and county. The specific CHOICE activity was included as a rating-level covariate. We report odds ratios and 95% CIs for the treatment effect for the separate year 1 and year 2 analyses, and for the change from year 1 to year 2 by group, and logistic regression coefficient and CI for tests of moderation in the combined years’ analyses. In these models, odds ratios are on an intuitive scale and therefore serve as unstandardized effect sizes. For quality of MI delivery, we tested the site average of each variable in a mixed-effects linear regression model, accounting for nesting within BGC and county. For dosage, we used similar models for respondents’ attendance from the control and intervention groups where the outcome was site-level attendance rate.
Youth proximal outcomes
We compared control and intervention youth on seven proximal outcomes across the three substances: cigarettes, alcohol, and marijuana. Because we had no distinct hypotheses about the seven different proximal outcomes, we combined them for each substance into a multivariate mixed model to improve the statistical power of the analysis. For each substance, we fit a mixed-effects logistic regression model where the observational unit was one binary proximal outcome (the lower-risk response) with random intercepts at the levels of youth, site, BGC, and county. Random slope models that we attempted to fit resulted in substantial convergence difficulties. We calculated intraclass correlations (ICCs) for youth intercepts at the level of BGC (the level of random assignment). BGC-level ICCs ranged from .000 to .027 across substances and years, median = .004. We report results in parallel structure as for the adherence data. Again, odds ratios are on an intuitive scale. Youth age, gender, and race/ethnicity (seven binary indicators with multiple selections permitted) were included as covariates in each model, as were baseline measures of all proximal outcomes for all substances.
Year 1/year 2 group comparisons
In year 1, there were no significant difference between groups in the probabilities of the three protocol adherence ratings (“not at all,” “partially,” “fully”) (see Table 2 for details). Table 2 includes by-condition descriptive and test statistics and an unstandardized odds ratio as an effect estimate for year 1 (leftmost) and year 2 (center). In year 2, intervention sites scored significantly higher on the 3-point adherence scale than control sites. For the intervention group, 2% of activities in year 2 were rated as not at all adherent, 11% as partially adherent, and 87% as fully adherent, similar to the 90% fully adherent ratings in the cluster-randomized trial of CHOICE . For the control group, the comparable numbers were 12%, 22%, and 66%, respectively.
Year 1 to year 2 within group change
As shown in the right-hand portion of Table 2, comparing years 1 and 2, the intervention group had significantly higher ratings of adherence to CHOICE activities in year 2 than in year 1. The control group showed no significant change.
Year 1 to year 2 interactions
The difference in change between the two groups was significant (rightmost column of Table 2). As noted, in year 2, the control group ratings were essentially unchanged from year 1. However, those in the intervention group increased their adherence to the protocol activities by almost four-fold in year 2 (defined as going from “not at all,” to “partially,” or “fully”; going from “partially” to “fully”).
Quality of CHOICE delivery
Year 1/year 2 group comparisons
Across all four MI quality variables, mixed-effects regression models showed no group differences in year 1, ps > .20 (Table 3). The median estimated Cohen’s d was − 0.18. In year 2, however, the intervention group had significantly higher MI quality ratings than the control group for two of the four quality variables. After FDR correction, these differences significantly favored the intervention group for the derived reflection question ratio and for percent MI adherent. The median d across the four ratings was 0.74. According to thresholds for proficiency established by the MITI  (reflection to question ratio = 1.0; percent open questions = 50%; percent complex reflections = 40%; and percent MI adherent = 90%), in year 2, the intervention sites were either near or over these thresholds on three of the four measures (.88, 64%, 32%, 95%, respectively). Control sites were over the threshold for percent open questions only (.81, 59%, 26%, 81%, respectively). Thus, results indicate that the intervention sites were delivering CHOICE with higher MI quality than the control sties in year 2. Further, MI adherent scores for the intervention group (year 1: 94%, year 2: 95%) were very similar to what was reported in the cluster-randomized trial of CHOICE (93%) .
Year 1 to year 2 within group change
Only the reflection question ratio improved from year 1 to 2 in the intervention group, p = .023. None of the measures showed significant year-to-year improvement in the control group.
Year 1 to year 2 interactions
None of the measures showed significantly different change over time between conditions after FDR. The median estimated partial ω2 for the interaction term was .028.
The control and intervention groups did not differ in their attendance in years 1 or 2 or in change between years 1 and 2, ps > .05.
Year 1/year 2 group comparisons
Table 4 has the proportions of youth endorsing a lower risk response for each proximal outcome by year, study group, and substance. Table 5 includes by-condition descriptive and test statistics and an unstandardized odds ratio as an effect estimate for year 1 (leftmost) and year 2 (center). In year 1, the mixed-effects logistic regression model found no significant difference between groups in the probabilities of endorsing the lower-risk response in the dichotomized proximal outcomes. The median odds ratio across the six comparisons was 0.94; based on absolute value of the logistic coefficient, the median effect size was OR = 1.14. In year 2, there were also no significant differences for the alcohol, cigarette, or marijuana outcomes after FDR correction. The median odds ratio was 1.12; median effect size OR = 1.23.
Year 1 to year 2 within group change
As shown in the right-hand portion of Table 5, between years 1 and 2, the control group showed fewer youth endorsing a lower risk response for the 6-month follow-up marijuana proximal outcomes in year 2 than in year 1. The intervention group showed fewer youth endorsing a lower risk response for the 3-month cigarette proximal outcomes in year 2 than in year 1.
Year 1 to year 2 interactions
As detailed in the rightmost column of Table 5, there were no significant differences between conditions in year-to-year change.
We conducted sensitivity analyses to determine whether distorted or arbitrary response sets from youth or inconsistent coding of CHOICE adherence could have resulted in different outcomes. First, to address the possibility of distorted or arbitrary response sets, we asked youth about their use of a non-existent drug called “derbisol.” Across all surveys and waves, 12 participants reported using derbisol at least once. Youth outcome analyses excluding these participants resulted in no differences in interpretation of results. Second, the range of reliability of the coders for CHOICE adherence was substantial. We repeated the CHOICE adherence analyses excluding ratings contributed by the two judges with coding reliability < .65. This analysis likewise resulted in no differences in interpretation of results.
The PREP study assessed GTO’s impact on an evidence-based program’s fidelity and youth proximal outcomes over 2 years. In year 1, intervention and control sites were similar on adherence. Both groups carried out about two-thirds of CHOICE activities in full and about one quarter of CHOICE activities in part, far below the original CHOICE trial. In year 2, as hypothesized, the intervention sites significantly improved adherence, implementing CHOICE activities fully 87% of the time (similar to sites implementing CHOICE in its original trial), while control sites were unchanged.
There was a similar pattern of results between years 1 and 2 for the MI delivery quality. None of the four MI variables were significantly different between the two groups in year 1. However in year 2, as hypothesized, the intervention sites had higher ratings on reflection to question ratio and percent MI adherent and had greater improvement on reflection question ratio from year 1 to 2 compared to the control sites. The intervention sites achieved the same percent MI adherence ratings as sites in the original CHOICE trial. Dosage (i.e., attendance) was not different between the groups in either year. The improvement in fidelity with implementation support documented here is similar to other studies testing implementation support models among alcohol and drug prevention evidence-based programs [77,78,79].
The similarity of the two groups’ fidelity scores in the first year could be because control sites received some GTO-like support simply by being in the study and because the CHOICE materials and training provide some guidance. We concluded this in part because we provided CHOICE training to three other youth-serving organizations in Los Angeles, at no cost, and then after having no contact with them for 6 months, asked them whether they ran any programming. None ran any CHOICE programming, suggesting that the control group’s act of participating in the study might induce some level of program implementation that would not otherwise occur. In the second year, with the addition of GTO’s quality improvement activities, in which plans were developed to specifically improve identified areas of weakness, the intervention group’s adherence and quality delivery ratings were much higher, as hypothesized. Thus, we conclude that working through the evaluation and quality improvement steps of GTO is important to achieving intended outcomes. This often takes time and requires more than 1 cycle of a program, since the benefits of quality improvement are not realized until they are applied to a subsequent program implementation.
Compared to the implementation findings, the youth proximal outcome results did not change much from baseline to 6 months in either year and thus did not show any clear differences between groups, contrary to our hypothesis. Direct outcome comparisons to the earlier CHOICE trial are not possible because that study involved youth with higher rates of substance use and the PREP youth tended to endorse responses to the proximal outcomes that were somewhat lower risk. Thus, one reason the proximal outcomes did not show larger change in our study could be due to a lack of variability in these outcomes—i.e., there was generally high endorsement of low risk responses in our sample. Similar to other studies of implementation support , improvements in implementation fidelity do not always translate into better program outcomes.
PREP’s findings are similar to the implementation results of EQUIPS. Like PREP, EQUIPS showed that sites using GTO had better fidelity results (i.e., adherence and delivery quality ratings) in year 2. In the EQUIPS study, we concluded that “in typical community-based settings, manuals and training common to structured evidence-based programs may be sufficient to yield….moderate levels of fidelity, but that more systematic implementation support is needed to achieve high levels of performance and fidelity [12, pg. 14].” The findings of the PREP study, using a different evidence-based program and measures of fidelity, appear to bolster that conclusion. However, the PREP study goes further, suggesting that GTO can help community-based practitioners carry out with proficiency a more complicated set of skills required by motivational interviewing. This is important as adolescents tend to be more satisfied with interventions that use motivational interviewing and facilitators that use these skills can be more effective in preventing or reducing substance use . Thus, supports like GTO could significantly improve the generally poor implementation of evidence-based programs among youth-serving organizations across a range of prevention programs.
Both EQUIPS and PREP were carried out in Boys and Girls Clubs and are thus generalizable to low-resourced, community-based settings. It is possible that organizations with greater resources and more staff could achieve even better results with GTO. These results were achieved with a modest amount of training and technical assistance time (about 26 h over the 2-year intervention period), which is similar to what government grant programs now offer . A cost analysis is underway from the PREP study that will provide more information about GTO’s return on investment.
There are some limitations that should be noted. First, it was difficult to evaluate GTO’s effects on substance use and proximal outcomes. The sample of youth in this study, ages 10 to 15, had very low base rates of substance use and high rates of endorsing prosocial responses on Proximal outcomes (more than in the original CHOICE trial). More youth in the nearby public school system (of the same age) were Latinx (74% vs. 64%) and fewer were African-American (9% vs. 17%) than the youth in our sample. A great deal of research has shown that African American youth tend to report less substance use than other races/ethnicities [82,83,84,85,86,87,88], so one reason for our lower rates of substance use and high positive proximal outcomes may be because we had more African American youth participate in the study. Alternatively, Boys and Girls Clubs are focused on helping youth make healthy choices across a variety of behaviors; and thus, these youth may have already been exposed to more preventive programming than youth in the CHOICE trials and the nearby public schools system. Second, sites did not have the full experience in doing a needs assessment or searching for and choosing an evidence-based program (GTO steps 1 and 3, respectively). Instead, club leaders agreed to carry out a single evidence-based program (i.e., CHOICE) prior to the study. Use of a single evidence-based program better isolates the effects of GTO between study groups. For all other GTO steps, each site individually carried out the related practices. Given the similarity among many universal alcohol and drug prevention programs, we believe GTO received a strong test in PREP. Understanding the impact of program choice on implementation and outcomes is an important topic for future studies. Third, we were only able to study two program cycles of CHOICE. Future studies should examine whether continued cycles using GTO confers even greater benefits to programs than we were able to document here. Fourth, staff were aware of the study group they were in because each staff person learned about the study during the consent process. It is possible that those in the intervention group were additionally motivated to conduct CHOICE with fidelity because of their knowledge of their group membership. Finally, smaller effects may have gone undetected given the number of sites was 29; substantial for an RCT, but modest for testing site-level outcomes. Future rigorous studies are needed in which the impact of implementation support is assessed on the large scale used in federally or state-funded initiatives [89,90,91,92].
Community-based practitioners using GTO to carry out an evidence-based alcohol and drug prevention program demonstrated better fidelity than practitioners not using GTO after 2 years. Findings replicate the implementation results of a previous GTO study using the same design, but with a different, more challenging evidence-based program, content domain, and fidelity measures. Improved implementation did not translate into better individual youth outcomes, in large part because the youth had very low drug and alcohol use and generally positive proximal outcomes from the start. However, given typically poor evidence-based program implementation nationwide, these findings are significant as they highlight that GTO can improve the implementation of programs that use complex delivery methods and operate across multiple domains.
This project is registered at ClinicalTrials.gov with number NCT02135991. The trial was first registered on May 12, 2014.
Boys and Girls Club(s)
Enhancing Quality Interventions Promoting Healthy Sexuality
Getting To Outcomes
Preparing to Run Effective Prevention
Randomized controlled trial
Johnston LD, O’Malley PM, Miech RA, Bachman JG, Schulenberg JE. Monitoring the future national survey results on drug use, 1975–2015: overview, key findings on adolescent drug use. Ann Arbor, MI: Institute for Social Research, The University of Michigan; 2016.
U.S. Department of Health and Human Services. Facing Addiction in America. The Surgeon General’s Report on Alcohol, Drugs, and Health. Washington, DC: U.S. Department of Health and Human Services; 2016.
Green L. From research to ‘best practices’ in other settings and populations. Am J Health Behav. 2001;25(3):165–78.
Wandersman A, Florin P. Community interventions and effective prevention: bringing researchers/evaluators, funders and practitioners together for accountability. Am Psychol. 2003;58(6/7):441–8.
Ennett ST, Ringwalt CL, Thorne J, Rohrbach LA, Vincus A, Simons-Rudolph A, et al. A comparison of current practice in school-based substance abuse prevention programs with meta-analysis findings. Prev Sci. 2003;4:1–14.
Hallfors D, Godette D. Will the 'principles of effectiveness' improve prevention practice? Early findings from a diffusion study. Health Educ Res. 2002;17(4):461–70.
Kumar R, O'Malley PM, Johnston LD, Laetz VB. Alcohol, tobacco, and other drug use prevention programs in U.S. schools: a descriptive summary. Prev Sci. 2013;14(6):581–92. https://doi.org/10.1007/s11121-012-0340-z.
Ringwalt C, Hanley S, Vincus AA, Ennett ST, Rohrbach LA, Bowling JM. The prevalence of effective substance use prevention curricula in the nation's high schools. J Prim Prev. 2008;29(6):479–88. https://doi.org/10.1007/s10935-008-0158-4.
Ringwalt C, Vincus AA, Hanley S, Ennett ST, Bowling JM, Rohrbach LA. The prevalence of evidence-based drug use prevention curricula in U.S. middle schools in 2005. Prev Sci. 2009;10(1):33–40. https://doi.org/10.1007/s11121-008-0112-y.
Chinman M, Acosta J, Hunter S, Ebener P. Getting To Outcomes®: evidence of empowerment evaluation and evaluation capacity building at work. In: Fetterman DM, Kaftarian S, Wandersman A, editors. Empowerment evaluation: knowledge and tools for self-assessment, evaluation capacity building and accountability. 2nd ed. Thousand Oaks, CA: Sage; 2015. p. 316–36.
D'Amico EJ, Tucker JS, Miles JNV, Zhou AJ, Shih RA, Green HDJ. Preventing alcohol use with a voluntary after school program for middle school students: results from a cluster randomized controlled trial of CHOICE. Prev Sci. 2012;13:415–25.
Chinman M, Acosta J, Ebener P, Malone PS, Slaughter ME. Can implementation support help community-based settings better deliver evidence-based sexual health promotion programs? A randomized trial of Getting To Outcomes(R). Implement Sci. 2016;11(1):78.
Jemmott JB, Jemmott LS, Fong GT. Abstinence and safer sex HIV risk-reduction interventions for African American adolescents: a randomized controlled trial. J Am Med Assoc. 1998;279:1529–36.
Livet M, Wandersman A. Organizational functioning: facilitating effective interventions and increasing the odds of programming success. In: Fetterman DM, Wandersman A, editors. Empowerment evaluation in practice. New York: Guilford; 2005. p. 123–54.
Ajzen I, Fishbein M. Attitude-behavior relations: a theoretical analysis and review of empirical research. Psychol Bull. 1977;84:888–918.
Bandura A. Health promotion by social cognitive means. Health Educ Behav. 2004;31(2):143–64.
Fishbein M, Ajzen I. Attitudes toward objects as predictive of single and multiple behavioral criteria. Psychol Rev. 1974;81:59–74.
Fishbein M, Ajzen I. Belief, attitude, intention, and behavior: an introduction to theory and research. Reading, MA: Addison-Wesley; 1975.
Acosta J, Chinman M, Ebener P, Malone PS, Paddock S, Phillips A, et al. An intervention to improve program implementation: findings from a two-year cluster randomized trial of assets-Getting To Outcomes. Implement Sci. 2013;8:87.
Smelson DA, Chinman M, McCarthy S, Hannah G, Sawh L, Glickman M. A cluster randomized hybrid type III trial testing an implementation support strategy to facilitate the use of an evidence-based practice in VA homeless programs. Implement Sci. 2015;10:79. https://doi.org/10.1186/s13012-015-0267-4.
Durlak JA, DuPre EP. Implementation matters: a review of research on the influence of implementation on program outcomes and the factors affecting implementation. Am J Community Psychol. 2008;41:327–50.
Chinman M, Imm P, Wandersman A. Getting to Outcomes 2004: Promoting Accountability Through Methods and Tools for Planning, Implementation, and Evaluation. Santa Monica, CA: RAND Corporation; 2004. Contract No.: TR-101
Wandersman A, Imm P, Chinman M, Kaftarian S. Getting to outcomes: a results-based approach to accountability. Eval Program Plann. 2000;23:389–95.
Chinman M, Hunter SB, Ebener P, Paddock S, Stillman L, Imm P, et al. The Getting To Outcomes demonstration and evaluation: an illustration of the prevention support system. Am J Community Psychol. 2008;41:206–24.
Chinman M, Acosta JD, Ebener P, Sigel C, Keith J. Getting To Outcomes: a guide for teen pregnancy prevention. Santa Monica, CA: RAND Corporation; 2016.
Imm P, Chinman M, Wandersman A, Rosenbloom D, Guckenburg S, Leis R. Using the “Getting to Outcomes” Approach to Help Communities Prevent Underage Drinking. Santa Monica, CA: RAND Corporation; 2007. Contract No.: RB-9262
Fisher D, Imm P, Chinman M, Wandersman A. Getting To Outcomes with developmental assets: ten steps to measuring success in youth programs and communities Minneapolis: MN: Search Institute; 2006.
Chinman M, Tremain B, Imm P, Wandersman A. Strengthening prevention performance using technology: a formative evaluation of interactive Getting To Outcomes. Am J Orthopsychiatry. 2009;79:469–81 PMC2859836.
Chinman M, Acosta J, Ebener P, Malone PS, Slaughter M. A cluster-randomized trial of Getting To Outcomes’ impact on sexual health outcomes in community-based settings. Prev Sci. In Press.
Hawkins JD, Oesterle S, Brown EC, Arthur MW, Abbott RD, Fagan AA, et al. Results of a type 2 translational research trial to prevent adolescent drug use and delinquency: a test of communities that care. Arch Pediatr Adolesc Med. 2009;163(9):789–98. https://doi.org/10.1001/archpediatrics.2009.141.
Spoth R, Redmond C, Shin C, Greenberg M, Clair S, Feinberg M. Substance-use outcomes at 18 months past baseline: the PROSPER Community-University Partnership Trial. Am J Prev Med. 2007;32(5):395–402. https://doi.org/10.1016/j.amepre.2007.01.014.
Rohrbach LA, Gunning M, Sun P, Sussman S. The project towards no drug abuse (TND) dissemination trial: implementation fidelity and immediate outcomes. Prev Sci. 2010;11(1):77–88. https://doi.org/10.1007/s11121-009-0151-z.
Miller WR, Rollnick S. Motivational interviewing: preparing people for change. 2nd ed. New York: Guilford Press; 2002.
Rollnick S, Miller WR, Butler CC. Motivational interviewing in health care: helping patients change behavior. New York: Guilford Press; 2008.
D'Amico EJ, Ellickson PL, Wagner EF, Turrisi R, Fromme K, Ghosh-Dastidar B, et al. Developmental considerations for substance use interventions from middle school through college. Alcohol Clin Exp Res. 2005;29:474–83.
Chinman M, Acosta J, Ebener P, Burkhart Q, Clifford M, Corsello M, et al. Intervening with practitioners to improve the quality of prevention: one year findings from a randomized trial of assets-Getting To Outcomes. J Prim Prev. 2013;34:173–91.
Chinman M, Ebener P, Burkhart Q, Osilla KC, Imm P, Paddock SM, et al. Evaluating the impact of getting to outcomes-underage drinking on prevention capacity and alcohol merchant attitudes and selling behaviors. Prev Sci. 2014;15(4):485–96. https://doi.org/10.1007/s11121-013-0389-3.
Hawkins JD, Oesterle S, Brown EC, Arthur MW, Abbott RD, Fagan AA, et al. Results of a type 2 translational research trial to prevent adolescent drug use and delinquency: a test of communities that care. Arch Pediatr Adolesc Med. 2009;163(9):789–98.
Spoth R, Redmond C, Shin C, Greenberg M, Clair S, Feinberg M. Substance-use outcomes at 18 months past baseline: the PROSPER Community-University Partnership Trial. Am J Prev Med. 2007;32(5):395–402.
Chinman M, Acosta J, Ebener P, Malone PS, Slaughter ME. Can implementation support help community-based settings better deliver evidence-based sexual health promotion programs? A randomized trial of Getting To Outcomes(R). Implement Sci. 2016;11(1):78. https://doi.org/10.1186/s13012-016-0446-y.
Chinman M, Acosta J, Ebener P, Malone PS, Slaughter ME. A cluster-randomized trial of Getting To Outcomes’ impact on sexual health outcomes in community-based settings. Prev Sci. 2018;19(4):437–48. https://doi.org/10.1007/s11121-017-0845-6.
D'Amico EJ, Green HDJ, Miles JNV, Zhou AJ, Tucker JA, Shih RA. Voluntary after school alcohol and drug programs: if you build it right, they will come. J Res Adolesc. 2012;22(3):571–82.
Aarons GA, Glisson C, Hoagwood K, Kelleher K, Landsverk J, Cafri G. Psychometric properties and U.S. national norms of the evidence-based practice attitude scale (EBPAS). Psychol Assess. 2010;22(2):356–65. https://doi.org/10.1037/a0019188.
McDonald RP. Test theory: a unified treatment. Mahwah: Lawrence Erlbaum; 1999.
Dunn TJ, Baguley T, Brunsden V. From alpha to omega: a practical solution to the pervasive problem of internal consistency estimation. Br J Psychol. 2014;105(3):399–412.
Krippendorff K. Content analysis: an introduction to its methodology. 2nd ed. Thousand Oaks, CA: Sage; 2004.
Bandura A. Self-efficacy: toward a unifying theory of behavioral change. Psychol Rev. 1977;84:191–215.
Kahneman D, Tversky A. Choices, values, and frames. New York: Cambridge University Press; 2000.
Bandura A. Self-efficacy: the exercise of control. New York: Freeman; 1997.
Miller WR, Rollnick S. Motivational interviewing: helping people change. 3rd ed. New York: Guilford Press; 2012.
D'Amico EJ, Edelen M. Pilot test of Project CHOICE: a voluntary after school intervention for middle school youth. Psychol Addict Behav. 2007;21(4):592–8.
Dusenbury L, Brannigan R, Hansen W, Walsh J, Falco M. Quality of implementation: developing measures crucial to understanding the diffusion of preventive interventions. Health Educ Behav. 2005;20:308–13.
Moyers TB, Martin T, Manuel JK, Miller WR, Ernst D. Revised Global Scales: Motivational Interviewing Treatment Integrity 3.1.1 (MITI 3.1.1). Albuquerque: NM University of New Mexico; 2010.
Madson MB, Campbell TC. Measures of fidelity in motivational enhancement: a systematic review. J Subst Abus Treat. 2006;31(1):67–73. https://doi.org/10.1016/j.jsat.2006.03.010.
Moyers TB, Martin T, Manuel JK, Hendrickson SM, Miller WR. Assessing competence in the use of motivational interviewing. J Subst Abus Treat. 2005;28(1):19–26. https://doi.org/10.1016/j.jsat.2004.11.001.
Cambell M, Carr C, DeVellis B, Switzer B, Biddle A, Amamoo MA, et al. A randomized trial of tailoring and motivational interviewing to promote fruit and vegetable consumption for cancer prevention and control. Ann Behav Med. 2009;38(2):71–85.
Martino S, Ball SA, Nich C, Frankforter TL, Carroll KM. Community program therapist adherence and competence in motivational enhancement therapy. Drug Alcohol Depend. 2008;96(1):37–48.
Parsons JT, Golub SA, Rosof E, Holder C. Motivational interviewing and cognitive-behavioral intervention to improve HIV medication adherence among hazardous drinkers: a randomized controlled trial. J Accuir Immune Defic. 2007;46(4):443.
Turrisi R, Larimer ME, Mallett KA, Kilmer JR, Ray AE, Mastroleo NR, et al. A randomized clinical trial evaluating a combined alcohol intervention for high-risk college students. J Stud Alcohol Drugs. 2009;70(4):555.
McCambridge J, Day M, Thomas BA, Strang J. Fidelity to motivational interviewing and subsequent cannabis cessation among adolescents. Addict Behav. 2011;36(7):749–54.
Pollak KI, Coffman CJ, Alexander SC, Ostbye T, Lyna P, Tulsky JA, et al. Predictors of weight-related communication during primary care visits with overweight adolescents. Patient Educ Couns. 2014;96(3):327–32.
Woodin EM, Sotskova A, O'Leary KD. Do motivational interviewing behaviors predict reductions in partner aggression for men and women? Behav Res Ther. 2012;50:79–84.
Cicchetti DV. Guidelines, criteria, and rules of thumb for evaluating normed and standardized assessment instruments in psychology. Psychol Assess. 1994;94:284–90.
Johnston LD, O'Malley PM, Miech RA, Bachman JG, Schulenberg JE. Monitoring the future national survey results on drug use, 1975–2016: overview, key findings on adolescent drug use. Ann Arbor: Institute for Social Research, The University of Michigan; 2017.
D'Amico EJ, Hunter SB, Miles JN, Ewing BA, Osilla KC. A randomized controlled trial of a group motivational interviewing intervention for adolescents with a first time alcohol or drug offense. J Subst Abus Treat. 2013;45(5):400–8. https://doi.org/10.1016/j.jsat.2013.06.005.
D'Amico EJ, Miles JN, Stern SA, Meredith LS. Brief motivational interviewing for teens at risk of substance use consequences: a randomized pilot study in a primary care clinic. J Subst Abus Treat. 2008;35(1):53–61. https://doi.org/10.1016/j.jsat.2007.08.008.
D'Amico EJ, Parast L, Meredith LS, Ewing BA, Shadel WG, Stein BD. Screening in Primary Care: What Is the Best Way to Identify At-Risk Youth for Substance Use? Pediatrics. 2016;138(6). https://doi.org/10.1542/peds.2016-1717.
D'Amico EJ, Miles JN, Tucker JS. Gateway to curiosity: medical marijuana ads and intention and use during middle school. Psychol Addict Behav. 2015;29(3):613–9. https://doi.org/10.1037/adb0000094.
D'Amico EJ, Parast L, Shadel WG, Meredith LS, Seelam R, Stein BD. Brief motivational interviewing intervention to reduce alcohol and marijuana use for at-risk adolescents in primary care: a randomized clinical trial. J Consult Clin Psychol. In Press.
D’Amico EJ, McCarthy DM. Escalation and initiation of younger adolescents’ substance use: the impact of perceived peer use. J Adolesc Health. 2006;39:481–7.
D'Amico EJ, Miles JNV, Stern SA, Meredith LS. Brief motivational interviewing for teens at risk of substance use consequences: a randomized pilot study in a primary care clinic. J Subst Abus Treat. 2008;35:53–61.
Ellickson PL, Bell RM. Drug prevention in junior high: a multi-site longitudinal test. Science. 1990;247:1265–372.
Lakens D. Calculate Partial Eta Omega Epsilon Squared.xlsx. 2013. https://osf.io/kzsyg/ Accessed 15 Apr 2016.
Lakens D. Calculating_Effect_Sizes.xlsx (v.4). 2013. https://osf.io/vbdah/. Accessed 22 Nov 2017.
Hess MR, Kromrey JD. Effect_CI: a SAS® macro for constructing confidence intervals around standardized mean differences. 2003. http://analytics.ncsu.edu/sesug/2003/SD04-Hess.pdf. Accessed 16 Apr 2016.
Benjamini Y, Hochberg B. Controlling the false discovery rate: a practical and powerful approach to multiple testing. J R Stat Soc. 1995;57(1):289–300.
Fagan AA, Hanson K, Hawkins JD, Arthur MW. Bridging science to practice: achieving prevention program implementation fidelity in the community youth development study. Am J Community Psychol. 2008;41:235–49.
Rohrbach LA, Gunning M, Sun P, Sussman S. The project towards no drug abuse (TND) dissemination trial: implementation fidelity and immediate outcomes. Prev Sci. 2010;11:77–88.
Spoth R, Guyll M, Lillehoj CJ, Redmond C. PROSPER study of evidence-based intervention implementation quality by community–university partnerships. J Community Psychol. 2007;35(8):981–99.
D'Amico EJ, Houck JM, Hunter SB, Miles JN, Osilla KC, Ewing BA. Group motivational interviewing for adolescents: change talk and alcohol and marijuana outcomes. J Consult Clin Psychol. 2015;83(1):68–80. https://doi.org/10.1037/a0038155.
Mueller T, Tevendale HD, Fuller TR, House LD, Romero LM, Brittain A, et al. Teen pregnancy prevention: implementation of a multicomponent Community-Wide Approach. J Adolesc Health. 2017;60(3S):S9–S17. https://doi.org/10.1016/j.jadohealth.2016.11.002.
Ellickson PL, Martino SC, Collins RL. Marijuana use from adolescence to young adulthood: multiple developmental trajectories and their associated outcomes. Health Psychol. 2004;23(3):299–307. https://doi.org/10.1037/0278-618.104.22.1689.
Jackson KM, Sher KJ, Schulenberg JE. Conjoint developmental trajectories of young adult substance use. Alcohol Clin Exp Res. 2008;32(5):723–37. https://doi.org/10.1111/j.1530-0277.2008.00643.x.
Schulenberg JE, Merline AC, Johnston LD, O'Malley PM, Bachman JG, Laetz VB. Trajectories of marijuana use during the transition to adulthood: the big picture based on National Panel Data. J Drug Issues. 2005;35(2):255–79.
Setoh P, Qin L, Zhang X, Pomerantz EM. The social self in early adolescence: two longitudinal investigations in the United States and China. Dev Psychol. 2015;51(7):949–61. https://doi.org/10.1037/a0039354.
Tucker JS, Ellickson PL, Orlando M, Klein DJ. Cigarette smoking from adolescence to young adulthood: women's developmental trajectories and associates outcomes. Womens Health Issues. 2006;16(1):30–7. https://doi.org/10.1016/j.whi.2005.08.004.
Tucker JS, Orlando M, Ellickson PL. Patterns and correlates of binge drinking trajectories from early adolescence to young adulthood. Health Psychol. 2003;22(1):79–87.
Vega WA, Chen KW, Williams J. Smoking, drugs, and other behavioral health problems among multiethnic adolescents in the NHSDA. Addict Behav. 2007;32(9):1949–56. https://doi.org/10.1016/j.addbeh.2006.12.022.
Chervin DD, Philliber S, Brindis CD, Chadwick AE, Revels ML, Kamin SL, et al. Community capacity building in CDC’s community coalition partnership programs for the prevention of teen pregnancy. J Adolesc Health. 2005;37:S11–S9.
Lesesne CA, Lewis KM, White CP, White DC, Green DC, Duffy JL, et al. Promoting science-based approaches to teen pregnancy prevention: proactively engaging the three systems of the Interactive Systems Framework. Am J Community Psychol. 2008;41:379–92.
Philliber S, Nolte K. Implementation science: promoting science-based approaches to prevent teen pregnancy. Prev Sci. 2008;9(3):166–77.
Zief S, Shapiro R, Strong D. How states are implementing evidence-based teen pregnancy prevention programs through the personal responsibility education program. Washington, DC: Office of Planning, Research and Evaluation, Administration for Children and Families; 2014.
All the authors are funded by a grant from the National Institute on Alcohol Abuse and Alcoholism: Preparing to Run Effective Prevention (R01AA022353-01). We would like to acknowledge the contributions of Marylou Gilbert and Christian Lopez for their assistance with this project.
All the authors are funded by a grant from the National Institute on Alcohol Abuse and Alcoholism: Preparing to Run Effective Prevention (R01AA022353-01).
Availability of data and materials
De-identified datasets, used and/or analyzed during the current study, are available from the corresponding author on reasonable request.
Ethics approval and consent to participate
This study was approved by RAND’s Human Subjects Protection Committee (Federal Assurance Number: FWA00003425), which reviews and monitors all non-exempt research to ensure protection of subjects from research risks, pursuant to federal regulations. All subjects in this research study provided consent to participate.
Consent for publication
This paper does not contain any individually identifiable images or any other data. Thus, consent for publication was not obtained.
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Chinman, M., Ebener, P., Malone, P.S. et al. Testing implementation support for evidence-based programs in community settings: a replication cluster-randomized trial of Getting To Outcomes®. Implementation Sci 13, 131 (2018). https://doi.org/10.1186/s13012-018-0825-7
- Implementation support
- Evidence-based prevention