- Open Access
Lessons learned about the effective operationalization of champions as an implementation strategy: results from a qualitative process evaluation of a pragmatic trial
Implementation Science volume 15, Article number: 87 (2020)
Though the knowledge base on implementation strategies is growing, much remains unknown about how to most effectively operationalize these strategies in diverse contexts. For example, while evidence shows that champions can effectively support implementation efforts in some circumstances, little has been reported on how to operationalize this role optimally in different settings, or on the specific pathways through which champions enact change.
This is a secondary analysis of data from a pragmatic trial comparing implementation strategies supporting the adoption of guideline-concordant cardioprotective prescribing in community health centers in the USA. Quantitative data came from the community health centers’ shared electronic health record; qualitative data sources included community health center staff interviews over 3 years. Using a convergent mixed-methods design, data were collected concurrently and merged for interpretation to identify factors associated with improved outcomes. Qualitative analysis was guided by the constant comparative method. As results from the quantitative and initial qualitative analyses indicated the essential role that champions played in promoting guideline-concordant prescribing, we conducted multiple immersion-crystallization cycles to better understand this finding.
Five community health centers demonstrated statistically significant increases in guideline-concordant cardioprotective prescribing. A combination of factors appeared key to their successful practice change: (1) A clinician champion who demonstrated a sustained commitment to implementation activities and exhibited engagement, influence, credibility, and capacity; and (2) organizational support for the intervention. In contrast, the seven community health centers that did not show improved outcomes lacked a champion with the necessary characteristics, and/or organizational support. Case studies illustrate the diverse, context-specific pathways that enabled or prevented study implementers from advancing practice change.
This analysis confirms the important role of champions in implementation efforts and offers insight into the context-specific mechanisms through which champions enact practice change. The results also highlight the potential impact of misaligned implementation support and key modifiable barriers and facilitators on implementation outcomes. Here, unexamined assumptions and a lack of evidence-based guidance on how best to identify and prepare effective champions led to implementation support that failed to address important barriers to intervention success.
Implementation strategies are techniques designed to promote the uptake of clinical interventions into practice  by targeting key modifiable barriers and facilitators (called implementation determinants) to adoption of these interventions [2, 3]. Implementation science seeks to develop evidence to improve the accurate identification of (i) implementation barriers relevant to specific interventions [4,5,6], and (ii) implementation strategies likely to address such barriers [5, 7, 8]. While the evidence on effective implementation strategies is growing, much remains unknown about how best to select and enact these strategies for a given context. Currently, such planning usually focuses on which implementation strategies to use, typically informed by knowledge of/hypotheses about likely implementation barriers [1, 3, 9,10,11,12,13,14,15], existing evidence about the effectiveness of the strategies [1, 3, 9, 13,14,15,16], the experience of the implementation team , guidance from conceptual models [6, 18,19,20,21,22,23,24,25], and pragmatic concerns [1, 3, 9, 15] including available resources [4, 5, 13, 26]. However, these sources generally do not provide detailed guidance on how these strategies should be operationalized to maximize their impact.
The Study of Practices Enabling Implementation and Adaptation in the Safety Net (SPREAD-NET) involved a set of implementation strategies, selected based on evidence demonstrating their ability to support practice change, their potential scalability, and pragmatic considerations (e.g., cost). These strategies included (i) creating and distributing educational materials, (ii) conducting educational meetings and ongoing trainings, (iii) using train-the-trainer strategies, (iv) facilitation, and (v) identifying and preparing champions . The study hypothesized that more intensive implementation support (e.g., more trainings; adding practice facilitation) would lead to greater improvements in the targeted outcomes. As such, the study was designed to test the impact of additive support, not the efficacy of any individual implementation strategy. The main SPREAD-NET analyses—of association between study outcomes and degree of implementation support—found that more intensive implementation support was not associated with greater improvements in the targeted outcomes (rates of guideline-concordant cardioprotective prescribing) .
For the manuscript presented here, we re-analyzed the study’s mixed-methods data to identify factors associated with differences in study outcomes in individual community health centers. As initial results indicated that the interplay between study implementers (clinic staff tasked with leading change processes) and organizational context was key to the success of a given community health center, we then explored the organization-specific pathways through which these implementers promoted practice change. We note that the use of champions [29,30,31,32] was one of the parent study’s implementation strategies chosen a priori; the intention was that the study implementers would act as champions. In this paper, however, and per implementation science definitions [18, 27, 31], we use the term “champion” to describe only those study implementers who demonstrated a sustained commitment to implementation activities. In other words, all clinic staff tasked with leading change processes were study implementers, but only some acted as champions.
Little has been reported on how to effectively prepare people for the champion role, or on the specific pathways through which champions enact change [31, 33, 34]. This paper presents a detailed assessment of how study implementers/champions impacted prescribing behavior in varied settings, to better understand how this role is optimally operationalized, and thus advance the specification and preparation of champions as an implementation strategy.
Study setting and design
The SPREAD-NET study compared the effectiveness of increasingly intensive implementation strategies at supporting community health centers’ adoption of a suite of clinical decision support tools called the cardiovascular disease (CVD) bundle. The CVD bundle included point-of-care alerts and panel management data tools promoting cardioprotective prescribing guidelines for patients with diabetes . It was activated in the study clinics’ shared EHR in June 2015, and modified in May 2017, when CVD risk calculation  was added to the logic underlying its alerts.
Twenty-nine clinics managed by 12 community health centers in six states participated in the study (a single community health center may encompass multiple clinics). All study community health centers were members of OCHIN, Inc., a non-profit organization based in Portland, OR, that provides a shared Epic© EHR to > 600 US clinics. The community health centers were cluster-randomized into one of three arms; the arms received increasingly intensive implementation support designed to promote uptake of the CVD bundle (Table 1) (we recognize that clinical decision support can be considered an implementation strategy ; here, the cVD bundle is the innovation targeted by he strategies under comparison).
All study community health centers were required to identify a staff member to engage with the study team and lead the clinic’s efforts to support the use of the CVD bundle—a study implementer. Based on known attributes of effective champions [30, 31, 36,37,38], and our past experience , we suggested (but did not define or require) selecting implementers with involvement in and enthusiasm about quality improvement activities, credibility, and influence at the clinic, and interest in diabetes/cardiovascular care. One staff member could be the study implementer for up to three clinics in a single community health center. Each implementer was expected to participate in regular telephone check-ins with the study team, and those in arms 2 and 3 to attend the in-person training. Each community health center’s leadership chose study implementers based on these parameters and their own needs; some implementers volunteered for the role while others were assigned. Once selected, the research team explained the study implementers role to them as (i) acting as a resource/champion for any SPREAD-NET-related activities their clinic chose to implement (what/when/how up to each clinic), and (ii) acting as a liaison between the community health center and the study team. If the primary study implementer was not a clinician, the community health centers were asked to identify a clinician to serve as a role model/resource; that person was also invited to the in-person training. Of the 11 participating community health centers, five chose physicians as the primary study implementer. Of the remaining six community health centers, three chose physicians as the additional clinical implementation resource, one chose a nurse practitioner, and two opted not to fill the role.
The SPREAD-NET study followed a convergent mixed-methods design  in which quantitative and qualitative data were collected concurrently and merged for interpretation. As noted in the “Introduction/background” section, the study’s primary hypothesis—that implementation support provided with increasing intensity by study arm would be associated with additive improvements in the study outcomes—was not realized . The analyses presented here used the same data sources as in the main study analysis to investigate factors associated with significantly improved outcomes in individual study community health centers (rather than across study arms). Data sources are summarized below.
Quantitative data were extracted from OCHIN’s EHR database. These data were used to measure the proportion of patients in a given community health center who had diabetes mellitus (DM), were indicated for a statin per care guidelines (denominator), and had a prescription for a statin (numerator), calculated monthly.
The 3-year qualitative process evaluation was conducted from August 2015, through July 2018. Figure 1 illustrates the intersection of the study timeline and process evaluation data collection points.
The evaluation design was informed by the practice change model , a conceptual model that seeks to identify the elements and interrelationships critical to practice change, and followed a grounded theory approach [42, 43] in which we attempted to understand the ongoing implementation experience from the perspective of study clinic staff. Our team’s qualitative researchers called the primary implementers twice a month in the 6 months after the start of the intervention in July 2015; conversations were loosely based on a question guide but were designed to be flexible and follow the implementers’ lead. We subsequently reduced call frequency to monthly, then quarterly, to reduce the participation burden. Four hundred and thirteen calls were conducted over the 3-year evaluation. We also conducted 2-day site visits with eight of the 12 community health centers, at 9 to 22 months post-intervention start. Sites were purposively chosen to maximize variation in implementation processes. During these visits, we interviewed clinic leadership, clinicians, and staff about individual- and clinic-level prescribing practices, use of the CVD bundle, barriers and facilitators to its adoption, and organizational approaches to practice standardization. The number of interviews per site ranged from 7 to 15, determined by clinic size and staff availability. We conducted three debrief sessions with the study practice facilitator (a member of the research team who provided ongoing practice facilitation to arm 3 clinics; Table 1) to learn about her experiences at the arm 3 clinics. Shortly after the intervention’s end, we conducted additional phone interviews with five clinic leaders to gain additional perspectives on the implementation process. All calls and interviews were recorded and professionally transcribed for analysis.
The secondary analysis of the SPREAD-NET data presented here occurred in two rounds. We first used the coding structure created as part of the parent study to conduct a preliminary exploration of the qualitative data by community health center  to identify factors that distinguished between individual community health centers reporting more or less robust implementation activities. Robustness was characterized by the time and effort that a given organization put into implementation activities, and its willingness to adapt its approach as necessary; these factors were determined during the check-in calls with study implementers and interviews of clinic staff, and informed by on-site observation during site visits. Concurrently, quantitative analysis was performed to identify which community health centers (if any) demonstrated statistically significant improvements in guideline-concordant statin prescribing. Next, we compared those quantitative and qualitative results, then re-immersed ourselves in the qualitative data to better understand observed associations between improved prescribing and contextual/implementation variables. Details on the quantitative and qualitative analysis methods are provided below.
As with the main study analysis , we used a difference-in-difference (DiD)  approach to evaluate pre-/post-change in prescribing rate(s). For this analysis, we evaluated the change in statin prescribing rates by community health center (rather than study arm). All analyses reflect tests of statistical significance with a two-sided α of 0.05 and were conducted using SAS Enterprise Guide 7.15.
Coding was conducted iteratively throughout the parent study. A preliminary code list was created by the lead qualitative researcher (AB) 1 year into data collection. This draft list was reviewed and refined by the three-person qualitative team into a codebook; each team member then coded an identical set of transcripts and compared the results to identify items needing revision. After agreed-upon codes were applied consistently across all coders, independent coding assignments were made. The team double-coded five percent of the qualitative data as a quality check, resolving inconsistencies through team discussion and consensus and updating the codebook as necessary. Coding was conducted using the QSR NVivo software, guided by the constant comparative method [45, 46]. Coding was completed concurrently, but blinded to, quantitative analysis testing study hypotheses.
When the main study analyses found that increasingly intensive implementation support did not yield additive benefits , we designed and conducted the secondary analysis presented here to assess factors that might explain differences by community health center irrespective of amount of support received. We first identified factors potentially key to adoption success, based on knowledge of the clinics, data from two researchers involved in data collection throughout the study (AB, JD), and review of all qualitative data by a researcher who joined the study post-data collection (IG). Factors initially considered included community health center staff’s EHR proficiency/skills; community health center culture of quality improvement/practice change; competing quality improvement initiatives; organizational approach to standardization and emphasis on evidence-based care; staff/study implementer turnover; community health center-specific implementation approach; and study implementer characteristics and level of engagement. We then created community health center-specific reviews for each relevant code (e.g., implementation strategies and barriers; EHR optimization; point person; leadership and staff buy-in; standardization; practice change; context setting, culture, churn), and summarized this data by community health center. The 3-year data collection period allowed us to observe changes in implementation activities, contextual factors that impacted the implementation and implementer perspectives over time, all of which informed and enriched our understanding of the process of implementation at each site.
Based on this analysis, we identified a set of interconnected factors related to characteristics of the study implementer and the organizational context that differentiated those community health centers that did and did not demonstrate a significant increase in guideline-concordant statin prescribing in the quantitative results. We then conducted multiple cycles of immersion crystallization  in which qualitative team members read and re-read the transcripts and code reports, emerged to reflect on and discuss overarching themes and patterns, then delved back into the data to explore connections, confirm/disconfirm insights and interpretations, and flesh out our understanding of the essential attributes of study implementers and the mechanisms through which they successfully influenced prescribing practice at their organization.
The standards for reporting qualitative research (SRQR) was used to guide the reporting of qualitative findings (see Additional file) . This study was approved by the Kaiser Permanente Northwest Institutional Review Board.
Five community health centers demonstrated statistically significant increases in guideline-concordant statin prescribing over the course of the study (Table 2).
The intersection of organizational investment and study implementer
Several factors, in combination, were associated with significant pre-post increases in guideline-concordant statin prescribing among the community health centers noted in Table 2. All were related to the characteristics of the study implementers and the organizational support they received. They included the following:
Engagement: Interest in and willingness to promote the intervention
Influence: Sufficient social capital to foster trust and the authority to prioritize implementation and stimulate practice change
Credibility: Conferred through prescribing privileges
Capacity: Time—and understanding of diabetes and cardiovascular care—sufficient to effectively advocate for the intervention
At each community health center that demonstrated a significant improvement in study outcomes, one or two implementers emerged as the de facto champions (see case studies, below). As noted in the “Introduction/background” section, we use the term champion to describe only those study implementers who demonstrated a sustained commitment to implementation activities [18, 27, 31] (Table 3). Study implementers who emerged as champions demonstrated each of the above elements (though at one community health center, the implementers were clinical pharmacists whose role included advising providers on prescribing decisions).
We defined organizational support as the creation of an environment within which implementation activities could be expected to be taken seriously by clinic staff. In many cases, this support was indicated by the selection of a staff member with the potential to be an effective champion (as described above; many of the necessary qualities were suggested by the study team in the initial study communications). Community health centers that demonstrated organizational support also promoted the use of the CVD Bundle and/or guideline-concordant statin prescribing. While organizational support, or the lack thereof, could take many forms—as illustrated below—implementation success depended on both the presence of champions with the aforementioned attributes and the implicit or explicit backing of clinic leadership, and the interaction of the two.
Community health centers that did not show a significant improvement in prescribing rates lacked either an emergent champion, and/or organizational support. Implementers at these community health centers often were less engaged with the study/CVD bundle; did not have adequate influence at their clinic to be effective change agents for this intervention; and/or did not have a clinical background (and therefore the credibility to affect provider care decisions). Two community health centers (see Table 3) did have champions that met each of the elements noted above apart from prescribing privileges (both were RNs), but implementation approaches and organizational priorities precluded their ability to effectively promote the targeted practice change.
While the combination of the above implementer characteristics—engagement, influence, credibility, and capacity—and organizational support appeared key to enabling practice change, each community health center—champion combination followed unique implementation paths suited to the particular context. Case studies of the seven community health centers with emergent champions, five of whom achieved significant improvements and two who did not, illuminate this diversity.
Community health centers that significantly increased guideline-concordant statin prescribing for patients with diabetes
Community health center #1
Implementation efforts were led by a medical director [credibility], who had advocated for the organization’s participation in the study, and were effective primarily due to her efforts, position, and influence within the organization: she had a long history at the community health center and was respected and trusted by providers and staff [influence]. Despite practicing exclusively at the clinic where she served as medical director, she was instrumental in promoting the adoption of the targeted care guidelines at both clinics in the study. She was actively engaged in the implementation, conducting multiple trainings for providers and discussing the intervention, relevant guidelines, and the CVD bundle at organization-wide meetings [engagement, capacity]. She also led by example, consistently using the CVD bundle tools and developing relevant plan-do-study-act tests of change [engagement]. She worked closely with and directed the activity of an RN-turned EHR technical support staff member, who incorporated the CVD bundle into new staff trainings.
Community health center #5
Implementation efforts were directed by the organization’s medical director, who had made the decision to participate in the study, and supported by EHR technical staff at each clinic who were assigned to the implementer role by the medical director. This community health center achieved a significant increase in guideline-concordant statin prescribing through the sustained, active engagement of a leadership figure within an organizational structure that prioritized hierarchy and standardization [influence]. This medical director, a practicing physician [credibility], was actively involved in developing reports that identified patients in a given providers’ panel who were indicated for but not on the target medications. These reports were regularly distributed to providers, and the medical director met with providers individually to review them [engagement, capacity]. She also incorporated statin guideline prescribing performance into provider evaluations [engagement]. She standardized EHR screen configurations organization-wide to maximize viewing the point-of-care alerts, and had most non-CVD bundle alerts turned off to minimize alert fatigue [engagement]. She modified the statin dosing table provided in the SPREAD-NET toolkit to include locations where statins were available at low cost, and taped it to providers’ monitors for easy access [engagement, capacity]. Under her direction, links to a CVD risk calculator were added to the organization’s EHR before it became available as part of the CVD bundle [engagement]. She integrated explicit cardioprotective prescribing guidance into the community health center’s residents’ training [engagement, capacity]. Under her guidance and supported by the EHR technical support staff, each clinic conducted outreach to patients who were (over)due for diabetes care; the intensity of this effort varied by clinic and care team.
Community health center #6
Implementation efforts were enacted by an experienced, well-regarded physician [credibility] who had practiced at one of the clinics in this organization for many years and had previously served as medical director for the community health center [influence]. This provider was interested in research and volunteered to fill the study implementer role; her schedule already included 4 h a week of administrative time designated to community and research work [capacity]. This gave her the flexibility to innovate in her own practice and effect change through one-on-one engagement with other providers, including sharing alternative workflows for diabetes care, within an organizational culture that afforded substantial provider autonomy. In addition to experimenting with different workflows for diabetes care, this physician also met individually with all providers at both clinics, including during new provider onboarding, for a short, over-the-shoulder introduction of the study and the CVD bundle, and demonstration of the tools [engagement].
Community health center #8
Early implementation activities were limited to an introductory presentation to providers and occasional reminders during staff meetings. This community health center replaced their study implementer approximately 1 year into the study, shifting the role to a provider who was newer to the organization but brought a wealth of experience in diabetes care [credibility]; both providers were appointed to the implementer role by organizational leadership. The community health center’s increase in guideline-concordant statin prescribing appeared to be due to a combination of awareness-raising efforts by this clinician, coupled with an increasing organization-wide emphasis on standardization and quality improvement throughout the study period [influence]. During this time, the organization was working on becoming an accredited patient-centered medical home, with an attendant focus on practice change capacity. The new implementer brought renewed focus to the statin guidelines, often discussing prescribing recommendations at organization-wide provider meetings [engagement, capacity]. She also requested and received the addition of a link to a CVD risk calculator on staff computers, and notified providers when it was added to the CVD bundle [engagement, capacity].
Community health center #12
Implementation efforts were led by two clinical pharmacists, one of whom was the clinic’s director of clinical pharmacy. The pharmacy director chose to lead implementation activities at the main clinic; the second clinical pharmacist was asked to take on the role at the second clinic. Success at this site was largely due to the strength and credibility of the clinical pharmacy program at the organization, the dedicated time the pharmacists were able to spend supporting guideline-concordant statin prescribing [capacity], and parallel awareness-raising by clinical leaders that resulted in an increased emphasis on relevant prescribing guidelines by multiple influential actors at the organization. The clinical pharmacy department was well-resourced, trusted [influence], and guideline-focused. The clinical pharmacists incorporated updated statin guidelines into pharmacy protocols, met individually with patients to review their medications, reviewed patient charts for guideline-concordant prescribing [credibility], and followed up with providers in conversations or messages, as well as discussing the CVD bundle and relevant guidelines at clinic meetings [engagement]. The clinical pharmacists worked outside the daily patient encounter workflow, which challenged their ability to stimulate change at the provider and team level. However, the organization’s medical director and the chief operating officer, who had formerly worked as an RN at the organization, ensured that the statin guidelines were discussed and debated at provider meetings.
Community health centers that did not significantly increase guideline-concordant statin prescribing for patients with diabetes
Community health center #10
Implementation efforts were coordinated by an RN who served as the director of quality improvement for the community health center; she volunteered for the role and acted as the primary study implementer for the largest of the four clinics taking part in the study. The remaining three study implementers, all of whom held patient-facing administrative roles, were asked to manage implementation activities at the other three clinics. The four study implementers initially met multiple times to plan their implementation approach, introduced the CVD bundle to care teams at all four clinics, and identified and worked with a single provider to initiate a pilot project using the CVD bundle within his own patient panel. These efforts, however, never gained momentum. Organizational support was diffused by other activities occurring at the community health center during the study period, including simultaneous participation in other quality initiatives, the opening of a new clinic, and major construction projects and upgrades across the organization. Although the director of quality improvement did demonstrate engagement, influence, and capacity, she did not see patients herself and as an RN would have been unable to prescribe statins [lack of credibility]. The three other study implementers did not have clinical backgrounds and lacked the influence, credibility, and capacity (particularly a deep understanding of diabetes and cardiovascular care) to persuade providers to change their behavior; the additional clinician implementer, a practicing physician, did not play an active role in supporting implementation activities.
Community health center #3
Implementation efforts at this community health center, a single clinic, were enacted by an RN diabetes care manager who was assigned to the study implementer position. She conducted a few initial presentations about the CVD bundle at provider meetings and occasionally discussed the study’s EHR tools with individual providers. However, as this RN personally met with most of the clinic’s patients with diabetes in her role as a diabetes care manager, most of her implementation activities focused on her own activities. She integrated review of statin prescriptions into her pre-visit chart review process and follow-up visits with diabetic patients, and followed up with the primary care provider to discuss a plan of care based on clinical recommendations. The effectiveness of this approach was limited because (i) the EHR tools were designed to fire during open encounters, not during pre-visit chart reviews; and (ii) the approach relied almost entirely on a single person without prescribing privileges [lack of credibility], with little effort put into raising awareness or buy-in from providers. In addition, although this study implementer actively worked to identify patients with diabetes who could benefit from a statin [engagement] and had the influence and capacity to advocate for the intervention, she occasionally expressed distrust in the guidelines underlying the CVD bundle and/or the algorithm behind the EHR tools. Finally, although this clinic had a standardized process for piloting and approving new interventions and workflows, this process was not applied to the clinic’s participation in SPREAD-NET—which appeared to limit organizational awareness of and support for implementation activities.
Our study adds to implementation science by providing insight into the pathways through which champions may impact implementation outcomes, and advances understanding of how to identify and prepare implementers to be effective champions within their own particular environments. Notably, unlike the community health centers that did not demonstrate a significant change in prescribing behavior, the five community health centers that did have improved outcomes all had engaged and respected champions (study implementers) who were able to directly influence provider behaviors or alter institutional prescribing norms. This suggests that not only the selection of champions as an implementation strategy but also the appropriate operationalization of support (i.e., the identification and preparation of champions), are necessary for effective practice change; this is likely to also apply to other implementation strategies, and further research is needed to identify best practices for doing so. These findings also align with those of other studies that showed the potential for champions to successfully support introducing and maintaining practice change [30,31,32,33, 49,50,51,52].
Implementation science emphasizes that identifying effective implementation strategies involves understanding the context-specific causal pathways through which these strategies can have impact [3, 34]. The case studies presented here underscore this: effective champions were key to implementation success, but individual differences between study implementers and contextual differences between organizations produced different pathways to change. In some cases, change was effected through hierarchical directives and practice standardization; in others, the champion relied on trust-based relationships, advocacy, and leading by example. This underscores the importance of assessing and reporting how implementation strategies operate in a given setting.
These findings also help explain the overall study results (higher-intensity implementation support was not associated with better outcomes). A recent review  found that variation in impact across implementation studies is often due to misalignment between implementation strategies and key contextual barriers and facilitators. A similar phenomenon occurred here: the selected strategy (champions) was appropriate, but as operationalized it had little impact. It was recommended that study clinics appoint study implementers with enthusiasm, credibility, influence, and clinical knowledge, but—in an effort to allow for organizational autonomy—these elements were not required. In addition, the study’s implementation support focused on adoption of the targeted innovation, rather than on increasing the study implementers’ effectiveness. Had the focus been on supporting the development of the implementers’ leadership skills and engagement with the intervention and/or had it been required that study implementers fulfill certain criteria, study results may have been different.
Recognizing this led us to question the process through which we initially selected and operationalized the implementation strategies compared in the SPREAD-NET study. After a period of reflection  and extended team discussions, we recognized that a given community health center’s decision to participate in SPREAD-NET had triggered collective, unacknowledged assumptions by our study team regarding organizational support for the targeted change, and study implementer engagement in/capacity to effect that change. We assumed that the implementers designated by the community health centers would have the necessary attributes and qualifications to be effective champions, so the implementation support focused on the specifics of the innovation, and change management strategies. However, these assumptions did not hold true in all study sites, which appears to have influenced study outcomes.
Implications for implementation science
Social scientists have long argued that articulating tacit assumptions—beliefs that “you accept as true without question or proof” —is necessary to understand the impact of such assumptions on research processes and results [53, 55, 56]. However, consideration of the impact of assumptions on study design and outcomes is largely absent in the implementation of science literature. We contend that reflexivity, or the active querying of one’s own assumptions and related decisions during the design phase—particularly as these assumptions relate to likely barriers and facilitators to intervention uptake—is essential to avoid a mismatch between implementation support and determinants .
In addition, despite strong evidence on the importance of champions in implementation activities, little direction exists on how best to support/develop/prepare champions. Often, the literature implies that effective champions have certain intrinsic qualities that cannot be taught [30, 31, 37]. The few articles that do address increasing champions’ efficacy recommend fairly vague strategies such as creating and sustaining learning communities, ongoing mentoring and feedback, fostering the development of leadership and change management skills, and valuing and rewarding champions for their contribution, coupled with hands-on practice and content-specific training [36, 38]. Relevant recommendations note only the need to “identify and prepare individuals who dedicate themselves to supporting, marketing, and driving through an implementation” . A recent article on the attributes of effective champions suggested that many of the necessary skills can be learned, and that supporting the development of these skills may be key to successful implementation outcomes .
Our findings also indicate that implementation science theories and frameworks that involve the use of champions should be refined to include detailed specifications on both necessary intrinsic champion attributes and guidance on developing and supporting effective champions. The analysis presented here contributes to the growing knowledge base within the field regarding what makes an effective champion [31, 49, 57]; more research is needed to identify essential champion characteristics, distinguish between those that are context-dependent (e.g., status within the clinic hierarchy) versus those that can be taught, and identify specific, pragmatic techniques that effectively foster necessary skills—all while accounting for the impact of contextual factors on implementation approaches and outcomes.
All of the study community health centers volunteered to participate, and may have shared unique motivations that limit the generalizability of study findings. Qualitative data collection did not occur evenly across community health centers due to lack of engagement and staff turnover at some community health centers, as well as the mid-study closure of one organization. Findings are presented at the organizational rather than clinic level; it is possible that a single clinic within a community health center may have driven the change in cardioprotective prescribing, although we believe this is unlikely. A major finding of the original analysis was that aspects of the CVD bundle itself proved a barrier to implementation . It is possible that weaknesses in the tools yielded a situation in which only sites with strong study implementers were able to make significant progress; better tools might have necessitated less reliance on champions. In addition, while our study used a cluster-randomized design to minimize bias introduced by unrecognized confounders, our randomization scheme was based on available and readily quantifiable factors such as clinic size, urban/rural location, and the prevalence of diabetes. Randomizing by such factors did not, however, ensure equal distribution of the factors ultimately recognized to be associated with differences in study outcomes.
This analysis adds to implementation science’s call for better approaches to selecting and operationalizing implementation strategies suitable to a given context [2, 5, 9, 11, 15, 26]. Here, unexamined researcher assumptions, coupled with a lack of specification  regarding how to prepare effective champions, led to implementation support that failed to address key barriers to success. These results also increase our understanding of the causal mechanisms through which champions may influence implementation outcomes. Implementation practitioners require detailed, pragmatic, context-specific, evidence-based recommendations on how to select and execute implementation strategies; thus, future research should focus on generating evidence on how to support the growth of effective champions.
Availability of data and materials
The quantitative data analyzed during the current study will not be made publicly available but are available from the corresponding author on reasonable request. The qualitative data analyzed during the current study are not publicly available due to them containing information that could compromise research participant privacy; the codebook and data collection tools are available on request.
Electronic health record
Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci. 2013;8(1):139.
Leeman J, Birken SA, Powell BJ, Rohweder C, Shea CM. Beyond “implementation strategies”: classifying the full range of strategies used in implementation science and practice. Implement Sci. 2017;12(1):125.
Powell BJ, Fernandez ME, Williams NJ, Aarons GA, Beidas RS, Lewis CC, et al. Enhancing the impact of implementation strategies in healthcare: a research agenda. Front Public Health. 2019;7:3.
Ross J, Stevenson F, Dack C, Pal K, May C, Michie S, et al. Developing an implementation strategy for a digital health intervention: an example in routine healthcare. BMC Health Serv Res. 2018;18(1):794.
Go VF, Morales GJ, Mai NT, Brownson RC, Ha TV, Miller WC. Finding what works: identification of implementation strategies for the integration of methadone maintenance therapy and HIV services in Vietnam. Implement Sci. 2016;11(1):54.
Birken SA, Nilsen P. Implementation science as an organizational process. Health Care Manage Rev. 2018;43(3):181.
Waltz TJ, Powell BJ, Fernández ME, Abadie B, Damschroder LJ. Choosing implementation strategies to address contextual barriers: diversity in recommendations and future directions. Implement Sci. 2019;14(1):42.
Gagliardi AR, Alhabib S, and the members of the Guidelines International Network Implementation Working G. Trends in guideline implementation: a scoping systematic review. Implement Sci. 2015;10(1):54.
Powell BJ, Beidas RS, Lewis CC, Aarons GA, McMillen JC, Proctor EK, et al. Methods to improve the selection and tailoring of implementation strategies. J Behav Health Serv Res. 2017;44(2):177–94.
Boyd MR, Powell BJ, Endicott D, Lewis CC. A method for tracking implementation strategies: an exemplar implementing measurement-based care in community behavioral health clinics. Behav Ther. 2018;49(4):525–37.
Baker R, Camosso-Stefinovic J, Gillies C, Shaw EJ, Cheater F, Flottorp S, et al. Tailored interventions to address determinants of practice. Cochrane Database of Syst Rev. 2015;4. https://doi.org/10.1002/14651858.CD005470.pub3.
Mittman BS. Implementation science in health care. In: Brownson RC, Colditz GA, Proctor EK, editors. . New York: Oxford University Press; 2012.
Wensing M, Huntink E, van Lieshout J, Godycki-Cwirko M, Kowalczyk A, Jager C, et al. Tailored implementation of evidence-based practice for patients with chronic diseases. PLoS One. 2014;9(7):e101981.
Wensing M, Bosch M, Grol R. Selecting, tailoring, and implementing knowledge translation interventions. Wiley-Blackwell: Oxford, England; 2009.
Harrison MB, Legare F, Graham ID, Fervers B. Adapting clinical practice guidelines to local context and assessing barriers to their use. Cmaj. 2010;182(2):E78–84.
Gold R, Bunce AE, Cohen DJ, Hollombe C, Nelson CA, Proctor EK, et al. Reporting on the strategies needed to implement proven interventions: an example from a “real-world” cross-setting implementation study. Mayo Clinic proceedings. 2016;91(8):1074–83.
Kothari A, Rudman D, Dobbins M, Rouse M, Sibbald S, Edwards N. The use of tacit and explicit knowledge in public health: a qualitative study. Implement Sci. 2012;7(1):20.
Damschroder L, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4(50). https://doi.org/10.1186/1748-5908-4-50.
Rycroft-Malone J, Burton CR, Wilkinson J, Harvey G, McCormack B, Baker R, et al. Collective action for implementation: a realist evaluation of organisational collaboration in healthcare. Implement Sci. 2016;11:17.
Moullin JC, Dickson KS, Stadnick NA, Rabin B, Aarons GA. Systematic review of the exploration, preparation, implementation, sustainment (EPIS) framework. Implement Sci. 2019;14(1):1.
Tabak RG, Khoong EC, Chambers DA, Brownson RC. Bridging research and practice: models for dissemination and implementation research. Am J Prev Med. 2012;43(3):337–50.
Wandersman A, Duffy J, Flaspohler P, Noonan R, Lubell K, Stillman L, et al. Bridging the gap between prevention research and practice: the interactive systems framework for dissemination and implementation. Am J Community Psychol. 2008;41(3-4):171–81.
Ober AJ, Watkins KE, Hunter SB, Ewing B, Lamp K, Lind M, et al. Assessing and improving organizational readiness to implement substance use disorder treatment in primary care: findings from the SUMMIT study. BMC Fam Pract. 2017;18(1):107.
Weiner BJ. A theory of organizational readiness for change. Implement Sci. 2009;4(1):67.
Perry CK, Damschroder LJ, Hemler JR, Woodson TT, Ono SS, Cohen DJ. Specifying and comparing implementation strategies across seven large implementation interventions: a practical application of theory. Implement Sci. 2019;14(1):32.
Flottorp SA, Oxman AD, Krause J, Musila NR, Wensing M, Godycki-Cwirko M, et al. A checklist for identifying determinants of practice: a systematic review and synthesis of frameworks and taxonomies of factors that prevent or enable improvements in healthcare professional practice. Implement Sci. 2013;8:35.
Powell BJ, Waltz TJ, Chinman MJ, Damschroder LJ, Smith JL, Matthieu MM, et al. A refined compilation of implementation strategies: results from the Expert Recommendations for Implementing Change (ERIC) project. Implement Sci. 2015;10(1):21.
Gold R, Bunce A, Cowburn S, Davis J, Nelson J, Nelson C, et al. Does increased implementation support improve community clinics’ guideline-concordant care? Results of a mixed methods, pragmatic comparative effective trial. Implement Sci. 2019;14(1):100.
Hendy J, Barlow J. The role of the organizational champion in achieving health system change. Soc Sci Med (1982). 2012;74(3):348-355.
Shaw EK, Howard J, West DR, Crabtree BF, Nease DE Jr, Tutt B, et al. The role of the champion in primary care change efforts: from the State Networks of Colorado Ambulatory Practices and Partners (SNOCAP). J Am Board Fam Med. 2012;25(5):676–85.
Miech EJ, Rattray NA, Flanagan ME, Damschroder L, Schmid AA, Damush TM. Inside help: an integrative review of champions in healthcare-related implementation. SAGE Open Med. 2018;6:2050312118773261.
Soo S, Berta W, Baker G. Role of champions in the implementation of patient safety practice change. Healthcare Q. 2009;12(Sp):123–8.
Urquhart R, Kendell C, Geldenhuys L, Ross A, Rajaraman M, Folkes A, et al. The role of scientific evidence in decisions to adopt complex innovations in cancer care settings: a multiple case study in Nova Scotia, Canada. Implement Sci. 2019;14(1):14.
Lewis CC, Klasnja P, Powell BJ, Lyon AR, Tuzzio L, Jones S, et al. From classification to causality: advancing understanding of mechanisms of change in implementation science. Front Public Health. 2018;6:136.
Goff D, Lloyd-Jones D, Bennett G, Coady S, D’Agostino RB, Gibbons R, et al. ACC/AHA guideline on the assessment of cardiovascular risk. Circulation. 2013;129(25):S49–73.
Shifaza F, Evans D, Bradley H, Ullrich S. Developing evidence-based practice champions in the Maldives. Int J Nurs Pract. 2013;19(6):596–602.
Luz S, Shadmi E, Admi H, Peterfreund I, Drach-Zahavy A. Characteristics and behaviours of formal versus informal nurse champions and their relationship to innovation success. J Adv Nurs. 2019;75(1):85–95.
Warrick DD. Developing organization change champions. OD Practitioner. 2009;41:14–9.
Gold R, Bunce A, Cowburn S, Davis JV, Hollombe C, Nelson CA, et al. Cardiovascular care guideline implementation in community health centers in Oregon: a mixed-methods analysis of real-world barriers and challenges. BMC Health Serv Res. 2017;17(1):253.
Creswell JW, Plano Clark VL. Designing and conducting mixed methods research. 3rd ed. Thousand Oaks, CA: SAGE Publications; 2017. Chapter 7: Analyzing and Interpreting Data in Mixed Methods Research.
Cohen D, McDaniel RR Jr, Crabtree BF, Ruhe MC, Weyer SM, Tallia A, Miller WL, Goodwin MA, Nutting P, Solberg LI, Zyzanski SJ, Jaen CR, Gilchrist V, Stange KC. A practice change model for quality improvement in primary care practice. J Healthc Manag. 2004;49(3):155–68 discussion 169-170.
Lewis-Beck MS, Bryman A, Liao TF. The SAGE encyclopedia of social science research methods. Thousand Oaks, CA: SAGE Publications; 2003. Grounded Theory.
Strauss A, Corbin J. Basics of qualitative research – techniques and procedures for developing grounded theory. 2nd ed. London, England: Sage Publications; 1998.
Zhou H, Taber C, Arcona S, Li Y. Difference-in-differences method in comparative effectiveness research: utility with unbalanced groups. Appl Health Econ Health Policy. 2016;14(4):419–29.
Braun V, Clarke V. Using thematic analysis in psychology. Qual Res Psychol. 2006;3(2):77–101.
Glaser BG, Strauss AL. The constant comparative method of qualitative analysis. Social Problems. 1965;12(4):436–55.
Borkan J. Immersion/Crystallization. In: Crabtree BF, Miller WL, editors. Doing qualitative research. 2nd ed. Thousand Oaks, CA: Sage Publications; 1999. p. 179–94.
O’Brien BC, Harris IB, Beckman TJ, Reed DA, Cook DA. Standards for reporting qualitative research: a synthesis of recommendations. Acad Med. 2014;89(9):1245–51.
Bonawitz K, Wetmore M, Heisler M, Dalton VK, Damschroder LJ, Forman J, Allan KR, Moniz MH. Champions in context: which attributes matter for change efforts in healthcare? Implementation Science. 2020;15:62.
Sharkey S, Hudak S, Horn SD, Barrett R, Spector W, Limcangco R. Exploratory study of nursing home factors associated with successful implementation of clinical decision support tools for pressure ulcer prevention. Adv Skin Wound Care. 2013;26(2):83–92 quiz p.3-4.
Aagaard EM, Gonzales R, Camargo CA, Auten R, Levin SK, Maselli J, et al. Physician champions are key to improving antibiotic prescribing quality. Jt Comm J Qual Saf. 2010;36(3):109–16.
Urquhart R, Porter GA, Sargeant J, Jackson L, Grunfeld E. Multi-level factors influence the implementation and use of complex innovations in cancer care: a multiple case study of synoptic reporting. Implement Sci. 2014;9:121.
Fook J. Reflexivity as method. Palliative Care Unit: La Trobe University, Melbourne; 1999.
Cambridge Dictionary. Meaning of assumption in English [Webpage]. Available from: https://dictionary.cambridge.org/us/dictionary/english/assumption. .
White S, Fook J, Gardner F. Critical reflection in health and social care: open university press; 2006. Chapter 10: Using critical reflection in research and evaluation; p144-155.
Jones-Devitt S, Smith L. Critical thinking in health and social care. Thousand Oaks, CA: SAGE Publications; 2007. Chapter 2: The authenticity of knowledge; p16-28.
Ramos-Vidal, I., Palacio, J., Villamil, I. et al. Examining the effect of the individual characteristics of implementers and the interaction of multiple relationships on the structure of psychosocial intervention teams. Implement Sci. 2020;15(69).
The authors wish to acknowledge the contributions of Elisabeth Hicks, MA, Joanna E. Bulkley, PhD, and L. Kris Gowen, PhD. We also wish to thank OCHIN’s leadership, and the leadership and staff of the 29 OCHIN member clinics that volunteered to be part in this study.
Research reported in this publication was supported by the National Heart, Lung, and Blood Institute of the National Institutes of Health under Award Number R01HL120894.
Ethics approval and consent to participate
The study was performed in accordance with the ethical standards of the institutional and/or national research committee and with the World Medical Association Declaration of Helsinki (1964) and its later amendments or comparable ethical standards. The study was approved by the Institutional Review Board at Kaiser Permanente Northwest (Protocol: Pro00004392). The IRB waived the requirement to obtain informed consent.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
About this article
Cite this article
Bunce, A.E., Gruß, I., Davis, J.V. et al. Lessons learned about the effective operationalization of champions as an implementation strategy: results from a qualitative process evaluation of a pragmatic trial. Implementation Sci 15, 87 (2020). https://doi.org/10.1186/s13012-020-01048-1
- Implementation strategies
- Practice change
- Clinical champions