- Open Access
- Open Peer Review
Using a continuum of hybrid effectiveness-implementation studies to put research-tested colorectal screening interventions into practice
Implementation Sciencevolume 14, Article number: 53 (2019)
Few previous studies have applied the hybrid effectiveness-implementation design framework to illustrate the way in which an intervention was progressively implemented and evaluated across multiple studies in diverse settings.
We describe the design components and methodologies of three studies that sought to improve rates of colorectal cancer (CRC) screening using mailed outreach, and apply domains put forth by Curran et al.: research aims, research questions, comparison conditions, sample, evaluation methods, measures, and potential challenges.
The Hybrid 1 study (emphasis on effectiveness) was a patient-level randomized trial of a mailed fecal test and stepped phone-outreach intervention program delivered in an integrated healthcare system (21 clinics, 4673 patients). The primary outcome was effectiveness (CRC screening uptake). Implementation outcomes included cost-effectiveness and acceptability.
The Hybrid 2 study (shared emphasis on effectiveness and implementation) was a pragmatic cluster-randomized trial of mailed fecal immunochemical test (FIT) outreach implemented at safety net clinics (26 clinics, 41,000 patients). The intervention used electronic health record tools (adapted from Hybrid 1) and clinic personnel to deliver the intervention. Outcomes included effectiveness (FIT completion) and implementation (FIT kits delivered, clinic barriers and facilitators, cost-effectiveness).
Hybrid 3 study (emphasis on implementation) is a demonstration project being conducted by two Medicaid/Medicare insurance plans (2 states, 12,000 patients) comparing two strategies for implementing mailed FIT programs that addressed Hybrid 2 implementation barriers. Outcomes include implementation (activities delivered, barriers) and effectiveness (FIT completion).
The effectiveness-implementation typology successfully identified a number of distinguishing features between the three studies. Two additional features, program design and program delivery, varied across our studies, and we propose adding them to the current typology. Program design and program delivery reflect the process by which and by whom a program is designed and delivered (e.g., research staff vs. clinic/health plan staff).
We describe three studies that demonstrate the hybrid effectiveness to implementation continuum and make recommendations for expanding the hybrid typology to include new descriptive features. Additional comparisons of Hybrid 1, 2, and 3 studies may help confirm whether our hybrid typology refinements are generalizable markers of the pipeline from research to practice.
Strong evidence indicates that colorectal cancer (CRC) screening decreases CRC morbidity and mortality. However, only 62% of age-eligible adults are current for screening, with lower rates among low-income (47%), uninsured (25%), and Hispanic populations (47%) . Effective strategies for improving CRC screening uptake and decreasing disparities are needed, as is research on whether the most effective strategies can be adapted, implemented, and maintained in diverse settings.
Much has been written about difficulties in translating evidence-based practices and research-tested programs into routine care. From the time evidence is generated to the time it becomes part of everyday practice is estimated to take longer than 17 years . Reasons for this delay include that (1) resources used to implement an effective program in a research setting may not be available in community settings, (2) the original intervention was tested in a unique setting (such as an academic center) and under optimal conditions that are not generalizable to other settings, and (3) the research included patients that are not representative of the general population or did not test the program with disadvantaged groups (such as those with language barriers or who lack health insurance).
We describe the design components and methodologies of three studies that sought to move a research-tested program into practice. In our description, we apply the hybrid effectiveness-implementation design models as described by Curran et al. . The hybrid model provides a framework for designing studies that have dual purposes of testing intervention effectiveness and implementation strategies to accelerate knowledge creation and increase the relevance of research . Hybrid studies provide a pathway for rapidly moving knowledge from research to implementation. Hybrid 1 studies generally focus on the effectiveness of a clinical intervention in a setting where it will be implemented, while secondarily exploring some aspects of implementation (e.g., fidelity of intervention delivery, costs). The main emphasis in this type of study is the following: Is the clinical intervention effective? Hybrid 2 studies focus on both implementation and effectiveness outcomes, and generally ask the following: Does the implementation method facilitate implementation of the intervention and is it still effective? Hybrid 3 studies focus on comparing different implementation strategies for scaling up and maintaining an intervention. We present here three pragmatic studies designed along this continuum, with each sequentially informing the next research questions and study design . We recommend ways for augmenting the hybrid effectiveness-implementation design framework to include two additional categories: program design and program delivery.
We first described key characteristics of three studies using the hybrid effectiveness-implementation design continuum, including the design, setting, data sources, participants, intervention/program, and outcomes. We then used Curran et al.’s published domains (Table 1) to characterize these Hybrid 1, 2, and 3 studies. These domains include research aims, research questions, units of randomization, comparison conditions, sample, evaluation methods, measures, and potential challenges. Finally, we compared the characteristics of the three hybrid studies along these domains and considered how our studies fit with the hybrid typology. This analysis led us to identify two additional domains, program design and program delivery, that distinguished our three studies along the Hybrid 1 through 3 continuum, as described below.
Program design considers the initial design of the program or intervention, and further adaptations for varied settings and populations. In the effectiveness to implementation continuum of our studies, we identified varying levels of collaboration between the program sites and researchers. This continuum went from researchers determining the design, to a shared process, to researchers mainly serving as consultants and/or evaluators. Program delivery includes the specification of the who, what, where, when, and how the program is delivered in each setting and is in part informed by the work of Proctor and others in specifying implementation components [5, 6]. For example, the who in Proctor is called the actor—or the individual implementing the strategy, which might be the research team or individuals within the setting with or without research team support or training. The what, when, and how the strategy is delivered may also vary across studies, with some studies focusing mainly on effectiveness and consistent delivery of interventions, and other studies focusing on real-world implementation of various program delivery strategies.
Description of the Hybrid 1, 2, and 3 study examples
Multiple prior publications provide detailed descriptions of the Hybrid 1 and 2 studies presented here; the Hybrid 3 study is on-going [7,8,9]. A brief description of each study is provided below. We also describe the two new domains, program design and program implementation as these were distinctly different between the three studies.
The Hybrid 1 study (Systems of Support to Increase Colorectal Cancer Screening, SOS, R01 CA121125) was a 4-arm patient-level randomized trial (4675 patients). It was conducted from 2008 to 2011 in an integrated health care system that provides both health insurance and health care (primary and specialty care, including colonoscopies) [7, 10]. The research staff used automated data (electronic health record [EHR] data, laboratory data, and procedural CPT codes) to identify age-eligible patients not current for CRC screening and evaluate outcomes. The intervention was embedded in the health care system and used an EHR-linked registry to deliver a centralized program to encourage CRC screening uptake. Usual care (Arm 1) received annual birthday reminders of overdue screening tests including CRC screening. In addition to usual care, participants randomized to active interventions received stepped intensity interventions: information on CRC screening choices, mailed fecal kits, postage-paid return envelopes, and a mailed reminder if the kit was not returned (Arm 2); mailings plus brief telephone assistance if screening was not completed after the mailings (Arm 3); mailings plus brief assistance plus nurse navigation for those still not completing screening (Arm 4). The research team managed the database with a vendor service mailing the intervention components; clinic medical assistants and nurses worked from their regular clinic but had protected time (the study paid for about 4–8 h a week of their time) to provide telephone assistance and navigation to patients across the organization . The primary outcome was CRC screening adherence over 2 years in each of the progressive intensity arms compared to usual care (Arm 1). Secondary implementation outcomes included the reach of the intervention, cost-effectiveness from the health plan’s perspective, qualitative assessments from the patients’ perspective of how the intervention could be improved, and the intervention’s fit within the organization’s other quality improvement efforts to increase CRC screening [10,11,12,13,14]. The mailed program alone doubled CRC screening uptake, with incremental increases in the stepped intensity groups . Mailed interventions and calls from medical assistants or nurses from other clinics were acceptable to patients if the medical assistant or nurse had access to their EHR record and could communicate with the patient’s physician .
The Hybrid 2 study (Strategies and Opportunities to STOP Colon Cancer in Priority Populations, UH3 AT007782), was a pragmatic cluster-randomized trial conducted from 2014 to 2016 at 26 safety net clinics with over 40,000 patients in the first year who were age-eligible and overdue for CRC screening [8, 15]. EHR and laboratory data were used to identify eligible patients and to evaluate effectiveness outcomes. The intervention was adapted in part from the Hybrid 1 SOS study, in that it used EHR data to identify individuals overdue for CRC screening and a registry to automatically generate mailings. However, the research team, EHR vendor, and clinic staff collaboratively designed and implemented the program for use in community clinic settings and by diverse patients. Adaptations included translating patient letters into Spanish and other languages, using wordless fecal immunochemical test (FIT) pictographic instructions , and embedding the registry into the EHR (Reporting Workbench) to generate lists of patients overdue for CRC screening and to create mailings. An advisory committee made up of organizational and clinic program leaders, CRC screening experts, and patients reviewed all materials and suggested changes or enhancements. Clinic staff generated lists and completed mailings, with the research team only providing initial training, monthly clinic staff “champion” meetings to augment training and troubleshoot issues with the EHR-embedded tool, and facilitation of Plan-Do-Study-Act (PDSA) cycles to help the clinics improve program implementation . The primary outcomes were FIT completion (effectiveness) and implementation outcomes based on the Reach, Effectiveness, Adoption, Implementation, and Maintenance (RE-AIM) framework (e.g., percent of clinics participating, percent of letters mailed by each clinic, percent of clinics continuing the program) and the Consolidated Framework for Implementation Research (CFIR) to assess contextual factors related to implementation (e.g. barriers and facilitators). Additional implementation outcomes included program cost-effectiveness. The intervention overall led to significantly higher rates of FIT completion compared to usual care clinics , but the net difference was modest (net increase 3.4%), mainly because implementation varied greatly across intervention clinics, with the health center-level percent of eligible intervention participants mailed a FIT ranging from 6.5 to 68.2% . Among the patients who were mailed FITs, completion rates were close to the Hybrid 1 SOS study, 21% . Mixed methods summative analyses of barriers and facilitators to implementation are underway, but qualitative rapid cycle assessments during the study suggested that contributing factors were delays in start-up, personnel changes and staffing shortages, and challenges with the technologies and lab ordering processes .
The Hybrid 3 study began in 2015 and is an on-going evaluation of two different mailed FIT programs being implemented by two Medicaid/Medicare insurance plans in 2 states (BeneFIT, UDP-14-0012) , with age-eligible health plan members identified and outcomes evaluated using claims data (12,000 patients in year 1). The investigators initiated the relationship with the health plans with the goal of addressing implementation barriers experienced in the Hybrid 2 study, specifically that clinics often had insufficient staff or resources to consistently implement the program. One program is working collaboratively with clinics to integrate it into their usual workflow (e.g., a vendor mails the same FIT kit type used in clinical care, and completed tests are mailed back to the clinics, where staff place test orders and deliver follow-up care). The other is a centralized program whose vendor mails the FITs, which are sent to a centralized laboratory for processing, then sends the results to the providers and directly contacts patients with abnormal results. Mixed methods outcomes include percent of age-eligible patients receiving mailings, acceptability of and satisfaction with the program by the health plans, implementation barriers and facilitators, effectiveness (FIT and colonoscopy completion rates), and program costs. Contextual factors related to implementation and maintenance of the programs are assessed using the CFIR framework.
Hybrid design model characteristics of the 3 study examples
Table 2 compares each of the three reported hybrid studies across key design characteristics and shows differences with progression from Hybrid 1 to Hybrid 3 designs. SOS, the Type 1 hybrid study, was a person-level randomized controlled trial that invited almost all age-eligible individuals’ overdue for CRC screening and had few exclusions (CRC, inflammatory bowel disease, life-threatening health conditions). Outcomes were primarily effectiveness, but mixed methods were used to understand screening barriers and facilitators and to further refine study interventions in subsequent study years . Cost-effectiveness evaluations were performed from a health care organization perspective to assist with decisions about adoption of the overall program and its specific components (such as the incremental cost-effectiveness of adding phone assistance and navigation) . Challenges included the requirement of verbal consent by the institutional review board, with only 38% of those invited agreeing to participate . Non-white and Hispanic individuals, and individuals with lower levels of education, were less likely to participate than non-Hispanic white individuals and patients with higher levels of education.
STOP CRC, the Hybrid 2 study, sought to adapt and implement the SOS program within community health center clinics that care for populations with the lowest CRC screening rates. STOP CRC had almost no exclusions and consent was waived, making it possible to reach nearly 100% of the population that needed the intervention. This study measured both overall effectiveness of the program compared to control clinics, (i.e., FIT uptake) and variation in effectiveness across health centers . Implementation was measured as the proportion of eligible patients who were mailed a FIT (quantitative) at each health center and their barriers and facilitators to implementation (qualitative)  and maintenance (RE-AIM). Early in the study, it became obvious that implementation fidelity was the primary challenge, with some clinics delaying start-up until almost the end of the participant accrual interval. These delays and implementation variability led to reductions in program effectiveness.
In BeneFIT, the Hybrid 3 study, we are evaluating mailed FIT programs designed to decrease the burden of implementation delivery on clinics by enlisting Medicaid/Medicare health insurance plans . BeneFIT is a quasi-experimental, naturalistic study that compares two different implementation strategies developed by the health plans, without any usual care comparison group. The investigators serve as consultants to the health plans for both program design and delivery and are conducting a formal program evaluation. The health plans are responsible for conducting their programs, and both contracted with vendors to implement the mailings. The program is measuring implementation using mixed methods: quantitative (e.g., proportion of age-eligible enrollees mailed a FIT and delivered reminders, supporting activities by clinics, program costs) and qualitative interviews of key stakeholders to assess motivations, barriers, and facilitators of the program. Effectiveness outcomes are limited to what is easily measurable (e.g., FIT and colonoscopy completion rates using claims data). The Hybrid 3 study has no usual care comparison, as we knew already from the Hybrid 2 study that no mailings led to lower FIT completion rates.
Program design and program delivery
Program design and delivery, the two new domains proposed for the hybrid studies, were distinctly different between the three studies. In Hybrid 1, the research team designed and built the EHR-linked mailed fecal testing program, trained clinical staff, and paid for their time, which helped to ensure fidelity to and completion of the intervention components. In Hybrid 2, the research team collaborated with the EHR vendor and health centers to design and build the EHR-embedded mailed FIT program, but the program itself was delivered by the clinic staff, with the research team assisting only with training and quality improvement activities. The timing, dose, and components of the intervention varied across health centers. For example, some health centers mailed letters and kits only once during the intervention year, while others completed monthly mailings. Some health centers combined the introductory letter mailing with the kit mailing and some required that the completed FIT be returned to the clinic, both were associated with decreased program effectiveness. The research provided clinics with small incentives but did not pay for program implementation. In Hybrid 3, the research team serves as consultants to the health plans and assists with evaluation but provides no training or implementation assistance. Hybrid 3 was paid for entirely by the health plans. They received an incentive for participation, but this was mainly to offset costs related to research, data capture, qualitative assessments, and time spent in meetings with the research staff.
The US Preventive Services Task Force recommends CRC screening for adults aged 50–75 because of strong evidence that it decreases CRC incidence and mortality . However, screening rates remain suboptimal, particularly among disadvantaged populations . The three studies we present were designed sequentially (informed by each prior study) to move evidence-based, research-tested interventions to increase CRC screening in diverse settings. The Hybrid 1 study, SOS, had the main goal of testing intervention effectiveness. It was embedded in a health care system, used automated data to identify eligible participants, used an EHR-linked registry to generate mailings, and employed clinic staff that worked in the delivery system to outreach to consented patients. Cost-effectiveness assessments were designed to inform decision-making about the spread of the program, and mixed methods explored potential refinements of the program and feasibility of implementation. The Hybrid 2 study, STOP CRC, moved the intervention into a community health center setting with a more diverse population. Key differences from the Hybrid 1 study were (1) inclusion of almost all age-eligible people, as consent was not required; (2) the collaborative process undertaken to design the program, which involved the EHR vendor, health center staff, and researchers; and (3) the implementation of the program solely by health center staff. The research team gave equal attention to evaluating overall program effectiveness and implementation by the community health centers. The Hybrid 3 study, BeneFIT, compares two different implementation strategies developed by the health plans, and was designed in part to decrease the implementation burden experienced in the Hybrid 2 study, STOP CRC. Each study answers important effectiveness and implementation questions and generated new questions.
We encountered some difficulties in trying to fit our 3 studies neatly into the hybrid framework. The Hybrid 1 study, SOS, is described as testing the effectiveness of an intervention on increasing CRC screening. However, one might argue that SOS instead tested different implementation strategies (i.e., the stepped interventions were implementation approaches). SOS had both “proof of concept” and pragmatic features. It was among the first trials to use EHR data to identify people and generate interventions for individuals’ overdue for CRC screening. It was also pragmatic in that it was embedded in a health care system and relied on employed clinic staff. As the primary outcome of the SOS trial was effectiveness of the intervention compared to usual care, we consider it to be an effectiveness trial, embedded in a real-world setting, but others might consider it to be an implementation study comparing different implementation methods, or an effectiveness trial only. Others have acknowledged that the distinction between effectiveness and implementation trials can be blurred .
Curran et al.’s framework domains omitted some distinguishing features we observed in our Hybrid 1, Hybrid 2, and Hybrid 3 studies, such as the role of research staff versus clinic/health plan staff in designing (or adapting) and implementing the intervention . We think these features might be generalizable to other hybrid studies that progress along the research to practice continuum. Based on our three studies, we propose additional categories to the current hybrid typology.
Collaboration with the target setting is often a critical component of implementation studies and helps to ensure a smoother implementation process. In the Hybrid 1 study, the researchers designed program elements and dose, while considering the resources available in that specific setting. In the Hybrid 2 study, researchers and healthcare centers collaboratively adapted the Hybrid 1 study components based on the new settings’ resources. In the Hybrid 3 study, the health plans implementing the program adapted the program components to fit their resources, with the researchers serving as consultants only.
How an intervention is delivered is also a critical aspect of its transition from research to clinical practice. In our studies, the “agents” who delivered the intervention, as defined by Proctor et al. , shifted from tight to almost no research control. In the Hybrid 1 study, the research team delivered the mailings, and they also trained and monitored the clinical staff (nurses who made the follow-up calls). In the Hybrid 2 study, the research team assisted with training and implementation supports (PDSA cycles). In the Hybrid 3 study, the research team provided consultation services only, with the health plans delivering the program on their own, as would be the case once the program was institutionalized.
Our studies demonstrate that research-tested interventions for increasing CRC screening are rarely ready “right out of the box” for broad dissemination and may explain in part the long transition from research to practice. Even when the intervention is found to be effective, cost-effective, feasible, and acceptable, transitioning the intervention to day-to-day practice is a stepped and nuanced process, and one size may not fit all. Furthermore, contextual factors may lead to diminution of intervention effectiveness, resulting in additional research questions and further program adaptations.
Our analysis of the Hybrid 1, 2, and 3 SOS, STOP CRC, and BeneFIT studies has limitations. We describe the hybrid continuum of research to practice studies using only one area of research, CRC screening, and this continuum may be different than in other areas of health services implementation research. We also note that others may disagree with the hybrid category we assigned each study.
Our studies began prior to the hybrid design framework publication, and while our knowledge of the model may have influenced the design of our Hybrid Type 2 and 3 studies, we did not plan our progression a priori based on the framework. The progression of the three studies over a ten-year period was more organic, with each study answering questions posed by the prior study, and implementation progressively becoming more important than effectiveness, as the body of evidence on effectiveness grew. Lastly, the hybrid framework was intended to speed research into practice, which our studies may or may not demonstrate. Mailed FIT programs are burgeoning, but implementing sites face many implementation challenges and potential solutions that we and others continue to explore.
Research-tested interventions are rarely ready for widespread dissemination. The hybrid models provide a framework for a series of progressive studies that test not just effectiveness, but also implementation, with each step progressing towards real-world practice. We present three studies that illustrate this process. We also found that the study continuum from an effectiveness focus to a focus on implementation may be nuanced and difficult to categorize clearly. We propose adding two new features that might be useful in characterizing hybrid studies: program design and program delivery, with a continuum from researcher-led in Hybrid 1 studies; researcher-organizational collaboration in Hybrid 2 studies; to organization-led in Hybrid 3 studies. Our three hybrid studies’ design and delivery became increasingly pragmatic, adaptable, and informed by the contexts, individuals, and organizations delivering them. Additional comparisons of Hybrid 1, 2, and 3 may help to confirm whether these new features are generalizable markers of studies along the pipeline from research to practice and whether they speed the uptake of research into practice.
BeneFIT to increase Colorectal Screening in Priority Populations
Consolidated framework for implementation research
Electronic health record
Fecal immunochemical test
Reach, Effectiveness, Adoption, Implementation, Maintenance
Systems of Support to Increase Colorectal Cancer Screening and Follow-up
- STOP CRC:
Strategies and Opportunities to STOP Colon Cancer in Priority Populations
White A, Thompson TD, White MC, et al. Cancer screening test use - United States, 2015. MMWR Morb Mortal Wkly Rep. 2017;66:201–6.
Glasgow RE, Lichtenstein E, Marcus AC. Why don’t we see more translation of health promotion research to practice? Rethinking the efficacy-to-effectiveness transition. Am J Public Health. 2003;93:1261–7 PMCID: 1447950.
Curran GM, Bauer M, Mittman B, et al. Effectiveness-implementation hybrid designs: combining elements of clinical effectiveness and implementation research to enhance public health impact. Med Care 2012;50:217–226. PMCID: 3731143.
Brown CH, Curran G, Palinkas LA, et al. An overview of research and evaluation designs for dissemination and implementation. Annu Rev Public Health. 2017;38:1–22 PMCID: 5384265.
Proctor EK, Powell BJ, McMillen JC. Implementation strategies: recommendations for specifying and reporting. Implement Sci 2013;8. PMCID: 3882890.
Leeman J, Birken SA, Powell BJ, Rohweder C, Shea CM. Beyond “implementation strategies”: classifying the full range of strategies used in implementation science and practice. Implement Sci 2017 Nov 3;12(1):125. PMCID: 5670723.
Green BB, Wang CY, Horner K, et al. Systems of support to increase colorectal cancer screening and follow-up rates (SOS): design, challenges, and baseline characteristics of trial participants. Contemp Clin Trials. 2010;31:589–603 PMCID: 2956868.
Coronado GD, Vollmer WM, Petrik A, et al. Strategies and opportunities to STOP colon cancer in priority populations: pragmatic pilot study design and outcomes. BMC Cancer. 2014;14 PMCID: 3936821.
Coury JK, Schneider JL, Green BB, et al. Two Medicaid health plans' models and motivations for improving colorectal cancer screening rates. Transl Behav Med 2018. doi: https://doi.org/10.1093/tbm/iby094. [Epub ahead of print] PubMed PMID: 30445511.
Green BB, Wang CY, Anderson ML, et al. An automated intervention with stepped increases in support to increase uptake of colorectal cancer screening: a randomized trial. Ann Intern Med. 2013;158:301–11 PMCID: 3953144.
Murphy CC, Verson SW, Haddock NM, et al. Longitudinal predictors of colorectal cancer screening among participants in a randomized controlled trial. Prev Med. 2014;66:123–30 PMCID: 4209306.
Meenan RT, Anderson ML, Chubak J, et al. An economic evaluation of colorectal cancer screening in primary care practice. Am J Prev Med. 2015;48:714–21 PMCID: 4441758.
Green BB, BlueSpruce J, Tuzzio L, et al. Reasons for never and intermittent completion of colorectal cancer screening after receiving multiple rounds of mailed fecal tests. BMC Public Health. 2017;17:531 PMCID: 5450259.
Green BB, Anderson ML, Chubak J, et al. Colorectal cancer screening rates increased after exposure to the Patient-Centered Medical Home (PCMH). J Am Board Fam Med. 2016;29:191–200 PMCID: 4868497.
Coronado GD, Petrik AF, Vollmer WM, et al. Effectiveness of a mailed colorectal cancer screening outreach program in community health clinics: the STOP CRC cluster randomized clinical trial. JAMA Intern Med. 2018;178:1174–81.
Coronado G, Sanchez J, Devoe J, Green B. Advantages of wordless instructions on how to complete a fecal immunochemical test: lessons from patient advisory council members of a federally qualified health center. J Cancer Educ. 2014;29:86–90 PMCID: 3946071.
Coury J, Schneider JL, Rivelli JS, et al. Applying the Plan-Do-Study-Act (PDSA) approach to a large pragmatic study involving safety net clinics. BMC Health Serv Res. 2017;17:411 PMCID: 5477281.
Coronado GD, Burdick T, Kapka T, et al. Using an automated data-driven EHR-embedded program for mailing FIT kits: lessons from the STOP CRC pilot study. J Gen Pract. 2014;14 PMCID: 3936821.
Green BB, Bogart A, Chubak J, et al. Nonparticipation in population-based trial to increase colorectal cancer screening. Am J Prev Med. 2012;42:390–7 PMCID: 3549634.
Knudsen AB, Zauber AG, Rutter CM, et al. Estimation of benefits, burden, and harms of colorectal cancer screening strategies: modeling study for the US Preventive Services Task Force. JAMA. 2016;315:2595–609.
White A, Thompson TD, White MC, Sabatino SA, de Moor J, Doria-Rose PV, Geiger AM, Richardson LC. Cancer screening test use - United States, 2015. MMWR Morb Mortal Wkly Rep. 2017;66:201–6.
Proctor E, Silmere H, Raghavan R, et al. Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Admin Pol Ment Health. 2011;38:65–76 PMCID: 3068522.
Research reported in this publication was supported by the National Cancer Institute (NCI), R01CA121125; National Cancer Institute of the National Institutes of Health under Award Number UH3CA188640; and Centers for Disease Control (CDC) (U48DP005013). The content is solely the responsibility of the authors and does not necessarily represent the official views of the National Institutes of Health.
We would like to acknowledge Dr. David Chambers, Deputy Director for Implementation Science, Division of Cancer Control and Population Sciences, National Cancer Institute, and Dr. Geoffrey M. Curran Director, Center for Implementation Research, University of Arkansas for Medical Sciences, for reviewing the manuscript and providing input.
National Cancer Institute of the National Institutes of Health under Award No. UH3CA188640.
Availability of data and materials
Ethics approval and consent to participate
The study was approved by the Institutional Review Board of University of Washington (STUDY00000472).
Consent for publication
Not applicable, no consent was required.
This manuscript was supported by grants from the following:
National Cancer Institute (NCI): R01CA121125
National Institute of Health (NIH): Common Fund, through a cooperative agreement (U54 AT007748, UH2AT007782 and 4UH3CA18864002.) from the Office of Strategic Coordination within the Office of the NIH Director.
Centers for Disease Control (CDC): Health Promotion and Disease Prevention Research Center grant supported by cooperative agreement U48DP005013.
The views presented here are solely the responsibility of the authors and do not necessarily represent the official views of the NCI, NIH, or CDC.
From November 2014 to August 2015, Dr. Coronado served as a co-Investigator on an industry-funded study to evaluate patient adherence to an experimental blood test for colorectal cancer. The study was funded by EpiGenomics. From September 2017 to June 2018, Dr. Coronado served as the Principal Investigator on an industry-funded study to compare the clinical performance of an experimental FIT to an FDA-approved FIT. This study was funded by the Quidel Corporation. All other authors declare no conflicts of interest.
The other authors report no conflicts of interest other than the funding that was awarded to their institutions for these grants. All authors take responsibility for the design, writing, editing, and final version submitted.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.