Testing the use of practice facilitation in a cluster randomized stepped-wedge design trial to improve adherence to cardiovascular disease prevention guidelines: HealthyHearts NYC
© Shelley et al. 2016
Received: 25 May 2016
Accepted: 7 June 2016
Published: 4 July 2016
HealthyHearts NYC (HHNYC) will evaluate the effectiveness of practice facilitation as a quality improvement strategy for implementing the Million Hearts’ ABCS treatment guidelines for reducing cardiovascular disease (CVD) among high-risk patients who receive care in primary care practices in New York City. ABCS refers to (A) aspirin in high-risk individuals; (B) blood pressure control; (C) cholesterol management; and (S) smoking cessation. The long-term goal is to create a robust infrastructure for implementing and disseminating evidence-based practice guidelines (EBPG) in primary care practices.
We are using a stepped-wedge cluster randomized controlled trial design to evaluate the implementation process and the impact of practice facilitation (PF) versus usual care on ABCS outcomes in 250 small primary care practices. Randomization is at the practice site level, all of which begin as part of the control condition. The intervention consists of one year of PF that includes a combination of one-on-one onsite visits and shared learning across practice sites. PFs will focus on helping sites implement evidence-based components of patient-centered medical home (PCMH) and the chronic care model (CCM), which include decision support, provider feedback, self-management tools and resources, and linkages to community-based services.
We hypothesize that practice facilitation will result in superior clinical outcomes compared to usual care; that the effects of practice facilitation will be mediated by greater adoption of system changes in accord with PCMH and CCM; and that there will be increased adaptive reserve and change capacity.
KeywordsPractice facilitation Primary care Cardiovascular disease
In 2011, the US Department of Health and Human Services launched the Million Hearts Campaign with an explicit goal of preventing one million heart attacks and strokes by 2017 . A core component of the campaign is to improve implementation and systematic delivery of ABCS treatment guidelines: (A) aspirin in high-risk individuals; (B) blood pressure control; (C) cholesterol management through guideline-recommended use of lipid lowering medications (i.e., statins); and (S) smoking cessation. In order to accelerate adoption of Million Hearts guidelines in primary care practices as well as build capacity for their ongoing implementation and dissemination, the Agency for Healthcare Research and Quality (AHRQ) launched EvidenceNOW—a national initiative that aims to transform health care delivery in small primary care practices by building critical infrastructure to integrate the most up-to-date research into practice to improve the heart health of their patients .
There is a strong rationale for targeting small practices (i.e., fewer than 10 providers). First, small practices continue to provide primary care for a significant proportion of the population. More than half of primary care office visits occur in small practice settings [3–5]. Second, small practices face daunting challenges in redesigning their system and care processes to meet regulatory requirements for practice transformation. Most small practices lack the resources and staff expertise related to information systems, goal setting, data analysis, and practice redesign to coordinate a complex set of multilevel system changes without external assistance . One implementation strategy that may effectively overcome these barriers is practice facilitation (PF) [7, 8].
PF provides external expertise to help organizations make meaningful practice changes that are tailored to local context and in accord with current practice transformation models like the chronic care model (CCM) and patient-centered medical home (PCMH) [9, 10]. Implementation science literature suggests that PF’s emphasis on building organization capacity to adapt clinical evidence to the specific circumstances of the practice environment may be more effective in achieving sustainable improvements in guideline-concordant care and patient outcomes in primary care settings compared to single component (e.g., audit and feedback alone) and other multicomponent guideline implementation strategies (e.g., learning collaborative) [7, 8, 11–22]. A recent systematic review of PF interventions demonstrated that primary care practices with the support of a facilitator are almost three times more likely to implement evidence-based guidelines compared with usual care practices . Despite this evidence, the effectiveness of PF remains largely untested in small primary care practices, especially those that serve vulnerable and underserved minority populations.
HealthyHearts NYC (HHNYC), one of seven cooperatives funded through the EvidenceNOW initiative, is a stepped-wedge cluster randomized controlled trial designed to evaluate the impact of PF compared to usual practice on implementation of the Million Hearts ABCS treatment guidelines in 250 small primary care practices, where a significant proportion of vulnerable populations in NYC receive care.
The specific aims of the HealthyHearts NYC project are to (1) compare the effect of PF versus usual care on ABCS outcome measures; (2) identify baseline organizational characteristics (e.g., site level adaptive reserve, organizational change capacity) that are associated with intervention outcomes; (3) use qualitative methods to assess barriers and facilitators to implementing practice change and achieving ABCS outcomes among high- and low-performing practices in the study; and (4) evaluate the intervention-specific costs per patient and per site.
HHNYC is a partnership between New York University School of Medicine (NYUSOM) and the NYC Department of Health and Mental Hygiene (NYCDOHMH) Primary Care Information Project (PCIP), which has over 1300 small- to medium-sized primary care practices in its regional extension center . The study is being conducted in small practices in NYC that are part of PCIP’s practice network. NYC has the largest (8.5 million) and most diverse population of any city in the US (26 % Hispanic, 26 % black, and 13 % Asian) . Although NYC has achieved substantial decline in CVD-related deaths, heart disease still remains the number one cause of death and significant disparities by income and race/ethnicity persist. Residents of NYC’s poorest neighborhoods, where many of the study small practice sites are located, consistently have higher mortality rates from almost all diseases, including CVD, compared with residents in its wealthiest neighborhoods [25, 26]. For example, hypertension-related death rates among black New Yorkers are four times higher than among white New Yorkers (35 vs. 9 per 100,000 adults). Additionally, patients in PCIP network practices with a greater proportion of low income and minority patients in their panels are less likely to achieve target measures for blood pressure control and smoking cessation measures .
Stepped-wedge study design
Year 1 (months)
Year 2 (months)
Year 3 (months)
1 (80 sites)
2 (70 sites)
3 (70 sites)
4 (70 sites)
Study site eligibility
We recruited practices from PCIP’s practice network, which was established in 2005 to improve health outcomes in low-income communities in NYC. This network includes over 1300 small- to medium-sized practices with significant numbers of Medicaid or uninsured patients. To be eligible, practice sites were required to have <10 full-time equivalent (FTE) healthcare providers (physician, nurse practitioner, physician assistant); have implemented an electronic health record (EHR) for at least 1 year; sign an agreement with PCIP to participate in PCIP’s Hub Population Health System, an EHR query architecture; have no immediate future plans to participate in a CVD-related quality improvement initiative; have no plans to change EHRs in the next 18 months; be willing to identify a site coordinator to work with study staff on all aspects of the intervention; and be located in NYC . We will recruit 40 of the sites that are enrolled in the study to participate in the qualitative assessment described below. For the qualitative interviews, sites will be chosen to ensure representation across different geographic locations in NYC and practice size.
Data is being abstracted for patients who: (a) have at least one of the ABCS risk factors (i.e., eligible for aspirin, has a diagnosis of hypertension, meets criteria for blood cholesterol management, and/or is a current smoker); (b) have received care at the site in the last 12 months; and (c) are age >18 years. Patients eligible for aspirin are those with a documented ICD (ICD-9/10) code for ischemic vascular disease in the last 12 months. Similarly, patients with a diagnosis of hypertension must have a documented ICD code for the targeted risk factor. Patients with hyperlipidemia based on an ICD code or who otherwise meet criteria for cholesterol management are identified by laboratory results for LDL cholesterol as well as ICD codes. Smokers are identified by documentation of current smoking status in the EHR (e.g., meaningful use measure) during the past 24 months.
PCIP focused recruitment on small practices in their network, which comprises over 1300 independent primary care sites in NYC. After a preliminary assessment of eligibility, the sample was reduced to a pool of 547 practices that are focused solely or primarily on adult primary care and currently using an EHR with data sharing capacity. Recruitment started with targeted invitation letters to an initial group of practices via post and email, which included a link to a recorded informational webinar. Simultaneously, PCIP conducted telephone outreach to determine practice site interest, to conduct a more detailed eligibility screening survey, and to secure a verbal commitment to participate. Once practices agreed to participate, PCIP and NYU co-hosted multiple in-person “kickoff” events to educate sites about program details. At these events, eligible practices were asked to sign the study participation agreement. In order to eliminate the risk of contamination between multiple sites in the same practice, only one site per target practice was permitted to join the study.
Practice facilitation protocol
The duration of the intervention is 12 months for each wave, and it is being implemented at the site level. PF comprises the following components: (1) onsite visits by the practice facilitator with the medical director and/or staff twice in the first month and then monthly thereafter; (2) expert consultation (i.e., quarterly webinars and “ask the expert” email consultant service); and (3) facilitated opportunities for shared learning across intervention sites (e.g., peer-to-peer collaborative calls) . The practice facilitator’s main role is to coordinate and facilitate meetings with study sites and to assist primary care practices in setting practice change and performance goals and in developing strategies for implementing evidence-based components of the models for practice transformation, including CCM, PCMH, and Bodenheimer’s 10 building blocks of primary care [9, 10, 33]. For this purpose, a quality improvement (QI) manual informed by these models was developed by PCIP to drive delivery of the intervention. PFs also share best practices across sites, train practice staff on QI processes and methods to help them develop robust QI capacity (e.g., workflow mapping, generating performance reports), assist teams in testing system changes and interpreting outcomes, and assist with generating data for performance feedback and Plan-Do-Study-Act (PDSA) cycles. The practice facilitator conducts phone and email exchanges with the sites as needed between site visits during the 12-month intervention period.
Practice facilitator training and supervision
All practice facilitators received training at the State University of New York at Buffalo Practice Facilitator Certificate Program, a 92-h course with 13 weekly 1.5-h live online classes plus a 40-h fieldwork practicum and 26 h of reflective learning . Twenty-one modules, based on AHRQ’s Practice Facilitation Handbook, cover the full range of topics for trainees to achieve the four core competencies for practice facilitation: data use to drive improvement; interpersonal skills; Health IT optimization; and QI and change management methods (e.g., Plan-Do-Study-Act) . PCIP supplemented the Buffalo program with an additional 20 h of training that focused on increasing knowledge of the ABCS practice guidelines and optimizing the use of the EHR, including creating registries and dashboards, activating clinical decision support systems and other practice change components that correspond to the PF activities in the PCIP QI manual. ABCS toolkits were developed for the practice facilitators to follow that correspond directly to the training content. Three PF managers meet weekly with up to five facilitators each to provide oversight and opportunities for problem solving, identify additional training needs or site specific resources, and review the structured weekly activity reports and monthly narrative reports.
Using a mixed methods approach, we will conduct a process and outcomes evaluation of five interrelated domains: (1) primary outcome (reaching clinical goals for ABCS measures); (2) secondary outcome (change in practice capacity); (3) internal context; (4) external context; and (5) implementation fidelity (PF activities, practice change). We will also assess the cost of the intervention.
Outcome measures and data source
Outcome measures, data source, and data collection timeline
• Aspirin: percent of patients aged 18 years and older with ischemic vascular disease (IVD) with documented use of aspirin or other antithrombotic.
• Blood pressure management:
1) Percent of patients aged 18–85 who had a diagnosis of hypertension (HTN) and whose blood pressure (BP) was adequately controlled (<140/90).
2) Percent of patients aged 60–85 without diabetes (DM) or chronic kidney disease (CKD) with controlled BP (<150/90),
3) Percent aged 18–85 with DM or CKD with controlled BP (<140/90)
• Cholesterol management:
1) Percentage of patients aged >21 diagnosed with ASCVD who are on statin therapy.
2) Percentage of patients aged >21 with history of LDL >190 mg/dL without ASCVD who are on statin therapy.
3) Percentage of patients aged 40–75 with diabetes without ASCVD and LDL 70–189 mg/dL who are on statin therapy
• Smoking cessation support: Percentage of patients aged 18 years or older who were screened about tobacco use one or more times within 24 months and who received cessation counseling intervention if identified as a tobacco user.
Baseline, quarterly during intervention period, at the end of intervention (12 months), and 6 months post intervention
Secondary outcome: practice change capacity
• Organizational change process capacity
• Adaptive reserve
Provider and staff survey
Baseline, at the end of intervention (12 months) and 6 months post intervention
The secondary outcome is changes in primary care practice capacity captured through the use of the change process capacity questionnaire and the 23-item adaptive reserve scale [31, 32]. These measures will be obtained through provider and staff surveys at three time points: baseline, immediately following the 12-month intervention and 6 months post-intervention.
Process evaluation measures and data collection methods
Process measures, data sources and timeline for data collection
Internal practice-level context
• Payer mix
• Patient demographics
• Baseline integration practice transformation components (population management)
• PCMH certification
• Practice survey
• Site visits and key informant (KI) interviews with sample of providers and staff
• Surveys at baseline, 12 and 18 months
• Site visits and KI interviews baseline and 12 months
• Regulatory and financial environment (e.g., pay-for-performance, value based payment)
• Provider surveys
• KI interviews with sample of practice site providers and staff
• PF activities: dose, intensity and mode of delivery of PF intervention Practice change: Extent to which practice change components are implemented and in use
• PF activities: Web-based PF tracking system, Webinar registration and evaluation and collaborative call attendance
• Practice change: practice change survey and site visits/KI interviews
• Ongoing PF tracking using saleforce.com
• Practice change survey monthly
• Implementation and intervention cost
• Cost data collection template embedded in PF tracking system
• Ongoing and reviewed monthly
Internal and external context
These domains will be assessed using both qualitative and quantitative data. The lead clinician or lead practice administrator will be asked to complete a survey (in addition to the survey described previously to assess primary care capacity) that will assess practice characteristics (e.g., use of EHR, FTEs, PCMH status, patient demographics). The survey will be conducted at baseline, 12 and 18 months. In order to obtain a deeper understanding of internal and external context, we will conduct semi-structured interviews with at least one provider and one staff person at 40 sites spread over different waves. We will visit half of these sites early in their implementation of the intervention and again toward the end of the intervention period. The interviews will focus on the practice-specific context, the local and state health policy environment, and experience with practice facilitation. We will visit the other 20 sites one time post-intervention. These sites will be chosen to represent variability in response to the intervention in terms of changes in patient outcomes.
To gain a broader perspective on external health care context and ongoing and emerging practice transformation programs in New York State and nationally, we will conduct semi-structured interviews with 10 key informants who represent leaders in this area from the city and state.
Practice facilitation activities: The primary outcome measure for PF activities is the frequency, duration, and mode of contact. PF activities are measured with data from the web-based software tool (Salesforce.com) that PCIP is currently using to track their practice facilitators’ activities. Each site has a separate page in Salesforce that includes the relevant elements for documenting PF activities and practice change components. Practice facilitators complete documentation in Salesforce after every visit including the time spent at the practice site, who the practice facilitator met with at the sites, what activities were discussed, outcomes of the visits, and agreed upon next steps (e.g., set date for upcoming visit). We will also track provider participation in other components of the PF intervention (e.g., webinars).
Practice change: We will also use Salesforce to assess our targeted practice changes (e.g., use of clinical decision support). Practice facilitators complete a survey, which is embedded in Salesforce, after each visit (i.e., monthly). The survey uses a 6-point scale (not yet educated; educated/PDSA; educated/using; educated/not using/deferred; educated/not using/refused; and not applicable) to assess practice change components that are the target of the intervention and correspond to the main elements of PCIP’s QI manual, which we have mapped to CCM and PCMH [9, 10]. Facilitators will also complete monthly narrative reports that will provide an overview of all implementation activities, including perceptions about what components of the intervention did and did not work and approaches to adaptation of intervention components.
We will apply a practical, cost assessment methodology developed by Ritzwoller to assess the implementation costs of the intervention . This approach provides data on what resources would be needed to implement and replicate the intervention. To determine the implementation and intervention costs, we will collect resource costs based on health professionals’ time requirements (including training), the facilitator’s time, IT costs, and needed supplies or equipment. We will aggregate the data across individual cost capture templates based on the tasks and timeline of the project. Intervention-specific data are defined by asking the question “would this resource be needed to deliver the intervention in practice?” After intervention specific resources are differentiated from research and development, total intervention costs per patient and per practice can be calculated. We will also conduct sensitivity analyses to estimate the range of intervention costs in a variety of settings and circumstances.
where C it is the number of patients in site i who (a) had received aspirin, (b) had controlled blood pressure, (c) had controlled cholesterol, or (d) received smoking cessation counseling during period t, for t ∈ (0, 1, …, 11). Each period is 3 months, and t = 0 is the baseline quarter. I it PF is an indicator variable, and I it PF = 1 if site i has been assigned to the PF intervention at period t, I it PF = 0 otherwise. s is the time period when the PF intervention begins for site i. Z is a site-level indicator variable, Z = 1 if site i has three or more providers, Z = 0 otherwise. E it is the number of patients at site i who were eligible to be counted during period t. The eligible pool of patients E it will be different for each outcome and has been described in detail in Table 1. Log(E it ), considered the “offset” in the Poisson regression model, is a random effect for site i with mean 0 and variance σ b 2 . The estimation via model (1) takes into account a general time trend and allows for the intervention effects to grow over time following implementation of the intervention.
Our primary outcome of interest for each of the four models is proportion of patients who meet clinical goals (described above) at 12 months. In each model, this effect (after taking into account general time trends) would be captured by β 2 + 3β 3. For the four primary outcomes, we will conduct tests with these null and alternative hypotheses: H 0 : β 2 + 3β 3 = 0 vs. H A : β 2 + 3β 3 ≠ 0. The models will be fit using the LME4 package using R software.
Power/effect size analysis
With the pre-specified sample size of 250 sites and design (wedge design with four strata/waves starting 3 months apart), we investigated the intervention effects under each of the primary hypotheses in Aim 1, that are detectable with at least 80 % power of a two-sided significance test with α = 0.05 (Appendix). Conversely, we assessed the power to detect effects that are clinically meaningful and realistic. For each of the four outcomes, we are interested in the proportion of eligible patients in a site who meet guideline recommendations for ABCS clinical care in the fourth quarter of the PF intervention (i.e., month 12 of the intervention) relative to compliance proportions at baseline.
Using conservative (i.e., low) eligibility levels of an average of 50 patients per physician per quarter, we will have over 80 % power to detect a 2.0 % change, assuming a starting compliance rate of 20 % (increasing from 20.0 to 22.0 %). Finally, we have little information to estimate the current levels of variation between sites (σ b 2 ). However, the simulations suggest that power does not decrease when the assumed variance is increased from 0.10 to 0.25 on the log scale. If in fact the variance is lower, closer to 0.01 or 0.05, power is increased substantially to over 95 %.
We will identify baseline organizational characteristics (e.g., site level adaptive reserve, organizational change capacity) that are associated with primary outcomes. This will be accomplished by including separately each organizational characteristic in the main models used under Aim 1 as a main effect and in interactions with the PF indicator. We will conduct a further analysis to explore the possibility that changes in organizational level factors mediate the effect of the intervention on the primary compliance outcomes. Mediation analysis methods in the context of a stepped-wedge design have not been well developed. We will conduct a preliminary analysis to assess if the intervention is associated with changes in site level adaptive reserve, organizational change capacity, and practice changes (assessed with the practice change survey). If there is a detectable effect of the intervention on the organizational level factors, we will assess whether the changes in these organizational level factors are associated with changes in the primary compliance outcomes using the Baron and Kenny “product method” as a starting point of the analysis . If necessary, we will explore more complex structural equation models (SEM) adapted for Poisson outcomes as well as more recently developed causal inference mediation methods to evaluate the mediation.
We will use qualitative methods to conduct a more in-depth assessment of internal and external context including barriers and facilitators (among high and low performers) to implementing practice change and achieving ABCS outcomes. The transcriptions of the interviews and focus groups will be coded using Atlas.ti. The coding scheme will be developed by the evaluation team to focus on key dimensions identified both a priori (i.e., from the interview and focus group protocols) and those that emerge during site visits, interviews, and focus groups. Two coders will independently code at least 10 interviews and all focus groups. Double coding will continue until adequate inter-rater reliability is achieved.
Most of the quantitative data will be analyzed descriptively to provide documentation and description of the practices, the implementation components, fidelity, and context. Using the score calculated from the practice change survey, we will evaluate changes over time in implementation of the practice change components (i.e., one component of implementation fidelity) with repeated measures ANOVA. We will also explore the association between implementation fidelity and primary outcomes (Aim 1 and 2 analyses).
Half of the US adult population has one or more preventable risk factors for cardiovascular disease (CVD) including hypertension (HTN) and hyperlipidemia, but only 10 % are meeting all of their clinical goals . As a result, CVD remains the number one cause of death in the US . A recent analysis by Farley and colleagues concluded that optimal implementation of clinical guidelines for treating HTN and hyperlipidemia could prevent up to 100,000 deaths per year . Yet, less than half of the patients with one or more CVD risk factors are receiving guideline recommended care [42–45].
HHNYC aims to fill this current research-to-practice gap by evaluating the effectiveness of practice facilitation as a quality improvement strategy for implementing evidence-based guidelines for reducing CVD-related risk factors in primary care practices in New York City. The specific focus is to improve adherence to the ABCS, with a long-term goal of creating a robust infrastructure to disseminate and implement patient-centered outcomes research (PCOR) findings in primary care practices and improve practices’ capacity to implement other PCOR findings in the future.
The study findings have the potential to accelerate the tempo of implementation and dissemination of evidence-based strategies for addressing CVD risk factors in primary care settings serving populations with higher risks of CVD-related mortality.
There are several potential methodological challenges that may be encountered during the project. First, obtaining high survey response rates from the providers can be challenging. The project team has extensive experience conducting provider surveys with the target study sites with good results. We will incorporate several strategies that have been effective in previous research in the current study sites including using financial incentives [46, 47]. Second, it is likely that we will observe variation in implementation fidelity. We have defined the core elements of the intervention; however, based on our previous experience and review of the current implementation literature, we acknowledge that adaptations to the unique practice context will be necessary. This is an inherent component of practice facilitation and is meant to enhance adoption and sustainability. We will use rigorous methods to measure implementation fidelity of core elements and will also document adaptations to enhance external validity. Finally, unforeseen changes in the external context of health care delivery may affect implementation outcomes (e.g., changes in Medicaid reimbursement). The practice context in NYC is similar to that nationally, with simultaneous implementation of numerous practice transformation programs and projects. Notable statewide activities include the Medicaid PCMH incentive program to expand primary care capacity and improve care coordination, especially for patients with chronic disease; support of two state-designated regional extension centers to promote adoption of electronic health records (EHRs); the Medicaid and Medicare meaningful use incentive programs; and, recently, the $6 billion Delivery System Reform Incentive Payment Program (DSRIP), which creates new regional collaborations to reduce avoidable hospitalizations among Medicaid patients [48, 49]. Programs and funding related to practice transformation will continue to evolve. Our plan to collect data on external context via key informant interviews with Steering Committee members and pre and post intervention provider surveys will allow us to analyze for any historical threats to internal validity.
ABCS, aspirin, blood pressure, cholesterol, smoking cessation; AHRQ, Agency for Healthcare Research and Quality; CCM, chronic care model; CFIR, Consolidated Framework for Implementation Research; CVD, cardiovascular disease; EBPG, evidence-based practice guidelines; EHR, electronic health records; FTE, full-time equivalent; GLMM, generalized linear mixed model; HHNYC, HealthyHearts New York City; NYC, New York City; NYCDOHMH, New York City Department of Health and Mental Hygiene; NYUSOM, New York University School of Medicine; PCIP, Primary Care Information Project; PCMH, patient-centered medical home; PCOR, patient-centered outcomes research; PDSA, Plan-Do-Study-Act; PF, practice facilitation; QI, quality improvement; SEM, structural equation models
This project is supported by a grant from the Agency for Healthcare Quality and Research. The authors would like to acknowledge the external evaluation center at Oregon Health Sciences University (ESCALATES) and other cooperatives for the intellectual contribution to the final survey tools. We also want to thank Batel Blechter, Kelly Quinn, Stefanie Lindeman, Lily Jordahl, and Allison Pastel for their assistance in providing additional content related to the intervention and outcomes and editing and formatting the final paper.
This research is funded with a grant from the Agency for Healthcare Research and Quality (5 R18 HS023922-01).
The authors have made equal intellectual contributions to the conceptualization, design, and writing of the current protocol. KG was responsible for the analytic plan. All authors read and approved the final manuscript.
The authors declare that they have no competing interests.
Ethics approval and consent to participate
We have obtained IRB approval from the New York University School of Medicine. This publication does not include any individual level data.
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.
- Million Hearts. http://millionhearts.hhs.gov/. Accessed 6 April 2016.
- EvidenceNOW: Advancing Heart Health in Primary Care. http://www.ahrq.gov/professionals/systems/primary-care/evidencenow.html. Accessed 6 April 2016.
- Hing E, Schappert SM. Generalist and specialty physicians: supply and access 2009-2010. NCHS Data Brief. 2012;105:1–8.Google Scholar
- Hing E, Uddin S. Visits to primary care delivery sites—United States, 2008. NCHS Data Brief. 2010;47:1–8.Google Scholar
- Hsiao C, Cherry DK, Beatty PC, Rechtsteiner EA. National Ambulatory Medical Care Survey: 2007 Summary. Hyattsville: National Health Statistics Reports; 2010.Google Scholar
- Bodenheimer T, Hoangmai PH. Primary care: current problems and proposed solutions. Health Affairs. 2010; doi:10.1377/hlthaff.2010.0026Google Scholar
- Baskerville NB, Liddy C, Hogg W. Systematic review and meta-analysis of practice facilitation within primary care settings. Ann Fam Med. 2012;10(1):63–74.View ArticlePubMedPubMed CentralGoogle Scholar
- Knox L, Taylor EF, Geonnotti K, Machta R, Kim J, Nysenbaum J, Parchman, M. Developing and running a primary care practice facilitation program: a how-to guide. Rockville: Agency for Healthcare Research and Quality; 2011.Google Scholar
- Patient Centered Medical Home Resource Center. https://pcmh.ahrq.gov/. Accessed 6 April 2016.
- Chronic care model. http://www.improvingchroniccare.org/index.php?p=The_Chronic_CareModel&s=2. Accessed 6 April 2016.
- Liddy C, Hogg W, Russell G, Wells G, Armstrong CD, Akbari A, Dahrouge S, Taljaard M, Mayo-Bruinsma L, Singh J, Cornett A. Improved delivery of cardiovascular care (IDOCC) through outreach facilitation: study protocol and implementation details of a cluster randomized controlled trial in primary care. Implement Sci. 2011;6(1):6–110.View ArticleGoogle Scholar
- Shojania KG, Grimshaw JM. Evidence-based quality improvement: the state of the science. Health Aff. 2005;24(1):138–50.View ArticleGoogle Scholar
- Grimshaw JM, Shirran L, Thomas R, Mowatt G, Fraser C, Bero L, Harvey E, Oxman A, O’Brien MA. Changing provider behavior: an overview of systematic reviews of interventions. Med Care. 2001;39 Suppl 2:2–45.Google Scholar
- Grol R, Grimshaw J. From best evidence to best practice: effective implementation of change in patients’ care. Lancet. 2003;362(9391):1225–30.View ArticlePubMedGoogle Scholar
- Renders CM, Valk GD, Griffin S, Wagner EH, Eijk JT, Assendelft WJ. Interventions to improve the management of diabetes mellitus in primary care, outpatient and community settings. Cochrane Database Syst Rev. 2001; doi: 10.1002/14651858.CD001481Google Scholar
- Parchman ML, Noel PH, Culler SD, Lanham HJ, Leykum LK, Romero RL, Palmer RF. A randomized trial of practice facilitation to improve the delivery of chronic illness care in primary care: initial and sustained effects. Implement Sci. 2013; doi: 10.1186/1748-5908-8-93Google Scholar
- Goodwin MA, Zyzanski SJ, Zronek S, Ruhe M, Weyer SM, Konrad N, Esola D, Stange KC. A clinical trial of tailored office systems for preventive service delivery. The Study to Enhance Prevention by Understanding Practice (STEP-UP). Am J Prev Med. 2001;21(1):20.View ArticlePubMedGoogle Scholar
- Øvretveit J, Bate P, Cleary P, Cretin S, Gustafson D, McInnes K, McLeod H, Molfenter T, Plsek P, Robert G, Shortell S, Wilson T. Quality collaboratives: lessons from research. Qual Saf Health Care. 2002;11(4):345–51.View ArticlePubMedGoogle Scholar
- Strickland PA, Hudson SV, Piasecki A, Hahn K, Cohen D, Orzano AJ, Parchman ML, Crabtree BF. Features of the chronic care model (CCM) associated with behavioral counseling and diabetes care in community primary care. J Am Board Fam Med. 2010;23(3):295–305.View ArticlePubMedPubMed CentralGoogle Scholar
- Ruhe MC, Weyer SM, Zronek S, Wilkinson A, Wilkinson PS, Stange KC. Facilitating practice change: lessons from the STEP-UP clinical trial. Prev Med. 2005;40(6):729–34.View ArticlePubMedGoogle Scholar
- Solberg LI. If you’ve seen one quality improvement collaborative. Ann Fam Med. 2005;3(3):198–9.View ArticlePubMedPubMed CentralGoogle Scholar
- Grimshaw J, McAuley LM, Bero LA, Grilli R, Oxman A, Ramsay C, Vale L, Zwarenstein, M. Systematic reviews of the effectiveness of quality improvement strategies and programmes. Qual Saf Health Care. 2003;12(4):298–303.View ArticlePubMedPubMed CentralGoogle Scholar
- Primary Care Information Project. http://www1.nyc.gov/site/doh/providers/resources/primary-care-information-project.page. Accessed 6 April 2016.
- The Changing Racial and Ethnic Makeup of New York City Neighborhoods, 2011. http://furmancenter.org/files/publications/The_Changing_Racial_and_Ethnic_Makeup_of_New_York_City_Neighborhoods_11.pdf. Accessed 6 April 2016.
- Karpati A, Kerker B, Mostashari F, Singh T, Hajat A, Thorpe L, Bassett M, Henning K, Frieden T. Health disparities in New York City. New York City: New York City Department of Health and Mental Hygiene; 2004.Google Scholar
- Myers C, Olson C, Kerker B, Thorpe L, Greene C, Farley T. Reducing health disparities in New York City: health disparities in life expectancy and death. New York: New York City Department of Health and Mental Hygiene; 2010.Google Scholar
- New York City Department of Health and Mental Hygiene. https://a816-healthpsi.nyc.gov/epiquery/CHS/CHSXIndex.html. Accessed 22 June 2014.
- Buck MD, Anane S, Taverna J, Stubbs-Dame R, Singer J. The hub population health system: distributed ad hoc queries and alerts. J Am Med Inform Assoc. 2012; Jun;19(e1):e46-50. Epub 2011 Nov 9.Google Scholar
- Damshroeder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC. Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009;4:50.View ArticleGoogle Scholar
- Solberg LI. Improving medical practice: a conceptual framework. Ann Fam Med. 2007;5(3):251.View ArticlePubMedPubMed CentralGoogle Scholar
- Nutting PA, Crabtree BF, Stewart EE, Miller WL, Palmer RF, Stange KC, Jaén CR. Effect of facilitation on practice outcomes in the National Demonstration Project model of the patient-centered medical home. Ann Fam Med. 2010; doi: 10.1370/afm.1119Google Scholar
- Solberg L, Asche SE, Margolis KL, Whitebird RR. Measuring an organization’s ability to manage change: the change process capability questionnaire and its use for improving depression care. Am J Med Qual. 2008;23(3):193–200.View ArticlePubMedGoogle Scholar
- Knox L, Brach C. Practice facilitation handbook: training modules for new facilitators and their trainers. Rockville: Agency for Healthcare Research and Quality; 2013. http://www.ahrq.gov/sites/default/files/publications/files/practicefacilitationhandbook.pdf. Accessed April 6, 2016.Google Scholar
- Bodenheimer T, Ghorob A, Willard-Grace R, Grumbach K. The 10 building blocks of high-performing primary care. Ann Fam Med. 2014;12(2):166–71.View ArticlePubMedPubMed CentralGoogle Scholar
- Practice Facilitator Certificate Program. http://www.ahrq.gov/professionals/prevention-chronic-care/improve/system/pfcasestudies/fillmore.html. Accessed 6 April 2016.
- Parsons A, McCullough C, Wang J, Shih S. Validity of electronic health record-derived quality measurement for performance monitoring. J Am Med Inform Assoc. 2012;19(4):604–9.View ArticlePubMedPubMed CentralGoogle Scholar
- Ritzwoller DP, Sukhanova A, Gaglio B, Glasgow RE. Costing behavioral interventions: a practical guide to enhance translation. Ann Behav Med. 2009;37(2):218–27.View ArticlePubMedGoogle Scholar
- Baron RM, Kenny DA. The moderator-mediator distinction in social psychological research: conceptual, strategic, and statistical considerations. J Pers Soc Psychol. 1986;51:1173–82.View ArticlePubMedGoogle Scholar
- Yang Q, Cogswell ME, Flanders WD, Hong Y, Zhang Z, Loustalot F, Gillespie C, Merritt R, Hu FB. Trends in cardiovascular health metrics and associations with all-cause and CVD mortality among US adults. JAMA. 2012;307(12):1273–83.View ArticlePubMedGoogle Scholar
- Centers for Disease Control and Prevention. Million Hearts: strategies to reduce the prevalence of leading cardiovascular disease risk factors—United States, 2011. MMWR. 2011;60(36):1248–51.Google Scholar
- Farley TA, Dalal MA, Mostashari F, Frieden TR. Deaths preventable in the U.S. by improvements in use of clinical preventive services. Am J Prev Med. 2010;38(6):600–9.View ArticlePubMedGoogle Scholar
- Lenfant C. Shattuck lecture—clinical research to clinical practice—lost in translation? N Engl J Med. 2003;349(9):868–74.View ArticlePubMedGoogle Scholar
- Hajjar I, Kotchen TA. Trends in prevalence, awareness, treatment, and control of hypertension in the United States, 1988-2000. JAMA. 2003;290:199–206.View ArticlePubMedGoogle Scholar
- McGlynn EA, Asch SM, Adams J, Keesey J, et al. The quality of health care delivered to adults in the United States. NEJM. 2003;348:263.View ArticleGoogle Scholar
- Wagner EH, Austin BT, Davis C, Hindmarsh M, Schaefer J, Bonomi A. Improving chronic illness care: translating evidence into action. Health Aff (Millwood). 2001;20(6):64–78.View ArticleGoogle Scholar
- Berry CA, Mijanovich T, Albert S, Winther C, Paul M, Ryan MS, McCullough C, Shih S. Patient centered medical home among small urban practices serving low income and disadvantaged patients. Ann Fam Med. 2013;11 Suppl 1:82–9.View ArticleGoogle Scholar
- Kellerman WE, Herold J. Physician response to surveys: a review of the literature. Am J Prev Med. 2001;20:61–7.View ArticlePubMedGoogle Scholar
- Delivery System Reform Incentive Payment (DSRIP) Program. http://www.health.ny.gov/health_care/medicaid/redesign/dsrip/. Accessed 6 April 2016.
- State Innovation Models(SIM). https://www.health.ny.gov/technology/innovation_plan_initiative/sim.htm. Accessed 6 April 2016.