- Study protocol
- Open Access
- Open Peer Review
This article has Open Peer Review reports available.
MyPreventiveCare: implementation and dissemination of an interactive preventive health record in three practice-based research networks serving disadvantaged patients—a randomized cluster trial
- Alex H Krist1Email author,
- Rebecca A Aycock1,
- Rebecca S Etz1,
- Jennifer E Devoe2, 3, 4,
- Roy T Sabo1, 5,
- Robert Williams6,
- Karen L Stein1,
- Gary Iwamoto7,
- Jon Puro2,
- Jon Deshazo1,
- Paulette Lail Kashiri1,
- Jill Arkind2,
- Crystal Romney6,
- Miria Kano6,
- Christine Nelson2,
- Daniel R Longo1,
- Susan Wolver8 and
- Steven H Woolf1, 9
© Krist et al.; licensee BioMed Central Ltd. 2014
Received: 14 October 2014
Accepted: 24 November 2014
Published: 11 December 2014
Evidence-based preventive services for early detection of cancer and other health conditions offer profound health benefits, yet Americans receive only half of indicated services. Policy initiatives promote the adoption of information technologies to engage patients in care. We developed a theory-driven interactive preventive health record (IPHR) to engage patients in health promotion. The model defines five levels of functionality: (1) collecting patient information, (2) integrating with electronic health records (EHRs), (3) translating information into lay language, (4) providing individualized, guideline-based clinical recommendations, and (5) facilitating patient action. It is hypothesized that personal health records (PHRs) with these higher levels of functionality will inform and activate patients in ways that simpler PHRs cannot. However, realizing this vision requires both technological advances and effective implementation based upon clinician and practice engagement.
We are starting a two-phase, mixed-method trial to evaluate whether the IPHR is scalable across a large number of practices and how its uptake differs for minority and disadvantaged patients. In phase 1, 40 practices from three practice-based research networks will be randomized to add IPHR functionality to their PHR versus continue to use their existing PHR. Throughout the study, we will engage intervention practices to locally tailor IPHR content and learn how to integrate new functions into their practice workflow. In phase 2, the IPHR to all nonintervention practices to observe whether the IPHR can be implemented more broadly (Scalability). Phase 1 will feature an implementation assessment in intervention practices, based on the RE-AIM model, to measure Reach (creation of IPHR accounts by patients), Adoption (practice decision to use the IPHR), Implementation (consistency, fidelity, barriers, and facilitators of use), and Maintenance (sustained use). The incremental effect of the IPHR on receipt of cancer screening tests and shared decision-making compared to traditional PHRs will assess Effectiveness. In phase 2, we will assess similar outcomes as phase 1 except for effectiveness.
This study will yield information about the effectiveness of new health information technologies designed to actively engage patients in their care as well as information about how to effectively implement and disseminate PHRs by engaging clinicians.
Evidence-based preventive services for the early detection of cancer and other health conditions offer profound health benefits, yet Americans receive only half of indicated services . Colorectal, breast, cervical, and prostate cancers are the leading causes of U.S. cancer deaths, claiming 127,860 lives annually -. Though many forms of cancer screening can reduce mortality, a sizable proportion of the eligible population does not receive recommended screening tests. Only 54% of adults are up-to-date with colorectal cancer screening, 74% with breast cancer screening, and 80% with cervical cancer screening . Less than half of adults are up-to-date with clinical preventive services generally ,, and the gap is even more pronounced among low-income Americans and racial and ethnic minorities . A variety of barriers affect patients, clinicians, and health-care systems -. Patients may be unaware that they need screening, lack motivation to be tested, or face logistical challenges. Clinicians may not promote needed tests due to oversight, lack of time, and/or competing demands . The health system is fragmented and often lacks resources, financing, and support to close these gaps. Cancer screening has the added challenge of requiring shared decision-making - and individualized risk assessment  to determine at what age to start or stop screening, how often to rescreen, which test to use (e.g., colonoscopy versus stool blood test), or whether to screen at all (e.g., PSA testing) . Confusion is compounded by inconsistencies among guidelines -, weak evidence supporting some screening tests ,, and public over-enthusiasm for screening .
Some of these problems might be alleviated by health information technology , especially personal health records (PHRs). Some PHRs can offer patients direct access to their electronic health record (EHR) ,, which is empowering, speeds access to past screening dates and results, and enables patients to discover potential inaccuracies in their record. The next generation of PHRs could offer even higher functionality —they could be programmed to apply evidence-based guidelines to assess prevention needs and to incorporate personal data to shape individualized recommendations. The modern information age can enable patients to link prevention guidelines with evidence-based educational resources and decision aids, community services, logistical details, and reminders. Automation has the potential to ease the information burden on clinicians and empower patients with better information. Additionally, while early EHR adopters documented that because of the digital divide, at-risk populations less frequently used PHRs , the digital divide has been steadily closing—particularly for mobile applications . Many leading disparities experts now embrace technology as having potential to reduce disparities in care .
There is, however, a shortage of objective evidence that PHRs can achieve these lofty aims. A minority of patients have a PHR , and most PHRs lack the enhanced functionality described above. The PHRs that connect patients with their medical record often fail to explain content in lay language. Some PHRs offer cancer screening recommendations but rely on simple age- and gender-based logic, ignoring other risk factors. There is little empirical evidence about adoption of highly functional PHRs in typical clinical settings, whether a range of patients will use such systems, what functions patients will use, and how use will be incorporated into practice workflow and patient care.
Improving functionality of patient health records
We propose an innovative solution to the above problems based on a conceptual model that defined five levels of functionality to make information technology patient-centered . These levels include: (1) collecting patient reported information, (2) collecting existing clinical information from EHRs, (3) translating medical information into lay language, (4) providing individualized recommendations by applying information to evidence-based guidelines, and (5) facilitating informed patient action through embedded information resources and tools. We created an innovative application that can be added to existing PHRs, known as an Interactive Preventive Health Record (IPHR), which features these five levels of functionality and is currently integrated into several major EHRs and PHRs (see www.MyPreventiveCare.org) ,-.
In brief, the IPHR addresses 18 clinical preventive services, including screening tests, counseling services, preventive medications, and immunizations recommended by the U.S. Preventive Services Task Force (USPSTF) . Patients are able to access the IPHR through their practice’s PHR, through a single sign-on integration that links the IPHR to the patient’s EHR record—essentially, the IPHR will function as a seamless application within the existing practice PHR. The IPHR extracts hundreds of clinical data elements to individualize preventive recommendations. Patients also complete a health risk assessment to provide information not available in the EHR but necessary for making preventive service recommendations. Based on this information and the USPSTF guidelines, the IPHR applies programmed logic to generate a personally tailored list of recommendations. The interface offers patients’ hyperlinks to detailed personal messages that explain the preventive service and its rationale; relates relevant details in the patient’s history (e.g., prior laboratory test values and dates) to the personalized recommendations; incorporates motivational interviewing content; includes links to evidence-based educational material, decision aids, and local resources based on each patient’s profile; and summarizes the next steps. Message content is modeled after the U.S. Department of Health and Human Services www.Healthfinder.gov and incorporates feedback from patient usability tests, clinician focus groups, and longitudinal patient and clinician advisory boards where stakeholders function as co-investigators to help create and update content . Content can be further tailored to meet individual user’s needs (e.g., low health literacy content, large print, audio, culturally appropriate images, local resources, free or subsidized services, visual displays for mobile devices). After patients use the IPHR, the system automatically forwards a summary to the EHR inbox of the patient’s clinician for appropriate follow-up.
Model for functionalities of a Patient-Centered Health Information System
Level 1 Functionality: patient reported information
Collect information, such as self-reported demographic and risk factor information as well as patient reported outcomes
Level 2 Functionality: existing clinical information
Integrate patient reported information with existing clinical information from electronic health records and/or claims data
Level 3 Functionality: interpretation of information
Interpret information for the patient by translating clinical findings into lay language and delivering health information through a user-friendly interface
Level 4 Functionality: individualization of information
Provide individualized recommendations to the patient, such as screening reminders, based on the patient’s risk profile and on evidence-based guidelines
Level 5 Functionality: patient activation and engagement
Facilitated informed patient action integrated with primary and specialty care through the provision of vetted health information resources, decision aids, risk calculators, personalized motivational messages, and logistical support for appointments and follow-up
Three previous studies have evaluated the IPHR’s feasibility. The first was a randomized controlled trial to test whether mailing patients an invitation to use the IPHR increased delivery of services (R18 HS17046-01, Efficacy Trial, 2007–2010) . Two prospective, observational time-series analyses expanded IPHR implementation to an entire primary care practice population in eight practices (R21 HS018811-01, Adoption Trial, 2010–2012) and to six additional practices in Virginia (RFTO #17 290-07-100113, Implementation Trial, 2009–2011) . These studies culminated in the production of a How-to Guide for Using Patient-Centered Personal Health Records for Prevention, which is being disseminated nationally by the Agency for Healthcare Research and Quality (AHRQ) and which will inform this project’s implementation . Our findings demonstrate that the IPHR is technically feasible, increases delivery of preventive services for patients enrolled in a controlled trial, and can be fielded to an entire primary care population. However, these tests occurred in a small group of Virginia practices with a relatively homogenous patient population.
Through these studies, the IPHR was integrated into three EHRs (EpicCare™, Enterprise™, and Professional™) that represent 31% of the U.S. EHR market share . The IPHR has also been integrated into two commercial PHR platforms (MyChart™ and FollowMyHealth™). Currently, more than 70,000 patients and 190 clinicians are using the IPHR. With this integration, practices successfully incorporated the IPHR into their workflow—using it to prepare patients for visits, augment health behavior counseling discussions, explain test results, issue automatic patient reminders for overdue services, prompt clinicians about services patients need during encounters, and formulate personalized prevention plans. However, considerable variation was documented, from 2% to 60%, in the proportion of clinicians’ patients who used the IPHR.
In preparation for phase 1, we will develop baseline assessments, recruit and randomize practices, integrate the IPHR into the EHRs and PHRs of study sites, assist with practice training, and support local tailoring of IPHR content. Phase 1 (Implementation and Comparative Effectiveness Trial) will include deployment of the IPHR at 20 intervention practices and then contrasting the effectiveness of the IPHR versus traditional PHR at 20 control practices in terms of delivery of recommended screening test and promotion of shared decision-making. Phase 2 (Scalability Assessment) will be assessed by examining the three networks’ ability to extend the IPHR to up to 130 nonintervention practices. At all phases, the study will assess disparities in use and outcomes among disadvantaged patients. For the purpose of this study, disadvantaged patients are defined as minority populations (e.g., African-Americans, Hispanics, and Native Americans) and Medicaid beneficiaries (a surrogate for low income). This study has been approved by the VCU Internal Review Board (IRB HM15307) and contains no more than minimal risk to participants. The risks are limited to breaches of privacy and confidentiality.
Specific aim 1. To test the feasibility and scalability of implementing the IPHR in three diverse practice-based research networks with high proportions of minority and underserved patients.
Sub-aim 1a. The percent of practices within networks approached to participate in the study that agree to and are able to use the IPHR (Adoption).
Sub-aim 1b. The percent of practices’ adult patients who make an office visit, use the IPHR, and receive prevention recommendations during the first (Reach) 6 months and subsequent 6 months (practice-level Maintenance) after adoption.
Sub-aim 1c. The percent of patients who continue to use the IPHR 6 months after initial use (patient-level Maintenance).
Sub-aim 1d. The consistency, variation, and fidelity of IPHR delivery across networks, practices, clinicians, and staff; of practice workflow redesigns; and of reported barriers and facilitators to use (Implementation).
Phase 2—sub-aim 1e (Scalability Assessment). To measure Adoption, Reach, Implementation, and Maintenance of the IPHR when the above implementation is replicated in up to 130 practices.
Increased delivery of indicated preventive services (Hypothesis 1).
Improved shared decision-making for cancer screening (Hypothesis 2).
Specific aim 3 (Disparities Assessment). To assess whether IPHR use (Reach), preventive service delivery rates (Effectiveness), and perceptions of the technology differ for disadvantaged patients, defined as minorities and Medicaid beneficiaries, in either phase 1 or phase 2.
Health system characteristics
Supporting practice-based research network
Virginia Ambulatory Care Outcomes Research Network (ACORN)
Oregon Community Health Information Network (OCHIN)
Research Involving Outpatient Settings Network (RIOS Net)
Number of practices in network
Number of practices eligible for study
Electronic health record
Patient health record
Urban, sub-urban, and rural
Urban, sub-urban, and rural
Unique patients seen annually
Native American (%)
Self pay/indigent (%)
2 / 7
The ACORN network consists of more than 100 primary care practices reflecting the range of primary care in Virginia . The network includes practices from seven health systems; using dozen of EHRs; located in rural, suburban, and urban settings; and with a range of practice organizational and ownership structures. For the purposes of this study, we are focusing on nine practices from the Virginia Commonwealth University Health System (VCUHS), allowing us to integrate the IPHR with one EHR/PHR that is housed on one server. The VCUHS practices are located in inner city Richmond and serve a predominantly African-American population. RIOS Net is a voluntary collaboration of clinicians serving southwest New Mexico’s low-income, medically under-served, and culturally diverse communities . Members include 250 primary care clinicians from academic, Indian health, community health, and private clinics. Similar to ACORN, we will focus on recruiting 21 practices from the University of New Mexico health system that utilize the same EHR/PHR. The OCHIN is a nonprofit Health Center Controlled Network headquartered in Portland Oregon . It is a collaborative of over 70 primary health-care systems that operate more than 200 clinics across the United States. OCHIN members are community health centers (CHCs), including federally qualified health centers, rural sites, and school-based health centers. All OCHIN practices will be eligible for participation given that they use the same EHR/PHR housed on a central server.
Intervention and control conditions
Forty practices will be recruited to participate in phase 1, which will be conducted over the span of 1 year. Eligible practices are those who have had an EHR for at least 6 months, currently have a PHR, and agree to participate. The decision to participate will be made by practices and, by extension, their clinicians. A representative from each practice will provide consent to participate through a memorandum of understanding. Randomization will be done by the biostatistician and research coordinator using a modified blocked design. To preserve allocation concealment, practices will be matched into paired clusters based on their associated health system, percent of Medicaid patients, and percent of practice PHR users and then randomized into the intervention and control conditions on a 1:1 ratio.
Further inform and refine the study design;
Advance and locally tailor IPHR content to better meet users’ needs;
Identify additional PHR functions needed to improve care and decision-making;
Integrate the IPHR into workflow;
Interpret findings; and
Disseminate results locally and nationally.
From a technical perspective, the IPHR will be programmed to integrate with each of the three networks’ EHR/PHR. This will include ensuring that all the required patient data elements for making preventive care recommendations are passed from the EHR to the IPHR, creating links to the IPHR through the practice’s existing PHR, transmitting IPHR clinician summaries and alerts into the practice’s EHR, and locally tailoring IPHR content to meet each practice’s specific resources and needs.
Once technically integrated, all clinicians and staff at intervention sites will be asked to encourage adult patients age 18–75 to use their PHR with its new IPHR features. Three theories will be used to engage the study practices and facilitate cultural and workflow changes necessary to successfully adopt the new IPHR functionality. From Organizational Change Theory, study staff will help practices to: (1) convey a sense of priority for IPHR use, (2) form a guiding coalition, (3) develop and communicate a shared vision, (4) empower practices to act on their vision, (5) plan for short-term wins, and (6) consolidate improvements and institutionalize success ,-. Organizational Change Theory has been successfully employed for implementing a range of practice changes as well as for adopting the use of health information technology (HIT) ,-. In addition, drawing from Complex Adaptive Systems Theory, study staff will encourage practices to adopt a variety of solutions, experimenting with and evolving their implementation strategy according to individual practice needs . Finally, using concepts consistent with Diffusion of Innovations Theory, which identifies features that speed adoption, study staff will locally tailor the IPHR in order to ensure that the IPHR is advantageous to practices, provides a better way to deliver prevention, improves quality of care, improves information in the EHR, and is easy to use ,.
To operationalize these theories, practices will be provided with benchmarking, feedback, practice facilitation, online diaries, and learning collaboratives to set priorities, share experiences, and institutionalize successes. Specifically, three groups of practice champions will be assembled, one in each network. Groups will consist of a clinician, nurse, staff, or office manager from each intervention site. Each group will participate in eight learning collaboratives—four sessions in the 4 months prior to adopting the IPHR and four in the year after. Learning collaboratives will be led by the practice liaisons from the PBRNs and follow an agenda successfully used in two prior AHRQ studies . Throughout the process, study staff and champions will collaborate to determine strategies to facilitate implementation. A second site champion will also post biweekly diary entries via email throughout the project. This will be another method to engage stakeholders for input. The research team will read entries in real time, provide feedback to the champion, and discuss message content at research meetings. This methodology was successfully used to evaluate the implementation of 27 interventions in over 200 practices as part of a 5-year RWJF-funded initiative ,.
Practices randomized to the control condition will deliver “usual” cancer screening and “usual” PHR functionality. The networks currently use commercial EHRs/PHRs with levels 1, 2, and (partial) 4 functionality. Control practices will institute ordinary PHR updates. No practices will be in a control condition in phase 2, which follows completion of the randomized trial.
Twelve data sources will be used to address the questions in our three specific aims: EHR data, IPHR data, PHR data, practice records, network records, practice surveys, champion surveys, learning collaborative observations and transcripts, practice recruitment assessments, practice diaries, patient surveys, and patient phone interviews.
EHR, IPHR, and PHR databases
The EHR, IPHR, and PHR databases will play a central role in assessing all three aims. Every 6 months during phase 1 and annually during phase 2, network IT staff will transfer EHR data to the research database manager for all patients age 18–75 who make an office visit. Throughout the data transfer process, a standard data transfer protocol will be used that allows us to link patients and their practices across all 12 data sources while maximizing the patients’ privacy and confidentiality. EHR data elements will include patient identification number, age, gender, race-ethnicity, diagnostic codes, family history, orders, screening test dates, and test results. IPHR/PHR data elements will include patient identification number and dates of use. The EHR database will capture data for all patients, irrespective of whether they establish an IPHR or PHR account, whereas the IPHR and PHR databases will contain information only about users.
Practice records, network notes, and recruitment assessment
ACORN, OCHIN, and RIOS Net currently collect annual inventories that include basic information about clinicians—e.g., age, gender, degree, years in practice, full-time equivalent—modeled after the Primary Care Network Survey used by AHRQ . Additionally, network PIs and coordinators will complete an assessment when approaching practices to participate in phases 1 and 2 of the study. This information will be used for insights on IPHR adoption and implementation.
Practice surveys will be delivered electronically to all practice administrators, clinicians, and staff at intervention practices. Responses will be used as covariates to evaluate Implementation. Surveys will be collected at baseline and 1 year after adopting the IPHR. Questions will address perceptions of preventive care and shared decision-making, perceptions of health information technology, barriers and facilitators to implementing the IPHR, consistency of IPHR implementation, IPHR adaptation over time, and practice culture. We will use questions from (a) the Information Systems Expectations and Experiences (I-SEE) Survey to assess elements of the Technology Acceptance Model (TAM)—a validated pre- and post-implementation survey associated with increased uptake of HIT;  (b) an assessment of barriers for adopting HIT;  (c) the reciprocal learning scale—an organizational culture characteristic uniquely related to successful HIT implementation and organizational change;  and the Patient-Centered Medical Home Assessment (PCMH-A) that address practice workflow .
A small group of clinicians, nurses, and support staff (2–3 people) from intervention and control practices will complete a baseline workflow and HIT infrastructure assessment. Responses will be used as covariates in our Implementation assessment and responses will be shared with practice champions prior to adopting the IPHR to help inform and guide implementation efforts. The champion survey will include questions from the PCMH-IT and specific questions about preventive care delivery workflow .
To assess outcomes that cannot be determined from EHR data, surveys will be mailed to 4,000 patients, 100 patients from each of the 40 phase 1 intervention and control practices. A sample of 2,000 patients with office visits in the first 6 months of the trial and 2,000 with office visits in the second 6 months of the trial will be randomly selected to receive surveys. Because the purpose of the study centers on cancer screening, the survey sampling frame will be males age 50–75 and females age 40–75, stratified by age, gender, practice, and PHR use (to include a minimum of 50% PHR users for the subgroup analysis). The modified Dillman method will be used to optimize the response rate for the mail survey ,. Surveys will be mailed on practice stationery and in practice envelopes and include a $1 incentive .
As part of the comparative effectiveness evaluation (outcome #2), surveys will assess components of shared decision-making—knowledge, communication, decisional conflict, and locus of control. Surveys will include a separate set of questions for colorectal, breast, cervical, and prostate cancer screening. National Cancer Institute’s (NCI’s) Health Information National Trends Survey (HINTS) will assess knowledge gained by using the IPHR . Questions will address general knowledge about the cancers, screening tests, screening recommendations, and risks and benefits of screening. To assess whether the IPHR helps patients weigh values regarding risks and benefits, measures of both (a) process (patient-clinician communication) and (b) patient perception (decisional conflict) will be used. Questions from AHRQ’s Consumer Assessment of HealthCare Providers and Systems (CAHPS) survey will be used to measure communication. These questions address the quality of the clinician-patient interaction as well as how often clinicians explained cancer screening, listened carefully, provided understandable instructions, knew the patient’s medical history, and spent enough time with the patient . O’Connor’s low literacy decisional conflict scale will be used to measure patient perspective on whether values were weighed -. Decisional conflict assesses the patient’s uncertainty, how well informed they feel, whether they have clarified their values, and whether they feel supported. To assess whether the IPHR fosters decision-making engagement at the desired level, Degner’s locus of control metric will be used to assess patients’ desired and actual level of involvement in recent cancer screening decisions ,. The metric will ask patients to choose from one of five options on Degner’s continuum that best describes the relative role they or their clinicians played in making the cancer screening decision.
Learning collaborative transcripts and observations
The main purpose of the learning collaboratives, which will occur during the phase 1 implementation assessment, will be to operationalize organizational change theory and engage the practices in creating an IPHR implementation strategy. However, the learning collaboratives will also serve as a rich source of qualitative information to understand the practices’ implementation experience in real time. In addition, learning collaboratives will serve as a setting to engage the clinicians in the development IPHR content and implementation strategies. Practices will identify 1–2 champions per site to participate in the learning collaboratives. Accordingly, we will audio record all meetings, and the study facilitator will take meeting notes to supplement recordings.
The online diaries will be used to see practice implementation through the lens of real-time participants. The entries will identify changes to practice approaches and IPHR content/use that could enhance adoption more broadly. Practices will identify 1–2 diary keepers per site. Diary keepers will be trained on how to use the online diaries during locally based and virtual kickoff meetings in year 1. Diary keepers will be asked to make entries biweekly and share observations and experiences regarding IPHR tailoring, use and adaptation; general and IPHR-specific information about patient engagement in preventive care; and issues related to sustained use of the practice PHR and IPHR. Entries will be read in real time and discussed by the VCU project team. Regular responses will be posted to facilitate rich diary interactions.
Patient phone interviews
Interviews with patients will be conducted at the end of phase 1 including 24 IPHR users, 15 who only use the practice PHR, and 9 who do not use the IPHR or the practice PHR. They will be stratified by insurance type (Medicaid versus commercial) and race/ethnicity. The interviews will focus on understanding what made patients use, and continue to use, the IPHR and practice PHR; barriers for engaging online; and effectiveness and perceived value of the IPHR and practice PHR.
Overview of data collection methods and analysis
Aim 1: Implementation: To field the IPHR and evaluate use in terms of:
•IPHR/PHR database to measure which patients use the IPHR or PHR, and when and how often they use it
•Percent of approached practices that agree to use the IPHR (Adoption)
•Percent of patients age 18–75 with a visit who create an IPHR account in months 1–12 (Reach) and 13–36 (practice-level Maintenance) (monthly repeated measures analysis)
•EHR database to measure the number of potential IPHR users (denominator)
•Field notes to gather quantitative and qualitative insights on practice-level Adoption
•Percent of users who use the IPHR after 6 months (patient-level Maintenance)
•Mixed methods analysis of (quantitative) practice and clinician variation in Reach (two-level mixed-effects logistic regression) and (qualitative) consistency, variation, and fidelity of IPHR delivery (immersion/crystallization analysis of transcripts and diaries) (Implementation)
•Network records to measure practice (e.g., size) and clinician characteristics (e.g., age)
•Learning collaborative transcripts, practice survey, practice diaries, and patient interviews to assess IPHR implementation, including consistency and adaptation, and to qualitatively assess Reach, Effectiveness, and Maintenance
Data sources and analysis similar to phase 1 except phase 2 will not include collecting and analyzing learning collaborative transcripts, practice diaries, site visits, or patient interviews
Aim 2: To compare the Effectiveness of the IPHR vs. traditional PHR functions
•EHR database to measure delivery of recommended cancer screening tests
•Percent of patients up-to-date with all indicated cancer screening for all practice patients (intention to treat) and for PHR users (sub-group) (two-level logistic regression)
•IPHR/PHR database to identify users
•Shared decision-making outcomes (knowledge, communication, decisional conflict, and decision control) (three-level generalized mixed-effects regression)
•Patient survey of 4,000 randomly selected patients to measure elements of shared decision-making
•Patient, practice, and clinician facilitators and barriers associated with Effectiveness (mixed-method analysis)
•Learning collaborative transcripts, practice survey, practice diaries, and patient interviews to explore perceptions
Aim 3: Disparities Assessment: Difference in use, effect and perception of technology for disadvantaged populations
•EHR database to identify at risk patients (minorities and Medicaid beneficiaries) and to measure delivery of recommended cancer screening tests
•Comparison of Reach and Effectiveness for the disadvantaged versus general population (two-level mixed-effects logistic regression)
•Patient interviews to understand technology barriers and needs; technology impact; and unique issues for disadvantaged patients
•IPHR/PHR database to stratify levels of use by minority and Medicaid status
Whereas the ultimate benefit of the IPHR is to promote recommended preventive services, a secondary benefit of potentially equal importance is to promote shared decision-making. Showing patients their health information, explaining content in lay language, presenting guideline disagreements, and providing decision aids is intended to engage and activate patients to participate in screening decisions. Accordingly, effectiveness outcomes include both deliveries of preventive services and elements of the decision-making process defined by Sheridan et al. for the USPSTF (the patient [a] understands the condition, [b] understands the service, [c] weighs values regarding risks and benefits, and [d] is engaged in decision-making at the desired level) . Shared decision-making outcomes to be measured will include patient knowledge, clinician-patient communication, decisional conflict, and difference in desired and actual desired locus of decision-making control.
Our statistical calculations will account for up to three sources of variation: variation within subjects measured repeatedly over time, variation among the physicians who see those patients, and variation among the practices where those physicians see their patients. Note that the variability among physicians will be nested within practices. Patient-level data will be collected for Reach, Effectiveness, and Maintenance, while aggregate outcomes will be used at the practice level. Multi-level, mixed-effect models will be used to account for the above variation in the evaluation of our aims. A similar approach will be used for Adoption and Implementation, although the data collected will be practice-level information.
Specific aim 1
Sub-aim 1a. The percent of practices that agree to and are able to use the IPHR (Adoption). Because strategies to field the IPHR will be practice wide, involving a range of staff, Adoption is a practice-level decision. We will use recruitment assessment and network records to calculate descriptive statistics about practices approached, practices willing to use the IPHR, and practices that are able to use the IPHR.
Sub-aim 1b. The percent of patients who use the IPHR in the first ( Reach ) and second and third years (practice-level Maintenance) after adopting the IPHR. We define Reach as the percent of patients age 18–75 who make an office visit during the first 6 months after intervention sites adopt the IPHR (denominator) and who (i) sign up for the IPHR, (iii) complete the intake process, and (iii) receive prevention recommendations (each as separate numerators). IPHR data will provide the numerator, and EHR data will provide the denominator. Practice-level Maintenance applies the same definition for patients seen at intervention sites in months 6 through 24 after IPHR adoption.
Sub-aim 1c. The percent of patients who use the IPHR more than 6 months after initial use. We define patient-level Maintenance as the percent of patients who establish an IPHR account (denominator) and revisit the website at least once 6 months after initially establishing their IPHR account (numerator).
Sub-aim 1d. The consistency, variation, and fidelity of IPHR delivery across networks, practices, clinicians, and staff (Implementation). While it is a practice decision to adopt the IPHR, we expect variation between practices and among clinicians in how they promote and use the IPHR. Quantitatively, we will compare Reach and Effectiveness at the practice-level and clinician-level to quantify variation in implementation at each level, using a two-level mixed-effects logistic regression. From learning collaborative transcripts, practice diaries, and patient interview transcripts, we will use a grounded theory approach and immersion/crystallization techniques to understand how different practices implemented IPHR functions (both initially and over time) and their consistency (versus adaptation) in implementation over time -. Specifically, we will engage project participants in the learning process by sharing emerging insights and having them share in the interpretation of findings. We will use Atlas.ti software to organize, code, and analyze data to identify key findings and themes . The analytic process will include group participation in creating a codebook based on (1) previous experience implementing the IPHR, (2) input and advice from our advisory panel, and (3) key factors that emerge from the data through the immersion-crystallization process of reading and rereading the data, organizing the data around common attributes, and developing an understanding of themes implied by recurrent patterns. Preliminary findings from this analysis will inform phase 2 implementation and scalability.
Sub-aim 1e. Adoption, Reach, Implementation, and Maintenance in the Scalability Assessment. As the networks disseminate the IPHR across all practices during phase 2, we will measure similar outcomes as described in sub-aims 1a–d, although we will not observe learning collaboratives or hold patient interviews.
Specific aim 2
Specific aim 2, hypothesis 1. Compare elements of shared decision-making (knowledge, communication, decisional conflict, and decision control) in intervention versus control practices. Knowledge: From the patient surveys, we will calculate the percent of correct responses from the series of knowledge questions. We will use a three-level mixed-effects logistic regression for the intention-to-treat and subgroup comparisons. Process of weighing values (communication): Per the approach defined by CAHPS, we will score the clinician-patient communication questions on a 6-point scale and calculate both the average scores and “top box” scores (percent reporting the most positive responses) . We will use a linear mixed-effect model for the 6-point scale outcome and will use a generalized linear mixed-effect model for the “top box” outcome, in both the intention-to-treat and subgroup comparisons. Perception of weighing values (decisional conflict): Per protocol, we will score the overall decisional conflict (score 0–100) and the four sub-scores (uncertainty, informed, values, and support) ,. We will use mixed-effect multinomial logistic regression for the intention-to-treat and subgroup comparisons. Engagement at the desired level (locus of decision control): Using a mixed-effect multinomial logistic regression, we will compare the proportion of patients expressing differences in preferred and actual locus of decision-making control for intention-to-treat and subgroup comparisons ,.
Specific aim 2, hypothesis 2. Compare the percent of patients who are up-to-date on all indicated cancer screening tests at intervention and control practices. Our primary Effectiveness outcome for which the study is powered (see below) is the percent of eligible patients who are up-to-date with all indicated cancer screenings (all-or-none measure) . Our prior studies demonstrated that the all-or-none measure was the most sensitive to practice-level changes. However, we also plan to measure the percent of indicated cancer screenings that are up-to-date (composite measure) and the percent of eligible patients who are up-to-date with each individual cancer screening test . Eligibility for cancer screening and “up-to-date” status will be based on the USPSTF recommendations in effect at the time of the office visit. Prostate cancer screening, breast cancer screening age 40–50, and colorectal cancer screening age 75–85 will be analyzed separately but excluded from the composite and all-or-none measures. We will use a three-level mixed-effect logistic regression for the all-or-none, composite, and individual screening test comparisons.
Specific aim 3
Assess whether the use of the IPHR, benefits from the system, and perception of the technology differ for disadvantaged patients. To determine whether exposure to the IPHR affects disparities, we will compare the percentage of minorities and Medicaid beneficiaries versus the general population that use the IPHR (Reach) in both phase 1 (years 2–3) and phase 2 (years 4–5) using a three-level mixed-effect logistic regression. Similarly, we will calculate the difference in delivery of preventive services (Effectiveness), using a three-level mixed-effect logistic regression for the all-or-none, composite, and individual screening tests. Both multi-level analyses will account for practice-level and clinician-level variation. From qualitative material, with an emphasis on the post-implementation patient interviews, we will use a similar grounded theory approach as described in specific aim 1d to understand disadvantaged patients’ perceptions of and experience with the IPHR.
Two power analyses were conducted for the primary (percent of patients up-to-date with recommended cancer screenings) and secondary effectiveness (shared decision-making) outcomes of the study. Data from preliminary studies were used to estimate the anticipated effect size for each outcome. For the primary outcome, it was determined that 40 study sites (20 intervention and 20 control practices) will provide 80% power, alpha = 0.05, to detect an 8% (effect size 0.82) difference in being up-to-date with recommended cancer screening between intervention and control groups in the intention to treat analysis and 80% power, alpha = 0.05, to detect a 15% increase (effect size 0.82) in the subgroup analysis.
For the secondary outcome of the study, it was determined the 100 patients per study site, with stratified sampling to ensure that at least 50% are IPHR users (for the subgroup analysis), will provide 80% power, alpha = 0.05, to detect a 10% knowledge score difference (effect size 0.75–0.80) between IPHR users and nonusers for the intention-to-treat and subgroup analyses, respectively.
The second stage of EHR Meaningful Use will require clinicians to engage 50% of their patients online through PHRs beginning in 2015. It remains to be seen whether clinicians can accomplish this daunting task. It is also unknown whether these efforts will pay off in terms of improved care and health outcomes. We believe that patient-centered PHRs with higher levels of functionality, combined with practice redesign to make use of these functions, can help patients obtain recommended preventive care by linking them to their doctor’s records, explaining information in lay language, displaying tailored recommendations and educational resources, providing logistical support and tools to stimulate action, and generating reminders. Appropriate delivery of these evidence-based services should reduce the burdens of chronic disease and prevent premature death.
This project will measure whether making these resources available to primary care practices and patients promotes shared decision-making and increases the delivery of recommended preventive services compared to existing information systems. Specifically, it will yield needed information about the effectiveness of new health information technologies designed to more actively engage patients in their care as well as information about how to most effectively implement and disseminate PHRs. If the integration of an IPHR into primary care practice is proven effective and scalable, our study findings will have wide ranging practice and policy implications.
We incorporated several important design features to ensure that this study has the potential for maximum impact. For example, the two-phased design allows us to conduct a comparative effectiveness evaluation of current PHR functionality versus embedding more personalized IPHR content followed by a scalability assessment to determine the feasibility and process of broadly integrating IPHR functionality into a wide range of primary care practices. Second, our mixed-method design will provide a robust set of data to understand from multiple perspectives the issues associated with engaging patients in PHR use. This evaluation is proving timely given the recent Stage 2 Meaningful Use requirements. Third, because many of our study practices serve a range of disadvantaged populations, this study will allow us to examine disparities in IPHR and PHR use and how potential disparities in use of these HIT tools might influence outcomes in care. Finally, a core element of our study design is to engage key stakeholders throughout the research project as co-investigators—clinicians, patients, and experts—using a variety of robust methods to ensure that IPHR content meets users’ needs, implementation is integrated into practice workflow, outcomes assessed are meaningful for users, interpretation of findings is guided by users’ experiences, and the voice of our users is incorporated into dissemination activities. Through collaboration with our stakeholders, we hope to ensure our research directions address the prominent concerns of primary care and patients.
Collectively, the above research study design features strike a balance between the fidelity to our interventions that is typically required for internal validity and the need for quality improvement to embrace both individual variation and customization. This National Cancer Institute-funded study is forging the way in striking a balance between these tensions to ensure that our research can be translated to clinical practice in a timely manner.
We recognize that our study has several limitations. While the goal is to improve the delivery of preventive care for all patients, we expect that only a subset of patients will use the IPHR or even the practice’s PHR. We expect that in the future, more and more patients will seek health information online. For the purposes of this study, we are assessing how practices increase their patient’s portal use (Reach). We will also engage patients throughout IPHR refinement to ensure that content is accessible, understandable, and meaningful as well as create mobile ready content that can be accessed on a smartphone. Another study challenge is the broad geographical distribution of our study practices. This makes it more difficult to meaningfully engage local stakeholders, ensure study protocols are carried out as designed, and conduct observational assessments. We plan to use several strategies to mitigate this limitation such as standardizing study materials, using established PBRN infrastructure, building on prior practice relationships, developing methods to promote virtual participation in learning collaboratives (e.g., use of video meetings), and sharing experiences between the three PBRNs.
PHRs hold great potential to improve patient education, promote shared decision-making, facilitate more in-depth conversations, and generally engage patients in their care. However, more is needed from PHRs and from the patients, clinicians, and practices using the PHRs in order to achieve the desired outcomes. This study will evaluate whether making PHRs more patient-centered improves outcomes and it will generate needed evidence about how to engage patients online in primary care.
Funding for this study is provided by the National Cancer Institute (R01CA166375-01A1) and the National Center for Advancing Translational Sciences (UL1TR000058). The opinions expressed in this manuscript are those of the authors and do not necessarily reflect those of the funders.
For more information on the IPHR including implementation please visit: http://healthit.ahrq.gov/sites/default/files/docs/page/PreventiveCareHandbook_062912comp.pdf.
- McGlynn EA, Asch SM, Adams J, Keesey J, Hicks J, DeCristofaro A, Kerr EA: The quality of health care delivered to adults in the United States. N Engl J Med. 2003, 348 (26): 2635-2645. 10.1056/NEJMsa022615.View ArticlePubMedGoogle Scholar
- Jemal A, Siegel R, Xu J, Ward E: Cancer statistics, 2010. CA Cancer J Clin. 2010, 60 (5): 277-300. 10.3322/caac.20073.View ArticlePubMedGoogle Scholar
- U.S. Cancer Statistics Working Group: United States Cancer Statistics: 1999–2007 incidence and mortality web-based report. 2010. http://www.cdc.gov.uscs. Accessed Aug 2010.
- Cancer topics. 2011. http://www.nci.nih.gov/cancertopics. Accessed Aug, 2011.
- Behavioral Risk Factor Surveillance System: Prevalence data. 2010. http://apps.nccd.cdc.gov/brfss/index.asp. Accessed Aug, 2011.
- Bolen J, Adams M, Shenson D: Routine preventive services for older women: a composite measure highlights gaps in delivery. J Womens Health (Larchmt). 2007, 16 (5): 583-593. 10.1089/jwh.2007.CDC5.View ArticleGoogle Scholar
- U.S. Department of Health and Human Services: Healthy People 2010: Understanding and Improving Health. 2000, Government Printing Office, Washington, D.C, 2Google Scholar
- Kottke TE, Brekke ML, Solberg LI: Making “time” for preventive services. Mayo Clin Proc Aug. 1993, 68 (8): 785-791. 10.1016/S0025-6196(12)60638-7.View ArticleGoogle Scholar
- Jaen CR, Stange KC, Nutting PA: Competing demands of primary care: a model for the delivery of clinical preventive services. J Fam Pract. 1994, 38 (2): 166-171.PubMedGoogle Scholar
- Woolf SH, Krist AH, Rothemich SF: Joining Hands: Partnerships Between Physicians and the Community in the Delivery of Preventive Care. 2006, Center for American Progress, Washington D.C.Google Scholar
- Yarnall KS, Pollak KI, Ostbye T, Krause KM, Michener JL: Primary care: is there enough time for prevention?. Am J Public Health. 2003, 93 (4): 635-641. 10.2105/AJPH.93.4.635.View ArticlePubMedPubMed CentralGoogle Scholar
- Sheridan SL, Harris RP, Woolf SH: Shared decision making about screening and chemoprevention. A suggested approach from the U.S. Preventive Services Task Force. Am J Prev Med. 2004, 26 (1): 56-66. 10.1016/j.amepre.2003.09.011.View ArticlePubMedGoogle Scholar
- Braddock CH, Edwards KA, Hasenberg NM, Laidley TL, Levinson W: Informed decision making in outpatient practice: time to get back to basics. Jama. 1999, 282 (24): 2313-2320. 10.1001/jama.282.24.2313.View ArticlePubMedGoogle Scholar
- Edwards A, Elwyn G: Evidence-based patient choice. 2001, Oxford University Press, Oxford, Great BritianGoogle Scholar
- Breast cancer risk assessment tool. 2004. http://bcra.nci.nih.gov/brc/, 2004.
- Prevention and care management. 2011. http://www.ahrq.gov/clinic/prevenix.htm.
- Smith RA, Cokkinides V, Brawley OW: Cancer screening in the United States, 2009: a review of current American cancer society guidelines and issues in cancer screening. CA Cancer J Clin. 2009, 59 (1): 27-41. 10.3322/caac.20008.View ArticlePubMedGoogle Scholar
- Squiers LB, Holden DJ, Dolina SE, Kim AE, Bann CM, Renaud JM: The public’s response to the U.S. Preventive Services Task Force’s 2009 recommendations on mammography screening. Am J Prev Med. 2011, 40 (5): 497-504. 10.1016/j.amepre.2010.12.027.View ArticlePubMedGoogle Scholar
- Levin B, Lieberman DA, McFarland B, Andrews KS, Brooks D, Dash C, Giardiello FM, Glick S, Levin TR, Pickhardt PJ, Rex DK, Smith RA, Thorson A, Bond J, Johnson D, Johnson CD, Winawer SJ: Screening and surveillance for the early detection of colorectal cancer and adenomatous polyps, 2008: a joint guideline from the American Cancer Society, the US Multi-Society Task Force on Colorectal Cancer, and the American College of Radiology.Gastroenterology 2008.,Google Scholar
- Levin B, Smith RA, Feldman GE, Colditz GA, Fletcher RH, Nadel M, Rothenberger DA, Schroy PS, Vernon SW, Wender R: Promoting early detection tests for colorectal carcinoma and adenomatous polyps: a framework for action: the strategic plan of the National Colorectal Cancer Roundtable. Cancer. 2002, 95 (8): 1618-1628. 10.1002/cncr.10890.View ArticlePubMedGoogle Scholar
- Petitti DB, Teutsch SM, Barton MB, Sawaya GF, Ockene JK, DeWitt T: Update on the methods of the U.S. Preventive Services Task Force: insufficient evidence. Ann Intern Med. 2009, 150 (3): 199-205. 10.7326/0003-4819-150-3-200902030-00010.View ArticlePubMedGoogle Scholar
- Guyatt GH, Helfand M, Kunz R: Comparing the USPSTF and GRADE approaches to recommendations. Ann Intern Med. 2009, 151 (5): 363-10.7326/0003-4819-151-5-200909010-00016. author reply 363–364View ArticlePubMedGoogle Scholar
- Schwartz LM, Woloshin S, Fowler FJ, Welch HG: Enthusiasm for cancer screening in the United States. JAMA. 2004, 291 (1): 71-78. 10.1001/jama.291.1.71.View ArticlePubMedGoogle Scholar
- Taplin SH, Rollason D, Camp A, di Donato K, Maggenheimer E: Imagining an electronic medical record for turning cancer screening knowledge into practice.Am J Prev Med38(1):89–97,Google Scholar
- Tang PC, Ash JS, Bates DW, Overhage JM, Sands DZ: Personal health records: definitions, benefits, and strategies for overcoming barriers to adoption. J Am Med Inform Assoc. 2006, 13 (2): 121-126. 10.1197/jamia.M2025.View ArticlePubMedPubMed CentralGoogle Scholar
- Tang PC, Lee TH: Your doctor’s office or the Internet? Two paths to personal health records. N Engl J Med. 2009, 360 (13): 1276-1278. 10.1056/NEJMp0810264.View ArticlePubMedGoogle Scholar
- Krist AH, Woolf SH: A vision for patient-centered health information systems. JAMA. 2011, 305 (3): 300-301. 10.1001/jama.2010.2011.View ArticlePubMedGoogle Scholar
- Chaudhry B, Wang J, Wu S, Maglione M, Mojica W, Roth E, Morton SC, Shekelle PG: Systematic review: impact of health information technology on quality, efficiency, and costs of medical care. Ann Intern Med. 2006, 144 (10): 742-752. 10.7326/0003-4819-144-10-200605160-00125.View ArticlePubMedGoogle Scholar
- Zickuhr K, Smith A: Digital differences. 2012. http://pewinternet.org/Reports/2012/Digital-differences/Main-Report/The-power-of-mobile.aspx. Accessed June, 2012.Google Scholar
- Turner-Lee N, Smedley BD, Miller J: Minorities, mobile broadband and the management of chronic diseases. 2012, Joint Center for Political and Economic Studies, Washington DCGoogle Scholar
- Connecting for Health: Americans overwhelmingly believe electronic personal health records could improve their health. 2008. http://www.connectingforhealth.com/resources/ResearchBrief-200806.pdf. Accessed May, 2009.
- Krist AH, Peele E, Woolf SH, Rothemich SF, Loomis JF, Longo DR, Kuzel AJ: Designing a patient-centered personal health record to promote preventive care. BMC Med Inform Decis Mak. 2011, 11: 73-10.1186/1472-6947-11-73.View ArticlePubMedPubMed CentralGoogle Scholar
- Krist AH: Featured projects. Patients take a bite of prevention apple with web-based interactive personal health records. 2011. http://healthit.ahrq.gov/sites/default/files/docs/page/Krist_Success_Story_Final_7.29.2011.pdf. Accessed Aug, 2011.
- Krist AH, Woolf SH, Rothemich SF, Johnson RE, Peele E, Cunningham T, Longo DR, Bello G, Matzke GR: Randomized trial of an interactive preventive health record to enhance the delivery of recommended care. Ann Fam Med. 2012, 10 (4): 312-319. 10.1370/afm.1383.View ArticlePubMedPubMed CentralGoogle Scholar
- Preventive services. 2010. http://www.ahrq.gov/clinic/uspstfix.htm. Accessed July, 2010.
- HealthFinder.gov. 2014. http://healthfinder.gov/. Accessed February, 2014.
- Skinner CS, Campbell MK, Rimer BK, Curry S, Prochaska JO: How effective is tailored print communication?. Ann Behav Med Fall. 1999, 21 (4): 290-298. 10.1007/BF02895960.View ArticleGoogle Scholar
- Rimer BK: Improving the use of cancer screening for older women. Cancer. 1993, 72 (3 Suppl): 1084-1087. 10.1002/1097-0142(19930801)72:3+<1084::AID-CNCR2820721324>3.0.CO;2-F.View ArticlePubMedGoogle Scholar
- Glanz K, Lewis FM, Rimer B: Health Behavior and Health Education. 1997, Jossey-Bass, San Francisco, 2Google Scholar
- Mullen PD, Simons-Morton DG, Ramirez G, Frankowski RF, Green LW, Mains DA: A meta-analysis of trials evaluating patient education and counseling for three groups of preventive health behaviors. Patient Educ Couns. 1997, 32 (3): 157-173. 10.1016/S0738-3991(97)00037-2.View ArticlePubMedGoogle Scholar
- Solberg LI, Brekke ML, Fazio CJ, Fowles J, Jacobsen DN, Kottke TE, Mosser G, O'Connor PJ, Ohnsorg KA, Rolnick SJ: Lessons from experienced guideline implementers: attend to many factors and use multiple strategies. Jt Comm J Qual Improv. 2000, 26 (4): 171-188.PubMedGoogle Scholar
- Bodenheimer T, Wagner EH, Grumbach K: Improving primary care for patients with chronic illness. JAMA. 2002, 288 (14): 1775-1779. 10.1001/jama.288.14.1775.View ArticlePubMedGoogle Scholar
- Bodenheimer T, Wagner EH, Grumbach K: Improving primary care for patients with chronic illness: the chronic care model, part 2. JAMA. 2002, 288 (15): 1909-1914. 10.1001/jama.288.15.1909.View ArticlePubMedGoogle Scholar
- Wagner EH, Austin BT, Davis C, Hindmarsh M, Schaefer J, Bonomi A: Improving chronic illness care: translating evidence into action. Health Aff (Millwood). 2001, 20 (6): 64-78. 10.1377/hlthaff.20.6.64.View ArticleGoogle Scholar
- Glasgow RE, Orleans CT, Wagner EH: Does the chronic care model serve also as a template for improving prevention?. Milbank Q. 2001, 79 (4): 579-612. 10.1111/1468-0009.00222. iv-vView ArticlePubMedPubMed CentralGoogle Scholar
- Orleans CT, Schoenbach VJ, Wagner EH, Quade D, Salmon MA, Pearson DC, Fiedler J, Porter CQ, Kaplan BH: Self-help quit smoking interventions: effects of self-help materials, social support instructions, and telephone counseling. J Consult Clin Psychol. 1991, 59 (3): 439-448. 10.1037/0022-006X.59.3.439.View ArticlePubMedGoogle Scholar
- Street RL, Gold WR, Manning TE: Health promotion and interactive technology: theoretical applications and future directions. 1997, Lawrence Wrlbaum Associates, London, EnglandGoogle Scholar
- Glasgow RE, McKay HG, Boles SM, Vogt TM: Interactive computer technology, behavioral science, and family practice. J Fam Pract. 1999, 48 (6): 464-470.PubMedGoogle Scholar
- Krist AH, Woolf SH, Bello G, Sabo RT, Longo DR, Kashiri P, Etz RS, Loomis J, Rothemich SF, Peele JE, Cohn J: Engaging primary care patients to use a patient-centered personal health record. Ann Fam Med. 2014, 12: 418-426. 10.1370/afm.1691.View ArticlePubMedPubMed CentralGoogle Scholar
- Department of Family Medicine: Virginia Commonwealth University. A How-To Guide for Using Patient-Centered Personal Health Records to Promote Prevention. 2012. http://healthit.ahrq.gov/KRIST-IPHR-Guide-0612.pdf. Accessed June, 2012.
- American EHR Partners: Market share and top 10 rated ambulatory EHR products by practice size. 2011. http://www.americanehr.com/blog/2011/10/new-ehr-market-share-report-from-americanehr-available/. Accessed June, 2012.
- Reach Effectiveness Adoption Implementation Maintenance (RE-AIM). 2011. www.re-aim.org. Accessed Sept, 2011.
- Glasgow RE, Klesges LM, Dzewaltowski DA, Estabrooks PA, Vogt TM: Evaluating the impact of health promotion programs: using the RE-AIM framework to form summary measures for decision making involving complex issues. Health Educ Res. 2006, 21 (5): 688-694. 10.1093/her/cyl081.View ArticlePubMedGoogle Scholar
- Glasgow RE: RE-AIMing research for application: ways to improve evidence for family medicine. J Am Board Fam Med. 2006, 19 (1): 11-19. 10.3122/jabfm.19.1.11.View ArticlePubMedGoogle Scholar
- Borkan JM: Mixed methods studies: a foundation for primary care research. Ann Fam Med. 2004, 2 (1): 4-6. 10.1370/afm.111.View ArticlePubMedPubMed CentralGoogle Scholar
- ACORN: Ambulatory Care Outcomes Research Network.http://www.acorn.fap.vcu.edu/. Accessed Sept, 2014.,
- Devoe JE, Gold R, Spofford M, Chauvie S, Muench J, Turner A, Likumahuwa S, Nelson C: Developing a network of community health centers with a common electronic health record: description of the Safety Net West Practice-based Research Network (SNW-PBRN). J Am Board Fam Med. 2011, 24 (5): 597-604. 10.3122/jabfm.2011.05.110052.View ArticlePubMedPubMed CentralGoogle Scholar
- DeVoe JE, Likumahuwa S, Eiff MP, Nelson CA, Carroll JE, Hill CN, Gold R, Kullberg PA: Lessons learned and challenges ahead: report from the OCHIN Safety Net West practice-based research network (PBRN). J Am Board Fam Med. 2012, 25 (5): 560-564. 10.3122/jabfm.2012.05.120141.View ArticlePubMedPubMed CentralGoogle Scholar
- Devoe JE, Sears A: The OCHIN community information network: bringing together community health centers, information technology, and data to support a patient-centered medical village. J Am Board Fam Med. 2013, 26 (3): 271-278. 10.3122/jabfm.2013.03.120234.View ArticlePubMedGoogle Scholar
- Research Involving Outpatient Settings Network (RIOSnet).http://fcm.unm.edu/programs/rios-net/index.html. Accessed Sept, 2014.,
- OCHIN.http://www.ochin.org/. Accessed Sept, 2014.,
- Glanz K, Rimer BK: Theory at a glance. A guide for health promotion practice. 2005. 2nd:http://www.sbccimplementationkits.org/demandrmnch//wp-content/uploads/2014/02/Theory-at-a-Glance-%E2%80%93-A-Guide-For-Health-Promotion-Practice.pdf. Accessed May, 2009.Google Scholar
- Porras JI, Robertson PJ: Organization development theory: a typology and evaluation. Research in Organizational Change and Development. Edited by: Woodman RW, Passmore WA. 1987, JAI Press, Greenwich, CNGoogle Scholar
- Fischer LR, Solberg LI, Kottke TE: Quality improvement in primary care clinics. Jt Comm J Qual Improv. 1998, 24 (7): 361-370.PubMedGoogle Scholar
- Solberg LI: Improving medical practice: a conceptual framework. Ann Fam Med. 2007, 5 (3): 251-256. 10.1370/afm.666.View ArticlePubMedPubMed CentralGoogle Scholar
- Koeck C: Time for organisational development in healthcare organisations. Improving quality for patients means changing the organisation. BMJ. 1998, 317 (7168): 1267-1268. 10.1136/bmj.317.7168.1267.View ArticlePubMedPubMed CentralGoogle Scholar
- Senge PM: The Fifth Discipline: The Art & Practice of the Learning Organization. 1990, Currency Doubleday, New York, NYGoogle Scholar
- Kotter J: Winning at change. Leader to Leader, 10, 27–33. 1999. http://leadertoleader.org/leaderbooks/L2L/fall98/kotter.html. Accessed May, 2009.
- Feifer C, Ornstein SM, Jenkins RG, Wessell A, Corley ST, Nemeth LS, Roylance L, Nietert PJ, Liszka H: The logic behind a multimethod intervention to improve adherence to clinical practice guidelines in a nationwide network of primary care practices. Eval Health Prof. 2006, 29 (1): 65-88. 10.1177/0163278705284443.View ArticlePubMedGoogle Scholar
- Ovretveit J, Scott T, Rundall TG, Shortell SM, Brommels M: Implementation of electronic medical records in hospitals: two case studies. Health Policy. 2007, 84 (2–3): 181-190. 10.1016/j.healthpol.2007.05.013.View ArticlePubMedGoogle Scholar
- Creswell JW, Fetters MD, Ivankova NV: Designing a mixed methods study in primary care. Ann Fam Med. 2004, 2 (1): 7-12. 10.1370/afm.104.View ArticlePubMedPubMed CentralGoogle Scholar
- Scott JT, Rundall TG, Vogt TM, Hsu J: Kaiser Permanente’s experience of implementing an electronic medical record: a qualitative study. BMJ. 2005, 331 (7528): 1313-1316. 10.1136/bmj.38638.497477.68.View ArticlePubMedPubMed CentralGoogle Scholar
- Plesk P: Redesigning health care with insights from the science of complex adaptive systems. Crossing the Quality Chasm: A New Health System for the 21st Century. Edited by: Io M. 2001, National Academy Press, Washington, DC, 322-335.Google Scholar
- Rogers EM: Diffusion of Innovations. 2003, Free Press, New York, NY, 5Google Scholar
- Sussman S, Valente TW, Rohrbach LA, Skara S, Pentz MA: Translation in the health professions: converting science into action. Eval Health Prof. 2006, 29 (1): 7-32. 10.1177/0163278705284441.View ArticlePubMedGoogle Scholar
- Etz RS, Cohen DJ, Woolf SH, Holtrop JS, Donahue KE, Isaacson NF, Stange KC, Ferrer RL, Olson AL: Bridging primary care practices and communities to promote healthy behaviors. Am J Prev Med. 2008, 35 (5 Suppl): S390-397. 10.1016/j.amepre.2008.08.008.View ArticlePubMedGoogle Scholar
- Cohen DJ, Leviton LC, Isaacson N, Tallia AF, Crabtree BF: Online diaries for qualitative evaluation: gaining real-time insights. Am J Eval. 2006, 27: 163-184. 10.1177/1098214006288448.View ArticleGoogle Scholar
- Tierney WM, Oppenheimer CC, Hudson BL, Benz J, Finn A, Hickner JM, Lanier D, Gaylin DS: A national survey of primary care practice-based research networks. Ann Fam Med. 2007, 5 (3): 242-250. 10.1370/afm.699.View ArticlePubMedPubMed CentralGoogle Scholar
- Wakefield DS, Halbesleben JR, Ward MM, Qiu Q, Brokel J, Crandall D: Development of a measure of clinical information systems expectations and experiences. Med Care. 2007, 45 (9): 884-890. 10.1097/MLR.0b013e3180653625.View ArticlePubMedGoogle Scholar
- Ajami S, Bagheri-Tadi T: Barriers for Adopting Electronic Health Records (EHRs) by Physicians. Acta informatica medica: AIM: J Soc Med Informatics of Bosnia & Herzegovina: casopis Drustva za medicinsku informatiku BiH. 2013, 21 (2): 129-134.Google Scholar
- Leykum LK, Palmer R, Lanham H, Jordan M, McDaniel RR, Noel PH, Parchman M: Reciprocal learning and chronic care model implementation in primary care: results from a new scale of learning in primary care. BMC Health Serv Res. 2011, 11: 44-10.1186/1472-6963-11-44.View ArticlePubMedPubMed CentralGoogle Scholar
- Patient centered medical home assessment (PCMH-A).http://www.improvingchroniccare.org/downloads/pcmha.pdf. Accessed Sept, 2014.,
- Dillman D: Mail and Internet Surveys: The Total Design Method. 1999, John Wiley and Sons, Hoboken NJ, 2Google Scholar
- Dillman DA: Mail and Internet Surveys: The Tailored Design Method. 1999, John Wiley Company, Hoboken NJ, 2Google Scholar
- Edwards P, Roberts I, Clarke M, DiGuiseppi C, Pratap S, Wentz R, Kwan I: Increasing response rates to postal questionnaires: systematic review. BMJ. 2002, 324 (7347): 1183-10.1136/bmj.324.7347.1183.View ArticlePubMedPubMed CentralGoogle Scholar
- Health Information National Trends Survey. 2007. http://hints.cancer.gov/questions.aspx. Accessed Sept, 2011.
- CAHPS Clinician & Group Survey. 2011. https://cahps.ahrq.gov/Surveys-Guidance/CG/index.html. Accessed Sept, 2011.
- O'Connor AM: Validation of a decisional conflict scale. Med Decis Making. 1995, 15 (1): 25-30. 10.1177/0272989X9501500105.View ArticlePubMedGoogle Scholar
- O'Connor A: Decisional Conflict Scale. 1999, University of Ottawa, Ottawa, 4Google Scholar
- Linder SK, Swank PR, Vernon SW, Mullen PD, Morgan RO, Volk RJ: Validity of a low literacy version of the Decisional Conflict Scale.Patient Educ Couns.,Google Scholar
- Degner LF, Sloan JA: Decision making during serious illness: what role do patients really want to play?. J Clin Epidemiol. 1992, 45 (9): 941-950. 10.1016/0895-4356(92)90110-9.View ArticlePubMedGoogle Scholar
- Degner LF, Sloan JA, Venkatesh P: The control preferences scale. Can J Nurs Res Fall. 1997, 29 (3): 21-43.Google Scholar
- Measuring the use of the RE-AIM framework constructs checklist. 2011. http://www.re-aim.hnfe.vt.edu/resources_and_tools/measures/checklistdimensions.pdf. Accessed Sept, 2011.
- Krueger RA: Focus group: a practical guide for applied research. 2000, Sage Publications, Thousand Oaks, CA, 3Google Scholar
- Krueger RA: Analyzing and Reporting Focus Groups Results, (Focus Group Kit, Vol #6). 1997, Sage Publications, Thousand Oaks, CAGoogle Scholar
- Patton MQ: Qualitative Research and Evaluation Methods. 2001, Sage Publications, Thousand Oaks, CAGoogle Scholar
- Miles MB, Huberman M: Qualitative Data Analysis: An Expanded Sourcebook. 1994, Sage Publications, Newbury Park, CA, 2Google Scholar
- Devers KJ: How will we know “good” qualitative research when we see it? Beginning the dialogue in health services research. Health Serv Res. 1999, 34 (5 Pt 2): 1153-1188.PubMedPubMed CentralGoogle Scholar
- Richards L, Morse JM: README FIRST for a User's Guide to Qualitative Methods. 2007, SAGE Publications, Inc, Thousand Oaks, CA, 2Google Scholar
- Crabtree BF, Miller WL: Doing Qualitative Research. 1999, Sage Oaks, California, 2Google Scholar
- Ryan GW, Bernard HR: Techniques to identify themes. Field Methods. 2003, 15: 85-109. 10.1177/1525822X02239569.View ArticleGoogle Scholar
- Nolan T, Berwick DM: All-or-none measurement raises the bar on performance. JAMA. 2006, 295 (10): 1168-1170. 10.1001/jama.295.10.1168.View ArticlePubMedGoogle Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/4.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly credited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.