Skip to main content

Designing a valid randomized pragmatic primary care implementation trial: the my own health report (MOHR) project



There is a pressing need for greater attention to patient-centered health behavior and psychosocial issues in primary care, and for practical tools, study designs and results of clinical and policy relevance. Our goal is to design a scientifically rigorous and valid pragmatic trial to test whether primary care practices can systematically implement the collection of patient-reported information and provide patients needed advice, goal setting, and counseling in response.


This manuscript reports on the iterative design of the My Own Health Report (MOHR) study, a cluster randomized delayed intervention trial. Nine pairs of diverse primary care practices will be randomized to early or delayed intervention four months later. The intervention consists of fielding the MOHR assessment – addresses 10 domains of health behaviors and psychosocial issues – and subsequent provision of needed counseling and support for patients presenting for wellness or chronic care. As a pragmatic participatory trial, stakeholder groups including practice partners and patients have been engaged throughout the study design to account for local resources and characteristics. Participatory tasks include identifying MOHR assessment content, refining the study design, providing input on outcomes measures, and designing the implementation workflow. Study outcomes include the intervention reach (percent of patients offered and completing the MOHR assessment), effectiveness (patients reporting being asked about topics, setting change goals, and receiving assistance in early versus delayed intervention practices), contextual factors influencing outcomes, and intervention costs.


The MOHR study shows how a participatory design can be used to promote the consistent collection and use of patient-reported health behavior and psychosocial assessments in a broad range of primary care settings. While pragmatic in nature, the study design will allow valid comparisons to answer the posed research question, and findings will be broadly generalizable to a range of primary care settings. Per the pragmatic explanatory continuum indicator summary (PRECIS) framework, the study design is substantially more pragmatic than other published trials. The methods and findings should be of interest to researchers, practitioners, and policy makers attempting to make healthcare more patient-centered and relevant.

Trial registration NCT01825746

Peer Review reports


There is significant interest in transforming health information technology to be more patient-centered [13]. To further integrate these movements toward patient-centeredness and information technology-supported healthcare, one key element that has been missing is a set of brief, practical patient-reported items that are relevant to and actionable by patients and their healthcare team [4, 5]. In 2011, 93 national primary care, public health, health behavior, and psychosocial experts as well as patients engaged in a rigorous three-phase process to identify evidence-based, patient-reported measures that if routinely collected could be used to improve health and monitor health status [6]. The process included assembling all possible candidate measures, using a grid-enabled measures wiki platform to identify key measures for consideration [7], and a town hall meeting to finalize the measures. From this process, 17 brief, feasible screening questions evaluating 10 domains of health behaviors and psychosocial problems were identified [8].

Before these patient-reported measures are fully integrated into health risk assessments or routinely integrated into electronic health records (EHRs) and patient portals, research is needed to evaluate relevance, feasibility, and impact on care [911]. Randomized pragmatic implementation studies are one way to efficiently and rapidly make this evaluation [12]. Key to pragmatic studies is stakeholder involvement, of both clinicians and patients, and outcomes important to potential adoption settings. Table 1 summarizes the key characteristics of pragmatic trials compared to traditional efficacy trials in terms of measures, costs, focus and other dimensions. As shown, there are major differences in the formulation of study questions, methods issues receiving the greatest priority, outcomes and analysis, and level of stakeholder involvement. Despite the differences, if properly designed, both study designs can yield scientifically valid answers to posed research questions. The Consolidated Standards of Reporting Trials (CONSORT) group has also developed a reporting standard and a helpful summary figure, called the Pragmatic Explanatory Continuum Indicator Summary (PRECIS), to report the extent to which an intervention study is pragmatic versus explanatory on 10 dimensions [13, 14].

Table 1 Distinguishing differences between pragmatic and traditional clinical efficacy trials

The purposes of this manuscript are to describe the background, engagement of stakeholders, characteristics of participating primary care practices, study design in relation to the pragmatic explanatory continuum, key outcomes, and automated tool and feedback system developed for the My Health Outcomes Report (MOHR) study.


Overall design

The MOHR study is a paired, cluster (practice level) randomized, non-blinded, practical, implementation study that uses a delayed intervention design with 9 pairs of primary care practices (18 total). The design combines elements of pragmatic trials, implementation science, systems science, and mixed methods approaches, and engages stakeholders as partners throughout the study design, implementation, evaluation and subsequent dissemination. One practice in each pair will be randomized to early implementation, and the second to delayed implementation four months later. Implementation will consist of each practice defining a patient population and workflow and then fielding the systematic collection of the patient reported measures in daily care. The primary outcomes will be implementation measures and costs that are relevant to potential adopting and funding agencies [1519]. Patient goal-setting and clinical follow-up in the 10 health behavior and psychosocial domains will be compared between early and delayed intervention sites prior to implementation of the MOHR assessment in delayed intervention sites, thus delayed intervention practices effectively serve as control practices. Feasibility will be assessed by monitoring implementation in both early and delayed implementation sites (see Figure 1 for project overview and timeline). The project is funded through a unique partnership of grant supplementation from the National Cancer Institute (NCI), the Agency for Healthcare Research and Quality (AHRQ), and the Office of Behavioral Social Sciences Research (OBSSR).

Figure 1
figure 1

MOHR study overview and timeline. The study started in January 2013. The approximate implementation date for the early implementation practices is March 2013 and July 2013 for the delayed implementation practices. (IRB = institutional review board, EIS = early implementation sites, DIS = delayed implementation sites).

Research teams and study settings

To carry out this pragmatic trial, partnerships were created between local research teams at eight nationally distributed academic institutions that manage a practice-based research network (PBRN) or participate in the Cancer Prevention and Control Research Network (CPCRN) [20, 21]. Given the network affiliations, each research team has expertise in partnering with their practice and patient stakeholders. Accordingly, research teams are charged with helping their partner practices implement the patient-reported assessment and carry out the MOHR study protocol, while balancing the need for adaptation to ensure implementation is relevant to local culture and practice flow (see ‘Balancing fidelity and adaption,’ below) [22]. To coordinate activities across the project, a MOHR planning committee composed of representatives from all local research teams and practices will have joint bimonthly meetings for planning (pre-implementation) and to share successes and challenges (post-implementation). The organization and coordination of activities is depicted in Figure 2.

Figure 2
figure 2

MOHR study organization and coordination. The four general partners include funders, project coordination, local research teams, and study practices. The funders provide project coordinators general input into the study implementation and design. Overall project coordination includes the five working groups, MOHR planning committee, and the coordination center. The local research teams directly coordinate with the nine practice pairs to carry out and support the MOHR implementation and evaluation. (ACORN = Virginia Ambulatory Care Outcomes Research Network, UV = University of Vermont, UNC = University of North Carolina, UCLA = University of California, Los Angeles, and UTH = University of Texas Houston).

Research teams identified one or two pairs of primary care practices within their network that were similar with respect to practice type (e.g., FQHC or PBRN, family practice or internal medicine), practice ownership, geographic region, EHR infrastructure, and patient population served. Practice pairs were purposefully selected to represent the diversity of primary care settings and populations to ensure greater generalizability of results. Practices were randomly allocated to the early or delayed condition by coin toss. To conceal allocation, the researchers randomizing practices (i.e., project coordination members) were blinded to practice name; practices were labeled by number in paired blocks (by local research teams), and identity was revealed to all participants one week later during a joint call.

Table 2 presents the characteristics of the 18 participating practices. The PBRN practices tend to have higher proportions of insured patients and fewer ethnic minorities compared to the FQHCs. Most practices are small to medium in size, with one to six providers and seeing 1,500 to 10,000 unique adult patients annually. All practices, except two, have EHRs. While nearly half of practices have initiated or received patient-centered medical home designations, most report limited experience with behavioral health and varying levels of experience with quality improvement.

Table 2 Participating practice characteristics

Work groups

To facilitate rapid, collaborative decision making on important issues relevant to the MOHR study design and implementation, work groups were formed consisting of research team members, clinicians, funding agency representatives, and content experts. Each work group is responsible for effectively solving issues related to their specific MOHR study area and bringing proposals to the larger collaborative for approval. Work groups include: Web Tool and Practice Training Work Group – to design and test an electronic version of the comprehensive behavioral and psychosocial risk assessment,, for practices to use and to create training sessions to prepare practices to address topics identified by the MOHR assessment; Patient Experience Survey Work Group – to design the Patient Experience Survey (see ‘Outcomes assessment’ , below) for assessing main patient-level outcomes; Clinical Context Work Group – to identify and evaluate how key practice, patient, and environmental contextual issues impact implementation; Cost Collection and Analyses Work Group – to measure the practice costs of implementing the MOHR assessment; and Publication, Presentation, and Relations Work Group – to promote opportunities for participants to present and publish findings, coordinate with relevant stakeholder entities, and create a repository of materials, tools, and disseminated findings.


The intervention being studied is the systematic implementation and fielding of the MOHR health behavior and psychosocial assessment, including the delivery of brief customized patient and provider feedback reports. Each local research team worked with their practice pair champions and key stakeholders to determine the most effective workflow for administering the MOHR assessment, including: (a) which patients would be invited to complete the assessment, (b) when the assessment would be completed (e.g., two weeks before a visit, immediately before a visit), (c) where the assessment would be completed (home or practice), (d) whether the assessment would be electronic or paper based, and (e) how clinicians would receive MOHR feedback summaries within the context of office visits. This is consistent with the original NIH consensus panel’s goal of identifying and standardizing patient reported elements to be incorporated into EHRs while minimizing staff burden and enhancing consistency of implementation. As a result, the practice pairs were encouraged to field the electronic version of the MOHR assessment if possible [8].

Because it was not feasible to fully integrate the MOHR assessment into each of the diverse practice’s EHR and patient portal, a stand alone, publically available website ( was created. The MOHR website is partially integrated with the study practices’ existing EHRs and administers the 17 health behavior and psychosocial screening questions (including basic demographics) as defined by the NIH consensus panel [6, 8]. In response to positive depression, anxiety, alcohol, and drug screening questions, the website seamlessly prompts the patient to complete more in-depth assessments, including the Patient Health Questionnaire (PHQ9) [23], the Generalized Anxiety Disorder (GAD) questionnaire [24, 25], the Alcohol Use Disorders Identification Test (AUDIT-C) [26], and the Drug Abuse Screening Test (DAST-10) [27], respectively. The MOHR website scores the patient’s responses and characterizes them as being of ‘no concern’ , ‘some concern’, or ‘high concern’ based on norms and national guidelines. For measures with ‘some’ or ‘high concern’ , patients are asked to select which topics they are ready to change and/or discuss with their doctor and select the one topic they feel is most important to address [2831].

After completing this process, patients are taken to a feedback page to review, download and print (see Figure 3). Health domains in which they are doing well are highlighted and reinforced, initial recommendations for changes and improvements are given for domains where improvements are needed, and space is provided to list any questions, decide on and create up to three SMART goals (specific, measurable, achievable, realistic and timely), and outline follow-up steps to help achieve their goals [32, 33]. When patients have finished reviewing their feedback page and leave the website, a clinician summary is also sent to the patient’s primary care team.

Figure 3
figure 3

Patient summary and feedback from . The first page demonstrates the patient’s health behavior and psychosocial scores, level of concern, and whether the patient reported readiness to change and interest in talking with their doctor. The second page includes patient workspace for notes, creation of SMART health goals, and a follow-up plan.

Practice implementation preparation

Practices are being prepared to implement the MOHR assessment through a series of calls and meetings with the site champions, an educational webinar with an interactive ‘call in’ and case scenarios, support materials, and on-site consultation for follow-up. The webinars have been designed using the principles of adult learning, including multiple modalities, reinforcement, modeling and observation. The follow-up consultations have been modeled on academic detailing that entails a brief face-to-face intervention with the clinician, repeated at periodic intervals [3436]. Academic detailing has been found effective in improving other aspects of clinical care such as cancer screening and chronic disease management [3743].

The webinars present the evidence supporting the MOHR assessment, strategies for successful practice implementation of risk assessments, and approaches to counseling and supporting patients as they address identified problem behaviors and psychosocial issues. The webinars, as academic detailing, distill the evidence into brief, practical messages that are relevant to and actionable for primary care clinicians. The webinar and local research team consultations seek to change staff attitudes and beliefs toward implementing the MOHR assessment though persuasive communications, including case examples, tailored feedback and reinforcement [44]. Feedback and reinforcement will include weekly practice reports on the number of patients completing the MOHR assessment and bimonthly meetings between practice champions and local research teams to share cross project implementation experiences and review the local implementation process.

Data sources

Outcomes will be collected from three sources: practice appointment records, the Patient Experience Survey, and patient responses to the MOHR assessment. Appointment records will be used for calculating Reach (see ‘Primary outcomes’, below). Reach measures are collected when practices enter the intervention phase, months two to six for early intervention practices and months seven to nine for delayed intervention practices. Appointment records are queried by the practices (electronically or manually) for patients who meet inclusion criteria, are documented, de-identified, and transferred to a central data coordination center every two weeks. The Patient Experiences Survey will be used to compare the early versus delayed intervention practices on patient reports of patient-provider discussions, recommendations for referrals, and collaborative goal-setting across the 10 health domains. The patient responses to the MOHR assessment will be used for calculating patient health behavior and psychosocial changes over time. This will be assessed by repeated administration of the MOHR assessment for the cohort of patients who complete the assessment on the website and provide email addresses (see ‘Outcomes assessment’ , below).

The content of the Patient Experience Survey and fielding protocol were developed by the working group, building on earlier pilot phase experiences and practice input, with the intention of balancing scientific rigor and patient needs, practice capabilities, and resource availability across the diverse settings. The survey questions were standardized. The 5As framework (Assess, Advise, Agree, Assist and Arrange) served as a guide for developing questions for each health domain [45]. Questions from the Consumer Assessment of Healthcare Providers and Systems (CAHPS) assessing patient trust in their healthcare team and perceived provider communication style were also included [46]. Practices will mail the Patient Experience Survey to patients that the early and delayed intervention practices defined as eligible to receive the MOHR assessment two weeks after the patient’s office visit. To enhance survey response rates, surveys will be mailed using a modified Dillman technique in the practice’s envelope with a cover letter from the patient’s provider [47, 48]. Study information will be included with the survey, and completion of the survey will be deemed as consent for participating in the survey – our Institutional Review Board approved a waiver of written consent.

Outcomes assessment

The primary outcomes for this study are implementation measures [15]. The measures include the practices’ implementation Reach and Effectiveness [49]. Reach measures include the percent and characteristics of patients that: are defined as eligible to be invited to complete the MOHR assessment; are invited to complete the MOHR assessment; and complete the MOHR assessment. The primary Effectiveness measures are a comparison of early versus delayed intervention practice patient report on whether they: were asked about the health behavior and psychosocial domains; set a goal or action plan with their clinician; and arranged any follow-up contact regarding their plan. As early intervention practices will have implemented the MOHR assessment and delayed intervention practices will not yet have implemented the MOHR assessment, they will respectively serve as intervention and control conditions. The impact on health behaviors and psychosocial problems will be assessed by repeated administration of the MOHR assessment for the subset of patients providing an email address. Patient responses will be compared at baseline and four months after baseline. Multivariate regressions will assess for influences of practice implementation characteristics (e.g., electronic versus paper) and patient demographics on each outcome. The overall CONSORT diagram depicting study participation and data capture are depicted in Figure 4.

Figure 4
figure 4

MOHR study CONSORT flow diagram.

Power and sample size analysis

To determine power, we focused on the patient report as to whether any practice personnel worked with her/him to set specific goals for change for the 10 behavior and mental health topics. For the base power analysis, we assumed 9 matched pairs of sites, 300 patients mailed the Patient Experience Survey from each site, and a 50% response rate (2,700 total respondents). Imputation analyses will be conducted to evaluate the impact of missing data, but here we assume the loss of up to 50% of the sample for power calculation purposes. For the first power analysis, we focus on the binary event that at least one goal was established (as opposed to none). Assuming a delayed intervention response rate of 0.3, there is 90% power to detect a 17.5% improvement in the early versus delayed intervention group, and 80% power to detect a 15% improvement.

For the second power analyses, we focus on the continuous outcome of the number of goals established using a hierarchical analysis of covariance. We adjust for any baseline differences in variables related to the outcome as well as clustering effects of patients within clinic. Assuming an intraclass correlation coefficient (ICC) of 0.03, there is 90% power to detect a between conditions effect size of 0.47 with 9 sites per treatment and a 50% patient response rate, and 80% power to detect an effect size of 0.40. Assuming a larger ICC of 0.05, we have 80% power to detect an effect size of 0.51 and 90% power to detect a difference of 0.60, given the assumptions above. All power analyses were performed using the ‘Optimal Design Software’ package (version 3.0).

Contextual assessments

An important aspect of a pragmatic trial is that it is implemented in ‘real-world’ contexts, making it essential that these contexts be described to understand differential outcomes [50]. Accordingly, local research teams will collect data on multiple levels of contextual factors [5153]. Levels will include national/state/local, community, healthcare system, practice and patient [54]. Contextual factors will include practice culture and staffing, patient panel characteristics, community characteristics and resources, and organizational features [55]. Local research teams will collect relevant data through structured interviews with clinicians and through field observations of clinical, social and organizational processes and interactions as practices field the MOHR assessment at the baseline, middle and end of the intervention to assess changes in context. Quantifiable characteristics will be incorporated from initial practice descriptions made during recruitment. Teams will use a standardized template to systematically elicit and record contextual data.

Cost assessment

The MOHR cost assessment will evaluate both the implementation costs and the replication costs using cost assessment procedures that are minimally intrusive and burdensome [19]. Using standardized a procedure, local research teams will collect each practice’s cost estimate following brief training. This person will use a combination of procedures, including observation of the intervention and brief standardized structured questions of key practice implementation staff. Cost data will be collected at two time points – early in the implementation and once a steady state has been reached. Calculations will be made using established expenditure methods, and replication costs will be categorized based upon different practice and implementation characteristics [5659]. The cost assessments will be conducted from the practice perspective and will include costs of recruitment, promotion, training, supervision, implementation and follow-up. Data collected during the trial will be used to estimate the cost of a practice replicating the MOHR assessment setting under different conditions and assumptions. Research costs, downstream costs of increased or reduced healthcare utilization, and costs from the patient perspective will not be considered.

Balancing fidelity and adaption

To maintain internal validity, certain components of the intervention were standardized. All participating practices will implement these components, and fidelity to implementation will be assessed. However, to reflect the pragmatic nature of the trial, practices were encouraged to adapt aspects of the intervention to their workflow and setting [22]. This is necessary because the MOHR study purposively includes practices that vary widely with respect to patients (mostly minority versus non-minority), location (rural versus urban), and EHR and patient portal use. Therefore, to ensure successful implementations as well as high assessment and survey response rates, each practice pair considered its patients’ needs. The elements that needed to be standardized were decided by the overall project steering committee and working groups, while decisions about local tailoring were made by the practice pairs with support and direction from the local research team.

Standard study elements include the MOHR health behavior and psychosocial assessment questions, provision of patient and provider feedback summaries based on the MOHR assessment, goal-setting materials and brief training to support clinicians with counseling, and Patient Experience Survey questions. Practice pairs are collaboratively deciding which patients to target for the MOHR assessment, while attempting to include the largest percentage of adult patients that is feasible; timing of the MOHR assessment (one to two weeks prior to a visit at the patient’s home or immediately prior to a visit at the practice); mode of the MOHR assessment (website or paper), delivery of clinician summary (integrated into EHR, faxed, printed); and mode of administering the Patient Experience Survey (postal, web-based, telephone or a combination).

Trial status

The MOHR trial began January 2013 and is expected to complete data collection in November 2013. The trial is proceeding according to the timeline outlined in Figure 1. Currently, the early implementation clinics are in the early stages of fielding the MOHR assessment to their patients. The delayed intervention clinics will begin implementation between July and September 2013.


This manuscript describes the need and plans for an important ongoing pragmatic study of the administration and use of patient-reported items in primary care, illustrating key features of implementation design with both standardized and flexible aspects reflecting the dual realities of research and practice. A central tenant throughout the process has been ensuring the relevance of the intervention, measures, and methodologies to stakeholders [10, 12].

Figure 5 summarizes the characteristics of the MOHR study on the 10 CONSORT PRECIS dimensions. Overall, the MOHR study is highly pragmatic, receiving completely pragmatic scores on three dimensions (intervention flexibility, control flexibility, and practitioner expertise) and the next most pragmatic rating on the seven remaining dimensions. These scores are substantially more pragmatic than other published trials [60].

Figure 5
figure 5

Depiction of the MOHR study PRECIS characteristics. The MOHR study design receives completely pragmatic scores for intervention flexibility, control flexibility, and practitioner expertise. The MOHR study design receives the next most pragmatic rating on the remaining seven dimensions.

Implementation science involves controlled research, in this case a cluster randomized delayed intervention trial, but that is pragmatic and thus more directly relevant to real-world settings [9, 61, 62]. By engaging all of the stakeholders on the front-end about the importance of healthcare process and outcome metrics for documenting healthcare quality, the results should be more directly usable by both primary care and potential funding and reimbursement agencies. This is especially so when including practical measures that can be embedded into clinical systems [63], applying a feasible and flexible intervention that provides direct feedback based on these practical measures, and assessing outcomes of importance to patient, provider, and policy maker stakeholders, including costs.

This type of science is needed because ‘evidence alone is not enough’ [64, 65]. In addition to the best evidence, as in this case recommendations for primary care-based health behavior and psychosocial counseling, other elements are needed [66, 67]. As noted in the Evidence Integration Triangle framework, other necessary components for successful implementation are practical measures of progress that can be tracked longitudinally (in MOHR, these are the 17 items), and partnership engagement processes, as illustrated above [64].

A major advantage of the MOHR trial for advancing both scientific and practice frontiers is its diversity of settings. Participating primary care settings include practices that already have adopted state-of-the art e-health technologies for tracking patients’ care, enabling an examination of the merger of patient-centered assessment and intervention approaches and EHRs in high resourced settings [68, 69]. On the other end of the continuum, inclusion of a number of community health centers, typically with less access to state-of-the art technologies, has generated problem-solving about the best strategies for conducting assessment and interventions in contexts where staff are challenged to see large caseloads of more diverse, low income/low educated populations with characteristically complex medical and psychosocial needs, and lower levels of health literacy and numeracy [70]. The mantra remains consistent: involving relevant stakeholders is important for both the initial adoption as well as the long-term sustainability of evidence-based patient assessment and intervention. This requires an understanding of the realities of the practice flow, an appreciation for the context of patient lives, and how to design practical assessments and feedback mechanisms that can be implemented with minimal resources.

From a traditional perspective, there are several limitations or weaknesses to the MOHR design. Scientifically, there are several questions the trial will be unable to address. The MOHR assessment will not be linked at the individual level to the follow-up Patient Experiences Survey, and the design selected precludes the examination of the long-term behavioral or biological health impact of our intervention. In contrast to more traditional research studies, the project will not employ expert, highly trained staff in leading academic medical centers to deliver the intervention; rather the intervention will be implemented within real-world settings with tailoring of strategies to meet the needs of each site. Similarly, a decision was made to be as inclusive as possible in the patient intake rather than using strict eligibility criteria selecting for non-complicated, highly motivated patients who often fare better in health promotion programs. Finally, practice participation is not limited to practices with excellent state of the art EHRs serving technology savvy patients; thus, a difficult balance has been sought between maximizing the use of technology whenever possible and developing protocols applicable to a range of settings. All of these decisions could reduce the initial effect size, or produce different results across different contexts, subgroups or settings. However, despite this possibility, we suggest that the project remains a worthwhile investment that will provide invaluable information about ‘what works under what conditions for what groups for what outcomes’— in a relatively short time frame [71].

The study design has many strengths, including the highly pragmatic nature of the protocol, as evidenced by the PRECIS ratings. Given the stakeholder involvement throughout, the results should be highly relevant and sensitive to both the primary care context as well as to the growing emphasis on patient-centered care for improving population health and well being. Anticipating the relevance of the results to diverse primary care practitioners as well as different healthcare settings and networks, the rapid and iterative processes using mixed methods should provide in-depth feedback to more quickly inform practice and policy. Inclusion of data on reach, implementation under different conditions, as well as implementation and replication cost estimates should provide the types of transparent information needed by potential adopting settings. The tools and materials used in the MOHR project, including the standardized brief assessment items, the MOHR automated assessment and feedback tool, and the assessment procedures are all in the public domain and available for others to use or adapt ( Replications and extensions are encouraged, and especially studies having the scope and funding to evaluate issues such as longer-term sustainability, health behavior changes, and health outcomes.

The MOHR study serves as an example of a pragmatic, participatory design that leverages resources and tailors design elements to a range of local settings, while maintaining the essential elements and scientific rigor of a randomized comparison trial. This design is necessary, given that the MOHR study aims to assess whether it is feasible for a broad range of primary care practices to systematically patient reported health behavior and psychosocial assessments. Outcomes will include all necessary implementation elements including feasibility, impact on care, the influence of setting and context, and the expense of implementation. These findings should be of interest to researchers, practitioners, and policy makers attempting to make healthcare more patient-centered and relevant.


  1. Krist AH, Woolf SH: A vision for patient-centered health information systems. JAMA. 2011, 305 (3): 300-301. 10.1001/jama.2010.2011.

    Article  CAS  PubMed  Google Scholar 

  2. Finkelstein J, Knight A, Marinopoulos S, Gibbons C, Berger Z, Aboumatar H, Wilson R, Lau B, Sharma R, Bass E: Enabling patient-centered care through health information technology. 2012, Bethesda MD: Agency for Healthcare Research and Quality (US), (Evidence Reports/Technology Assessments, No. 206). Accessed Feb, 2013

    Google Scholar 

  3. Agency for Healthcare Research and Quality: Ambulatory safety and quality program: health IT portfolio. Enabling patient-centered care through health IT. 2013, Bethesda MD: Agency for Healthcare Research and Quality, Accessed Feb, 2013

    Google Scholar 

  4. Glasgow RE, Kaplan RM, Ockene JK, Fisher EB, Emmons KM: Patient-reported measures of psychosocial issues and health behavior should be added to electronic health records. Health Aff (Millwood). 2012, 31 (3): 497-504. 10.1377/hlthaff.2010.1295.

    Article  Google Scholar 

  5. Rabin BA, Purcell P, Naveed S, Moser RP, Henton MD, Proctor EK, Brownson RC, Glasgow RE: Advancing the application, quality and harmonization of implementation science measures. Implementation science: IS. 2012, 7: 119-10.1186/1748-5908-7-119.

    Article  PubMed  PubMed Central  Google Scholar 

  6. Department of Health and Human Services, National Institute of Health, Office of Behavioral and Social Sciences Research, National Cancer Institute: Identifying core behavioral and psychosocial data elements for the electronic health record. 2011, Bethesda MD: Department of Health and Human Services, National Institute of Health, Office of Behavioral and Social Sciences Research, National Cancer Institute, Accessed Jan, 2013

    Google Scholar 

  7. National Cancer Institute: Grid-enabled measures database. 2011, Bethesda MD: National Cancer Institute, Accessed Jan, 2013

    Google Scholar 

  8. Estabrooks PA, Boyle M, Emmons KM, Glasgow RE, Hesse BW, Kaplan RM, Krist AH, Moser RP, Taylor MV: Harmonized patient-reported data elements in the electronic health record: supporting meaningful use by primary care action on health behaviors and key psychosocial factors. J Am Med Inform Assoc. 2012, 19 (4): 575-582. 10.1136/amiajnl-2011-000576.

    Article  PubMed  PubMed Central  Google Scholar 

  9. Brownson RC: Dissemination and implementation research in health: translating science to practice. 2012, New York: Oxford University Press, 1

    Book  Google Scholar 

  10. Glasgow RE: What does it mean to be pragmatic? Pragmatic methods, measures and models to facilitate research translation. Health Educ Behav. 2013, In press

    Google Scholar 

  11. Glasgow RE, Chambers D: Developing robust, sustainable, implementation systems using rigorous, rapid and relevant science. Clin Transl Sci. 2012, 5 (1): 48-55. 10.1111/j.1752-8062.2011.00383.x.

    Article  PubMed  Google Scholar 

  12. Tunis SR, Stryer DB, Clancy CM: Practical clinical trials: increasing the value of clinical research for decision making in clinical and health policy. JAMA. 2003, 290 (12): 1624-1632. 10.1001/jama.290.12.1624.

    Article  CAS  PubMed  Google Scholar 

  13. Zwarenstein M, Treweek S: What kind of randomized trials do we need?. CMAJ. 2009, 180 (10): 998-1000. 10.1503/cmaj.082007.

    Article  PubMed  PubMed Central  Google Scholar 

  14. Thorpe KE, Zwarenstein M, Oxman AD, Treweek S, Furberg CD, Altman DG, Tunis S, Bergel E, Harvey I, Magid DJ, Chalkidou K: A pragmatic-explanatory continuum indicator summary (PRECIS): a tool to help trial designers. CMAJ. 2009, 180 (10): E47-E57. 10.1503/cmaj.090523.

    Article  PubMed  PubMed Central  Google Scholar 

  15. Proctor E, Silmere H, Raghavan R, Hovmand P, Aarons G, Bunger A, Griffey R, Hensley M: Outcomes for implementation research: conceptual distinctions, measurement challenges, and research agenda. Adm Policy Ment Health. 2011, 38 (2): 65-76. 10.1007/s10488-010-0319-7.

    Article  PubMed  Google Scholar 

  16. Glasgow RE, Klesges LM, Dzewaltowski DA, Estabrooks PA, Vogt TM: Evaluating the impact of health promotion programs: using the RE-AIM framework to form summary measures for decision making involving complex issues. Health Educ Res. 2006, 21 (5): 688-694. 10.1093/her/cyl081.

    Article  PubMed  Google Scholar 

  17. Reach effectiveness adoption implementation maintenance (RE-AIM). 2013, Blacksburg VA: Virginia Tech, Accessed August, 2013

  18. Ritzwoller DP, Sukhanova A, Gaglio B, Glasgow RE: Costing behavioral interventions: a practical guide to enhance translation. Ann Behav Med. 2009, 37 (2): 218-227. 10.1007/s12160-009-9088-5.

    Article  PubMed  Google Scholar 

  19. Krist AH, Cifuentes M, Dodoo MS, Green LA: Measuring primary care expenses. J Am Board Fam Med. 2010, 23 (3): 376-383. 10.3122/jabfm.2010.03.090089.

    Article  PubMed  Google Scholar 

  20. Agency for Healthcare Research and Quality: Primary care practice-based research networks (PBRNs). 2013, Bethesda MD: Agency for Healthcare Research and Quality, Accessed Jan, 2013

    Google Scholar 

  21. National Cancer Institute: Cancer prevention and control research network (CPCRN). 2013, Bethesda MD: National Cancer Institute, Accessed Jan, 2013

    Google Scholar 

  22. Allen JD, Linnan LA, Emmons KM: Fidelity and its relationship to implementation effectiveness, adaption, and dissemination. Dissemination and implementation research in health: translating science to pratcice. Edited by: Brownson RC, Colditz G, Proctor E. 2012, New York: Oxford University Press, 281-304.

    Chapter  Google Scholar 

  23. Kroenke K, Spitzer RL, Williams JB: The PHQ-9: validity of a brief depression severity measure. J Gen Intern Med. 2001, 16 (9): 606-613. 10.1046/j.1525-1497.2001.016009606.x.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  24. Swinson RP: The GAD-7 scale was accurate for diagnosing generalised anxiety disorder. Evid Based Med. 2006, 11 (6): 184-10.1136/ebm.11.6.184.

    Article  PubMed  Google Scholar 

  25. Spitzer RL, Kroenke K, Williams JB, Lowe B: A brief measure for assessing generalized anxiety disorder: the GAD-7. Arch Intern Med. 2006, 166 (10): 1092-1097. 10.1001/archinte.166.10.1092.

    Article  PubMed  Google Scholar 

  26. Bush K, Kivlahan DR, McDonell MB, Fihn SD, Bradley KA: The AUDIT alcohol consumption questions (AUDIT-C): an effective brief screening test for problem drinking. Ambulatory Care Quality Improvement Project (ACQUIP). Alcohol Use Disorders Identification Test. Arch Intern Med. 1998, 158 (16): 1789-1795. 10.1001/archinte.158.16.1789.

    Article  CAS  PubMed  Google Scholar 

  27. Skinner HA: The drug abuse screening test. Addictive behaviors. 1982, 7 (4): 363-371. 10.1016/0306-4603(82)90005-3.

    Article  CAS  PubMed  Google Scholar 

  28. Whitlock EP, Orleans CT, Pender N, Allan J: Evaluating primary care behavioral counseling interventions: an evidence-based approach. Am J Prev Med. 2002, 22 (4): 267-284. 10.1016/S0749-3797(02)00415-4.

    Article  PubMed  Google Scholar 

  29. Glasgow RE, Goldstein MG: Introduction to and principles of health behavior change. Health promotion and disease prevention in clinical practice. Edited by: Woolf SH, Jonas S, Lawrence RS. 2008, Philadelphia, PA: Lippincott, Williams &Wilkin, 2

    Google Scholar 

  30. Krist AH, Woolf SH, Frazier CO, Johnson RE, Rothemich SF, Wilson DB, Devers KJ, Kerns JW: An electronic linkage system for health behavior counseling effect on delivery of the 5A's. Am J Prev Med. 2008, 35 (5 Suppl): S350-S358.

    Article  PubMed  Google Scholar 

  31. Glasgow RE, Davis CL, Funnell MM, Beck A: Implementing practical interventions to support chronic illness self-management. Jt Comm J Qual Saf. 2003, 29 (11): 563-574.

    PubMed  Google Scholar 

  32. Croteau J, Ryan D: BeWell@stanford. Acheiving your SMART health goals. 2013, Stanford CA: Stanford University, Accessed Jan, 2013

    Google Scholar 

  33. O'Neil J: SMART Goals, SMART Schools. Educational Leadership. 2000, Feb: 46-50.

    Google Scholar 

  34. Avorn J, Soumerai SB: Improving drug-therapy decisions through educational outreach. A randomized controlled trial of academically based "detailing". N Engl J Med. 1983, 308 (24): 1457-1463. 10.1056/NEJM198306163082406.

    Article  CAS  PubMed  Google Scholar 

  35. Soumerai SB, Avorn J: Principles of educational outreach ('academic detailing') to improve clinical decision making. JAMA. 1990, 263 (4): 549-556. 10.1001/jama.1990.03440040088034.

    Article  CAS  PubMed  Google Scholar 

  36. Hensley S: As drug bill soars, some doctors get an "unsales" pitch. Wall Street Journal. 2006, []

    Google Scholar 

  37. Simon SR, Majumdar SR, Prosser LA, Salem-Schatz S, Warner C, Kleinman K, Miroshnik I, Soumerai SB: Group versus individual academic detailing to improve the use of antihypertensive medications in primary care: a cluster-randomized controlled trial. Am J Med. 2005, 118 (5): 521-528. 10.1016/j.amjmed.2004.12.023.

    Article  PubMed  Google Scholar 

  38. Simon SR, Smith DH, Feldstein AC, Perrin N, Yang X, Zhou Y, Platt R, Soumerai SB: Computerized prescribing alerts and group academic detailing to reduce the use of potentially inappropriate medications in older people. J Am Geriatr Soc. 2006, 54 (6): 963-968. 10.1111/j.1532-5415.2006.00734.x.

    Article  PubMed  Google Scholar 

  39. Dietrich AJ, Sox CH, Tosteson TD, Woodruff CB: Durability of improved physician early detection of cancer after conclusion of intervention support. Cancer Epidemiol Biomarkers Prev. 1994, 3 (4): 335-340.

    CAS  PubMed  Google Scholar 

  40. Ashford A, Gemson D, Sheinfeld Gorin SN, Bloch S, Lantigua R, Ahsan H, Neugut AI: Cancer screening and prevention practices of inner-city physicians. Am J Prev Med. 2000, 19 (1): 59-62. 10.1016/S0749-3797(00)00152-5.

    Article  CAS  PubMed  Google Scholar 

  41. Sheinfeld Gorin S, Gemson D, Ashford A, Bloch S, Lantigua R, Ahsan H, Neugut A: Cancer education among primary care physicians in an underserved community. Am J Prev Med. 2000, 19 (1): 53-58. 10.1016/S0749-3797(00)00153-7.

    Article  CAS  PubMed  Google Scholar 

  42. Sheinfeld Gorin S, Ashford A, Lantigua R, Hossain A, Desai M, Troxel A, Gemson D: Effectiveness of academic detailing on breast cancer screening among primary care physicians in an underserved community. J of Am Board Fam Med. 2006, 19: 110-121. 10.3122/jabfm.19.2.110.

    Article  Google Scholar 

  43. Shankaran V, Luu TH, Nonzee N, Richey E, McKoy JM, Graff Zivin J, Ashford A, Lantigua R, Frucht H, Scoppettone M, Bennett CL, Sheinfeld GS: Costs and cost effectiveness of a health care provider-directed intervention to promote colorectal cancer screening. J Clin Oncol. 2009, 27 (32): 5370-5375. 10.1200/JCO.2008.20.6458.

    Article  PubMed  PubMed Central  Google Scholar 

  44. Bouton ME: A learning theory perspective on lapse, relapse, and the maintenance of behavior change. Health Psychol. 2000, 19 (1 Suppl): 57-63.

    Article  CAS  PubMed  Google Scholar 

  45. Ranney L, Melvin C, Lux L, McClain E, Morgan L, Lohr K: Evidence report/technology assessment No. 140. Tobacco Use: prevention, cessation, and control. 2006, Rockville, MD: Prepared by the RTI International-University of North Carolina Evidence-Based Practice Center under Contract No. 290-02-0016

    Google Scholar 

  46. Agency for Healthcare Research and Quality: CAHPS clinician & group surveys. 2013, Bethesda MD: Agency for Healthcare Research and Quality, Accessed Mar, 2013

    Google Scholar 

  47. Dillman DA: Mail and internet surveys: the tailored design method. 1999, Hoboken NJ: John Wiley Company, 2

    Google Scholar 

  48. Edwards P, Roberts I, Clarke M, DiGuiseppi C, Pratap S, Wentz R, Kwan I: Increasing response rates to postal questionnaires: systematic review. BMJ. 2002, 324 (7347): 1183-10.1136/bmj.324.7347.1183.

    Article  PubMed  PubMed Central  Google Scholar 

  49. Kessler RS, Purcell EP, Glasgow RE, Klesges LM, Benkeser RM, Peek CJ: What does It mean to "employ" the RE-AIM model?. Eval Health Prof. 2013, 36 (1): 44-66. 10.1177/0163278712446066.

    Article  PubMed  Google Scholar 

  50. Stange KC, Glasgow RE: Considering and reporting important contextual factors. Methods Brief for the AHRQ Initiative in Patient-Centered Medical Home (PCMH). Edited by: Agency for Healthcare Research and Quality. Rockville: MDIn press

  51. Kaplan HC, Brady PW, Dritz MC, Hooper DK, Linam WM, Froehle CM, Margolis P: The influence of context on quality improvement success in health care: a systematic review of the literature. Milbank Q. 2010, 88 (4): 500-559. 10.1111/j.1468-0009.2010.00611.x.

    Article  PubMed  PubMed Central  Google Scholar 

  52. Kirsh SR, Lawrence RH, Aron DC: Tailoring an intervention to the context and system redesign related to the intervention: a case study of implementing shared medical appointments for diabetes. Implementation science :IS. 2008, 3: 34-10.1186/1748-5908-3-34.

    Article  PubMed  PubMed Central  Google Scholar 

  53. Moos RH: Social contexts: transcending their power and their fragility. Am J Community Psychol. 2003, 31 (1–2): 1-13.

    Article  PubMed  Google Scholar 

  54. Frohlich KL, Potvin L, Chabot P, Corin E: A theoretical and empirical analysis of context: neighbourhoods, smoking and youth. Soc Sci Med. 2002, 54 (9): 1401-1417. 10.1016/S0277-9536(01)00122-8.

    Article  PubMed  Google Scholar 

  55. Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC: Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implementation science: IS. 2009, 4: 50-10.1186/1748-5908-4-50.

    Article  PubMed  PubMed Central  Google Scholar 

  56. Drummond MF, Jefferson TO: Guidelines for authors and peer reviewers of economic submissions to the BMJ. The BMJ Economic Evaluation Working Party. Bmj. 1996, 313 (7052): 275-283. 10.1136/bmj.313.7052.275.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  57. Drummond MF, Sculpher GW, Torrance GL, Stoddart GW: Methods for the economic evaluation in health care programmes. 2005, Oxford: Oxford University Press, 3

    Google Scholar 

  58. Carande-Kulis VG, Maciosek MV, Briss PA, Teutsch SM, Zaza S, Truman BI, Messonnier ML, Pappaioanou M, Harris JR, Fielding J: Methods for systematic reviews of economic evaluations for the guide to community preventive services. Task Force on Community Preventive Services. Am J Prev Med. Jan. 2000, 18 (1 Suppl): 75-91.

    Article  CAS  Google Scholar 

  59. Boardman A, Greenberg D, Vining A, Weimer D: Cost benefit analysis: concepts and practice. 2006, Saddle River, NJ: Prentice Hall, 3

    Google Scholar 

  60. Glasgow RE, Gaglio B, Bennett G, Jerome GJ, Yeh HC, Sarwer DB, Appel L, Colditz G, Wadden TA, Wells B: Applying the PRECIS criteria to describe three effectiveness trials of weight loss in obese patients with comorbid conditions. Health Serv Res. 2012, 47 (3 Pt 1): 1051-1067.

    Article  PubMed  Google Scholar 

  61. Glasgow RE, Vinson C, Chambers D, Khoury MJ, Kaplan RM, Hunter C: National institutes of health approaches to dissemination and implementation science: current and future directions. Am J Public Health. 2012, 102 (7): 1274-1281. 10.2105/AJPH.2012.300755.

    Article  PubMed  PubMed Central  Google Scholar 

  62. Klesges LM, Estabrooks PA, Dzewaltowski DA, Bull SS, Glasgow RE: Beginning with the application in mind: designing and planning health behavior change interventions to enhance dissemination. Ann Behav Med. 2005, 29 (Suppl): 66-75.

    Article  PubMed  Google Scholar 

  63. Glasgow RE, Riley WT: Pragmatic measures: what are they and why do we need them. Am J Prev Med.

  64. Glasgow RE, Green LW, Taylor MV, Stange KC: An evidence integration triangle for aligning science with policy and practice. Am J Prev Med. 2012, 42 (6): 646-654. 10.1016/j.amepre.2012.02.016.

    Article  PubMed  PubMed Central  Google Scholar 

  65. Green LW, Ottoson JM, Garcia C, Hiatt RA: Diffusion theory and knowledge dissemination, utilization, and integration in public health. Annu Rev Public Health. 2009, 30: 151-174. 10.1146/annurev.publhealth.031308.100049.

    Article  PubMed  Google Scholar 

  66. U.S. Preventive Services Task Force: Preventive services. 2013, Accessed June, 2013

    Google Scholar 

  67. Whitlock EP, Polen MR, Green CA, Orleans T, Klein J: Behavioral counseling interventions in primary care to reduce risky/harmful alcohol use by adults: a summary of the evidence for the U.S. Preventive services task force. Ann Intern Med. 2004, 140 (7): 557-568. 10.7326/0003-4819-140-7-200404060-00017.

    Article  PubMed  Google Scholar 

  68. Krist AH, Woolf SH, Rothemich SF, Johnson RE, Peele JE, Cunningham TD, Longo DR, Bello GA, Matzke GR: Interactive preventive health record to enhance delivery of recommended care: a randomized trial. Ann Fam Med. 2012, 10 (4): 312-319. 10.1370/afm.1383.

    Article  PubMed  PubMed Central  Google Scholar 

  69. Krist AH, Peele E, Woolf SH, Rothemich SF, Loomis JF, Longo DR, Kuzel AJ: Designing a patient-centered personal health record to promote preventive care. BMC Med Inform Decis Mak. 2011, 11: 73-10.1186/1472-6947-11-73.

    Article  PubMed  PubMed Central  Google Scholar 

  70. Chin MH, Drum ML, Guillen M, Rimington A, Levie JR, Kirchhoff AC, Quinn MT, Schaefer CT: Improving and sustaining diabetes care in community health centers with the health disparities collaboratives. Med Care. 2007, 45 (12): 1135-1143. 10.1097/MLR.0b013e31812da80e.

    Article  PubMed  Google Scholar 

  71. Pawson R, Greenhalgh T, Harvey G, Walshe K: Realist review–a new method of systematic review designed for complex policy interventions. J Health Serv Res Policy. 2005, 10 (Suppl 1): 21-34. 10.1258/1355819054308530.

    Article  PubMed  Google Scholar 

Download references


Funding for the MOHR project was provided by the National Cancer Institute, Agency for Healthcare Research and Quality, Office of Behavioral and Social Sciences Research, and National Center for Advancing Translational Sciences (CTSA Grant Number ULTR00058). Dr. Stange's time is supported in part by a Clinical Research Professorship from the American Cancer Society and by the National Cancer Society through the Intergovernmental Personnel Act. The opinions expressed in this manuscript are those of the authors and do not necessarily reflect those of the funders.

For their support and guidance of the MOHR project we would like to thank Janice Genervro PhD MSW, David Meyers MD, Robert M Kaplan PhD, Holly Jimison PhD, Richard Wood MHA, Kayla Fair RN MPH, Jennifer Leeman PHD MPH, Catherine L Rohweder DrPH, Alexis Moore MPH, Maria E Fernandez PhD, John Quillin PhD CGC, Melissa Hayes BS, Paul Estabrooks PhD, Christine Nelson PhD RN, and John Heintzman MD MPH.

Most of all we would like to thank our practices for their partnership, insights and hard work: Vienna Primary and Preventive Medicine, Little Falls Family Practice, Charles City Regional Health Services, Chester Family MedCare, Carilion Family Medicine – Roanoke, Carilion Family Medicine – Southeast, Berlin Family Health, Milton Family Practice, Humbolt Open Door Clinic, McKinleyville Community Health Center, St. Johns’ Well Child and Family Center, Spring Branch Community Health Center - Pitner Clinic, Spring Branch Community Health Center - Hillendahl Clinic, HealthPoint Community Health Centers - Navasota, HealthPoint Community Health Centers - Hearne, Murfeesboro Clinic, and Snow Hill Clinic.

For more information on the MOHR project including of the electronic MOHR assessment, patient and healthcare team feedback reports, project updates, information for research and policy makers, lists of publications and presentation, please visit: and

The MOHR Study Group is a collaborative partnership among the study investigators and their respective institutions.

Author information

Authors and Affiliations



Corresponding author

Correspondence to Alex H Krist.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors’ contributions

All authors collaborated in the design of the study and led or participated in described working groups. AH, BG and RG provided initial drafts of the manuscript and integrated feedback from all authors, the remainder of whom are listed alphabetically. All authors read and approved the final manuscript.

Authors’ original submitted files for images

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Krist, A.H., Glenn, B.A., Glasgow, R.E. et al. Designing a valid randomized pragmatic primary care implementation trial: the my own health report (MOHR) project. Implementation Sci 8, 73 (2013).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: