Open Access
Open Peer Review

This article has Open Peer Review reports available.

How does Open Peer Review work?

No more ‘business as usual’ with audit and feedback interventions: towards an agenda for a reinvigorated intervention

  • Noah M Ivers1Email author,
  • Anne Sales2,
  • Heather Colquhoun3,
  • Susan Michie4,
  • Robbie Foy5,
  • Jill J Francis6 and
  • Jeremy M Grimshaw7
Implementation Science20149:14

https://doi.org/10.1186/1748-5908-9-14

Received: 17 September 2013

Accepted: 14 January 2014

Published: 17 January 2014

Abstract

Background

Audit and feedback interventions in healthcare have been found to be effective, but there has been little progress with respect to understanding their mechanisms of action or identifying their key ‘active ingredients.’

Discussion

Given the increasing use of audit and feedback to improve quality of care, it is imperative to focus further research on understanding how and when it works best. In this paper, we argue that continuing the ‘business as usual’ approach to evaluating two-arm trials of audit and feedback interventions against usual care for common problems and settings is unlikely to contribute new generalizable findings. Future audit and feedback trials should incorporate evidence- and theory-based best practices, and address known gaps in the literature.

Summary

We offer an agenda for high-priority research topics for implementation researchers that focuses on reviewing best practices for designing audit and feedback interventions to optimize effectiveness.

Keywords

Audit and feedback Synthesis Best practice Implementation Optimization

Background

Audit and feedback (A&F) involves providing a recipient with a summary of their performance over a specified period of time and is a common strategy to promote the implementation of evidence-based practices. A&F is used widely in healthcare by a range of stakeholders, including research funders and health system payers, delivery organizations, professional groups and researchers, to monitor and change health professionals’ behaviour, both to increase accountability and to improve quality of care. A&F is an improvement over self-assessment [1] or self-monitoring [2] as it can provide objective data regarding discrepancies between current practice and target performance, as well as comparisons of performance to other health professionals. The recognition of suboptimal performance can act as a cue for action, encouraging those who are both motivated and capable to take action to reduce the discrepancy.

The effectiveness of A&F has been evaluated in the third update of a Cochrane review, which included 140 randomized trials of A&F conducted across many clinical conditions and settings around the world. The review found that A&F leads to a median 4.3% absolute improvement (interquartile range 0.5% to 16%) in provider compliance with desired practice [3]. One-quarter of A&F interventions had a relatively large, positive effect on quality of care, while another quarter had a negative or null effect. The challenge of identifying factors that differentiate more and less successful A&F interventions is exacerbated by poor reporting of both intervention components and contextual factors in the literature [4]. Furthermore, most A&F interventions tested in RCTs are designed without explicitly building on previous research or extant theory [5, 6]. As a result, there has been little progress with respect to identifying the key ingredients for a successful A&F intervention or understanding the mechanisms of action of effective A&F interventions in healthcare [7]. Given the now established effectiveness of A&F, its popularity as an intervention is likely to continue; it is imperative that we improve our understanding of how, when and why it achieves large effects.

A&F has been used as a method for changing behaviour in educational [8], organizational [911], cognitive [12], and health psychology [13, 14]; research in these areas has produced hypotheses about mechanisms of action. The effectiveness of A&F may be influenced by the characteristics of the targeted behavior, the recipients, their context, and the components of the A&F intervention itself [9, 15, 16]. To optimize the effectiveness of A&F, we need to understand how to account for these characteristics to design better interventions. The most recent Cochrane review advanced knowledge about A&F by using theory underpinning A&F to guide its evidence synthesis. For example, it incorporated re-analyses [14, 17] of the previous Cochrane review informed by control theory [13] and feedback intervention theory [10], resulting in the finding that feedback is more effective when accompanied by both explicit goals and an action plan. This provides, for the first time, clear advice as to key components that should be incorporated. While this theory-based synthesis made good progress, we need to build on this to identify other key components that will allow practitioners and policy-makers to adopt or adapt A&F with sufficient confidence of achieving high impacts on service delivery and outcomes.

Just as the need to identify research gaps and prioritize research agendas is accepted for clinical research [18], it is also a prerequisite for making substantial progress in implementation research [19]. Rather than continuing with ‘business as usual,’ we believe that a coordinated effort and a systematic approach are now required to determine how best to design and deliver A&F. To this end, a group of international experts from a range of disciplinary backgrounds met for two days in December, 2012. The objectives were: to develop guidance based on current best practices for those developing A&F interventions; and to identify high-priority research topics and opportunities for optimizing A&F effectiveness. Participants were purposively sought based on their contributions to A&F or related literature in healthcare or their roles as potential knowledge users. The list of participants and agenda of the meeting are provided in the (see Additional file 1). This paper summarizes the outcomes of that meeting.

Discussion

Objective 1. Guidance for developing A&F interventions

In the process of developing guidance for A&F intervention development, three linked concepts were identified: identifying known best practices; applying relevant theory when operationalizing best practices; and considering intervention components as factors to be manipulated.

1a. Best practices in design of A&F interventions

Although the ‘ideal’ design for A&F interventions depends on the recipient, their context, and the targeted behaviour, ‘best practices’ can be guided by theory and evidence. The consensus judgment of ‘best practices’ for designing A&F interventions are summarized in Table 1, and evidence in support of these practices is reviewed below.
Table 1

Tentative ‘best practices’ when designing A&F interventions

Audit components

Data are valid

 

Data is based on recent performance

 

Data are about the individual/team’s own behavior(s)

 

Audit cycles are repeated, with new data presented over time

Feedback components

Presentation is multi-modal including either text and talking or text and graphical materials

 

Delivery comes from a trusted source

 

Feedback includes comparison data with relevant others

Nature of the behaviour change required

Targeted behavior is likely to be amenable to feedback

 

Recipients are capable and responsible for improvement

Targets, goals, and action plan

The target performance is provided

 

Goals set for the target behaviour are aligned with personal and organizational priorities

 

Goals for target behaviour are specific, measurable, achievable, relevant, time-bound

 

A clear action plan is provided when discrepancies are evident

1a.i. A&F components

The meta-regression in the Cochrane review of A&F indicated that feedback is more effective when it is presented both verbally and in writing than when using only one modality and when the source (i.e., person delivering the feedback) is a supervisor or senior colleague rather than unknown investigators. Conversely, feedback is less effective when it comes from a regulatory body, agreeing with qualitative work suggesting that feedback perceived to be potentially punitive appeared less effective than a supportive approach [15, 20]. Feedback intervention theory [10] suggests that feedback that captures the recipient’s attention and directs it toward a specific behaviour is more likely to be successful, but when A&F comes from a regulatory body, the recipient may be more likely to activate affective processes (e.g., distress) which distract attention from the specific task requiring change. Diffusion of innovation theory [21] would suggest that feedback from opinion leaders is likely to be more effective, but this was not investigated by the Cochrane reviews of opinion leaders [22] or the A&F review. Aside from exerting social pressure, in-person delivery of feedback could allow for tailoring of the message according to recipient characteristics. This type of process appears to offer a potentially important and readily testable hypothesis. Finally, the A&F review suggested that repeated feedback cycles led to greater improvements in performance than once-only feedback. In part, this may be because recipients are more likely to perceive the data as relevant and accurate when it is delivered closer in time to their own performance [9].

Although the Cochrane review of A&F did not make a specific recommendation, there is a theoretical basis for including comparison data in feedback as a means of motivating recipients to change behavior. Specifically, the theory of planned behaviour emphasizes the potential for social pressure to influence intentions to change professional practice [23]; this could help explain why feedback delivered by senior colleagues tends to be more effective than feedback delivered by unknown investigators. When feedback is delivered via written reports, a preferred approach to the inclusion of normative data for comparison has been identified through a large head-to-head trial [24]. In particular, recipients who received feedback with ‘achievable benchmarks of care’ [25] (i.e., comparing to the top 10% of peers) had greater improvements in processes of care than those receiving feedback where the median performance of peers was highlighted.

1a.ii. Nature of the behaviour change required

The Cochrane review indicated that feedback is most effective when targeting performance with large room for improvement. This may be attributed to larger discrepancy with the comparator, resulting in greater intention to take action, or may reflect the absence of ceiling effects, or both. In addition, the likelihood of feedback identifying a discrepancy between actual and ideal care leading to changed behaviour may depend upon the degree to which recipients believe they have control over that behavior and are committed to change [11]. Feedback should describe achievable behaviors that are ideally desired by the recipient and implementable by the member(s) of the healthcare team responsible for enacting the desired change [15]. This may, in part, explain the tentative findings in the Cochrane review that A&F was less effective when the targeted behaviour changes were more complex (e.g., when special skills are required to implement the desired change).

An observational study showed that clinical practice recommendations rated as incompatible with clinician values and norms were associated with lower compliance at baseline but greater behavior change following feedback than recommendations perceived as highly compatible [26]. This illustrates the potential for A&F to have an effect even upon targeted behaviours reported by clinicians not to be compatible with their values [20, 24].

1a.iii. Targets, goals, and action planning

The Cochrane review findings that feedback interventions were more effective when they included explicit goals and an action plan are not surprising [3]. This is consistent with theories [11, 13] positing that goals can make feedback more salient and action plans specify and facilitate the steps required to achieve goals. When goals are more specific, this process is more effective [9]; ideal goals are commonly considered to be specific, measurable, achievable, relevant and time-bound [27]. Feedback intervention theory proposes that action plans help focus recipients’ attention more productively on the task [10]. In healthcare, providers often manage multiple competing goals with their patients [28]; when A&F successfully directs attention toward specific tasks, it may influence prioritization of these goals.

From a cognitive perspective, receipt of feedback that indicates a discrepancy between expected and actual performance may generate a response described as cognitive dissonance [12]. In this case, participants may attempt to reduce the discrepancy by planning to improve [13] or by undermining the saliency of the discrepancy by discounting the data (or the goals targeted) as invalid [16]. Facilitating the former and reducing the latter is an important aspect of designing effective A&F interventions.

1b. Applying relevant theory to improve design and increase contribution to the literature

While the optimal design of A&F will vary to some extent for each purpose or setting, the strategies chosen for each component (Table 1) should be justified both by empirical evidence and theory [29]. If theory is used, it allows the evaluation of mechanisms of action, thereby contributing to an understanding of why any observed effects occurred. Although a single, all-encompassing, highly predictive theory of A&F is unlikely in the foreseeable future, a number of theories and theoretical constructs are informative [12]. For example, at the individual level, increasing self-efficacy [3032] and goal commitment [11, 33] is likely to increase performance. Organizational and system-level factors that may affect recipients’ ability to act upon a discrepancy highlighted by feedback could also be considered, including competing priorities for the recipients, teams and organizations [34]. This has implications not only with respect to potential contextual effect modifiers, but for the design of the intervention itself: the ideal number and nature of goal-directed behaviours targeted by the feedback may vary by setting and by the organizational level at which the feedback is directed. For example, managers may be able to assimilate and act on a greater number of topics than frontline clinicians, but may lack direct control over specific behaviors.

1c. Manipulating levers: applying best practices within real-world constraints

A&F interventions consist of multiple components, each requiring attention during the design stage; ensuring the inclusion of all desirable components and designing each component in the optimal fashion is rarely possible. We recommend considering individual intervention components as ‘levers’ to be manipulated when working within setting-specific or study design constraints. For example, if circumstances dictate that the delivery of feedback cannot be repeated in a reasonable timeframe, extra attention should be paid to other aspects of the intervention to ensure that discrepancies between actual and desired practice are salient and actionable.

In addition, co-interventions, tailored to overcome identified barriers and boost facilitators, may help if feedback alone seems unlikely to activate the desired response [35]. Ideally, co-interventions that are likely to enhance the process and outcome of A&F interventions should be selected, based on careful consideration of factors that are likely to influence the target behaviours. This gives the opportunity to investigate potential synergy between multiple implementation strategies. For example, while it may seem intuitive to combine payment incentives with A&F, in recipients who are already highly motivated to achieve the goals but have limited perceived control over the outcomes, this combination may not be effective [33]. Alternatively, co-interventions that increase recipients’ skills in quality improvement may be useful because skill development and resultant self-efficacy may increase the likelihood that feedback will lead to improved performance [11, 36]. Since multifaceted interventions featuring A&F may offer little benefit over A&F alone [3], cost-effectiveness of co-interventions should also be considered.

Objective 1 summary – best practices for implementing effective A&F interventions

To summarize our guidance for developing A&F interventions, the best practices outlined in Table 1 should be considered and applied when possible. In particular, prior to conducting A&F, the nature of the behaviour change required to improve performance should be elucidated in order to develop clear action plans. Theory and evidence should be made explicit with respect to proposed design choices when manipulating the design of individual intervention components, and the expected mechanisms of action should be justified to aid interpretation.

Objective 2. High-priority research topics and opportunities for optimizing audit and feedback

There is a need for future research to turn its focus from evaluating A&F interventions against usual care to explicitly evaluating different types of A&F. In particular, confirmatory head-to-head trials examining different strategies for implementing the tentative recommendations in Table 1 would provide a valuable contribution to knowledge and practice. However, even if the components featured in Table 1 were optimized, additional components of A&F (that might be currently known or unknown) or co-interventions would be worthy of testing. Additionally, further knowledge regarding potential effect modifiers is needed.

Examples of potentially fruitful research topics that could lead to enhancing the effectiveness of A&F are listed in Table 2. These topics are organized by factors related to: a) contextual and recipient characteristics that may moderate A&F effectiveness, b) the design of the A&F intervention itself, and c) the characteristics of the targeted behaviour change. Knowledge about organizational resources and the nature of the team responsible for achieving the outcome of interest could help explain variability in the effects of A&F interventions [37]. The complexity of clinical practice recommendations and the degree to which benefits can readily be observed may play an important role [21, 3840]. More information is needed to understand whether some behaviours are more amenable to change through A&F and/or how to adapt the design of A&F accordingly.
Table 2

Example research topics for optimizing audit and feedback

Factors related to context and/or recipient

Characteristics of the recipient

• Engagement in audit and/or in feedback design

 

• Goal orientation of recipients

 

• Degree of motivation to improve performance

 

• Training of recipients to understand and act on feedback

 

• Profession of recipient and/or multi-disciplinary feedback

Characteristics of the setting

• Location (e.g., hospital versus clinic, national setting)

 

• Organizational resources

 

• Size of the team responsible for outcomes of interest

Co-interventions

• Time and/or standardized support to reflect upon feedback

 

• Impact of combining A&F with one of the following:

 

• Incentives or penalties (financial, CME, licensing)

 

• Tools and practise aids (clinical decision tool)

 

• Education (academic detailing, group learning)

 

• Practice redesign (coaches, facilitation, mentorship)

Factors related to intervention design

Nature of delivery of the information

• Mode of delivery of feedback (e.g., paper, electronic, face-to-face)

 

• Length, duration

 

• Perceived credibility of the source and/or competence of the presenter

 

• Different sources (peer versus supervisor versus external group)

 

• Frequency of feedback

 

• Role of social pressure, dissemination/visibility of information to peer-group

Nature of the content

• Sign of the message (positive versus negative)

 

• Graded feedback (starting positive)

 

• Type of benchmarks and/or comparison information

 

• Type action plans or correct solution information

 

• Level of aggregation of feedback data (individual versus team)

 

• Role for intermediate outcomes/process measures versus patient-level outcomes

Characteristics of the targeted behaviours

• Perceived importance of the target relative to other priorities

 

• Observability of improvement (whether impact of using a new practice can be seen quickly)

 

• The degree to which the recommended practice requires changes in habits and routines

 

• Complexity of targeted behaviours (number of indicators reported or behavioural changes required and skill level necessary for desired behaviour change)

There are opportunities to embed research within ongoing organizational initiatives at low cost in settings where routine data are available and where mechanisms for audit are in place, and this should be considered in the planning and conduct of A&F studies [41]. In particular, A&F researchers could partner with those providing routine feedback to healthcare providers to test variations of intervention components that might increase effectiveness. Administrators of quality improvement initiatives involving A&F may be keen to participate in head-to-head trials comparing different A&F designs because, in this case, all providers receive an intervention, while management learns about how their programs may be improved. Thus, settings where valid data for audit and mechanisms for feedback are already established are potential implementation research laboratories, enabling relatively rapid testing of multiple intervention components.

There is evidence suggesting that physicians respond differently than other health professionals to implementation interventions [42], and it is possible that this is also true for A&F. Even within the same broad professional discipline, there may be differential responses related to values and contexts of different groupings (e.g., primary and secondary care providers). The Cochrane review found little difference in effect size between A&F targeting physicians and non-physicians, but 86% of the trials focused on physicians. It is plausible that the effectiveness of A&F and key effect modifiers may vary in less-studied settings where non-physicians are the primary providers, such as home care or long-term care.

Opportunities for advancing the science related to A&F may be found through consideration of different study designs. When evaluating the addition of a single component, such as adding action plans to feedback, a two-arm, head-to-head trial is a logical option [43]. However, where there is an opportunity to consider many A&F-related factors at once, a multi-phase optimization strategy including a fractional factorial trial offers an efficient approach [44]. In such a study design, an investigator could simultaneously test various components of A&F content and delivery of the intervention. Existing A&F-relevant theories should be used to select the intervention components to be studied and to generate hypotheses regarding the interaction between such components and contextual characteristics, including those of the recipient and of the targeted behaviour. In addition, a process evaluation within a trial can be used to investigate and develop program-specific theory [45]. Examining mechanisms of action and qualitative studies prior to trials may help identify barriers and facilitators to change that can inform intervention design [15].

Objective 2 summary – opportunities to improve knowledge regarding how to optimize A&F

To summarize, we believe that future A&F research should prioritize studies that aim to determine how the effect of the intervention may be optimized rather than assessing whether the intervention can increase the implementation of evidence-based practices. This justification is similar to the arguments made in support of the shift toward comparative effectiveness research elsewhere in healthcare. Table 2 provides example topics in which greater understanding is needed to determine how best to utilize A&F. On the basis of the considerations reviewed above, we recommend that, prior to directing further research effort toward new trials of A&F versus usual care, investigators consider the following questions:
  1. 1.

    How could a proposed study add to present empirical knowledge? For example, is there a good justification for considering a particular intervention component and a clear hypothesis regarding how it will alter the effectiveness of A&F?

     
  2. 2.

    To what degree can the components of the A&F intervention be sufficiently defined and operationalized across different settings? For example, how costly is the intervention to replicate, and is there potential for sustainability and spread?

     
  3. 3.

    Has some pilot work been conducted? For example, is the feedback interpreted as intended, and has the design been iteratively optimized prior to scaling up?

     

Summary

A&F has the potential to significantly improve quality of care, but despite more than 140 published RCTs, important knowledge gaps remain regarding when it will work best and why, and how to design reliable, effective A&F interventions across settings and provider groups. The lack of systematic, coordinated research in this area, as well as gaps in study reporting, contribute to this problem. To determine changes required to the ‘business as usual’ approach focusing on trials of A&F versus usual care, we gathered a multidisciplinary group of international researchers and knowledge users to develop an agenda for future research in the field. Participants recognized the tension between research and action and urged the research community to capitalize on opportunities to simultaneously contribute to implementation science while conducting implementation. The discussions we summarize here were driven by the findings from the most recent Cochrane review of A&F interventions in healthcare, and the collective experiences and research programs of the meeting participants. The limitations of the recommendations described here reflect the limitations of the studies reviewed. In addition, our considerations only reflect opinions expressed at the meeting. We deliberately sought to invite a wide range of individuals to the meeting to maximize diversity, but with a maximum of 30 attendees, there may have been important perspectives omitted. We acknowledge that there is no present universally accepted, comprehensive list of A&F intervention components.

We have summarized best practice recommendations for the design of A&F based on currently available evidence and have proposed examples of high priority research questions that could move the field forward. Future A&F interventions should feature well-described and carefully justified components, with the theoretical rationale made explicit. Rather than comparisons of A&F versus usual care, head-to-head comparisons of interventions with embedded process evaluations are encouraged to determine how to optimize effect size. Researchers, funders and other stakeholders should assess new A&F studies critically to evaluate the extent to which they contribute new knowledge to the field, and endeavor to draw on best evidence when designing and implementing interventions.

Declarations

Acknowledgments

The meeting leading to this paper was supported by a planning grant from the Canadian Institutes of Health Research (CIHR) entitled, ‘Improving the Effectiveness of Audit and Feedback Interventions in Health Care,’ and supplemented with financial and administrative support from Knowledge Translation Canada (KT Canada). The positions taken in this paper should not be attributed to CIHR or KT Canada. NI is supported by Fellowship awards from CIHR and the Department of Family and Community Medicine, University of Toronto. JMG holds a Canada Research Chair in Health Knowledge Transfer and Uptake. HC is supported by a CIHR Fellowship Award.

The authors recognize and thank all participants at the Improving the Effectiveness of Audit and Feedback Interventions in Health Care meeting: Donna Angus, Richard Baker, Tupper Bean, Jamie Brehaut, Craig Campbell, Mark Chignell, George Collier, Janet Curran, Lynn Dionne, Mary Dixon-Woods, Signe Flottorp, Simon French, Michael Halasy, Gro Jamtvedt, Cara Litvin, Sumit Majumdar, Denise O’Connor, Steve Ornstein, Alison Paprica, David Price, Sue Wells, Michel Wensing, and Merrick Zwarenstein.

Authors’ Affiliations

(1)
Department of Family and Community Medicine, Institute for Health Systems Solutions and Virtual Care, Women’s College Hospital, University of Toronto
(2)
Center for Clinical Management Research, VA Ann Arbor Healthcare System, University of Michigan
(3)
Ottawa Hospital Research Institute, Clinical Epidemiology Program, The Ottawa Hospital
(4)
Centre for Outcomes Research and Effectiveness, Department of Clinical, Educational and Health Psychology University College London
(5)
Leeds Institute of Health Sciences, University of Leeds
(6)
School of Health Sciences, City University London
(7)
Clinical Epidemiology Program, Ottawa Health Research Institute, University of Ottawa

References

  1. Davis DA, Mazmanian PE, Fordis M, Van Harrison R, Thorpe KE, Perrier L: Accuracy of physician self-assessment compared with observed measures of competence: a systematic review. JAMA. 2006, 296 (9): 1094-1102. 10.1001/jama.296.9.1094.View ArticlePubMedGoogle Scholar
  2. Audet AM, Doty MM, Shamasdin J, Schoenbaum SC: Measure, learn, and improve: physicians’ involvement in quality improvement. Health Aff (Millwood). 2005, 24 (3): 843-853. 10.1377/hlthaff.24.3.843.View ArticleGoogle Scholar
  3. Ivers N, Jamtvedt G, Flottorp S, Young JM, Odgaard-Jensen J, French SD, O’Brien MA, Johansen M, Grimshaw J, Oxman AD: Audit and feedback: effects on professional practice and healthcare outcomes. Cochrane Database Syst Rev. 2012, 6: CD000259Google Scholar
  4. Glasziou P, Meats E, Heneghan C, Shepperd S: What is missing from descriptions of treatment in trials and reviews?. BMJ. 2008, 336 (7659): 1472-1474. 10.1136/bmj.39590.732037.47.View ArticlePubMedPubMed CentralGoogle Scholar
  5. Sales A, Smith J, Curran G, Kochevar L: Models, strategies, and tools. Theory in implementing evidence-based findings into health care practice. J Gen Intern Med. 2006, 21 (Suppl 2): 43-49.Google Scholar
  6. Colquhoun HL, Brehaut JC, Sales A, Ivers N, Grimshaw J, Michie S, Carroll K, Chalifoux M, Eva KW: A systematic review of the use of theory in randomized controlled trials of audit and feedback. Implement Sci. 2013, 8 (1): 66-10.1186/1748-5908-8-66.View ArticlePubMedPubMed CentralGoogle Scholar
  7. Foy R, Eccles MP, Jamtvedt G, Young J, Grimshaw JM, Baker R: What do we know about how to do audit and feedback? Pitfalls in applying evidence from a systematic review. BMC Health Serv Res. 2005, 5: 50-10.1186/1472-6963-5-50.View ArticlePubMedPubMed CentralGoogle Scholar
  8. Shute V: Focus on formative feedback. Rev of Educational Research. 2008, 78: 153-189. 10.3102/0034654307313795.View ArticleGoogle Scholar
  9. Ilgen DR, Fisher CD, Taylor MS: Consequences of individual feedback on behavior in organizations. J Appl Psychol. 1979, 64 (4): 349-371.View ArticleGoogle Scholar
  10. Kluger AN, DeNisi A: The Effects of Feedback Interventions on Performance: A Historical Review, a Meta-Analysis, and a Preliminary Feedback Intervention Theory. Psychol Bull. 1996, 119 (2): 254-284.View ArticleGoogle Scholar
  11. Locke EA, Latham GP: Building a practically useful theory of goal setting and task motivation. A 35-year odyssey. Am Psychol. 2002, 57 (9): 705-717.View ArticlePubMedGoogle Scholar
  12. Brehaut JC, Eva KW: Building theories of knowledge translation interventions: use the entire menu of constructs. Implement Sci. 2012, 7: 114-5908-7-114-View ArticleGoogle Scholar
  13. Carver CS, Scheier MF: Control theory: a useful conceptual framework for personality-social, clinical, and health psychology. Psychol Bull. 1982, 92 (1): 111-135.View ArticlePubMedGoogle Scholar
  14. Gardner B, Whittington C, McAteer J, Eccles MP, Michie S: Using theory to synthesise evidence from behaviour change interventions: the example of audit and feedback. Soc Sci Med. 2010, 70 (10): 1618-1625. 10.1016/j.socscimed.2010.01.039.View ArticlePubMedGoogle Scholar
  15. Hysong SJ, Best RG, Pugh JA: Audit and feedback and clinical practice guideline adherence: Making feedback actionable. Implement Sci. 2006, 1: 9-10.1186/1748-5908-1-9.View ArticlePubMedPubMed CentralGoogle Scholar
  16. van der Veer SN, de Keizer NF, Ravelli AC, Tenkink S, Jager KJ: Improving quality of care. A systematic review on how medical registries provide information feedback to health care providers. Int J Med Inform. 2010, 79 (5): 305-323. 10.1016/j.ijmedinf.2010.01.011.View ArticlePubMedGoogle Scholar
  17. Hysong SJ: Meta-analysis: audit and feedback features impact effectiveness on care quality. Med Care. 2009, 47 (3): 356-363. 10.1097/MLR.0b013e3181893f6b.View ArticlePubMedPubMed CentralGoogle Scholar
  18. Li T, Vedula SS, Scherer R, Dickersin K: What comparative effectiveness research is needed? A framework for using guidelines and systematic reviews to identify evidence gaps and research priorities. Ann Intern Med. 2012, 156 (5): 367-377. 10.7326/0003-4819-156-5-201203060-00009.View ArticlePubMedPubMed CentralGoogle Scholar
  19. Eccles MP, Armstrong D, Baker R, Cleary K, Davies H, Davies S, Glasziou P, Ilott I, Kinmonth AL, Leng G, Logan S, Marteau T, Michie S, Rogers H, Rycroft-Malone J, Sibbald B: An implementation research agenda. Implement Sci. 2009, 4: 18-5908-4-18-Google Scholar
  20. Kluger AN, Van Dijk D: Feedback, the various tasks of the doctor, and the feedforward alternative. Med Educ. 2010, 44 (12): 1166-1174. 10.1111/j.1365-2923.2010.03849.x.View ArticlePubMedGoogle Scholar
  21. Rogers E: Diffusion of innovations. 1995, New York: Free PressGoogle Scholar
  22. Flodgren G, Parmelli E, Doumit G, Gattellari M, O’Brien MA, Grimshaw J, Eccles MP: Local opinion leaders: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2011, 8 (8): CD000125-PubMedGoogle Scholar
  23. Ajzen I: The theory of planned behavior. Organ Behav Hum Decis Process. 1991, 50 (2): 179-211. 10.1016/0749-5978(91)90020-T.View ArticleGoogle Scholar
  24. Kiefe CI, Allison JJ, Williams OD, Person SD, Weaver MT, Weissman NW: Improving quality improvement using achievable benchmarks for physician feedback: a randomized controlled trial. JAMA. 2001, 285 (22): 2871-2879. 10.1001/jama.285.22.2871.View ArticlePubMedGoogle Scholar
  25. Weissman NW, Allison JJ, Kiefe CI, Farmer RM, Weaver MT, Williams OD, Child IG, Pemberton JH, Brown KC, Baker CS: Achievable benchmarks of care: the ABCs of benchmarking. J Eval Clin Pract. 1999, 5 (3): 269-281. 10.1046/j.1365-2753.1999.00203.x.View ArticlePubMedGoogle Scholar
  26. Foy R, MacLennan G, Grimshaw J, Penney G, Campbell M, Grol R: Attributes of clinical recommendations that influence change in practice following audit and feedback. J Clin Epidemiol. 2002, 55 (7): 717-722. 10.1016/S0895-4356(02)00403-1.View ArticlePubMedGoogle Scholar
  27. Doran GT: There’s a S.M.A.R.T. way to write management’s goals and objectives. AMA Forum. Management review. 1981, 70 (11): 35-36.Google Scholar
  28. Presseau J, Sniehotta FF, Francis JJ, Campbell NC: Multiple goals and time constraints: perceived impact on physicians’ performance of evidence-based behaviours. Implement Sci. 2009, 4: 77-5908-4-77-View ArticleGoogle Scholar
  29. Cane J, O’Connor D, Michie S: Validation of the theoretical domains framework for use in behaviour change and implementation research. Implement Sci. 2012, 7: 37-5908-7-37-View ArticleGoogle Scholar
  30. Ashford SJ, Cummings LL: Feedback as an individual resource: Personal strategies of creating information. Organ Behav Hum Perform. 1983, 32 (3): 370-398. 10.1016/0030-5073(83)90156-3.View ArticleGoogle Scholar
  31. Bandura A, Cervone D: Self-evaluative and self-efficacy mechanisms governing the motivational effects of goal systems. J Pers Soc Psychol. 1983, 45 (5): 1017-1028.View ArticleGoogle Scholar
  32. Johnson RE, Chang CH, Lord RG: Moving from cognition to behavior: What the research says. Psychol Bull. 2006, 132 (3): 381-415.View ArticlePubMedGoogle Scholar
  33. Hysong SJ, Simpson K, Pietz K, SoRelle R, Broussard Smitham K, Petersen LA: Financial incentives and physician commitment to guideline-recommended hypertension management. Am J Manag Care. 2012, 18 (10): e378-e391.PubMedPubMed CentralGoogle Scholar
  34. Presseau J, Francis JJ, Campbell NC, Sniehotta FF: Goal conflict, goal facilitation, and health professionals’ provision of physical activity advice in primary care: An exploratory prospective study. Implement Sci. 2011, 6: 73-10.1186/1748-5908-6-73.View ArticlePubMedPubMed CentralGoogle Scholar
  35. Baker R, Camosso-Stefinovic J, Gillies C, Shaw EJ, Cheater F, Flottorp S, Robertson N: Tailored interventions to overcome identified barriers to change: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2010, 3: CD005470Google Scholar
  36. Bonetti D, Johnston M, Pitts NB, Deery C, Ricketts I, Tilley C, Clarkson JE: Knowledge may not be the best target for strategies to influence evidence-based practice: using psychological models to understand RCT effects. Int J Behav Med. 2009, 16 (3): 287-293. 10.1007/s12529-008-9017-1.View ArticlePubMedGoogle Scholar
  37. Ivers N, Barnsley J, Upshur R, Shah B, Tu K, Grimshaw J, Zwarenstein M: My job is one patient at a time’: perceived discordance between population-level quality targets and patient-centered care inhibits quality improvement. Canadian Family Physician. 2013, in pressGoogle Scholar
  38. Grilli R, Lomas J: Evaluating the message: the relationship between compliance rate and the subject of a practice guideline. Med Care. 1994, 32 (3): 202-213. 10.1097/00005650-199403000-00002.View ArticlePubMedGoogle Scholar
  39. Grol R, Dalhuijsen J, Thomas S, Veld C, Rutten G, Mokkink H: Attributes of clinical guidelines that influence use of guidelines in general practice: observational study. BMJ. 1998, 317 (7162): 858-861.View ArticlePubMedPubMed CentralGoogle Scholar
  40. Shekelle PG, Kravitz RL, Beart J, Marger M, Wang M, Lee M: Are nonspecific practice guidelines potentially harmful? A randomized comparison of the effect of nonspecific versus specific guidelines on physician decision making. Health Serv Res. 2000, 34 (7): 1429-1448.PubMedPubMed CentralGoogle Scholar
  41. Glynn RJ, Brookhart MA, Stedman M, Avorn J, Solomon DH: Design of cluster-randomized trials of quality improvement interventions aimed at medical care providers. Med Care. 2007, 45 (10 Supl 2): 38-43.View ArticleGoogle Scholar
  42. Godin G, Belanger-Gravel A, Eccles M, Grimshaw J: Healthcare professionals’ intentions and behaviours: a systematic review of studies based on social cognitive theories. Implement Sci. 2008, 3: 36-10.1186/1748-5908-3-36.View ArticlePubMedPubMed CentralGoogle Scholar
  43. Ivers N, Tu K, Francis J, Barnsley J, Shah B, Upshur R, Moineddin R, Grimshaw J, Zwarenstein M: Feedback GAP: study protocol for a cluster-randomized trial of goal setting and action plans to increase the effectiveness of audit and feedback interventions in primary care. Implement Sci. 2013, 8: 42-10.1186/1748-5908-8-42.View ArticleGoogle Scholar
  44. Collins LM, Chakraborty B, Murphy SA, Strecher V: Comparison of a phased experimental approach and a single randomized clinical trial for developing multicomponent behavioral interventions. Clin Trials. 2009, 6 (1): 5-15. 10.1177/1740774508100973.View ArticlePubMedPubMed CentralGoogle Scholar
  45. Dixon-Woods M, Bosk CL, Aveling EL, Goeschel CA, Pronovost PJ: Explaining Michigan: developing an ex post theory of a quality improvement program. Milbank Q. 2011, 89 (2): 167-205. 10.1111/j.1468-0009.2011.00625.x.View ArticlePubMedPubMed CentralGoogle Scholar

Copyright

© Ivers et al.; licensee BioMed Central Ltd. 2014

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.

Advertisement