Open Access

Improving the delivery of care for patients with diabetes through understanding optimised team work and organisation in primary care

  • Martin P Eccles1Email author,
  • Gillian Hawthorne2,
  • Marie Johnston3,
  • Margaret Hunter1,
  • Nick Steen1,
  • Jill Francis4,
  • Susan Hrisos1,
  • Marko Elovainio5 and
  • Jeremy M Grimshaw6, 7
Implementation Science20094:22

DOI: 10.1186/1748-5908-4-22

Received: 19 December 2008

Accepted: 27 April 2009

Published: 27 April 2009

Abstract

Background

Type 2 diabetes is an increasingly prevalent chronic illness and is an important cause of avoidable mortality. Patients are managed by the integrated activities of clinical and non-clinical members of the primary care team. Studies of the quality of care for patients with diabetes suggest less than optimum care in a number of areas.

Aim

The aim of this study is to improve the quality of care for patients with diabetes cared for in primary care in the UK by identifying individual, team, and organisational factors that predict the implementation of best practice.

Design

Participants will be clinical and non-clinical staff within 100 general practices sampled from practices who are members of the MRC General Practice Research Framework. Self-completion questionnaires will be developed to measure the attributes of individual health care professionals, primary care teams (including both clinical and non-clinical staff), and their organisation in primary care. Questionnaires will be administered using postal survey methods. A range of validated theories will be used as a framework for the questionnaire instruments. Data relating to a range of dimensions of the organisational structure of primary care will be collected via a telephone interview at each practice using a structured interview schedule. We will also collect data relating to the processes of care, markers of biochemical control, and relevant indicator scores from the quality and outcomes framework (QOF). Process data (as a proxy indicator of clinical behaviours) will be collected from practice databases and via a postal questionnaire survey of a random selection of patients from each practice. Levels of biochemical control will be extracted from practice databases. A series of analyses will be conducted to relate the individual, team, and organisational data to the process, control, and QOF data to identify configurations associated with high quality care.

Study registration

UKCRN ref:DRN120 (ICPD)

Background

In the UK, Type 2 diabetes is an increasingly prevalent chronic illness (prevalence now over 3% [1], equating to approximately 50 patients per full time general practitioner). It is an important cause of avoidable mortality. Patients are managed by the integrated activities of clinical and non-clinical members of the primary care team. There are National Institute for Health and Clinical Effectiveness (NICE) guidelines defining standards of care, but studies of the quality of care for patients with diabetes suggest less than optimum care in a number of areas [2]. Quality and outcomes framework (QOF) data for 2004 and 2005 suggest high rates of measurement of clinical and biochemical parameters but lower rates of acting on the results. Only 67% of practices achieved the target of 50% of patients having their hemoglobin A1c (HbA1c) within the target range; the similar figure for blood pressure (BP) (target of 55%) was 70%. Some of this variability will reflect patient physiology/behaviour, but it will also reflect variable clinical management behaviours. General practices, although relatively small organisations, have become more complex in terms of their structures and functions. Few studies have examined factors underlying the organisation and delivery of care in general practice, and only one UK study related them to clinical data [3] demonstrating that team climate, booking interval, and practice size together explained 31% of the variance in diabetes management. These authors, along with others [4, 5], identified the need for a better understanding of the relationship between the quality of patient care and individual and organisational factors.

A consistent finding in health services research is that the transfer of research findings into practice is unpredictable and can be a slow and haphazard process [6]. Studies in the USA and the Netherlands suggest that 30 to 40% of patients do not receive care according to current scientific evidence, and 20 to 25% of care provided is not needed or potentially harmful [7, 8]. A review of quality of care studies (including diabetes) from UK primary care concluded that 'in almost all studies the process of care did not reach the standards set out in national guidelines or set by the researchers themselves'[2]. In our recently completed randomized controlled trial (RCT), only 25% of diabetic patients received statins and less than 50% had a foot examination. Recognition of this quality gap has led to increased interest in active implementation strategies and implementation research (the scientific study of methods to promote the systematic uptake of research findings into routine clinical practice) over the past fifteen years [911]. It has been demonstrated that interventions can be effective, but less information is provided to guide the choice or optimise the components of complex interventions in practice [12]. The effectiveness of interventions varies across different clinical problems, contexts, and organizations, but studies provided scant theoretical or conceptual rationale for their choice of intervention [13], and only limited descriptions of the interventions and contextual data [14]. We have argued previously [15] that the understanding of potential barriers and enablers to implementation is limited. The challenge for implementation researchers is to develop and evaluate a theory-based approach that will offer a generalisable framework for research and support the choice and development of interventions. Such a framework is also needed for the interpretation of implementation study results.

We have conducted a range of relevant work, including: studies of quality of care in diabetes in primary care [1618]; pragmatic randomised controlled trials of various quality improvement (QI) strategies, including computerisation of guidelines [19], educational messages attached to test results [20, 21], an enhanced area wide diabetes register [22], and outreach visiting [2325]; studies introducing QI strategies into health service organisations [26, 27]; and studies exploring the role of theory in the design and conduct of QI studies [15, 2830]. We are currently conducting a cluster RCT of a QI intervention targeting general practitioners (GPs) caring for patients with diabetes in Newcastle. We have also conducted research in organizational development and behavioural change related to health care organizations in primary and secondary care [3133].

Aim

The aim of this study is to improve the quality of care for patients with diabetes cared for in primary care by identifying individual, team, and organisational factors that predict the implementation of best practice.

Objectives

  1. 1.

    To measure attributes of individual health care professionals (HCPs), teams, and their organisation in primary care.

     
  2. 2.

    To measure a range of dimensions of organisational structure in primary care.

     
  3. 3.

    To measure the process of care, markers of biochemical control, and QOF scores.

     
  4. 4.

    To relate the data from objectives one and two to the data from objective three, and thereby identify configurations associated with high quality care.

     

Methods

Design

The overall design is a predictive study where a series of attributes of individuals, their teams, and organisations are measured, and their ability to predict quality of care over the subsequent 12 months is explored in the analysis.

Participants and setting

The study will be based in 100 general practices within the UK's MRC general practice research framework (GPRF). There are two major reasons why the use of such a network of practices is necessary. First, in this study it is important to have high-quality data on clinical behaviours. In our previous work, we have used routinely available (prescribing) data as our measure of performance, and we have experienced problems with the specificity of such measures [34]. Second, it is of fundamental importance to this study that we obtain as near as possible complete 'sets' of questionnaires from all relevant members of the primary care team in order to allow the production of robust collective team level values. Although in previous studies we have been successful in obtaining sufficient responses, these have usually been across a large number of practices with only one or two responses per practice. Practices will be recruited by postal invitation with telephone follow-up via the GPRF.

Participants will be all clinical and non-clinical members of the primary care team within each practice. We will collect information about the organisation and team functioning of the primary care team from both clinical and non-clinical team members and information relating to clinical behaviours from clinical team members. Six different aspects of care delivery for patients with diabetes will be examined: glycaemic control, BP control, foot examination, weight management, patient education, and self monitoring. Anonymous clinical records of patients in the study practices will provide the data on clinical variables. A random sample of 100 patients per practice will provide data on clinician behaviours.

Predictor variables

Data relating to practice organisational structure

We will collect details of practice structures and function informed by previous studies [3, 5]: practice demographics (including practice list size; training status of the practice, and postcodes covered); routine booking intervals for patient consultations; staffing levels of practice staff (numbers of, and number of sessions worked by, doctors, practice employed nurses, and administrative staff); skill mix (ratio of doctors to non-medical clinical staff, and of clinical to administrative staff); organisation of care for the clinical conditions (including specialisation within the clinicians).

Data relating to individual staff's perceptions of the clinical care of their patients with type 2 diabetes, team functioning, and practice organisation

Choosing theories

The theoretical frameworks included in this study have been carefully chosen after a process of critical consideration [35]. Multiple theories are required as no one theory covers all of the relevant domains of behaviour [36]. We have chosen theories that predict behaviour change and that have standard methods of operationalisation. Especially in the context of diabetes care, a combination of individual and team/organisational measures has great potential for added understanding.

Theories on human behaviour (and especially adult behaviour change) can be categorised in many different ways. We propose to use theories that cover individuals' cognitions, habitual behaviour, and team performance and decision-making. Individuals' cognitions about the six clinical behaviours will be measured using situation-outcome, outcome and self-efficacy expectancies from social cognitive theory [37], attitude, subjective norms, perceived behavioural control and intention from the theory of planned behaviour (TPB) [38], the self-report habit index [39], and action planning and coping planning [40]. We will also measure self-reported past behaviour.

Based on Bandura's social cognitive theory [37], we will measure three kinds of expectancies – situation-outcome, outcome, and self-efficacy expectancies. TPB proposes that the strength of an individual's intention (or motivation) to engage in a behaviour, and the degree of control they feel they have over that behaviour (perceived behavioural control, or PBC) are the proximal determinants of engaging in it [38]. In turn, intention is influenced by three variables: attitudes towards the behaviour, subjective norms, and PBC. Anticipated regret (an extension to TPB) reflects the notion that the anticipation of regret that would follow the adoption or non-adoption of a behaviour influences its adoption. Including this construct improves the predictive value of TPB [41]. The self-report habit index [39] is a 12-item measure that breaks down the habit construct into a number of features (perceptions of frequency, automaticity, and self-identity). Gollwitzer [42] has identified implementation intentions as explicit plans about when and where a goal intention will be achieved [42]. A relatively new concept in health behaviour research experimental studies suggest that people who have formulated plans are more likely to translate their intentions into action than those who have not [43, 44]. The concept has recently been further developed by Sniehotta [40], who proposes the two distinct dimensions of plans: action planning (planning the initiation of a behaviour) and coping planning (planning what to do when barriers to action are encountered in order to maintain changed behaviours).

Individual staff's perceptions of team performance and decision making will be measured through their ratings of: work characteristics, as defined by the demand-control model [45] ('job control'); characteristics of employee interaction, as defined by team climate research [46]; and the characteristics of decision making and managerial procedures, as defined by organisational justice research [47, 48].

Developing theory-based measures

Using the methods that we have described previously [35], measures will be developed for each predictor variable for the survey. Wherever possible we will use existing measures as a starting point in this process and will follow the standard procedures that have been described to develop measures of these theoretical constructs. Questionnaire items will be rated on seven-point scales with appropriately worded anchors, usually 'strongly disagree' to 'strongly agree'. The questionnaire will be piloted for clarity and acceptability to both clinical and non-clinical staff.

Individuals' cognitions about work characteristics (both in general and in relation to diabetes care provision) will be measured using Karasek's job decision latitude scale [45] and job demands scale [49], and Siegrist's effort-reward imbalance measure [50]. Cognitions about the team will be measured using the shortened version of the original team climate inventory [46, 51]. Cognitions about the organisation will be measured using the organizational justice evaluation scale [52, 53].

Dependent variables

Data relating to the process of care, markers of biochemical control, and QOF scores

Data for dependent variables will be collected from four sources. We will study both the performance of clinical behaviours (measuring HbA1c, BP) and the associated biochemical/physiological measurement (level of HbA1c; level of BP) accepting that there will be variability in the latter measures reflecting patient physiology and behaviour. However, these are the sort of criteria by which quality of care is judged. We will also measure clinician self-report and patient report of clinician behaviour. All dependent variables will relate to the same twelve-month period (12 months from the completion of the initial theory-based questionnaire).

Data held within practice computers on the performance of clinical behaviours and measures of physiology

These data will be collected in two ways. First, for all registered patients with diabetes we will collect data on the total number of patients with diabetes in the practice and the number who have had: a foot check; BP, HbA1c, cholesterol, and weight measured; level of systolic and diastolic BP, level of HbA1c, level of cholesterol, and body mass index; diabetes-related medication (hypoglycaemic drugs, lipid-lowering drugs, weight-reducing drugs); advice about self-monitoring and education.

Second, we will collect QOF scores for diabetes and for additional items that reflect aspects of good organisation and that, on the basis of QOF scores, can be expected to discriminate between practices (e.g., Records 18 (The practice has up-to-date clinical summaries in at least 80 percent of patient records); Education 6 (The practice conducts an annual review of patient complaints and suggestions to ascertain general learning points which are shared with the team); Med 9 (A medication review is recorded in the notes in the preceding 15 months for all patients being prescribed repeat medicines.). These data are available as categorical variables [54].

Clinician self-reported behaviour

Two measures of clinicians' self-reported behaviour that can be regarded as proxies for actual behaviour will be included in the initial theory-based questionnaire – behavioural simulation and behavioural intention. Behavioural simulation will be measured using the method we have used previously [27]. From literature and expert consensus we will identify elements reported to influence management of the clinical conditions. From this, clinical scenarios will be constructed describing patients presenting in primary care. Respondents will be asked to make decisions on the management of the patients described. Behavioural intention will be measured by three items worded in a standard manner (e.g., I intend to control the BP of my patients with diabetes rated on a seven-point scale from 'strongly disagree' to 'strongly agree'). Responses to the three items will be summed [55]. A third measure of self-reported behaviour will be used. After a period of 12 months a questionnaire will be sent to clinical staff asking them about their behaviour in relation to the same six aspects of care over the preceding twelve months.

Patient-report of clinicians' behaviour

Three dependent variables representing proxy measures of clinicians' behaviour will be generated from a questionnaire survey of a random sample of 100 patients with diabetes from each practice. A self-administered questionnaire will be developed that will ask patients if, over the previous 12 months, they were: offered advice about self-management, and if so, what this advice entailed; received advice about losing or controlling their weight, and if so, what this advice entailed; received or were offered education, and if so what this education entailed.

The content of the questionnaire will include selected sections of the questionnaire used by the healthcare commission in their 2006 national survey of people with diabetes (e.g., self-management and knowledge, education and training). This will provide data that are interpretable in the context of a pre-existing national survey. These questions will be supplemented with specific knowledge questions developed in collaboration with our study team and our local diabetes UK voluntary support group. Patients will also be asked about their uptake of the advice given, and what they did or currently do in terms of acting on it. The final questionnaire will be piloted with patients from the voluntary group.

Administration of data collection

Data relating to practice organisational structure will be collected by telephone interview with the study contact at each recruited practice. This will also identify recipients of the questionnaires and will include members of the administrative staff (e.g., receptionist staff) and clinicians who are attached rather than employed (district nurses), although we are aware that a previous study found it was not possible to gather usable data from this latter group [3].

Questionnaires will be delivered to the practices where the research nurse (or other nominated study contact) will be responsible for their distribution, collection, and return. Reminders will be sent to practices at two-week intervals.

For clinical data held within practice computers, the research nurses will run computer queries to extract data on the total number of patients with diabetes in the practice and the number who have had the relevant clinical actions performed.

Data relating to clinician self-reported behaviour will be measured in the postal questionnaire (above) and, in addition, after a period of 12 months, a second, short questionnaire will be sent to clinical staff asking them about their behaviour in relation to the same six aspects of care over the preceding twelve months.

For the patient-reported data, questionnaires will be sent out to the 100 randomly selected patients from each practice by the nominated study contact at each participating practice. To protect patient confidentiality, a reply paid envelope, addressed to the Institute of Health and Society at Newcastle, will be provided for each patient to return their completed questionnaire directly to the research team. No reminders will be used.

Sample size

For the practice-held, QOF, and clinician self-reported data, the analysis program 'G Power' [56] (using the method described by Cohen [57]) has been used to investigate the sample size required for testing alternative regression models. In this approach, the effect size is defined as the proportion of variance accounted for by a set of predictor variables (anticipated to be 14) relative to the residual variance proportion. In this context, effect sizes of 0.02, 0.15 and 0.35 are considered to be 'small', 'medium' and 'large', respectively. For the analysis of practice level variables (practice behaviour and the aggregated intention and behavioural simulation variables) with 100 practices we will have 80% power to detect an effect size of 0.21 assuming a type one error rate of 5% and 14 predictor variables. Assuming a (worst case) response rate of 75%, we will have data from all relevant staff (100 practices, four GPs, three practice nurses, one practice manager, two receptionists) the estimated sample size will be around 750 which, assuming a type one error rate of 5% and 14 predictor variables will give us 80% power to detect an effect size of 0.03. Thus, in the analyses of individual level data we will have 80% power to detect effect sizes in the small to medium range, and at the practice level we will have 80% power to detect effect sizes in the medium to large range.

For the patient-reported data, making the conservative assumption that the estimated proportion is around 50% if we sample 25 patients per practice, the standard error associated with our estimate for each practice is 10%. We will approach 100 randomly selected patients per practice (10,000 patients in total), allowing for a 25% response rate to achieve a final sample size of 25 per practice.

Statistical analysis

The study will generate data on:
  1. 1.

    Individuals' cognitions about clinical behaviours, clinical conditions, work, team and organisational setting.

     
  2. 2.

    Organisational structure and function.

     
  3. 3.

    Clinical behaviours relating to teams of clinicians.

     
  4. 4.

    Biochemical and physiological measurements that are the consequence of the clinical behaviours, but include consequences of patient physiology and behaviour.

     

The analysis will explore the relationships between the predictor variables (one and two) and the dependent variables of respondents' intentions to perform behaviours, respondent's behavioural simulation, (both from survey instruments), and the clinical behaviours (three) as well as biochemical and physiological measures (four).

Where appropriate, data will be analysed as the standardised (by list size) number of patients who have had the investigation/procedure of interest performed; standardised number of patients who have their level of the investigation/procedure of interest within the accepted target range. From the patient survey we will calculate the proportion of patients in each practice who have a particular attribute (e.g., have been given education or training in self-monitoring).

Where appropriate (i.e., more than two items per measure), the reliability of the measures will be assessed prior to analysis using Cronbach's alpha to assess internal reliability and confirmatory factor analysis to identify and discard redundant items. The construct validity of the measures will be assessed prior to analysis by examining correlations between predictor variables that are expected to be similar (convergent validity) and dissimilar (discriminant validity).

We will examine mediating and moderating effects of the predictor variables on the dependent variables. In principle, we will be dealing mainly with dependent variables measured at the level of the practice and predictor variables measured at either the level of the practice or the level of the individual. The theoretical models will be tested using standard multiple regression analysis and structural equation modelling. For predictor variables that are measured at the level of the individual, we will produce a summary statistic for the practice. We will use the practice mean and, in the multiple regression analysis, will weight by practice size. We will explore other summary statistics with weightings reflecting the relevant roles and responsibilities of the respondents.

In addition to the main study analysis, the patient survey will be reported in its own right, including a comparison with the healthcare commission data (available from their website), a comparison of the level of agreement between patients' and health professionals' perceptions and a comparison of patient perceptions and physiological measures of diabetes control.

Ethics committee review

The study has been approved by Newcastle and North Tyneside Research Ethics Committee Two (REC Ref Number 07/H0907/102).

Declarations

Acknowledgements

The study is funded by Diabetes UK http://www.diabetes.org.uk/.

Authors’ Affiliations

(1)
Institute of Health and Society, University of Newcastle upon Tyne
(2)
Newcastle Diabetes Centre, Newcastle General Hospital
(3)
College of Life Sciences and Medicine, University of Aberdeen
(4)
Health Services Research Unit, University of Aberdeen
(5)
National Institute for Health and Welfare
(6)
Ottawa Health Research Institute
(7)
Department of Medicine, University of Ottawa

References

  1. NHS Information Centre: [http://www.ic.nhs.uk/webfiles/QOF/2006-07/QOF0607_Practice%20Prevalence.xls]
  2. Seddon ME, Marshall MN, Campbell SM, Roland MO: Systematic review of studies of quality of clinical care in general practice in the UK, Australia and New Zealand. QHC. 2001, 10 (3): 152-158. 10.1136/qhc.0100152...PubMedPubMed CentralGoogle Scholar
  3. Bower P, Campbell S, Bojke C, Sibbald B: Team structure, team climate and the quality of care in primary care: an observational study. Quality and Safety in Health Care. 2003, 12 (4): 273-279. 10.1136/qhc.12.4.273.View ArticlePubMedPubMed CentralGoogle Scholar
  4. Davies HTO, Nutley SM, Mannion R: Organisational culture and quality of health care. QHC. 2000, 9: 111-119. 10.1136/qhc.9.2.111.PubMedPubMed CentralGoogle Scholar
  5. Elwyn G, Ryhdderch M, Edwards A, Hutchings H, Marshall M, Myres P, Grol R: Assessing organisational development in primary medical care using a group based assessment: the TM Maturity Matrix. Quality and Safety in Health Care. 2004, 13: 287-294. 10.1136/qshc.2003.008540.View ArticlePubMedPubMed CentralGoogle Scholar
  6. Agency for Health Research and Quality: Translating research into practice (TRIP)-II. 2001, Washington DC: Agency for Health Research and QualityGoogle Scholar
  7. Grol R: Successes and failures in the implementation of evidence-based guidelines for clinical practice. Med Care. 2001, 39 (8 Suppl 2): 1146-1154.Google Scholar
  8. Schuster M, McGlynn E, Brook RH: How good is the quality of health care in the United States?. Milbank Q. 1998, 76: 517-563. 10.1111/1468-0009.00105.View ArticlePubMedPubMed CentralGoogle Scholar
  9. Bero L, Grilli R, Grimshaw JM, Harvey E, Oxman AD, Thomson MA: Closing the gap between research and practice: an overview of systematic reviews of interventions to promote implementation of research findings by health care professionals. BMJ. 1998, 317: 465-468.View ArticlePubMedPubMed CentralGoogle Scholar
  10. NHS Centre for Reviews and Dissemination: Getting evidence into practice. Effect Health Care. 1999, 5 (1): 1-16.Google Scholar
  11. Grimshaw J, Shirran L, E TR, Mowatt G, Fraser C, Bero L: Changing provider behaviour: an overview of systematic reviews of interventions. Med Care. 2001, 39 (8 Suppl 2): II-2-II-45.Google Scholar
  12. Foy R, Eccles M, Jamtvedt G, Grimshaw J, Baker R: What do we know about how to do audit and feedback?. BMC Health Services Research. 2005, 5: 50-10.1186/1472-6963-5-50.View ArticlePubMedPubMed CentralGoogle Scholar
  13. Davies P, Walker A, Grimshaw J: Theories of behaviour change in studies of guideline implementation. Proceedings of the British Psychological Society. 2003, 11 (1): 120-Google Scholar
  14. Grimshaw J, Thomas RE, Maclennan G, Fraser C, Ramsay C, Vale L, Whitty P, Eccles M, Matowe L, Shirren L, Wensing M, Dijkstra R, Donaldson C: Effectiveness and efficiency of guideline dissemination and implementation strategies. Health Technol Assess. 2004, 8 (6): 1-72.View ArticleGoogle Scholar
  15. Eccles M, Grimshaw J, Walker A, Johnston M, Pitts N: Changing the behaviour of healthcare professionals: the use of theory in promoting the uptake of research findings. J Clin Epidemiol. 2005, 58: 107-112. 10.1016/j.jclinepi.2004.09.002.View ArticlePubMedGoogle Scholar
  16. Bradshaw C, Eccles MP, Choi HY: Can General Practitioners assess diabetic control? Can anybody do any better?. Diabetic Med. 1991, 8: 543-546.View ArticlePubMedGoogle Scholar
  17. Bradshaw C, Steen IN, Eccles M: Glycated haemoglobin levels in patients with diabetes in one general practice over a ten year period. Diabetic Med. 1995, 12: 628-631.View ArticlePubMedGoogle Scholar
  18. Bradshaw C, McColl E, Eccles M, Bryce C, Sampson R: Can we improve the education for type 2 diabetes patients in general practice?. Pract Diabetes Int. 1999, 16: 241-245. 10.1002/pdi.1960160807.View ArticleGoogle Scholar
  19. Eccles M, McColl E, Steen N, Rousseau N, Grimshaw J, Parkin D, Purves I: Effect of computerised evidence based guidelines on management of asthma and angina in adults in primary care: cluster randomised controlled trial. BMJ. 2002, 325 (7370): 941-944. 10.1136/bmj.325.7370.941.View ArticlePubMedPubMed CentralGoogle Scholar
  20. Eccles M, Steen N, Grimshaw J, Thomas L, McNamee P, Soutter J, Wilsdon J, Matowe L, Needham G, Gilbert F, Bond S: Effect of audit and feedback, and reminder messages on primary-care radiology referrals: a randomised trial. Lancet. 2001, 357: 1406-1409. 10.1016/S0140-6736(00)04564-5.View ArticlePubMedGoogle Scholar
  21. Thomas RE, Croal BL, Ramsay C, Eccles M, Grimshaw J: Effect of enhanced feedback and brief educational reminder messages on laboratory test requesting in primary care: a cluster randomised trial. Lancet. 2006, 367 (9527): 1990-1996. 10.1016/S0140-6736(06)68888-0.View ArticlePubMedGoogle Scholar
  22. Eccles M, Hawthorne G, Whitty P, Steen N, Vanoli A, Grimshaw J, Wood L: A randomised controlled trial of a patient based Diabetes Recall and Management System: the DREAM Trial: A study protocol [ISRCTN32042030] (21 March 2002). BMC Health Services Research. 2002, 2: 5-10.1186/1472-6963-2-5.View ArticlePubMedPubMed CentralGoogle Scholar
  23. Freemantle N, Nazareth I, Eccles M, Wood J, Haines A: Trialists tE-bOE: A randomised trial of the effect of educational outreach by community pharmacists on prescribing in UK general practice. Br J Gen Pract. 2002, 52: 290-295.PubMedPubMed CentralGoogle Scholar
  24. Hall L, Eccles M, Steen N, Barton R, Campbell M: Pragmatic randomised control trial of outreach visiting as a strategy to implement guidelines for the eradication of Helicobacter pylori in primary care, evaluated using routinely available PACT data. 1999, University of Newcastle: Centre for Health Services Research, 100:Google Scholar
  25. Hall L, Eccles M, Barton R, Steen N, Campbell M: Is untargeted outreach visiting in primary care effective? A pragmatic randomized controlled trial. J Pub Health Med. 2001, 23 (2): 109-113. 10.1093/pubmed/23.2.109.View ArticleGoogle Scholar
  26. Hall L, Eccles M: Case study of an inter-professional and inter-organisational programme to adapt, implement and evaluate clinical guidelines in secondary care. Clin Perform Qual Health Care. 2000, 8 (2): 72-82. 10.1108/14664100010343863.View ArticlePubMedGoogle Scholar
  27. Parr JH, Bradshaw C, Broderick W, Courtenay H, Eccles M, Murray E, Royle J, Whitty P: Improving the use of aspirin in myocardial infarction: a district strategy. British Journal of Clinical Governance. 1999, 4: 24-27. 10.1108/14664109910297155.View ArticleGoogle Scholar
  28. Bonetti D, Eccles M, Johnston M, Steen IN, Grimshaw J, Baker R, Walker A, Pitts N: Guiding the design and selection of interventions to influence the implementation of evidence-based practice: an experimental simulation of a complex intervention trial. Soc Sci Med. 2005, 60: 2135-2147. 10.1016/j.socscimed.2004.08.072.View ArticlePubMedGoogle Scholar
  29. Bonetti D, Pitts NB, Eccles M, Grimshaw J, Steen N, Glidewell L, Thomas R, Maclennan G, Clarkson JE, Walker A: Applying psychological theory to evidence-based clinical practice: identifying factors predictive of taking intra-oral radiographs. Soc Sci Med. 2006, 63: 1889-1899. 10.1016/j.socscimed.2006.04.005.View ArticlePubMedGoogle Scholar
  30. Eccles M, Foy R, Bamford C, Hughes J, Whitty P, Steen N, Grimshaw J: A trial platform to develop a tailored theory based intervention to improve professional practice in the disclosure of a diagnosis of dementia. Implementation Science. 2006, 1: 7-10.1186/1748-5908-1-7.View ArticlePubMedPubMed CentralGoogle Scholar
  31. Elovainio M, Makela M, Sinervo T, Kivimaki M, Eccles M, Kahan J: Effects of job characteristics, team climate, and attitudes towards clinical guidelines. Scandinavian Journal of Public Health. 2000, 28 (2): 117-122.View ArticlePubMedGoogle Scholar
  32. Elovainio M, Kivimaki M, Eccles M, Sinervo T: Team climate and procedural justice as predictors of occupational strain. J Appl Psychol. 2002, 32 (2): 359-374. 10.1111/j.1559-1816.2002.tb00220.x.Google Scholar
  33. Kivimaki M, Elovainio M, Vahtera J, Ferrie JE: Organisational justice and health of employees: prospective cohort study. Occup Environ Med. 2003, 60: 27-34. 10.1136/oem.60.1.27.View ArticlePubMedPubMed CentralGoogle Scholar
  34. Eccles MP, Grimshaw J, Johnston M, Steen IN, Pitts NB, Thomas R: Applying psychological theories to evidence-based clinical practice: Identifying factors predictive of managing upper respiratory tract infections without antibiotics. Implementation Science. 2007, 2: 26-10.1186/1748-5908-2-26.View ArticlePubMedPubMed CentralGoogle Scholar
  35. Walker A, Grimshaw JM, Johnston M, Pitts N, Steen N, Eccles MP: PRocess modelling in ImpleMEntation research:selecting a theoretical basis for interventions to change clinical practice. BMC Health Services Research. 2003, 3: 22-10.1186/1472-6963-3-22.View ArticlePubMedPubMed CentralGoogle Scholar
  36. Michie S, Johnston M, Abraham C, Lawton R, Parker D, Walker A: 'Psychological Theory' Group. Making psychological theory useful for implementing evidence based practice: a consensus approach. Quality & Safety in Health Care. 2005, 14 (1): 26-33. 10.1136/qshc.2004.011155.View ArticleGoogle Scholar
  37. Bandura A: Self-efficacy: towards a unifying theory of behaviour change. Psychological Review. 1977, 84: 191-215. 10.1037/0033-295X.84.2.191.View ArticlePubMedGoogle Scholar
  38. Ajzen I: The theory of planned behaviour. Organizational Behaviour and Human Decision Processes. 1991, 50: 179-211. 10.1016/0749-5978(91)90020-T.View ArticleGoogle Scholar
  39. Verplanken B, Orbell S: Reflections on past behavior: a self-report index of habit strength. Journal of Applied Social Psychology. 2003, 33: 1313-1330. 10.1111/j.1559-1816.2003.tb01951.x.View ArticleGoogle Scholar
  40. Sniehotta FF, Schwarzer R, Scholz U, Schuz B: Action planning and coping planning for long-term lifestyle change: theory and assessment. Eur J Soc Psychol. 2005, 35: 565-576. 10.1002/ejsp.258.View ArticleGoogle Scholar
  41. Godin G, Gagnon H, Alary M, Noel L, Morissette MR: Correctional officers' intention of accepting or refusing to make HIV preventive tools accessible to inmates. AIDS Educ Prev. 2001, 13: 462-473. 10.1521/aeap.13.5.462.24145.View ArticlePubMedGoogle Scholar
  42. Gollwitzer PM: Goal achievement: the role of intentions. European review of social psychology. Edited by: Stroebe W, Hewstone M. 1993, Chichester, UK: Wiley, 141-185.Google Scholar
  43. Orbell S, Hodgkins S, Sheeran P: Implementation intentions and the theory of planned behaviour. Personality and Social Psychology Bulletin. 1997, 23: 945-954. 10.1177/0146167297239004.View ArticleGoogle Scholar
  44. Gollwitzer PM: The volitional benefits of planning. The psychology of action: linking cognition and motivation to behaviour. Edited by: Gollwitzer Pm, Bargh JA. 1996, New York: Guilford Press, 287-312.Google Scholar
  45. Karasek RA: Job content questionnaire and user's guide. 1979, University of Southern Carolina. Los AngelesGoogle Scholar
  46. West MA, Poulton BC: A failure of function: teamwork in primary health care. J Interprof Care. 1997, 11: 205-216.View ArticleGoogle Scholar
  47. Kramer RM, Tyler TR: Trust in organizations: frontiers of theory and research. 1996, London: SageGoogle Scholar
  48. Elovainio M, Kivim„ki M, Helkama K: Organizational justice evaluations, job control and occupational strain. J Appl Psychol. 2001, 86: 418-424. 10.1037/0021-9010.86.3.418.View ArticlePubMedGoogle Scholar
  49. Karasek : Job content questionnaire and user's guide. Volume Revision 1.1.1985. 1985Google Scholar
  50. Siegrist J: Adverse health effects of high-effort low-reward conditions. Journal of Occupational Health Psychology. 1996, 1: 27-41. 10.1037/1076-8998.1.1.27.View ArticlePubMedGoogle Scholar
  51. Kivimaki M, Elovainio M: A short version of the Team Climate Inventory: development and psychometric properties. Journal of Occupational Health Psychology. 1998, 72: 241-246.Google Scholar
  52. Moorman RH: Relationship between organizational justice and organizational citizenship behaviors: do fairness perception influence employee citizenship?. J Appl Psychol. 1991, 76: 845-855. 10.1037/0021-9010.76.6.845.View ArticleGoogle Scholar
  53. Elovainio M, Kivimaki M, Vahtera J: Organizational justice: evidence of a new psychosocial predictor of health. Am J Public Health. 2002, 92: 1058-10.2105/AJPH.92.1.105.View ArticleGoogle Scholar
  54. NHS Information Centre: [http://www.ic.nhs.uk/webfiles/QOF/2007-08/NewFilesGS/QOF0708_Pracs_Diabetes.xls]
  55. Francis J, Eccles MP, Johnston M, Walker AE, Grimshaw JM, Foy R, Kaner EFS, Smith L, Bonetti D: Constructing questionnaires based on the theory of planned behaviour. A manual for health services researchers. 2004, Newcastle upon Tyne: Centre of Health Services Research, University of Newcastle upon TyneGoogle Scholar
  56. Erdfelder E, Faul F, Buchner A: GPOWER: A general power analysis program. Behavior Research Methods, Instruments and Computers. 1996, 28: 1-11.View ArticleGoogle Scholar
  57. Cohen J: Statistical power analyses for the behavioral sciences. 1977, Acad PressGoogle Scholar

Copyright

© Eccles et al; licensee BioMed Central Ltd. 2009

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.