This article has Open Peer Review reports available.
Implementing health research through academic and clinical partnerships: a realistic evaluation of the Collaborations for Leadership in Applied Health Research and Care (CLAHRC)
© Rycroft-Malone et al; licensee BioMed Central Ltd. 2011
Received: 17 February 2011
Accepted: 19 July 2011
Published: 19 July 2011
The English National Health Service has made a major investment in nine partnerships between higher education institutions and local health services called Collaborations for Leadership in Applied Health Research and Care (CLAHRC). They have been funded to increase capacity and capability to produce and implement research through sustained interactions between academics and health services. CLAHRCs provide a natural 'test bed' for exploring questions about research implementation within a partnership model of delivery. This protocol describes an externally funded evaluation that focuses on implementation mechanisms and processes within three CLAHRCs. It seeks to uncover what works, for whom, how, and in what circumstances.
Design and methods
This study is a longitudinal three-phase, multi-method realistic evaluation, which deliberately aims to explore the boundaries around knowledge use in context. The evaluation funder wishes to see it conducted for the process of learning, not for judging performance. The study is underpinned by a conceptual framework that combines the Promoting Action on Research Implementation in Health Services and Knowledge to Action frameworks to reflect the complexities of implementation. Three participating CLARHCS will provide in-depth comparative case studies of research implementation using multiple data collection methods including interviews, observation, documents, and publicly available data to test and refine hypotheses over four rounds of data collection. We will test the wider applicability of emerging findings with a wider community using an interpretative forum.
The idea that collaboration between academics and services might lead to more applicable health research that is actually used in practice is theoretically and intuitively appealing; however the evidence for it is limited. Our evaluation is designed to capture the processes and impacts of collaborative approaches for implementing research, and therefore should contribute to the evidence base about an increasingly popular (e.g., Mode two, integrated knowledge transfer, interactive research), but poorly understood approach to knowledge translation. Additionally we hope to develop approaches for evaluating implementation processes and impacts particularly with respect to integrated stakeholder involvement.
Despite considerable investment in the generation of research, for the most part it is not routinely used in practice or policy [1–4]. In the United Kingdom (UK), a national expert group reviewed the implementation research agenda and recommended sustained and strategic investment in research and infrastructure aimed at increasing our capability and capacity to maximise the impact of health research . The group also recommended that implementation researchers and implementation research should be embedded within health services [6–11]. In response to the recommendations of Clinical Effectiveness Research Agenda Group (CERAG), there has been a major investment in nine partnerships between higher education institutions and local health services within the English National Health Service (NHS) [12, 13]. The Collaborations for Leadership in Applied Health Research and Care (CLAHRC) are funded by the National Institute for Health Research (NIHR) to produce and implement research evidence through sustained interactions between academics and services (see Additional File 1 for more information about the CLAHRC concept). The establishment of the CLAHRCs and their explicit remit for closing the gap between research and practice provides a natural 'experiment' for exploring and evaluating questions about research implementation within a partnership model. This protocol describes one of four externally funded evaluations of CLAHRC (NIHR SDO 09/1809/1072).
Implementing research in practice
Health services are more or less informed by the findings of research [14–19]. The Cooksey Report  distinguishes between two gaps in knowledge translation: the 'first' gap between a scientist's bench to product/process/service, and the 'second' gap, their routine use in practice. It is the second gap that has been neglected and provides the focus for our evaluation. Specifically, we are interested in exploring implementation in its broadest sense. This breadth includes acknowledging that information and knowledge comes in many forms, such as research, audit data, patient and public involvement, and practice know how, which variably inform decision making and service delivery. We treat research implementation and knowledge translation as related concepts, sharing a largely common literature and theory base. Both concern closing the gap between what is known from research and implementation of this by stakeholders pursuing improved health outcome and experiences.
Implementation is a slow, complex and unpredictable process [14, 15, 21–27]. The rational-logical notion that producing research, packaging it in the form of guidelines and assuming it will automatically be used is now outdated. There is a substantial body of evidence showing that using research involves significant and planned change involving individuals, teams, organisations and systems [14, 22–24, 28–33]. One meta-synthesis of case studies showed that adopting knowledge depends on a set of social processes that include sensing and interpreting new evidence, integrating it with existing evidence; reinforcement (or not) by professional networks, which in turn is mediated by local context , including the contribution that patients and the public make.
Context is emerging as a significant influence on knowledge flow and implementation. Micro, meso and macro contextual influences  include factors such as financial and human resources [14, 15, 31], structure , governance arrangements , culture [27, 35–38], power [38, 39], and leadership [22, 23, 28, 33, 35, 40]. Such factors appear to influence an organisation's capacity to manage, absorb, and sustain knowledge use . However we do not know whether some contextual factors are more influential than others, or how they operate and change over time.
Networks and communities of practice  may also play an important role in both the flow and use of evidence [14, 23, 41–45]. Multi-disciplinary communities of practice have been found to transform research evidence through interaction and collective sense making, such that other forms of knowledge (e.g., practice know how) become privileged [44, 45]. Whilst communities of practice are intuitively appealing, there is little empirical research to support claims that they actually increase knowledge uptake in health services [46–48]. There is evidence to suggest that communities of practice show promise as a means of creating and sharing knowledge that has meaning for practitioners , however little is known about the mechanisms by which this may occur. There is an opportunity within this study to explore the relevance of communities of practice to the implementation of research, what mechanisms and processes may be at work, and the role that patients and the public may play in this.
'Boundary objects' may facilitate or inhibit knowledge flow [50–54]. Typically boundary objects are representations, abstractions, or metaphors that have the power to 'speak to' different communities of practice by sharing meaning and learning about each others' perspectives and by acting as (temporary) anchors or bridges [50–54]. The theory of 'boundary objects' has importance in exploring the translation of meaning from one setting to another. Objects have the capability to be understood by actors in more than one setting, for example, between different departments, doctors and nurses, researchers and users, and practitioners and patients. We are interested in finding out whether such boundary objects exist in the CLAHRCs and the NHS communities they serve-and if they do, what do they look like and how are they being used, particularly in relation to implementation.
To date, funders and policy makers have focused on the generation of research knowledge to the relative neglect of how research is used in practice. A number of NHS initiatives including Academic Health Science Centres, Health Innovation and Education Clusters, and Quality Observatories are emerging that could help bridge research and practice. However the CLAHRCs have an explicit remit for closing the gap in translation. Implementation has generally been studied through one-off, retrospective evaluations that have not been adequately theorised, which leaves many questions unanswered. This study is a theory driven, longitudinal evaluation of research implementation within CLAHRCs and will address some critical gaps in the literature about increasing applied health research use.
We are exploring how research is implemented within CLAHRCs through the following aims and objectives.
To inform the NIHR SDO programme about the impact of CLAHRCs in relation to one of their key functions: 'implementing the findings from research in clinical practice.'
To make a significant contribution to the national and international evidence base concerning research use and impact, and mechanisms for successful partnerships between universities and healthcare providers for facilitating research use.
To work in partnership so that the evaluation includes stakeholder perspectives and formative input into participating CLAHRCs.
To further develop theory driven approaches to implementation research and evaluation.
To identify and track the implementation mechanisms and processes used by CLAHRCs and evaluate intended and unintended consequences (i.e., impact) over time.
To determine what influences whether and how research is used or not through CLAHRCs, paying particular attention to contextual factors.
To investigate the role played by boundary objects in the success or failure of research implementation through CLAHRCs.
To determine whether and how CLAHRCs develop and sustain interactions and communities of practice.
To identify indicators that could be used for further evaluations of the sustainability of CLAHRC-like approaches.
Methodology and methods
This study is a longitudinal three-phase, multi-method evaluation, which deliberately aims to explore the boundaries between knowledge use in practice. The evaluation, as expressed by the funder, is being conducted for the process of learning, not for judgement. Given the processual and contextual nature of knowledge use and our objectives, realistic evaluation is our overarching methodology . Realistic evaluation is an approach that is underpinned by a philosophy of realism that recognises reality as a construction of social processes. Thus realists attempt to understand complex social interactions/interventions. Complex social interventions according to Pawson and Tilley [60, 61] are comprised of theories, involve the actions of people, consist of a chain of steps or processes that interact and are rarely linear, are embedded in social systems, prone to modification and exist in open, dynamic systems that change through learning. As such, realistic evaluation offers a means of understanding network-based approaches such as CLAHRCs, which by their nature are social systems, involve the actions of people and groups, and which are likely to change over time. Realistic evaluation is also a useful approach for capturing contextual influences and changes at multiple levels over time because of the cyclical approach to evaluation.
Others have successfully used realistic evaluation to evaluate complex, system, and network orientated initiatives [e.g., [62, 63]] and in implementation related research [64–66]. For example Greenhalgh and colleagues  evaluated a whole-system transformation in four large healthcare organisations in London. They identified implementation mechanisms and sub-mechanisms, with associated enabling and constraining factors, which included networks (hard and soft), evidence, structures, contracts, governance, and roles http://axisto.com/webcasting/bmj/berlin-2009/plenary-3/index.htm). Additionally, Sullivan and colleagues  successfully used realistic evaluation to evaluate a national initiative in which they specified the types and levels of collaborative activity necessary to deliver Health Action Zone objectives. Rycroft-Malone et al.[64–66] conducted a realistic evaluation of the mechanisms and impact of protocol-based care within the NHS. There are growing numbers of researchers engaged in realistic evaluation research (for example [67–69]), this evaluation provides a further opportunity to test and develop the approach.
Within realism, theories are framed as propositions about how mechanisms act in contexts, to produce outcomes. Realistic evaluation is particularly relevant for this study because it aims to develop explanatory theory by acknowledging the importance of context to the understanding of why interventions and strategies work. Programmes (i.e., CLAHRC implementation) are broken down so that we can identify what it is about them (mechanisms) that might produce a change (impact), and which contextual conditions (context) are necessary to sustain changes. Thus, realistic evaluation activity attempts to outline the relationship between mechanisms, context, and outcomes.
Instrumental use: the direct impact of knowledge on practice and policy in which specific research might directly influence a particular decision or problem.
Conceptual use: how knowledge may impact on thinking, understanding, and attitudes.
Symbolic use: how knowledge may be used as a political tool to legitimatise particular practices.
This proposal has been developed by a team including participants from four CLAHRCs (RB, CT, GH, GM, and SA). Their involvement from the outset ensures the evaluation is addressing questions of interest, is feasible, and offers opportunities for mutual learning and benefit. We recognise that those being evaluated being part of the evaluation team, whilst consistent with an interactive approach [73–77] calls for particular attention to issues of rigour. Sociological and anthropological research, utilisation-focused evaluation, and participant action research have a longstanding tradition of including 'insiders' [78, 79]. An insider perspective will provide insight and enable us to crosscheck face validity of data against the experience of operating within a CLAHRC context. Our approach is consistent with the principles upon which the CLAHRCs were created, and the proposed methods have their own criteria for rigour and integrity [80, 81]. However, we acknowledge that the evaluation, through its activities and formative input might influence how participating CLAHRCs approach implementation over time. We have therefore built in a process for monitoring any cross fertilisation of ideas and their potential impact (see section below for more information).
Phases and methods
In keeping with utilisation-focused evaluation principles  our plan integrates ongoing opportunities for interaction between the evaluation team, three participating CLAHRCs, and the wider CLAHRC community to ensure findings have programme relevance and applicability.
Realistic evaluation case studies
The three participating CLARHCS provide an opportunity to study in-depth comparative case studies of research implementation . We have focussed on three CLAHRCs because it would not be practically possible to capture the in-depth data required to meet study aims and objectives across all nine CLAHRCs. However, there are opportunities throughout the evaluation for the wider CLAHRC community to engage in development and knowledge sharing activities (participating CLAHRCs are described in more detail in Additional Files 2, 3 and 4).
A 'case' is implementation [theme/team] within a CLAHRC and the embedded unit, particular activities/projects/initiatives related to a tracer issue . These cases represent a natural sample of the CLAHRCs as each has planned a different approach to implementation. Sampling is based on a theoretical replication argument; it is anticipated that each CLAHRC will provide contrasting results, for predictable reasons .
Phase one: Hypotheses generation (up to 18 months)
In this first phase, we will: develop good working relationships and establish ways of working with participating CLAHRCs; develop an evaluation framework that will provide a robust theoretical platform for the study; and map mechanism-context-outcome (MCO) links and generate hypotheses, i.e., what might work, for whom, how, and in what circumstances.
Establishing ways of working
We recognise the importance of establishing good working relationships and clear ways of working with the CLAHRC communities. During the early stages of this project, we are working with CLAHRCs to agree on ways of working and have developed a memorandum of understanding to which each party is happy to commit (see Additional File 5).
Development of evaluation framework and mapping mechanism-context-outcome links
In order to explore and describe the links between research and its implementation a 'theoretical map' of what CLAHRCs have planned concerning implementation is needed, which is incorporated into the study's evaluation framework. We will collect documentary evidence such as strategy documents, proposals and implementation plans, and other evidence. Drawing on the research implementation literature, we will discuss implementation and internal evaluation plans with each CLAHRC. Once gathered, we will analyse and synthesise the data using concept mining, developing analytical themes and framework development. The framework will yield what approaches and mechanisms each CLAHRC intends to be used for implementation, in what settings, with whom and to what affect.
Using the output of the documentary analysis, we will hold discussions with relevant stakeholders (i.e., CLAHRC participants, NHS staff linked to CLAHRC projects, service user group, research team) to develop and refine MCO links, i.e., the evaluation's hypotheses (for example, 'The translation and utilisation of knowledge in and through CLAHRCs and the resulting range of impacts will be dependent upon the different types of knowledge that are given attention and valued' and 'The impact of translation and implementation of knowledge in and through CLAHRCs will be dependent upon the adoption and use of appropriate facilitation approaches, including individuals in formal and informal roles'). We will then ensure that the hypotheses are shared across all nine CLAHRCs. This will provide another opportunity to scrutinise the credibility and representativeness of our hypotheses across contexts, and also to share knowledge that could be used more widely by CLAHRC programme participants.
To provide a focus for testing the hypotheses, we will work with the three CLAHRCs to determine what topics would be appropriate to become tracer issues. Criteria of choice will include the potential to have greatest impact in practice, examples from the increased uptake of existing evidence as well as new evidence being generated through CLAHRCs, and that might provide the most useful formative information for CLAHRCs and summative data for this evaluation. We anticipate that at least one of the tracer issues will be common to all three CLAHRCs to enable greater comparison.
Using available documents and our discussion with CLAHRC teams, we will map the clinical and implementation issues being addressed within and across each CLAHRC. Once these have been mapped, we will reach consensus with them about which topics become tracer issues. Tracer issues may not necessarily be clinical issues, but it is likely that the projects we focus on for in-depth study will have a particular clinical focus (e.g., nutrition care, diabetes, stroke, kidney disease, long-term conditions). For example, one tracer issue could be change agency, the focus of in-depth study within a particular CLAHRC could then be the role of knowledge brokering in the implementation of improved service delivery for patients with chronic kidney disease.
Phase two: Studying research implementation over time-testing hypotheses (up to 28 months)
We will test the hypotheses developed in phase one against what happens in reality within each CLAHRC case and tracer issue (i.e., what is working (or not), for whom, how, and in what circumstances) over time. We will focus on specific projects/initiatives/activities within the tracer issues and conduct in-depth case studies on these.
To facilitate description, explanation, and evaluation, within each site multiple data collection methods will be used in order to identify different impacts or types of knowledge use as shown in Additional File 6. During phase one, we will negotiate the details and timings of phase two data collection activity, which will be dependent on the stages of CLAHRC development and other factors that are influencing CLAHRCs (e.g., health service re-organisations). Being guided by our evaluation framework, objectives, and MCOs, we will aim to capture data at critical points in the implementation pathways of tracer issues. We plan for data collection and analysis to be iterative and cyclical; checking our observations against MCOs, and feeding this information back to participating sites as formative input (what seems to be working (or not), for whom, how, and in what circumstances). There will be four rounds of data collection and MCO refining over 28 months.
We will draw on the following data collection methods as appropriate for each in-depth study.
We will conduct semi-structured interviews with stakeholders at multiple levels within and across the particular project/initiative (e.g., role of knowledge brokering in the implementation of improved service delivery for patients with chronic kidney disease). A sampling framework for interviews will be developed based on a stakeholder analysis . Using both theoretical and criterion sampling, we will determine which stakeholders are 'essential,' 'important,' and/or 'necessary' to involve . We will commence interviews with a representative sample of essential stakeholders, and further stakeholders will be interviewed from the other two categories based on theoretical sampling. Criterion sampling will be used to ensure the inclusion of a variety of stakeholders with criteria being developed to include different roles, length of involvement for example, in CLAHRCs.
Interviews will focus on perceptions about what is influencing implementation efforts, the content of which will be informed by MCOs and evaluation framework, as well as participant-driven issues. We are interested in exploring stakeholder perceptions of both the intended and unintended consequences or impact of implementation. As appropriate, interviews will be conducted either face-to-face or by telephone, and will be audio-recorded. The number of interviews conducted will be determined on a case-by-case basis, but is likely to be up to 20 in each case studied at each round of data collection.
Focussed observation of a sample of tracer issue community of practice activities and team interactions (e.g., between implementers and users, planning and implementation meetings) will be undertaken at appropriate points throughout this phase. We will identify a range of 'events' that could be observed and map these against our objectives to identify appropriate sampling. These observations will focus on interactions and be informed by an observation framework developed from Spradley's  nine dimensions of observation, including space, actors, activities, objects, acts, events, time, goals, and feelings. Observations will be written up as field notes.
Routine and project-related data
As appropriate to the topic and outcomes of interest, we will draw on data being gathered by CLAHRCs, which they are willing to share. It is difficult to anticipate which data may be informative at this stage, but it could include implementation plans, ethics and governance applications, findings from specific implementation efforts and measures of context, minutes of meetings, internal audit data, cost data, and evidence of capacity and capability building (e.g., research papers, staff employment, new roles, research activity). We will negotiate access to such information on a case-by-case basis.
Publicly available data
Because CLAHRCs are regional entities and over time their impact might be realised at a population level, publically available information relevant to the tracer issues from Public Health Observatories and the Quality and Outcome Framework for general practitioners (for example) in participating CLAHRC areas could be a useful source of information. These data could be mined and tracked over time, and compared to data from non-CLAHRC areas; specifically, we are interested in exploring data from regions that were not successful in the CLAHRC application process. Whilst we recognise there will be a time lag in realising an impact of CLAHRC activity, these data have the potential to help our understanding about the effect of CLAHRCs on population health outcomes.
We will gather and analyse documentary material relevant to: implementation, generally in relation to CLARHC strategy and approaches, and specifically with respect to the tracer issue and related project/initiative' context of implementation (e.g., about wider initiatives, success stories, critical events/incidents, outputs, changes in organisation.); and CLAHRC internal evaluation plans. These materials may include policies, minutes of meetings, relevant local/national guidance, research/development/quality improvement papers, newspaper stories, job adverts, and reports (e.g., about the CLAHRC programme more widely). These will provide information with which to further contextualise findings, provide insight into influences of implementation, and help explanation building.
Evaluation team reflection and monitoring
Including key CLAHRC staff as research collaborators and the provision of formative learning opportunities will enable CLAHRCs to critically review (and potentially adapt) their implementation strategy and activities. In this respect, knowledge will be produced within a context of application, which requires nuanced approaches to establishing research quality . The insider perspective from members of the research team will provide additional insights and enable us to crosscheck face validity of findings against the experience of operating within a CLAHRC context. A range of benchmarks (e.g., immersion in the field, member-checking, audit trail) are available to demonstrate transparency in the interpretation of study findings. However, additional strategies to establish research quality are required that accommodate for the (potential) adaptation of CLAHRC's implementation programmes occurring through the cycle of learning and teaching described earlier. An information management strategy (including accurate record keeping, document version control, and information flow charts) will be established to allow a real time record of (codifiable) information sharing within the research team and with CLAHRCs. Once information flows are established, then it will be possible to explore the impacts of specific information sharing (e.g., progress reports) in targeted interviews. Research team meetings will provide an important opportunity to adopt a reflexive approach to the discussion of the potential and actual impacts of findings within CLAHRCs through recording and observations of these meetings, and the maintenance of an evaluation team critical event diary. We will take a reflexive approach to meetings and ensure consideration of how our approach and/or contact may have influenced CLAHRC activity. As metadata, this information will be used in two ways: as a contribution to understanding implementation processes and influences; and to evaluate our decisions and actions to better understand how to conduct evaluations such as this in the future.
Phase three: Testing wider applicability (up to six months)
Closing the realistic evaluation loop (Figure 2), we will test the wider applicability of findings emerging from phases one and two (see section below for analysis process) with a wider community. We will hold a joint interpretative forum-an opportunity for different communities to reflect on and interpret information from data collection efforts-enabling the surfacing of different viewpoints and knowledge structures for collective examination .
Members from relevant communities, including participants from all nine CLAHRCs, representatives from other initiatives such as Academic Health Science Centres, researchers and practitioners, service user representatives, policy makers, funders, commissioners, and managers interested in research implementation and impact will be invited. We will use our international networks to broaden the scope of attendance beyond the UK.
Using interactive methods and processes, and facilitated by an expert, we will test out our emerging theories about what works, for whom, how, and in what circumstances. Participants will be given the opportunity to challenge and interpret these from the position of their own frame of reference. We will capture workshop data through appropriate multimedia, such as audio recording, images, and documented evidence. These data will be used to refine theory.
This phase will provide an opportunity to maximize the theoretical generalisability of findings, will serve as a knowledge-transfer activity, and provide an opportunity to develop the potential for international comparison. The outputs of the forum will also be translated into a web-based resource for open access.
The focus of analysis will be on developing and refining the links between mechanisms, context and outcomes (i.e., hypotheses testing and refining) to meet study objectives. As a multi-method comparative case study, we will use an analysis approach that draws on Yin , Miles and Huberman , and Patton . As this is a longitudinal evaluation, teasing out MCO configurations/interactions will involve an ongoing process of analysis, and be undertaken by various members of the team to ensure the trustworthiness of emerging themes. For each MCO, evidence threads will be developed from analysing and then integrating the various data; the fine-tuning of MCOs is a process that ranges from abstraction to specification, including the following iterations.
We will develop the theoretical propositions/hypotheses (with CLAHRCs in phase one around objectives, theories, and conceptual framework)-these MCOs are at the highest level of abstraction-what might work, in what contexts, how and with what outcomes, and are described in broad/general terms, e.g., 'CLAHRC partnership approach' (M1), is effective (O1) at least in some instances (C1, C2, C3).
As data are gathered through phase two, data analysis and integration facilitates MCO specification ('testing') that will be carried out in collaboration with CLAHRCs. That is, we will refine our understanding of the interactions between M1, O1, C1, C2, and C3. For example, data analysis shows that in fact there appear to be particular approaches to partnerships (now represented by M2), that have a specific impact on increased awareness of research evidence by practitioners (now represented by O2), only in instances in teams where there is multi-disciplinary working (an additional C, now represented by C4). This new MCO configuration (i.e., hypothesis) can then be tested in other settings/contexts/sites seeking disconfirming or contradictory evidence.
Cross-case comparisons will determine how the same mechanisms play out in different contexts and produce different outcomes. This will result in a set of theoretically generalisable features addressing our aims and objectives.
Consistent with comparative case study each case is regarded as a 'whole study' in which convergent and contradictory evidence is sought and then considered across multiple cases. A pattern matching logic, based on explanation building will be used [81, 87]. This strategy will allow for an iterative process of analysis across sites, and will enable an explanation about research implementation to emerge over time, involving discussions with the whole team. Analysis will first be conducted within sites, and then to enable conclusions to be drawn for the study as a whole, findings will be summarised across the three sites [81, 82]. Our evaluation and theoretical framework will facilitate data integration.
While some ambiguity exists in relation to the definitions of quality improvement, implementation research, and evaluation projects in relation to the need for formal ethical approval [88, 89], this study will be generating primary data. Following the principles of good research practice [90, 91], ethical approval will be sought from a multi-site research ethics committee for data collection from phase two onwards. The nature of the evaluation as an iterative and interactive process may necessitate a phased application to research ethics in order to provide the necessary detail for each round of data collection.
In line with good research practice , we will adhere to the following principles.
Whilst CLAHRCs as a whole are contractually obliged to engage in external evaluation activities, the participation of individuals in this study is voluntary. Participants will be provided with written information about the evaluation and details of the nature and purpose of the particular data-collection activities before being asked to provide written consent to participate. They will have the right to withdraw consent at any point without giving a reason. We recognise that in research of this nature, there is always scope for exposing issues of concern, for example, poor quality of practice or service failings. Should issues of this nature occur in the course of data collection, the participant would be made aware that the researcher, following research governance and good research practice guidance [90–92], would discuss these in the first instance with the study principal investigator and further action taken as necessary.
Confidentiality and anonymity
Participants will be known to the researchers gathering primary data, but beyond this, they will be assigned codes and unique identifiers to ensure and maintain anonymity. Where individuals are recognisable due to information provided in, for example, audio-recorded interviews, at the point of transcription a process of anonymising will be used to ensure that they are not recognisable. As it may be possible to identify staff who hold unique or unusual roles if their job title were used in the written reporting of data, alternative ways of recording these will be used, such a providing a general title to protect their anonymity. Details of the codes will be stored according to good practice and research governance requirements [90, 91].
Data management and storage
Documentary data, interview transcriptions, and fieldwork diaries will be stored securely. Only the principal investigator and research fellow will have access to primary data. Back-up copies of interviews will be stored separately, but in the same manner and all data kept on a password-protected computer.
There have been discussions with CLAHRC directors at an early stage about ensuring burden and disruption are minimised, and this has been formalised in the memorandum of understanding (see additional file 5). We will therefore negotiate and agree the practicalities of data collection at each phase and round of data collection at a local level. Our study design allows us to take a flexible approach with the potential for amendment as necessary to reflect changing circumstances in each CLAHRC. Wherever possible, our evaluation will complement those being undertaken internally by each CLAHRC and with the three other NIHR SDO Programme evaluation teams.
The rationale underpinning the investment in the CLAHRC initiative and the theory on which they have been established is that collaboration between academics and practitioners should lead to the generation of more applied research, and a greater chance that research will be used in practice . Despite a growing interest and belief in this theory , it has yet to be fully tested. This study has been designed to explore the unknown, as well as build on what is already known about research implementation within a collaborative framework through a theory and stakeholder driven evaluation.
Currently there are plans for a radical change in the way that healthcare is commissioned, planned, and delivered within the NHS . Policy changes will mean fundamental shifts to the way some CLAHRCs are managed and funded, which have the potential to create a very different context for them, and a significantly different evaluation context for us. For example, the introduction of competition within a local health economy may result in fragmentation and a tendency to be less open and collaborative-the antitheses of the philosophy upon which CLAHRCs were established. Realistic evaluation provides an ideal approach for monitoring how such policy changes impact on CLAHRC over time. As the evaluation progresses and the MCOs are tested and refined, we will pay attention to the impact that these wider political changes have in terms of acting as barriers or enablers to knowledge generation, implementation, and use.
In addition, the local response to the current governmental debate about NHS funding as one aspect of widespread public sector revisions, is as yet unknown. It is inevitable that in a time of financial austerity the CLAHRCs will face challenges about how they interpret and manage decisions about their joint remit for research and implementation. This, in turn, may impact on our evaluation, depending on the nature and extent of, for example, reductions, amendments, or cessation of the planned projects undertaken in the CLAHRCs. A pragmatic and flexible approach to undertaking research in 'real world' settings, and in particular in health care, is increasingly recognised as not only realistic, but necessary .
As described earlier, this is a longitudinal and interactive evaluation, which has some potential advantages. Realistic evaluation is iterative and engages stakeholders throughout the process. This will ensure we are able to adapt to ongoing changes to circumstances and facilitate the development of robust and sustained working relationships with the CLAHRCs. Engaging CLAHRC members in the development of the proposal and ongoing delivery of the research should ensure an appropriately focussed evaluation, contextually sensitive approaches to data collection, and opportunities for sharing and verifying emerging findings.
This evaluation was funded to provide information for learning, not for judgement. The purpose of the evaluation is formative, focusing on processes and a range of potential and actual impacts from implementation and use of knowledge as they occur over the lifespan of the evaluation and beyond the initial funding period of the CLAHRCs (2008 to 2013). The outputs of the study will be both theoretical and practical, and therefore opportunities for formative learning have been built in.
There are a number of ways the findings from this evaluation may contribute to knowledge about implementation. CLAHRCs provide a rare opportunity to study a natural experiment in real time, over time. The idea that collaboration, partnership, and sustained interactivity between the producers and users of knowledge lead to the production of more applicable research and increases the likelihood that research will be used in practice, has grown in popularity within the implementation science healthcare community. Whilst this is the theory, in practice we do not know whether this is the case, what the facilitators and barriers are to this way of working, or what the intended and unintended consequences may be. Our evaluation is designed to capture the processes and impacts of collaborative approaches for implementing research in practice, and therefore should contribute to the evidence base about an increasingly popular (e.g., mode two, integrated knowledge transfer, interactive research), but poorly understood approach to knowledge translation. Additionally, we have specific research questions about the role particular collaborative mechanisms, such as communities of practice and boundary objects play. Addressing these questions has the potential to increase our understanding of these mechanisms as potential implementation interventions, and inform future evaluation studies.
To date, much of the research exploring implementation processes and impacts has been conducted with a focus on isolated and one-off projects or initiatives, such as the implementation of a guideline or procedure. This means that we know little about implementation within sustained and organisational initiatives. As a longitudinal study that is focused at multiple levels within large regional entities, this evaluation could add to what we know about organisation level implementation initiatives over a sustained period of time.
Finally, we hope to contribute to methods for evaluating implementation processes and impacts. We have described why realistic evaluation is appropriate for this study; however, there are limited examples of its use in the published literature. This is an ideal opportunity to apply, and potentially develop, this approach, particularly with respect to integrated stakeholder involvement.
Case study research generates findings that are theoretically transferrable to other similar settings, but does not provide generalisable data, and therefore trying to generalise findings to other contexts either in the UK or in international settings should be undertaken with caution and acknowledgement of its provenance.
Each data collection method has its own limitations, but the benefit of using several data sources as triangulation of methods can largely overcome these by providing multiple perspectives on phenomena. To enhance the trustworthiness of data, the researchers will use a reflective approach to conducting the study, and this will be further explored and recorded as part of the project learning.
This article presents independent research commissioned by the National Institute for Health Research (NIHR) Service Delivery and Organisation Programme (SDO) (SDO 09/1809/1072). The views expressed in this publication are those of the authors and not necessarily those of the NHS, NIHR, or the Department of Health. The funder played no part in the study design, data collection, analysis and interpretation of data or in the submission or writing of the manuscript. The NIHR SDO Programme is funded by the Department of Health.
Heledd Owen for inserting and formatting references.
- Schuster ME, McGlynn E, Brook RH: How good is the quality of healthcare in the United States?. Milbank Quarterly. 1998, 76: 517-563. 10.1111/1468-0009.00105.View ArticlePubMedPubMed CentralGoogle Scholar
- Grol R: Success and failures in the implementation of evidence-based guidelines for clinical practice. Medical Care. 2001, 39 (8 Suppl 2): 1146-1154.Google Scholar
- McGlynn EA, Asch SM, Adams J, Keesey J, Hicks J, DeCristofaro A, Kerr EA: The quality of care delivered to adults in the United States. New England Journal of Medicine. 2003, 348 (26): 2635-2645. 10.1056/NEJMsa022615.View ArticlePubMedGoogle Scholar
- Clinical Effectiveness Research Agenda Group (CERAG): An Implementation Research agenda Report. 2008, (last accessed 13 February 2011), [http://preview.implementationscience.com/content/supplementary/1748-5908-4-18-s1.pdf]Google Scholar
- Eccles M, Armstrong D, Baker R, Clearly K, Davies H, Dvaies S, Glasziou P, Illott I, Kinmonth AL, Leng G, Logan S, Marteau T, Michie S, Rogers H, Rycroft-Malone J, Sibbald B: An implementation research agenda. Implementation Science. 2009, 4: 18-10.1186/1748-5908-4-18.View ArticlePubMedPubMed CentralGoogle Scholar
- Denis JL, Beaulieu MD, Hebert Y, Langley A, Lozeau D, Pineault R, Trottier LH: Clinical and Organizational Innovation in Healthcare Organizations. 2001, Ontario: Canadian Health Services Research Foundation/Fondation Canadienne de la recherché sur les services de santéGoogle Scholar
- Antil T, Desrochers M, Joubert P, Bouchard C: Implementation of an innovative grant programme to build partnerships between researchers, decision-makers and practitioners: the experience of the Quebec Social Research Council. Journal of Health Services Research & Policy. 2003, 8 (Suppl 2): 35-43.View ArticleGoogle Scholar
- Denis JL, Lomas J: Editorial: Convergent evolution: the academic and policy roots of collaborative research. Journal of Health Services Research & Policy. 2003, 8 (Suppl 2): S2:1-S2:6.Google Scholar
- Goering P, Butterill D, Jacobson N, Sturtevant D: Linkage and exchange at the organizational level: a model of collaboration between research and policy. Journal of Health Services Research & Policy. 2003, 8 (Suppl 2): S2:14-S2:19.View ArticleGoogle Scholar
- Bowen S, Martens P, The Need to Know Team: Demystifying knowledge translation: learning from the community. Journal of Health Services Research & Policy. 2005, 10 (4): 203-211. 10.1258/135581905774414213.View ArticleGoogle Scholar
- Gagliardi AR, Fraser N, Wright FC, Lemieux-Charles L, Davis D: Fostering knowledge exchange between researchers and decision-makers: Exploring the effectiveness of a mixed-methods approach. Health Policy. 2008, 86: 53-63. 10.1016/j.healthpol.2007.09.002.View ArticlePubMedGoogle Scholar
- Baker R, Robertson N, Rogers S, Davies M, Brunskill N, Khunti K, Steiner M, Williams M, Sinfield P: The Natinoal Institute of Health Research (NIHR) Collaboration for Leadership in Applied Health Research and Care (CLAHRC) for Leicestershire, Northamptonshire and Rutland (LNR): a programme protocol. Implementation Science. 2009, 4: 72-10.1186/1748-5908-4-72.View ArticlePubMedPubMed CentralGoogle Scholar
- NIHR: Collaborations for leadership in applied health research and care. Call for proposals to establish pilots. 2007, [http://www.sdo.nihr.ac.uk/files/researchcall/1072-brief.pdf]Google Scholar
- Dopson S, Fitzgerald L, Ferlie E, Gabbay J, Locock L: No magic targets! Changing clinical practice to become more evidence based. Health Care Management Review. 2002, 27 (3): 35-47.View ArticlePubMedGoogle Scholar
- Rycroft-Malone J, Kitson A, Harvey G, McCormack B, Seers K, Titchen A, Eastbrooks C: Ingredients for Change: Revisiting a conceptual model. Qual Saf Health Care. 2002, 11: 174-180. 10.1136/qhc.11.2.174.View ArticlePubMedPubMed CentralGoogle Scholar
- Lomas J, Culyer T, McCutcheon C, McAuley L, Law S: Conceptualizing and combining evidence for health system guidance. Canadian Health Services Research Foundation (CHSRF). 2005, [http://www.chsrf.ca]Google Scholar
- Graham ID, Logan J, Harrison MB, Straus SE, Tetroe J, Caswell W, Robinson N: Lost in knowledge translation: Time for a map?. The Journal of Continuing Education in the Health Professionals. 2006, 26: 13-24. 10.1002/chp.47.View ArticleGoogle Scholar
- Hirschkorn KA: Exclusive versus everyday forms of professional knowledge: legitimacy claims in conventional and alternative medicine. Sociol Health Illn. 2006, 28 (5): 533-57. 10.1111/j.1467-9566.2006.00506.x.View ArticlePubMedGoogle Scholar
- Davies HTO, Nutley S, Walter I: Why 'knowledge transfer' is misconceived for applied social research. Journal of Health Services Research & Policy. 2008, 13 (3): 188-190. 10.1258/jhsrp.2008.008055.View ArticleGoogle Scholar
- Treasury HM: A review of UK health research funding: Sir David Cooksey. 2006, London. HM Treasury, [http://18.104.22.168/d/pbr06_cooksey_final_report_636.pdf]Google Scholar
- Davies HTO, Nutley S, Smith PC: What Works? Evidence-based policy and practice in public services. 2000, Bristol: The Policy PressView ArticleGoogle Scholar
- Greenhalgh T, Robert G, McFarlane F, Bate P, Kyriakidou O: Diffusion of Innovations in Service Organisations: Systematic Review and Recommendations. The Millbank Quarterly. 2004, 82 (4): 581-629. 10.1111/j.0887-378X.2004.00325.x.View ArticleGoogle Scholar
- Dopson S, Fitzgerald L, (Eds): Knowledge to Action? Evidence-based health care in context. 2005, Oxford: Oxford University PressGoogle Scholar
- Harrison MB, Graham ID, Lorimer K, Griedberg E, Pierscianowski T, Brandys T: Leg-ulcer care in the community, before and after implementation of an evidence-based service. CMAJ. 2005, 172 (11): 1447-1452. 10.1503/cmaj.1041441.View ArticlePubMedPubMed CentralGoogle Scholar
- Davies H, Powell A, Rushmer R: Healthcare professionals' views on clinician engagement in quality improvement. A literature review. 2007, London, The Health FoundationGoogle Scholar
- Nutley SM, Walter I, Davies HTO: Using Evidence: How research can inform public services. 2007, Bristol: The Policy PressGoogle Scholar
- Estabrooks CA, Scott S, Squires JE, Stevens B, O'Brien-Pallas L, Watt-Watson J, Profetto-McGarth J, McGilton K, Golden-Biddle K, Lander J, Donner G, Boschma G, Humphrey CK, Williams J: Patterns of research utilization on patient care units. Implementation Science. 2008, 3: 31-10.1186/1748-5908-3-31.View ArticlePubMedPubMed CentralGoogle Scholar
- Van de Ven A: Central Problems in the Management of Innovation. Management Science. 1986, 32 (5): 590-607. 10.1287/mnsc.32.5.590.View ArticleGoogle Scholar
- Nutley S, Davies HTO: Making a Reality of Evidence-Based Practice: Some Lessons from the Diffusion of Innovations. Public Money and Management. 2000, 35-42.Google Scholar
- Iles V, Sutherland K: Organisational change: A review for health care managers, professionals and researchers. 2001, London: National Co-ordinating Centre for NHS Service Delivery and OrganisationGoogle Scholar
- Sheldon TA, Cullum N, Dawson D, Lankshear A, Lowson K, Watt I, West P, Wright D, Wright J: What's the evidence that NICE guidance has been implemented? Results from a national evaluation using time series analysis, audit of patients' notes, and interviews. British Medical Journal. 2004, 329 (7473): 999-10.1136/bmj.329.7473.999.View ArticlePubMedPubMed CentralGoogle Scholar
- Eccles M, Grimshaw J, Walker A, Johnston M, Pitts N: Changing the behaviour of healthcare professionals: the use of theory in promoting the uptake of research findings. Journal of Clinical Epidemiology. 2005, 58: 107-112. 10.1016/j.jclinepi.2004.09.002.View ArticlePubMedGoogle Scholar
- Mitton C, Adair CE, McKenzie E, Patten SB, Waye Perry B: Knowledge Transfer and Exchange: Review and Synthesis of the Literature. The Millbank Quarterly. 2007, 85 (4): 729-768. 10.1111/j.1468-0009.2007.00506.x.View ArticleGoogle Scholar
- McNulty T, Ferlie E: Reengineering health care: The complexities of organisational transformation. 2002, Oxford: Oxford University PressGoogle Scholar
- McCormack B, Kitson A, Harvey G, Rycroft-Malone J, Seers K, Titchen A: Getting Evidence into Practice: The meaning of 'context'. Journal of Advanced Nursing. 2002, 38 (1): 94-104. 10.1046/j.1365-2648.2002.02150.x.View ArticlePubMedGoogle Scholar
- Scott T, Mannion R, Davies H, Marshall M: Healthcare Performance & Organisational Culture. 2003, Radcliffe Medical Press: OxfordGoogle Scholar
- Kitson A, Rycroft-Malone J, Harvey G, McCormack B, Seers K, Titchen A: Evaluating the successful implementation of evidence into practice using the PARIHS framework: theoretical and practical challenges. Implementation Science. 2008, 3: 1-10.1186/1748-5908-3-1.View ArticlePubMedPubMed CentralGoogle Scholar
- Scott SD, Estabrooks CA, Allen M, Pollock C: A Context of Uncertainty: How context shapes nurses' research utilization behaviours. Qualitative Health Research. 2008, 18 (3): 347-357. 10.1177/1049732307313354.View ArticlePubMedGoogle Scholar
- Ferlie E, Fitzgerald L, Wood M, Hawkins C: The nonspread of innovations: The mediating role of professionals. Academy of Management Journal. 2005, 48 (1): 117-134.View ArticleGoogle Scholar
- Lavis JN, Robertson D, Woodside JM, McLeod CB, Abelson J, The Knowledge Transfer Study Group: How Can Research Organizations More Effectively Transfer Research Knowledge to Decision Makers?. The Millbank Quarterly. 2003, 81 (2): 221-248. 10.1111/1468-0009.t01-1-00052.View ArticleGoogle Scholar
- Wenger E: communities of practice: Learning, Meaning & Identity. 1998, Cambridge University Press: New YorkView ArticleGoogle Scholar
- Locock L, Dopson S, Chambers D, Gabbay J: Understanding the role of opinion leaders in improving clinical effectiveness. Social Science and Medicine. 2001, 53: 745-757. 10.1016/S0277-9536(00)00387-7.View ArticlePubMedGoogle Scholar
- Swan J, Scarbrough H, Robertson M: The Construction of 'communities of practice' in the Management of Innovation. Management Learning. 2002, 33 (4): 477-496. 10.1177/1350507602334005.View ArticleGoogle Scholar
- Gabbay J, le May A, Jefferson H, Webb D, Lovelock R, Powell J, Lathlean J: A case study of knowledge management in multi-agency consumer-informed 'communities of practice': implications for evidence-based policy development in health and social services. Health: An Interdisciplinary Journal for the Social Study of Health, Illness and Medicine. 2003, 7 (3): 283-310. 10.1177/1363459303007003003.Google Scholar
- le May A: communities of practice in Health and Social Care. 2009, West Sussex: Blackwell Publishing LtdGoogle Scholar
- Braithwaite J, Westbrook JI, Ranmuthugala G, Cunningham F, Plumb J, Wiley J, Ball D, Huckson S, Hughes C, Johnston B, Callen J, Creswick N, Georgiou A, Betbeder-Maibet L, Debono D: The development, design, testing, refinement, simulation and application of an evaluation framework for communities of practice and social-professional networks. BMC Health Services Research. 2009, 9: 162-10.1186/1472-6963-9-162.View ArticlePubMedPubMed CentralGoogle Scholar
- Li L, Grimshaw J, Nielson C, Judd M, Coyle PC, Graham ID: Use of communities of practice in business and health care sectors: A systematic review. Implementation Science. 2009, 4: 27-10.1186/1748-5908-4-27.View ArticlePubMedPubMed CentralGoogle Scholar
- Li L, Grimshaw J, Nielson C, Judd M, Coyle PC, Graham ID: Evolution of Wenger's concept of community of practice. Implementation Science. 2009, 4: 11-10.1186/1748-5908-4-11.View ArticlePubMedPubMed CentralGoogle Scholar
- Gabbay J, le May A: Evidence-based guidelines or collectively constructed 'mindlines?' Ethnographic study of knowledge management in primary care?. British Medical Journal. 2004, 329: 1013-10.1136/bmj.329.7473.1013.View ArticlePubMedPubMed CentralGoogle Scholar
- Star SL, Griesemer JR: Institutional Ecology. 'Translations' and Boundary Objects: Amateurs and Professionals in Berkeley's Museum of Vertebrate Zoology, 1907-39. Social Studies of Science. 1989, 19 (1989): 387-420.View ArticleGoogle Scholar
- Guston DH: Stabilizing the Boundary between US Politics and Science: The Rôle of the Office of Technology Transfer as a Boundary Organization. Social Studies of Science. 1999, 29 (1): 87-111. 10.1177/030631299029001004.View ArticlePubMedGoogle Scholar
- Carlile PR: A Pragmatic View of Knowledge and Boundaries: Boundary objects in new product development. Organization Science. 2002, 13 (4): 442-455. 10.1287/orsc.13.4.442.2953.View ArticleGoogle Scholar
- Swan J, Bresnen M, Newell S, Robertson M: The object of knowledge: The role of objects in biomedical intervention. Human Relations. 2007, 60 (12): 1809-1837. 10.1177/0018726707084915.View ArticleGoogle Scholar
- McGivern G, Dopson S: Inter-epistemic power and transferring knowledge objects in a biomechanical network. Organization Studies. 2010, 31 (12): 1667-1286. 10.1177/0170840610380808.View ArticleGoogle Scholar
- ICEBeRG Group: Designing theoretically-informed implementation interventions. Implementation Science. 2006, 1: 4-[http://www.implementationscience.com/content/1/1/4]View ArticleGoogle Scholar
- Rycroft-Malone J: Theory and Knowledge Translation: Setting some co-ordinates. Nursing Research. 2007, 56 (4S): S78-S85.View ArticlePubMedGoogle Scholar
- Rycroft-Malone J, Harvey G, Seers K, Kitson A, McCormack B, Titchen A: An exploration of the factors that influence the implementation of evidence into practice. Journal of Clinical Nursing. 2004, 13: 913-924. 10.1111/j.1365-2702.2004.01007.x.View ArticlePubMedGoogle Scholar
- Rycroft-Malone J, Seers K, Titchen A, Kitson A, Harvey G, McCormack B: What counts as evidence in evidence based practice?. Journal of Advanced Nursing. 2004, 47 (1): 81-90. 10.1111/j.1365-2648.2004.03068.x.View ArticlePubMedGoogle Scholar
- Pawson R, Tilley N: Realistic Evaluation. 1997, London: Sage PublicationsGoogle Scholar
- Pawson R, Greenhalgh T, Harvey G, Walshe K: Realist Synthesis: an introduction. ESRC Research Methods Programme: University of Manchester RMP: Methods Paper 2/2004, [http://www.ccsr.ac.uk/methods/publications/documents/RMPmethods2.pdf]
- Sridharan S, Campbell B, Zinzow H: Developing a Stakeholder-Driver Anticipated Timeline of Impact for Evaluation of Social Programs. American Journal of Evaluation. 2006, 27 (2): 148-162. 10.1177/1098214006287990.View ArticleGoogle Scholar
- Sullivan H, Barnes M, Matka E: Building collaborative capacity through 'theories of change'. Early lessons from the evaluation of health action zones in England. Evaluation. 2002, 8 (2): 205-226. 10.1177/1358902002008002514.View ArticleGoogle Scholar
- Greenhalgh T, Humphrey C, Hughes J, Macfarlane F, Butler C, Pawson R: How do you modernize a Health Service? A realist evaluation of whole-scale transformation in London. Millbank Quarterly. 2009, 87 (2): 391-417. 10.1111/j.1468-0009.2009.00562.x.View ArticleGoogle Scholar
- Rycroft-Malone J, Fontenla M, Bick D, Seers K: Protocol-Based Care Evaluation Project. Final Report; 2008a. NIHR Service Delivery & Organisation Programme SDO/78/2004, [http://www.sdo.nihr.ac.uk/files/project/78-final-report.pdf]
- Rycroft-Malone J, Fontenla M, Bick D, Seers K: Protocol-based care: Impact on roles and service delivery. Journal of Evaluation in Clinical Practice. 2008, 14: 867-873. 10.1111/j.1365-2753.2008.01015.x.View ArticlePubMedGoogle Scholar
- Rycroft-Malone J, Fontenla M, Bick D, Seers K: A Realistic Evaluation: the case of protocol-based care. Implementation Science. 2010, 5 (38):Google Scholar
- Byng R, Norman I, Redfern S: Using realistic evaluation to evaluate a practice-level intervention to improve primary healthcare for patients with long-term mental illness. Evaluation. 2005, 11 (1): 69-93. 10.1177/1356389005053198.View ArticleGoogle Scholar
- Marchal B, Dedzo M, Kegels G: A realist evaluation of the management of well-performing regional hospital in Ghana. Health Service Research. 2010, 10: 24-10.1186/1472-6963-10-24.View ArticleGoogle Scholar
- Pittam G, Boyce M, Sesker J, Lockett H, Samele C: Employment advice in primary care: a realistic evaluation. Health and Social Care in the Community. 2010, 18 (6): 598-606. 10.1111/j.1365-2524.2010.00929.x.View ArticlePubMedGoogle Scholar
- Weiss CH: The many meanings of research utilization. Public Administration Review. 1979, 39 (5): 426-431. 10.2307/3109916.View ArticleGoogle Scholar
- Wilkinson JE: Research impact - hard hitting or subtle change?. Worldviews on Evidence-based Nursing. 2010, 7 (1): 1-3.View ArticlePubMedGoogle Scholar
- Wilkinson JE, Johnson N, Wimpenny P: Models and approaches to inform the impacts of implementation of evidence-based practice. Evaluating the Impact of Implementation of Evidence-Based Practice. Edited by: Bick D, Graham I. 2010, Oxford: Wiley BlackwellGoogle Scholar
- Gibbons M, Limoges C, Nowotny H, Schwartzman S: The new production of knowledge - The dynamics of science and research in contemporary societies. 1994, London. SageGoogle Scholar
- Denis JL, Lehoux P, Hivon M, Champagne F: Creating a new articulation between research and practice through policy? The views and experiences of researchers and practitioners. Journal of Health Services Research & Policy. 2003, 8 (Suppl 2): S2:44-S2:50.Google Scholar
- Martens PJ, Roos NP: When Health Services Researchers and Policy Makers Interact: Tales from the Tectonic Plates. Healthcare Policy. 2005, 1 (1): 72-84.PubMedPubMed CentralGoogle Scholar
- Kothari A, Birch S, Charles C: 'Interaction' and research utilisation in health policies and programs: does it work?. Health Policy. 2005, 71: 117-125. 10.1016/j.healthpol.2004.03.010.View ArticlePubMedGoogle Scholar
- Cargo M, Mercer SL: The Value and Challenges of Participatory Research: Strengthening Its Practice. Annu Rev Public Health. 2007, 29 (24): 1-24.Google Scholar
- Guba EG, Lincoln YS: 4th Generation Evaluation. 1989, Newbury Park: SageGoogle Scholar
- Kemmis S: Participatory Action Research and the Public Sphere. EDUCATIONAL ACTION RESEARCH. 2006, 14 (4): 459-476. 10.1080/09650790600975593.View ArticleGoogle Scholar
- Seale C: The quality of qualitative research. 1999, Sage: LondonView ArticleGoogle Scholar
- Yin RK: Case study research - design and methods. 2003, Thousand Oaks: Sage, 3Google Scholar
- Patton M: Utilization-Focused Evaluation. 2008, Thousand Oaks, CA: Sage Publications, 4Google Scholar
- Andrews GJ, Evans J: Understanding the reproduction of health care: towards geographies in healthcare work. Progress in Human Geography. 2008, 32 (6): 759-780. 10.1177/0309132508089826.View ArticleGoogle Scholar
- Spradley JP: Participant Observation. 1980, Orlando: Harcourt Brace Jovanovich College PublishersGoogle Scholar
- Nowotny H: Democratising expertise and socially robust knowledge. Science and Public Policy. 2003, 30 (3): 151-156. 10.3152/147154303781780461.View ArticleGoogle Scholar
- Bartunek J, Trullen J, Bonet E, Sauquet A: Sharing and expanding academic and practitioner knowledge in health care. Journal of Health Services Research & Policy. 2003, 8 (Suppl 2): S2:62-S2:68.View ArticleGoogle Scholar
- Huberman AM, Miles MB: Data management and analysis methods. Collecting and interpreting qualitative materials. Edited by: Denzin NK, Lincoln YS. 1998, Sage: Thousand Oaks, CA, 179-210.Google Scholar
- Kass N, Pronovost PJ, Sugarman J, Goeschel C, Lubomski L, Faden R: Controversy and quality improvement: lingering questions about ethics, oversight and patient safety research. The Joint Commission Journal on Quality and Patient Safety. 2008, 34 (6): 349-353.PubMedGoogle Scholar
- Flaming D, Barrett-Smith L, Brown N, Corocan J: Ethics? But it's only quality improvement!. Healthcare Quarterly. 2009, 12 (2): 50-54.View ArticlePubMedGoogle Scholar
- Department of Health (DH): Research Governance Framework for Health and Social Care. 2005, [http://www.dh.gov.uk]2Google Scholar
- Symons T: Good clinical practice and the regulatory requirements for clinical trials: a refresher session. 2010, NISCHR CRC and Symons Associates Clinical Research ConsultancyGoogle Scholar
- UK Research Integrity Office (UKRIO). [http://www.ukrio.org.uk]
- Canadian Institutes of Health Research (CIHR): Evidence in Action, Acting on Evidence. 2006, CIHR Institute of Health Services and Policy Research, [http://cihr-irsc.gc.ca/e/documents/ihspr_ktcasebook_e.pdf]Google Scholar
- Department of Health (DH): Equity and Excellence: Liberating the NHS. 2010, White paper: London DH, [http://www.dh.gov.uk/en/publicationsandstatistics/Publications/PublicationsPolicyAndGuidance/DH_118602]Google Scholar
- Rycroft-Malone J, Dopson S, Degner L, Hutchinson AM, Morgan D, Stewart N, Estabrooks C: Study protocol for the translating research in elder care (TREC): building context through case studies in long-term care project (project two). Implementation Science. 2009, 4: 53-10.1186/1748-5908-4-53.View ArticlePubMedPubMed CentralGoogle Scholar
- Department of Health (DH): Report of the High Level Group on Clinical Effectiveness chaired by Professor Sir John Tooke. 2007, London: DH, [http://www.dh.gov.uk/en/PublicationsandStatistics/Publications/PublicationsPolicyAndGuidance/DH_079799]Google Scholar
This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (http://creativecommons.org/licenses/by/2.0), which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.