Skip to main content

Data for Improvement and Clinical Excellence: a report of an interrupted time series trial of feedback in home care



There is substantial evidence about the effectiveness of audit with feedback, but none that we know have been conducted in home care settings. The primary purpose of the Data for Improvement and Clinical Excellence – Home Care (DICE-HC) project was to evaluate the effects of an audit and feedback delivered to care providers on home care client outcomes. The objective of this paper is to report the effects of feedback on four specific quality indicators: pain, falls, delirium, and hospital visits.


A 10-month audit with feedback intervention study was conducted with care providers in seven home care offices in Alberta, Canada, which involved delivery of four quarterly feedback reports consisting of data derived from the Resident Assessment Instrument – Home Care (RAI-HC). The primary evaluation employed an interrupted time series design using segmented regression analysis to assess the effects of feedback reporting on the four quality indicators: pain, falls, delirium, and hospitalization. Changes in level and trend of the quality indicators were measured before, during, and after the implementation of feedback reports. Pressure ulcer reporting was analyzed as a comparator condition not included in the feedback report. Care providers were surveyed on responses to feedback reporting which informed a process evaluation.


At initiation of feedback report implementation, the percentage of clients reporting pain and falls significantly increased. Though the percentage of clients reporting pain and falls tended to increase and reporting of delirium and hospital visits tended to decrease relative to the pre-intervention period, there was no significant effect of feedback reporting on quality indicators during the 10-month intervention. The percentage of clients reporting falls, delirium, and hospital visits significantly increased in the 6-month period following feedback reporting relative to the intervention period. About 50% of the care providers that read and understand the feedback reports found the reports useful to make changes to the way clients are cared for.


Routinely collected data used over time for feedback is feasible in home care settings. A high proportion of care providers find feedback reports useful for informing how they care for clients. Since reporting on the frequency of quality indicators increased in the post-intervention period, this study suggests that ongoing use of audit with feedback to enhance health outcomes in home care may promote improved reporting on standardized instruments.

Peer Review reports


Home care consists of an array of services designed to meet individual client needs [1, 2]. Based on census data from 2003, it was estimated that 550,000 unique clients aged 65 and older received home care in Canada [3], and by 2006, this number grew to nearly one million home care clients at any given time [2]. Home care is increasing in both numbers and in level of client acuity [4, 5] across Canada and in most other developed countries. It is typically part of the broader continuing care sector, which also includes long-term care (facility living) and assisted or supportive living within the Canadian healthcare system. Home care is not included in the Canada Health Act as a necessary medical service and as such receives variable public funding across Canadian jurisdictions through provincial governments [6]. The objective of home care is to provide cost-effective ongoing, home-based care for clients of all age groups following hospital discharge, for palliation, aging in place, or for those that are cognitively and physically impaired [5]. Home care is provided based on assessed need. The assessment is done by case managers, most of whom are registered nurses, but may also be an allied health professional. Case managers have complex caseloads that vary depending on the type of client, for example, whether they are short-term post hospital care, pediatric, or long-term maintenance clients [7]. Caseloads are typically culturally diverse and include clients with a myriad of diagnoses. Following assessment, case managers care plan with the client and family, coordinate care and services required, monitor the client response to care, and ensure client safety among other duties [1, 6, 8].

Case managers in Canada are the gate keeper to home care. Care is delivered most frequently by healthcare aides, followed by licensed practical nurses, registered nurses, and allied health professionals. Direct care is either provided directly through a government home care program or more typically in many Canadian jurisdictions through a contracted services provider agency that works closely with the case manager from intake to discharge. In contractual arrangements, the case manager determines and allocates care while the service provider agency hires, supervises, and manages all aspects of the staffing component. The case manager determines the plan of care, and the service provider agency carries out the plan of care.

This project arm is part of a larger group of DICE studies [9,10,11,12] and focused on the home care setting specifically, as opposed to long-term care (LTC) or other healthcare settings. The LTC project was reported elsewhere [11].

This arm of the study focused on the feedback as a quality improvement practice in the home care setting [9]. In home care, quality improvement initiatives are increasing in number but still lag somewhat behind other continuing care sectors [13]. The availability of standardized tools such as the RAI-HC has made data more readily available. However, systems to use the data are often not imbedded in organizations which makes it challenging to compare outcomes so trials of quality improvement interventions are rarely conducted compared with other healthcare settings [13].

The evidence for specific interventions to support the implementation of evidence-based practices in healthcare settings is mixed at best. In addition to the findings of a recent systematic review by our team [13], a study by Markle-Reid et al. on nurse-led health promotion in southern Ontario illustrated that the use of standardized screening tools can improve the effectiveness of interventions in home care [14]. Another intervention that shows promise is the use of audit with feedback intervention (referred to as feedback intervention) [15, 16].

Feedback interventions have demonstrated a modest effect to promote desired behavioral changes among healthcare providers, across settings and provider types [15, 16]; and although proposed in 1989 as having positive implications for future nursing care [17], very few [18] feedback intervention studies have been conducted in home care settings. The lack of data on care processes and outcomes have posed a major barrier to using feedback interventions in home care settings. However, this has been somewhat improved with the implementation of the RAI-HC, a standardized assessment tool used in many jurisdictions around the world.

The probable mechanism by which feedback interventions has a main effect is in providing healthcare providers with information about their own performance [19–21]. It is thought that knowledge of one’s performance, particularly among those who have not received data-based feedback on their past performance, could contribute to an increase in motivation to change certain behaviors. Feedback reports have often been used in conjunction with other interventions to enhance a desired effect [15]. The purpose of this study was to adopt feedback reporting as a sole intervention in home care since providers often work alone in the community. As a result, there is less opportunity to interact with colleagues and other providers where they may receive feedback on performance. Feedback reporting constitutes a method by which home care providers can receive information on client outcomes directly related to provider care. The strengths of this feedback intervention lie in the widespread availability of data on outcomes using RAI-HC, the relative simplicity with which it is constructed, and that it does not require additional cost for audit data to construct feedback reports [22–24].

This paper reports the summative client outcomes of a feedback intervention that was delivered to all staff in seven home care settings and contains a process evaluation on uptake of feedback reporting. The primary hypothesis of this study was that a consistent, long duration, client-focused feedback intervention would improve client outcomes among four important quality indicators: pain, falls, delirium, and hospital visits.


An interrupted time series design was employed to assess the overall effect of feedback reports in the home care setting. Based on the underlying conceptual model built on the theory of planned behavior [25], it was hypothesized that feedback reporting would influence intention to change behavior through multiple paths including attitudes and social norms [11]. We are unable to measure to provider behavior directly but report on elements of process evaluation. The full protocol for the Data for Improvement and Clinical Excellence project is available [9]. Methodology pertaining to the specific elements of DICE-HC is reported here.

Settings and sample

Site selection was guided by the degree of RAI-HC implementation. Ultimately, seven home care offices in both rural and urban areas within Alberta participated in the study. All home care offices were part of Alberta Health Services which is the single health region in Alberta responsible for all health service delivery including home care. At the time of the study, there were five zones. These were South Zone covering areas south of the city of Calgary, Calgary Zone, Central Zone, Edmonton Zone, and North Zone covering all areas north of Edmonton. Participants included direct care staff: case managers, nurses, healthcare aides, and allied health professionals, as well as managers and home care professional practice leaders.

The intervention: quality indicator feedback reports

The feedback reports were initially developed during a pilot study conducted in Edmonton area in 2007. The feedback reports were refined for home care using client quality indicator outcomes selected based on the RAI-HC. The outcomes of the RAI-HC indicate potential quality issues that may need further review and are used by care providers to monitor trends and improve care [26]. The quality indicators for this study were selected based on clinical importance. The quality indicators that were determined to be of most value to this project were determined using a consensus building process with a group of six decision-maker partners based on the following criteria:

  1. a)

    What is known from the literature about what is important

  2. b)

    The strategic initiatives in progress or likely to be initiated in the period of the project

  3. c)

    The comparability of the indicators (i.e., were the indicators comparable/similar to those used in long-term care)

  4. d)

    What was useful to clinicians

  5. e)

    Where changes are likely to be observed

The decision-maker group consisted of home care clinical and administrative leaders who were part of the research team. They had a variety of clinical or administrative backgrounds including nursing, rehabilitation, and quality improvement. The consensus results showed that pain, falls, delirium, and hospital visits were the most important quality indicators to be included in the feedback report.

Pain was selected as the primary quality indicator for consistency throughout the continuing care sectors assessed. Pain was also the primary indicator assessed in the DICE-LTC project [10, 11] as provider awareness of client pain may prompt pain alleviation measures like a referral or prescription of medication. Pain was quantified by taking into account both the frequency and intensity with which the client complained or showed evidence of pain. A client was reported to be in pain if he or she had either (a) less than daily pain but it was horrible or excruciating; (b) one to more periods daily and it was moderate to severe; or (c) one to more periods daily and it was excruciating.

Falls were reported in terms of frequency, i.e., the number of times a client has fallen in the last 90 days or since the last assessment if less than 90 days. In the present study, a client was recorded to have a fall if he or she had at least one fall during the current 90-day period.

A client reported delirium if either (a) there was sudden or new onset/change in mental function of the client over the last 7 days and/or (b) the client experienced agitation or disorientation in the last 90 days, or since the last assessment if less than 90 days, such that his or her safety is endangered or the client required protection by another person.

Clients reported a hospital visit if one occurred in the last 90 days or if since last assessment, they used at least one of the three services described: (a) admission to hospital with an overnight stay; (b) visit to emergency room without an overnight stay; (c) emergent care—including unscheduled nursing, physician, or therapeutic visits to the office or home.

Finally, the presence of pressure ulcer was assessed as a related quality indicator not included in the feedback report. Pressure ulcer was selected as a reference measure because it is sensitive to change that may occur independently of the response to feedback reporting. Clients were reported to have a pressure ulcer if they had at least stage 2 of this skin condition.

Feedback report generation

Feedback reports specific to each home care office were then generated using the selected quality indicators based on data from the RAI-HC and included the percentages of clients with pain, falls, delirium, and hospital visits (Additional file 1). Client assessments using the RAI-HC are routinely conducted once annually or when there is a change in health status and were implemented during the study period under several circumstances: initial assessment, follow-up assessment, routine assessment at fixed intervals, review within the 30-day period prior to discharge from the home care program, review at return from hospital, and change in status. If a client had more than one assessment in a quarter, only the most recent assessment was recorded.

Feedback report intervention

The first feedback report was based on the most recent data for one full year divided into four annual quarters, and in the succeeding reports, subsequent quarterly data were added with the previous first/earliest quarter being omitted. Feedback reports included data on 1 year of historical performance from the home care office staff to whom it was distributed and those data from other participating home care offices. Reports were distributed in June 2011, September 2011, December 2011, and March 2012. The intervention involved giving the feedback reports to the individual care provider either in person or via electronic distribution.

The in-person intervention was implemented in Edmonton, North, and Central Zone home care offices and electronically for South and Calgary Zones [9]. If it was not possible for staff to be in attendance in person, for example, if staff were out on home visits, on a different shift, or on a day off, reports were left in the office at an arranged location for pickup and review.

Electronic distribution was done using the Continuing Care Desktop that was developed by the Centre for Health Evidence in collaboration with Alberta Health and Wellness. All employees in continuing care in Alberta have access to the Continuing Care Desktop, and employee access is managed by each home care office. Site leaders were notified by email when the reports became available on the Continuing Care Desktop.

Response to feedback reporting: Process evaluation

In addition to the feedback intervention, data on intention to change behavior based on feedback were also obtained. We provide brief description of the methods for the process evaluation here as the detailed protocol was published previously [9]. A sample survey is provided in the protocol paper [11]. One week after feedback reports were distributed, employees were asked to complete a survey electronically on the Continuing Care Desktop. The survey includes questions on demographics and an assessment of employee response to the feedback report. Questions on intention to change behavior regarding clients that are having pain were answered on a 7-point rating scale where 1 indicated lack of importance, disagreement, or unlikeliness and 7 indicated high importance, agreement, or likeliness. The survey section aimed to assess intention to change behavior was only completed by direct care providers.


The study was designed as an interrupted time series where the RAI-HC data were available for 22 months (partitioned into 6 pre-intervention months, 10 intervention months, and 6 post-intervention months). A segmented regression analysis was employed to statistically evaluate the magnitude of the effect of the feedback reports on the quality indicators of interest. RAI-HC data available from each of the 22 months were grouped so that the pre-intervention period had 6 monthly time points (December 1, 2010, to May 31, 2011), the intervention period had 10 monthly time points (June 1, 2011, to March 31, 2012), and the post-intervention period had 6 monthly time points (April 1, 2012, to September 30, 2012) for interrupted time series. Each quality indicator was assessed before, during, and after the feedback report intervention for changes in level and slope in specific time series. A predicted regression line was fitted to each segment for quality indicators. The relationship between time and the quality indicator within each segment in the model was assumed to be linear. The specified linear regression model is as follows:

$$ {Y}_t={b}_0+{b}_1*{\mathrm{time}}_t+{b}_2*{\left(\mathrm{feedback}\right)}_t + {b}_3*{\left(\mathrm{time}\ \mathrm{after}\ \mathrm{feedback}\right)}_t+{b}_4*{\left(\mathrm{without}\ \mathrm{feedback}\right)}_t + {b}_5*{\left(\mathrm{time}\ \mathrm{after}\ \mathrm{without}\ \mathrm{feedback}\right)}_t+{e}_t $$


  • Y t is the percentage of clients with the quality indicator in month t

  • time is the time in months at time t from the start of the observation period, it ranges from 1 to 22 months

  • feedback is an indicator for time t occurring before (feedback = 0) or after (feedback = 1) the feedback report, which was implemented in June 2011 in the series

  • time after feedback is the number of months after the intervention at time t, coded 0 before the feedback report and (time − 6) after the feedback report;

  • without feedback is an indicator for time t occurring before (without feedback = 0) or after (without feedback = 1) without the feedback report, which was after March 2012 in the time series

  • time after without feedback is the number of months after the intervention at time t, coded 0 before the end of the feedback report and (time − 16) after the end of the feedback report

  • e t is the error term at time t that represents the unexplained random variation in the model

In this model, regression parameters are estimated as follows:

  • b 0, percentage of clients with the health outcome at baseline

  • b 1, trend prior to the feedback report (baseline trend)

  • b 2, change in level immediately after the first feedback report

  • b 3, change in trend after the feedback report distribution

  • b 4, change in level immediately after the last feedback report

  • b 5, change in trend after the end of the feedback report distribution

Serial autocorrelation of the error terms was tested in the regression model using the Durbin-Watson (DW) statistic. Theoretically, possible values of the DW statistic range from 0 to 4. Observations close to midpoint of 2.00 indicate no serious autocorrelation. If serial correlation was detected in the data, correction would be made using the Prais-Winsten estimator.

Data on client demographics was collected at four points during the study: December 2010, June 2011, March 2012, and September 2012. Statistical analyses were conducted using STATA version 12 (StataCorp, College Station, TX), and statistical significance was defined at a level of α = 0.05.


Resident characteristics

Over the entire study period, data from 548 participants are reported. The average age of clients was similar at study baseline (December 2010), study completion (September 2012), feedback report initiation (June 2011), and conclusion (March 2012). Home care clients were predominantly female (Table 1).

Table 1 Client demographics at study initiation, initiation and conclusion of feedback report period, and study conclusion


Each quality indicator in the model was estimated using simple linear regression with an additional test for first-order autocorrelation. Serial autocorrelation was only present in pain (DW statistic = 1.45), and so Prais-Winsten estimator was used to correct for this quality indicator.

Segmented regression analysis

Figure 1a–e depicts the five time series quality indicator outcomes. Findings of the segmented regression analysis are summarized in Table 2. Analyses showed a mixed pattern of results across the four quality indicator outcomes of interest. The implementation of feedback reporting was not associated with the baseline trend uniformly across quality indicators. As well, the change in the trend across quality indicators after feedback report implementation were not uniformly associated with the implementation of feedback reporting indicating that audit with feedback in home care may only impact select quality indicators.

Fig. 1
figure 1

ae Time series quality indicator outcomes

Table 2 Parameter estimates for percentage of clients with pain, falls, delirium, hospital visits, and pressure ulcer


The number of clients reporting pain ranged from 30 to 50% between December, 2010, and September, 2012. The coefficient estimate for the change in level of the proportion of clients reporting high pain scores at initiation of the feedback report (b 2) indicates a significant increase in reporting of pain. The negative value of the regression coefficient estimate for the feedback report phase (b 3) of the study indicates a decrease in the proportion of clients with high pain scores relative to baseline trend, but the measure is not significant. There were no other statistically significant coefficient estimates for pain (Fig. 1a).


The number of clients reporting a fall ranged from 20 to 40% between December, 2010, and September, 2012. As in the case of pain, there was a significant increase in the coefficient estimate for the change in level of the proportion of clients who had a fall at feedback report initiation (b 2). The negative value of the regression coefficient estimate for the feedback report period of the study (b 3) indicates a decrease (not significant) in the proportion of clients reporting a fall relative to baseline trend. In the post-intervention period, the regression coefficient estimate (b 5) indicates a statistically significant increase for falls relative to the feedback report period. There were no other statistically significant coefficient estimates for pain (Fig. 1b).


The number of clients reporting delirium ranged from 4 to 14% over the observation period. There were no significant changes in the percentage of clients reporting delirium in the pre-intervention or feedback reporting periods, and there were no changes in the level of feedback reporting at the initiation or conclusion of feedback reporting. However, the positive value of the regression coefficient estimate for the post-feedback report period of the study (b 5) indicates a significant increase in the proportion of clients reporting delirium relative to the 10-month intervention period (Fig. 1c).

Hospital visits

The number of clients reporting hospital visits ranged from 35 to 60% over the observation period. Though the negative slope of the regression line for the feedback reporting period indicates a decrease in hospital visits, the value of the regression coefficient estimate for this period (b 3) suggests no change in reporting of hospital visits relative to the baseline trend. However, there was a significant increase in the proportion of clients reporting hospital visits in the post-feedback report period (b 5) relative to the 10-month feedback report period. There were no other statistically significant coefficient estimates for hospital visits (Fig. 1d).

Pressure ulcer

The percentage of clients reporting pressure ulcer was low throughout the 22-month observation period and did not exceed 6%. There was no statistically significant change in the trend of the percentage of clients reporting a pressure ulcer before, during, or after the feedback reporting phases or in the level of pressure ulcer reported at implementation or conclusion of feedback reporting (Fig. 1e).

Process evaluation

There were a total of 300 responses to feedback report surveys in the study (Table 3). Response rate to feedback report survey ranged from 46% for the first survey (distributed after the first feedback report) to 56% for the third survey (distributed after the third feedback report). Several types of care providers completed the surveys: healthcare aides constituted the majority of respondents to the first survey (25%), case managers constituted the majority of respondents to the second (39%) and third surveys (38%), and registered nurses were the most abundant respondents to the fourth survey (33%). Overall the four survey periods, 87% of respondents indicated they read more than half of the report, 86% indicated understanding more than half of the report, 43% found the report useful, 38% discussed the report with another staff member, and 35% found the report useful to make changes in the way they take care of clients. One third of the respondents indicated that the feedback reports were discussed at staff meetings. Based on the feedback reports, the respondents who indicated that they desired changes made in the way clients are taken care of cited primarily that the information is useful to other care providers (not necessarily themselves) and changes should be made in the way clients are assessed and assisted in daily living, the policies that affect clients, and the daily schedule of clients. The mean respondent scores (±standard deviation) for intention to assess of monitor clients’ level of pain during each shift for surveys 1, 2, 3, and 4 were 4.58 (±1.72), 4.95 (±1.83), 5.47 (±1.44), and 5.17 (±1.72), respectively on the 7-point rating scale.

Table 3 Process evaluation outcomes of an audit with feedback in home care


This study reports on an audit with feedback intervention delivered over 10 months across seven home care offices. It was expected that the percentage of clients who experienced high pain, falls, delirium, or hospital visits would decrease during the intervention phase. It was also expected that the decline in quality indicators would be sustained after the end of the feedback report distribution. Feedback reporting showed that there were decreases in pain, fall, delirium, and hospital visits over the intervention period but that there was no difference in the trend for reporting quality indicators in the intervention period relative to the pre-intervention period.

Contrary to expectation, the decreases in pain, falls, delirium, and hospital visits observed during the intervention period were not sustained beyond the 6-month post-intervention period. The change in trend for pain, falls, delirium, and hospital visits all increased in the post-feedback report period relative to the intervention period. These data suggest that audit with feedback intervention may need to be an ongoing process to sustain potential improvement in pain, falls, delirium, and hospital visits or that the intervention may require inclusion of additional elements to maintain or continue to improve intervention-related increases in the quality indicators of interest.

According to this study, the initiation of feedback reporting was associated with an increased level in the proportion of clients reporting pain, falls, and possibly hospital visits, but not with delirium. On the other hand, there were no changes in the levels of pain, falls, delirium, or hospital visits reported immediately following intervention (after distribution of final feedback report).

Although there was large month-to-month variability in fall and hospital visits reported in the pre-feedback period, there was no significant linear trend for any of the four quality indicators of interest over the 6 months prior to intervention. Finally, pressure ulcer served as an appropriate reference quality indicator as there was no change in reporting on level or trend of this outcome among any of the study periods.

Facilitators and barriers to audit with feedback in home and long-term care settings

Some of the differences in outcomes observed between the present study in home care and the previously published study in long-term care may be a result of the frequency of report distribution, the source of the feedback, and the perceived sign (whether positive, neutral, or negative) of the information in the report. Feedback reporting occurred monthly in the DICE-LTC study and quarterly in the DICE-HC study. Due to the small magnitude of effects observed, the results of both studies do not provide evidence as to the optimal frequency of feedback reporting in continuing care. Furthermore, feedback report delivery was done in person in the DICE-LTC study, and some feedback reports were delivered electronically in the DICE-HC study. Electronic distribution can reach a great proportion of staff in a timely manner, though electronic distribution may also be ignored without verbal notification or filtered from view in email. In-person distribution relies heavily on a “site champion” to engage and make staff aware of feedback reports. This is especially important as staff in home care working remotely may not come to respective home care offices frequently to pick up feedback reports in mail.

It was proposed that the post-feedback report survey may have unintentionally served as a co-intervention to feedback reports unintentionally causing the decrease in reports of high pain scores in long-term care. Since the audit with feedback study in home care also involved post-feedback report surveys and actually demonstrated an increase in reports of high pain scores at the beginning of the intervention period, it is unlikely that the post-feedback report surveys acted as a co-intervention tool in LTC or in the home care study reported on here.

Process evaluation

More than 90% of the staff read more than half of the first two feedback reports. Only about 80% of the staff read more than half of the final feedback report. The decrease in feedback report reading over time may indicate that the staff felt that report distribution intervals too frequent to notice differences or changes in the trend of quality indicators. Alternatively, positive or negative trends of quality indicators in feedback report may have influenced the desire of the staff to continue reading feedback reports over time. The staff in home care discussed reports with others less often than the staff in LTC. This is because home care staff may not have the same opportunity to discuss feedback reports with peers and colleagues that staff in LTC had due to the nature of the work environment. Providers in home care are usually based remotely in the community where interaction with colleagues cannot occur very often.

More than 50% of the respondents in the DICE-LTC study were healthcare aides compared to less than 25% in DICE-HC. The range of abilities and role boundaries of healthcare aides differ from that of other professions. There is variable education and training for different care provider professions (e.g., licensed practical nurse, healthcare aide, social worker). As such, a recent study of audit with feedback in home care on pain and falls in Ontario, Canada, suggests that feedback reporting may benefit from being catered to specific provider groups [27]. The findings from the DICE-HC study support this notion as almost half of the respondents indicated that they found reports useful overall, but only about one third of the respondents found the report useful to make changes to the way they cared for clients. This indicates that relevant information regarding client care may be present in the feedback report, but the care provider reading the report does not necessarily believe that they can positively influence the specific client outcome. This may be because the care provider reading the report believes that the activities performed within the role, competencies, and scope of practice of their profession do not have great influence on a specific quality indicator. Improvement in quality indicators are often achieved as a result of collaborative care planning in teams, not solely through individual efforts. It would be of interest to explore aspects of interprofessional collaboration and team climate in home care as it relates client outcomes in future study in order to discern which care professionals and how specific care professionals believe they can impact quality indicators.


Best practice recommendations made from the present study may be limited by the quasi-experimental design. The degree on RAI-HC implementation across provincial home care offices did not allow for inclusion of a contemporaneous control wherein the measurement of quality indicators could be obtained from home care offices not undergoing audit with feedback intervention. Interrupted time series study designs may be restricted by the number of times at which the dependent variable is assessed. In order to satisfy the requirements for the use of an interrupted time series design, quarter-annual report data were instead analyzed monthly so that three times as many time segments were available for analyses. It is possible that in addition to changes in care planning behavior, the intervention prompted changes to the attention given to specific parameters in administering the RAI-HC. It was a challenge to deliver feedback reports electronically as email notifications can often be ignored and the process requires buy-in from a site champion; as such, the collective study data may be more representative of the Edmonton, North, and Central Zones where feedback reports were distributed in person. Irrespective of perceived study limitations, this is one of the longest and most intensive studies of an audit with feedback intervention conducted in the home care setting.


Routinely collected data when used over time for feedback is a feasible approach for a quality improvement initiative. There is a need to devise strategies for implementation, effective delivery, and follow-up of the intervention in addition to the resources required to achieve beneficial outcomes, for example, more resources for distribution and real-time discussion of results. Pain, falls, delirium, and hospital visits increased in the post-feedback report period after decreasing during the intervention period. Thus, this study supports the ongoing use of audit with feedback using client data from the RAI-HC so that care givers can continue to provide client-centered service for optimal outcomes in the home care setting.



Data for Improvement and Clinical Excellence – Home Care


Data for Improvement and Clinical Excellence – Long-Term Care


Long-term care


Resident assessment instrument – home care


  1. Fraser KD, Estabrooks CA. How do case managers make resource allocation decisions? A systematic review of the literature. Med Decis Mak. 2008;28:394–410. doi:10.1177/0272989X07312709.

    Article  Google Scholar 

  2. The Canadian Home Care Association: Integrating and coordinating mental health and home care in Canada: a scan of promising practices and existing service delivery models. (2008). Accessed 22 May 2015.

  3. Petrucka P. The Canadian health care system. In: Potter PA, Perry AG, Ross-Kerr JC, Wood MJ, editors. Canadian fundamentals of nursing—revised reprint. 4th ed. Toronto: Elsevier; 2010. p. 20–2.

    Google Scholar 

  4. Fraser KD, Estabrooks CA, Allen M, Strang VA. The relational nature of case manager resource allocation decision-making: an illustrated case. Care Manag J. 2010;11(3):151–6.

    Article  PubMed  Google Scholar 

  5. Markle-Reid M, Browne G, Gafni A. Nurse-led health promotion interventions improve quality of life in frail older home care clients: lessons learned from three randomized trials in Ontario, Canada. J Eval Clin Pract. 2013;19(1):118–31. doi:10.1111/j.1365-2753.2011.01782.x.

    Article  PubMed  Google Scholar 

  6. Ivers N, Jamtvedt G, Flottorp S, et al. Audit and feedback: effects on professional practice and healthcare outcomes. Cochrane Database Syst Rev. 2012;6:CD000259. doi:10.1002/14651858.CD000259.pub3.

    Google Scholar 

  7. Jamtvedt G, Young JM, Kristoffersen DT, et al. Audit and feedback: effects on professional practice and health care outcomes. Cochrane Database Syst Rev. 2006;2:CD000259. doi:10.1002/14651858.CD000259.pub2.

    Google Scholar 

  8. Fisher D. Quality assurance and the AIDS patient. Caring. 1989;8(11):52–4.

    CAS  PubMed  Google Scholar 

  9. Meijers JM, Candel MJ, Schols JM, et al. Decreasing trends in malnutrition prevalence rates explained by regular audits and feedback. J Nutr. 2009;139(7):1381–6. doi:10.3945/jn.108.102608.

    Article  CAS  PubMed  Google Scholar 

  10. Kluger AN, DeNisi A. Feedback interventions: toward the understanding of a double-edged sword. Curr Dir Psychol Sci. 1998;7(3):67–72.

    Article  Google Scholar 

  11. Kluger AN, Van Dijk D. Feedback, the various tasks of the doctor, and the feedforward alternative. Med Educ. 2010;44(12):1166–74. doi:10.1111/j.1365-2923.2010.03849.x.

    Article  PubMed  Google Scholar 

  12. Gardner B, Whittington C, McAteer J, et al. Using theory to synthesise evidence from behaviour change interventions: the example of audit and feedback. Soc Sci Med. 2010;70(10):1618–25. doi:10.1016/j.socscimed.2010.01.039.

    Article  PubMed  Google Scholar 

  13. Hirdes JP. Addressing the health needs of frail elderly people: Ontario’s experience with an integrated health information system. Age Ageing. 2006;35(4):329–31. doi:10.1093/ageing/afl036.

    Article  PubMed  Google Scholar 

  14. Hirdes JP, Sinclair DG, King J, et al. From anecdote to evidence: complex continuing care at the dawn of the information age in Ontario. 2014. Accessed 8 Jun 2015.

  15. Slamet D. Business case for RAI 2.0 implementation: a jurisdictional review of outcomes observed from RAI v2.0 implementation. Vancouver: Fraser Health Region; 2005.

    Google Scholar 

  16. Sales AE, Fraser K, Baylon MA, et al. Understanding feedback report uptake: process evaluation findings from a 13-month feedback intervention in long-term care settings. Implement Sci. 2015;10:20. doi:10.1186/s13012-015-0208-2.

    Article  PubMed  PubMed Central  Google Scholar 

  17. Fraser KD, Sales AE, O’Rourke HM, Schalm C. Data for improvement and clinical excellence: protocol for an audit with feedback intervention in home care and supportive living. Implement Sci. 2012;7:4. doi:10.1186/1748-5908-7-4.

    Article  PubMed  PubMed Central  Google Scholar 

  18. Sales AE, Schalm C, Baylon MA, Fraser KD. Data for improvement and clinical excellence: report of an interrupted time series trial of feedback in long-term care. Implement Sci. 2014;9:161. doi:10.1186/s13012-014-0161-5.

    Article  PubMed  PubMed Central  Google Scholar 

Download references


We gratefully acknowledge the intellectual input from the full research team for this project: Anusha Aduatta, Marian Anderson, Anne-Marie Bostrom, Thorsten Duebel, Kari Elliott, Carole Estabrooks, KDF, Gloria Gao, Vivien Lai, Kaila Lapins, Lili Liu, Suzanne Maisey, Anastasia Mallidou, Lynne Mansell, Colleen Maxwell, Joshua Murray, Iris Neumann, and Sharon Warren.


We acknowledge funding for this project from the Canadian Health Services Research Foundation and the Alberta Heritage Foundation for Medical Research. The funding agency was not involved in drafting this manuscript, and the agency was not involved in the conduct of the project.

Availability of data and materials

The datasets used during the current study are available from the corresponding author on reasonable request.

Authors’ contributions

KDF was the primary investigator of the overall project. AES contributed to the study design. MAB conducted the statistical analyses. CS contributed to the study design. JJM contributed to the data analysis and manuscript preparation. All authors read and approved the final manuscript.

Competing interests

The authors declare that they have no competing interests.

Consent for publication

Not applicable.

Ethics approval and consent to participate

Approval for the study was obtained from the University of Alberta Health Research Ethics Board. All participants provided informed consent.

Publisher’s Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Author information

Authors and Affiliations


Corresponding author

Correspondence to Kimberly D. Fraser.

Additional file

Additional file 1:

DICE Project Feedback Report, December 2011

Rights and permissions

Open Access This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (, which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Fraser, K.D., Sales, A.E., Baylon, M.A.B. et al. Data for Improvement and Clinical Excellence: a report of an interrupted time series trial of feedback in home care. Implementation Sci 12, 66 (2017).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: