Skip to main content
  • Research article
  • Open access
  • Published:

Using a summary measure for multiple quality indicators in primary care: the Summary QUality InDex (SQUID)



Assessing the quality of primary care is becoming a priority in national healthcare agendas. Audit and feedback on healthcare quality performance indicators can help improve the quality of care provided. In some instances, fewer numbers of more comprehensive indicators may be preferable. This paper describes the use of the Summary Quality Index (SQUID) in tracking quality of care among patients and primary care practices that use an electronic medical record (EMR). All practices are part of the Practice Partner Research Network, representing over 100 ambulatory care practices throughout the United States.


The SQUID is comprised of 36 process and outcome measures, all of which are obtained from the EMR. This paper describes algorithms for the SQUID calculations, various statistical properties, and use of the SQUID within the context of a multi-practice quality improvement (QI) project.


At any given time point, the patient-level SQUID reflects the proportion of recommended care received, while the practice-level SQUID reflects the average proportion of recommended care received by that practice's patients. Using quarterly reports, practice- and patient-level SQUIDs are provided routinely to practices within the network. The SQUID is responsive, exhibiting highly significant (p < 0.0001) increases during a major QI initiative, and its internal consistency is excellent (Cronbach's alpha = 0.93). Feedback from physicians has been extremely positive, providing a high degree of face validity.


The SQUID algorithm is feasible and straightforward, and provides a useful QI tool. Its statistical properties and clear interpretation make it appealing to providers, health plans, and researchers.

Peer Review reports


Assessment of the quality of primary care is becoming a clear priority in national healthcare agendas. To evaluate the care provided to patients with chronic illnesses and clinical conditions that affect large segments of the population, numerous quality indicators and performance measures have been developed. For example, performance measurements by the US Centers for Medicare and Medicaid Services (CMS) Physician Focused Quality Initiative, including the Doctor's Office Quality Project, Doctor's Office Quality Information Technology Project, and Vista-Office Electronic Health Record, are being implemented nationally to assess the care of Medicare beneficiaries, support clinicians in providing appropriate treatment, prevent avoidable health problems, and evaluate the concept of pay-for-performance [1]. Other examples include performance measures endorsed by the US National Committee for Quality Assurance, the National Quality Forum, the American Medical Association/Physician Consortium for Performance Improvement, and the Ambulatory Care Quality Alliance [1].

Implementation of research into clinical practice has been facilitated through multiple QI strategies, including audit and feedback [2, 3]. Providing feedback to clinicians on their performance related to specific indicators is one of the components used to improve the quality of care provided. In situations such as this, where numerous quality indicators are utilized, it has been argued that there may be instances in which fewer numbers of more comprehensive indicators are preferable [4]. For example, during quality improvement (QI) projects involving multiple process and/or outcome measures within multiple clinical domains, efforts to improve quality in one area may yield a decline in quality in another area. In such circumstances, a summary measure may provide clinicians and researchers with a better sense of whether their efforts (or lack thereof) result in net increases or decreases in quality.

Several earlier publications have discussed algorithms used to summarize quality measures in different arenas of the healthcare system. For example, CMS has developed a system for summarizing quality indicators for hospitals [5], and investigators with RAND Corporation have created a mechanism for assessing overall quality of care provided to various communities around the US [6, 7]. The US Department of Veterans Affairs (VA) has developed similar evidence-based measures, incorporating a "prevention index" and a "chronic disease index" as a means of encouraging better provider performance [8]. Likewise, several papers have addressed statistical methodology (e.g., latent variable models [9], factor analysis [4], and Bayesian hierarchical regression models [10]) for physician, hospital, or health plan 'profiling,' in which an index is created that compares the overall quality of care provided among various physicians. Global statistical tests have also been proposed for comparisons of multiple correlated outcomes, typically used within the clinical trials setting; however, their use in composite quality indices has been minimal [1114]. Although generally such sophisticated statistical methods provide summaries across multiple quality domains and account for correlation among the individual measures of quality, with the exception of the CMS, RAND, and VA methodologies, the composite indices proposed in those papers do not have a direct clinical interpretation. Additionally, these methods may be inadequate when the composite score includes individual indicators that are not applicable to selected groups of patients.

This paper outlines the construction, validation, and use of the Summary Quality Index (SQUID), a composite measure summarizing the quality of care provided by primary care providers. It was developed in the Practice Partner Research Network (PPRNet), a practice-based research network, for use in a QI demonstration project. PPRNet is a network of ambulatory primary care clinicians throughout the US who use a common electronic medical record (Practice Partner, Seattle, WA). Data from outpatient encounters (e.g., demographics, diagnoses, medications, laboratory results, and vital signs) are remitted quarterly to PPRNet staff at the Medical University of South Carolina, where the data are prepared for analysis and summarized in practice performance reports. Throughout this process, only active adult patients over 18 years old are included. Within PPRNet, a patient is considered active at any point in time if he/she has had a progress note recorded in the electronic medical record in the prior 12 months; a patient is considered to have an active medication if it was prescribed in the prior 12 months. As of the third quarter 2005, 89 practices were represented. Although the SQUID has been developed within the PPRNet setting, the algorithm used to create it is generalizable to many other healthcare settings.

As a part of the QI demonstration project entitled Accelerating the Translation of Research into Practice (A-TRIP), an intervention which spanned 42 months (January 2003 through June 2006), this group of PPRNet clinicians has been provided with quarterly reports on 36 unique quality indicators (see Table 1). Thirty-one of these indicators are process measures, while five are outcome measures. As is customary with performance measurement [15], the indicators were chosen based on the ability of providers to act on them, supporting evidence and national prevention and disease management guidelines [1628], and availability of data from the EMR. Chosen indicators are in the following domains: prevention and management of hypertension (HTN), coronary heart disease (CHD), stroke, diabetes mellitus (DM), and respiratory/infectious disease, cancer screening, immunizations, substance abuse and mental health, nutrition and obesity, and inappropriate prescribing in elderly patients. The A-TRIP QI demonstration project was comprised of three specific types of interventions: practice performance reports (audit and feedback), optional semi-annual site visits to practices for academic detailing and participatory planning, and optional annual network meetings to share 'best practice' approaches. The logic and supporting theory of the A-TRIP intervention has been published elsewhere [29]. The purpose of this paper is to summarize the development of a statistically robust and clinically meaningful composite summary measure that would help the research team and individual practices evaluate the overall progress of a QI demonstration project.

Table 1 A-TRIP quality indicators and eligibility criteria


The algorithm for creating the composite quality measure was developed during the A-TRIP project, which was approved by the Institutional Review Board of the Medical University of South Carolina. The algorithm for creating the SQUID from the 36 quality measures includes 1) determining which patients are eligible for which process and outcome measures; 2) determining which patients have met their desired clinical targets; and 3) calculating SQUIDs for each patient and for each practice.

Determining which patients are eligible for which process and outcome measures

The first step in the SQUID algorithm involves counting the number of process and outcome measures for which the patient is eligible. For example, only patients with DM are eligible for hemoglobin A1c (A1C) monitoring. An indicator variable is thus created, with a one indicating that a given patient is eligible (i.e., has DM) for the particular measure of interest (i.e., A1C monitoring), and a zero indicating that the patient is not eligible (i.e., does not have DM). These indicator variables are denoted by Ei, where E1 is an indicator variable reflecting eligibility for the first unique measure, E2 is an indicator variable reflecting eligibility for the second unique measure, etc., and where 'i' ranges from one to thirty-six, the total number of unique process and outcome measures. The total number of measures for which a patient is eligible is thus E = i = 1 36 E i MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaadaaeWbqaaiabdweafnaaBaaaleaacqWGPbqAaeqaaaqaaiabdMgaPjabg2da9iabigdaXaqaaiabiodaZiabiAda2aqdcqGHris5aaaa@36BC@ . Note that patients with greater numbers of diseases/medical conditions will be eligible for more process and outcome measures, and thus the total (E) may be used subsequently in analyses that need to adjust for the level of patient complexity. Also, all adult patients over 18 years old are eligible for at least six process measures, including blood pressure (BP), total cholesterol and high density lipoprotein (HDL) cholesterol monitoring, tetanus/diphtheria vaccine, depression, and alcohol screening.

Determining which patients have met their desired targets

The next set of indicator variables reflects whether or not the patient has met the targets for the eligible quality measures. For process measures, the target has been met if the process has been performed within some pre-specified time frame (e.g., past six months, past year). For outcome measures, the target has been met if the measure of interest is under (in the case of BP, low density lipoprotein [LDL], triglycerides, and A1C) or over (for HDL) the guideline recommendation. These targets may vary according to the patients' co-morbidities. For example, the BP control target is less than 140/90 mmHg for patients with HTN and less than 130/80 mmHg for patients with DM. Patients with both HTN and DM need to meet the more stringent target (i.e., less than 130/80 mmHg). The relevant indicator variables (Mj's) are then summed so that M, the total number of process/outcome targets that a patient has met, is defined as M = j = 1 E M j . MathType@MTEF@5@5@+=feaafiart1ev1aaatCvAUfKttLearuWrP9MDH5MBPbIqV92AaeXatLxBI9gBaebbnrfifHhDYfgasaacH8akY=wiFfYdH8Gipec8Eeeu0xXdbba9frFj0=OqFfea0dXdd9vqai=hGuQ8kuc9pgc9s8qqaq=dirpe0xb9q8qiLsFr0=vr0=vr0dc8meaabaqaciaacaGaaeqabaqabeGadaaakeaadaaeWbqaaiabd2eannaaBaaaleaacqWGQbGAaeqaaaqaaiabdQgaQjabg2da9iabigdaXaqaaiabdweafbqdcqGHris5aOGaeiOla4caaa@36E3@ .

Calculating SQUIDs at the patient and practice level

Once E and M have been determined for each patient, the patient-level SQUID is simply calculated by dividing M (measures met) by E (eligible measures), thus reflecting the proportion of relevant targets achieved for that patient. Because the SQUID is a proportion, it ranges from 0.0% to 100.0%. Note that the SQUID incorporates both individual process and outcome indicators, as has been done for specific clinical domains in other studies [30, 31].

Another feature of the SQUID is that it can be calculated at the patient, provider, or practice-level. The practice-level SQUID is calculated as the average of all the patient-level SQUIDs among active patients in the practice. The practice-level SQUID thus reflects the average proportion of relevant targets achieved for patients in the practice. In A-TRIP, provider-level SQUIDs were not reported; however, these could easily be calculated in other settings.

Use of the SQUID in QI

Once the patient-level and practice-level SQUIDs were developed, they were incorporated into practice quality performance reports provided to A-TRIP practices on a quarterly basis. From January 2003 to April 1, 2005, participating A-TRIP practices received quarterly performance reports that only encompassed performance on the individual quality measures. After April 1, 2005, practice reports included a statistical process control chart that summarized the practices' performance on their practice-level SQUID. These charts, similar to ones used for the individual quality measures, mapped the practices' SQUID scores on a monthly basis over the past 24 months. Practices were provided these reports as part of the A-TRIP project through the end (i.e., June 2006) of the QI project, and final analyses of the A-TRIP project included an assessment of the change in practice-level SQUID scores over the 3.5 year study time frame. The analysis of the change in SQUID scores during A-TRIP was also presented to providers during the 2005 and 2006 A-TRIP network meetings, which were designed to help providers improve quality by listening to 'best practice' approaches and by discussing their ideas with one another. In fact, the 'best' practices were determined, in part, by performance on their SQUID scores.

In addition to the practice-level reports, throughout the A-TRIP QI effort practices have been provided with patient-level reports, similar to a patient registry. These reports consist of Excel spreadsheets with embedded filters and macros that can help the practice identify their patients not at goal on individual quality measures. Starting with the 2nd quarter 2005 practice-level reports, the patient-level SQUID was added to these reports. By having this overall quality score calculated for individual patients, practices were then able to identify, for example, their patients with the lowest SQUID scores (i.e., those patients with the lowest overall quality scores). They could also go a step further and identify the most complex patients (using the SQUID denominator) with low SQUID scores, to identify their more complex patients in need of improved care.

Measuring SQUID reliability, responsiveness, validity, internal consistency, and distributional properties

Various statistical properties of the SQUID such as reliability, responsiveness, validity, and distributional properties were also of interest. Reliability refers to the degree to which two SQUID measures at different points close in time are correlated with one another, while responsiveness refers to whether the index detects clinically meaningful changes over time. To the extent that the patients' electronic medical record data is accurate, the measure is, by definition, essentially perfectly reliable. Responsiveness was investigated by examining the absolute increase in patients' and practices' SQUID values during this 15-month study period. Because the practices were participating in a QI project, it would be expected that patient-level and practice-level SQUIDs would increase significantly over time. Change over time was assessed for statistical significance using paired t-tests, linear regression, and the Wilcoxon sign test, as appropriate.

Validity refers to the degree to which the measure accurately reflects that which is being measured. Although several types of validity exist, we focused on face validity (i.e., a subjective assessment of whether the SQUID measures that which it was intended to measure). This property was assessed through an e-mail listserv for PPRNet members and through informal interviews with providers who participated in site visits or who attended the 2005 PPRNet A-TRIP network meeting in Seattle, WA.

Other statistical properties were also examined. It has also been recommended that performance measures based on multiple measures need to have good internal consistency, indicating that the individual items are measuring similar constructs [32]. Internal consistency was measured using Cronbach's alpha coefficient among the practices' third quarter 2005 scores on the individual quality indicators that comprise the SQUID. The intraclass correlation coefficient to determine the proportion of patient-level SQUID variation explained by practice membership was also calculated, by using a mixed linear regression model (SAS V9.1, Cary, NC), treating practice as a random effect. The distribution of E (the total number of eligible measures) was examined across the patient population to provide a general sense of its distribution, including the most frequent values observed and the associated variability. Lastly, histograms were created for patient- and practice-level SQUIDs from third quarter 2005 for use in determining their distributional properties, as this type of information may provide further insight into the overall nature of the variation in the quality of care provided. All analyses were performed with SAS 9.1 (Cary, NC).


The third quarter 2005 population studied included 330,966 active adult patients in 89 active PPRNet primary care practices. Table 2 lists key descriptive statistics for these practices and patients within the practices. Most (78.7%) of the practices were family practices, with multiple providers. Of the diseases/conditions of interest, the most frequently reported were hypertension (24.6%) and hyperlipidemia (21.2%). A histogram reflecting the distribution of the total number of eligible indicators (E) is shown in Figure 1. Although E has a distribution that is skewed to the right, the way our indicators are defined, each adult has an E value that is 6 or greater. The median of E is 9, and the mean is 10.6 (s.d. = 4.9).

Table 2 Characteristics of 89 active A-TRIP practices as of September 30, 2005
Figure 1
figure 1

Histogram of third quarter 2005 patient-level total number of eligible indicators (n = 350,307 patients).

The responsiveness of patient and practice-level SQUIDs is highlighted in Table 3. Among patients who were active during the entire 15-month time period, the mean SQUID increased 3.6% (from 40.0% to 43.6%). Among all active patients (during the quarter of interest), the mean SQUID increased 3.2% (from 35.1% to 38.3%). Among practices that were active during the entire 15-month time period, the mean practice-level SQUID increased 3.8% (from 34.8% to 38.6%), with 88% of practices exhibiting a positive increase in their practice-level SQUID score. Additionally, analyses across the entire 3.5 year A-TRIP study indicated an adjusted average annual improvement in the SQUID of 2.43% (95% confidence interval 2.24% to 2.63%), an improvement that was consistent throughout the entire study. These changes were all significantly different from zero (p < 0.0001). The reason why the mean practice-level SQUIDs among patients active for the entire study are lower than the mean patient-level SQUIDs is due to patient turnover. The practice-level SQUIDs incorporate data from many patients who later became inactive during the time period, as well as new patients who join the practice. Since these two groups of patients did not have continual contact with their practice during the 15-month time period, their SQUID scores tended to be lower than the patients who were active throughout the study, thus reducing the values of the overall practice-level SQUIDs.

Table 3 Quarterly means, standard deviations (s.d), correlations among patient-level SQUIDs

When the SQUID algorithm and preliminary findings were presented to clinicians participating in site visits or attending the 2005 and 2006 PPRNet A-TRIP network meetings, feedback was favorable. During site visits, providers and staff reviewed practice-level SQUIDs to further assess their performance on A-TRIP measures. One practice used the trend of increasing SQUIDs to reinforce their focus on improving process measures related to preventive care (i.e., updating aspirin prescriptions in applicable patients, and sending letters to patients overdue for mammograms or colonoscopies). Another practice observed a decreasing trend in practice-level SQUIDs related to growth of their practice, and used their past performance as motivation for providing quality care to an influx of new patients. In general, providers appreciated the fact that the SQUID was an index that had a direct interpretation of the overall quality of care provided in their practices.

When PPRNet e-mail listerv members were asked to provide feedback on the SQUID, several interesting responses emerged, as they commented on how it was used in their practices. Direct quotes from this informal feedback request from physicians include:

"The SQUID ... provides an over-all indication of whether or not a practice is on a 'trajectory of improvement'. We find that there is 'psychic value' to knowing that."

"It's nice to have along with the [other] two graphs comparing us to the rest of the group. We just use it as an overall assessment of how we're doing."

" [We] have been using it as some information for my patients on how the practice does as a whole and for negotiations with insurers."

"We have used this extensively. I presented our data to the corporate fall conference. People were quite impressed. The insurance companies we work with also are excited about our improvements. We use the summary to give an overall view to ourselves (providers), the associates (staff), and others in our network. We follow this measure closely as a gauge of our progress. It would be interesting to use it for specific patients. We could have it to encourage compliance and congratulate successes for certain patients. I envision presenting a graph of that particular person's progress to him/her."

"Last year we had an influx of patients who work for [company X] and were being seen by other docs. Our summary indicator dipped and then came back up – the people at [company X] were most happy. It is a great lead-off slide for presentations...It is the future for medicine."

Patients' third quarter 2005 SQUIDs correlated relatively well (p < 0.0001) with their most recent systolic (r = -0.17) and diastolic (r = -0.23) BP (DM and HTN patients only), LDL (r = -0.26) (DM and CHD patients only), HDL (r = 0.17) (DM patients only), triglycerides (r = -0.16) (DM patients only), and A1C (r = -0.24) (DM patients only) measurements. The directionality of these associations also provide evidence of construct validity for the SQUID; that is, better overall quality was associated with lower values of BP, A1C, LDL, and triglyceride measures as well as higher values of HDL. The Cronbach's alpha coefficient among the practices' scores on the individual quality indicators was found to be 0.93, indicating excellent internal consistency. Although a low internal consistency would not necessarily be indicative of a poor composite measure, the fact that the SQUID does have a high Cronbach's alpha coefficient suggests that it is comprised of indicators measuring a common underlying quality construct.

A histogram of the third quarter 2005 patient-level SQUID values is shown in Figure 2. Note that approximately 4% of patients had SQUID values of zero, and the relatively bimodal distribution, with peaks between 15% and 20% and between 50% and 55%. A histogram of the third quarter 2005 practice-level SQUID values is shown in Figure 3. In contrast to the patient-level SQUIDs, the practice-level SQUID distribution was uni-modal. The average practice-level SQUID was 37.9%, with a standard deviation of 10.7%. The practice-level SQUIDs ranged from 12.3% to 68.3%, and the intra-class correlation coefficient, reflecting the proportion of SQUID variation explained by practice membership, was 23.8%.

Figure 2
figure 2

Histogram of third quarter 2005 patient-level SQUIDs (n = 350,307 patients).

Figure 3
figure 3

Histogram of third quarter 2005 practice-level SQUIDs (n = 89 practices).


This paper describes the Summary Quality Index (SQUID), a composite measure of healthcare quality in the primary care setting. The SQUID has several advantages compared with other composite quality measures. The algorithm is straightforward, and the resulting index satisfies the qualities of good performance measures and good outcome measures. Within the setting of A-TRIP, a QI demonstration project, it has been shown to be a reliable, responsive, and valid measure of healthcare quality. Feedback from clinicians suggests that this type of measure is quite appropriate and acceptable for primary care settings. They appreciate its use for tracking a summary measure of quality over time, and are excited about its potential for appealing internally to their clinical and clerical staff, as well as externally to insurers, corporate officials, and even their patients.

Having a patient-level composite measure is advantageous for several reasons, most notably it allows for comparisons across groups of patients with specific conditions (e.g., diabetes), demographics (e.g., the elderly), or types of care (e.g,. preventive or chronic) [33]. In fact, a subset of the A-TRIP quality indicators relevant to diabetes care has already been used in the development of the Diabetes-SQUID [30], which is ideal for studying ways to improve care for diabetes patients in the primary care setting. During the A-TRIP project, making the patient-level SQUIDs available to the clinicians responsible for the patients' care has allowed those clinicians to identify their most clinically complicated patients (i.e., based on the SQUID denominator values) along with their patients with the greatest need for care improvement (i.e., those with low SQUID scores). Using a composite measure may also be quite useful within QI projects involving multiple process and/or outcome measures within multiple clinical domains. Because efforts to improve quality in one area may yield declines in other areas, a summary measure may provide interested parties with a better sense of the resulting net increases or decreases in performance.

Because the SQUID can be calculated at the patient level, or aggregated to a higher level, such as that of the provider, practice, or health plan, it is useful from a variety of perspectives. As mentioned earlier, practices may use the patient-level SQUID to identify patients in most need of certain types of care. However, they may also use their practice-level SQUID as a marker of QI over time, or to compare their progress against that of other practices in their network. Health plans might use provider-level SQUIDs to rank providers or track progress over time, and researchers or QI organizations might use practice-level SQUIDs to rank practices or track them over time.

Because the denominator (referred to as 'E' in the algorithm) used in calculating the SQUID reflects the total number of relevant indicators for a given patient, a rather intuitive "complexity" adjustor is created in the process of calculating each patient's SQUID. Although this value (E) does not reflect the severity or duration of any individual patient conditions, it does reflect an overall level of complexity for that patient, because it includes a number of unique chronic conditions that are commonly treated in the primary care setting. This denominator can serve as a covariate in patient-level regression models for the purposes of complexity adjustment (analogous to risk adjustment), or it can be averaged across patients to serve as a complexity adjustor in provider or practice-level analyses.

This approach to quantifying overall quality of care is emerging as a useful tool in practice, in QI, and in research. Other algorithms mentioned in the literature for composite quality measures have typically been aimed at some aggregated level (rather than at the patient level), such as those used in physician or health plan profiling [4, 5, 9, 10]. With the exception of the method described by CMS for quantifying multiple quality measures for hospitals [5], these algorithms involve the creation of some composite index that typically has no direct clinical interpretation. One set of methods that has been mentioned in the medical literature for combining multiple patient-level outcomes is the use of global statistical tests [1114]. These tests can be an excellent way to account for correlated outcomes among patients in clinical trials; however, their effectiveness is limited when one or more of the outcomes is not relevant for significant numbers of patients (e.g., gender-specific measures such as whether a Pap test has been done in the past 3 years). The SQUID algorithm is similar to ones developed by CMS, RAND Corporation, and the VA [5, 7, 8]. The CMS methodology, however, has only been applied to the hospital setting, rather than at a patient or physician level, and likewise the VA aggregate indices are used as performance measures across groups of patients. The RAND methodology is broader in nature but relies on patient surveys and medical record abstracts.

There is much debate about the manner in which quality of healthcare should be measured [34]. For example, there are aspects of quality such as patient satisfaction, access to care, certain health outcomes, and efficiency that are not easily measured using electronic medical record or administrative data. Additionally, there is no consensus on whether quality should be measured as a single construct or as multiple domains [35]. Thus balancing what is practical and economical with what is desirable from various perspectives (e.g., patients, providers, insurers, and researchers) will likely continue to be a source of controversy.

The SQUID satisfies the criteria for a good outcome measure that can be used in clinical research studies, including being appropriate, reliable, responsive, precise, interpretable, acceptable, and feasible [36]. The SQUID also satisfies criteria for a desirable performance measure, as defined in a consensus document of the American Medical Association, the Joint Commission on Accreditation of Healthcare Organizations, and the National Committee for Quality Assurance [37]. These criteria included being of high priority for maximizing the health of persons or populations, financially important, able to demonstrate variation in care and/or the potential for improvement, based on established clinical recommendations, potentially actionable by users, and meaningful and interpretable to users. Another strength of this approach is that the SQUID can be easily adapted to reflect revisions in evidence for individual quality indicators.

The actual practice-level SQUID descriptive statistics (mean: 37.9%; standard deviation: 10.7%; range: 12.3% to 68.3%; intraclass correlation coefficient: 23.7%) may seem as a cause for concern, especially when compared to the RAND study's finding that adults in 12 metropolitan areas in the US received 54.9 percent of recommended care, ranging from 51% (Little Rock) to 59% (Seattle) [7]. However, the SQUID calculations for the PPRNet practices do rely on documentation of process of care within certain specific areas of the electronic medical record compared to patient telephone surveys and chart review by the RAND investigators. Thus we may have underestimated the true quality provided in these practices, due to some physicians opting to record data in the records in a manner (i.e., within the text of a progress note) that is not obtainable via the current PPRNet data extraction process.

The intraclass correlation coefficient for the patient-level SQUID (i.e., 23.8%) may seem relatively high in comparison with ICCs for outcomes of other studies [38, 39]. However, because these practices were all involved to with a QI project during this time period, and since practices were allowed to determine the extent to which they participated in A-TRIP, we expected high variability in patient healthcare quality and that practice membership would explain much of this variation.

There are several limitations of this summary quality measure. Currently, each of the individual processes and outcomes comprising the SQUID is equally weighted, and it could be argued that certain process or outcome indicators should be weighted more heavily. Certain indicators may be viewed as being more clinically important than others, and other indicators may be easier to achieve than others. It is also possible that certain individual processes or outcomes may interact with one another, having synergistic or even antagonistic effects on overall quality; however, examining the influence of such interactions was beyond the scope of this study. Although it is possible that indicator-specific weights could be incorporated into the SQUID's summation formulas, deriving them would typically require building some type of group consensus or using statistical methodology such as factor analysis or item response theory methods [40]. One of the difficulties of these empirical approaches in the context of our patient population is the fact that many patients are not eligible for multiple measures; thus trying to determine how indicators cluster together or whether certain indicators are more difficult than others would require much more in-depth analyses that took into consideration eligibility differences among patients. Even if such analyses were conducted, resulting in a revised weighting scheme for each indicator, we would argue that such a process would result in a loss in the ease of interpretability of the SQUID, a factor we feel is key in communicating with an extremely varied audience that includes providers with varied levels of training and expertise (doctors, physician assistants, nurses), office staff, and even patients. Item weighting (or possibly item reduction) may, however, help address another potential limitation of the SQUID, that some of the individual indicators are correlated with one another. For example, practices that do well in measuring patients' total cholesterol routinely also tend to do well in measuring their patients' HDL and LDL cholesterol levels. Future research into possible weighting and/or item reduction schemes for the individual indicators could help sort out these issues. Additionally, as a general performance measure, the SQUID algorithm does not account for patient allergies or other contraindications to immunizations or medications; thus it would be virtually impossible for a practice to achieve a practice-level SQUID score of 100%. This fact is communicated to practices during site visits and network meetings, and practices are given a sense of what is practically achievable via internally derived SQUID benchmarks. One other potential limitation of the SQUID is the multi-modal nature of its distribution at the patient-level and the fact that it is bounded by 0% and 100%. Thus caution should be used when analyzing SQUID data from small numbers of patients. Lastly, although the SQUID may be useful in detecting general trends over time in quality, specific problematic areas within a given practice are likely more easily identified via individual- or condition-specific indicators.


The SQUID has been a helpful tool in quantifying overall quality within the A-TRIP demonstration project. Providers have used the practice-level SQUIDs to assess overall performance on quality indicators in 8 clinical domains, and they have used the patient-level SQUIDs to identify the patients in most need of attention. A-TRIP research investigators have used it to identify practices making the largest gains in overall QI. The ability to identify these 'best practices' allows us to encourage dialogue between practices during annual A-TRIP network meetings, in which physicians, nurses, and other office staff share ideas to improve the quality of the care they provide. The SQUID values have also served as the primary outcomes in the final analyses of the A-TRIP project. Thus, it has benefit to patients, practitioners, insurers, and researchers.


  1. Centers for Medicare and Medicaid Services: Physician Focused Quality Initiative. 2006, []

    Google Scholar 

  2. Grimshaw J, McAuley LM, Bero LA, Grilli R, Oxman AD, Ramsay C, Vale L, Zwarenstein M: Systematic reviews of the effectiveness of quality improvement strategies and programmes. Qual Saf Health Care. 2003, 12: 298-303. 10.1136/qhc.12.4.298.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

  3. Davis DA, Thomson MA, Oxman AD, Haynes RB: Changing physician performance. A systematic review of the effect of continuing medical education strategies. JAMA. 1995, 274: 700-705. 10.1001/jama.274.9.700.

    Article  CAS  PubMed  Google Scholar 

  4. Zaslavsky AM, Shaul JA, Zaborski LB, Cioffi MJ, Cleary PD: Combining health plan performance indicators into simpler composite measures. Health Care Financ Rev. 2002, 23: 101-115.

    PubMed  PubMed Central  Google Scholar 

  5. Centers for Medicare and Medicaid Services: HQI Composite Ranking Index Calculation. 2006, []

    Google Scholar 

  6. Kerr EA, McGlynn EA, Adams J, Keesey J, Asch SM: Profiling the quality of care in twelve communities: results from the CQI study. Health Aff. 2004, 23: 247-256. 10.1377/hlthaff.23.3.247.

    Article  Google Scholar 

  7. McGlynn EA, Asch SM, Adams J, Keesey J, Hicks J, DeCristofaro A, Kerr EA: The quality of health care delivered to adults in the United States. New Engl J Med. 2003, 348: 2635-2645. 10.1056/NEJMsa022615.

    Article  PubMed  Google Scholar 

  8. Perlin JB, Kolodner RM, Roswell RH: The Veterans Health Administration: quality, value, accountability, and information as transforming strategies for patient-centered care. Healthc Pap. 2005, 5: 10-24.

    Article  PubMed  Google Scholar 

  9. Landrum MB, Bronskill SE, Normand ST: Analytic methods for constructing cross-sectional profiles of health care providers. Health Serv Outcomes Res Methodol. 2000, 1: 23-47. 10.1023/A:1010093701870.

    Article  Google Scholar 

  10. Normand ST, Glickman ME, Gastonis CA: Statistical methods for profiling providers of medical care: issues and applications. J Am Stat Assn. 1997, 92: 803-814. 10.2307/2965545.

    Article  Google Scholar 

  11. O'Brien PC: Procedures for comparing samples with multiple endpoints. Biometrics. 1984, 40: 1079-1087. 10.2307/2531158.

    Article  PubMed  Google Scholar 

  12. Tilley BC, Pillemer SR, Heyse SP, Li S, Clegg DO, Alarcon GS: Global statistical tests for comparing multiple outcomes in rheumatoid arthritis trials. MIRA Trial Group. Arthritis Rheum. 1999, 42: 1879-1888. 10.1002/1529-0131(199909)42:9<1879::AID-ANR12>3.0.CO;2-1.

    Article  CAS  PubMed  Google Scholar 

  13. Pocock SJ, Geller NL, Tsiatis AA: The analysis of multiple endpoints in clinical trials. Biometrics. 1987, 43: 487-498. 10.2307/2531989.

    Article  CAS  PubMed  Google Scholar 

  14. Lefkopoulou M, Ryan L: Global tests for multiple binary outcomes. Biometrics. 1993, 49: 975-988. 10.2307/2532240.

    Article  CAS  PubMed  Google Scholar 

  15. Improvement PCP: Introduction to Physician Performance Measurement Sets: Tools Developed by Physicians for Physicians. 2006, []

    Google Scholar 

  16. Guidelines for the evaluation and management of heart failure. Report of the American College of Cardiology/American Heart Association Task Force on Practice Guidelines (Committee on Evaluation and Management of Heart Failure). Circulation. 1995, 92: 2764-2784.

  17. Centers for Disease Control and Prevention: Prevention of pneumococcal disease: recommendations of the Advisory Committee on Immunization Practices. Mor Mortal Wkly Rep CDC Surveill Summ. 1997, 46: 1-24.

    Google Scholar 

  18. Chobanian AV, Bakris GL, Black HR, Cushman WC, Green LA, Izzo JL, Jones DW, Materson BJ, Oparil S, Wright JT, Roccella EJ: The Seventh Report of the Joint National Committee on Prevention, Detection, Evaluation, and Treatment of High Blood Pressure: the JNC 7 report. JAMA. 2003, 289: 2560-2572. 10.1001/jama.289.19.2560.

    Article  CAS  PubMed  Google Scholar 

  19. Third Report of the National Cholesterol Education Program (NCEP) Expert Panel on Detection, Evaluation, and Treatment of High Blood Cholesterol in Adults (Adult Treatment Panel III) final report. Circulation. 2002, 106: 3143-3421.

  20. Smith SC, Blair SN, Criqui MH, Fletcher GF, Fuster V, Gersh BJ, Gotto AM, Gould KL, Greenland P, Grundy SM, .: Preventing heart attack and death in patients with coronary disease. Circulation. 1995, 92: 2-4.

    Article  PubMed  Google Scholar 

  21. Prystowsky EN, Benson DW, Fuster V, Hart RG, Kay GN, Myerburg RJ, Naccarelli GV, Wyse DG: Management of patients with atrial fibrillation. A Statement for Healthcare Professionals. From the Subcommittee on Electrocardiography and Electrophysiology, American Heart Association. Circulation. 1996, 93: 1262-1277.

    Article  CAS  PubMed  Google Scholar 

  22. Diabetes Quality Improvement Project, Initial Measure Set (Final Version). 2004, []

  23. Smith RA, Saslow D, Sawyer KA, Burke W, Costanza ME, Evans WP, Foster RS, Hendrick E, Eyre HJ, Sener S: American Cancer Society guidelines for breast cancer screening: update 2003. CA Cancer J Clin. 2003, 53: 141-169.

    Article  PubMed  Google Scholar 

  24. American Cancer Society: ACS Cancer Detection Guidelines. 2006, []

    Google Scholar 

  25. Pearson TA, Blair SN, Daniels SR, Eckel RH, Fair JM, Fortmann SP, Franklin BA, Goldstein LB, Greenland P, Grundy SM, Hong Y, Miller NH, Lauer RM, Ockene IS, Sacco RL, Sallis JF, Smith SC, Stone NJ, Taubert KA: AHA Guidelines for Primary Prevention of Cardiovascular Disease and Stroke: 2002 Update: Consensus Panel Guide to Comprehensive Risk Reduction for Adult Patients Without Coronary or Other Atherosclerotic Vascular Diseases. American Heart Association Science Advisory and Coordinating Committee. Circulation. 2002, 106: 388-391. 10.1161/01.CIR.0000020190.45892.75.

    Article  PubMed  Google Scholar 

  26. American Lung Association: Cold and Flu Guidelines: Influenza. 2006, []

    Google Scholar 

  27. U.S.Preventive Services Task Force: Recommendations and Rationale: Screening for Depression. 2006, []

    Google Scholar 

  28. Beers MH: Explicit criteria for determining potentially inappropriate medication use by the elderly. An update. Arch Intern Med. 1997, 157: 1531-1536. 10.1001/archinte.157.14.1531.

    Article  CAS  PubMed  Google Scholar 

  29. Feifer C, Ornstein SM, Jenkins RG, Wessell A, Corley ST, Nemeth LS, Roylance L, Nietert PJ, Liszka H: The logic behind a multimethod intervention to improve adherence to clinical practice guidelines in a nationwide network of primary care practices. Eval Health Prof. 2006, 29: 65-88. 10.1177/0163278705284443.

    Article  PubMed  Google Scholar 

  30. Ornstein SM, Nietert PJ, Jenkins RG, Wessell AM, Nemeth LS, C. F, Corley ST: Improving diabetes care through a multi-component quality improvement model in a practice-based research network. Am J Med Qual. 2007, 22: 34-41. 10.1177/1062860606295206.

    Article  PubMed  Google Scholar 

  31. Beaulieu ND, Horrigan DR: Putting smart money to work for quality improvement. Health Serv Res. 2005, 40: 1318-1334. 10.1111/j.1475-6773.2005.00414.x.

    Article  PubMed  PubMed Central  Google Scholar 

  32. McGlynn EA, Asch SM: Developing a clinical performance measure. Am J Prev Med. 1998, 14: 14-21. 10.1016/S0749-3797(97)00032-9.

    Article  CAS  PubMed  Google Scholar 

  33. Eddy DM: Performance measurement: problems and solutions. Health Aff. 1998, 17: 7-25. 10.1377/hlthaff.17.4.7.

    Article  CAS  Google Scholar 

  34. Brook RH, McGlynn EA, Shekelle PG: Defining and measuring quality of care: a perspective from US researchers. Int J Qual Health Care. 2000, 12: 281-295. 10.1093/intqhc/12.4.281.

    Article  CAS  PubMed  Google Scholar 

  35. Kahn KL, Tisnado DM, Adams JL, Liu H, Chen W, Hu FA, Mangione CM, Hays RD, Damberg CL: Does ambulatory process of care predict health-related quality of life outcomes for patients with chronic disease?. Health Serv Res. 2007, 42: 63-83. 10.1111/j.1475-6773.2006.00604.x.

    Article  PubMed  PubMed Central  Google Scholar 

  36. Fitzpatrick R, Davey C, Buxton MJ, Jones DR: Evaluating patient-based outcome measures for use in clinical trials. Health Technol Assess. 1998, 2: i-iv, 1-74.

    Google Scholar 

  37. Physician Consortium for Performance Improvement: Desirable attributes for performance measures: a consensus document from the American Medical Association, The Joint Commission on Accreditation of Healthcare Organizations, and The National Committee for Quality Assurance. 1999, []

    Google Scholar 

  38. Ornstein SM, Jenkins RG, Nietert PJ, Feifer C, Roylance LF, L. N, S. C, Dickerson LM, Bradford WD, C. L: Multi-method quality improvement intervention vs. quarterly performance reports to improve preventive cardiovascular care: a cluster randomized trial. Ann Intern Med. 2004, 141: 523-532.

    Article  PubMed  Google Scholar 

  39. Parker DR, Evangelou E, Eaton CB: Intraclass correlation coefficients for cluster randomized trials in primary care: the cholesterol education and research trial (CEART). Contemp Clin Trials. 2005, 26: 260-267. 10.1016/j.cct.2005.01.002.

    Article  PubMed  Google Scholar 

  40. Hays RD, Morales LS, Reise SP: Item response theory and health outcomes measurement in the 21st century. Med Care. 2000, 38: II28-II42. 10.1097/00005650-200009002-00007.

    Article  CAS  PubMed  PubMed Central  Google Scholar 

Download references


This study was supported by a grant from the Agency for Healthcare Research and Quality (Grant 1 U18 HS013716).

Author information

Authors and Affiliations


Corresponding author

Correspondence to Paul J Nietert.

Additional information

Competing interests

The author(s) declare that they have no competing interests.

Authors' contributions

PJN participated in the SQUID methodology development, development of the feedback tools, statistical analysis, manuscript drafting, and critical review of the manuscript. AMW participated in the methodology development, manuscript drafting, and critical review of the manuscript. RGJ participated in the methodology development, development of the feedback tools, manuscript preparation, and critical review of the manuscript. CF participated in the methodology development, statistical analysis, and critical review of the manuscript. LN participated in the methodology development and critical review of the manuscript. SMO participated in the SQUID methodology development, development of the feedback tools, manuscript drafting, and critical review of the manuscript.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2

Authors’ original file for figure 3

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and permissions

About this article

Cite this article

Nietert, P.J., Wessell, A.M., Jenkins, R.G. et al. Using a summary measure for multiple quality indicators in primary care: the Summary QUality InDex (SQUID). Implementation Sci 2, 11 (2007).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI: