Skip to main content

The meaning and measurement of implementation climate



Climate has a long history in organizational studies, but few theoretical models integrate the complex effects of climate during innovation implementation. In 1996, a theoretical model was proposed that organizations could develop a positive climate for implementation by making use of various policies and practices that promote organizational members' means, motives, and opportunities for innovation use. The model proposes that implementation climate--or the extent to which organizational members perceive that innovation use is expected, supported, and rewarded--is positively associated with implementation effectiveness. The implementation climate construct holds significant promise for advancing scientific knowledge about the organizational determinants of innovation implementation. However, the construct has not received sufficient scholarly attention, despite numerous citations in the scientific literature. In this article, we clarify the meaning of implementation climate, discuss several measurement issues, and propose guidelines for empirical study.


Implementation climate differs from constructs such as organizational climate, culture, or context in two important respects: first, it has a strategic focus (implementation), and second, it is innovation-specific. Measuring implementation climate is challenging because the construct operates at the organizational level, but requires the collection of multi-dimensional perceptual data from many expected innovation users within an organization. In order to avoid problems with construct validity, assessments of within-group agreement of implementation climate measures must be carefully considered. Implementation climate implies a high degree of within-group agreement in climate perceptions. However, researchers might find it useful to distinguish implementation climate level (the average of implementation climate perceptions) from implementation climate strength (the variability of implementation climate perceptions). It is important to recognize that the implementation climate construct applies most readily to innovations that require collective, coordinated behavior change by many organizational members both for successful implementation and for realization of anticipated benefits. For innovations that do not possess these attributes, individual-level theories of behavior change could be more useful in explaining implementation effectiveness.


This construct has considerable value in implementation science, however, further debate and development is necessary to refine and distinguish the construct for empirical use.

Peer Review reports


Katherine Klein and Joann Sorra's [1] theory of innovation implementation has become increasingly prominent in the field of implementation science. The article in which the theory first appeared has been cited 258 times since its publication in 1996. Reflecting the theory's popularity in health and human services research, one-third of the 258 citing articles focus on innovation implementation in hospitals, physician practices, community health centers, substance abuse organizations, mental health agencies, and child welfare organizations. The theory's appeal derives partly from its simplicity. Klein and Sorra [1] identified two key determinants of effective implementation: implementation climate, or the extent to which intended users perceive that innovation use is expected, supported, and rewarded; and innovation-values fit, or the extent to which intended users perceive that innovation use is consistent with their values. Although innovation-values fit seems to have garnered more attention, especially among mental health and substance abuse researchers [29], implementation climate is arguably the more important construct, both in terms of its role in Klein and Sorra's [1] theory and for its potential to bring theoretical and empirical coherence to the growing body of research on organizational 'facilitators and barriers' of effective implementation.

Klein and Sorra [1] developed the implementation climate construct based on an extensive review of the determinants of effective information technology implementation. They observed that organizations use a wide variety of policies and practices to promote innovation use. Examples include training, technical support, incentives, persuasive communication, end-user participation in decision making, workflow changes, workload changes, alterations in staffing levels, alterations in staffing mix, new reporting requirements, new authority relationships, implementation monitoring, and enforcement procedures. Not only do organizations vary in their use of specific 'implementation policies and practices,' but the effectiveness of these policies and practices varies from organization to organization and innovation to innovation. In some contexts, for example, the provision of high-quality training is crucial for implementation success. In other contexts, the provision of highly valued rewards, not training, makes the difference. In light of such diversity in organizational practice and variability in effectiveness, Klein and Sorra [1] developed the construct of implementation climate to shift attention to the collective influence of the multiple policies and practices that organizations employ to promote innovation use. Implementation climate is a shared perception among intended users of an innovation, of the extent to which an organization's implementation policies and practices encourage, cultivate, and reward innovation use. The stronger the implementation climate, they assert, the more consistent high-quality innovation use will be in an organization, provided the innovation fits intended users' values. Moreover, if implementation climates of equal strength can result from different combinations of implementation policies and practices, as Klein and Sorra [1] claim, then a focus on implementation climate could bring theoretical parsimony and greater cumulativeness to scientific knowledge about the organizational determinants of innovation implementation.

Despite the construct's potential value to the field of implementation science, several conceptual and methodological problems threaten to undermine its theoretical distinctiveness and empirical utility. First, the construct has suffered from theoretical neglect. Less than a third of the 258 articles citing Klein and Sorra's [1] work discuss implementation climate, and many that do refer to the construct do so only in passing. Second, researchers have sometimes treated implementation climate as synonymous with related, yet distinct constructs such as receptive organizational context [10, 11], supportive organizational context [12], and organizational culture [13]. Third, notwithstanding the widespread appeal of Klein and Sorra's [1] theory, the construct of implementation climate has been assessed empirically in only six studies [1419], one of which was qualitative assessment [15]. Regrettably, three of the five quantitative studies exhibit levels of analysis problems (i.e., the statistical models were mis-specified), a flaw that raises concerns about the interpretation and value of the research findings. Finally, and not surprisingly, given the dearth of empirical research just noted, no standard instrument exists for measuring implementation climate. Few instruments have been used more than once, each instrument differs somewhat in content, and none has been systematically assessed for reliability and validity at the appropriate (organizational) level of analysis.

In this article, we clarify the meaning of implementation climate and distinguish it from other constructs important in implementation science. In addition to exploring conceptual matters, we discuss the levels of analysis issue and other measurement considerations upon which the proper testing of the theory and the utility of the construct in implementation research depend. Our intent in exploring these conceptual and methodological concerns is to promote further scholarly discussion of this important construct and foster the cumulative production of knowledge about the organizational determinants of effective implementation.


What is implementation climate?

Klein and Sorra [1, p. 1060] define implementation climate as 'targeted employees' shared summary perceptions of the extent to which their use of a specific innovation is rewarded, supported, and expected within an organization.' Six features of this definition have important conceptual and methodological implications.

First, and most importantly from a conceptual standpoint, implementation climate has a specific strategic focus: innovation implementation. Unlike organizational climate, culture, or context, implementation climate does not describe a general state of affairs in an organization. As early as 1975, Schneider [20] recognized that climate, as an abstract construct, seems to include organizational members' perceptions of anything and everything that occurs in an organization. Giving the construct a strategic focus narrows attention to organizational members' perceptions of those organizational policies, practices, and procedures that promote a specific behavior or outcome (e.g., innovation implementation). This not only sharpens the construct's conceptual boundaries, Schneider argues [20, 21], it also increases the construct's predictive validity by emphasizing perceptions that are psychologically proximal to the behavior or outcome of interest (e.g., implementation). Since Schneider's critique [20], scholars have proposed, theorized, and assessed climates for service [2225], safety [26, 2633], creativity [3438], and justice [3943]. Although disparate in their strategic focus, these climates 'for something,' like implementation climate, focus on organizational members' shared perceptions of policies, practices, and procedures that orient behavior toward a specific organizational goal.

Second, implementation climate not only focuses on innovation implementation, but is also innovation-specific. Following Schneider [20], Klein and Sorra [1] insist that multiple implementation climates can exist simultaneously in an organization. Thus, a strong implementation climate can exist for one innovation (e.g., clinical decision support) and not another (e.g., patient-centered medical homes) if organizational members perceive differences in the extent to which innovation use is expected, supported, and rewarded. Although conceptually distinct, implementation climates for different innovations could be empirically correlated if the same implementation policies and practices pertain to multiple innovations, or the broader organizational climate, culture, or context that exists in the organization exerts a strong and pervasive influence on organizational members' perceptions and actions.

Third, Klein and Sorra [1] use the term 'targeted employees' to refer to those organizational members who are expected either to use an innovation directly (e.g., front-line staff) or to support an innovation's use (e.g., information technology specialists, supervisors). We use the term 'organizational members' rather than targeted employees because, in healthcare, the expected users of an innovation are not always employed by the implementing organization (e.g., private-practice physicians with hospital privileges). As we discuss later, the idea that implementation climate embraces the perceptions of both expected innovation users and innovation supporters has implications for sampling and measurement.

Fourth, implementation climate refers to organizational members' shared perceptions, not to their individual or idiosyncratic views. Climate researchers have long recognized that climate is a multilevel construct [20, 21, 4451]. It can be conceived and assessed at the organizational, unit, group, or individual level of analysis. Klein and Sorra [1] construe implementation climate as an organization-level construct and focus on organizational members' shared perceptions because innovation implementation in organizations is often a collective endeavor, with many people contributing something to the implementation effort. Electronic health records, chronic care models, open access scheduling, patient-centered medical homes, rapid response teams, quality improvement programs, and patient safety systems are examples of innovations that exhibit implementation complexity (i.e., implementation tasks must be coordinated across people, departments, shifts, or locations) and outcome interdependence (i.e., anticipated benefits depend on collective, not just personal, innovation use). For such innovations, implementation problems are likely to arise if some expected users and supporters perceive that innovation use is expected, supported, and rewarded, while others do not. We discuss this point further in a later section.

Fifth, implementation climate refers to organizational members' 'summary' perceptions of the extent to which the innovation use is expected, supported, and rewarded. Similar to other climate researchers [20, 22, 47, 50, 52], Klein and Sorra see implementation climate as a gestalt perception of the multiple and various policies and practices that an organization puts into place to promote innovation use. The focus on gestalt perceptions is consistent with their view that implementation policies and practices are cumulative, compensatory, and equifinal. Generally speaking, the more implementation policies and practices the organization uses, the better; however, the presence of some high-quality policies and practices could compensate for the absence, or low quality, of other policies and practices. For example, high-quality in-person training could substitute for poor-quality program manuals. Finally, as suggested earlier, different mixes of policies and practices can produce equivalent implementation climates. This implies that implementation climate should be measured as a composite of organizational members' perceptions of implementation policies and practices.

Finally, implementation climate focuses on organizational members' perceptions, not their attitudes. Like other climate researchers [17, 49, 53], Klein and Sorra [1] emphasize that climate perceptions are descriptive, not evaluative, in content. This means that implementation climate is not synonymous with organizational members' satisfaction with or appraisal of the innovation itself (e.g., perceived need, level of evidence) or the organization's implementation policies and practices (e.g., satisfaction with training or technical assistance). We discuss the measurement implications of this point in a later section.

What generates implementation climate?

Organizations can create a positive climate for implementation by employing a variety of policies and practices to enhance organizational members' means, motives, and opportunity for innovation use (see Figure 1). For example, organizations can create a positive climate by making sure that expected innovation users have easy access to high-quality training, technical assistance, and documentation (all of which enhance knowledge and skills); engaging expected users and supporters in decision making about innovation design and implementation, providing incentives for innovation use, and providing feedback on innovation use (all of which enhance motivation), and by making the innovation easily accessible or easy to use, giving expected users time to learn how to use the innovation, and redesigning work processes to fit innovation use (all of which increase opportunities or remove obstacles). Klein and Sorra use the shorthand phrase 'implementation policies and practices' to refer to the array of strategies that organizations put into place to promote innovation use. Implementation policies and practices can be temporary measures that intentionally or naturally disappear when the consistency and quality of innovation use reaches desired levels. Alternatively, they can remain in place long after initial or early implementation in order to support and reinforce continued innovation use.

Figure 1
figure 1

Implementation climate: its antecedents, consequences, and modifiers. Dashed lines indicate relationships discussed by Klein and Sorra (1996), but not discussed in this article. a. Strategic accuracy of innovation adoption (not discussed in this article) refers to the innovation's 'fit' with the strategic problem its adoption is intended to solve. b. Innovation effectiveness (not discussed in this article) refers to the benefits an organization receives as a result of its implementation of a given innovation.

Although implementation policies and practices are the primary basis for implementation climate perceptions, broader organizational features like organization climate, culture, or context may also play a role. Theory and research on the subject is limited. However, in their study of teachers' use of new computer technology in science education, Holahan et al.[16] found that organizational receptivity toward change was positively associated with implementation climate, and implementation climate fully mediated the effect of organizational receptivity toward change on teachers' innovation use. Similarly, building on his empirical work on service climate in banks [22], Schneider [21] proposed that service climate is influenced not just by specific organizational routines to promote good customer service, but also by 'deeper' organizational attributes, such as general human resource practices. More research is needed, but it may be the case that implementation climate arises from an amalgam of implementation policies and practices and broader organizational features. This amalgam is likely to be complex. An organization that values innovation and experimentation, for example, might not need to offer specific rewards or incentives for innovation use. Cultural values alone might be sufficient to support a positive implementation climate. On the other hand, an organization that values tradition and caution might find it essential to offer specific rewards or incentives for innovation use. These rewards or incentives would have to be powerful to counteract the dampening effect of the organization's culture on implementation climate.

Klein and Sorra [1] suggest several processes through which organizational members develop, or could develop, shared implementation climate perceptions. First, shared perceptions could result from organizational members' shared experiences with, observations of, and discussions about the organization's implementation policies and practices. Consistent leadership messages and actions could also promote common understandings among organizational members of the goals, tasks, roles, and performance expectations associated with innovation use [28, 29, 5456]. Finally, broader organizational processes like attraction, selection, socialization, and attrition might also play a role [17, 57, 58]. By increasing the similarity in organizational members' backgrounds, experiences, values, and beliefs, these broader organizational processes increase the likelihood that organizational members will hold similar perceptions of the organization's implementation policies and practices. Conversely, organizational members are unlikely to hold common perceptions of implementation policies and practices when intra-organizational units have limited opportunity to interact and share information, when leaders communicate inconsistent messages or act in inconsistent ways, or when organizational members do not have similar backgrounds, experiences, values and beliefs.

With its emphasis on shared perception, the construct of implementation climate implies a high level of agreement in organizational members' perceptions of implementation policies and practices. The degree of 'within-group agreement' should be tested, not assumed, because, as just indicated, organizational members can vary in their perceptions of implementation policies and practices. The absence of shared perception, or put differently, the presence of high 'within-group variability,' implies that implementation climate does not exist. In other words, there is no shared meaning about the organization's implementation policies and practices [45, 57].

High within-group variability, however, can be theoretically meaningful in its own right. In recent years, climate researchers have distinguished climate strength (the degree of within-group variability in perceptions) from climate level (the average magnitude of perceptions), and proposed that the former moderates the effect of the latter [24, 39, 54, 56, 59]. Building on Mischel's [60] idea of situational strength, they argue that people behave more uniformly in situations that provide clear, powerful cues about the desirability of potential behaviors. By contrast, individual differences govern behavior when situations provide ambiguous or weak cues. It follows that when implementation climate is both strong (i.e., shared) and positive, organizational members are collectively more likely to use an innovation. Conversely, when implementation climate is both strong (i.e., shared) and negative, they are collectively less likely to use an innovation. When implementation climate is weak (i.e., not shared), organizational members are likely to vary in their innovation use as a function of individual differences (e.g., personality traits, personal values) or, in complex organizations, group differences (e.g., inter-unit variability in implementation climate). The moderating effect of climate strength on climate level has not been tested in implementation research, but it does receive support from studies of service climate and team climate [24, 39, 54, 59].

What outcomes result from positive implementation climate?

Klein and Sorra [1, p. 1058] propose that implementation climate is positively associated with implementation effectiveness, which they define as 'the overall, pooled or aggregate consistency and quality of [organizational members'] innovation use.' Like implementation climate, these authors conceive implementation effectiveness as an organization-level construct. Although they recognize that individuals and groups can vary in their innovation use, they emphasize organizational members' pooled or aggregate innovation use. This emphasis is consistent with their theoretical focus on innovations that require active, coordinated use by many organizational members (e.g., electronic health records). For such innovations, they argue, implementation is more effective--and more likely to generate anticipated benefits--when all expected users use the innovation consistently and well than when some expected users use the innovation consistently and well while others use it inconsistently or poorly.

Few studies have quantitatively tested Klein and Sorra's [1] theory of innovation implementation in organizations. However, there is some evidence to support their prediction that implementation climate is positively associated with implementation effectiveness. For example, Holahan et al.[16] found that implementation climate was positively associated with both the quality and consistency of teachers' use of new computer technologies in science education in 69 K-12 schools in New Jersey. Klein et al.[61] found that the implementation climate was positively associated with consistent, high-quality use of advanced computerized manufacturing technology in 39 plants located across the United States. However, Klein et al. measured implementation climate as the extent to which innovation implementation was perceived to be important (or a priority) in the organization. This slippage between the construct's conceptual and operational definitions renders the meaning of the study's findings ambiguous. Consistent with Klein and Sorra's [1] predictions, Dong et al.[14] found in their study of large-scale information systems implementation that implementation effectiveness was highest when implementation climate was positive and innovation-values fit was present. Likewise, Osei-Bryson et al.[18] found in their study of enterprise resource planning systems that implementation climate was significantly associated with implementation effectiveness. It is important to note that the latter two studies measured and analyzed implementation climate at the individual level of analysis rather than the organizational level of analysis at which the implementation construct is formulated. Caution should be exercised in attributing their study results to the organizational level of Klein and Sorra's [1] theory. Doing so could result in drawing erroneous conclusions or, in the language of multi-level organizational research, committing a fallacy of the wrong level [57, 6265].

What is the appropriate level of analysis for implementation climate?

Levels issues arise when incongruence occurs between or among the level of theory, the level of measurement, or the level of statistical analysis [45, 57, 64]. Implementation climate is one of many constructs that are potentially relevant to implementation science that can be conceptualized at an organizational level of theory even though the source of data for the construct resides at the individual level (i.e., the level of measurement). Other constructs that fit this description include leadership, culture, power, participation, and communication.

In proposing constructs where the level of theory and the level of measurement do not match, researchers should specify the composition model or functional relationship that links the lower-level data to the higher-level construct [45, 57, 64, 66, 67]. Several composition models exist [67]. In the case of implementation climate, Klein and Sorra [1] propose a functional relationship of homogeneity--that is, they posit that organizational members share sufficiently similar perceptions of implementation climate that they can be characterized as a whole. Because both implementation climate and implementation effectiveness are formulated as organization-level constructs, an appropriate test of the relationship between these constructs should take place at the organizational level of analysis. Before proceeding with such an analysis, however, it is important to verify that the data conform to the level of the theory--that is, that the functional relationship specified in the composition model holds for the data in question [57, 64]. This means ensuring that sufficient within-group agreement exists to justify aggregating individuals' implementation climate perceptions to the organizational level of analysis.

Implementation scientists can use several measures to verify that sufficient within-group agreement exists, including rwg, eta-squared and two intraclass correlation coefficients, ICC(1) and ICC(2). As Klein and Kozlowski [45] note, each offers a different, yet complementary assessment. Rwg answers the question: how high is within-group agreement on a given variable for a given unit (e.g., organization)? Eta-squared and ICC(1), by comparison, answer the question: to what extent does a measure vary between-units versus within-units? ICC(2) answers the question: how reliable are the unit means within a sample? An extensive literature describes the statistical assumptions, merits, limitations, and interpretative rules of thumb for these measures [45, 66, 6874]. Climate researchers often assess within-group agreement using multiple measures [17, 24, 25, 27, 28, 52, 61, 75, 76]. However, different measures can produce different results depending on the number of units, the number of respondents per unit, and the amount and distribution of missing data between and within units [6874, 77, 78].

The rwg differs from the other three measures discussed here in that it assesses within-group variability for individual units (e.g., organizations). The others compare within-group variability to between-group variability across an entire sample of units. The advantage of the rwg is that it allows researchers to assess the extent to which units vary in the level of within-group agreement in implementation climate perceptions. What, though, should a researcher do with those units for which the rwg does not exceed 0.70, the rule-of-thumb value for justifying aggregation of individual perceptions to the unit-level? Klein et al. argue that such units should be excluded from further analysis because the implementation climate is not present in these units: no shared meaning exists [45, 57]. If the data from these units do not conform to the level of theory, including these units in a statistical analysis of between-group differences can prove misleading. Construct validity issues arise [45, 57, 66]. For example, if one-half of the members of a unit describe the implementation climate as positive and the other one-half describe it as negative, then the average of members' perceptions of implementation climate describes none of the members' views. One could examine whether units with higher within-group agreement in implementation climate perceptions differ from those with lower within-group agreement on outcomes such as variability in organizational members' innovation use. However, such an analysis would represent a shift in the research question under investigation.

How should implementation climate be measured?

Implementation scientists wishing to assess implementation climate face a twofold measurement dilemma: no standard instrument exists for measuring implementation climate, and existing instruments contain items specific to information systems implementation that have questionable relevance for implementation research in health and human services (e.g., access to internet resources, 'help desk' availability). Although existing instruments could be adapted, changes in item content or item wording could reduce the instruments' comparability and alter their psychometric properties. For those interested in developing implementation climate measures, five guidelines follow from the conceptual discussion above (see Appendix 1 for an example of how we are following these guidelines in a study).

First, climate researchers stress that climate measures should be descriptive in content, not evaluative, in order to distinguish climate from related constructs, like attitudes or satisfaction [17, 49, 53]. Survey items should ask organizational members to indicate 'whether relatively objective and neutral descriptions of the work environment are accurate or inaccurate,' rather than asking them to 'rate evaluative (positive or negative) descriptions of their work environment, in light of their own values, experiences, and expectations' [17: p. 6]. Descriptive item examples include: 'Supervisors praise employees for using [innovation] properly,' 'Employees have enough time to do their work and learn new skills associated with [innovation],' and 'Technical assistance is readily available for [innovation].' Evaluative item examples include 'I'm discouraged from using [innovation],' 'I think [innovation] is a waste of time and money for our organization,' and 'I'm satisfied with the technical assistance for [innovation].' While this advice has merit, Klein et al.[17] note that writing purely descriptive items is difficult because, in describing relatively positive or negative policies or practices (e.g., praise, expectation, monitoring), descriptive items take an evaluative tone. They suggest that climate researchers view the descriptive-evaluative distinction as a continuum rather than a dichotomy, yet stay on the descriptive side of the continuum.

Second, theory and research suggest that the wording of survey items can influence not only the variability in a construct, but also the relationship between a construct and outcomes [17, 44]. Specifically, items with group (e.g., organizational) referents rather than individual referents may increase the within-group agreement and between-group variability in climate measures. Glick [49] argues that survey items that direct respondents' attention to their individual experiences (e.g., 'I' or 'my') encourage them to look within and ignore the experiences of others; conversely, items that direct respondents' attention to groups or higher units (collectivities) encourage them to consider the common or shared experience of others. In their study of not-for-profit community service organizations, Baltes et al.[44] found that psychological climate measures that differed only in their referents (individual versus organizational) were not only empirically distinguishable from one another, but each uniquely predicted job satisfaction. Moreover, discrepancies in employees' climate perceptions measured with organizational and individual referents (e.g., differences in employees' perceptions of the 'average' or 'typical' employees' experience versus their own experience) also predicted job satisfaction. The findings, and others [17], suggest that survey items that differ only in referent may in fact assess closely related but nevertheless subtly different constructs. Emphasis should be placed, therefore, on items with group (organizational) rather than individual referents.

Third, researchers should assess implementation climate with items that directly measure the extent to which innovation use is perceived to be expected, supported, and rewarded. This guideline contradicts the current practice of assessing the construct with items that measure perceptions of the availability and adequacy of various implementation policies and practices [14, 16, 18, 19]. Current practice ignores the equifinality of implementation policies and practices. If different mixes of policies and practices can generate equivalent implementation climates, then there is little reason to expect consistent relationships between specific implementation policies and practices and implementation climate. In some organizations, for example, the availability and adequacy of supervisor praise for innovation use could serve as a good indicator (indirect measure) of implementation climate. In other organizations, say those that rely primarily on financial incentives to reward innovation use, the availability or adequacy of supervisor praise would make a poor, or even irrelevant, indicator of implementation climate. A better approach for measuring implementation climate, we suggest, is to develop items that focus directly on perceived expectations, support, and rewards for innovation use. With regard to an open-access scheduling innovation, for example, direct measures could include 'Physicians in this practice are expected to use open-access scheduling,' 'Physicians in this practice have the support they need to use open-access scheduling,' and 'Physicians in this practice are recognized for using open-access scheduling.' What is important in measuring implementation climate in this example is that physicians share the perception that innovation use is expected, supported, and rewarded; less important are the specific policies or practices that generate that perception.

Fourth, as a summary or global perception, implementation climate should be measured as a multi-item scale based on a factor analysis of items that exhibit high internal consistency. In their study of innovation implementation in manufacturing plants, for example, Klein et al.[61] conducted factor analyses and examined the alpha-coefficients among climate items at both the individual level and organizational level before computing an implementation climate scale and subjecting the resulting scale to within-group agreement analysis. Similarly, Holahan et al.[16] found that their 30 implementation climate items demonstrated high internal consistency. Although they did not run a factor analysis, they too computed a mean scale at the individual level before assessing within-group variability and aggregating teachers' climate perceptions to the school level. Neither theory nor research indicates how researchers should proceed if implementation climate items do not cohere into a single scale. Does implementation climate exist if, for example, organizational members perceive that innovation use is expected and supported, but not rewarded? If so, what are the implications of such a climate for implementation effectiveness?

Finally, Klein and Sorra [1] suggest that the 'targeted employees' whose perceptions should be assessed in measuring implementation climate include not only those expected to use an innovation directly (e.g., front-line staff), but also those expected to support an innovation's use by others (e.g., information technology specialists, supervisors). However, researchers conducting empirical studies, including Klein et al.[61], have not included the perceptions of expected supporters in their measurement of implementation climate. We also favor focusing only on the perceptions of expected users because we believe, the perceptions of expected supporters have an indirect effect, as opposed to direct effect, on innovation use. When expected supporters perceive that innovation use is not expected, supported, or rewarded, they are likely to omit or put into place poor-quality implementation policies and practices. Top managers, for example, might withhold resources. Supervisors might send mixed signals. Information technology specialists might provide lackluster technical support. In our view, the actions or non-actions of expected supporters influence innovation use by creating a favorable or unfavorable implementation climate for expected users. It is the implementation climate perceptions of expected users that are more psychologically proximal to, and therefore, like to be more predictive of, the consistency and quality of expected users' innovation use.


Over the last decade, impressive efforts have been made to catalogue the features of innovations, organizations, and environments that influence innovation implementation [79, 80]. While the volume of research on implementation is slim compared to that on adoption, the list of such factors is large and shows no signs of shrinking. These efforts to catalogue facilitators and barriers of implementation are to be applauded, especially if they stimulate the construction of testable theories to explain implementation success, or encourage the development of useful models to guide implementation processes. The challenge for building research evidence in implementation science, however, is that often, perhaps even most of the time, there are multiple ways to achieve the same outcomes. For example, there are at least three ways that organizations can create a good fit between the knowledge and skills of expected users and those demanded for consistent, high-quality use of a technically complex innovation. Organizations can raise expected users' knowledge and skills to the level required by the innovation; lower the innovation's technical complexity to match expected users' current knowledge and skills; or hire, promote, or transfer organizational members who already possess the required level of knowledge and skills. If equifinality is an essential feature of organizations, as it is of most social systems, then efforts to link specific policies and practices to implementation success are likely to produce equivocal results. Sometimes training will be associated with implementation success; sometimes it will not. Researchers could focus on identifying the conditions under which organizations use specific implementation policies and practices, such as training. Alternatively, they could focus on the cumulative impact of implementation policies and practices by examining whether positive implementation climate (regardless of how such a climate is achieved) is associated with implementation success. These options are not mutually exclusive, since they address different, and arguably important, research questions. A focus on implementation climate, however, would facilitate the comparison of implementation effectiveness across organizations that use different mixes of policies and practices to promote consistent, high-quality innovation use.

Ultimately, the value of the implementation climate construct depends on its predictive utility. We conclude, therefore, with some thoughts on how to advance empirical investigation and theoretical inquiry. First, since the construct and the theory in which it figures are pitched at the organizational level, a longitudinal multi-organizational research design provides the best means for assessing the construct's scientific worth. Although sample size and statistical power considerations make it tempting to test the theory at the intra-organizational level, caution should be exercised in using clinics, departments, or organizational divisions as units of analysis. This approach might be defensible if a reasonable case can be made that the clinics, departments, or divisions in question represent distinct (i.e., independent) units of implementation. As noted earlier, though, measuring the construct and testing the theory at the intra-organizational level introduces the risk of committing the fallacy of the wrong level. Pragmatically, implementation climate might not demonstrate enough between-group variability among intra-organizational units to permit the observation of a significant association with implementation effectiveness.

Second, implementation scientists should keep in mind the type of innovation that Klein and Sorra's (1996) theory of implementation effectiveness seeks to predict and explain. Theories, like tools, have a bounded range of application. Given the theory's context of origin--the study of information systems and technology implementation in manufacturing settings--the construct of implementation climate is perhaps most useful for studying complex innovations in health and human service delivery. By complex, we mean innovations that require collective, coordinated behavior change by many organizational members in order to successfully implement them and realize some or all of the anticipated benefits of innovation use. Put differently, implementation climate is likely to prove useful in studying innovations that exhibit moderate to high levels of task interdependence and outcome interdependence. Conversely, implementation climate is not likely to prove useful in studying innovations that individual health and human service providers can adopt, implement, and use on their own with relatively modest training and support and for which they and their patients or clients can realize anticipated benefits regardless of what other providers do. For such innovations, individual or interpersonal theories of behavior change may offer more explanatory power than organization theories of innovation implementation.

Third, good measurement practice, particularly in the development of new measures, is essential for building scientific knowledge. The measurement guidelines offered above could promote consistency across studies. Yet, implementation scientists might still find it challenging to develop measures of implementation climate that are sufficiently tailored to make them predictive in specific innovation-implementation contexts, yet not so tailored that they could not be used in other innovation-implementation contexts without substantial modification. The construction of instruments that directly measure implementation climate perceptions could mitigate this tension, but it cannot eliminate it entirely. If no single instrument will meet implementation scientists' needs, then perhaps the field of self-efficacy research offers a useful model. Health behavior scientists have developed self-efficacy instruments for smoking, physical activity, and other health behaviors that are reliable and valid within their domain of application [8188]. Although item content is tailored, the instruments are based on theory and have enough features in common that scholars can accumulate scientific knowledge across health problems.

Finally, implementation scientists should continue to develop the implementation climate construct. Several questions merit further theoretical, and empirical, attention. Is it useful, for example, to distinguish implementation climate strength from implementation climate level? Do some implementation policies and practices--or, for that matter, some broader features of organizational context--influence the strength of implementation climate but not the level of implementation climate? Likewise, are the three aspects of implementation climate (i.e., expected, supported, and rewarded) equally important? Does their relative importance depend on the implementation context and, if so, how? Lastly, is implementation climate a theoretically meaningful construct at the individual level? If so, how does an individual-level analogue relate to the organization-level construct or to other important constructs in implementation science?

Appendix 1

Implementation climate and organizational performance in the Community Clinical Oncology Program

In a current study, we are examining the association of implementation climate, innovation values fit, and organizational performance in the Community Clinical Oncology Program (CCOP). Established in 1983, the CCOP is a three-way partnership between the NCI's Division of Cancer Prevention (NCI/DCP), selected cancer centers and clinical cooperative groups ('CCOP research bases'), and community-based networks of hospitals and physicians ('CCOP organizations') to conduct Phase III clinical trials [89, 90]. In this partnership, NCI/DCP provides overall direction and funding; CCOP research bases design clinical trials; and CCOP organizations assist with patient accruals, data collection, and dissemination of study findings. As of December 2010, 47 CCOP organizations located in 28 states, the District of Columbia, and Puerto Rico participated in NCI-sponsored clinical trials. The CCOP includes 400 hospitals and more than 3,520 community physicians. In FY 2010, the CCOP budget totaled $93.6 million. The median CCOP organization award was $850,000.

CCOP organizations are led by a physician principal investigator who provides local program leadership. CCOP staff members include a program coordinator, research nurses or clinical research associates, data managers, and regulatory specialists. These staff members coordinate the selection of new clinical trial protocols for CCOP participation, disseminate protocol updates to the participating physicians, and collect and submit study data [15, 90, 91]. CCOP-affiliated physicians accrue or refer participants to clinical trials, and typically include medical, surgical and radiation oncologists, general surgeons, urologists, gastroenterologists, and primary care physicians. Through their membership in CCOP research bases, CCOP-affiliated physicians also participate in the development of clinical trials by proposing study ideas, providing input on study design, and, occasionally, serving as principal investigator for a clinical trial [15, 90, 91].

In the fall of 2011, we will survey a stratified random sample of 900 CCOP-affiliated physicians to obtain data on their perceptions of implementation climate, innovation-values fit, and other constructs. We will measure implementation climate with six items referenced to the respondent's CCOP organization:

  1. 1.

    Physicians are expected to enroll a certain number of patients in NCI-sponsored clinical trials.

  2. 2.

    Physicians are expected to help the CCOP meet its patient enrollment goals in NCI-sponsored clinical trials.

  3. 3.

    Physicians get the research support they need to identify potentially eligible patients for NCI-sponsored clinical trials.

  4. 4.

    Physicians get the research support they need to enroll patients in NCI-sponsored clinical trials (e.g., consenting patients).

  5. 5.

    Physicians receive recognition for enrolling patients in NCI-sponsored clinical trials.

  6. 6.

    Physicians receive appreciation for enrolling patients in NCI-sponsored clinical trials.

Respondents will use a five-point scale to indicate whether they disagree, somewhat disagree, neither agree nor disagree, somewhat agree, or agree with each statement.

Our measurement approach is consistent with the measurement guidelines described in this paper. Specifically, the items are: descriptive versus evaluative in focus; group-referenced rather than individually referenced; direct measures of climate perceptions rather than indirect measures of specific implementation policies and practices; multiple in number for the three dimensions of implementation climate (i.e., expected, supported and expected); and targeted toward respondents who are expected to use the innovation directly (i.e., physicians).

Like Klein and Sorra's (1996) theory, our conceptual model emphasizes organization-level constructs. Therefore, we will conduct statistical tests to assess the extent to which responses to individual-level scales constructed from factor analysis show sufficient within-CCOP agreement to justify aggregation to the CCOP organization level. Specifically, we will compute eta-squared, ICC(1), ICC(2), and rwg. We will compare the values of these statistics to recommended cut-off values and values reported in other studies using individual-level variables aggregated to the organizational level [31, 49]. If on balance the statistical tests justify data aggregation, we will construct CCOP-organization-level averages for implementation climate, innovation-values fit, and other organization-level constructs for which data are obtained at the individual level of measurement. Using regression analysis, we will examine the association of these variables with CCOP organizational performance, measured as number of patients enrolled in treatment trials by the CCOP organization. If the statistical tests do not justify aggregation, we will revise our hypotheses to focus on implementation climate strength and incorporate in our statistical models variables that measure intra-CCOP variability of individual responses (e.g., coefficient of variation).


  1. 1.

    Klein KJ, Sorra JS: The challenge of innovation implementation. Academy of Management Review. 1996, 21: 1055-1080.

    Google Scholar 

  2. 2.

    Aarons GA: Measuring provider attitudes toward evidence-based practice: Consideration of organizational context and individual differences. Child and Adolescent Psychiatric Clinics of North America. 2005, 14: 255-+-

    Article  PubMed  PubMed Central  Google Scholar 

  3. 3.

    Aarons GA, Fettes DL, Flores LE, Sornmerfeld DH: Evidence-based practice implementation and staff emotional exhaustion in children's services. Behaviour Research and Therapy. 2009, 47: 954-960. 10.1016/j.brat.2009.07.006.

    Article  PubMed  PubMed Central  Google Scholar 

  4. 4.

    Aarons GA, Palinkas LA: Implementation of evidence-based practice in child welfare: Service provider perspectives. Administration and Policy in Mental Health and Mental Health Services Research. 2007, 34: 411-419. 10.1007/s10488-007-0121-3.

    Article  PubMed  Google Scholar 

  5. 5.

    Chorpita BF, Regan J: Dissemination of effective mental health treatment procedures: Maximizing the return on a significant investment. Behaviour Research and Therapy. 2009, 47: 990-993. 10.1016/j.brat.2009.07.002.

    Article  PubMed  Google Scholar 

  6. 6.

    Oser C, Knudsen H, Staton-Tindall M, Leukefeld C: The adoption of wraparound services among substance abuse treatment organizations serving criminal offenders: The role of a women-specific program. Drug and Alcohol Dependence. 2009, 103: S82-S90.

    Article  PubMed  PubMed Central  Google Scholar 

  7. 7.

    Oser CB, Knudsen HK, Staton-Tindall M, Taxman F, Leukefeld C: Organizational-level correlates of the provision of detoxification services and medication-based treatments for substance abuse in correctional institutions. Drug and Alcohol Dependence. 2009, 103: S73-S81.

    Article  PubMed  Google Scholar 

  8. 8.

    Smith BD, Mogro-Wilson C: Multi-level influences on the practice of inter-agency collaboration in child welfare and substance abuse treatment. Children and Youth Services Review. 2007, 29: 545-556. 10.1016/j.childyouth.2006.06.002.

    Article  Google Scholar 

  9. 9.

    Smith BD, Mogro-Wilson C: Inter-agency collaboration: Policy and practice in child welfare and substance abuse treatment. Administration in Social Work. 2008, 32: 5-24.

    Article  Google Scholar 

  10. 10.

    Brennan P, Claber O, Shaw T: The Teesside Cancer Family History Service: change management and innovation at cancer network level. Familial Cancer. 2007, 6: 181-187. 10.1007/s10689-007-9125-0.

    Article  PubMed  Google Scholar 

  11. 11.

    Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC: Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implementation Science. 2009, 4:

    Google Scholar 

  12. 12.

    Allen NE, Lehrner A, Mattison E, Miles T, Russell A: Promoting systems change in the health care response to domestic violence. Journal of Community Psychology. 2007, 35: 103-120. 10.1002/jcop.20137.

    Article  Google Scholar 

  13. 13.

    Ruppel CP, Harrington SJ: Sharing knowledge through intranets: A study of organizational culture and intranet implementation. Ieee Transactions on Professional Communication. 2001, 44: 37-52. 10.1109/47.911131.

    Article  Google Scholar 

  14. 14.

    Dong LY, Neufeld DJ, Higgins C: Testing Klein and Sorra's innovation implementation model: An empirical examination. Journal of Engineering and Technology Management. 2008, 25: 237-255. 10.1016/j.jengtecman.2008.10.006.

    Article  Google Scholar 

  15. 15.

    Helfrich CD, Weiner BJ, McKinney MM, Minasian L: Determinants of implementation effectiveness - Adapting a framework for complex innovations. Medical Care Research and Review. 2007, 64: 279-303. 10.1177/1077558707299887.

    Article  PubMed  Google Scholar 

  16. 16.

    Holahan PJ, Aronson ZH, Jurkat MP, Schoorman FD: Implementing computer technology: a multiorganizational test of Klein and Sorra's model. Journal of Engineering and Technology Management. 2004, 21: 31-50. 10.1016/j.jengtecman.2003.12.003.

    Article  Google Scholar 

  17. 17.

    Klein KJ, Conn AB, Smith DB, Sorra JS: Is everyone in agreement? An exploration of within-group agreement in employee perceptions of the work environment. Journal of Applied Psychology. 2001, 86: 3-16.

    CAS  Article  PubMed  Google Scholar 

  18. 18.

    Osei-Bryson KM, Dong LY, Ngwenyama O: Exploring managerial factors affecting ERP implementation: an investigation of the Klein-Sorra model using regression splines. Information Systems Journal. 2008, 18: 499-527. 10.1111/j.1365-2575.2008.00309.x.

    Article  Google Scholar 

  19. 19.

    Pullig C, Maxham JG, Hair JF: Salesforce automation systems - An exploratory examination of organizational factors associated with effective implementation and sales force productivity. Journal of Business Research. 2002, 55: 401-415. 10.1016/S0148-2963(00)00159-4.

    Article  Google Scholar 

  20. 20.

    Schneider B: Organizational climates - essay. Personnel Psychology. 1975, 28: 447-479. 10.1111/j.1744-6570.1975.tb01386.x.

    Article  Google Scholar 

  21. 21.

    Schneider B: The climate for service: an application of the climate construct. Organizational climate and culture. Edited by: Schneider B. 1990, San Francisco: Jossey-Bass, 383-412. 1

    Google Scholar 

  22. 22.

    Schneider B, Bowen DE: Employee and customer perceptions of service in banks - replication and extension. Journal of Applied Psychology. 1985, 70: 423-433.

    Article  Google Scholar 

  23. 23.

    Schneider B, Parkington JJ, Buxton VM: Employee and customer perceptions of service in banks. Administrative Science Quarterly. 1980, 25: 252-267. 10.2307/2392454.

    Article  Google Scholar 

  24. 24.

    Schneider B, Salvaggio AN, Subirats M: Climate strength: A new direction for climate research. Journal of Applied Psychology. 2002, 87: 220-229.

    Article  PubMed  Google Scholar 

  25. 25.

    Schneider B, White SS, Paul MC: Linking service climate and customer perceptions of service quality: Test of a causal model. Journal of Applied Psychology. 1998, 83: 150-163.

    CAS  Article  PubMed  Google Scholar 

  26. 26.

    Zohar D: Safety climate in industrial-organizations - theoretical and applied implications. Journal of Applied Psychology. 1980, 65: 96-102.

    CAS  Article  PubMed  Google Scholar 

  27. 27.

    Zohar D: A group-level model of safety climate: Testing the effect of group climate on microaccidents in manufacturing jobs. Journal of Applied Psychology. 2000, 85: 587-596.

    CAS  Article  PubMed  Google Scholar 

  28. 28.

    Zohar D: The effects of leadership dimensions, safety climate, and assigned priorities on minor injuries in work groups. Journal of Organizational Behavior. 2002, 23: 75-92. 10.1002/job.130.

    Article  Google Scholar 

  29. 29.

    Zohar D, Luria G: Climate as a social-cognitive construction of supervisory safety practices: Scripts as proxy of behavior patterns. Journal of Applied Psychology. 2004, 89: 322-333.

    Article  PubMed  Google Scholar 

  30. 30.

    Hughes LC, Chang Y, Mark BA: Quality and strength of patient safety climate on medical-surgical units. Health Care Management Review. 2009, 34: 19-28.

    Article  PubMed  Google Scholar 

  31. 31.

    Zohar D: Safety climate in industrial organizations -- theoretical and applied implications. Journal of Applied Psychology. 1980, 65: 96-102.

    CAS  Article  PubMed  Google Scholar 

  32. 32.

    Zohar D, Livne Y, Tenne-Gazit O, Admi H, Donchin Y: Healthcare climate: A framework for measuring and improving patient safety. Critical Care Medicine. 2007, 35: 1312-1317. 10.1097/01.CCM.0000262404.10203.C9.

    Article  PubMed  Google Scholar 

  33. 33.

    Zohar D: Thirty years of safety climate research: Reflections and future directions. Accident Analysis and Prevention. 2010, 42: 1517-1522. 10.1016/j.aap.2009.12.019.

    Article  PubMed  Google Scholar 

  34. 34.

    Hunter ST, Bedell KE, Mumford MD: Climate for creativity: A quantitative review. Creativity Res J. 2007, 19: 69-90. 10.1080/10400410709336883.

    Article  Google Scholar 

  35. 35.

    Amabile TM, Conti R, Coon H, Lazenby J, Herron M: Assessing the work environment for creativity. Acad Manage J. 1996, 39: 1154-1184.

    Article  Google Scholar 

  36. 36.

    Ekvall G, Ryhammar L: The creative climate: Its determinants and effects at a Swedish university. Creativity Res J. 1999, 12: 303-310. 10.1207/s15326934crj1204_8.

    Article  Google Scholar 

  37. 37.

    Isaksen SG, Lauer KJ, Ekvall G: Situational outlook questionnaire: A measure of the climate for creativity and change. Psychological Reports. 1999, 85: 665-674. 10.2466/PR0.85.6.665-674.

    Article  Google Scholar 

  38. 38.

    Mathisen GE, Einarsen S: A review of instruments assessing creative and innovative environments within organizations. Creativity Res J. 2004, 16: 119-140. 10.1207/s15326934crj1601_12.

    Article  Google Scholar 

  39. 39.

    Colquitt JA, Noe RA, Jackson CL: Justice in teams: Antecedents and consequences of procedural justice climate. Personnel Psychology. 2002, 55: 83-109. 10.1111/j.1744-6570.2002.tb00104.x.

    Article  Google Scholar 

  40. 40.

    Naumann SE, Bennett N: A case for procedural justice climate: Development and test of a multilevel model. Acad Manage J. 2000, 43: 881-889.

    Article  Google Scholar 

  41. 41.

    Akgun AE, Keskin H, Byrne JC: Procedural Justice Climate in New Product Development Teams: Antecedents and Consequences. Journal of Product Innovation Management. 2010, 27: 1096-1111. 10.1111/j.1540-5885.2010.00773.x.

    Article  Google Scholar 

  42. 42.

    Lin SP, Tang TW, Li CH, Wu CM, Lin HH: Mediating effect of cooperative norm in predicting organizational citizenship behaviors from procedural justice climate. Psychological Reports. 2007, 101: 67-78.

    PubMed  Google Scholar 

  43. 43.

    Mayer D, Nishii L, Schneider B, Goldstein H: The precursors and products of justice climates: Group leader antecedents and employee attitudinal consequences. Personnel Psychology. 2007, 60: 929-963. 10.1111/j.1744-6570.2007.00096.x.

    Article  Google Scholar 

  44. 44.

    Baltes BB, Zhdanova LS, Parker CP: Psychological climate: A comparison of organizational and individual level referents. Human Relations. 2009, 62: 669-700. 10.1177/0018726709103454.

    Article  Google Scholar 

  45. 45.

    Klein KJ, Kozlowski SWJ: From micro to meso: critical steps in conceptualizing and conducting multilevel research. Organizational Research Methods. 2000, 3: 211-236. 10.1177/109442810033001.

    Article  Google Scholar 

  46. 46.

    Guion RM: Note on organizational climate. Organ Behav Hum Perf. 1973, 9: 120-125. 10.1016/0030-5073(73)90041-X.

    Article  Google Scholar 

  47. 47.

    James LR, Jones AP: Organizational climate - review of theory and research. Psychological Bulletin. 1974, 81: 1096-1112.

    Article  Google Scholar 

  48. 48.

    Jones AP, James LR: Psychological Climate - Dimensions and Relationships of Individual and Aggregated Work-Environment Perceptions. Organ Behav Hum Perf. 1979, 23: 201-250. 10.1016/0030-5073(79)90056-4.

    Article  Google Scholar 

  49. 49.

    Glick WH: Conceptualizing and measuring organizational and psychological climate - pitfalls in multilevel research. Academy of Management Review. 1985, 10: 601-616.

    Google Scholar 

  50. 50.

    Reichers AE, Schneider B: Climate and culture: an evolution of constructs. Organizational climate and culture. Edited by: Schneider B. 1990, San Francisco: Jossey-Bass, 5-39. 1

    Google Scholar 

  51. 51.

    Litwin GH, Stringer RA: Motivation and organizational climate. 1968, Boston,: Division of Research, Graduate School of Business Administration, Harvard University

    Google Scholar 

  52. 52.

    Patterson MG, West MA, Shackleton VJ, Dawson JF, Lawthom R, Maitlis S, Robinson DL, Wallace AM: Validating the organizational climate measure: links to managerial practices, productivity and innovation. Journal of Organizational Behavior. 2005, 26: 379-408. 10.1002/job.312.

    Article  Google Scholar 

  53. 53.

    Hellrieg D, Slocum JW: Organizational Climate - Measures, Research and Contingencies. Acad Manage J. 1974, 17: 255-280.

    Article  Google Scholar 

  54. 54.

    Gonzalez-Roma V, Peiro JM, Tordera N: An examination of the antecedents and moderator influences of climate strength. Journal of Applied Psychology. 2002, 87: 465-473.

    Article  PubMed  Google Scholar 

  55. 55.

    Kozlowski SWJ, Doherty ML: Integration of climate and leadership - examination of a neglected issue. Journal of Applied Psychology. 1989, 74: 546-553.

    Article  Google Scholar 

  56. 56.

    Luria G: Climate strength - How leaders form consensus. Leadership Quarterly. 2008, 19: 42-53. 10.1016/j.leaqua.2007.12.004.

    Article  Google Scholar 

  57. 57.

    Klein KJ, Dansereau F, Hall RJ: Levels issues in theory development, data collection, and analysis. Academy of Management Review. 1994, 19: 195-229.

    Google Scholar 

  58. 58.

    Schneider B, Reichers AE: On the Etiology of Climates. Personnel Psychology. 1983, 36: 19-39. 10.1111/j.1744-6570.1983.tb00500.x.

    Article  Google Scholar 

  59. 59.

    Gonzalez-Roma V, Fortes-Ferreira L, Peiro JM: Team climate, climate strength and team performance. A longitudinal study. Journal of Occupational and Organizational Psychology. 2009, 82: 511-536. 10.1348/096317908X370025.

    Article  Google Scholar 

  60. 60.

    Mischel W: Toward a cognitive social learning reconceptualization of personality. Psychological Review. 1973, 80: 252-283.

    CAS  Article  PubMed  Google Scholar 

  61. 61.

    Klein KJ, Conn AB, Sorra JS: Implementing computerized technology: An organizational analysis. Journal of Applied Psychology. 2001, 86: 811-824.

    CAS  Article  PubMed  Google Scholar 

  62. 62.

    Glick WH: Organizations are not central tendencies - shadowboxing in the dark, round 2 - response. Academy of Management Review. 1988, 13: 133-137.

    Google Scholar 

  63. 63.

    Glick WH, Roberts KH: Hypothesized interdependence, assumed independence. Academy of Management Review. 1984, 9: 722-735.

    Google Scholar 

  64. 64.

    Rousseau D: Issues of level in organizational research: multilevel and cross-level perspectives. Research in organizational behavior. Edited by: Cummings LL, Staw BM. 1985, Greenwich, Conn.,: JAI Press, 7: 1-37.

    Google Scholar 

  65. 65.

    Dansereau F, Cho J, Yammarino FJ: Avoiding the "fallacy of the wrong level". Group & Organization Management. 2006, 31: 536-577. 10.1177/1059601106291131.

    Article  Google Scholar 

  66. 66.

    James LR: Aggregation bias in estimates of perceptual agreement. Journal of Applied Psychology. 1982, 67: 219-229.

    Article  Google Scholar 

  67. 67.

    Chan D: Functional relations among constructs in the same content domain at different levels of analysis: A typology of composition models. Journal of Applied Psychology. 1998, 83: 234-246.

    Article  Google Scholar 

  68. 68.

    Bliese PD, Halverson RR: Group size and measures of group-level properties: An examination of eta-squared and ICC values. Journal of Management. 1998, 24: 157-172.

    Article  Google Scholar 

  69. 69.

    Brown RD, Hauenstein NMA: Interrater agreement reconsidered: An alternative to the r(wg) indices. Organizational Research Methods. 2005, 8: 165-184. 10.1177/1094428105275376.

    Article  Google Scholar 

  70. 70.

    Cohen A, Doveh E, Eick U: Statistical properties of the r(WG(J)) index of agreement. Psychological Methods. 2001, 6: 297-310.

    CAS  Article  PubMed  Google Scholar 

  71. 71.

    Cohen A, Doveh E, Nahum-Shani I: Testing Agreement for Multi-Item Scales With the Indices rWG(J) and ADM(J). Organizational Research Methods. 2009, 12: 148-164.

    Article  Google Scholar 

  72. 72.

    James LR, Demaree RG, Wolf G: Estimating within-group interrater reliability with and without response bias. Journal of Applied Psychology. 1984, 69: 85-98.

    Article  Google Scholar 

  73. 73.

    LeBreton JM, James LR, Lindell MK: Recent issues regarding r(WG), r*(WG), r(WG)(J), and r*(WG)(J). Organizational Research Methods. 2005, 8: 128-138. 10.1177/1094428104272181.

    Article  Google Scholar 

  74. 74.

    Newman DA, Sin HP: How Do Missing Data Bias Estimates of Within-Group Agreement? Sensitivity of SDWG, CVWG, rWG( J), rWG( J)*, and ICC to Systematic Nonresponse. Organizational Research Methods. 2009, 12: 113-147.

    Article  Google Scholar 

  75. 75.

    Glisson C, Landsverk J, Schoenwald S, Kelleher K, Hoagwood KE, Mayberg S, Green P, Res Network Youth Mental H: Assessing the Organizational Social Context (OSC) of mental health services: Implications for research and practice. Administration and Policy in Mental Health and Mental Health Services Research. 2008, 35: 98-113. 10.1007/s10488-007-0148-5.

    Article  PubMed  Google Scholar 

  76. 76.

    Glisson C, Schoenwald SK, Kelleher K, Landsverk J, Hoagwood KE, Mayberg S, Green P: Therapist turnover and new program sustainability in mental health clinics as a function of organizational culture, climate, and service structure. Adm Policy Ment Health. 2008, 35: 124-133. 10.1007/s10488-007-0152-9.

    Article  PubMed  Google Scholar 

  77. 77.

    Dansereau F, Cho J, Yammarino FJ: Avoiding the "Fallacy of the wrong level" - A within and between analysis (WABA) approach. Group & Organization Management. 2006, 31: 536-577. 10.1177/1059601106291131.

    Article  Google Scholar 

  78. 78.

    van Mierlo H, Vermunt JK, Rutte CG: Composing Group-Level Constructs From Individual-Level Survey Data. Organizational Research Methods. 2009, 12: 368-392.

    Article  Google Scholar 

  79. 79.

    Damschroder LJ, Aron DC, Keith RE, Kirsh SR, Alexander JA, Lowery JC: Fostering implementation of health services research findings into practice: a consolidated framework for advancing implementation science. Implement Sci. 2009, 4: 50-10.1186/1748-5908-4-50.

    Article  PubMed  PubMed Central  Google Scholar 

  80. 80.

    Greenhalgh T: Diffusion of innovations in health service organisations: a systematic literature review. 2005, Malden, Mass.: BMJ Books/Blackwell Pub.

    Book  Google Scholar 

  81. 81.

    Dishman RK, Motl RW, Saunders R, Felton G, Ward DS, Dowda M, Pate RR: Self-efficacy partially mediates the effect of a school-based physical-activity intervention among adolescent girls. Preventive Medicine. 2004, 38: 628-636. 10.1016/j.ypmed.2003.12.007.

    Article  PubMed  Google Scholar 

  82. 82.

    Leung DYP, Chan SSC, Lau CP, Wong V, Lam TH: An evaluation of the psychometric properties of the Smoking Self-Efficacy Questionnaire (SEQ-12) among Chinese cardiac patients who smoke. Nicotine & Tobacco Research. 2008, 10: 1311-1318. 10.1080/14622200802238928.

    Article  Google Scholar 

  83. 83.

    Dishman RK, Motl RW, Saunders RP, Dowda M, Felton G, Ward DS, Pate RR: Factorial invariance and latent mean structure of questionnaires measuring social-cognitive determinants of physical activity among black and white adolescent girls. Preventive Medicine. 2002, 34: 100-108. 10.1006/pmed.2001.0959.

    CAS  Article  PubMed  Google Scholar 

  84. 84.

    Finkelstein J, Lapshin O, Cha E: Feasibility of Promoting Smoking Cessation Among Methadone Users Using Multimedia Computer-Assisted Education. Journal of Medical Internet Research. 2008, 10:

    Google Scholar 

  85. 85.

    Etter JF, Bergman MM, Humair JP, Perneger TV: Development and validation of a scale measuring self-efficacy of current and former smokers. Addiction. 2000, 95: 901-913. 10.1046/j.1360-0443.2000.9569017.x.

    CAS  Article  PubMed  Google Scholar 

  86. 86.

    Robinson-Smith G, Johnston MV, Allen J: Self-care self-efficacy, quality of life, and depression after stroke. Archives of Physical Medicine and Rehabilitation. 2000, 81: 460-464. 10.1053/mr.2000.3863.

    CAS  Article  PubMed  Google Scholar 

  87. 87.

    Lev EL, Owen SV: A measure of self-care self-efficacy. Research in Nursing & Health. 1996, 19: 421-429. 10.1002/(SICI)1098-240X(199610)19:5<421::AID-NUR6>3.0.CO;2-S.

    CAS  Article  Google Scholar 

  88. 88.

    van der Ven NCW, Weinger K, Yi J, Pouwer F, Ader H, van der Ploeg HM, Snoek FJ: The confidence in diabetes self-care scale - Psychometric properties of a new measure of diabetes-specific self-efficacy in Dutch and US patients with type 1 diabetes. Diabetes Care. 2003, 26: 713-718. 10.2337/diacare.26.3.713.

    Article  PubMed  PubMed Central  Google Scholar 

  89. 89.

    Carpenter WR, Weiner BJ, Kaluzny AD, Domino ME, Lee SY: The effects of managed care and competition on community-based clinical research. Med Care. 2006, 44: 671-679. 10.1097/01.mlr.0000220269.65196.72.

    Article  PubMed  Google Scholar 

  90. 90.

    Minasian LM, Carpenter WR, Weiner BJ, Anderson DE, McCaskill-Stevens W, Nelson S, Whitman C, Kelaghan J, O'Mara AM, Kaluzny AD: Translating research into evidence-based practice: the National Cancer Institute Community Clinical Oncology Program. Cancer. 2010, 116: 4440-4449. 10.1002/cncr.25248.

    Article  PubMed  PubMed Central  Google Scholar 

  91. 91.

    Weiner BJ, McKinney MM, Carpenter WR: Adapting clinical trials networks to promote cancer prevention and control research. Cancer. 2006, 106: 180-187. 10.1002/cncr.21548.

    Article  PubMed  Google Scholar 

Download references


This work was supported by funding from the National Cancer Institute (1 R01 CA124402). The author would like to thank Megan Lewis for her thoughtful comments and suggestions.

Author information



Corresponding author

Correspondence to Bryan J Weiner.

Additional information

Competing interests

The authors declare that they have no competing interests.

Authors' contributions

BJW conceived the idea for the manuscript and took the lead in drafting it. MB, DB, and MJ conducted the background research that informed the manuscript, contributed ideas about the meaning of the construct, made editorial and substantive changes to manuscript drafts. All authors read and approved the final manuscript.

Bryan J Weiner, Charles M Belden, Dawn M Bergmire and Matthew Johnston contributed equally to this work.

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Rights and permissions

This article is published under license to BioMed Central Ltd. This is an Open Access article distributed under the terms of the Creative Commons Attribution License (, which permits unrestricted use, distribution, and reproduction in any medium, provided the original work is properly cited.

Reprints and Permissions

About this article

Cite this article

Weiner, B.J., Belden, C.M., Bergmire, D.M. et al. The meaning and measurement of implementation climate. Implementation Sci 6, 78 (2011).

Download citation

  • Received:

  • Accepted:

  • Published:

  • DOI:


  • Implementation Policy
  • Organizational Member
  • Shared Perception
  • Climate Perception
  • Innovation Implementation