Background

The context of managing the knowledge base for healthcare is complex. Healthcare organizations are composed of multi-level and multi-site interlacing networks that, despite central command and control structures, have strong front-line local micro-systems involved in interpreting policy direction [1]. The nature of healthcare knowledge is characterized by proliferation of information, fragmentation, distribution, and high context dependency. Healthcare practice requires coordinated action in uncertain, rapidly changing situations, with the potential for high failure costs [2]. The public sector context includes the influence of externally imposed performance targets and multiple stakeholder influences and values, the imperative to share good practice across organisational boundaries, and a complex and diverse set of boundaries and networks [3]. Having strong mechanisms and processes for transferring information, developing shared meanings, and the political negotiation of action [4, 5] are therefore crucially important in public sector/healthcare settings, but it is not surprising that there are reports of problems in the organizational capacity of the public sector to effectively manage best practice innovation [611], particularly around issues of power and politics between different professional groups [1217].

The development of capacity to implement evidence-based innovations is a central concept in UK government programmes in healthcare [18]. Strategies to improve evidence-based decision making in healthcare have only recently shifted emphasis away from innovation as a linear and technical process dominated by psychological and cognitive theories of individual behaviour change [19], toward organisational level interventions [20], with attention shifting toward the development of inter-organisational clinical, learning, and research networks for sharing knowledge and innovation [2123], and attempts to improve capacity for innovation within the public sector [24].

Organisational capacity refers to the organisation's ability to take effective action, in this context for the purpose of continually renewing and improving its healthcare practices. Absorptive and receptive capacities are theorized as important antecedents to innovation in healthcare [25]. Broadly, the concept of absorptive capacity is the organization's ability to recognise the value of new external knowledge and to assimilate it, while receptive capacity is the ability to facilitate the transfer and use of new knowledge [2631]. Empirical studies have identified some general antecedent conditions [3234], and have tested application of the concept of absorptive capacity to healthcare [35, 36], although receptive capacities are less well studied. Empirically supported features of organisational context that impact on absorptive and receptive capacities in healthcare include processes for identifying, interpreting, and sharing new knowledge; a learning organisation culture; network structures; strong leadership, vision, and management; and supportive technologies [25].

Public sector benchmarking is widely promoted as a tool for enhancing organisational capacity via a process of collaborative learning [37]. Benchmarking requires the collation and construction of best practice indicators for institutional audit and comparison. Tools are available to measure the organizational context for evidence-based healthcare practice [3841], and components of evidence-based practice (EBP) including implementation of organisational change [4245], research utilization (RU) [46], or research activity (RA) [47]. While organisational learning (OL) and knowledge management (KM) frameworks are increasingly being claimed in empirical studies in healthcare [4853], current approaches to assessing organisational capacity are more likely to be underpinned by diffusion of innovation or change management frameworks [54].

Nicolini and colleagues [2] draw attention to the similarity between the KM literature and the discourse on supporting knowledge translation and transfer in healthcare [5557], as well as between concepts of OL and the emphasis on collective reflection on practice in the UK National Health Service [58, 59], but suggest that 'ecological segregation' between these disciplines and literatures means that cross-fertilisation has not occurred to any great extent. OL and KM literatures could be fruitful sources for improving our understanding of dimensions of organizational absorptive and receptive capacity in healthcare. We therefore aimed to support the development of a metric to audit the organizational conditions for effective evidence-based change by consulting the wider OL and knowledge literatures, where the development of metrics is also identified as a major research priority [60], including the use of existing tools in healthcare [2].

Definitions of KM vary, but many include the core processes of creation or development of knowledge, its movement, transfer, or flow through the organisation, and its application or use for performance improvement or innovation [61]. Early models of KM focused on the measurement of knowledge assets and intellectual capital, with later models focusing on processes of managing knowledge in organisations, split into models where technical-rationality and information technology solutions were central and academic models focusing on human factors and transactional processes [62]. The more emergent view is of the organisation as 'milieu' or community of practice, where the focus on explanatory variables shifts away from technology towards the level of interactions between individuals, and the potential for collective learning. However, technical models and solutions are also still quite dominant in healthcare [63].

Easterby-Smith and Lyles [64] consider KM to focus on the content of the knowledge that an organisation acquires, creates, processes, and uses, and OL to focus on the process of learning from new knowledge. Nutley, Davies and Walker [54] define OL as the way organisations build and organise knowledge and routines and use the broad skills of their workforce to improve organisational performance. Early models of OL focused on cognitive-behavioural processes of learning at individual, group, and organisational levels [6567], and the movement of information in social or activity systems [68]. More recent practice-based theories see knowledge as embedded in culture, practice, and process, conceptualising knowing and learning as dynamic, emergent social accomplishment [6972]. Organisational knowledge is also seen as fragmented into specialised and localised communities of practice, 'distributed knowledge systems' [73], or networks with different interpretive frameworks [74], where competing conceptions of what constitutes legitimate knowledge can occur [75], making knowledge sharing across professional and organization boundaries problematic.

While the two perspectives of KM and OL have very different origins, Scarbrough and Swan [76] suggest that differences are mainly due to disciplinary homes and source perspectives, rather than conceptual distinctiveness. More recently, there have been calls for cognitive and practice-based theories to be integrated in explanatory theories of how practices are constituted, and the practicalities of how socially shared knowledge operates [77, 78]. Similarly, there have been calls for integrative conceptual frameworks for OL and knowledge[79, 80], with learning increasingly defined in terms of knowledge processes [81, 82].

Practice models have their limitations, particularly in relation to weaknesses in explaining how knowledge is contested and legitimated [83]. In a policy context that requires clinical decisions to be based on proof from externally generated research evidence, a comprehensive model for healthcare KM would need to reflect the importance of processes to verify and legitimate knowledge. Research knowledge then needs to be integrated with knowledge achieved from shared interpretation and meaning within the specific social, political, and cultural context of practice, and with the personal values-based knowledge of both the individual professional and the patient [84]. Much public sector innovation also originates from practice and practitioners, as well as external scientific knowledge [85, 86]. New understandings generated from practice then require re-externalising into explicit and shared formal statements and procedures, so that actions can be defended in a public system of accountability.

Our own preference is for a perspective where multiple forms of knowledge are recognised, and where emphasis is placed on processes of validating and warranting knowledge claims. Attention needs also be directed towards the interrelationship between organisational structures of knowledge governance, such as leadership, incentive and reward structures, or the allocation of authority and decision rights, and the conditions for individual agency [8789]. Our own focus is therefore on identifying the organizational conditions that are perceived to support or hinder organizational absorptive or receptive capacities, as a basis for practical action by individuals.

The indicators for supportive organisational conditions are to be developed by extracting items from existing tools, as in previous tools developed to measure OL capability [90]. Existing tools are used because indicators are already empirically supported, operationalised, and easily identified and compared, and because our primary focus is one of utility for practice [91], by specifying 'the different behavioural and organisational conditions under which knowledge can be managed effectively' [92p ix]. Measurement tools that were based on reviews of the literature in the respective fields of KM and learning organisations were chosen as comparison sources to assess the comprehensiveness of the current tools in healthcare, and to improve the delineation of the social and human aspects of EBP in healthcare. If this preliminary stage proves fruitful in highlighting the utility of widening the pool for benchmark items, future work aims to compare the source literatures for confirming empirical evidence, with further work to test the validity and reliability of the benchmark items.

Methods

A structured literature review was undertaken to collate measurement tools for organisational context from the domains of research use or RA in healthcare, or for KM or OL in the management or organisational science literature.

Search and screening

A search of electronic databases from inception to March 2006 was carried out on MEDLINE, CINAHL, AMED, ZETOC, IBSS, Web of Science, National Research Register, Ingenta, Business Source Premier, and Emerald. Measurement tools were included if they were designed to measure contextual features of whole organisations, or sub-units such as teams or departments. Tools needed to include at least one item relating to organisational factors influencing RU, RA, KM, or OL. To be included, papers had to report a structured method of tool development and psychometric testing.

Data extraction and analysis

Individual reviewers (BF, PB, LT) extracted items relating to organisational context from each measurement tool. Items were excluded if they focused solely on structural organisational factors not amenable to change (e.g., organisational design, size; inter-organisational factors) and environment (e.g., political directives); or characteristics of the commercial context that were not applicable in a public service context. Some tools had items expressed as staff competencies (e.g., 'Staff in our organization have critical appraisal skills...') or organisational processes (e.g., 'Our organization has arrangements with external expertise...' [93]). Items such as these were included and interpreted in terms of the availability of an organisational resource (e.g., facilities for learning critical appraisal skills, or availability of external expertise). However, some items were not expressed in a way that could be inferred as an organisational characteristic (e.g., 'Our employees resist changing to new ways of doing things' [94]), and were excluded.

Category analysis

Initially, similar items from different measurement tools were grouped together, e.g., 'I often have the opportunity to talk to other staff about successful programmes...' [95] and 'employees have the chance to talk among themselves about new ideas...' [96]. After an initial failed attempt to categorize all items using an existing diffusion of innovation framework [25], the review team constructed categories of organisational attributes by grouping items from across all the measurement instruments, and refining, expanding, or collapsing the groupings until a fit was achieved for all extracted items. The material is illustrated in Table 1 by items allocated to two attributes: involving the individual, and shared vision/goals (tool source in brackets – see Table 2[97104]). While broadly similar, it can be seen that items from the different domains are expressed differently, and there was some judgement involved in determining the similarity of meaning across domains. It can also be seen that for some categories, particular domains of tool did not contribute any items, while other domains contributed multiple items.

Table 1 Example of categorisation of items extracted from measurement tools
Table 2 Measurement tools included for item extraction

We conducted three rounds of agreement with the fit of items to categories: an initial round using categories derived from the diffusion of innovation framework by Greenhalgh and colleagues [25], which was rejected because of the lack of fit for numerous items; a second round with our own constructed categorization framework built from grouping items; and a third and final round for reviewers to check back that all items from their measurement tools had been included and adequately categorized in the constructed framework. Between each round, joint discussions were held to agree refinements to categories and discuss any disagreement. Using this process, agreement was reached between all reviewers on the inclusion and categorization of all items. An independent reviewer (LP) then checked validity of extraction, categorization, and merging by tracing each composite attribute back to the original tool, agreeing its categorization, then reviewing each tool to ensure that all relevant items were incorporated. Items queried were re-checked.

Results

Thirty tools were identified and appraised [see Additional file 1]. Based on the inclusion criteria for tool development and testing, 18 tools with 649 items in total were selected. These are listed in Table 2, with information on development and psychometric testing [see Additional File 2] The number of the tool from Table 2 will be used in subsequent tables.

In total, 261 items related to organisational context were extracted from the measurement tools. For two tools [105, 106], the full text of each item was not available, so the names of the categories of measurement for which results were reported were used as items, e.g., organisational climate for change.

Final model

Figure 1 illustrates the final category structure constructed to account for all of the items from the measurement tools. Seven broad categories gave a best fit for the items. The central white circle of the diagram shows three core categories of vision, leadership, and a learning culture. The middle ring shows four categories of activity: 'knowledge need and capture' and 'acquisition of new knowledge' (relating to organisational absorptive capacity); and 'knowledge sharing' and 'knowledge use' (related to organisational receptive capacity). The outer ring illustrates the organisational attributes contributing to each category.

Figure 1
figure 1

Model of categories and organisational attributes.

Tool item analysis

Table 3 summarises the organisational attributes for each category. Attributes are based on a composite of items extracted from the tools across the four domains. An example of a single tool item is given to illustrate the source material for each attribute.

Table 3 Details of attributes in each category, and example of tool items

The marked areas in Table 4 identify the measurement tool source of each organisational attribute. The percentages are derived from the number of times an item is included in a category, compared with the total possible in each domain, e.g., there were two items from RA tools included in the learning culture category, out of a possible total of 16 items. The results for each category are discussed below:

Table 4 Categorisation of measurement tool items

Learning culture

OL and KM tools were the most frequent source of these attributes, with seven out of nine tools covering attributes in this category, although none of the tools covered all of the attributes. Three RA/RU tools covered the attribute of 'involving the individual', with one of the RU tools also including the attribute of 'valuing the individual'. Each attribute was sourced from between three and five tools across all domains. The most representation was sourced from KM tools.

Vision

Eight out of nine of the OL/KM tools, and five out of nine RA/RU tools included attributes from this category. The most common attribute was 'shared vision/goals' (eight tools), and the least common was 'policies and infrastructures' (three tools). The most representation was sourced from OL tools.

Leadership

All of the domains included some reference to attributes of management or leadership. Five out of nine RA/RU tools and four out of nine KM/OL tools included items related to leadership. The most representation was in RA tools.

Knowledge need

All of the OL tools and three out of four of the KM tools included items related to attributes of this category. They were less commonly sourced from RA and RU tools. The most common attribute was 'learning from experience' (seven tools). The most representation was sourced from OL tools.

Acquiring new knowledge

Attributes in this category were more commonly sourced in RA/RU tools. Attributes were sourced from between five and nine tools out of the total of 18 tools across all domains, and each attribute was covered in each domain, except 'accessing information', which was not covered in any KM tool. The most representation was sourced from RU tools.

Knowledge sharing

Most OL/KM tools included multiple attributes from this category, all RA tools included one or two items, but only two out of five RU tools included one attribute. 'Promoting internal knowledge transfer' was the most common attribute, included in 13 out of 18 tools, with 'promoting external contacts' included in seven tools. The other items were included in five tools. The most representation for this category was sourced from OL tools.

Knowledge use

Overall, this was the largest and most populated category. The most common attributes referred to were 'encouraging innovation', included in 14 out of 18 tools, and 'role recognition/reward', referred to in 13 tools. Each of the other attributes was also referred to in at least eight tools. All attributes were sourced from all domains. The most representation for this category was sourced from RA tools.

Analysis of tool coverage

Table 4 also summarises how well each tool domain covers the constructed categories and attributes. The results for each domain are discussed below:

RA tools

The category with the most representation in the RA tools was 'knowledge use', with items in the category of 'acquiring new knowledge' and 'vision' also well represented. The categories of 'knowledge need' and 'knowledge sharing' were less well reflected across the RA tools. Two attributes of 'recognising and valuing existing knowledge' and 'knowledge transfer technology' did not appear in any RA tool. Five attributes appeared in only one of the tools. Four attributes of 'developing expertise, role recognition and reward','support/access to expertise', and'access to resources' were common to all tools. Two tools had relatively good coverage of the attributes: the ABC survey [107], with 14 out of 26 attributes covered, and the KEYS Questionnaire [93] with 15 out of 26 attributes covered.

RU tools

This was the domain with the least coverage overall, commonly centered in the categories of 'acquiring new knowledge' and 'knowledge use'. The other categories were poorly represented. The attribute of 'accessing information' was common to all tools, with 'role recognition/reward', and 'support/access to expertise' common to four out of five tools. The tool which covered the most attributes (10 out of 26) was the RU Survey Instrument [105, 108].

KM tools

The KM tools covered all of the categories, with more common representation in the categories of 'learning culture', 'knowledge need', 'knowledge sharing' and 'knowledge use', but individual tools varied in their emphasis. The categories of 'leadership' and 'acquisition of new knowledge' were the least well represented. Two attributes were included in all four tools: 'promoting internal knowledge transfer', and 'encouraging innovation'.'Learning climate' and 'access to resources' were included in three out of four tools. Five attributes were not represented in any tool: 'involving the individual','policies and infrastructures','managerial attributes','accessing information', and 'supporting teamwork'. The tool with the best overall coverage of the attributes (13 out of 26) was the KM Questionnaire [109].

OL tools

OL tools covered all categories, and generally had more consistent coverage than other domains of the categories 'vision', 'knowledge need' and 'knowledge sharing'. Single attributes relating to 'promoting internal knowledge transfer', and 'encouraging innovation' were covered in all five tools, with the attributes of 'communication', 'shared vision and goals','learning from experience', and 'promoting external contacts/networks' covered in four out of five tools. 'Key strategic aims','policies and infrastructures','questioning culture', 'accessing information', and 'exposure to new information' were only covered in one out of the five tools. The OL Scale [110] covered 17 out of the 26 possible attributes. The other four tools covered between 8 and 11 attributes.

Comparison of support for benchmark items: what can EBP tools learn from the KM and OL literature?

While each of the composite attributes is supported by items extracted from at least three measurement tools, there are differences in emphasis across the domains. To consider the potential contribution of the newer domains of KM/OL, the number of items from these domains have been pooled and compared against the number of items sourced from the domains commonly represented in the healthcare literature, i.e., RA/RU. Figure 2 illustrates that the KM and OL literature focus more on 'learning culture', 'vision', 'knowledge need', and 'knowledge sharing'. The RA and RU literatures have a stronger emphasis on 'leadership', 'acquiring new knowledge', and 'knowledge use'.

Figure 2
figure 2

Comparison of RA/RU versus KM/OL measurement tools.

Discussion

The importance of understanding context has been reiterated by the High Level Clinical Effectiveness group [18]. This project was developed in response to perceived limitations in the conceptualisation and measurement of organisational context for EBP in healthcare. We wanted to move away from the rather narrow focus on RU and change management to include wider process and practice-based perspectives from the KM and OL literature. Our analysis of existing measurement tools has confirmed differences in emphasis across the domains. Measurement tools for RA and RU focus more on access to new information, leadership, and resources for change, and less on recognizing, valuing, and building shared knowledge. This is congruent with the culture of 'rationality, verticality, and control' [[6] p660] in healthcare, but the lack of attention to social context may be one reason why attempts to improve practice by influencing the behaviour of individual practitioners have variable results [111].

The emphasis in KM and OL tools on shared vision, learning culture, and sharing existing knowledge reflects a more socially mediated view of knowledge. If it is groups and networks that generate the meaning and value to be attached to evidence, organisational efforts to improve EBP would need to do more to shift towards supporting horizontal knowledge transfer. Networks have emerged as a recent UK government strategy for moving health research into action by creating clusters that break down disciplinary, sectoral, and geographic boundaries, but communication structures alone are unlikely to be successful for knowledge transfer across specialized domains [4, 112] without additional mechanisms to support the transfer of practice and process knowledge [73].

Since this search was conducted, three additional tools to measure organisational context in quality improvement related areas have been reported [113115]. Each of these tools has strengths, including attributes such as feedback that are not included in our model, but none have comprehensive coverage of all of the attributes identified in this study.

A potential weakness in using existing tools as sources in this study is that they might not reflect the latest theories and concepts, because tool development tends to lag behind conceptual development. This might result in inadvertent bias towards earlier more technical models, and we acknowledge that the existing tools do largely adopt a structuralist perspective. While the items contained in the existing measurement tools can only ever provide a rather simplistic reflection of complex phenomena, we felt that including them was better than not attempting to express them at all. However, while new perspectives are worth investigating, the unquestioning interdisciplinary transfer of theory also needs care. Compared with the public sector, there are differences in the types of problems, the availability of information and resources, and the motivations for evidence uptake and use. KM theory supposes an identified knowledge need, scarce information, and a workforce motivated by external incentive in a resource-rich environment. EBP on the other hand requires compliance with externally produced information for predominantly intrinsic reward, with high innovation costs in a resource-limited environment. A number of studies have identified some of the difficulties of knowledge sharing in the public sector [1, 6, 11, 16, 48, 116, 117]. Organisational theory may not transfer well into healthcare if EBP is viewed as a process of social and political control to promote compliance with centrally derived policy, rather than a generative process to make best use of available knowledge.

Conclusion

Assessing organisational absorptive and receptive capacity with the aim of improving organizational conditions is postulated as a first step in supporting a research informed decision-making culture. Foss [87] suggests the emergence of a new approach referred to as knowledge governance: the management of the mechanisms that mediate between the micro-processes of individual knowledge and the outcomes of organisational performance. But what would this mean in practice? The kinds of support which KM and OL tools include as standard, but that are not well reflected in existing tools to measure context in healthcare, would include effort to detect and support emergent and existing communities of practice; encourage and reward individuals and groups to ask questions; discuss and share ideas across knowledge communities; and support the progression, testing, and adopting of new ideas by embedding them in systems and processes.

The processes by which individual- and group-level knowledge are collated into organisational level capability to improve care are less clear. If social networks of individuals are to be facilitated to undertake repeated, ongoing, and routine uptake of evidence within their daily practice, we also need to extend our thinking even further toward considering the organisational contextual features that would support the collective sense-making processes of key knowledge workers.