Introduction

The Category Normalised Citation Impact (CNCI) indicator puts citations in context. By normalising citation counts by publication year, document type and subject category, reasonable comparisons between the influence of papers within and without a given subject field are permitted. It is a standard indicator for national and institutional research performance comparisons (Jappe, 2020) and has been used to inform government policy (BIS, 2009), hire and promote academic staff (Holden et al., 2005) and allocate resources (Carlsson, 2009). CNCI, which is used in Clarivate products, is not the only indicator of its type; Elsevier, for example, offer the Field-weighted citation impact (FWCI).

The need for citation normalisation is well justified (Schubert et al., 1988), given field (e.g., Garfield, 1979) and national cultural (Adams, 2018) differences in citation growth, and continues to be a central theme of bibliometrics and scientometrics (Waltman & van Eck, 2019). CNCI, however, considers only three factors in its normalisation. There are arguments that other factors should be considered when normalising citations such as the referencing behaviour of citing publications or citing journals (e.g., source normalisation: Waltman & van Eck, 2013; citing-side normalisation: Zitt & Small, 2008), or the use of citation percentiles (Bornmann, 2020; Bornmann & Williams, 2020). A weighted CNCI, based on citations within a short time window and a fixed longer window has also been suggested (Wang & Zhang, 2020; Zhang & Wang, 2021). Others (Adams et al., 2007, 2019a, 2023) suggest that indicators should instead be used as part of a profile, not as a standalone metric. CNCI also relies on citations acknowledging the usefulness of work (Garfield, 1955) and the broad correlation between citation accumulation and peer judgements of research significance (i.e., higher cited papers are judged to be more significant e.g., Moed, 2005, 2017). Though this is generally true, citations can be made for numerous reasons (e.g., Garfield, 1977; Small, 1982) as well as those of a nefarious nature (e.g., Heneberg, 2016; Bartneck & Kokkelmans, 2011; Fister Jr et al. 2016).

A further complication is that academic research has become increasingly team-oriented (Bozeman & Youtie, 2017) and international in the last 40 years (Adams, 2013; Narin et al., 1991). Motivations for international collaboration include knowledge transfer, access to equipment and financial aid and it is generally viewed positively (e.g., Hicks & Katz, 1996; Katz & Martin, 1997). Bilateral partnerships, which flourished into the 2000s, are beginning to be superseded by more multilateral partnerships (Adams & Szomszor, 2022; Adams et al., 2019b) and the formation of a global research network (Adams & Szomszor, 2022; Wagner & Leydesdorff, 2005), which is increasingly necessary to tackle global scale issues such as pandemics (e.g., COVID-19) and climate change, and large-scale technical research projects such as CERN.

Due to the normalisation process, the benchmark value of CNCI is one, which represents world average. A paper with a CNCI value of 2 would be above world average, having twice the number of citations as the average paper; a paper with a CNCI value of 0.5 would be below world average, having half the number of citations of the average paper. However, because of this single value outcome, CNCI offers only a snapshot of a document’s performance. How this performance credit is subsequently attributed to authors (and therefore institutions and countries), especially as more collaborative work generally results in higher CNCI (Adams et al., 2019b; Glänzel & Schubert, 2004; Narin et al., 1991; Thelwall, 2020; Waltman & van Eck, 2015), creates issues.

Numerous counting methods have been devised (see Gauffriau, 2021) – 32 since 1981 – to assign publication credit to authors. The most common of these is fractional counting, where an entity is assigned a fractional share of paper credit (e.g., Aksnes et al., 2012; Burrell & Rousseau, 1995; Egghe et al., 2000; van Hooydonk, 1997), which itself has many variations (e.g., Sivertsen et al., 2019; Waltman & van Eck, 2015). Fractional counting has been applied to institutions (e.g., Leydesdorff & Shin, 2011), countries (e.g., Glänzel & De Lange, 2002), and used by national research councils to evaluate work (e.g., Sweden: Kronman et al., 2010). However, as author numbers on publications increase, sometimes into the hundreds or even thousands, the assignment of fractional credit and, consequently any evaluation of such work, becomes futile to the detriment of research management.

Given the collaborative nature of research, it appears appropriate to consider the type of collaboration when attempting to apportion credit. Collaborative CNCI (collab CNCI) was formulated (Potter et al., 2020, 2022) with this is mind, and as an approach that does not presume credit (as fractional counting does). This alternative approach assigns documents to one of five types (discussed later), depending on the level of collaboration, and normalises citation counts by this, in addition to the three parameters standard CNCI already considers.

Results have demonstrated, at both the institutional and national level, that collab CNCI values are generally lower than standard CNCI, due to the additional normalisation, and agree well with fractional values (Adams et al., 2022; Potter et al., 2020, 2022). Collab CNCI can also highlight situations where an institution’s domestic articles outperform its internationally collaborative articles (Adams et al., 2022; Potter et al., 2022). Additionally, division of papers into the five collaboration types provides an in-depth profile, rather than a single metric, view of an entity’s research profile for assessment. Without this additional normalisation and subdivision, crucial aspects of an entity’s research profile would be overlooked and potentially impact management decision making.

Here, building upon our previous work (Potter & Kovač, 2023; Potter et al., 2022), these three alternative CNCI indicators (standard, collaboration, and fractional) are used in concert to analyse and explain institutional and, by extension, regional CNCI changes over a ten-year period. This demonstrates the ‘profiles not metrics’ approach and offers insights into the practicality of combining approaches and its importance in performance evaluation and informing research management decision making.

Methods

This work used the same data source as Potter et al. (2022) and Potter and Kovač (2023). This consisted of documents with the type ‘article’ published in the Web of Science Core Collection between 2009 and 2018 inclusive. Building upon Potter and Kovač (2023), this set was filtered to include 30 institutions from nine predefined regions: North America (defined as just USA and Canada), Mexico and South America (MexSAm), Western Europe (WEur), Eastern Europe (EEur), Nordic Europe (Nordic), Middle East-North Africa-Turkey (MENAT), Southeast Asia (SEAsia), China, and Asia–Pacific (APAC). Institutions for each region were chosen semi-randomly to reflect high, middle and low volume article-producing institutions, with a minimum threshold of 5000 (co)authored articles over the 10-year period. Africa had only 11 institutions covering three countries (Nigeria, South Africa, Uganda) that met the minimum criteria, but was included for completeness. The dataset included almost 5.5 M articles. Institutions were collated from author address information available within Web of Science. Specifically, these were the ‘Affiliation’ data provided as part of the Author Information in the Web of Science metadata for a given document. This ‘Affiliation’ provides a ‘unified’ institution name that merges names from Web of Science addresses including name variants, such as previous names, affiliated sub-organisations and spelling variants, and is a combination of background research by Clarivate editorial staff and feedback from organisations. Though some address information was incomplete, this did not affect result interpretation due to the relatively small set of incomplete data (e.g., ~ 3% of articles when considering the top 20 institutions by article count).

CNCI for each institutions’ articles was calculated in the standard manner: citations counts were normalised by document type, year of publication, and Web of Science subject category, and is hence referred to as standard (s-) CNCI. The Web of Science subject categories cover 254 subjects with journals sometimes assigned to multiple ones; in such cases, the average CNCI of all an article’s categories was assigned to the article.

Collab CNCI, defined by Potter et al. (2020), and sometimes referred to as c-CNCI in the following text, was calculated in the same manner as standard CNCI but with the additional normalisation by collaboration type. The five possible collaboration types are: domestic single institutional, domestic multi-institutional, international bilateral, international trilateral, and international quadrilateral plus. International collaborations were based only on the number of unique countries; number of unique institutions was irrelevant. In theory, a paper with a single author could fall into any one of these collaboration types, depending on their affiliations. It is important to note that highly multilateral work (i.e., international quadrilateral plus papers) was rare. This set of papers accounted for only 4% of all output though this is likely to increase in the future.

The standard and collab CNCI values for an article were assigned to all institutions on that article. For example, if an article with three institutions had a CNCI of 1.5, each institution was awarded an article count of 1.0 and a CNCI value of 1.5. An institution’s mean CNCI was then calculated by summing the CNCI of all the articles on which the institution was recorded in an author address and dividing this by the number of such articles, over the entire dataset.

Fractional CNCI, also referred to as f-CNCI in the following text, was calculated at the author level following Waltman and van Eck (2015). This method assigns credit to each author, institution, country etc. on a publication based upon the total, deduplicated number of each. This credit is then multiplied by the article’s CNCI to calculate the fractional CNCI value. An institution’s mean CNCI was then calculated by summing the CNCIs on which that institution was recorded in an author address and dividing this by the sum of its total fractional CNCI value. The reader is referred to Waltman and van Eck (2015), as well as Potter et al. (2020), for more thorough descriptions and methodology of each indicator. To investigate temporal changes between each of the CNCI approaches, the absolute differences in CNCI values between the years 2009 and 2018 were compared.

Results

Temporal CNCI indicator changes

Figure 1 presents CNCI values for all institutions across all three indicator variants for the years 2009 and 2018. All regions, bar North America and Western Europe, show a general increase in s-CNCI over the period. For the other regions, some tend to show a far greater spread in s-CNCI values in 2018 compared to 2009, particularly Eastern Europe. However, some of this spread may be due to the choice of institutions. These results contrast those seen for fractional and collab CNCI, where there is little difference in overall trends between the two years. North America and Western Europe appear to show a general decline in fractional and collab. China is the only region where the general trend is an increase in indicator values. Overall, fractional and collab CNCI have largely subdued any improvements seen in s-CNCI.

Fig. 1
figure 1

A comparison of standard, fractional and collaboration CNCI values between 2009 and 2018 for selected institutions covering 10 global regions. Regional labels refer to: USA and Canada, Eastern Europe (EEur), Middle East-North Africa-Turkey (MENAT), China, Asia–Pacific (APAC), Southeast Asia (SEAsia), Nordic Europe (Nordic), Western Europe (WEur), Africa, and Mexico and South America (MexSAm)

Between the two reference years, some institutions distance themselves from their peers for s-CNCI, notably University of Adelaide, Australia; Shanghai Institute of Biological Sciences, China; and Niels Bohr Institute, Copenhagen, Denmark. However, when considering fractional and collab CNCI, many of these are no longer unique (e.g., another institution has a similar value) or fall back into the main group. Only the University of Adelaide remains in a unique position across all three indicators. Massachusetts Institute of Technology (MIT), by contrast, declines across all three indicators though remains the top performing institution. Uniquely, King Abdullah University of Science and Technology (KAUST), Saudi Arabia, was an anomaly in s-CNCI, though not the other indicators, in 2009 (incidentally the year in which the university was officially established). However, by 2018, it had become an outlier for fractional and collab CNCI. Though it’s s-CNCI value increased, it is now joined by King Abdulaziz University, Saudi Arabia, as stand out performers in the MENAT region.

Figure 2 converts the data in Fig. 1 to absolute differences plotting s-CNCI and f-CNCI against c-CNCI to show relative relationships between the indicators. When comparing absolute s-CNCI and c-CNCI changes over the ten-year period, institutions generally have a greater s-CNCI improvement (s-CNCI = 1.158 × c-CNCI). When comparing f-CNCI and c-CNCI changes, institutions generally have greater c-CNCI improvements (f-CNCI = 0.904 × c-CNCI).

Fig. 2
figure 2

Absolute difference in standard (s-) CNCI (top) and fractional (f-) CNCI (bottom) against collaboration (c-) CNCI values

Chinese institutions strongly follow the identity (x = y) line (gradient of 0.97 and 0.89 for s-CNCI and f-CNCI, respectively, vs c-CNCI; Supplementary Fig. 1, Supplementary Table 1) with the vast majority showing increases in all three CNCI variants and consequently plotting in the first quadrant (+ x, + y). African institutions show a similar trend, but with the caveat of only 11 institutions included. Southeast Asia and MENAT institutions also tend to follow the identity line, though more institutions fall into the second (+ x, − y) and third (− x, − y) quadrants showing decreases in at least one variant; North American institutions mainly plot in these quadrants too. APAC, Western Europe, Nordic Europe, and Mexico/South America institutions are generally concentrated around the origin; Mexico/South America tend to follow the identity line, while the other regions are more concentrically clustered, with Western Europe more dispersed among the four quadrants with many of its institutions showing decreasing values. Eastern Europe institutions tend to plot in the first quadrant, but with a far greater s-CNCI increase than c-CNCI increase.

Institutions of note include Northwestern Polytechnical University (NPU), Xi’an, China, which exhibits significant increases across all CNCI variants (0.78 for s-CNCI, 0.71 for f-CNCI, and 0.74 for c-CNCI) and MIT, Cambridge, USA, which exhibits a decline in all variants (− 0.24 in s-CNCI, − 0.49 in f-CNCI, and − 0.38 in c-CNCI). Some institutions have mixed outcomes, for example, Tulane University, New Orleans, USA, has an s-CNCI and c-CNCI increase of 0.49 and 0.16, respectively, but a decline of − 0.17 in f-CNCI. However, for most institutions with mixed outcomes, one change is usually minor (i.e., < 0.05).

Fractional and collab CNCI rely, to differing extents, on the number of collaborating countries and institutions on papers. Figure 3 illustrates the division between article outputs and their citation share for each collaboration type for selected institutions.

Fig. 3
figure 3

Total articles and total citations by collaboration type for selected institutions over the period 2009-2018. Collaboration types: domestic single institutional (dom:single), domestic multi-institutional (dom:multi), international bilateral (int:bilat), trilateral (int:trilat) and quadrilateral plus (int:quad+). Note different scales

Over 50% of Niels Bohr Institute output is international quadrilateral plus. Additionally, these papers account for 60% of all citations received. Overall, the institute’s output is dominated by international collaboration (~ 90%). There is no domestic single output as the institute is part of the University of Copenhagen.

NPUs output is heavily domestic (77%). International quadrilateral plus output is negligible (< 1%). Notably, the share of citations per group closely follows the article share, though international papers received relatively slightly more. Seoul National University is also highly domestic (72%), mainly domestic multi (54%). Domestic single and international bilateral have similar output (~ 18%). International quadrilateral plus accounts for ~ 6% of output but nearly 19% of citations.

MIT has an almost equal split between domestic (49%) and international (51%) output. International quadrilateral plus articles account for 20% of citations compared to 12% of output.

Table 1 compares the average number of institutional partners at each collaboration level in 2009 and 2018 for the same institutions presented in Fig. 2. Almost all institutions increase their average partnerships across all collaboration types. Some increases are incremental (NPU international trilateral partnerships increases from 5.3 to 5.4), while others increase by one whole institution (e.g., University of Adelaide international trilateral from 5.0 to 6.0); for international quadrilateral plus, this can be tens of partners (e.g., Tulane University from ~21 to ~88; Niels Bohr Institute from ~17 to 109; note that NPU had no international quadrilateral plus papers in 2009). Similar patterns are seen for median values (Supplementary Table 2); these values are almost identical to the means apart from international quadrilateral plus where medians are generally far lower than means (e.g., Niels Bohr Institute), though there are exceptions (MIT and Seoul National University in 2009). Table 1 also illustrates that the relative percentage of international papers has increased for all shown institutions. International papers were the main output of MIT and University of Adelaide in 2018. NPUs international output more than doubled from 12 to 28%, though is still very much domestically focused. Across the complete dataset, 20% of articles had only a single institution, 26% had two institutions; 22% had five or more institutions; 4% had 10 or more.

Table 1 Average number of institutional partners for each collaboration type for selected institutions. Numbers are inclusive of the institution analysed

Discussion

General trends

Well established research regions (such as North America and Western Europe) generally saw stagnation or declines in CNCI values, however their values remained competitive compared to other regions’ institutions. For example, MIT, which is likely seen as a prestigious institution to partner with, declined in all three indicators over the period but remained the top ranked institution in 2018. Such trends are likely a consequence of well performing institutions only having marginal increases in indicator values or decreases as they cannot continue to perform at the same level. This is analogous to the Red Queen Race in game theory where one must run fast to stay in the same place and even faster to move forward. Africa and Mexico/South America institutional gains in CNCI were generally nullified when looking at f-CNCI or c-CNCI. This is corroborated by previous results, though at the national level, that showed fractional counting generally benefits large research economies (Potter et al., 2020). Eastern Europe and China do see increases in c-CNCI over the period in question. This suggests that their research (unlike that of Africa and Mexico/South America), when analysed by collaboration type, is performing better than their peers. Overall, results illustrate that international collaboration is generally increasing over the period, further corroborating the results of Potter et al. (2022) as well as others (e.g., Adams & Gurney, 2018; Ribeiro et al., 2018).

Chinese citation patterns

Chinese institutions appear relatively immune to using different CNCI indicators showing similar increases across all three; accounting for collaboration either fractionally or by collaboration type seems to make little difference, in contrast to other world regions (Figs. 1, 2, Supplementary Fig. 1, Supplementary Table 1). Several factors could explain this phenomenon. First, China’s scientific output has grown rapidly over the last 20 years (Leydesdorff & Zhou, 2005; Stahlschmidt & Hinze, 2018; Wang, 2016) making it the second largest scientific producer globally (behind USA) in the most recent year of this analysis (2018). Secondly, Chinese publication citation characteristics differ from the worldwide citation distribution (Stahlschmidt & Hinze, 2018); China researchers tend to cite their fellow citizens more often (Bakare & Lewison, 2017; Shehatta & Al-Rubaish, 2019), as well as citing authors from some Asian countries more than others (Tang et al., 2015). Additionally, Chinese authored research tends to have longer reference lists than the world average (Stahlschmidt & Hinze, 2018). These all provide more opportunities for Chinese citations. Finally, articles authored by Chinese researchers are increasing their presence in the references of other countries’ publications, while, for example, USA authored articles, are experiencing a slow decline (Khelfaoui et al., 2020).

Highly cited researchers

Unusual increases in CNCI indicator values may also be due in part to hiring policies, particularly hiring of Highly Cited Researchers. These researchers, selected by the Institute for Scientific Information (ISI) at Clarivate, have demonstrated significant and broad influence reflected in their publication of multiple highly cited papers over the last decade. These highly cited papers rank in the top 1% by citations for a field (or fields) and publication year in the Web of Science. Two institutions that showed anomalously high CNCI values leaving behind their peers in their respective regions were KAUST and University of Adelaide. These institutions saw their number of highly cited researchers increase from 5 to 13 (KAUST) and 5 to 10 (University of Adelaide) between 2014 and 2018 (R. Fry, personal communication, November 7, 2022). While further assessment is outside the focus of this study, the hiring and potential citation impact of HCRs could positively influence an institution’s CNCI.

Collaboration and research focus

Changes and differences in CNCI indicator values could also be due to individual institution’s research foci. For example, institutions focused on large, highly multi-lateral projects would likely receive low fractional credit regardless of how many citations such work received. For example, the Niels Bohr Institute (Figs. 2 and 3, Table 1), whose papers with the most institutions (494, 298, 297) where all CERN-based collaborations (ATLAS Collaboration 2018a, 2018b, 2018c). The increasing trend of highly multi-institutional, multi-lateral papers is seen in Table 1. How these papers perform relative to their peers (i.e., same collaboration type) would influence its c-CNCI value (e.g., NPU—Fig. 2). Additionally, institutions that mainly publish with no international co-authors may see the likelihood of their work being noticed by others diminished, but this could depend on the size of their national research base (e.g., NPU located in China—Figs. 2, 3). We should expect collaboration to be influenced by many factors, of which research discipline will be one because of links between the research and the need for teams and facilities. As fields of research vary between institutions this could impact the type and size of collaborations formed. These topics will be covered in future analyses.

These results, therefore, suggest that the collaborations institutions make may influence CNCI changes over time. This could include the hiring of Highly Cited Researchers, who may relatively quickly amass citations directly influencing CNCI values. Additionally, collab CNCI demonstrates how research performs relative to its collaboration group. As previously shown (Potter et al., 2022), such analysis could highlight cases where an institution’s domestic research outperforms its more international and collaborative output. However, such insights should not discourage increased collaboration – it merely highlights that CNCI as a single metric does not provide a thorough view of performance and that other factors may be influential and should be considered by research and policy managers.

Conclusions

The standard CNCI indicator provides only a snapshot of an institution’s performance hiding underlying factors. By including fractional and collaboration CNCI alongside the standard approach a greater ‘profiles not metrics’ view can be produced demonstrating the complementarity of these indicators. However, to fully appreciate and understand the meaning of CNCI value changes, research and policy managers must understand the formulation behind each CNCI indicator, recognising the (dis)advantages. Additionally, in an increasingly collaborative world, it is vital to understand how an institution’s collaboration profile is changing over time, as collaboration complicates analysis requiring more nuanced approaches. Furthermore, collaboration partnerships and citation patterns which directly influence CNCI calculations must be considered. Evaluating and comparing the performance of institutions is a main driver of funding and research direction decisions. Managers must understand all aspects of CNCI indicators, including secondary factors, to be able to make better informed decisions.