Introduction

Early key quantitative research findings of spatial scientometrics, a relatively young field drawing both from traditional scientometrics and geography of science (Frenken et al., 2009) showed that high-prestige research output—results published in journals covered by the Science Citation Index (SCI)—is highly clustered in a few core countries both with regards to quantity and impact (Frame et al., 1977; Narin & Carpenter, 1975). Subsequent investigations into the spatial aspect of scientific knowledge production continued focusing on national differences in science production (Csomós, 2019), while critical social scientific studies contributed to the discussion, reflecting on the uneven power relations and biases in global academia (Demeter, 2019a, 2020; Goyanes & Demeter, 2020; Háló & Demeter, 2022).

In this paper, we continue exploring the intersection of scientometrics and critical sociology through an empirical analysis of geopolitical differences amongst the most productive scholars in communication studies. Over the past century, communication has become an important and emerging area of scientific inquiry in global academia due to major social and technological changes (Marinho & Mariño, 2018). Communication science is generally understood as the study of how people use messages to create meanings within and across various contexts, channels, and media (Craig, 2008), however, the field is still facing challenges in defining its comprehensive identity due to its multidisciplinary foundations as well as the rapidly evolving media and socio-technological environment (Nordenstreng, 2004, 2007; Waisbord, 2019). As the paper at hand applies scientometric means to conceptualize the field, for the purpose of this study, communication science is to be understood as a category of SCOPUS; the major scientometric databases that provide data for the presented analyses.

The study contributes to the field by being the first analysis to measure geopolitical biases among the most productive scholars in communication studies through SCOPUS view counts. Besides testing the more traditional citations per document formula, we also measure scholarly performance through views per document and citations per view and examine how geographical location (derived from a scholar’s institutional affiliation) affects the impact of scholars ranked top 500 by SciVal in Communication. Comparing researchers from 11 countries and 3 world regions, we investigate if there are statistically significant differences between the frequency of their publications being accessed as well as cited.

Scientific excellence through the lenses of country-level metrics

Country-level scientometric analyses, as well as its variants focusing on geographic regions (Bornmann & Waltmann, 2011; Godin & Ippersiel, 1996) or organizations (universities, clinics, hospitals, public and private research institutions as centers of knowledge production in a metropolitan region; Matthiessen & Schwartz, 1999; Matthiessen et al., 2010; van Noorden, 2010) remained widely popular in scientometrics over the years. Focusing on the quantities and impact of academic knowledge production, subsequent country-level research can be grouped into two major categories based on the source of data they draw on. The first group deals with interactions occurring within scientific platforms, measuring the acknowledgment and further use of results among members of the scientific community, while the second aims to measure impact outside the peer-review realm and look at activities in various social and other digital media sources via alternative metrics (altmetrics as commonly referred to; Priem et al., 2010; Repiso et al., 2019; Wasike, 2021).

A major traditional indicator for the former—besides the number of published papers—is citation count, a metric considered to indicate the influence, perceived quality, and, ultimately, academic/social value of published research. Research on the spatial distribution of received citations gained a considerable impetus by showing the applicability of Merton’s classical concept of the Matthew effect (a cumulative advantage in science) at the country level: Bonitz et al. (1997). On the one hand, it was found that papers authored by US, UK, and another few European core country affiliations received more citations than what could have been expected based on the impact factor of the journal they were published in.

On the other hand, authors affiliated with any other country received fewer citations than expected.Footnote 1 The authors coined this systemic discrepancy in expected citations as the ‘Matthew Effect for Countries’ (MEC), showing that the accumulation of citations is radically uneven between nations. Later results with a global (Pan et al., 2012; Perc, 2014; Schmoch & Schubert, 2008; Zanotto et al., 2016) or a European (Makkonen & Mitze, 2016) focus confirmed that the regional distribution of global science is more of a question of presence versus absence than the competition in a field of weaker and stronger but roughly comparable players. Since both quantity and impact are concentrated in traditional Western countries, and other players’ partial success can be directly linked to their westernizing efforts (Demeter et al., 2022a, 2022b; Háló, 2022), critical approaches to these inequalities are often discussed in terms of de-Westernization in communication/media research (Curran & Park, 2000; Gunaratne, 2005; Thussu, 2009; Waisbord & Mellado, 2014; Wang, 2011).

Similar trends arise on the individual level. Individual success of an academic is closely tied to editing (Goyanes & de-Marcos, 2020) and publishing in top-tier journals (Efranmanesh et al., 2017), as well as affiliations with top-tier universities (Cowan & Rossello, 2018). As research on both native-English (Collyer, 2014) and European (Fumasoli et al., 2015) countries has shown, scholars educated at elite Western universities are favored in all of these segments of academic life (Demeter & Toth, 2020). Recent results on the systemic conditions of academic excellence in Communication studies conclude that besides personal traits, being excellent requires “following the research conventions, interests and values of a research community and conducting “sound research” based on tacit rules of the scientific craft, which were acquired at elite universities.” (Goyanes & Demeter, 2021, p. 76).

The majority of elite universities being Western, it is not surprising that highly cited individual authors are similarly concentrated in specific geographical areas. Data could be refined even further to expose smaller geographical centers within countries. In one of the earliest country-level citation assessments, Batty (2003) studied 1222 top scientists from ISI’s Highly Cited database, covering twelve scientific fields. They found that 815 of these are from the US and 100 are from the UK (the third being Germany with 62, followed by European countries, as well as Japan and Australia in the first 10 places). Also, when looking closer at their institutional concentration, the top 20 institutions among these affiliations (in terms of the number and percentage of scientists cited), which housed nearly 30% of the 1222 highly cited scientists, were from the US (18) or the UK (2). These do not necessarily mirror the domestic efficiency of US education investments, as foreign-born and foreign-educated high-performing scientists systematically migrate to the US due to its high GDP and large Research and Development (R&D) spending (Corley & Sabharwal, 2007; Franzoni et al., 2014; Hunter et al., 2009; Stephan & Levin, 2001). This elite brain drain is another country-level example of cumulative advantage in academia. Economic wealth correlates with scientific production in the case of European countries too (Almeida et al., 2009; Lee et al., 2011). Drawing on the ISI Thomson National Science Indicators on Diskette database (NSIOD-2003), Horta and Veloso (2007) concluded that high-income countries of the EU are generally more successful in amassing publications than low-income ones, and their research output is of the highest impact. However, even including the UK (which is no longer part of the EU), the total international visibility and impact of EU research are still lower than that of the US (Horta & Veloso, 2007).

Regarding the share of publications, citations received, and the average number of citations received per paper, the dominance of the US and the UK is still evident. Counting all publications in English in the ISI Web of Science (WoS) database (Science Citation Index Expanded, Social Sciences Citation Index, and the Arts & Humanities Citation Index) for the period 2003–2010, the analysis of Pan et al. (2012) showed that the US produced around 28.12% of WoS-indexed publications (UK: 6.51 percent) and received 38.22% of all WoS citations (UK: 7.45%) For perspective, the same indicators are 23.65% (WoS publications%) and 26.03% (WoS citations%) for the Western European region and 3.49% and 1.76% for the Eastern European region.

The average number of citations per paper is around 10 for US and UK papers, while it varies between 6 and 12 for countries in Western Europe. However, among Eastern European countries, Hungary had the highest average citations per paper (7.31), and Romania had the lowest (3.30). In general, high-income European countries thus seem to be more efficient than lower-income European countries, and their output is of higher impact.

A trend towards more highly cited publications among European countries – even at the cost of quantity in terms of published papers—was indicated by Leydesdorff et al. (2014) among 28 EU nations. Applying integer counting to allocate publications, his analysis allowed for the assumption that, ceteris paribus, 10% of a nation’s internationally co-authored publications can be expected to be within the top 10% of the most highly cited publications, while mutatis mutandis, 1% of these publications to be amongst the top-1%. It is noteworthy that not all European countries were found to pull their weight equally: while Western European countries managed to contribute according to or above expectations to both the top 1% and top 10% most highly cited publications, among the countries from the Eastern European region, only Latvia and Estonia scored above expectations in the top 1% and top 10%. All other Eastern European countries performed below expectations in both sets (Leydesdorff et al., 2014).

Furthermore, an epistemic vulnerability of all non-US science was shown by empirical data by Bornmann et al. (2018). The authors calculated that—in addition to the US dominating the top 1% highest cited publications by a 24% share—44% of all references in the top 1% of publications are made to US-authored outputs. This means that US research is the main source of both US and non-US top research from which prolific researchers worldwide draw their knowledge base. The US also has higher-than-expected citations compared to its already high publication volume, and only a few Western European countries show similar patterns (Netherlands: 1.70% share and 2.47% presence in the reference list of top 1% articles; Switzerland: 1.12% and 1.77%; UK: 5.57% and 7.79%; Sweden: 1.15% and 1.49%). Eastern European countries that managed to produce at least 1% of the top 1% highly cited articles all show the opposite tendency; their accumulated knowledge is being used less than expected in top research compared to their publication volume (Bornmann et al., 2018).

Further differences between relative comparative advantages and disadvantages in major disciplines have been shown by Harzing and Giroud (2014). Looking at the share of a country's papers in social sciences relative to all papers between 1994 and 2012, the US, the UK, the Netherlands, and Norway have a medium comparative advantage in this field, while France, Italy, Hungary, Poland, Russia, and Ukraine have a medium comparative disadvantage, and Austria, Germany, Switzerland, Finland, Ireland, Belgium, Sweden, and Denmark have a low comparative disadvantage. Regarding communication research, a seminal study by Lauf (2005) found that communication journals indexed in WoS, as well as the most high-ranking players of this already meticulously screened and selective group are not only mostly US-published, but their gatekeeping processes and content are also heavily dominated by the US. Facing the data, Lauf proposed measures pointing towards a possible international diversification of communication studies by publishing more results coming from so-far underrepresented regions and accepting major regional or national journals into WoS.

Some measures of Thomson Reuters (TR) implemented at this time resonated to the later proposition as the company started to examine more than 10,000 non-US and non-UK journals in 2007, which mainly published research from a particular region or country and covered non-mainstream topics of mostly regional relevance, a portion of them finding ways into the main Thomson Reuters indices like SCI (Science Citation Index), SSCI (Social Science Citation Index), and A&HCI (Arts & Humanities Citation Index). Although these measures had a positive effect on the coverage of Eastern European contributions (Leydesdorff & Wagner, 2009), they left top-ranked journals unaffected due to the lack of implementation of the so-called pillar stone impact criteria during the vetting process (Aman, 2015). Though citation analysis may have been applied in some cases, the importance of a regional journal was not measured by its citation impact. Instead, its inclusion was decided by the specificity of its content that could enrich the source materials already visible to a broader international community of researchers (Testa, 2009). This led to the inclusion of low-impact, quasi-invisible Eastern-European Communication journals into WoS’s main databases (Tóth, 2018). It can even be said that, in a sense, TR deepened existing inequalities in the representation of scholars of the “West and the Rest” (Ferguson, 2012) by inflating the number of lower-impact journals in WoS with regional ones, making the bottom more available while the top continued to be as closed as ever for the “Rest”.

Nevertheless, the regional expansion of WoS during this period and the slow internationalization of some leading communication journals’ editorial boards resulted in the shrinking of the dominance of the US (and the UK) in communication studies. Comparing two five-year periods (1998–2002 and 2013–2017), Demeter (2018c) showed that many journals had increased the number of their non-US editorial staff, while the ratio of articles produced by US-affiliated authors have decreased from 66 to 50 percent, a major share flowing toward Western European and developed Asian countries. However, Eastern European countries, in comparison, only have around 1% of the world’s total WoS publication output in communication science (Demeter, 2018c).

In the soft sciences, cultural and epistemic differences across geographical regions have a strong role and influence on what literature are considered relevant and important, and what will eventually be cited and canonized in a particular field (Tóth & Demeter, 2021). Though this study focuses mainly on the science production and recognition of three global regions (Western and Eastern Europe, and the U.S.), additional insights can be gained by considering how other regions fare in global comparisons.

Based on Lauf (2005) and Demeter’s (2019b) research, we know that in the field of communication, the list of top 40 countries with the highest share in high-prestige publication output remained, with little variance in order, the same in the past 50 years. With the U.S. leading the ranks, the rest of the list is comprised of 21 European and 18 non-European countries. Among these 18 non-European countries, only 6 have 1% or higher share in the total publication output. These best performers (China, Taiwan, South Korea, Canada, Australia and New Zealand) all have high GDP and the last three also share language and cultural ties through colonial British-American history. The above economic and cultural factors are heavily present in the remaining top-40 countries as well, with the addition of a third cluster comprising of countries with a Hispanic colonial heritage. The African continent’s visibility is very low compared to other world regions -not counting South Africa, the overwhelming majority of research by African scholars remain invisible to the international community (Chasi & Rodny-Gumede, 2018), which contributes largely to the “suffocating whiteness of communication studies” originally understood in the frames of racial neoliberalization (Calvente et al., 2020).

To sum up, despite the weakening of the US and the UK’s dominance, and the rapidly rising share of China from high-prestige publication outputs, the MEC is still accurately applicable in Communication Studies to show geopolitical inequalities. Most of the field’s research output of international prestige and recognition is produced by a handful of developed nations, and the contribution of large regions of the world remains invisible (Demeter, 2018a, 2018b; Goyanes et al., 2022). Based on these results, we expect the following:

H1

There are statistically significant differences in received citations per documents between geographical regions.

To test if the difference exists between the most productive scholars across countries, the corresponding sub-hypothesis will be:

H1a

There are statistically significant differences between citation per document within the most productive scholars across countries.

View counts from SCOPUS—a so far neglected metric

Similar to the majority of scientometric research, the largest and most comprehensive journal-level analyses of national diversity in the field of communication studies (see Demeter’s more recent (Demeter, 2019b) and Lauf’s earlier results (Lauf, 2005) used Web of Science data. The MEC was also originally shown by (now Clarivate, then) Thomson Reuters data. However, scientometric analysis of Communication Studies also often considers Elsevier’s SCOPUS (e.g., Demeter, 2017, 2018a, 2018b, 2018c; Tóth, 2018; Trabadela-Robles et al., 2020). The general argument for using SCOPUS is that Elsevier’s product represents soft science fields better, indexes more non-article publication type-items, and has more non-English content (Archambault et al., 2006; Li et al., 2014). Both are important for fields where—compared to hard sciences – regional language articles, monographs, chapters, and conference papers have more relevance in disseminating research results. Even though potential biases stemming from geographical deficiencies in WoS’s journal coverage (overrepresentation of English-speaking countries, especially in the Social Science Citation and Arts & Humanities Citation Indexes) were mitigated by the introduction of the Emerging Sources Citation Index (ESCI) in 2015, it leaves the national distribution of top journals unaffected, as ESCI journals do not have impact factor and are not ranked by JCR (Journal Citation Reports).

While there is flexibility indicated with regard to the source of data, scientometric research rarely considers testing article-level variables accessible from scientific abstracting and indexing services for visibility and impact other than publications and citations. When in need of other metrics, they turn to sources outside the confines of abstracting and indexing services. These altmetrics date back to the Altmetrics Manifesto (Priem et al., 2010) and aim to cover research impact outside the peer-reviewed realm and work not only with “citations” coming from the digital public space – mentions and backlinks pointing to the original research from social media networks, wikis, (micro)blogs and various other sources—but also to emphasize website analytics like the number of views and downloads as impact metrics.

In the digital era, when scientific papers are mostly also published online in some form, it is only natural that usage indicators like the number of views (Bollen et al.,; 2009; Perneger, 2004), downloads (Gorraiz et al., 2014; Moed & Halevi, 2016) and bookmarks (Bar-Ilan et al., 2012; Mohammadi & Thelwall, 2014) are attracting interest in traditional scientometric research as possible forecasters for future citations. Reviewing previous studies, Thelwall (2018) summarized that a positive correlation exists between most altmetric and citation count; the association being stronger (0.5–0.8) at Mendeley reader counts (bookmarks) and weaker (0.1–0.3) at Tweets, Facebook wall posts, blog citations, Google + citations, Reddit citations, and other media mentions. Studies focusing on Twitter exposure found that tweets predict citation rates (Eysenbach, 2011; Peoples et al., 2016), and tweeted articles receive more citations versus those with no tweets (Vaghjiani et al., 2021). Recently, Breitzman (2021) showed that usage in the first six months correlates with a citation index after five years; therefore, these early usage counts can be used to identify papers early that will likely be highly cited, given enough time for other researchers to use them.

Notwithstanding, only a handful of studies analyzed altmetrics in communication science (Torres-Salinas et al., 2013; Repiso et al., 2019; Wasike, 2021; Özkent, 2022) so our knowledge is limited. However, a strong correlation between Mendeley readers/tweets received and WoS citations was found by Repiso et al. (2019), and at least two recent studies (Wasike, 2021; Özkent, 2022) indicated a positive correlation between exposure to social media and article citations in the case of articles in top communication-based journals.

In an environment where strategic altmetrics manipulation (Thelwall, 2021; Zimmermann, 2013), as well as potential national (Kemp, 2022; Singh, 2020; Thelwall & Kousha, 2015) and age (Mohammadi et al., 2015; Sugimoto et al., 2017) biases are already making the application of these measures increasingly complex for research evaluation, any reliable metrics less susceptible for manipulation should be welcomed. It seems that view count measures integrated into abstracting and indexing databases have a few advantages over altmetrics. View counts represent actual instances of a given document being visited within the database, calculated from the sum of abstract views and full-text link clicks, while some altmetrics, most importantly Mendeley bookmarks are referring to potential and not actual readers (Delgado-López-Cózar & Martín-Martín, 2016). Views from a scientific database like SCOPUS are also more likely to be generated by a scholarly audience and more difficult to inflate through digital marketing because of access costs. In comparison, even relatively hard-to-inflate altmetrics generated by a scholarly audience such as Mendeley views can be cost-efficiently tweaked by spamming multiple new profiles to bookmark the very same paper. Other social media mentions, links, and clicks are even easier to adjust and manipulate. Another advantage of accessing articles from a scientific database is that on these platforms, there is no additional context driving the perception of its importance besides traditional article-level metrics: no popularizing summaries and less opportunity to embed the results into cultural or political issues or offer value-laden contextualizations. The number of views can be influenced by a default algorithmic ordering for relevance or recentness or by the author’s name, but not by network effects influential within social media.

To sum up, internal view counts are better suited to measure interest from the scientific community towards an article; raised by the research itself, compared to meticulously designed situations consciously created to drive attention toward a specific content in digital space. The view count metric in SCOPUS is considered internal because it shows how many times an article has been viewed in SCOPUS from the results screen and thus cannot be directly influenced by traffic from anywhere outside the database. With that said, no studies so far have exploited the potential of using this metric for analysis. Therefore, we intend to fill this gap and offer our contribution to the literature by comparing countries and their respective geographical regions based on how many views their indexed documents get while also analyzing how the number of views reflects on received citations. We anticipate the followings:

H2

There are statistically significant differences between views per document between geographical regions.

To test if the difference exists between the most productive scholars across countries, the corresponding sub-hypothesis will be:

H2a

There are statistically significant differences between views per document within the most productive scholars across countries.

H3

There are statistically significant differences between received citations per view between the examined geographical regions.

To test if the difference exists between the most productive scholars across countries, the corresponding sub-hypothesis will be:

H3a

There are statistically significant differences between citation per view within the most productive scholars across countries.

The importance of this study, underpinned by the three hypotheses above, lies in measuring the impact conceptualized through the number of published papers, number of citations, and number of views. While citations per document is a common formula in Matthew-effect analyses (e.g., Pan et al., 2012; Perc, 2014; Schmoch & Schubert, 2008; Zanotto et al., 2016), this will be the first study to measure geopolitical biases among the most productive scholars in communication studies via applying views per document and citations per view measures through SCOPUS data. These formulas make it possible to measure research impact normalized to the number of published papers, the number of citations, and the number of views. It is important to note, therefore, while the number of overall papers tends to raise the number of overall citations when we assess the means for citations per paper, an increase in citations cannot be explained by an increase in publications alone.

Methods

Sample and applied measures

In this paper, we empirically analyzed geopolitical biases among the most productive scholars in communication. We applied three formulas (citations per document, views per document, and citations per view) to measure scholarly performance amongst the top-performing scholars in 11 countries (the UK, France, Germany, Italy, Spain, Poland, Hungary, Russia, Romania, Ukraine, and the US) and 3 corresponding world regions (Western Europe, Eastern Europe, and the US). The reasons for selecting these countries have to do with their broad impact in the literature of communication and because they fairly represent the diversity of communication research in different geographical areas. In addition, their inner connections and cultural bonds enabled us to recode these countries in bigger geographical regions with similar background. The sample was drawn from the SciVal TOP 500 list in Communication for the period between 2017 and 2020.

First, we assessed individual-level publication count, citation count, and view count metrics based on SciVal data for the TOP performing 500 scholars (2017–2020) in the 11 analyzed countries. Second, applying these measures, the three main indicators of our study (I1: citations per document; I2: views per document; I3: citations per view) were formulated as dependent variables. Finally, for H1-H3, country-level affiliation of the scholars (SciVal) served as the independent variable, while in the case of H1a-H3a, the independent variable was the aggregated region-level scholarly affiliation (WE: UK, France, Germany, Italy, Spain; EE: Poland, Hungary, Russia, Romania, Ukraine; US: US).

It is important to note that the US appears both as a country- and a region-level affiliation in the analysis. On the one hand, this can be explained by the sheer size of the US academic population renders the country comparable for analysis to specific European regions. On the other hand, and more importantly, our analysis – being based on individual-level non-additive indicators – allows for comparisons to be made between different sample sizes (WE: 2500; EE: 2082; US: 500). In fact, the analysis design even enables performance differences of countries and regions to be investigated and contrasted to each other at the same time. The above argument – very importantly – also allows for countries that do not have 500 top-performing scholars listed in SciVal to enter the analysis. These occurrences indicate that between 2017 and 2020, the number of scholars in these countries publishing at least one Scopus-indexed was below 500 (i.e., Hungary: 409; Romania: 251; Ukraine: 423). Descriptive statistics of the countries under analysis and their corresponding values for our three dependent variables are reported in Table 1.

Table 1 Descriptive statistics of the three dependent variables across countries under analysis

Analysis strategy

Our study posed three hypotheses and three corresponding sub-hypotheses. A preliminary analysis of our three dependent variables (I1-I3) suggested that they were positively skewed amongst researchers. A follow-up analysis further indicated that they do not fit a normal distribution across countries and geographical regions (Shapiro–Wilk test; p = 0.000). Accordingly, we ran the non-parametric equivalent of a one-way ANOVA (i.e., the Kruskall-Wallis H test, also commonly known as one-way ANOVA on ranks). For H1a, H2a, and H3a, we considered the three impact measures (i.e., citations per document, views per document, citations per view) as dependent variables and the country of the most productive scholars as the independent variable. Similarly, for testing H1, H2, and H3, we considered geographical location (US, Western Europe, and Eastern Europe) as the independent variable and citations per document, views per document, and citations per view as the dependent variables. Pairwise comparisons were performed using Dunn’s (1964) procedure with a Bonferroni correction for multiple comparisons.

Results

H1 stated that there are statistically significant differences between citations per document across geographical regions. A Kruskal–Wallis test was implemented for Western Europe (n = 2,500), Eastern Europe (n = 2082), and the United States (n = 500). Distributions of I1 scores were similar for all geographical regions, as assessed by visual inspection of a boxplot. Medians for I1 scores were significantly different between countries, χ2 (2) = 685.303, p = 0.000. The post hoc analysis for citations per document revealed statistically significant differences between Easter Europe (Mdn = 1.000) and Western Europe (Mdn = 1.7500) (p = 0.000), between Eastern Europe and the United States (3.8571) (p = 0.000), and between Western Europe and the United States (p = 0.000). Pairwise comparisons are graphically represented in Fig. 1. H1 was supported.

Fig. 1
figure 1

Pairwise Comparison between geographical regions I1 (citations per document). Each node shows the sample average of I1. Orange lines represent statistically significant associations, while black lines represent non-significant ones

H2 stated that there are statistically significant differences between views per document across geographical regions. Distributions of I2 scores were similar for all geographical regions, as assessed by visual inspection of a boxplot. Medians for I2 scores were statistically significantly different between countries, χ2 (2) = 84.853, p = 0.000. The post hoc analysis for views per document revealed statistically significant differences between Eastern Europe (Mdn = 12.0000), the United States (Mdn = 16.5714) (p = 0.000), and between Western Europe (Mdn = 12.3333) and the United States (p = 0.000), but not between Eastern Europe and Western Europe (p = 0.356). Pairwise comparisons are graphically represented in Fig. 2. H2 was supported.

Fig. 2
figure 2

Pairwise Comparison between geographical regions I2 (views per document). Each node shows the sample average of I2. Orange lines represent statistically significant associations, while black lines represent non-significant ones

H3 stated that there are statistically significant differences between citations per view across geographical regions. Distributions of I3 scores were similar for all geographical regions, as assessed by visual inspection of a boxplot. Medians for I3 scores were statistically significantly different between countries, χ2 (2) = 706.822, p = 0.000. The post hoc analysis for citations per view revealed statistically significant differences between Eastern Europe (Mdn = 0.0345) and Western Europe (Mdn = 0.1304) (p = 0.000), between Eastern Europe and the United States (Mdn = 0.2304) (p = 0.000), and between Western Europe and the United States (p = 0.000). Pairwise comparisons are graphically represented in Fig. 3. H3 was supported.

Fig. 3
figure 3

Pairwise Comparison between geographical regions I3 (citations per view). Each node shows the sample average of I3. Orange lines represent statistically significant associations, while black lines represent non-significant ones

H1a stated that there are statistically significant differences between citations per document across country-level affiliations of the most cited scholars. A Kruskal–Wallis test was implemented for the United Kingdom (n = 500), France (n = 500), Germany (n = 500), Italy (n = 500), Spain (n = 500), Poland (n = 500), Hungary (n = 408), Russia (n = 500), Romania (n = 251), and the United States (n = 500). Distributions of I1 scores were similar for all countries, as assessed by visual inspection of a boxplot. Medians for I1 scores were statistically significantly different between countries, χ2 (10) = 845.423, p = 0.000. The post hoc analysis for citations per document is reported in Table 2 and graphically represented in Fig. 4.

Table 2 Differences between countries in I1 (citations per document)
Fig. 4
figure 4

Pairwise Comparison between countries for I1 (citations per document). Each node shows the sample average of I1. Orange lines represent statistically significant associations, while black lines represent non-significant ones

H2a stated that there are statistically significant differences between views per document across country-level affiliations of the most cited scholars. Distributions of I2 scores were similar for all countries, as assessed by visual inspection of a boxplot. Medians for I2 scores were statistically significantly different between countries, χ2(10) = 734.375, p = 0.000. The post hoc analysis for views per document is reported in Table 3 and graphically represented in Fig. 5.

Table 3 Differences between countries in I2 (views per document)
Fig. 5
figure 5

Pairwise Comparison between countries for I2 (views per document). Each node shows the sample average of I2. Orange lines represent statistically significant associations, while black lines represent non-significant ones

H3a stated that there are statistically significant differences between citations per view across country-level affiliations of the most cited scholars. Distributions of I3 scores were similar for all countries, as assessed by visual inspection of a boxplot. Medians for I3 scores were statistically significantly different between countries χ2 (10) = 877.709, p = 0.000. The post hoc analysis for citations per view is reported in Table 4 and graphically represented in Fig. 6.

Table 4 Differences between countries in I3 (citations per view)
Fig. 6
figure 6

Pairwise Comparison between countries for I3 (citations per view). Each node shows the sample average of I3. Orange lines represent statistically significant associations, while black lines represent non-significant ones

Discussion

Prior to this research, the scholarly community already had a good picture of the geopolitically unequal distribution of scientific impact in Communication Studies, as shown through numerous analyses of the numbers and distribution of published documents and received citations (Demeter, 2017, 2019b; Lauf, 2005). In line and in broad agreement with studies comparing groups of knowledge-producing agents via citations per document from the past two decades, our results confirm that a statistically significant difference exists between the citations of top scholars in Communication studies from the United States, Western Europe, and Eastern Europe. Our findings provide three interrelated contributions within this line of inquiry.

First, through a new empirical analysis, this study shows that the top US scholars are the most highly cited, followed by their Western and Eastern European peers, while country-level comparisons revealed that the top US scholars have by far the highest impact per document. Adding to our understanding of a significant Americanization of the field discussed in other studies (Chakravartty et al., 2018; Demeter et al., 2022a, 2022b; Gunaratne, 2010; Waisbord & Mellado, 2014), our results direct the attention toward the need of a further de-Westernization in communication studies.

In Western and Eastern Europe, we found insufficient evidence that the impact of scholars is significantly different based on their region. The impact of top-performing scholars from Romania and Hungary is not that different from those of France (in our sample, the country with the weakest median citation value from the Western European region). While having a similar impact as their French colleagues, we also found insufficient evidence for top Ukrainian performers being that far behind their Spanish and Italian counterparts. Ukraine, with one of the highest median citation values among Eastern European countries, also differs from other members of the Eastern European region analyzed here. Germany and the UK – top impact countries from the Western European region – are similar to each other, but while data from the UK shows very strong evidence for them being different from every other country in their respective region, the evidence for Germany being different from Spain or Italy is weak.

Second, the number of citations and published papers are common measures of spatial scientometrics and have a long tradition of measuring performance in Communication studies, including the more recent ones of those of the top-performing scholars (Demeter et al., 2022c; Goyanes, 2022). Our study, while making use of these more common metrics, also focused on so far unaddressed metrics, views received to articles written by top-performing scholars within a major abstracting and indexing service, compared to the number of publications and citations they have. Focusing on this metric enabled us to suggest a description of inequalities less distorted by self-marketing (either from the part of the author, their institution, or the journal they published in) or extra epistemic values embedded in (or to be harvested from) the published research. We aimed at mitigating the effects of an increased digital presence to visibility, trying to tie this metric more tightly to its appeal for an actual scientific community using SCOPUS. For the number of internal view counts to increase, the paper has to be visited through the SCOPUS interface as a result of browsing or searching the database, an activity typically performed by researchers and, to a lesser extent, librarians and science administrators. To summarize, this metric is coming from the activity of a more tightly targeted audience; therefore, future analyses using this metric can potentially describe geopolitical biases inside academia with a better relevant resolution than traditional altmetrics.

The question remains that how these newly developed formulas would fair in different research areas. This paper only used them for analyzing communication studies, and their measures may be weaker for other fields with different characteristics. It can be argued that the new metrics would work better with research areas where the consensus of the scientific community is stronger on excellent publications being generally indexed in SCOPUS (or Web of Science if examining “usage counts” instead of “view counts”) and that a researcher should aim at publishing mainly in indexed and highly ranked venues. Since coverage for book-oriented disciplines in the two main indexing database is weaker compared to article-oriented disciplines, and it takes significantly more time to produce and publish a book than a journal article or conference paper, it is not expected that the new metrics would work as well with research fields where monographs are significantly more important for a researcher than with those where journal articles or conference proceedings are. This is something to consider in the future when one would want to use these metrics in their analyses of the humanities (and even then; for example analytical philosophy or archaeology could be a better fit than classical literature or history). There are also major differences in the ease of access to SCOPUS around the globe, some linked to financial constraints, some to research cultures or current events (see f.e. the shrinking access to full-text international journals in Russia). Interest from a scientific community without the financial or institutional background necessary for SCOPUS subscription will not necessarily manifest in view counts when researchers browse Sci-Hub or similar shadow libraries instead. Overal, it is plausible that the newly introduced metrics could work better or worst with areas of different pubication characteristics than communication studies, and applicability would be influenced by access characteristics of different world regions to commercial abstracting and indexing databases. However, tools for accurately weighing these factors when including the "citation per view" indicator in the analysis are yet to be developed and tested.

Third, looking at the view count calculated from SCOPUS for articles written by SCIVAL Top 500 communication scholars in the U.S., our analysis revealed statistically significant differences between Eastern Europe and the United States and between Western Europe and the United States. Notwithstanding, we found no significant difference between Eastern and Western Europe in the case of views per publication. It is immediately apparent that based on these internal view counts, Eastern European communication research is more accessed than used. If we compare the countries ranked by citations per view and citations per document side by side, both columns show approximately the same relative order, while the highest median values for views per article belong to Ukraine and Spain. The latter, same as Poland and Russia, is especially weak in converting views to actual citations. For reasons currently unknown, research published by Eastern European and Spanish scholars is less cited but visited more than or similar to those published by American or Western European scholars. These results are especially concerning as they indicate a strong Western bias in citation practices within the academic sphere. One possible explanation is that—although reading inclusively—scholars prefer to cite their peers of the same diaspore, and as there is an apparent overrepresentation of Western scholars in international publishing (Demeter, 2017, 2018a, 2020), this inevitably leads to a biased over-citation of the same groups of authors in top journals. Furthermore, as the quality of references—that is, the journal these were originally published in – is indicative of the quality of the citing paper, non-Western scholars are also called forth to prioritize these sources to be able to publish in leading journals. Further investigations into the socio-cultural factors behind these dynamics of Western scholars’ over-citation are therefore prompted.

Future research may explore geographical or institutional diversity in the source of view counts, which would be useful to refine our results further. Some countries like the U.S. give more weight to metrics drawn from Web of Science, while others like Poland and Hungary rely more on SCOPUS in their performance evaluations and promotion policies in Communication Studies. This may not only create large differences in what type of journal indexing researchers prioritize when sending papers for publication—constituting a limitation of this study in itself -, but may also affect Scopus view counts coming from specific countries. However, this data is proprietary to ELSEVIER and currently not accessible to the authors. The data could be used to show differences between regions and countries in SCOPUS access, and use and control our results for regional overrepresentations in view counts.

A qualitative analysis should also investigate how academics who published SCOPUS-indexed articles use the database in different regions and what decision-making patterns they follow when citing an article found in the database.