Introduction

Using online services to get more information about the quality of a job applicant or an employee is increasingly gaining popularity, as information gathered through platforms like, for example, Facebook or LinkedIn can be automatically processed without the need of investing personnel expenditure (Sattelberger 2015). Furthermore, the process suggests neutrality and fairness (Gapski 2015), implying an objective treatment of the individual, since every person is evaluated by the same success measures. Using a bibliometric measure to evaluate the performance of scientists, for example, is common practice (Nature 2017). One of the most frequently used of such metrics is the h-index (Ball 2007; Saleem 2011). It computes a value depending on the number of papers published by a scientist and their respective impact on other researches such that:

A scientist has index h, if h of his or her \(N_{{\text {p}}}\) papers have at least h citations each and the other (\(N_{{\text {p}}}-h\)) papers have \(\le h\) citations each (Hirsch 2005).

The validity of such a one-dimensional evaluation of scientists relies heavily on a complete list of all publications the scientist authored, as well as all publications which cite any of those. Since manually gathering such a list is impossible, the utilization of citation databases is inevitable. The most popular platforms offering the h-index are: AMiner (2017a), Google Scholar (2017), ResearchGate (2017), Scopus (2017) and Web of Science (2017). Searching for the h-index of multiple names on all of those platforms it can be seen that a number of inconsistencies between the different platforms like wrong assignments of works to an author (entity resolution errors), missing, doubled or false database entries and many more sources of error (MacRoberts and MacRoberts 1989) can lead to huge discrepancies in the resulting h-indices (and make it almost impossible to correctly estimate the exact value). Additionally, it is generally known that citation behaviour often varies greatly between different scientific disciplines. Various studies have already proven this, but the investigations are either directed at a single platform where the names and associated scientific disciplines can be extracted from the citation database itself (Batista et al. 2006), which limits the validity of the results due to the error-proneness of such an approach, or to multiple platforms with a very small set of names (Bar-Ilan 2008).

For promotion committees at a university looking up the h-index value of a potential employee on at least one of the platforms, it is important to know to what extent those errors may influence the results. Therefore, we compare the aforementioned five platforms offering the computation of the h-index, following four big research questions:

  • RQ1 (Differences between platforms): To what extent are there differences of h-index value distributions for a given sample between the platforms?

  • RQ2 (Difference between values for the individuals): For a given set of persons, how big are the individual discrepancies of h-index values?

  • RQ3 (Differences between scientific disciplines): How much do the h-indices for Nobel Prize winners differ depending on their scientific discipline on the various platforms?

  • RQ4 (Comparison to ground truth): Of what magnitude are the differences compared to the persons manually assessed h-indices?

To answer these questions, we first introduce the platforms in “The platforms” section. We then present the results of our study for which we gathered the h-indices on said platforms for different sets of names and determine the aggregated gaps in “Study I” section. Afterwards we evaluate how these differences are reflected in the h-indices of the nobel prize winners when we differentiate by scientific discipline (“Study II” section). Finally, we compute the reasonably accurate h-indices for a test group of 25 names by hand in order to compare them to the results delivered by the platforms and to inspect the respective deviations, which is presented in “Study III” section. In “Threats to validity” section, we will analyze potential threats to validity and in “General discussion” section, we will conclude the results in a general discussion.

The platforms

In this section, we discuss the platforms used in our studies and their individual properties. In Table 1 those properties are summarized and extended by additional common aspects. We are especially interested in the question of whether scientists create their own accounts or whether an account can also be automatically created by the system or by someone else.

AMiner (AM) started as a research project led by Dr. Jie Tang at Tsinghua University, China. Based on the comparably lower number of unique surnames in China than in other countries, one of its primal goals is to differentiate between multiple people with the same name, but at the same time it ignores the fact that a person may publish under several name variations (e.g. different name abbreviations) (Tang et al. 2008). The database is constructed by crawling a variety of different web-sources, which leads to automatically constructed profiles without a permission or notice. This procedure and additional manually constructed profiles may lead to duplicated and non-scientific entries. The platform also allows the user to manually correct mistakes and complete profiles. Manual modification does not require any form of authentication or validation and therefore allows easy manipulation of profiles.

Google Scholar (GS) is a free to use platform for scientific investigation provided by Alphabet. It screens websites for a certain kind of formatting and checks the indexing of online documents to decide whether they are scientific publications. Due to the susceptibility to errors of this way of data extraction, many unscientific contributions are listed on GS (Petersen et al. 2014). A profile has to be constructed manually, automatically constructed profiles are created only for deceased people such as Sigmund Freud or Albert Einstein.

ResearchGate (RG) is a social network for researchers and scientists, focusing on the person as central entity instead of their work. In general, accounts can be manually created by a scientist or automatically constructed. Only on manually created accounts, the h-index is shown. If such an account is abandoned later, the indexing algorithm automatically adds further papers and increases the h-index accordingly, though the respective person can configure a mandatory manual validation.

Scopus (SP) is a platform that offers a number of services which are strongly limited depending on whether the user has a purchased or free access. Since the free version provides all functionality important for this study, the following description focuses on the latter. The database is constructed by extraction of bibliometric information from a specific set of journals (on the platform this set is called content coverage) this set is publicly visible. In case of missing publications or mistakes, the automatically constructed profiles cannot be edited by the user, a time-consuming support system has to be used instead. Whether and to what extent user feedback will actually influence the database is unclear.

Web of Science (WoS) is a database set up in a similar way as SP, by screening publications from a limited set of journals. It has to be noted that the platform does not support profiles of any kind to provide bibliometric information of authors. Instead, it allows the dynamic construction of a so-called citation report, which contains any publication released by a person with the name searched for. From this report, falsely assigned publications can be excluded manually to correct the displayed h-index. This process leads to a comparatively low error rate. However, the strongly limited selection of journals considered results in rather low values (Piwowar 2013; Nature 1965).

All of the presented platforms lack transparency regarding the limited validity of the h-indices they provide. Even well-known basic aspects that have to be considered when it comes to the h-index, for example the incomparability of values between different scientific fields, are nowhere to be found, let alone explanations about less-known potential sources of errors. Though all of them list an explanation about how they gathered their data, they do not note that they cannot guarantee for complete publication coverage. Additionally, none of the platforms seems to consider a differentiation between scholarly articles and others beyond the method they use for crawling the data, at least there is no information available indicating such a differentiation. Last but not least the crucial information to what extent authors actually administrate their profile is not visible. Thus, we were interested in how much the h-indices would differ for a wide range of scientists. In the following, we will describe three different studies consisting of various lists of scientists for which we gathered their h-index, and the resulting variance of a person’s h-index over the different platforms.

Table 1 Important aspects of the platforms

Study I

Description

In order to find out the differences between the platforms as posed in RQ1 we collected several sets of names and developed a scraper visiting each platform once per name, extracting the respective h-index and saving it to a database for elaboration. Subsequently, the maximum deviation of the average h-index of individuals compared to the h-index found on the platforms was analyzed in order to address RQ2.

Method

Step 1 Setup sets of names: To encounter multiple possible threats to validity we chose four sets of names with different criteria to analyze as described in the following.

  • \({D}_{{\text {topGS}}}\) contains 1360 names of the researchers with an h-index of at least 100 according to the Webometrics Ranking of World Universities initiative of the Cybermetrics Lab research group in Spain (Webometrics 2017), who get their data from GS. Since some of the names are only available in an abbreviated form, we removed them to increase the chance of correct evaluation. After doing so, 1295 names were left.

  • \({D}_{{\text {topAM}}}\) contains 139 names of the researchers with an h-index of at least 100 according to AM (2017b).

  • \({D}_{{\text {nobel}}}\) contains the 632 names of the researchers who won a Nobel Prize in chemistry, physics, medicine or economics (Nobel 2017).

  • \({D}_{{\text {TUKL}}}\) contains the 56 names of the current docents of a technical department at a German university.

Step 2 Collect h-indices: Most of the platforms do not offer an API to allow easy automatic access. As a consequence, the scraper visits the respective websites via browser by using the Selenium testing environment (Razak and Fahrurazi 2011) to search each name from a sequentially loaded subset of names and extracts the returning h-indices for the first result found respectively, to simulate an employers behavior. The h-indices are stored in an Elasticsearch database, which allows easy access, modification and visualization via Kibana.

Step 3 Refine the results: Since the crawling procedure is prone to failure due to occasional loading errors, delays and further problems beyond our control, some subsets delivered very little results. To avoid a lack of applicable data those subsets were reprocessed. Names with accented or special characters that might not be processed correctly have been excluded from the study as well.Footnote 1

Step 4 Evaluation: Sometimes, a name cannot be found on all platforms discussed. Therefore we restrict the database for our analysis to those names which were found on all platforms (see Table 2). The found h-indices are split by platform and visualized as box plots for each dataset, respectively.

Table 2 The number of names in each dataset, for which an h-index on all platforms could be found

Results

The box plots in Fig. 1 unveil notable differences between the platforms for each set of names. The key values for \({D}_{{\text {topGS}}}\) as presented in Table 3 show that the average h-index differs up to a factor of 3 between AM and GS, the median between RG and GS even by a factor of 12. The Inter Quartile Ranges (IQR) for the values from AM and RG are at least twice as high of those for GS, SP and WOS (see Fig. 1 and Table 3).

Except for AM, the plots for \({D}_{{\text {topAM}}}\) look very similar (see Fig. 2 and Table 4).

For \({D}_{{\text {nobel}}}\), the distribution between the plots look much like those for \({D}_{{\text {topGS}}}\) except that they are lower (see Fig. 3 and Table 5). This could be explained by the different criteria for the selection of scientists (citation based on \({D}_{{\text {topGS}}}\) vs. Nobel Prize based on \({D}_{{\text {nobel}}}\)).

\({D}_{{\text {TUKL}}}\) focuses on scientists from only one department of a university. Accordingly, the resulting values are lower and lead to different distributions. While h-indices from GS are still greater than the ones from other platforms, AM, RG and SP yield similar values compared to each other (see Fig. 4 and Table 6).

To inspect the worst-case impact of the findings for individuals, we examined the maximum and minimum h-indices from any of the platforms for each of the 1052 names in all datasets sorted by their average h-index on all five platforms (see Fig. 5). It turns out that the discrepancies between maximum and minimum are considerably high.

Fig. 1
figure 1

Box plots of the h-indices for the names in \({D}_{{\text {topGS}}}\)

Table 3 Key values describing the results of \({D}_{{\text {topGS}}}\)
Table 4 Key values describing the results of \({D}_{{\text {topAM}}}\)
Table 5 Key values describing the results of \({D}_{{\text {nobel}}}\)
Table 6 Key values describing the results of \({D}_{{\text {TUKL}}}\)

Discussion

Scientific evidence that different platforms yield different results when it comes to the h-index values and therefore address RQ1 has already been provided by others (Falagas et al. 2008). Consequently, we did not expect to find very similar h-index values, however, the extent of differences was considerably higher than expected. Especially the h-index values from GS generally seem to be higher than on the other platforms. The values found for \({D}_{{\text {topAM}}}\) are approximately as high for AM as for RG. This is due to the fact that the dataset focuses on names with a high h-index on AM itself. WOS is considered a well-known and well-used tool for scientific literature research, but due to its relatively small coverage of journals (Reuters 2008), the gathered h-indices are comparably low.

Fig. 2
figure 2

Box plots of the h-indices for the names in \({D}_{{\text {topAM}}}\)

Fig. 3
figure 3

Box plots of the h-indices for the names in \({D}_{{\text {nobel}}}\)

Fig. 4
figure 4

Box plots of the h-indices for the names in \({D}_{{\text {TUKL}}}\)

The full scope of the individual discrepancies of h-index values (RQ2) becomes apparent when considering the individual maximum and minimum h-index value that can be found on any of the five platforms. Naturally this discrepancy is somehow related to the actual h-index, which however can only be calculated with great effort. Therefore we have chosen the average h-index of a person on all five platforms as x-axis and plotted the minimum and maximum against it (see Fig. 5). The results clearly show that the discrepancies between the h-indices on the different platforms are enormous even for scientists with a small h-index. The potential harm of consulting the h-index on the wrong platform is considerably high, since an h-index that is too small can have a negative impact on a scientist’s career, whereas an h-index that is too large can lead to unfair competition and thus to an advantage for scientists who deserve it less than their competitors. In the area of individual observation, further studies in quantitative terms do not exist yet. Thus, our results harden the impression that the individual database issues have a higher negative impact than assumed by studies examining small datasets like Bar-Ilan (2008).

Fig. 5
figure 5

Average h-index against the respective maximum (blue) and minimum (red) h-index on all five platforms for every person in the datasets. (Color figure online)

Since a comparison with the approximately correct real h-indices disambiguates the magnitude of differences even further, we elaborate them for the scientists in dataset \({D}_{{\text {TUKL}}}\) in study III.

Study II

Description

With the help of data set \({D}_{{\text {nobel}}}\), an analysis of the importance of the various scientific disciplines of the Nobel laureates for their h-index (RQ3) was carried out on several platforms. Since only 89 names are too few for a meaningful analysis, we investigated which circumstances have led to such a reduction from previously 632 names.

Method

Step 1 Reevaluation of Study I: To investigate why only 89 names remained, we examined the refined results from Study I with regard to how many results could be found on each platforms. It turns out that on Google Scholar noticeably less names could be found than on the others, therefore Google Scholar is excluded from this study. 289 names remain for which results on the other four platforms could be found. Step 2 Division into scientific disciplines: The results found are split into the four Nobel Prize categories considered. Of the 289 remaining laureates 78 have a Nobel Price in chemistry, 90 in medicine, 71 in physics and 50 in economics. People who have won a Nobel Prize in two different categoriesFootnote 2 are listed once per category.

Results

The results for the fields of chemistry and medicine are very similar to the overall results for \({D}_{{\text {nobel}}}\), while those for physics and economics are significantly lower (see Figure 6).

Discussion

The finding corresponds to the results of comparable analyses (like for example Iglesias and Pecharromán (2007)), but here, too, the significance of the data must be critically assessed, since the data set is limited to the names of Nobel laureates who do not necessarily have an average publication behaviour. Additionally, winning a Nobel Prize necessarily results in an increase of popularity of the scientist, which could trigger atypical citation rates for the respective scientific discipline.

Fig. 6
figure 6

Box plots of the h-indices for the Nobel laureates in chemistry, medicine, physics and economics

Discussion

Study III

Method

Of great interest is the question of how well the platforms approximate the truth, i.e., the real h-index of a person (RQ4). Of course, assessing this real h-index is a problem on its own: The person needs to know all her publications and all references made to them from all valid, scientific documents. To tackle this question, we collaborated with 27 out of the 32 people in \({D}_{{\text {TUKL}}}\) to evaluate their h-index by first using the Publish or Perish tool with subsequent assessment by hand. The Publish or Perish tool by Anne Harzing (2007) is based on Google scholar which tends to assign too many papers to persons rather than too few. By going through the list of publications assigned by the tool to the scientists, we deleted those that clearly did not belong to the scientists. The process led to applicable results for 25 scientists.Footnote 3 It is still possible that publications or citations were missing. Thus, —even with an assessment by hand—the h-index might still not be the real one, however, as it was assessed together with the scientists we assume that the deviation should not exceed 2–5 points.

Results

The results are sorted by the manually assessed h-index and displayed in Fig. 7. Table 7 lists the exact values and shows the sample standard deviation \(d_p\) for each platform p computed by

$$\begin{aligned} d_p=\sqrt{\frac{\sum ^N_{i=1}(h_{p,s_i}-\overline{h}_{s_i})^2}{N-1}} \end{aligned}$$
(1)

where N represents the number of scientists, \(h_{p,s_i}\) represents the h-index of scientist \(s_i\) on platform p and \(\overline{h}_{s_i}\) represents the respective manually assessed h-index. To investigate the extent of differences even further, we also compared the manually assessed h-index for each scientist with the respective minimum and maximum value among all platforms and the sample standard deviation \(d_s\) for each scientist s computed by

$$\begin{aligned} d_s=\sqrt{\frac{\sum ^N_{i=1}(h_{s,p_i}-\overline{h}_{s_i})^2}{N-1}} ,\end{aligned}$$
(2)

where N represents the number of platforms, \(h_{s,p_i}\) represents the h-index of scientist s on platform \(p_i\) and \(\overline{h}_{s_i}\) represents the respective manually assessed h-index.

Fig. 7
figure 7

Plotted results of the manually assessed h-index and the corresponding h-indices on the five platforms for the 25 names from \({D}_{{\text {TUKL}}}\)

Discussion

As it can be seen, all platforms except for GS yield a too low h-index value in most cases, while GS overestimates it (see Fig. 7). This observation aligns with the fact that GS provides the most comprehensive database (Jordan 2014; Google Scholar 2017), including multiplied entries and unscientific publications. The frequently correct value estimations might be an indication for accurately maintained profiles on the platform, which is a common thing to do as a university lecturer in Germany. On every platform at least one of the researchers receives a too high h-index value and another a too low value compared to their manually assessed h-index values (see Table 7). Thus, it cannot generally be said that one platform always delivers higher or lower results than expected. Additionally, there is at least one extreme outlier on each platform (GS and SP positive, AM, RG and WOS negative) for this dataset (see Table 8) making it inevitable to properly inspect the results of a name to ensure a minimal deviation.

Table 7 Results of the manually assessed h-index and the corresponding h-indices on the five platforms for the 25 names from \({D}_{{\text {TUKL}}}\) and the sample standard deviation for each platform as described in Eq. 1
Table 8 Manually assessed h-index with the respective minimum and maximum value among all platforms and the sample standard deviation for each of the 25 names from \({D}_{{\text {TUKL}}}\) as described in Eq. 2

Threats to validity

To be able to assess the quality of the results and the limits of their interpretation, we follow a schematic guideline by Drost (2011). This guideline mentions various threats to validity that can be checked in order to gain a better insight into the interpretability of empirical results.

Internal validity: Obviously, many names have not been found on all five platforms, which were therefore removed from the study. This applies especially to the names from \({D}_{{\text {Nobel}}}\), which might lead to biased results. Thus, we have based our findings on the aggregated results of all datasets. Additionally, taking h-indices computed by hand only from \({D}_{{\text {TUKL}}}\) as reference might be biased to a certain extent, which is why those can explicitly be taken solely as first evidence.

Construct validity: Another issue is the way information has been extracted from the platforms. The self-programmed crawler is designed to enter the names and extract the results based on dynamically rendered HTML-structures. Furthermore, for the first study, we always took the first result on that platform with that name while some of them give a selection menu of all persons with the same name. For an automatic tool, it is not possible to choose the most likely correct name.

External validity: Our conclusions are based on 1052 names from a variety of four different subgroups, which is why the generalizability to all scientists of the findings might be debatable. Since the results between each subgroup do not vary much, this problem should not be an issue here.

Overall, the list of possible biases is small and most of the items are inevitable for an automated study.

General discussion

Our results raise the concern, whether promotion committees at universities should use any of the tested platforms to obtain the h-index of a candidate.

In direct comparison between the platforms, study I shows that the h-index values yield massive differences (RQ1). Within all datasets, major discrepancies between the respective average h-indices as well as the interquartile ranges can be observed. These large-scale results strongly substantiate the findings of the small-scale studies mentioned in the introduction and broaden the results of the more specific, disciplinary studies (e.g. Schreiber 2007; Engqvist and Frommen 2008).

The quantitative approach in the second part of study I minimizes the potential impact of random fluctuations and leaves overwhelming evidence that it is impossible to rely on the results of the h-index provided by only one platform. It validates the previously observed alarming state of the art and indicates a change for the worse.

Study II showed that the ground truth h-index of 25 scientists (manually assessed) deviated strongly from most of their various online scores to an extent that is not tolerable.

Independent of the strong differences between the platforms, a society that uses such platforms must consider the error tolerance inherent in such automatic performance evaluation processes.

While working on this article we came across three major ethical problems which are caused by the current use of the h-index computed by various online platforms:

First of all, there is a significant problem in defining the input. Most platforms do not give clear information about their definition of a ’scientific article’. It is, however, obvious that the different platforms work on very different data bases of articles, based on their definition of that concept. Secondly, a lot of profiles of scientists on AMiner, Google Scholar, ResearchFate and Scopus are created without their consent or even knowledge. Thus, an individual usually does not know how many platforms collect information about himself or herself. If these are used in a job application, for example, an incomplete profile could lead to a rejection without the actual scientific output justifying it. So if a person does not even know that a platform has computed such important metrics, they should at least be informed about it, as well as given the opportunity to correct errors and add data to their profile. Unfortunately, this possibility is often only accessible via a complicated route.

Last but not least, the automated assignment of scientific articles to the right scientists poses a big challenge to a computer. This problem is known as the entity recognition problem, i.e. the attempt to identify the correct entity from properties, in this example from a paper author to the correct individual (O’Neil 2016).

This ties in with a broad ethical discussion about the use of algorithms in various social processes (Lischka and Klingel 2017) and the question when and how algorithms must be made accountable for society (Diakopoulos 2014). As an overview of the responsibilities and sources of error when dealing with algorithms, we refer to Zweig (2018).

Finally, we conclude that as long as the variation between the most commonly used systems to automatically evaluate scientific performance and the real values is as high as measured for this article the systems are neither fair nor neutral and should not be used to assess academic job candidates.