These shortcomings are a matter of common knowledge, and it is all the more surprising now that Altmetric.com, one of the major altmetrics providers, has introduced its so-called “altmetric score”. This composite indicator can only be seen as somewhat paradoxical, since the altmetrics movement rather intends to overcome the flaws of citation-based indicators than to repeat and reinforce them. Certainly an “all-in-one” indicator is, whatever tempting, not desirable to accomplish this mission. Nevertheless the Altmetric.com doughnut with its central altmetric score (see Fig. 1) is a success story and is now being increasingly encountered in research communication channels.
According to the information provided on the Altmetric.com website, this score is a composite quantitative measure of the attention that a scholarly article has received, and it is based on three main factors:
-
1.
Volume: the score for an article rises as more people mention it (restriction: only one mention from each person per source is counted).
-
2.
Sources: these are categorized and each category of mention contributes differently to the final score (a table with the different weights is available).
-
3.
Authors: here it counts who and how often and to whom someone mentions something.
Combined, the score intends to be a weighted approximationFootnote 1 of all the attention Altmetric.com has picked up for a research output. Altmetric.com clearly states: “The score is useful when looking at several outputs together to quickly identify the level of online activity surrounding a particular research output—it is not a measure of the quality of the research, or the researcher”.
Everyone is familiar with the last part of this statement, which has been repeated like a mantra by bibliometricians concerning the use of their quantitative indicators, particularly the journal impact factor.
Granted, altmetrics are in their infancy, but some issues need to be addressed right from the start before this altmetric score is used in an irresponsible and harmful way for bibliometric purposes.
The first issue is transparency, because despite of providing some explanatory information the calculation of this indicator is far from transparent (e.g., rounding scores to integers, using score modifiers). Being a proprietary indicator, only Altmetric.com rules how weights are determined and how this score is calculated for each publication.
The quite different aspects of scholarly communication retrieved from the web are weighted according to rather random criteria not relying on scientific principles. Finally all aspects are mixed up, resulting in a score with questionable validity and significance. A single number may be convenient, but leaves the door wide open for abusive behaviour, since the multidimensionality and complexity of all the compiled information is suddenly wiped away. Neither normalization nor standardization of the compiled data seems to bother the provider of this metric. Moreover restrictions are not properly exposed. Simple questions arise like “What is a person?”, “Are persons equivalent to users?” “Do they need to register?” “Does Altmetric.com consider users or simply IP-addresses?” “How are data retrieved?” “How reliable are these data?”
Furthermore, there is a discrepancy between the traced data sources and the weighted ones for the calculation of the score. For example, data from reference managers are not considered. Admittedly reference managers are only responsible for captures, and captures and mentions are different expressions of measuring different things. Nevertheless it is not comprehensible to totally neglect captures in the total score if it really aims “to quickly identify the level of online activity surrounding a particular research output”. If that is truly the case, are captures not even more interesting than mentions for achieving this goal?
Data consistency is another weak spot. The data sources are not even clearly defined. For example, let us focus on data collected in Wikipedia. Which Wikipedia language edition is considered? Has an entry in the English language Wikipedia the same weight than an entry in any other language version?
The completeness of data is another issue. Many relevant sources are so far not included. Certainly Altmetric.com needs permission to trace data sources, and it is assumingly much easier to include Wikipedia than other encyclopaedias like Britannica. However, the incompleteness of data distorts the results and the significance of this measure. There is a bias towards the included sources, whereas missing sources are disadvantaged.
Another issue is the varying degree of altmetrics availability at publication level. It is incomprehensible for scientists why this information is sometimes provided, but lacking in other cases. Transparent information about when and where to expect altmetrics would be most desirable.
Last but not least, altmetrics data are clearly unstable and irreproducible. Even Altmetric.com has to admit this fact and indicates on their website that “from time to time you might notice that the score for your paper fluctuates, or goes down”. This would at least require thorough and transparent documentation, in order to trace and understand all score changes. However, such documentation would be bulky, difficult to maintain and is therefore not available.