Percentage-based Author Contribution Index: a universal measure of author contribution to scientific articles
- 4.7k Downloads
Deciphering the amount of work provided by different co-authors of a scientific paper has been a recurrent problem in science. Despite the myriad of metrics available, the scientific community still largely relies on the position in the list of authors to evaluate contributions, a metric that attributes subjective and unfounded credit to co-authors. We propose an easy to apply, universally comparable and fair metric to measure and report co-authors contribution in the scientific literature.
The proposed Author Contribution Index (ACI) is based on contribution percentages provided by the authors, preferably at the time of submission. Researchers can use ACI to compare the contributions of different authors, describe the contribution profile of a particular researcher or analyse how contribution changes through time. We provide such an analysis based on contribution percentages provided by 97 scientists from the field of ecology who voluntarily responded to an online anonymous survey.
ACI is simple to understand and to implement because it is based solely on percentage contributions and the number of co-authors. It provides a continuous score that reflects the contribution of one author as compared to the average contribution of all other authors. For example, ACI(i) = 3, means that author i contributed three times more than what the other authors contributed on average. Our analysis comprised 836 papers published in 2014-2016 and revealed patterns of ACI values that relate to career advancement.
There are many examples of author contribution indices that have been proposed but none has really been adopted by scientific journals. Many of the proposed solutions are either too complicated, not accurate enough or not comparable across articles, authors and disciplines. The author contribution index presented here addresses these three major issues and has the potential to contribute to more transparency in the science literature. If adopted by scientific journals, it could provide job seekers, recruiters and evaluating bodies with a tool to gather information that is essential to them and cannot be easily and accurately obtained otherwise. We also suggest that scientists use the index regardless of whether it is implemented by journals or not.
KeywordsCo-authorship Author contribution Publication metric Scientific integrity
Deciphering the role and quantifying the amount of work provided by different co-authors of a particular paper has been a recurrent problem for the scientific community [1, 2, 3]. The position in the list of authors is commonly used to infer co-authors’ contribution and a number of systems have been proposed on this basis. They range from simple calculations based on the rank of the authors such as harmonic authorship credit, fractional authorship credit and inflated authorship  to more complex credits (e.g. ), some even taking into account the controversial journal impact factor . However, these metrics are essentially ‘one fits all’ approaches that assume the contribution of each author based on their position in the author list and attributes subjective and unfounded values to these positions. As such, they do not attempt to represent and quantify ‘true’ contribution. Despite the growing interest in resolving the issue of authorship contributions in scientific disciplines [3, 4, 6], no standard ranking system has been widely recognised or adopted by scientific journals. With this lack of consensus, some journals have implemented a compulsory or recommended section dedicated to reporting authors’ contribution.
A review of the top 150 ecology journals referenced in ISI Web Of Knowledge revealed that 13.3% of them require information on author contributions (Additional file 1). Authors are usually asked to briefly describe which task was conducted by which co-author. Although this information is valuable, it does not provide an objective, straightforward and universal measure of author contribution. For example, ‘data collection’ for a review article may in some cases simply involve searching a database by filtering papers using specific key words, while it may be a very time-consuming task in field ecology, and a highly technical task in computational ecology. So ‘data collection’ can mean very different things depending on the field of study or the type of paper. In addition, individual tasks are often conducted by multiple authors but there is no way of knowing whether and to what extent one author has contributed more to them. Although some authorship contribution systems propose graded contributions for each task (e.g. lead, equal, supporting role in the CRedIT system , or the three-tier criterion proposed by the International Committee of Medical Journal Editor ), the lack of a quantitative value means these systems lack accuracy and produce qualitative data that is challenging to analyse or compare. The second common limitation is the complexity of the proposed metrics, which deters authors from providing data in the first place. A third major issue is the lack of fairness where often the lead or corresponding author can unilaterally decide on the order of the co-authors and the description of their contribution.
To address these shortcomings, we propose an easy to apply, universally comparable and fair tool to measure and report author contribution.
A simple and accurate measure: percentage contributions
Percentages are straightforward and can be universally applied independent of research field, the number of co-authors or the nature of the paper (e.g. experimental, review, perspective, etc.). Because the authors of a paper are the best placed to make a judgement call about the value of each contribution, it is essential that percentage contributions are determined by authors rather than by a model based solely on the authors’ rank. Although disagreement may occur between co-authors, clarifying contribution among co-authors in the early stages of the research process is likely to ease potential tension  and in some cases prompt ‘real collaboration’. We propose that co-authors discuss and agree on their respective contributions prior to submitting their manuscript and these figures be provided by the corresponding author at the submission stage. By confirming their authorship, all co- authors confirm their agreement with their contributions and that of all other authors. This ensures that every published paper displays percentage contributions that have been discussed and agreed upon by every co-author.
A number of guidelines and best practices have been proposed for authors’ contributions (e.g. ). A possible starting point is to divide 100% by the number of authors and then estimate whether and to what extent each author provided more or less work than the others. The use of author-provided percentages has been proposed before to reflect the contribution of co-authors accurately (e.g. ) but with limited guidance about how to implement it. Verhagen et al.  developed the Quantitative Uniform Authorship Declaration (QUAD) approach, where each author is attributed percentage contributions in four categories: Conception and design, data collection, data analysis and conclusion and manuscript preparation. More recently, a very similar approach was proposed, based on scores rather than percentages with the more specific aim of deciding which contributor deserves authorship and which does not . Clement  also suggests the use of four categories, albeit slightly different ones (ideas, work, writing and stewardship). However, an overly complicated metric may deter authors from applying it, and the criteria used in calculations must be consistent or have comparable importance across research fields, which may not be applicable to every type of article. In addition, many authors suggest that contributions should be restricted to an arbitrary threshold, for example 50% of the average contribution , 10% of the total work  or a threshold chosen by the authors . Such limitation is likely to introduce major inconsistencies between papers, journals and fields of research, thereby preventing comparison. In addition, these thresholds limit the number of co-authors, which may affect interdisciplinary research and act as incentives to leave out minor contributors, potentially increasing ghost authorship (i.e. the omission of collaborators who did contribute to the work).
A universally comparable metric: percentage-based author contribution index (ACI)
where for author i,
Ci = contribution of author i in percentage (must be > 0 and < 1)
n = total number of authors including i (must be > 1).
ACI(i) reflects the contribution of author i as compared to the average contribution of all other authors. It is superior to one when the contribution of author i is larger than the average contribution of all other authors and inferior to one when the contribution of author i is less than the average contribution of all other authors. For example, on a paper written by three authors, where author i contributed 60% of the paper, ACI(i) = 3, meaning that author i contributed three times more than what the other authors contributed on average. Another useful metric is log10(ACI), which is positive when the author’s contribution is larger than the average contribution of all other authors and negative when the author’s contribution is less than the average. This metric is particularly useful to normalise data for further comparison and statistical analyses.
A fair tool: assisting job seekers, recruiters and performance-based evaluations
The scientific community seems to have reached the consensus that journal impact factors are not an accurate measure of the value of a particular article or the value of its author(s) . One of the main reasons is that a very highly ranked journal may publish few articles that are heavily cited, but it may also publish a number of papers that have very little impact. In recent years, article-based impact has been preferred to journal impact factor. For example, the Hirsch index (h-index), which is based on the number of citations of one’s papers, is now widely used to gauge the output of a scientist. However, the h-index can also be manipulated  and it does not provide a measure of the amount of work produced by each co-author, which means guest or honorary authorship (i.e. inclusion of authors who did not contribute to the work) cannot be accounted for. This issue of guest authorship has been denounced widely in medical and clinical science , but other research fields are not immune to the problem (e.g. in ecology , environmental science , geography , geology , etc.). Since genuine authors suffer no cost when they add co-authors, papers tend to have more and more co-authors [21, 22]. With percentage contributions, the amount of work invested in a paper is a finite value (100%). Therefore, when more authors are added as a ‘gift’, they all need to be attributed a percentage of the work. In this zero-sum game, either it will be visible that guest authors have contributed an extremely small proportion of the work—and should receive very little recognition—or the genuine authors will have to give away large chunks of their well-deserved credit.
To ensure fair values are reported, we propose that co-authors discuss and agree on their respective contributions prior to submitting their manuscript and these figures be provided by the corresponding author at the submission stage. By confirming authorship, all co-authors confirm their agreement with their contributions and that of all other authors. This ensures that every published paper displays percentage contributions that have been discussed and agreed upon by every co-author.
The sentence-based descriptions of authors’ contribution that are used by some journals provide an indication of tasks performed by each author. Maintaining and generalising this practice as a detailed record of the role of each author is essential to increasing transparency. However, obtaining a clear idea of the amount of work a scientist is actually providing is difficult if one needs to read through all the authors’ contribution sections and weigh in the topic, the type of paper, the number of co-authors, etc. The proposed index has the potential to complement descriptive authorship information sections with a quantitative measure, which is much easier to analyse, summarise and compare across a large number of publications. For example, ACI can help in sifting through the numerous papers published by one scientist to select only those where this particular author has made a major contribution. This short list of papers could then be analysed in more details, for example using the written author contribution sections.
ACI can provide valuable information for performance-based evaluation processes and could be implemented in existing reporting systems. This includes internal evaluation for career advancement, as well as research productivity evaluation for funding purposes and national-scale ranking schemes (such as the Performance-based Research Fund (PBRF) system in New Zealand or the Research Excellence Framework (REF) in the UK). It is also in the advantage of the candidate to be able to demonstrate his/her actual contribution to a potential recruiter who may ask ‘What have you done on all these papers listed on your CV?’ One could answer such a question by analysing the distribution of a scientist’s ACI and its evolution through time or by calculating and comparing his/her average ACIs in experimental, review and perspective papers. ACI could also be used as an additional metric in network-based collaboration analyses (e.g. ) or to further inform composite citation indicators (e.g. ).
A first look: testing ACI
Here, we provide an analysis of ACI calculated from work published in the past 3 years (January 2014–December 2016) based on contribution percentages provided by scientists who volunteered to respond to an online survey (Additional file 2). This survey aims at (1) demonstrating that authors can provide percentage contribution for their work and are willing to do so, (2) demonstrating that there is an interest from the scientific community to provide more transparency in author contribution, (3) exemplifying basic calculations that can be easily performed using ACI values and (4) demonstrating that ACI can be used to understand the scientific production of scientists in relation to their career advancement. Because the contribution percentages were provided after publication and without discussion among co-authors, these values may not be as accurate as if they had been agreed upon by all co-authors prior to publication . Hence, the aim of this exercise was not to produce a highly accurate dataset, and therefore, the following analysis should be regarded as illustrative.
Respondents were asked to provide their percentage contributions for all the articles they co-authored in the past 3 years. Additional information collected was country of origin, time since research active and job description (Additional file 2). The URL address and information about the online survey were distributed through electronic mailing lists, ecological society newsletters and social media. Responses were collected between September 4, 2016, and January 8, 2017. During this timeframe, 97 ecology scientists from 19 different countries completed the survey. Respondents were anonymous and no filtering of respondents was performed other than removing incomplete entries. Data with contradictory or obviously inaccurate information (for example multi-authored publications where the respondent claims 100% of the work) were removed. The final dataset comprised data for 836 publications from 97 ecology researchers. Author contribution indices (ACIs) of the respondents were calculated for each publication using Eq. (1). ACIs were log-transformed (log10(ACI)) to meet the assumptions of normality and all statistical analyses were conducted in R . Respondents were asked to provide information about the number of years they have been research active. This was defined as the time from first year of PhD study or first published peer-reviewed paper, whichever came first. Linear regression and F statistics were used to analyse ACI in relation to the number of years as an active researcher.
Respondents were categorised in different job positions as follows: postgrad, a postgraduate student; postdoc, a postdoctoral fellow or other non-permanent staff; ECR, a tenure or permanent early-career researcher; PI, a mid-career principal investigator; prof, an associate professor of full professor; other. ACI was analysed in relation to job description using ANOVA with a priori contrasts between non-permanent supervised staff (postgraduate and postdoc) and permanent and independent researchers (ECR, PI and professors), as well as between all early career researchers (postgraduate, postdoc and ECR) and established researchers (mid-career PI and professors).
Results and discussion
In our dataset, ACI varies in relation to the number of years as a researcher (F1,835 = 87.57, p < 0.0001); however, the correlation remains weak (r = 0.306), due to wide variability in ACI (Fig. 4c). With regards to job descriptions, average ACI was higher for non-permanent supervised staff (postgraduate students and postdoctoral fellows) compared to permanent and independent researchers (early career researchers, mid-career principal investigators and professors) (ANOVA t = 2.327, p = 0.0229). A similar difference was observed when comparing all early career researchers (post-graduate students, postdoc and early career researchers) to established researchers (mid-career principal investigators and Professors) (ANOVA t = 2.546, p = 0.0132) (Fig. 4). As researchers become independent and establish their own research team, they probably start supervising their own students and postdocs and their ACI is likely to decrease accordingly.
It is possible to link ACI with article impact through the number of citations, altmetrics or any other article-based impact metrics, for example, by dividing ACI by the number of citations for a particular paper. By averaging ACIs, one could also summarise the output of a given scientist as a single number as suggested with other indices (e.g. ). However, we do not recommend such practice as it would largely mask the scientist’s output profile, thereby deceiving the purpose of ACI, and it would not be a meaningful way to compare scientist outputs as scientists with very different profiles may reach a very similar average ACI. In our dataset, there are numerous cases in which individuals at very different stages of their career reached a similar average ACI (Fig. 4, Additional file 3).
Limitations and alternative applications
The proposed model requires authors to score precisely their contribution but does not provide a framework for making decisions around the value brought by each co-author. The focus of ACI is on the reporting and dissemination of a quantitative, accumulative and comparable measure of their contributions, after authors have agreed on their percentage contributions. The way these contributions are calculated and agreed upon is likely to vary greatly from one paper to another and does not equate to ‘time spent working on the paper’. For example, tasks requiring specific skills or expertise that are essential to the realisation of the work may rank highly, despite not being the most time-consuming. This apparent subjectivity should, however, be put in perspective with that of existing ranking systems. In any ranking systems, the reader can only place faith in the authors’ good judgement and ethical practices. Authors are the best placed to judge contribution to their own work and must be given the freedom, flexibility and adaptability to choose the percentage contributions that they believe are the most appropriate on a case by case basis.
It is important to note that the issues of arbitrary decisions and unethical behaviour, which may often arise from power dynamics between co-authors (e.g. between senior and junior scientists), will not be immediately resolved by using ACI. However, because percentage contributions have to be ratified by every co-author, groups are compelled to collegially discuss and score their contribution prior to submission, something that usually does not happen. It is expected that the mere fact of openly discussing contribution will reduce unilateral decisions and help to obtain thoughtful and more accurate contribution scores. With increased transparency, both among co-authors and between authors and readers, ethically questionable behaviours may not disappear, but they will become less common and easier to detect and report.
Another crucial limitation of ACI lies in the fact that a critical mass of data is required for the proposed metric to reach its true potential. Yet, its implementation by a significant number of journals or publishers will take a long time, if it is adopted at all. To circumvent this limitation and also generate ACI data from past literature and papers published in journals that will not implement the index, we propose that reference list repositories such as ResearchGate (researchgate.net) or ORCID (orcid.org) could provide an option for authors to record their percentage contributions to their publications. Because these values may not be vetted by all co-authors (as opposed to percentage contributions provided at the time of submission), several levels of verification should be displayed for each paper. Percentage contributions could be considered as (1) unverified if only one co-author provides them, (2) partially verified if at least a second co-author confirms the numbers and (3) fully verified if all co-authors of a paper confirm the numbers. In addition, we suggest authors to provide percentage contributions in their mansucripts, even if these are not specifically requested by the journal they publish in.
There are many examples of author contribution indices that have been proposed but none has really been adopted by scientific journals. Many of the proposed solutions are either too complicated, not accurate enough or not comparable across articles, authors and disciplines. The author contribution index presented here addresses these three major issues and if adopted by scientific journals, it could significantly clarify the contribution of co-authors. This index is currently implemented in the recently launched journal Rethinking Ecology . We hope that ACI will be adopted by many other journals to increase transparency in co-authored work and attribute accurate credit to authors. Although the current paper uses ecology as the focus, the proposed index is readily applicable to other scientific fields. Ambiguity around co-author contribution is a systemic issue in scientific publication, and accurate reporting of author contribution would benefit a very large variety of scientists in all fields.
The current paper focuses on the reporting of one percentage contribution and one ACI per paper and per author. Another possibility is to provide percentage contributions (and calculate ACI) for a variety of specific tasks. This would provide even more detailed information by specifying ‘who did how much of what’ on a given paper. However, task-specific ACI would only be useful if the same set of tasks is systematically reported, as proposed in the QUAD or the CReDIT approach, to allow comparisons between papers and meta-analyses.
The proposed ACI index has the potential to contribute to more transparency in the science literature, and if adopted by scientists, it could provide job seekers, recruiters and evaluating bodies with a tool to gather information that is essential to them and cannot be easily and accurately obtained otherwise.
We thank the 97 ecologists who responded to the online survey on author contribution.
No funding was associated with the preparation of this manuscript. Publication fees were funded by CNRS.
Availability of data and materials
The data produced from the online survey is available on the Figshare server (DOI: https://doi.org/10.6084/m9.figshare.5139613).
SB designed the concept and drafted the manuscript. SB and TI prepared the figures. SB, TI, JMO and JS arranged the ethical approval and gathered data from scientists in their respective country. MCL reviewed the existing journals’ policy. SB, TI, JMO, JS and MCL contributed to the writing of the final version of the manuscript. Percentage contributions are SB: 55%, TI: 15%, MCL:10%, JMO: 10%, JS: 10%. All authors read and approved the final manuscript.
Ethics approval and consent to participate
Application for human ethics were granted by the Unitec Research Ethics Committee (UREC) (application number: 2016-1062) and the University of Georgia Institutional Review Board (IRB) (application number: STUDY00004056). Our online survey was anonymous and included the following sentence: ‘By answering you agree that the data you provide is used for the above-mentioned analysis and subsequent publication’.
Consent for publication
The authors declare that they have no competing interests.
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
- 3.Boyer S, Lefort M-C, Winder L. Rethinking ecology—challenging current thinking in ecological research. Rethink Ecol 2016;1:1–8.Google Scholar
- 6.Teixeira da Silva JA, Dobránszki J. Multiple authorship in scientific manuscripts: ethical challenges, ghost and guest/gift authorship, and the cultural/disciplinary perspective. Sci Eng Ethics. 2015;22(5):1457–72.Google Scholar
- 7.McNutt M, Bradford M, Drazen J, Hanson B, Howard B, Jamieson KH, et al. Transparency in authors’ contributions and responsibilities to promote integrity in scientific publication. bioRxiv. 2017. doi: 10.1101/140228, https://www.biorxiv.org/content/early/2017/05/18/138875.
- 9.Primack R, Cigliano J, Parsons C. Editorial: Co-authors gone bad—how to avoid publishing conflicts five examples of conflicts among co-authors. Biological Conservation. 2014;176:277-80.Google Scholar
- 14.Lariviere V, Kiermer V, MacCallum CJ, McNutt M, Patterson M, Pulverer B, et al. A simple proposal for the publication of journal citation distributions. bioRxiv. 2016. http://biorxiv.org/content/early/2016/09/11/062109.abstract. doi: 10.1101/062109.
- 15.Purvis A. The h index: playing the numbers game. In Proc. R. Soc. B (Vol. 205, pp. 581-598). Chicago. Trends Ecol Evol 2006;21:422.Google Scholar
- 16.Kovacs J. Honorary authorship epidemic in scholarly publications? How the current use of citation-based evaluative metrics make (pseudo)honorary authors from honest contributors of every multiauthor article. J Med Ethics. 2012;39(8):509–12.Google Scholar
- 22.Baethge C. Publish together or perish. Dtsch Arztebl Int. 2008;105:380–3.Google Scholar
- 24.Ioannidis JPA, Klavans R, Boyack KW. Multiple citation indicators and their composite across scientific disciplines. PLoS Biol. 2016;14:1–17.Google Scholar
Open AccessThis article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver (http://creativecommons.org/publicdomain/zero/1.0/) applies to the data made available in this article, unless otherwise stated.