Many fake papers are produced by paper mills: a successful business model

Scientists who publish fake results are nothing new. A well-known example of this is a study that was intended to prove a link between the mumps-measles-rubella vaccination and autism (Wakefield et al. 1998), which turned out to be fraudulent (Godlee et al. 2011). In recent years, however, the problem has worsened dramatically. Falsified scientific papers are no longer isolated incidents but contaminate the scientific record on a large scale. In the scientific community, these papers have become known as fake papers. They contain intentionally manipulated data or are completely fictitious. Large numbers of fake papers are produced by so-called paper mills, also referred to more bluntly as criminal science publishing gangs (Byrne and Christopher 2020; Christopher 2021; Else and Van Noorden 2021; Sabel and Seifert 2021; Seifert 2021a; Van Noorden 2023a). Paper mills offer fake papers from various scientific disciplines and advertise them online. People can then buy authorship of a paper of their interest. The price depends on the authorship position and the impact factor of the journal in which the fake paper will be published later. Paper mills also handle the entire publication process (Abalkina 2023). The services of these paper mills make it easy for people with little (or even no) scientific experience to become authors of fake papers, which in many cases are done well enough to pass peer review and get published (Else and Van Noorden 2021). Key to the commercial success of paper mills are misguided scientists who misuse their expertise to falsify science and commercially exploit dysfunctional academic evaluation systems.

How can I inform myself about fake papers? Unfortunately, much of the relevant information is not found in original peer-reviewed papers, but rather in other publication formats

This review provides an overview of important sources on the topic of fake papers and paper mills (Tables 1, 2, 3, 4, 5, 6, 7). These sources were categorized as follows: peer-reviewed original research (Table 1), reviews (Table 2), non-peer-reviewed preprints (Table 3), editorials (Table 4), journalistic perspective articles, news, views, and opinions (Table 5), blogs (Table 6), and further sources (Table 7). A related overview on this topic was published in 2021 (Van der Heyden 2021). Since then, there have been many new publications on this topic, so this review now provides an updated overview. However, this review does not claim to be complete and is limited to a deadline of June 2024.

Table 1 Original peer-reviewed papers dealing with “ truthl fake paper research”
Table 2 Reviews dealing with fake papers
Table 3 Non-peer-reviewed preprints dealing with fake papers
Table 4 Editorials dealing with fake papers
Table 5 Other article formats (including journalistic articles) dealing with fake papers
Table 6 Science blogs dealing with fake papers
Table 7 Further information sources on fake papers

Why people publish fake papers: from desperate doctors and established scientists seeking extra prestige

In the academic world, the number of publications is used as one benchmark to evaluate scientists (Abbott et al. 2010; Kumar et al. 2009; Rawat and Meena 2014; Walker et al. 2010). Particularly in China, but not exclusively, there are strong incentives to publish scientific papers, especially in high-impact factor journals. These incentives include jobs, tenure, promotion, grants, and even personal financial rewards (Abritis and Mccook 2017; Hvistendahl 2013; Quan et al. 2017; Tian et al. 2016; Zhang and Wang 2024). Without publications, promotion is often not possible in China, not even for medical doctors (Hvistendahl 2013; Tian et al. 2016). However, a lack of time and resources can make it difficult for many scientists to publish enough papers. And of course, for a medical doctor with full patient responsibility, the conduct of high-quality research is almost impossible. Unfortunately, under these circumstances, some people are tempted to publish fake papers. Data from real studies are then falsified, or entire studies are freely invented (Byrne and Christopher 2020; Else and Van Noorden 2021; Gopalakrishna et al. 2022; Miyakawa 2020; Seifert 2021a). The fact that many “authors” of fake papers use the expensive services of paper mills to do this shows two aspects: First, the pressure or incentive to publish is apparently so high that they accept the high costs of a paper mill (Abalkina 2023; Hvistendahl 2013). Second, these people are apparently so inexperienced with scientific publications or have so little time that they do not produce the fake papers themselves. It is also possible that the paper mills claim to use real research data (Hvistendahl 2013) and that paper mill customers therefore assume that the papers they buy are based on real studies.

Apart from the professional and financial benefits, there may be other reasons why someone would publish a fake paper. Diederik Stapel is a scientist who has been proven to have fabricated data in 58 of his papers (Palus 2015). He continued to fabricate study data long after his professional career was already considered glorious. He described his reasons for doing so as complex but also stated ambition and the desire to present excellent data in his papers (Bhattacharjee 2013). Eric Poehlman (Dahlberg and Mahler 2006) and Hwang Woo-suk (Wade and Sang-Hun 2006) are other well-known examples of scientists who have faked some of their work despite their advanced careers. These two also do not fit the narrative described above of desperate scientists or doctors who see no other option for their careers than to hire a paper mill. Hwang Woo-suk, for example, claimed the prestige of having cloned human embryonic stem cells (Wade and Sang-Hun 2006), so his ambitions presumably went beyond simply “surviving” in the scientific business. Thus, it seems that in individual cases, scientific prestige is also a motivation to fake results. In such cases, paper mills are not involved in the fake paper production.

How big is the fake paper problem? Estimates vary greatly and reflect methodological difficulties

It is hard to estimate the proportion of fake papers in the scientific record. Currently, different numbers are circulating. Two journals that had noticed submissions from paper mills estimated the proportion of such submissions at about 5–10% and 5%, respectively (Heck et al. 2021; Seifert 2021a). However, such estimates of individual journals are not representative, as paper mills attack certain journals more frequently (COPE & STM 2022). The proportion of submitted fake papers varied widely across disciplines and journals, with rates up to 46% (COPE & STM 2022). Another study found that about 0.01% of all scientific papers published in 2019 were retracted because they originated from a paper mill (Candal-Pedreira et al. 2022). Since by far not all fake papers are detected and retracted, it can be assumed that this proportion underestimates the actual amount of published fake papers. A news article reported that in 2023, more than 10,000 publications were retracted globally and across all disciplines, most of them for reasons indicating that they were fake papers. The article also shows a gradual increase in retraction rates over the last 20 years, which exceeded 0.2% in 2023 (Van Noorden 2023a). This increase may be due to both an increasing number of published fake papers and greater awareness of the problem. Another article reported on an unpublished study that found strong textual similarities with already known fake papers in 1.5 to 2% of all papers published in 2022. The suspicion was expressed that many of these papers could be fake papers too, but it was also pointed out that the actual proportion of fake papers cannot be determined without individual examination of these suspected cases (Van Noorden 2023b). In a screening of around 20,000 articles published in various journals between 1995 and 2014, problematic images were found in 3.8% of cases (Bik et al. 2016). Another study of more than 1300 randomly selected open-access articles published by PubMed Central in January 2014 found problematic images in 6% of cases (Bucci 2018). Many of the problematic images in these studies were intentionally fabricated, leading to suspicions that the publications in question might be fake papers. One study even estimated that 28.8% of all biomedical papers published in 2020 were potential fake papers. However, one of the three criteria used to classify publications as suspicious in this study was the use of non-institutional email addresses (Sabel et al. 2023). A different study concluded that this criterion is not suitable for identifying fake papers (Wittau and Seifert 2023), so the 28.8% is most likely a considerable overestimate. The same applies to another version of this study, in which the authors reduced their estimate to 11% but retained the methodology (Sabel et al. 2023). In a study focusing on rejected and withdrawn papers in Naunyn-Schmiedebergs Arch Pharmacol that were finally published elsewhere, the percentage of verified fake papers was below 1% (Wittau et al. 2023).

The fairy tale that high-impact journals are not affected by fake papers

In the scientific community, it is often assumed that fake papers are only a problem of low-medium scientific journals, but this is simply not true. A recent systematic analysis revealed that the problem is very pertinent in high-impact factor oncology journals (Pathmendra et al. 2024). These links list articles that have been retracted by Nature (https://www.nature.com/nature/articles?type=retraction; accessed June 17, 2024) and Science (https://pubmed.ncbi.nlm.nih.gov/?term=%28%22Science+%28New+York%2C+N.Y.%29%22%5BJournal%5D%29+AND+%28%22Retraction+of+Publication%22%5Bpt%5D%29&size=200; accessed June 17, 2024). Many of the retractions occurred due to honest errors, but fraud was also a recurring reason. These cases show that even the most respected and prestigious journals, having access to the best reviewers globally, can be effectively fooled by fake paper authors. In the case of this top category of journals, the desire of editors to publish the most impactful, sensational, and attention-drawing papers may, at least in some cases, impede the objective assessment of papers.

The damage that fake papers produce in science

As the extent of the fake paper problem is not clear, there are no studies to date that capture its impact. This section is therefore speculative and only lists a few probable consequences and examples.

Contamination of the scientific record by fake papers undermines the reliability of scientific studies. This particularly affects scientists who base their studies on the published results of other scientists. Fake papers likely contribute to the reproducibility crisis (Miyakawa 2020) and therefore lead to a waste of resources in science and industry. A well-known example of this can be found in Alzheimer’s research. One of the most highly regarded Alzheimer’s studies (Lesné et al. 2006), which has helped shape the direction of Alzheimer’s research and treatment development in recent years, has been suspected of being falsified for some time. If this suspicion is confirmed, this paper could have contributed to a waste of millions or even billions of dollars and the efforts of numerous people (Piller 2022). Another famous example can be found in research on room-temperature superconductors. Three papers that had caused a sensation in this research area have been recently retracted (Dasenbrock-Gammon et al. 2023; Durkee et al. 2021; Snider et al. 2020). The reasons for the retractions were doubts about the integrity of the data (Castelvecchi 2023; Garisto 2023). There are no numbers on how much money and time other scientists have wasted because of these studies. However, since the three studies together were cited very often (Dasenbrock-Gammon et al. 2023; Durkee et al. 2021; Snider et al. 2020) and there were various attempts to reproduce the results (Castelvecchi 2023; Garisto 2023), it can be assumed that they led to a substantial waste of resources. Of course, these are two extreme examples of the consequences scientific fraud can have. Most fake papers will presumably be much less influential. It is also conceivable that fake papers not only lead to unjustified research efforts in a particular area but also discourage honest scientists from conducting research in areas where results already appear to be available.

In addition to the harm to science and the economy, fake papers could also have a direct impact on society. If scientific studies lose credibility in the public eye, people who deny scientific evidence may feel confirmed, and public discourse on important social issues such as climate change or vaccination may be damaged (Hopf et al. 2019).

Medical fake papers could even lead to negative consequences for patients (Van Noorden 2023c). One example of this is a case in which a researcher is alleged to have fabricated data in over 90 publications, many of which contained a positive assessment of the risk–benefit ratio of infusions containing hydroxyethyl starch solutions. These publications influenced, for example, the guidelines of several medical societies in the UK. After the retraction of these publications, the data from the remaining studies indicated a worse risk–benefit ratio of hydroxyethyl starch solutions. At that time, however, many patients worldwide had already been treated with them under a different assumption (Marcus 2018). In another case, a researcher is believed to have faked numerous clinical trials investigating the benefits of various drugs and supplements to prevent bone fractures. Consequently, the results of meta-studies were distorted, a Japanese guideline may have been influenced, and real clinical trials were initiated based on the fabricated results (Else 2019; Kupferschmidt 2018). Another example is that some questionable studies have claimed that the drug ivermectin is effective against COVID-19. These studies were included in some meta-analyses, which came to incorrect conclusions regarding the efficacy of this drug against COVID-19 (Lawrence et al. 2021). Further meta-analyses showed that ivermectin is probably of no use in the prevention and treatment of COVID-19 infections (Popp et al. 2022). Nevertheless, many patients were treated with the drug because of the previous results (Mega 2020). The questionable studies were not explicitly stated to be intentional fakes. Instead, a more cautious wording was used, for example, stating that these studies were problematic, that there were several irregularities in the data, and that there were concerns about misconduct and a lack of research integrity (Lawrence et al. 2021; Popp et al. 2022).

How fake papers can be detected: there is no simple single method

It is necessary to identify fake papers to retract them and to prevent their publication. Unfortunately, this is not easy. There are currently different approaches to detecting fake papers, all of which have their limitations. These approaches can be broadly divided into those that identify problems in the content of papers and those that focus on non-content-related anomalies. Some of them are time-consuming, while others can be used automatically on a large scale.

Content-related approaches to detect fake papers

An easy way to weed out some potential fake papers at an early stage is to ask for original data at the time of submission. Many fake authors seem unprepared for this request and cannot provide original data (Miyakawa 2020; Seifert 2021a). A much more complicated approach is to identify manipulated images (Byrne and Christopher 2020; Seifert 2021a). Identifying such manipulations by hand is time-consuming and requires substantial skill. Van der Heyden (2021) and two videos by STM (STM 2022, 2023) describe possible procedures for recognizing image manipulation. Software is now being tested that can be used to automatically search for problematic images on a larger scale (Bucci 2018; Van Noorden 2022). Another indication of possible fake papers is errors in nucleotide sequences, which can be detected semi-automatically (Park et al. 2022).

Non-content-related approaches to detect fake papers

In addition to missing, incorrect, or manipulated data, there are other signs that a paper may be fraudulent. One approach focuses on peer reviews. It seems to happen that paper mills successfully pose as reviewers. This way, they can provide positive reviews of their own papers. Identical reviews that come from different reviewers may indicate fake reviewers. It may also be conspicuous if different reviewers submit their reviews for the same paper at the same time. All papers handled by these reviewers should then be checked again in detail to see whether they are fake papers (Abalkina and Bishop 2023; Day 2022). Naunyn-Schmiedebergs Arch Pharmacol therefore ignores author suggestions for referees, at least reducing the probability of fake reviews.

Another method focuses on the writing style of papers. To reduce the effort of writing a fake paper, some people use artificial intelligence to either create new text or rewrite text from other publications and claim it as their own. Sometimes, this results in very strange expressions, which have been called “tortured phrases.” Systematically scanning for such phrases can therefore identify potential fake papers (Cabanac et al. 2021). Even without the occurrence of “tortured phrases,” papers written by artificial intelligence can be identified automatically and with high accuracy by the writing style (Desaire et al. 2023a, 2023b). Naunyn-Schmiedebergs Arch Pharmacol uses software for the detection of language manipulation and rejects all papers with signs of attempted plagiarism. It is also possible to compare papers automatically with already proven fake papers. A strong textual similarity can indicate that the examined papers originate from the same paper mill (Van Noorden 2023b). A further approach is to scan for papers that have been submitted to different journals but with different author lists. The background is that paper mills exploit the fact that journals are unaware of submissions to other journals. This allows them to change the customers for the purchased authorships while submitting papers to different journals (Wittau et al. 2023). It is also possible to look for papers that have previously been publicly offered for sale by a paper mill (Abalkina 2023). Finally, there are several anomalies and metadata-based features that may serve as additional indicators of fake papers (Abalkina 2023; Dadkhah et al. 2023; Sabel et al. 2023; Seifert 2021a; Wittau and Seifert 2023).

To date, there is no perfect method for detecting fake papers. The methods presented here are unfortunately only applicable under certain conditions. For this reason, tools are now being developed that combine many different methods. However, it is not publicly stated how exactly these tools are supposed to identify fake papers (Else 2022). Although paper mills will likely adapt to methods of detecting fake papers, the use of such tools will hopefully make their work more difficult in the future.

Training scientists, authors, editors, and reviewers to detect fake papers is crucial

The most important step in keeping the scientific record clean of fake papers is to prevent them from getting published. Ideally, fake papers are already identified in a pre-review screening by tools or experts for detecting fakes, thus relieving the burden on scientific reviewers. Training opportunities for reviewers to recognize fraud would also be helpful (Korte and van der Heyden 2017). Retracting already published fake papers is also a very important step, but unfortunately less effective, because it does not prevent the fake papers from being cited in the meantime, and the citing does not even stop after the retraction (Candal-Pedreira et al. 2020; Wittau and Seifert 2023). Publishing notes of concern is a good way to be transparent about problems in a publication until a detailed investigation of a potential fake paper case has been completed (Seifert 2021b).

Retractions should be viewed as positive tools to restore trust

Handling fake papers requires a huge effort and commitment on the part of the journals. Molecular Therapy, Biology Open, the International Journal of Cancer, and Naunyn–Schmiedeberg’s Archives of Pharmacology are examples of journals that openly address the problem of fakes within their own submissions and publications and take care of it (Frederickson and Herzog 2021; Hackett and Kelly 2020; Heck et al. 2021; Seifert 2021a). All these journals have already retracted fake papers. Unfortunately, there are also several respected journals that disregard evidence that several of their publications are fake and do not take appropriate action (Wittau et al. 2023). But the longer they do not act, the more severely will their reputation be damaged in the long run. There are also bad cases where papers are retracted silently and without any explanation. Specifically, all evidence of the paper’s existence is removed from the journal’s website (Teixeira da Silva and Daly 2024). We can only encourage all journals to publish retraction notes proactively, even if there will be many of those. From the experience with Naunyn-Schmiedebergs Arch Pharmacol, we can tell that such measures are appreciated by the scientific community and lead to an increase in honest high-quality submissions. But if journals follow the policy of ignoring the fake paper problem and fail to publish retraction notes although suspicious papers are published in this journal, the reputation of the respective journal will be damaged in the scientific community by spreading the word informally among scientists.

Software tools provided by publishers to fight fake papers

Publishers should be responsible for ensuring that editorial boards can maintain the scientific integrity of their journals. They should support them with tools, training, and dedicated research integrity resources to which editors and reviewers can turn when encountering suspicious papers. Such integrity tools could also be made contactable for readers. In this way, the readership could be actively involved in post-publication peer review (Korte and van der Heyden 2017). Leading scientific journal publishers such as Springer Nature are aware of the urgency to develop software tools to detect fake papers more effectively. For example, the software tools Geppetto und SnappShot will support editors and referees alike to detect more easily text sophisticated manipulations and image manipulations, respectively (https://group.springernature.com/fr/group/media/press-releases/new-research-integrity-tools-using-ai/27200740; last accessed June 17, 2024). Such long-awaited software tools will constitute a major support for editors and reviewers to catch fake papers in the future that would have smoothly passed review despite diligent efforts. Internal author-flagging software systems in editorial systems will also be helpful to identify suspicious authors and catch fake papers.

Even more effective would be blacklists of suspicious authors shared across different publishers and journals because switching between different journals and publishers is an effective strategy for fake authors not being caught (immediately). Currently, fake paper authors aggressively abuse the absence of flagging systems to their advantage because they know that there are high barriers to the implementation of such data bases including reasons of discrimination, fairness, personal data protection, and political correctness.

Substantial cultural changes at different levels are needed to weed out fake papers

To effectively combat fake papers, editors and reviewers need to become more aware of the problem and more willing to act against it. There is no point in identifying fake papers if they are not subsequently retracted or at least flagged by the journals (Wittau et al. 2023). We noted that despite the fake paper problem in Naunyn-Schmiedebergs Arch Pharmacol, the number of submissions and published honest papers has increased substantially over the past years in this journal. Apparently, honest authors value the approach of the journal to retract every single fake paper that the Editor-in-Chief becomes aware of. Thus, retraction of every fake paper pays off in many respects, for the scientific community, the honest authors, and the reputation of the affected journal. Fortunately, there are journals that share this perspective. At a recent symposium, the editors-in-chief of several journals met to discuss different strategies against fake papers and declared their intention to work together on this issue (Benyó et al. 2024).

Unfortunately, paper mills will not give up their business model as long as it remains financially lucrative and as long as there are well-paying customers. Presumably, they will continually adapt their methods to make their papers undetectable by current fake paper scanning methods (Abalkina 2023). To solve the fake paper problem in the long term, it will be necessary to address the cause of this problem: academic systems that put scientists under too much pressure to publish and reward them based on publication metrics. Revising academic systems globally would not only help against fake papers. Many (presumably) honestly conducted studies cannot be reproduced either (Baker 2016; Begley and Ellis 2012). In two surveys of scientists, most respondents stated that pressure to publish would be one of the reasons for this. The surveys did not explain further how this should be connected (Baker 2016; Samuel and König-Ries 2021). Perhaps the connection is that the thoroughness with which research is planned, conducted, and analyzed suffers from the need to produce many publications in a short period of time (Diaba-Nuhoho and Amponsah-Offeh 2021). Furthermore, the practice of evaluating scientists based on their publication metrics was cited as detrimental to scientific innovation. According to this theory, many scientists would respond by focusing on established research areas, as they feel they are more likely to get publications and citations this way. Novel and unconventional ideas would therefore be less likely to be pursued (Bhattacharya and Packalen 2020; Foster et al. 2015). The final example to be mentioned here is that focusing too much on a professor’s publications can lead to professors being hired at universities with insufficient consideration of their teaching skills (Rawat and Meena 2014). Opinions differ on the extent to which this happens (Abbott et al. 2010). If professors then also have little time for teaching because they must focus on producing more publications, the quality of academic education suffers (Rawat and Meena 2014; Miller et al. 2011).

Getting rid of the fixation on publication metrics is crucial in the fight against fake papers

To sum up, there are various reasons why the scientific community needs to rethink its fixation on publication metrics when it comes to filling professor positions in academia and awarding research grants. In a recent study, using leading German pharmacologists as a case study, it has been shown that bibliometric parameters are arbitrary and can be interpreted (and hence, manipulated) one way or the other (Fox and Seifert 2024). If one just chooses the “right” bibliometric parameter, even a “bottom” scientist can be propelled to the “top,” and a “top” scientist can be sent to the “bottom.” Eliminating the mantra of bibliometric parameters as a proxy of perceived scientific excellence in academia is probably the best long-term solution to the fake paper problem, as it eliminates the reason for their publication and thus avoids an arms race between the methods of forgery and the methods of detection. In medicine, one would speak of treating causally instead of just symptomatically. Fortunately, there is an awareness of problematic standards in the assessment of research. The Declaration on Research Assessment (DORA) (https://sfdora.org/) is a statement and initiative advocating better approaches to research assessment. The statement has now been signed by over 25,000 individuals and organizations. Establishing such structural changes will of course take time. It will be crucial to instill the mindset in the education of the next generation of scientists that truthful scientific findings are the highest value, not metrics.

Fighting fake papers is the responsibility of the entire scientific community

Appropriate education of scientists by universities and research institutions would not only help against fake papers. According to Tang (2024), there are also unintentionally questionable research practices due to a lack of knowledge of good scientific practice. Better education of scientists in good scientific practice and the acquisition of truthful knowledge would therefore also increase the quality of scientific publications whose authors have not intentionally fabricated data, but use questionable practices such as p-hacking or selective reporting to be able to tell a supposedly “convincing story.” Uneducated authors may assume that journal editors like such “convincing stories” to increase the acceptance probability of the paper. Rather, Naunyn-Schmiedebergs Arch Pharmacol accepts even “incomplete” stories, provided that the limitations of the study are discussed and future directions of research are properly outlined, e.g. with a scheme highlighting open questions. With this policy, we hopefully reduce the pressure of authors to create a “convincing story.” Reporting of true data has priority in our journal.

Authors must ensure that their work complies with good scientific practice at all times. Institutions should promote and monitor compliance with good scientific practice. The journals are responsible for thoroughly reviewing papers before they are published. In case of reasonable suspicion of scientific misconduct in an already published paper, they must investigate this conscientiously and transparently. Finally, readers can participate in post-publication reviews. Fighting fake papers and promoting truthful research is thus the responsibility of the entire scientific community.

Unfortunately, truthful research on fake papers will be needed

Since fake papers will remain a major problem at least in the intermediate future, further research into how they can be identified is urgently needed. Such “truthful fake paper research” needs to be conducted with large datasets so that the findings on this topic are no longer based on assumptions, anecdotal evidence, and small sample sizes. The original research papers listed in Table 1 provide a solid methodological and conceptual starting point for “truthful fake paper research.” Naunyn-Schmiedebergs Arch Pharmacol has become a target of paper mills and is committed to fighting them and publishing original papers and reviews on research integrity and fake paper research. Dishonest paper mill authors publishing in Naunyn-Schmiedebers Arch Pharmacol should know that we will follow up on every single case of suspected fake science and that such papers will not prevail in the long run. New software tools will support our efforts in due course.