Introduction

Choosing the right journal is one of the key decisions made during a research project. Judging from surveys reported by Fry et al. (2009), around 90% of UK researchers, regardless of the discipline, consider peer reviewed journals as a “very important” channel for communicating research output. Publishing in prestigious journals is crucial for career development and, typically, for individual satisfaction. Journal choice is also important for the community and society at large, because articles published in reputable journals are more widely read and cited, as well as taken more seriously. Clearly, it is thus very desirable that high-quality articles appear in top journals. Whereas one would expect just that, given how selective they are, there is some evidence to the contrary. In particular, retractions seem to be more common in top journals (Fang & Casadevall, 2011).

Low-quality or outright fraudulent papers being accepted in prestigious outlets is just one of many reasons, for which the current practice of academic publishing has been criticized (e.g. Bornmann, 2012); and indeed there has been an overhaul of practices in recent years. Notably, open-access journals have emerged as important players in the last decade or so, with more and more public agencies and universities wanting research results they have helped generate to be freely available rather than hidden behind a paywall (Björk, 2017). It is also a widespread perception that academics are under mounting pressure to publish and to publish well (van Dalen, 2021), even if this requires some ethical or intellectual maneuvering (Kiai, 2019). As a medical researcher quoted by Fry et al. (2009) put it, “Increasingly there has seemed to be no point in doing anything other than aiming for top class American publications.” Likewise, Heckman and Moktan (2020) note that “Pursuit of T5 publications (T5 standing for the top five journals) has become the obsession of the next generation of economists.” All the more, it is important to understand how researchers navigate this dynamic landscape.

One question particularly worth asking is that about the possible differences in the publishing strategies of males and females. There is some evidence that males tend to be more competitive, self-confident and risk-seeking (Barber & Odean, 2001; Croson & Gneezy, 2009). One likely (and advantageous) consequence of these characteristics in the world of academia is that they are more inclined to cite their own work (Ghiasi et al., 2016). It is also very plausible that males are more willing to target top journals. This approach on the part of males may perhaps partly explain gender difference in publication records (Holman et al., 2018; Jappelli et al., 2017; Madison & Fahlman, 2021; Mauleón et al., 2013; Mayer & Rathmann, 2018) and career paths (Weisshaar, 2017), including drop-out rates (Astorne-Figari & Speer, 2018).

Literature review

For this overview we focused on studies exploring how an author’s characteristics might affect their choice of journal. Here we touch upon theoretical and empirical papers in order to identify predictions that could potentially be tested given our approach.

How authors choose journals

Pepermans and Rousseau (2016) have reviewed the literature on this subject, and present three general categories that potentially determine an author’s choice of journal:

  1. 1.

    Author’s characteristics: “CV value of the publication, the author’s evaluation of the editor, and past submission success”;

  2. 2.

    Journal’s characteristics: “the quality of the review process, publication delays, submission rejection risk, service level of the journal, technical features of the journal, author charges, local visibility, professionalism, influence, credibility and prestige of the journal, and the likelihood of acceptance”;

  3. 3.

    Other features: “impact on scientists and practitioners, the potential impact or visibility of the article, communication strategies, negotiations with coauthors, and philosophical and ethical issues”.

Other categories have also been proposed. For instance, Björk and Öörni (2009) in their “net value of submission” model classify 30 factors under four categories: journal prestige, readership, performance and infrastructure. Lee et al. (2020) review 33 factors identified in the literature and their survey of authors who publish in information science and library journals. They come up with 15 dominant factors which can be divided among five categories: peer review, prestige, infrastructure, readership and performance. It appears that the factors highlighted and analyzed in different studies tend to overlap to a large extent (see Frandsen, 2019 for a literature review).

Research also shows that the relative importance of these factors tends to vary across disciplines (Bröchner & Björk, 2008; Dalton, 2013; Gordon, 1984; Lee et al., 2020; Lusk & Hudson, 2009; Luukkonen, 1992; Wijewickrema & Petras, 2017). For instance, Wijewickrema and Petras (2017) report, that compared to social science researchers, the authors in the discipline of medicine tend to pay more attention to the impact factor and whether a given journal is included in indexing services. Also taken into account are the prestige of the publisher, and online submission and tracking services.

Overall, it appears that a journal’s attributes like reputation, readership, decision lags,Footnote 1 acceptance rate and review quality tend to be the top determinants of where scholars submit their work (Björk & Öörni, 2009; Oster, 1980; Pepermans & Rousseau, 2016; Śpiewanowski & Talavera, 2021; Unutmaz Durmuşoğlu et al., 2021). Unutmaz Durmuşoğlu and Durmuşoğlu (2021), for instance, use a dataset of published articles referring to “environmental risk” in order to identify which factors have the greatest impact on the journal in which a given paper is published. The results of their model show that the acceptance rate tends to be the most important criterion that the authors consider when selecting a suitable outlet. This is followed by scope similarity (the most cited journals in a given paper tend to be the ones where the paper is eventually published) and time to first decision.

There have been relatively few studies focusing on author’s characteristics. Apart from research excellence and past publication experience, it is likely that individual features such as gender, risk tolerance, overconfidence, and time preferences could play a role in determining submission strategies. We address these below.

Overconfidence, risk and time preferences

It has been argued that positivity bias, notably overconfidence and excessive optimism, may significantly influence the choice of a journal (Crack et al., 2011). There is some evidence that people tend to overestimate their ability in most situations. In our case, the authors may systematically overestimate the quality of their papers and thus choose overly prestigious journals that are very unlikely to accept them. Overconfidence could be balanced by such factors as risk aversion and impatience, because targeting lower-ranked journals results in a more certain reward; and one that is sooner. The model of Heintzelman and Nocetti (2009), for instance, predicts that risk-neutral and patient authors submit their work to top-tier journals, while impatient and risk-averse authors submit to second- or third-tier journals depending on the paper’s quality.

Gender differences

While overconfidence, risk acceptance, and patience are generally unobservable, they may have visible correlates, notably gender. Gender differences on these dimensions might lead to differences in submission strategies. A number of studies using different samples, including students, stock traders, and professionals (Barber & Odean, 2001; Charness & Gneezy, 2012; Thöni & Volk, 2021) find greater overconfidence and greater risk-seeking in men, although null results were also sometimes reported (Deaves et al., 2010; Hardies et al., 2013; see also the critical review of literature on risk by Filippin & Crosetto, 2016). The situation is less clear in the case of time preference: although several studies (Bettinger & Slonim, 2007; Castillo et al., 2011; Dittrich & Leipold, 2014) report that men tend to be more impatient, Golsteyn et al. (2014) find the opposite and several studies observe no difference (Thöni & Volk, 2021).

Although moderate gender differences in risk taking and overconfidence plausibly exist in the general population, it is an open question if they survive self-selection and socialization typical for academia (Hardies et al., 2013; Nekby et al., 2008). In their Study 1, Drupp et al. (2020) find that the gender risk gap in academics is reduced (and disappears in older individuals) when their professional (rather than private) identity is made salient by asking, in a preceding task, about the professional rather than private life. Such a manipulation affected women’s (but not men’s) risk aversion, the professions’ seemingly prevailing norm of high risk acceptance modifying their home-grown preferences. It should be noted however, that the authors fail to replicate their results in Study 2; it is also difficult to say if they would carry over to non-financial risk taking.

Concerning the authors whose papers get rejected (which is to say—almost all of us), gender difference in willingness to compete after a negative feedback (observed, in student samples, by Buser (2016) and Buser and Yuan (2019)) might affect future submission patterns. Specifically, females might be willing to submit to lower-ranked journal than males after a rejection, thus decreasing their chances for a publication in a prestigious outlet. Overall, thus, there are some reasons to hypothesize that male academics will tend to be more ambitious in their choice of journals. Yet, as of now there has only been limited and mixed evidence. Gravert and Sørensen (2020) surveyed nearly 100 young economists graduating at top-50 universities (RePec world ranking), asking about the submission order of the top paper they had written during their PhD curriculum (known in economics as job market papers); and in so doing they elicited risk, time, and competitive preferences, as well as grit. No gender differences in submission strategies were found; this is perhaps not surprising given that gender differences in measured traits were also small or non-existent.

This finding contrasts with the results obtained by Djupe et al. (2019) (see also Teele & Thelen, 2017) who compared the submission strategies of male and female political scientists. Based on their self-reports, relatively few female authors submit their work to the top three journals and relatively many female authors submit to a journal “that is most likely to accept” their work. This, however, could be partly due to systematic differences in the contents. For example, males tend to write/submit more quantitative articles (Beneito et al, 2021). Likewise, it is not completely clear whether and to what extent risk and time preferences play a role in these submission differences. Djupe et al. (2019) also suggest that co-authorship tends to amplify gender differences: it increases the number of submissions and publications of males to a greater extent than those of females.

A clearly understudied question is if males and females devote the same amount of time and attention to planning their submissions. It is a common perception that male scholars have more time for research, mostly because they devote less time to academic service (Guarino & Borden, 2017) and domestic chores (Jolly et al., 2014; Mandel et al., 2020). Interestingly though, the evidence that men actually do more hours of research is very limited (Bentley & Kyvik, 2013; Ceci et al., 2014). Still, it is plausible that, within the same time devoted to research, men (being more competitive on average) think more about possible publication strategies (while women perhaps focus more on the research for its own sake). That may lead men to consider a larger number of possible outlets, as well as to submit their papers to more prestigious journals (as it seems to pay off); but also to journals giving a greater chance of success for the paper in question, and without compromising on the prestige of the journal. We are not aware of earlier research that would shed light on these questions.

Team composition

Gender may also play a role if the decision is made by a team. The presence of a man (but not necessarily men only) may trigger risk taking (Bansak et al., 2011; Bogan et al., 2013), and perhaps partly because men are relatively willing to lead their teams (Born et al., 2020). On the other hand, in the case of an investment bank portfolio, more women in the executive team were associated with higher portfolio risks (Berger et al., 2014). Castillo et al. (2020) also suggest that women take more risks as the fraction of males increases in the room, even if their presence or actions do not affect the payoffs. Also, being in a team was found to enhance women’s willingness to compete (Healy & Pate, 2011). Hence, risk-taking and the competitive preferences of females in teams seem to greatly depend on the context. Elsaid and Ursel (2011) find that more women in an executive board is associated with a higher likelihood of selecting a female CEO and that changing a CEO from a man to a woman decreases corporate risk-taking.

Summary of prior literature

To summarize, the existing literature clearly hints at ways in which submission strategies may be affected by the gender of authors. However, direct verifications are scattered and troubled by small sample sizes, focus on top journals/universities only, and/or possible uncontrollable difference in quality of papers. We seek to reduce the latter problem by sampling journals rather than authors. We can thus compare submission strategies for papers that ended up at equally prestigious journals.

Methods

Journals, papers and author data

In this project, we focused on three scientific fields: economics (econ), environmental sciences (env), and history (hist). We chose these three because they represent three major areas of science (social sciences, natural sciences, and humanities, respectively). Subject to this constraint, when choosing the fields, we avoided disciplines that are known to be dominated by one gender, such as mathematics (Huang et al., 2020; Larivière et al., 2013). Additionally we chose fields with a relatively low number of authors per paper (Abramo & D’Angelo, 2015). Identifying gender effects would have been much harder in disciplines in which most papers have a large team of authors. While we would have liked to cover more fields, it would have been difficult given that our approach is time-consuming.

First, we downloaded the list of all journals from the “Scientific Journal Rankings” (scimagojr, SJR) in these subject areas. Next, for every journal, we looked for the list of very recent publications (published between January 1, 2020 and the date of data collection, July 28, 2020) on Scopus and exported the data available. We then dropped all the publications for which no corresponding author (henceforth, CA) was indicated and those for which the CA’s email address was missing. From the resulting database, we drew a random sample of 4800 English-language publications (1800 in econ, 1500 in env and 1500 in hist; the sample of econ papers was a bit larger because we used two different ways of data collection; see the survey procedures under “survey data” below).

In addition, based on an automated python script using the title of publication, we looked for the profiles of all the authors of these 4800 articles on Google Scholar and ORCID in order to obtain additional information: full name (of the co-authors only, as we had CA’s full name from Scopus database), the list of publications (including the title and year of publication), the number of citations and the Hirsch index (h-index). Using this data, we calculated additional measures. First, “years active”, a measure of experience and seniority (which may differ by gender), was calculated as the number of years that passed (by 2020) since the CA’s first publication (listed on Google Scholar).

Second, we created a variable measuring the share of CA’s publications that were in English. We did so with the help of python package langdetect applied to each publication title.

Third, we created a dummy variable indicating whether the CA is a “theorist”, a proxy for the type of research and subfield represented (which, again, is likely to correlate with gender). Clearly, the term “theorist” is not well defined or easily operationalizable. We constructed a crude indicator as follows. Based on common sense and visual inspection of all CAs’ publications, we created a list of keywords likely to signal at least some theoretical contribution in a publication, consisting of the terms understand*, interpret*, theor*, teor*, mathem*, matem*, concept*, method*, model, framework, where “*” stands for any string of letters (so that e.g. “theoretical” but also German or Dutch “theoretisch*” are included). We then categorised the CA as a “theorist” if they had at least one publication the title of which contained at least one of these keywords.

Finally and most importantly, we tried to identify the gender of all the authors based on names from Google Scholar profiles by using a python package gender-guesser, version 0.4.0. We took a conservative approach by treating the “mostly male” and “mostly female” verdict of gender-guesser as inconclusive. Unfortunately, this means that gender remains uncertain for as many as 34.14% of the sampled authors. Because of the important role that gender plays in our investigation, among those CAs who responded to our questions, but whose gender was uncertain, we established it manually. Towards this goal, a research assistant was instructed to visit the Google Scholar and/or professional website of the author in question and locate relevant pronouns (“In his research, prof. Jones uses…” etc.) or establish the gender, if possible, by the author’s photograph. This procedure is very time-consuming and running it for all the authors would have produced little additional insight; it was thus restricted to those CAs who responded. Both sources of gender information were used to calculate the number of male and female authors per article.

Survey data

Between August and December 2020, we contacted all the CAs by sending an e-mail from the university account of one of the authors, asking about past rejections and potential alternative journals for that specific published paper (see the email template in Appendix 1). In the e-mail, we stated that we were asking for these data mostly so as to investigate the usefulness of automated journal suggesters. We believe it reduced the probability that the responders would see the reported past rejections and alternative outlets as possibly signaling their confidence, risk preference etc. This method has been approved by the ethics committee at the Faculty of Economic Sciences at the University of Warsaw.

Starting out with the field of economics, we first checked whether there would be differences in the response rate if the email contained a link to a web survey containing our questions, instead of asking them directly in the body of the e-mail. Assigning a randomly selected half of papers to each method, we observed a slightly lower response rate for the link (22.1% vs. 25.6%). Moreover, e-mail responses usually included more information about publication strategies and the use of automated journal suggesters. We thus used this approach for the other two disciplines.

The overall response rate across all disciplines was 21.9%, resulting in 1111 usable responses. The responses to our questions were recorded and the authors were thanked for their input. The statistics of the authors who responded are provided in Table 1. Selection patterns are addressed in Table 5 in Appendix 1.

Table 1 Statistics of the authors who responded (1111 authors)

Predictions

Measures

We proposed three measures for our authors’ submission strategies. First, we looked at the number of past rejections of a given paper. It was assumed that a greater number of rejections corresponds to a higher (over)confidence, because it suggests that the paper was repeatedly sent to an overly prestigious journal, apparently because its authors believed it was good enough to have a chance of being accepted.

Second, confidence could also be measured by journal rank change after a rejection, defined as the difference between the ranks of the journal where the paper ends up and where it is submitted initially.

$$Rank\,change=Rank\,published-Rank\,first\,submission$$

We made recourse to the Scientific Journal Rankings (SJR) rank which ranges from 0.1 (least prestigious journals) to 88.192 (Ca-A Cancer Journal for Clinicians). As it turns out, in the sample of 1111 papers whose CAs responded, the highest rank was 17.134 and the vast majority were ranked between 0 and 3. As expected, rank changes tend to be negative, with a large negative rank change meaning that the paper ended up in a much less prestigious journal than the author(s) initially had hoped for, again signaling (over)confidence. Naturally, we set \(Rank\,change = 0\) if the article is published at the first attempt.

The third measure was based on the reported number of alternative journals. A low value suggests that the author(s) had not thought that much about their follow-up strategy in the case of a rejection at the current journal, either because they did not have much time for this or because they were very confident it would be accepted.

Gender

Following the literature, we expect men to be more overconfident, meaning more past rejections, greater rank reduction (more negative rank change) and fewer alternatives. In other words, men initially would tend to submit to “better” journals and, consequently, would be more likely to get rejected (perhaps several times) and end up in a much less prestigious one. Being more confident about the acceptance, men would also tend to have fewer alternative journals.

On the other hand, given the findings that women tend to be more responsive to negative feedback, an alternative hypothesis could be formulated for the rank change measure, namely that it is more negative for women. Finally, assuming that men are more eager to engage in publishing as a competitive game and that they may devote more time and attention to their submission strategy, they may end up considering more alternative journals.

Findings

In this section we succinctly report the results, focusing on the three measures indicated earlier: rank change, the number of rejections and the number of alternatives. We run the regressions using “general to specific” approach, meaning iterative elimination of variables from an initial general model with a large number of explanatory variables. We use the Poisson regression, which is appropriate given that the dependent variables are (small) integers. Table 2 summarizes results for the number of past rejections. Not surprisingly, there are significant differences among the three disciplines: compared to econ, hist and env scholars tend to have fewer past rejections. Also, more experienced authors, as measured by the h-index, tend to have slightly fewer rejections, however the effect is not significant for the number of years they have been active professionally. The share of English language papers and our simple “CA theorist” measure are also insignificant for past rejections. Most importantly, articles with a male CA tend to have a higher number of past rejections, confirming our hypothesis about male overconfidence. Likewise, teams of authors with a higher female ratio (the number of authors identified as women divided by the number of authors identified as men or women) tend to have fewer rejections. On the other hand, the number of authors per se seems to have a smaller and less robust effect on the number of rejections.

Table 2 Poisson regression results for the number of rejections

Turning now to rank change (Table 3), we see that, again, there are differences between the fields: environment and history researchers tend to have lower rank reductions (in other words, less negative rank changes) than economics authors. Seeking to verify our hypotheses, we observe no evidence that characteristics of the CA or the team that we tested affect journal rank change.

Table 3 OLS regression results for rank change

Looking at the number of alternative journals as dependent variable (Table 4), we see that male CAs tend to report more alternatives. That would imply that our hypothesized competitive strategizing effect is stronger than the overconfidence effect, although the possibility of gender-specific misreporting of alternatives is also difficult to rule out (see possible limitations in the discussion and conclusions section below). The number of authors has no effect on the number of alternatives. By contrast, past experience, as measured by years of professional activity (but not h-index) has a slight negative impact on the number of alternatives (this significance remains in place in all other specifications we run; one model is reported here aiming to keep the table concise). Again, our “CA theorist” measure is not significant. However, interestingly, “CA share of English titles” measure has a slight positive significance for the number of alternatives, meaning that higher share of English language publications is associated with higher number of alternatives. One explanation is that those who publish predominantly in English simply have more outlets to consider. Another possibility could be that their target audience is wider, so they are not oriented towards one particular country, hence, again, they have more journals to consider.

Table 4 Poisson regression results for the number of alternatives

Discussion

Clearly, our novel approach to investigating journal choice is not free of limitations and caveats. Most importantly, we conducted a survey with no incentives for truth-telling and we were not in a position to confirm the veracity of the collected data using publishers’ records. Participants may have genuine difficulty recalling past rejections and alternative plans. They may also strategically omit some items (or, a less likely possibility—make them up) in order to make a better impression. For example, they may be ashamed to admit that a paper ending up in a mediocre journal was initially submitted to an excellent one. On the other hand, such a situation need not always mean overconfidence or bad planning; so it would seem that the shame involved in revealing it to a stranger who does not know the project would be limited. This is especially true given that we (truthfully) stated that the project is about journal suggesters, which shifts the focus from a journals’ prestige (vertical comparisons) to their scope (horizontal comparisons) and from the efficacy of the responders (the authors), to that of the mindless robots/algorithms (the suggesters). Moreover, some authors may be ashamed that they have been rejected by a bad journal, in which case these two effects could partially cancel each other out in the calculations of our rank change measure. By contrast, the number of past rejections, which is arguably strikingly small, was most probably often underreported and should thus be cautiously treated as a lower bound. It is also quite possible that participants did not always mention the journals in the right chronological order, which may have affected our rank change measure. Besides, there are limitations in terms of data availability; for example, we can only identify gender and professional experience in a subset of authors.

It may be noted that, within each of our three fields, gender distribution varies by subfield. If our dependent variables also depend substantially on subfield, estimates of gender effects obtained without proper control for sub-discipline may thus be biased. The challenge is that it would be difficult to define subfields in a manner consistent across our three very different disciplines. Moreover, numbers of observations would necessarily be very unequal and in some cases very low. Besides, it is very difficult to categorize the authors into a single subfield, as many of them can belong to more than one (e.g., involved in micro- and macro-economics, or environment and economics simultaneously). One approach would be to categorize the papers instead of the authors using pre-defined SJR journal categories, however the same problem emerges: many publications in our sample are in journals that belong to several categories, often from different fields (e.g., Political Geography is categorized under “history” as well as “social and political science” categories). Nevertheless, the analyses involving “CA theorist” measure should partially address this concern, as some subfields are more “theoretical” than others. The same is true for the “CA share of English titles” variable. While evidently not accounting for subtle differences between various subfields, our identification strategy is consistent across the three disciplines and compatible with the sample size. The key observation is that controlling for “CA theorist” and “CA share of English titles” in the models does not substantially change the estimate for “CA male”, meaning that gender differences persist irrespective of the field represented (more vs. less theoretical, more vs. less international).

Our finding of greater confidence in males (as measured by the number of past rejections) confirms earlier reports coming from other domains, such as finance and management (Barber & Odean, 2001; La Rocca et al., 2020). A few instances of the authors’ replies to the open-ended question about the journal selection criteria also confirm the importance of journal choice factors discussed in the literature (Lee et al., 2020). For instance, one of our respondents mentions: “For every journal article I write, I create a “ranking” of possible journals to submit it to, which I rate according to impact factor, how well it fits the subject and style of my article”. Reading through the comments, one also gets the impression that experienced researchers know the relevant target journals in their field, while younger scholars tend to rely on the advice of colleagues or supervisors (Dalton, 2013).

Conclusions

The aggregate picture that our alternative measures yield is that male authors (and male-dominated teams) pursue a more aggressive publication strategy, continuing with their submissions to relatively good journals and planning more potential submission targets in the face of past rejections. Indeed, the number of females in a team is negatively related to the number of rejections and, weakly, to the number of alternatives. Likewise, the male gender of the CA is associated with more rejections and alternatives.

Even if males experience more rejections overall, they also have more chances. Given the inherently random nature of the refereeing process, their strategy thus gives them a greater probability of ending up in a prestigious journal. It might also give them a greater possibility for achieving publication in some journal. Indeed, suppose an author has a threshold of journal prestige below which they will not submit. The faster they lower the prestige of the targeted journal, the more likely they will fall below the threshold, thus increasing the chance that the paper never gets published at all. The fact that males are more forward-looking, and consider more alternative outlets, also gives them an edge here. Moreover, submissions are not independent. Once an author has publications in highly-ranked journals, their improved reputation is likely to help publish more in the future. These differences can thus contribute to the observed gender difference in publishing records.

These effects could also have a bearing for the choice of the subfield. In economics for example, leading, prominent journals of some specialties (like econometrics, macroeconomics, and financial economics) are extremely selective, which might partly explain why women tend to shy away from these subfields, choosing demography, labor, or poverty economics instead. As Beneito et al. (2021) show, the fraction of female researchers is below 20% in the former three and above 40% in the latter three subfields. Choosing to focus on demography rather than econometrics is tantamount to recognizing that one will never publish in Econometrica; in view of our findings, it is very plausible that such a concession usually comes more easily for women. Again, a larger study, allowing for a finer operationalisation of subfields (rather than just our crude “CA theorist” measure), could shed some light on this conjecture.

Likewise, there are very few women (about 14%) to be found in highly mathematical, prestigious subfields of computing, such as programming languages, numerical & scientific computing, or the theory of computer science. Less coveted subfields, such as interdisciplinary computing or human–computer interaction see more balanced numbers, although women still only comprise 20% of the whole.

Among other things, these observations emphasize the importance of mentoring programs (Blau et al., 2010). They could encourage female scholars to be not only more confident but also more strategic and forward-looking in their journal choice strategies.