1 Introduction

Crowdsourcing is no longer just a fashion or buzzword (Cricelli et al. 2022; Pavlidou et al. 2020), but a popular topic of research and a powerful tool for organizations (Karachiwalla and Pinkow 2021). It is attracting increasing interest in management research and practice as a promising and distinctive business model (Kohler and Nickel 2017) and a new paradigm with a wide scope of application (Palacios et al. 2016).

Crowdsourcing has proved to be useful in solving organizational problems, introducing innovative products and services to the market (Greco et al. 2016), creating value (Lenart-Gansiniec and Sułkowski 2020), and managing knowledge and organizational learning (Lenart-Gansiniec 2021). Recently, scholars recognized that by mobilizing a large and diverse community leveraged by technology use, crowdsourcing may also become a promising way of creating scientific knowledge (Beck et al. 2022a). This can be achieved in particular by carrying out research tasks, accessing unlimited sources of knowledge, improving the quality and speed of research project implementation and lowering costs (Lukyanenko et al. 2019).

The use of crowdsourcing for research is seen as a natural consequence of the evolution of crowdsourcing per se and its limitless possibilities (Beck et al. 2022a). The openness of researchers to society and the inclusiveness of research does not only fit in with general postulates for the democratization of science, or its technicization, or platformization. In practice, conducting research using crowdsourcing also allows researchers to transcend small, hermetic intra-university research teams (Uhlmann et al. 2019) to embrace asynchronous, transnational, interdisciplinary teams consisting of researchers, non-scientists and anyone interested in performing various research tasks commissioned by academics (Franzoni et al. 2022). In this paper, the use of crowdsourcing to carry out a task or research tasks is referred to as crowdsourcing in science.

For academics, crowdsourcing may be a strategy for working effectively (Lukyanenko et al. 2019), a viable alternative to research projects (Aristeidou et al. 2017), a tool for research (Law et al. 2017; Uhlmann et al. 2019), or even an alternative model for doing science (Uhlmann et al. 2019). A crowd can perform a variety of tasks related to the seven stages of the research process (Beem et al. 1965): identifying the general study area (Beck et al. 2019; Beck et al. 2022a); selecting a topic and developing a focus (Beck et al. 2022a, b); deciding on the approach (Doan et al. 2011; Ipeirotis et al. 2009; Uhlmann et al. 2019); formulating a plan (Beck et al. 2019; Schlagwein and Daneshgar 2014); analysing data (Benoit et al. 2016); and presenting findings (Uhlmann et al. 2019). In addition, crowdsourcing in science facilitates communication with various stakeholders and the search for collaborators for joint research, and reduces the costs of conducting research. All in all, there are numerous reasons for using crowdsourcing in science, and such practices have indeed become widespread.

Crowdsourcing in science is of interest across a range of disciplines and has been discussed variously in literature on the environment, ecology, geography, astronomy, biology (Parrick and Chapman 2020), health-related research studies (Bassi et al. 2020), medicine, physics (Beck et al. 2019), economics (Hossain and Morgan 2006), sociology (Salganik et al. 2006), psychology (Fishbach et al. 2011), and behavioural science (Chandler et al. 2014). However, the growing number of publications certainly does not facilitate scientific recognition of crowdsourcing in science. Each of these different disciplines uses definitions for crowdsourcing in science that align with its dominant paradigms. While enriching the debate, this diversity of views can be an obstacle in the delivery of consistent findings and the development of science (Klimas and Czakon 2022). The lack of a solid understanding and adoption of an integrative definition makes it difficult to “advance the theory and practice” (Stock and Boyer 2009). It is also challenging for management and policy makers of higher education institutions to propose plans or strategies to support academics interested in reaching out for crowdsourcing in science.

Finally, the growing popularity of crowdsourcing in science is also resulting in an avalanche of related concepts. Researchers often interchangeably use such terms as: crowdsourcing science (Uhlmann et al. 2019), online citizen science (Aristeidou and Herodotou 2020), open innovation in science (Beck et al. 2022a), crowd science (Franzoni and Sauermann 2014), participatory science (Heaton et al. 2016), collaborative science (Herschlag 2020), virtual citizen science (Leeuwis et al. 2018), crowd research (Vaish et al. 2018), scientific crowdsourcing (Can et al. 2017), science 2.0 (Bücheler and Sieg 2011), and academic crowdsourcing (Hedges and Dunn 2017). Hence, discrepancies and overlaps emerge with regard to what academic crowdsourcing is considered to be.

At the same time, however, researchers seem to consistently share the view that crowdsourcing in science does not coincide with concepts such as citizen science or participatory science (Ciasullo et al. 2022; Palumbo et al. 2022). In citizen science, volunteer non-scientists support researchers and are involved in the collection, classification, analysis and reporting of large volumes of data. We agree with Levy and Germonprez (2017, p. 29) that “crowdsourcing is not rooted in citizen interference in the scientific process”. In crowdsourcing in science initiatives, tasks are directed to scientists and people from outside the scientific community (Franzoni et al. 2022). Besides, citizen science has a strong tradition in open science, conservation and biodiversity (Palumbo et al. 2022), while crowdsourcing comes from open innovation and business (Beck et al. 2022a). In turn, participatory science is an umbrella concept for a wide range of activities and ways (including citizen science) of involving non-scientists in a research task or tasks. We believe that crowdsourcing in science and participatory science cannot be equated with one another because crowdsourcing in science is a narrower concept.

Research on crowdsourcing in science has so far focused on its importance for science and for society, the functionality of science-dedicated crowdsourcing platforms (Schlagwein and Daneshgar 2014), the circumstances in which scientific crowdsourcing is feasible, desirable or useful for scientists (Law et al. 2017), and the role of researchers in crowdsourcing (Silberzahn and Uhlmann 2015). The majority of the literature is focused narrowly on technical, procedural and efficacy questions, such as quality control measures, potential and effectiveness, crowdsourcing platform requirements (Schlagwein and Daneshgar 2014), barriers to crowdsourcing projects carried out by researchers (Law et al. 2017), identifying the type of community funding appropriate for research, the potential of the virtual community, ways of motivating the virtual community (Baruch et al. 2016), and the characteristics of the virtual community taking part in research projects (Franzoni and Sauermann 2014). Crowdsourcing in science has also been examined in terms of virtual community involvement, the quality of results obtained thanks to their work, the possibilities for controlling them, and outcomes for society (Behrend et al. 2011). Furthermore, existing studies tackle open innovation in science (Beck et al. 2022a), explore the potential possibilities of using crowdsourcing (Correia et al. 2018), analyse citizen participation in online platforms based only on a Brazilian article and cases (Santini and Carvalho 2019), and review ten empirical studies (Aristeidou et al. 2017).

Although the literature provides important information about the conceptualization of crowdsourcing in science, this has been conducted from the various perspectives of life sciences, medical sciences, human and cultural sciences, technical sciences and scientific information (Lukyanenko et al. 2019). Despite the rapid spread of crowdsourcing in scientific practice, and an impressive growth in empirical and theoretical studies in recent years (Sauermann et al. 2019), equally rapid developments in conceptual and theoretical framing have not emerged. The reasons for this situation are disciplinary boundaries (Beck et al. 2022a). One of the conceptualization challenges of crowdsourcing in science is its closeness to other concepts such as: scientific cooperation, team research and problem-solving, and the ideas of open science (Guazzini et al. 2015). The boundaries between crowdsourcing and crowdfunding are also imprecise (Bouncken et al. 2015). Wiggins and Crowston (2011, p. 1) note that “crowdsourcing is insufficiently defined”. Others believe that the understanding of crowdsourcing in science is still problematic (Dunn and Hedges 2012), and that this calls for more conceptual rigor.

Scholars have called for a conceptual framework for crowdsourcing in science (Bassi et al. 2020, p. 9) in order to allow for a better understanding of how it may best be used and by whom (Law et al. 2017). It is useful here to look at the characteristics of crowdsourcing of science. Previous studies referred to crowdsourcing per se (Estellés-Arolas and González-Ladrón-de-Guevara 2012) without reflecting on the specificity of its application to science. Crowdsourcing in science is distinctive from crowdsourcing aimed at innovation processes in terms of the characteristics and configuration of the crowd and the specificity of research tasks (Franzoni et al. 2022).

For all these reasons, we believe that the interest in crowdsourcing in science and the need to advance conceptual rigor in this area is necessary and well-motivated. The current lack of conceptual clarity regarding the notion of crowdsourcing and its characteristics makes it challenging to derive insights applicable to such areas of investigation, and to effectively translate research findings into practice. This encourages a systematic review of the literature, a synthesis of conceptual advances, and the development of a coherent conceptual framework so as to bring together fragmented knowledge, and better understand the roots and recent developments of crowdsourcing in science (Eklund et al. 2019).

Our study is in line with calls for greater conceptual rigor and consensus among researchers of crowdsourcing in science by proposing an integrative definition for the term. Therefore, the aim of this study is to establish an integrative definition of crowdsourcing in science, that is to rigorously identify its characteristics and to develop a conceptual framework useful in guiding further research. The identified research gap gives rise to the following research questions (RQs):

RQ1. What are the core characteristics of crowdsourcing in science?

RQ2. To what extent are these characteristics crucial for the definition of crowdsourcing in science?

Encouraged by invitations, “independent literature reviews will continue to be written as a result of their necessity, importance, relevance, and urgency when it comes to advancing knowledge” (Kraus et al. 2022, p. 14), and given the exploratory aims of our study and the complexity of the issue under scrutiny, in order to address our research questions by synthesizing an integrative definition of crowdsourcing in science and developing a conceptual framework, we adopted a systematic mixed-study review (SMSR) (Magistretti et al. 2020; Tranfield et al. 2003), which is particularly useful in addressing complex phenomena in various contexts (Hong and Pluye 2019). Our SMSR examines crowdsourcing in science as a “between-domain hybrid” concept (Kraus et al. 2022, p. 7), which results from its interdisciplinary nature (Bücheler and Sieg 2011). This review type consists of integrating qualitative and quantitative analysis techniques to a systematically extracted body of literature. Based on an SMSR, 62 articles published in peer-reviewed journals extracted from the Web of Science database are thematically analysed (Braun and Clarke 2006) to synthesize an integrative definition of crowdsourcing in science.

For a better understanding of crowdsourcing in science and to suggest theoretical frames, we adopted the theoretical framework of value co-creation (Galvagno and Dalli 2014), according to which numerous parties are encouraged to become involved in a specific task, while each of those who perform and commission the task obtains benefits on this account. This is consistent with the empirical manifestation of crowdsourcing in science, which allows academics to engage with other researchers and individuals from outside the academic community in the creation of scientific knowledge (Franzoni et al. 2022). This provides the opportunity to create value for stakeholders, which is part of the commitment “both by the government and senior members of the research community to real, active, meaningful and consistent public engagement and dialogue” (Burchell et al. 2009, p. 9). The theoretical value co-creation framework effectively addresses the repeated postulates of taking into account the interests of the community, the applicability of scientific knowledge, and maintaining a balance between the use of material and non-material resources when creating scientific knowledge.

Our research advances knowledge about crowdsourcing in science in two ways: by developing a rigorous definition which captures its core characteristics, and by offering a conceptual framework including the initiator, crowd, process and technology. Hence we contribute to a better understanding of how to best use crowdsourcing for research and determine its actors and related tasks. Secondly, our integrative definition may foster research in various disciplines, while increasing the transparency of research on crowdsourcing in science.

This remainder of the paper is organized as follows. The next part focuses on outlining the essence of crowdsourcing in science and distinguishing it from related concepts, in particular: citizen science, open citizen science and crowd science. The methodological approach is presented in the third part. The fourth part focuses on findings, in particular the characteristics of crowdsourcing in science including crowd, initiator, process and technology. In this section, we provide an integrative definition for crowdsourcing in science. The last part of the article indicates the theoretical contribution, implications for policy makers, main limitations and directions of future research.

2 Conceptual background

The term 'crowdsourcing' was coined by Howe (2006, p. 1) to mean the “act of taking a function once performed by employees and outsourcing it to an undefined (and generally large) network of people in the form of an open call”. Since then, interest in crowdsourcing has increased rapidly. Research on crowdsourcing has been conducted on industry, non-profit organizations, the public sector, governments and local governments (Lenart-Gansiniec 2021; Lenart-Gansiniec and Sułkowski 2020). Crowdsourcing is a process or a means of leveraging crowd input or contributions to projects and activities. Crowdsourcing is combined with problem solving, the transparency and openness of organizations, the improvement of business processes, open innovation, competitive advantage, scarce and valuable resources, data collection, mapping and sharing, idea generation, and crisis management and disasters (Sari et al. 2019).

The phenomenon of crowdsourcing is well recognized. Researchers have identified and analysed its dimensions, components, typology, attributes, motivation and ways that crowds and initiators work, crowdsourcing initiative mechanisms, potential benefits, and the ethical aspects of crowdsourcing (Williamson 2016). In turn, crowdsourcing in science, along with related approaches, has initiated debates about the way crowdsourcing might be perceived (Uhlmann et al. 2019). In particular, researchers have rooted crowdsourcing in science in: (1) citizen science, (2) open citizen science, and (3) crowd science. Firstly, citizen science is thought to be “a form of research collaboration involving members of the public in scientific research projects to address real-world problems” (Wiggins and Crowston 2011, p. 1). Some have suggested that citizen science is linked with citizen participation in some elements of the research process (Chilvers and Kearnes 2020). Mäkipää et al. (2020) suggest that citizen science comprises relationships between scientists and the public in scientific research: the public can be involved in collecting and analysing data, proposing research questions, conducting monitoring, disseminating research results, and obtaining feedback. Others suggest that citizen science is particularly useful in ornithology (Greenwood 2007), ecology (Bonney et al. 2009), and monitoring biodiversity (Szabo et al. 2010), for example mapping locations where bird species are present and there is a relative abundance of birds (Brotons et al. 2007). Citizen science does not require any dedicated internet platforms: “citizen science is not necessarily ‘open science’, a term that refers to open source-like practices in formal scientific research settings” (Wiggins and Crowston 2011, p. 2). Secondly, online citizen science involves the transferring of citizen science projects to the Internet (Holliman and Curtis 2014). Some researchers of online citizen science use alternative terms such as citizen cyberscience or virtual citizen science (Wiggins and Crowston 2011). Curtis (2018, p. 5) differentiates citizen science from online citizen science and claims that “as a result of these developments, some citizen science projects are conducted entirely via the Internet and participants help to analyse large sets of data that have been provided by the project scientists”. Finally, Franzoni and Sauermann (2014) suggest the term “crowd science”, which refers to large-scale projects involving thousands of participants via a platform. To solve problems, participants mainly use input data made available by researchers. According to Franzoni and Sauermann (2014), crowd science combines approaches such as “citizen science”, “networked science”, and “massively-collaborative science” (Nielsen 2011; Wiggins and Crowston 2011). Some researchers combine crowdsourcing in science with crowd science (Scheliga et al. 2018). This view is justified because “volunteer engagement in knowledge creation is by no means a new phenomenon, but it is one that has gained momentum with the emergence of online technologies” (Scheliga et al. 2018, p. 516). In this approach, crowdsourcing in science becomes a distinctive, alternative way of conducting research, and equating crowdsourcing in science with related concepts becomes an oversimplification. Crowdsourcing in science refers to the practice of acquiring the services, ideas or content needed through online collaboration and participation. In contrast, in citizen science-based projects, amateur volunteers are only involved in collecting data and monitoring certain situations.

The literature stresses that crowdsourcing in science is an alternative to research projects (Vachelard et al. 2016), a digitized version of citizen science which promotes research collaboration between scientists and volunteers (Eklund et al. 2019), a form of citizen science (Scheliga et al. 2018), a strategy for researchers' work (Lukyanenko et al. 2019), a tool for research (Law et al. 2017), open innovation in scientific practices (Beck et al. 2022a), a new way of conducting scientific and research activity and an example of open science and co-production of scientific knowledge, an alternative way to research projects (Lukyanenko et al. 2019), a participatory method of enquiry (Tauginienė et al. 2020), organizing virtual participation in knowledge production (Wiggins and Crowston 2011), or a form of public engagement (Scheliga et al. 2018). Scholars agree that crowdsourcing in science is a concept that requires in-depth analysis because the direct transfer of findings on issues based on citizen science or open innovation in science does not take into account the specificity of crowdsourcing in science (Beck et al. 2022a, b; Uhlmann et al. 2019). Despite emerging contributions, the conceptualization of crowdsourcing in science continues to be fragmented (Beck et al. 2022a). Uhlmann et al. (2019, p. 727) indicate that crowdsourcing holds great potential, but lacks the conceptual frameworks that might prove useful in harnessing it. Therefore, it is important to precisely define the key features that characterise crowdsourcing in science (cf. Rubin and Babbie 2008).

3 Methodology

In order to generate an integrative definition of crowdsourcing in science and to develop a conceptual framework, we adopted a systematic mixed-study review, which consists of integrating qualitative and quantitative analysis techniques and applying them to a systematically extracted body of literature (Hong and Pluye 2019). In order to identify and review how crowdsourcing in science is defined in academic research, we used the systematic process outlined by Tranfield et al. (2003). We performed an in-depth review of the 62 publications identified. In particular, we focused on the identified vulnerabilities and on the recommendations of other researchers (Kitchenham 2004; Webster and Watson 2002). Next we mapped, linked and synthesized 42 extracted definitions of crowdsourcing in science.

3.1 Data collection

Our mixed systematic literature review is oriented towards a structured search for a definition of crowdsourcing in science. We limited our search to one electronic database, Web of Science, for several reasons. Firstly, as compared to other available databases, such as ProQuest, Scopus or Emerald, Web of Science is more established, due to its robustness, convenient interface, and the availability of different sorting, ranking and refining features (de lo Santos et al. 2020). Access to the Web of Science is easier than to other databases, as it is possible through the Internet library systems of most universities—which is important for the reproducibility of our review. Secondly, compared to other electronic databases, including Scopus (Jacsó 2011; Linnenluecke et al. 2019), Web of Science provides an appropriate and comprehensive collection of relevant scientific literature because it is a holistic database that aggregates the content of around 3,400 journals across 58 social science disciplines. Its use has been widespread in published research using bibliometric techniques (Rey-Martí et al. 2016) including in crowdsourcing research (Malik et al. 2019). Web of Science comprises “access to multidisciplinary research studies (i.e., peer-reviewed and grey literature) from the fields of sciences, social sciences, arts and humanities” (Skarlatidou et al. 2019, p. 4). Thirdly, many significant publications providing new definitions of concepts at an early stage of development are based on one electronic database only, that is Web of Science (Granstrand and Holgersson 2020; Kaartemo 2017; Trischler and Li-Ying 2022).

We decided not to search for additional sources of grey literature such as Google Scholar.com as its value is still debated in the literature: “Google Scholar does not support many of the features required for systematic searches (…) Google Scholar's coverage and recall is an inadequate reason to use it as principal search system in systematic searches” (Gusenbauer and Haddaway 2020, p. 211). For the objectives of our study, Google Scholar does not provide a systematic search function, which may impinge on the reproducibility and clarity of our systematic literature review.

In order to illustrate the publication search criteria and the source inclusion and exclusion criteria, we used the PRISMA guidelines (Preferred Reporting Items for Systematic Reviews and Meta-Analyses) (Moher et al. 2010) (Fig. 1).

Fig. 1
figure 1

PRISMA flow diagram

We started the collection of existing definitions from the literature with a basic search in the Web of Science electronic database on October 15, 2020. We assumed that the overall structure of searching and filtering the literature would be broad and inclusive and focused on gathering different literature streams in order to comprehensively collect all articles that may contain reports of crowdsourcing in science. As crowdsourcing in science is interdisciplinary (Bücheler and Sieg 2011) and due to the fact that the literature postulates conducting interdisciplinary research into crowdsourcing involving “sociology, psychology, management science, economics, computer science, and artificial intelligence” (Lenart-Gansiniec 2022, p. 18), we did not narrow the database search by subject.

The publication search timeframe was limited to between 2006 and October 2021. The starting point refers to the publication year of Howe’s (2006) seminal article on crowdsourcing. As the terminology is not standardized within the topic domain of crowdsourcing in science, two different sets of keywords were used to extract relevant resources. Considering that the search for the right terminology becomes a search for certain ideas and concepts, our search was optimized to return as many relevant documents as possible (Boell and Cecez-Kecmanovic 2014). Consequently, a keyword search was performed to extract publications from databases for analysis. To ensure comprehensibility, we employed the following search terms:

TITLE: (“online citizen science” OR “crowdsourcing citizen science” OR “crowdsourced science” OR “crowdsourcing science” OR “citizen cyberscience” OR “virtual citizen science” OR “crowd science” OR “crowd research” OR “scientific crowdsourcing” OR “science 2.0” OR “crowdsourcing in science” OR “crowdsourcing research” OR “crowdsourcing for science” OR “academic crowdsourcing”) AND YEAR PUBLISHED: (2006–2021) AND DOCUMENT TYPES: (Article) Timespan: All years. Indexes: SCI-EXPANDED, SSCI, AandHCI, CPCI-S, CPCI-SSH, BKCI-S, BKCI-SSH, ESCI, CCR-EXPANDED, IC. Initial searches yielded 551 studies. Subsequently, in the second stage of the search, we used additional criteria to narrow down the articles to those containing "crowdsourcing in science" in the title, abstract and keywords. As summaries are the key to fully understanding the argumentation of the original article, we read the titles and abstracts of each of the identified articles. Most of them were not strictly related to crowdsourcing in science. We adopted the following three exclusion criteria: (1) publications that showed no relationship to crowdsourcing in science; (2) publications where there was no definition of crowdsourcing in science; (3) publications related to citizen science and online / virtual citizen science. Therefore, we excluded 449 articles, leaving a sample of 102 articles. In the next step, we imposed inclusion criteria: full-text, peer-reviewed English-language scientific articles. However, books, book chapters, conference materials, reviews and editorial introductions were excluded. The purpose of this procedure was to ensure a higher level of quality control of the publications obtained. On this basis, we isolated a total of 78 articles. In the last step, we re-verified the publications in terms of compliance with the adopted exclusion criteria, which contributed to the exclusion of 18 articles from further analyses and the identification of the final set of 60 publications.

To ensure high-quality methodological rigor and standards, journal articles were included, but not editorials, conference papers, books or book chapters (Thyer 2010). Additionally, we searched for articles devoted to crowdsourcing in science that had been published in the top ten higher education journals: Assessment and Evaluation in Higher Education, Higher Education, Higher Education Research and Development, International Journal of Educational Technology in Higher Education, Internet and Higher Education, Journal of Computing in Higher Education, Journal of Higher Education, Research in Higher Education, Review of Higher Education, and Studies in Higher Education. The selection was guided by recommendations made by Tight (2018) and by the results offered in the Scimago Journal and Country Rank 2021. However, this search did not produce additional results in the form of publications.

To avoid the potential risk of missing any key definitions and to update our search, we reiterated our Web of Science search procedure in October 2021with the adopted inclusion and exclusion criteria. After removing duplicates, we found 10 additional publications on crowdsourcing in science. After the abstracts were verified according to the exclusion criteria, 2 studies were included. Our final data set consisted of 62 research articles covering a range of different disciplines such as business and economics, education, psychology, biology, education and communication studies. We identified 42 definitions of the term “crowdsourcing in science” in our literature database.

3.2 Data analysis

The aim of our study was to establish an integrative definition for crowdsourcing in science and to develop a conceptual framework. Therefore, the definitions that were identified through the systematic literature review were subject to content analysis (Berelson 1952). Content analysis is both systematic and rigorous, as well as highly flexible (White and Marsh 2006). In this context, “content analysis is a research method that provides a systematic and objective means to make valid inferences from verbal, visual, or written data in order to describe and quantify specific phenomena” (Downe-Wambolt 1992, p. 314). Content analysis can take on a quantitative or qualitative approach, applied either inductively or deductively. Our research concentrated on development of the conceptualization of crowdsourcing in science. We adopt a mixed methods approach as it effectively addresses concerns of subjectivity voiced with regard to coding and thematic analysis, as well as concerns of shallow interpretation directed towards quantitative analyses. The aim was to become immersed in the data, which is why the written material was read through several times. In our study, we travelled back and forth between frequency analysis, coding and thematic analysis. We therefore analysed all the collected definitions in accordance with the guidelines of Wacker (2004). We carefully and repeatedly read all the definitions in terms of conceptual clarity and their recognition of the basic properties and features of the concept under consideration. We placed emphasis on the semantic aspect of the identified definitions of crowdsourcing in science. For this reason, we started with qualitative content analysis, which is concerned with theory generation and the exploration of meanings, and focuses on the systematic classification process of coding and identifying themes or patterns (Shannon and Hsieh 2005, p. 1278), while deriving rich meanings and insights from the text. That is, the contents were summarized under themes, without coding but with notes. The aim was to become immersed in the data through several rounds of reading (Polit and Beck 2006). The purpose of the qualitative content analysis was to break down the words from individual definitions into their component parts based on the rules of grammatical analysis, which in practice meant the separation of verbs from nouns and adjectives.

After making sense of the data, we used deductive coding to look for pre-defined and existing terms. Each component was assessed for its semantic purpose and then coded. We structured our findings according to three characteristics of crowdsourcing (Estellés-Arolas and González-Ladrón-de-Guevara 2012, p. 6): “(1) crowd: who forms it., what they have to do, what they get in return; (2) initiator: who it is, what they get in return for the work of the crowd; (3) process: the type of process it is, the type of call used, the medium used”. Moreover, we identified an additional component, that is “technology”. We believe that these four building blocks have reach a saturation point for covering all aspects of the definition of crowdsourcing in science.

The next stage of the analysis consisted of summarizing relevant pieces of information from the 42 definitions. A series of summaries were performed by two investigators in order to avoid discrepancies, the inclusion of incorrect data, or the omission of significant data. During the analysis, each action taken was noted down. The information collected was re-verified, combined and synthesized. These activities were aimed at maintaining coherence so as to coordinate the linking of topics. As a final step, the findings were summarized and related to the key characteristics of crowdsourcing in science.

We assessed the internal validity of our synthesis through investigator triangulation, which is recommended in mixed approaches (Archibald et al. 2015). These activities were carried out by two independent researchers, each of them tasked with conducting the same analyses, which made it possible to compare the results. After completing the task, the two independent researchers met and calibrated the obtained results together. This allowed the risk of observer bias and other cognitive biases of the researchers to be reduced. In the last step, we assessed the internal validity of our synthesis by investigator triangulation. This ensured the transparency and replicability of our content analysis.

In the last stage, we analysed the identified components of crowdsourcing in science based on frequency analysis (see Appendix). This procedure was aimed at illustrating the abilities of the components of crowdsourcing in science. Additionally, we used NVivo 1.5 to automate this process. This procedure led us to develop a theoretical framework.

4 Findings

Before moving onto the analysis of the extant definitions of crowdsourcing in science, we present here an overview of the different characteristics of crowdsourcing in science, including crowd, initiator, process and technology. For each characteristic, we provide a detailed overview of the findings and insights, linked with the systematic literature review results.

4.1 Characteristics of crowdsourcing in science

In this section, we present the results of content analysis conducted with regard to the characteristics of crowdsourcing in science. We identified four elements: initiator, crowd, process and technology. These are referred to as the characteristics of crowdsourcing.

4.1.1 The initiator

Generally, the initiators in crowdsourcing in science are academics (Schildhauer and Voss, 2014). This category may include researchers (Del Savio et al. 2016; Edgar et al. 2016; Majima et al. 2017; Williams 2013), scientists (Curtis 2018), or academics (Pan and Blevis 2011). Some studies suggest that crowdsourcing in science is particularly popular among “early-career researchers from less well-known institutions, underrepresented demographic groups, and countries that lack economic resources and may never have a fair chance to compete” (Uhlmann et al. 2019, p. 713). This may suggest that crowdsourcing is used as a viable way of addressing resource constraints. On the other hand, Beck et al. (2022a) suggest that older male researchers demonstrate more openness to crowdsourcing in science. However, they also state that in the context of crowdsourcing in science, there is some discrepancy in the set of personal characteristics, including gender and age. This could suggest that reputation may play an important role in successfully carrying out crowdsourcing.

4.1.2 The crowd

The crowd is generally viewed as “constituted by a large collectivity of people who find themselves in direct spatial contact and who react spontaneously, without reflection, and imitatively to common stimuli and the co-presence of others” (Lenart-Gansiniec 2016, p. 31). In this context, the crowd is made up of geographically dispersed anonymous members of a virtual community who possess varied knowledge, skills and experience (Karachiwalla and Pinkow 2021). The crowd is also defined by the openness to virtually anyone who wishes to participate (Estellés-Arolas, González-Ladrón-De-Guevara 2012). The literature on crowdsourcing does not provide any clear definition of crowds, in particular for crowdsourcing in science. Instead, it formulates a description of the crowd in the context of crowdsourcing in science. Our analysis revealed two forms of the crowd in crowdsourcing in science. The majority of the articles available indicate that the crowd consists of the following individuals: amateur or scientists (Table 1).

Table 1 Crowd forms in crowdsourcing in science

The majority of authors do agree on a general definition of the term “crowd” and indicate that apart from the crowd composition and type of people involved, heterogeneity and the skills possessed are also important. Some have suggested that the notion of the crowd refers to a large (Beck et al. 2019; Poetz and Schreier 2009; Wexler 2011; Wiggins and Crowston 2011), indeterminate (Wexler 2011), undefined (Steelman et al. 2014; Wiggins and Crowston 2011), anonymous (Steelman et al. 2014), geographically distributed (Pan and Blevis 2011), varying in knowledge and heterogeneity, numerous group (Petersen 2013). Others claim that it is difficult to determine who most frequently participates in crowdsourcing in science initiatives (Peer et al. 2017). This stems from the fact that crowd participants do not represent any particular population, and differ from one another nationally. Some authors believe that the crowd mainly consists of Americans and Indians (Berinsky et al. 2012), or Europeans and South-East Asians (Sulser et al. 2014). Others, meanwhile, refer to the Japanese (Majima et al. 2017). According to Ipeirotis (2009), workers in crowdsourcing in science initiatives are predominantly female (70%), aged 21–35 (54%), and from 60 countries outside of the United States. Ross et al. (2010), found evidence that on aggregate, approximately 52% of the crowd are female.

Therefore, we conclude that the crowd in crowdsourcing in science refers to a large, indeterminate, undefined and geographically distributed group of amateurs and professionals whose characteristics of age, gender, skills and knowledge are determined by the specifics of the crowdsourcing in science initiative. In the literature, it is noted that crowdsourcing in science aims to provide opportunities for those outside major research institutions to contribute to the development of interesting projects, thus increasing inclusivity, contributions and investment yield (Uhlmann et al. 2019).

With reference to crowd motivation for participating in crowdsourcing in science initiatives, we identified a few motivators, which may be grouped into:

  • extrinsic motivations: receiving satisfaction from a given type of economic or social recognition (Petersen 2013; Wang and Yu 2019), winning a prize for participation (Keating and Furberg 2013), reputation (Pan and Blevis 2011);

  • intrinsic motivations: self-esteem, development of individual skills (Petersen 2013; Wang and Yu 2019), an individual's inborn desire and feelings of competence, satisfaction (Keating and Furberg 2013), enjoyment (Keating and Furberg 2013; Pan and Blevis 2011), sense of pride, enjoyment in their activities on the platform, desire to take part in science, the possibility of collaboration, new knowledge acquisition, gaining knowledge (Wang and Yu 2019), contribution to the accumulation of scientific knowledge (Parrick and Chapman 2020), need for cognition (Berinsky et al. 2012), experience (Wang and Yu 2019), affiliation to a team and tenure (Pan and Blevis 2011), helping scientists, participation in social activities by joining a community, learning (Behrend et al. 2011; Parrick and Chapman 2020), personal interest, and fun (Parrick and Chapman 2020).

4.1.3 The crowdsourcing process

Crowdsourcing refers to a collaborative process (Pan and Blevis 2011; Beck et al. 2022a), which may involve: problem solving (Bassi et al. 2020), creativity, innovation, knowledge capture (Franzoni et al. 2022), collection of information (Edgar et al. 2016) or data (Franzoni and Sauermann 2014; Law et al. 2017; Majima et al. 2017; Sheehan 2018), theory development (Mason and Suri 2012), support in the identification of new research problems (Beck et al. 2022a, b), interactions among group members leading to knowledge development (Law et al. 2017), and workflow tasks, assets, processes and outputs (Hedges and Dunn 2017). Our analysis covers several tasks that the initiator can delegate to the crowd (Table 2).

Table 2 Task types in crowdsourcing in science

In prior research, tasks related to the research process were especially highlighted.

We structure our findings according to the seven stages of the research process (Beem et al. 1965) (Fig. 2).

Fig. 2
figure 2

Crowdsourcing research process

Apart from the tasks listed above and performed by the crowd in crowdsourcing in science, authors have demonstrated that crowdsourcing may prove itself to be helpful in the case of peer reviewing (Uhlmann et al. 2019), communication between volunteers and scientists (Bonney et al. 2016; Parrick and Chapman 2020; Scheliga et al. 2018), replicating published findings (Moshontz et al. 2018; Uhlmann et al. 2019), deciding which findings to pursue further (Uhlmann et al. 2019), contributing to scientific knowledge (Parrick and Chapman 2020), knowledge dissemination, knowledge integration, knowledge application, knowledge absorption, knowledge consumption (Wang and Yu 2019), and knowledge creation (Friesike et al. 2015).

4.1.4 Technology

All authors that mention the medium to be used clearly refer to Internet platforms designed for collaborative research (Law et al. 2017; Stewart et al. 2017). These platforms provide an interface and act as an intermediary for cooperation between the researcher and members of the virtual community of non-professionals and professionals (Schildhauer and Voss 2014; Stewart et al. 2017; Woodcock et al. 2017). The platforms are tools to be used to facilitate accessing knowledge from the crowd, and ​ “provide an online space for collective place-based knowledge, experience and wisdom to be captured, shared, exchanged, contested and negotiated” (Corbett and Cochrane 2019, p. 2). Additionally, crowdsourcing platforms play an intermediary role in the relationships and communication between researchers and those performing social tasks (Eklund et al. 2019). They also support initiators in crowd management, the creation and processing of tasks delegated to the crowd, assuring quality and providing motivational systems, and profiling potential crowd members (Hirth et al. 2017). Some authors suggest that platforms for crowdsourcing in science are alternatives to labour market platforms (Law et al. 2017; Shank 2016), while technological support is provided by science 2.0 (Bücheler and Sieg 2011), and Web 2.0 (Bonney et al. 2009; Goodchild, 2007; López and Olvera-Lobo 2018).

Such crowdsourcing in science platforms include Amazon Mechanical Turk (AMT; https://www.mturk.com/), which is designed for supporting research, collecting data (Shank 2016; Sheehan 2018) or conducting experiments (Mortensen et al. 2018). AMT is recommended for social sciences research (Behrend et al. 2011). Employees are remunerated for their work (Lovett et al. 2018). Other platforms dedicated to supporting research are Prolific Academic (https://www.prolific.co/), Crowdflower (https://visit.figure-eight.com/People-Powered-Data-Enrichment_T) and Clickworker (https: //www.clickworker.com/). These three examples of crowdsourcing platforms are gaining in importance and are becoming ever more popular tools for conducting research (Gleibs 2017), as indicated by publications based on data collected via crowdsourcing in science platforms (e.g., Baldwin et al. 2015; Greenaway et al. 2015; Hui et al. 2015).

In addition to platforms, there are also academic social networking sites such as ResearchGate or Academia, which are not strictly crowdsourced. However, they assist researchers in establishing contact with one another, as well as obtaining help from large groups regarding the correctness of research questions or for seeking answers to methodological issues (Heinrich et al. 2015; Tarrell et al. 2013). All in all, crowdsourcing in science platforms and social networking sites for researchers are both tools that support research and facilitate relatively inexpensive access to unlimited resources of knowledge and skills (Bassi et al. 2020).

Employing crowdsourcing in science is closely connected with the type of call used to propose a task to the crowd. Our literature analysis reveals that a call takes the form of an ‘open call to participate’ (Eklund et al. 2019; Franzoni and Sauermann 2014; Law et al. 2017; Mason and Suri 2012; Schrögel and Kolleck 2019; Wiggins and Crowstron 2011), and an open research question (Beck et al. 2022a, b; Petersen 2013). Many authors believe that an open call can be one of the following two types: (1) the crowd as a large, undefined network of people (Lukyanenko et al. 2020; Schildhauer and Voss 2014; Shapiro et al. 2013; Wiggins and Crowstron 2011); (2) groups of topic experts (Schildhauer and Voss 2014; Uhlmann et al. 2019), specialists from different disciplines (Lakhani et al. 2007), other researchers, research labs or experts (Schildhauer and Voss 2014).

4.2 Integrative definition of crowdsourcing in science

The conceptual history of crowdsourcing in science differs from the conceptual history of crowdsourcing. Crowdsourcing as such is focused on providing innovative solutions to organizational problems, creating new products or services, or collecting users' opinions about a brand, as well as creating and capturing value (Tucci et al. 2018). Crowdsourcing in science, meanwhile, is associated with the creation of scientific knowledge and the involvement in this process of society in general, open science and an increase in the impact of science on public opinion (Beck et al. 2022a, b; Wiggins and Crowston 2011). With regard to crowdsourcing, there is a consensus among researchers as to what crowdsourcing actually is (Kietzmann 2017). However, crowdsourcing in science has not been precisely defined because it is usually understood through the prism of the benefits that can be obtained and the tasks transferred by scientists to the crowd (Uhlmann et al. 2019), and is also associated with online citizen science. In recent years, several studies have offered definitions or descriptions of crowdsourcing in science, with blurred lines between the two. We have therefore generously interpreted what could be considered as definitions in our systematic review of 62 publications on crowdsourcing in science and their key references, identifying 42 more or less unique definitions.

Additionally, empirical descriptions of crowdsourcing in science initiatives often show the importance of the relationship between different characteristics. For example, Bassi et al. (2020) and Wexler (2011) believe that all characteristics are necessary for crowdsourcing in science, and that they are related to one another. Moreover, the connection between individual characteristics in crowdsourcing in science is more or less omitted from the accepted definitions, despite its importance. We have included these links in the conceptual framework of crowdsourcing in science (Fig. 3).

Fig. 3
figure 3

Conceptual framework of crowdsourcing in science

Moreover, we have identified several different ways of capturing each of the characteristics. For instance, many definitions of crowdsourcing in science lack a clear indication of exactly who the crowd is. In crowdsourcing, this is usually the general Internet public (Estellés-Arolas and González-Ladrón-de-Guevara 2012), but in crowdsourcing in science, the crowd includes both non-professionals: the general Internet public (Bassi et al. 2020) and professionals such as the scientific community (Franzoni and Sauermann 2014). Therefore, we ran a coding procedure driven by the three characteristics.

Thus, we map the various meanings attributed to the crowd, the various tasks seen as essential to the process, and the roles of the initiators. Interestingly, we find one additional characteristic of crowdsourcing in science that emerges from the various ways of referring to technology, such as: “online”, “web”, “internet”, “platform”, “digital”. After coding and aggregating the four crowdsourcing characteristics, we re-ran frequency analyses for the 1st and 2nd order categories and found the process in 92.85% of definitions, the crowd in 78.57% of definitions, technology in 23.80% of definitions, and the initiator in 19.04% of definitions. By running NVivo 1.5, the five highest frequency keywords in the data mined definitions from between 2010 to 2020 were found to be as follows: research (19), crowd (15), data (13), large (13), open (11), process (11), tasks (11), online (10), science (10), and new (8).

Conceptualizations of social behaviour are expected to balance generality, simplicity and accuracy (Weick 1979). Based on the 42 definitions and the characteristics of crowdsourcing in science, we propose a broad, integrative definition of crowdsourcing in science which includes characteristics designated as differentia specifica: the crowd (forming the crowd, the task to be completed by the crowd, what the crowd gets in return), the initiator (who they are, what the initiator gets in return), the process (the type of process, the type of call to use), and technology (the medium). We thus propose the definition: Crowdsourcing in science is a collaborative online process through which scientists involve a group of self-selected individuals of varying, diverse knowledge and skills, via an open call on the Internet and/or online platforms, to undertake a specified research task or set of tasks.

Our definition starts with the notion of “process”, understood to be a series of steps and interactions designed to complete a research project. We establish that crowdsourcing in science is a collaborative process (Tavanapour and Bittner 2017). By collaborative process, we mean an effort in which people work together. However, crowdsourcing does not imply collective work, but tasks performed by individuals. Single or multiple tasks may be crowdsourced, ranging from data collection, through data analysis, to problem solving.

The collaborative process is designed by the researcher, who makes some tasks available to the crowd, while keeping control over others, and over the research process. Hence, prior literature consensually attributes a leading role to the researcher, as opposed to bottom-up, democratized (Uhlmann et al. 2019) research formulation. By involving the crowd, researchers solicit inputs from a large number of both professionals and non-professionals, who self-elect to contribute. Hence, the crowd refers to an unspecified set of participants working individually on tasks without interacting with each other or with the initiator. This process is manageable thanks to technology, which helps participants to connect and carry out tasks in a geographically distributed way at low cost. We derive a conceptual framework for crowdsourcing in science from our integrative definition. Our framework for crowdsourcing in science relates to a collection of references to existing behaviour units, and has specified process start and end conditions. We also took into account the links between characteristics postulated by other researchers (Bassi et al. 2020; Wexler 2011).

5 Discussion and conclusion

This paper establishes an integrative definition of crowdsourcing in science by rigorously identifying its characteristics, and developing a conceptual framework useful in guiding further research. Crowdsourcing in science is in line with the postulates of recombining the creation of scientific knowledge towards its democratization and inclusiveness. It is consistent with the postulates of Society 5.0 (Palumbo et al. 2022). Therefore, crowdsourcing in science becomes a novel, alternative strategy for conducting scientific research, which uses advanced Internet technologies to involve other researchers and people from outside the scientific community in the implementation of a research task or tasks. Due to the potential of crowdsourcing in science, it is not surprising that it has attracted attention in various disciplines. Clarifying the concept of crowdsourcing is important both due to the rapidly increasing use of such practices across multiple disciplines, and because specifying the characteristics of a phenomenon (Podsakoff et al. 2016) helps develop robust categories comprehensible to academics (Suddaby 2010). Thus, we have responded to the challenges presented in the literature regarding the need to better understand and define what crowdsourcing is.

Consistent definitions are necessary, as they allow researchers to communicate and to better understand a given concept. Therefore, we did not narrow down our search only to the field of management. Instead, we build on and integrate the findings of other researchers dealing with crowdsourcing in science, taking into account the lenses of various disciplines. Using rigorous methods, we responded to calls on the need to develop a solid nomenclature related to crowdsourcing in science and to provide a conceptual framework. However, this would not have been possible without distinguishing the components of crowdsourcing in science, as imposed by the guidelines and recommendations for creating definitions (Wacker 2004; Suddaby 2010).

Firstly, crowdsourcing in science is not a homogeneous concept. Therefore, our rigorous analyses have allowed us to distinguish the components: initiator, crowd, process and technology. Bearing in mind the results of the frequency analysis, researchers who offer a definition of crowdsourcing in science most often focus their attention on the process (82 times), then the crowd (39 times), technology (16 times) and initiator (10 times). Identifying the components sheds light on the difficulties and confusion surrounding crowdsourcing in science. Moreover, it indicates that in crowdsourcing in science, it is not only technology that is important, but also the human and process aspects.

Secondly, we believe that each of these elements is equally important and should be included in the integrative definition and conceptual framework. We provide an integrative definition for researchers of various disciplines or fields of science, while increasing the transparency of research on crowdsourcing in science. Our findings are in line with the current challenges in the literature regarding the need to provide a theoretical framework (Law et al. 2017; Eklund et al. 2019). We extend the findings of Beck et al. (2022a) and we take the theory of value co-creation perspective, which allows us to propose a unification of crowdsourcing in the conceptual framework of science. Nevertheless, it should also be noted that crowdsourcing in science should where possible be distinguished from other synonymous concepts such as citizen science, crowd science and participatory science.

The definition we propose is based on a systematic analysis of the literature and the existing definitions of crowdsourcing in science. We think of crowdsourcing in science as a collaborative online process through which non-professional and/or professional scientists incorporate a group of individuals of varying diversity of knowledge and skills, via an open call to the Internet and/or online platforms, to undertake a science-related task. Our definition indicates that crowdsourcing in science is a complex concept in which initiator, crowd, process and technology are important.

Our in-depth analysis of literature extracted from the Web of Science database was conducted by combining a mixed systematic literature review with content and frequency analysis. Our study contributes to the development of theory and has multiple practical implications that open the way for rigorously harnessing the potential of crowdsourcing in science. Taken together, our results have theoretical, practical and policy implications.

5.1 Implications for theory

Our work contributes to advancing an understanding of crowdsourcing in science, as presented by authors such as Beck et al. (2022a). Despite the progressive interdisciplinarity of knowledge and research in the field of crowdsourcing in science, its understanding remains problematic (Beck et al. 2022a, b; Sauermann et al. 2019), which encourages attempts to advance conceptual clarity (Eklund et al. 2019). Moreover, research on crowdsourcing in science is advised because of its potential in the context of scientific development and the transition from individual work to team building and distributed collaboration, which leads to an increase in the scale and impact of scientific research (Uhlmann et al. 2019).

This led us to update prior frameworks based on the collaborative approach involving scholars from multiple disciplines, and rooted in the context of science, including problems, antecedents, boundary conditions and impact (Beck et al. 2022a). Our review of 42 crowdsourcing in science definitions shows that they tend to focus on potential crowd-directed tasks, and less frequently take into account key characteristics, thus displaying a substantial degree of fragmentation.

The mixed method approach we used shows that crowdsourcing in science is a heterogeneous concept. Our analysis was initially structured around the three characteristics of crowdsourcing: the initiator, the crowd and the process. We also identified a fourth characteristic of crowdsourcing, adding technology to the three existing ones in the literature. A higher number of attributes increases the sharpness of definitions because it reduces the number of cases. We include all characteristics in our integrative definition. We provide a conceptual framework helpful in further advancing research on crowdsourcing in science by refining its characteristics, guiding scrutiny of its empirical manifestations, and developing typologies.

Secondly, based on the theory of value co-creation (Galvagno and Dalli 2014), we provide a detailed insight into the various ways each of the four characteristics are approached in the literature. By doing so, we outline design decisions for successfully running crowdsourcing in science. Some researchers compare crowdsourcing in science initiatives to “digital sweatshops” (Zittrain 2009) and “share-the-scraps economy” (Reich 2015). This leads to doubts about the quality of the results, and the very fact of a researcher initiating a project based on crowdsourcing in science may be poorly received in the immediate environment. Researchers who reach for crowdsourcing in science may even be ostracized by the scientific community (Shirk et al. 2012). Therefore, ensuring rigor is of the utmost importance. Design and assessment decisions range from identifying the specific task to be outsourced, selecting ways of soliciting individuals, to reaching out to appropriate contributors. Hence, we contribute to advancing the debate on design and rigorous assessment of crowdsourcing in science.

5.2 Implications for practice

Identification of the basic components of crowdsourcing in science casts light on the explanation of the complexity of this concept. It helps not only to connect an interdisciplinary view, but also to develop a coherent stream of research. Crowdsourcing in science is not only about technology or inviting the crowd into different stages of the research process. The identified components and the working definition suggest that using crowdsourcing in science requires the development of an initiative. In addition to preparing an invitation for the crowd to participate, it is necessary to focus on the very process of crowdsourcing in science. When choosing the right platform for a given crowdsourcing initiative, the initiator should remember that understanding crowd motivation is very important. Both financial and non-financial incentives may be effective. Additionally, clear protocols for the crowd, crowd-aimed briefings or training, user-friendly tools and interfaces, effective data validation methods, and the provision of crowd feedback are likely to increase in importance.

Secondly, academics should not ignore or disregard the participation of the crowd and other researchers in their research. The crowd may become a valuable member of project teams, and researchers themselves may become a part of the postulates for creating interdisciplinary teams and democratizing science. In addition, crowdsourcing in science project initiators may utilise the enormous resources of human knowledge and creativity with relatively low financial outlays and in a shorter time. Moreover, according to the demands of "Responsible Research and Innovation" included in the European Commission's policy and the guidelines of Science in Society of the Horizon 2020 strategy—scientists should be aware of and take into account the participation of citizens in their scientific research. It would be a missed opportunity if academic researchers failed to use these new research strategies to increase scientific collaboration and citizens’ involvement, as these are factors which may ultimately contribute to the inclusion of society at large in research, and increase citizens’ confidence in science.

5.3 Implications for policy makers

Our findings contribute to the lively discussions on crowdsourcing in science and may give rise to important policy implications. All in all, the most important component of crowdsourcing in science refers to its initiator. As the results of our frequency analysis, so far researchers have paid least attention to this component. Yet the success of crowdsourcing depends on researchers themselves (Hui et al. 2015), their intrinsic motivations, beliefs about the potential and effectiveness of crowdsourcing (Riesch and Potter 2014), including their ability to collaborate with individuals who are not their peers and who might be in geographically distant locations. Therefore, it is necessary and important to prepare researchers for using crowdsourcing in science, in particular to show its benefits—thanks to which they will see the possibilities and legitimacy (Taeuscher et al. 2021) of its use in scientific work. Universities are encouraged to support academics in reaching for crowdsourcing in science through active support structures in their ecosystems and also skills development initiatives, and to create a culture based on public participation and scientific ethos.

5.4 Limitations and future research agenda

Our research has some limitations. The first results from the specificity of the content analysis carried out. This is an objective and precise technique, recommended for the conceptualization and identification of the characteristics of a given concept, and allows for a systematic analysis of the text. However, as a descriptive method it may omit the motives underlying a given pattern. Secondly, our systematic literature review only considers peer-reviewed publicly available scientific articles. We have not included post-conference materials or books in our analyses. Therefore, future research may consider the inclusion of post-conference materials and books in the systematic literature review. Thirdly, our research is based on the assumption that we need the characteristics of crowdsourcing in science so as to develop an integrative definition and conceptual framework. This may limit the results obtained, as the conceptual framework may depend on the types of crowdsourcing in science (Bassi et al. 2020). This opens the way for conceptual efforts carried out both from theoretical perspectives, thus detached from extant literature, as well as empirical scrutiny that would single out manifestations of the phenomenon rather than integrating extant views of these manifestations.

Crowdsourcing in science requires separate research streams. A clarification of what crowdsourcing in science actually is will allow researchers to take the next step, in this case towards the operationalization of this concept. Further research progress is needed in the field of crowdsourcing in science. In particular, we pay attention to the need to provide measurement scales that enable assessment of the level of academics' awareness of the potential of crowdsourcing in science.