Towards Artificial Intelligence Augmenting Facilitation: AI Affordances in Macro-Task Crowdsourcing

Crowdsourcing holds great potential: macro-task crowdsourcing can, for example, contribute to work addressing climate change. Macro-task crowdsourcing aims to use the wisdom of a crowd to tackle non-trivial tasks such as wicked problems. However, macro-task crowdsourcing is labor-intensive and complex to facilitate, which limits its efficiency, effectiveness, and use. Technological advancements in artificial intelligence (AI) might overcome these limits by supporting the facilitation of crowdsourcing. However, AI’s potential for macro-task crowdsourcing facilitation needs to be better understood for this to happen. Here, we turn to affordance theory to develop this understanding. Affordances help us describe action possibilities that characterize the relationship between the facilitator and AI, within macro-task crowdsourcing. We follow a two-stage, bottom-up approach: The initial development stage is based on a structured analysis of academic literature. The subsequent validation & refinement stage includes two observed macro-task crowdsourcing initiatives and six expert interviews. From our analysis, we derive seven AI affordances that support 17 facilitation activities in macro-task crowdsourcing. We also identify specific manifestations that illustrate the affordances. Our findings increase the scholarly understanding of macro-task crowdsourcing and advance the discourse on facilitation. Further, they help practitioners identify potential ways to integrate AI into crowdsourcing facilitation. These results could improve the efficiency of facilitation activities and the effectiveness of macro-task crowdsourcing.


Introduction
Artificial intelligence (AI) holds the potential to transform collaborative activities such as crowdsourcing (Griffith et al. 2019;Introne et al. 2011;Kiruthika et al. 2020;Manyika et al. 2016;Seeber et al. 2020). In crowdsourcing, a crowd collaborates to solve a task in a digital participative environment, such as an online platform (Estellés-Arolas and González-Ladrón-de-Guevara 2012). Thereby, the crowd may be diverse, including individuals from diverse disciplinary backgrounds (Cullina et al. 2015;Dissanayake et al. 2019). When a crowd is dedicated to tackling complex and interdependent tasks collaboratively, the practice is referred to as macro-task crowdsourcing (Robert 2019;Schmitz and Lykourentzou 2018). Macro-tasks are tasks that are difficult or sometimes impossible to decompose into smaller (interdependent) subtasks (Robert 2019). The use of crowdsourcing to address macro-tasks is rarely straightforward and requires a specific skill set and knowledge of the crowd (Schmitz and Lykourentzou 2018). A prominent example of macro-tasks are wicked problems. Wicked problems are highly complex and thus require the involvement of many different stakeholders (Alford and Head 2017;Head and Alford 2015;Ooms and Piepenbrink 2021). Global challenges that are very broad in scope, such as the advancement of the sustainable development goals as defined by the United Nations (2015), may be understood as wicked problems of current relevance (McGahan et al. 2021). In response to these problems, existing macro-task crowdsourcing initiatives such as OpenIDEO or Futures CoLab elaborate on sustainability-related improvements and solution approaches Kohler and Chesbrough 2019).
For macro-task crowdsourcing to realize its potential and tackle such complex problems, structure, guidance, and support are needed to coordinate the collaborating crowd workers (Adla et al. 2011;Azadegan and Kolfschoten 2014;Shafiei Gol et al. 2019). If this need is satisfied through unbiased (human) observation and intervention, it is known as facilitation (Adla et al. 2011;Bostrom et al. 1993). Although facilitation has already been widely analyzed in other contexts such as group interaction (Bostrom et al. 1993), face-to-face meetings (Azadegan and Kolfschoten 2014), and open innovation (Winkler et al. 2020), it has barely been investigated in macro-task crowdsourcing. AI is seen as a system's ability to interpret and learn from external data to achieve a predetermined goal . With AI breaking human text challenges (Wang et al. 2019), new potentials arise, especially for textbased applications like crowdsourcing. The high transformative potential of AI gives rise to the question: Can AI support the facilitation of macro-task crowdsourcing? If it can, the quality of crowdsourcing results might be improved, leading to better outcomes and results. For example, an AI with semantic text understanding could recognize novel or innovative-yet-unrecognized ideas and highlight these as focal points for further discussion within the crowd (Toubia and Netzer 2017). Furthermore, by relieving the bottleneck of labor-and knowledge-intensive facilitation, macro-task crowdsourcing could be applied to more wicked problems.
AI and facilitation may be closely interwoven in macro-task crowdsourcing. Facilitation, in the specific context of macro-task crowdsourcing, requires human facilitators as well as technological advancements which support the former by fulfilling a large variety of burdensome activities (Briggs et al. 2013;de Vreede and Briggs 2019;Franco and Nielsen 2018;Khalifa et al. 2002;Seeber et al. 2016;Winkler et al. 2020). Among other duties, a facilitator is responsible for understanding the problem to be tackled by the macro-task crowdsourcing, motivating and guiding the crowd and its dialogues, and making sense of the outcome. Lately, AI -as one specific technological advancement -has been investigated for its supportive potential (Rhyn and Blohm 2017;Seeber et al. 2016;Tavanapour and Bittner 2018a). AI carries various functionalities, including text mining or natural language processing, that can support macro-task crowdsourcing facilitation. For instance, intelligent conversational agent systems could guide the crowd through the crowdsourcing process (Derrick et al. 2013;Ito et al. 2021) or issue detailed instructions to crowd workers in the form of specific tasks (Qiao et al. 2018). The evaluation of the workers' contributions could also be drastically simplified by designing appropriate systems that leverage the potential of text mining and natural language generation to automatically generate reports or summaries (Füller et al. 2021;Rhyn et al. 2020). Such AI-augmented facilitation systems could improve human facilitation (Adla et al. 2011;Siemon 2022), (partially) automate facilitation processes Jalowski et al. 2019;Kolfschoten et al. 2011), or even wholly replace the facilitator with an AI agent (de Vreede and Briggs 2019).
Although AI has considerable potential in macro-task crowdsourcing and assisting human problem solving (Rhyn and Blohm 2017;Schoormann et al. 2021;Seeber et al. 2020), there are only a few AI-related contributions in the literature on macrotask crowdsourcing or crowdsourcing facilitation. A holistic understanding of how AI could be applied to facilitate problem-solving in on-or offline groups is missing. However, such a holistic understanding is necessary to guide further research on crowdsourcing and facilitation and inform practitioners as to how crowdsourcing initiatives might be improved. We set out to investigate how AI can and may enable macro-task crowdsourcing facilitation. Therefore, we pose the following research questions (RQs), which address both the identified lack of research into macro-task crowdsourcing facilitation and the need for a holistic understanding of AI in this given context: RQ1: Which activities comprise macro-task crowdsourcing facilitation?

RQ2: What action possibilities does AI afford for macro-task crowdsourcing facilitation?
We apply a two-stage, bottom-up approach to establish a theory-driven understanding validated and refined using practical insight to answer these research questions. In our approach, we turn to affordance theory (Volkoff and Strong 2017), which is known to help develop better theories in IT-associated transformational contexts (Ostern and Rosemann 2021). Given AI's high potential to transform crowdsourcing, affordance theory can be seen as an established, suitable, and meaningful lens to theorize the relationship between the technological artifact of AI and the goaloriented actor -namely, the facilitator (Lehrer et al. 2018;Markus and Silver 2008;Ostern and Rosemann 2021;Volkoff and Strong 2013). In the first stage, we develop initial sets of macro-task crowdsourcing facilitation activities and AI affordances.
Both sets are based on a structured search and review of extant scholarly knowledge. The second stage validates and refines our facilitation activities and AI affordances. We observe two real-world macro-task crowdsourcing initiatives and perform six interviews with experts from the crowdsourcing facilitation and AI domain, thus, including insights from practice.
For RQ1, our results provide a detailed understanding of macro-task crowdsourcing facilitation comprising 17 facilitation activities. We answer RQ2 by developing a set of seven AI affordances relevant to macro-task crowdsourcing facilitation. We also detail manifestations of the affordances that demonstrate actionable practices of AI-augmented macro-task crowdsourcing facilitation. Our findings increase the scholarly understanding of macro-task crowdsourcing facilitation and the application of AI therein. Furthermore, the results will help practitioners to evaluate potential ways of integrating AI in crowdsourcing facilitation. These results will increase the efficiency of facilitation activities and, ultimately, increase the effectiveness of macro-task crowdsourcing.
The remainder of the paper is structured as follows: Sect. 2 provides theoretical background on macro-task crowdsourcing facilitation, AI-augmented facilitation, and affordance theory. We outline our research process in Sect. 3. Section 4 presents the macro-task crowdsourcing facilitation activities, the AI affordances, and the manifestations of AI in macro-task crowdsourcing facilitation. After discussing the implications and limitations of our results in Sect. 5, we conclude with a brief summary in Sect. 6.

Macro-Task Crowdsourcing
Crowdsourcing is an umbrella term that can have many meanings. The concept was first introduced in an article in Wired magazine (Howe 2006b). Elsewhere, Howe (2006a) defines crowdsourcing as "the act of a company or institution taking a function once performed by employees and outsourcing it to an undefined (and generally large) network of people in the form of an open call." Since then, understandings of crowdsourcing have evolved. Estellés-Arolas and González-Ladrón-de-Guevara (2012) proposed a holistic definition that we will use in this paper: "Crowdsourcing is a type of participative online activity in which an individual, an institution, a non-profit organization, or company proposes to a group of individuals of varying knowledge, heterogeneity, and number, via a flexible open call, the voluntary undertaking of a task." A panoply of different crowdsourcing types exists, ranging from corporate to social or public contexts (Vianna et al. 2019). In a corporate context, open innovation is used to strategically manage knowledge flows between an external crowd and a firm to improve the firm's innovation processes (Bogers et al. 2018). With crowdfunding, entrepreneurs can receive funding, via an open call, from funders who may receive a private benefit in return (Belleflamme et al. 2014). Organizations can use micro-task crowdsourcing to outsource low-complexity tasks (e.g., image tagging, or phone number verification) completed by independent crowd workers (Hossain and Kauranen 2015;Schenk and Guittard 2011). More complex tasks (e.g., invention or software engineering) require collaboration among crowd workers (Kittur et al. 2013). Flash organizations, for example, are computationally built structures comprised of a crowd automatically arranged into a hierarchy, where participants are assigned to smaller units focused on complex tasks according to their particular skills (Valentine et al. 2017). The structure of the resultant crowd organization can adapt over time, allowing it to efficiently collaborate and achieve open-ended goals relating to complex tasks (Retelny et al. 2014;Valentine et al. 2017). Real-world problems can be approached by using citizen science, a participative way of performing research involving experts and non-experts (Hossain and Kauranen 2015;Wiggins and Crowston 2011). For example, Fritz et al. (2019) underline the scientific value of citizens' contributions of data which helped to track the progress of the United Nations' sustainable development goals.
As a step beyond predominant crowdsourcing types, we define macro-task crowdsourcing based on Leimeister (2010) Macro-task crowdsourcing leverages the collective intelligence of a crowd through facilitated collaboration on a digital platform to address complex or wicked problems.
The problems being addressed with the help of macro-task crowdsourcing may range from open innovation product design, to software development, or to grand social challenges (Kohler and Chesbrough 2019;McGahan et al. 2021) like climate change (Introne et al. 2011). Many of these are rooted in wicked problems characterized by their high complexity and their need to elicit broad stakeholder involvement (Alford and Head 2017;Ooms and Piepenbrink 2021). Macro-task crowdsourcing differs from existing crowdsourcing types in several ways. Although the boundaries between macro-task and, for example, micro-task crowdsourcing are blurred, there are some distinguishing characteristics, which are presented in Table 1.
The fact that the problem cannot easily be broken down into smaller constituent parts means it requires a high level of crowd diversity -i.e., providing multiple perspectives from experts with different levels of expertise and knowledge in various disciplines (Lykourentzou et al. 2019;Robert 2019). Due to the complexity of the underlying problem and the broad stakeholder involvement, a guiding, moderating, and neutral central agent is necessary, which we refer to as a facilitator . It is important to note that the results produced by the crowd will not necessarily be the final solution to the overarching problem. Existing macro-task crowdsourcing initiatives such as Climate CoLab (Introne et al. 2013), Futures CoLab , and OpenIDEO (Kohler and Chesbrough 2019) tend to produce valuable but non-conclusive approaches to addressing a wicked problem from one specific angle. These approaches have evolved and matured during several guided phases Introne et al. 2013), making macro-task crowdsourcing even more reliant on a facilitator and a clear understanding of its role within the crowdsourcing initiative.
The panoply of different crowdsourcing types has produced a variety of terminologies with synonyms and ambiguities now requiring unification. Figure 1 depicts an abstract view of existing terms and definitions within crowdsourcing. Generally, we use the term macro-task crowdsourcing initiative to refer to an overarching set of online activities that aim to address a problem (Estellés-Arolas and González-Ladrón-de-Guevara 2012). We refer to a crowdsourcing exercise as a whole process of crowdsourcing techniques (Vukovic and Bartolini 2010) that may be applied multiple times or in combination with other exercises as part of a macro-task crowdsourcing initiative.
The context of an exercise is highly relevant. A macro-task crowdsourcing initiative may conduct multiple exercises in various (e.g., geographical) environments, using different strategies or workflows (e.g., to address the problem) with different infrastructural prerequisites (e.g., hard-and software). The nature of the problem being tackled by an exercise influences how a task is designed and, therefore, how contributions are generated (Zuchowski et al. 2016).
In an exercise, three different groups of people participate. The requestor is an organization or an individual that seeks help, the worker is part of a help-offering crowd capable of (partially) addressing or solving the requestor's problem (Pedersen et al. 2013). The facilitator acts as a crucial intermediary who tries to understand the requestor and facilitates the crowd of workers to reach a predefined goal concerning the problem (Franco and Nielsen 2018;Gimpel et al. 2020;Rippa et al. 2016). From an activity-driven perspective, exercises consist of three major process phases: preparation, execution, and resolution. While preparation refers to "breaking down a problem or a goal into lower level, smaller sub-task" (Vukovic and Bartolini 2010), execution describes the elaboration on the task by a diverse crowd of workers (Zuchowski et al. 2016) supported and guided by one or more facilitators. Evaluating and synthesizing the workers' contributions finishes the last process phase, termed The requestor or the digital platform's algorithm performs repetitive and standardized patterns of actions The facilitator or facilitating teams guide process phases with high degrees of freedom for the workers Generated Outcome Aggregable and structurable solutions to the problem Approaches to addressing the problem, which are difficult to synthesize resolution (Lopez et al. 2010). IT enables all participants to collaborate online in a distributed or decentralized way. Typically, a digital platform is used to capture and store the interactions and communication between individuals (Lopez et al. 2010). Interactions on a digital platform for macro-task crowdsourcing can include rating, creation, solving, and processing (Geiger and Schader 2014). Sometimes tools like video communication software are used to run the exercise more efficiently or effectively. To steer the exercise within the given context, governance, using a dedicated strategy, creates suitable boundary conditions for the people (Blohm et al. 2018;Pedersen et al. 2013). While rules can define norms or desired conducts, roles govern responsibilities and accountabilities, and culture creates a desirable and productive atmosphere for collaboration.
Each exercise results in different types of outcomes (Zuchowski et al. 2016). Contributions represent manifestations of work on the communicated and processed task. We see tacit knowledge gained during the exercises as learnings that could, for instance, be achieved by reflections or feedback. We distinguish these from consequences, which represent immutable conclusions that have been caused by performing the exercise (e.g., unsatisfied workers who will not contribute to future exercises). Finally, every stakeholder of the macro-task crowdsourcing initiative can perceive value in the exercise. While the requestor could, for example, see value in the synthesized contributions, a worker could perceive value in social recognition within the crowd.
Behind each of these terms, there is a whole range of activities that, taken together, should be carefully aligned with a goal during the macro-task crowdsourcing initiative, in order to contribute to the overarching problem. Thus, facilitation is important to ensure proper alignment and goal orientation. Thereby, facilitators play an essential role, particularly -yet, not only -during the exercises.

Facilitation in Crowdsourcing
In the crowdsourcing domain, crowdsourcing governance aims to facilitate workers in performing their tasks and steer them toward a solution (Pedersen et al. 2013;Shafiei Gol et al. 2019). According to Shafiei Gol et al. (2019), whether crowdsourcing governance is centralized or decentralized, the task is to control and coordinate workers on the crowdsourcing platform. This involves activities such as defining the task (Zogaj and Bretschneider 2014), providing proper incentives , ensuring the quality of the contributions (Blohm et al. 2018), and managing the community and its culture (Zuchowski et al. 2016). Crowdsourcing governance is often analyzed in environments involving paid work or smaller tasks (Blohm et al. 2018;Shafiei Gol et al. 2019). Hence, activities like controlling costs and standardizing procedures also gain relevance (Shafiei Gol et al. 2019). Despite extensive frameworks Shafiei Gol et al. 2019;Zogaj et al. 2015), crowdsourcing governance is often conceptualized on the organizational and platform level, which could explain why it is also referred to as a management activity (Blohm et al. 2018;Jespersen 2018;Pohlisch 2021;Zogaj and Bretschneider 2014). The increasing complexity of the problems under investigation means increasingly sophisticated governance strategies are required to deliver successful crowdsourcing initiatives (Blohm et al. 2018;Boughzala et al. 2014;Pedersen et al. 2013). Since macro-task crowdsourcing initiatives are known for their complex (sometimes even wicked) underlying problems in a collaborative environment, utilizing facilitation can be a suitable and effective governance strategy. Facilitation is primarily focused on the crowd, enabling workers to collaborate on complex tasks and, ultimately, reach an overarching goal Kim and Robert 2019;Lykourentzou et al. 2019).
To tackle increasingly complex -often wicked -problems using macro-task crowdsourcing, the facilitation of groups is both highly relevant and very challenging (Khalifa et al. 2002;Shafiei Gol et al. 2019). Following Bostrom et al. (1993), the main aim of facilitation is to ensure unified goal orientation among collaborating workers. This challenging task can require various social and technical skills or abilities to support problem-solving (Antunes and Ho 2001). Researchers have explored several types of facilitation specifically tailored to collaborative settings. Adla et al. (2011) differentiate between four overlapping types: Technical facilitation mainly aims to support participants with technology issues. Group process facilitation strives to ensure all members of a group jointly reach overarching goals such as motivation or moderation. Process facilitation assists by coordinating participants or structuring meetings. Finally, content facilitation focuses on, and introduces changes to, the content under discussion. Facilitators serve as experts practicing techniques to support problem-solving processes (Winkler et al. 2020), for example, in face-to-face meetings (Azadegan and Kolfschoten 2014;Bostrom et al. 1993). Besides completing a burdensome amount of work before, during, and after the collaboration (Vivacqua et al. 2011), facilitators must also evince particular character and behavioral traits (Dissanayake et al. 2015a). Training and experience (Clawson and Bostrom 1996), appearance and behavior within a group (Franco and Nielsen 2018;Ito et al. 2021;McCardle-Keurentjes and Rouwette 2018), and the handling of feedback and reflection (Azadegan and Kolfschoten 2014;de Vreede et al. 2002) play an essential role here. Thereby, facilitators maintain a delicate balance between situations in which they moderate and observe the group and instances in which they intervene -for instance, due to content-related issues (Khalifa et al. 2002) -without compromising the outcome of the exercise (Dissanayake et al. 2015b). To better assist the group and balance the workload, multiple facilitators with different foci may sometimes be involved, making it possible to split the work among the facilitators (Franco and Nielsen 2018) and maintain a good relationship with all the workers (Liu et al. 2016). However, some scholars note that face-to-face facilitation techniques may be less effective when applied in distributed or virtual environments (Adla et al. 2011). Hence, it is difficult for crowdsourcing facilitators to rely on facilitation knowledge established in other contexts. This difficulty could be rooted in the fundamentally different nature of collaboration on a crowdsourcing platform Nguyen et al. 2013).
Building upon the current, broad understanding of crowdsourcing governance Pedersen et al. 2013;Shafiei Gol et al. 2019) and facilitation (Antunes and Ho 2001;Bruno et al. 2003;Kolfschoten et al. 2011;Maister and Lovelock 1982;Zajonc 1965) offered in the literature -in particular, an existing definition by Bostrom et al. (1993) -we define macro-task crowdsourcing facilitation thus: Facilitation in macro-task crowdsourcing initiatives comprises all observing and intervening activities used before, during, and after a macro-task crowdsourcing exercise to foster beneficial interactions among crowd workers aimed at making (interim-) outcomes easier to achieve and ultimately align joint actions with predefined goals.
Despite substantial knowledge of crowdsourcing governance and facilitation, an overarching and integrated understanding of relevant facilitation activities in macrotask crowdsourcing is missing. Therefore, it is challenging for facilitators to delimit their competencies in crowdsourcing endeavors involving many participants and perspectives (Zhao and Zhu 2014).

Advances in AI-Augmented Facilitation
AI uses technologies and algorithms to simulate and replicate human behavior or achieve intelligent capabilities (Alsheibani et al. 2018;Simon 1995;Stone et al. 2016;Te'eni et al. 2019). AI may be defined as a "[…] system's ability to correctly interpret external data, to learn from such data, and to use those learnings to achieve specific goals and tasks through flexible adaptation […]" . Although more general definitions exist -such as those by Rai et al. (2019) and Russell and Norvig (2021) -in this paper, we follow the definition by Kaplan and Haenlein (2019). Their socio-technical system perspective focuses on the interrelationship between humans and AI, which is highly relevant in the context of macro-task crowdsourcing facilitation. While AI has been a subject established in science for over seven decades Rzepka and Berger 2018;Simon 1995), in recent years, it has received increasing attention in both research and practice (Bawack et al. 2019;de Vreede et al. 2020;Hinsen et al. 2022;Hofmann et al. 2021;Leal Filho et al. 2022;Pumplun et al. 2019;Rai 2020). AI is expected to disrupt the interplay between user, task, and technology (Maedche et al. 2019;Rzepka and Berger 2018) and the nature of work Iansiti and Lakhani 2020;Nascimento et al. 2018). This expectation is accompanied by many unrealistic expectations, and the timeless question of "[W]hat can AI do today?" (Brynjolfsson and McAffe 2017). There is a stream of AI research that answers this question using terminology usually related to humans or animals, including intelligence, learning, recognizing, and comprehending (Asatiani et al. 2021;Benbya et al. 2021;Rai et al. 2019), that explicitly considers human-inspired AI and humanized AI . For example, Hofmann et al. (2020) answer the question of "what can AI do today?" by providing a structured method to create AI use-cases applicable to various domains. Thereby, they distinguish seven abstract functions, defined in Table 2, through which AI can occur as a solution: perceiving, identification, reasoning, predicting, decision-making, generating, and acting (Hofmann et al. 2020). However, such approaches may also lead to the over-humanization of AI and should not distort the fact that AI systems are human-made artifacts, not humans.
AI differs in the presence of cognitive, emotional, or social intelligence . To be able to best support -or even replace -the facilitator would require an AI to hold all three types of intelligence, thereby resulting in a selfconscious and self-aware humanized AI (de Vreede and Briggs 2019; Kaplan and Haenlein 2019). Humanized AIs are not yet available in the facilitation domain, which could either be due to the complexity of collaboration (Kolfschoten et al. 2007) or the limited capabilities of current AI systems (Briggs et al. 2013;Kaplan and Haenlein 2019;Sousa and Rocha 2020). Hence, scholars from the facilitation domain focus on projects and approaches to building human-inspired AIs (Seeber et al. 2018). We refer to these as AI-augmented facilitation systems, which could have a vast impact on team collaboration (Maedche et al. 2019;Seeber et al. 2018Seeber et al. , 2020. For instance, Derrick et al. (2013) and Ito et al. (2021) propose the first results of conversational AI capable of issuing instructions to team members or responding to workers' contributions. Further inspired by the widespread application of AI (Dwivedi et al. 2021;Wilson and Daugherty 2018), researchers have also begun to explore more specifically AI's potential use in crowdsourcing facilitation Rhyn and Blohm 2017;Tavanapour and Bittner 2018a). For instance, some approaches seek to automate facilitation activities and decision-making by integrating AI such as "Acquiring and processing data from the real world to produce information" Identification "Extracting and identifying specific objects from data" Reasoning "Explaining underlying relationships and structures in data" Prediction "Estimating future events or conditions on a continuous scale" Decision-making "Choosing between known, discrete alternatives" Generating "Producing or creating something" Acting "Executing goal-oriented actions (e.g., movement, navigate, control)" text mining or natural language processing . Most of these AIaugmented approaches are prototypes, suggesting that further investigation of possible AI-augmented facilitation may be warranted (Askay 2017;Ghezzi et al. 2018;Robert 2019).

Affordance Theory
Within our research, we use affordance theory as a conceptual lens. Affordances are action possibilities that characterize the relationship between a goal-oriented actor and an artifact within a given environment (Burlamaqui and Dong 2015;Gibson 1977;Markus and Silver 2008). The concept of affordances was initially introduced in ecological psychology to describe how animals perceive value and meanings in things within their environment (Gibson 1977 . Thereby, affordances describe the relationship between an actor and an information technology to determine goal-oriented action possibilities available to the actor and using specific information technology at hand (Faik et al. 2020;Markus and Silver 2008;Volkoff and Strong 2017). Actors can perceive or actualize affordances (Ostern and Rosemann 2021). Perceiving affordances requires that the actor holds a certain level of awareness regarding the information technology and is, hence, able to identify its potential uses (Burlamaqui and Dong 2015; Volkoff and Strong 2017). The information about a perceived affordance can lead actors to an affordance's actualization. Herein, the actor makes efforts to realize the affordance, unleashing the value it holds in relation to the actor's goal (Ostern and Rosemann 2021).
To analyze the "cues of potential uses" (Burlamaqui and Dong 2015) of AI in a specific environment, researchers often turn to affordance theory (Burlamaqui and Dong 2015; Kampf 2019; Volkoff and Strong 2017). In our endeavor, the particular environment is macro-task crowdsourcing with the facilitator as the actor and AI as the specific information technology. This confluence of technology and actor in our macro-task crowdsourcing context is a complex socio-technical phenomenon, where affordance theory can help better understand the interrelationships. With RQ2, we aim to exploratively investigate the relationship between the actor and the technology, revealing the action possibilities of AI in macro-task crowdsourcing facilitation. In line with the original definition by Gibson (1977) and following technology-related affordance literature (Faik et al. 2020;Leonardi 2011;Norman 1999;Steffen et al. 2019;Vyas et al. 2006), we focus on perceived affordances throughout our research endeavor. Hence, we define affordances in our context as perceived action possibilities arising from AI in macro-task crowdsourcing facilitation that do not necessarily need to be performed (Askay 2017). We see these perceived affordances as necessary to compose the nucleus of AI's intersubjective meaning for facilitators (Suthers 2006). The most salient perceived affordances will ultimately support collaboration among the crowdsourcing workers.

Research Design
Our research set out to address the lack of knowledge on macro-task crowdsourcing facilitation and the need for a holistic understanding of how AI might augment facilitation in this context. Thereby, we followed a two-stage, bottom-up approach to establish a theory-driven understanding that we then validated and refined from a practical perspective. In our approach, we turned to affordance theory as an established lens to theorize the relationship between the technological artifact, AI, and the goal-oriented actor, the facilitator (Lehrer et al. 2018; Markus and Silver 2008; Ostern and Rosemann 2021; Volkoff and Strong 2013). Our approach served to identify macro-task crowdsourcing facilitation activities and AI affordances in macro-task crowdsourcing facilitation. Firstly, in the initial development stage, we conducted two literature searches. We identified 17 macro-task crowdsourcing activities and 116 statements about AI in macro-task crowdsourcing being further processed to manifestation (i.e., specific action possibilities) that substantiate AI's potential use for macro-task crowdsourcing facilitation. From this, we identified seven AI affordances for macrotask crowdsourcing. Secondly, we iteratively refined our results in the validation & refinement stage through two observed macro-task crowdsourcing initiatives and six semi-structured interviews with experts from the AI and crowdsourcing facilitation domain. Figure 2 depicts the overarching research design, which yielded seven AI affordances for macro-task crowdsourcing.

Initial Development Stage
We developed an initial set of AI affordances in three steps. The aim in the first two steps was to gain an understanding of facilitation and AI within macro-task crowdsourcing. Thereby, we developed macro-task crowdsourcing facilitation activities necessary for performing the third step, which served to combine relevant insights from literature into an initial set of AI affordances.
In step I) Facilitation activities list, we conducted a structured literature search to extract activities that describe macro-task crowdsourcing facilitation. In an initial broad search, we identified the journal 'Group Decision and Negotiation' as an adequate source of broad, foundational knowledge about facilitation (Laengle et al. 2018). A searched for the term 'facilitation' in this journal returned a total of 176 papers, which we sequentially screened by title, abstract, and full text to Create an initial set of AI affordances in macro-task crowdsourcing Gain holistic understanding of macro-task crowdsourcing facilitation Build broad foundation of AI manifestations in macro-task crowdsourcing I) Facilitation activities list: Structured literature search on facilitation and extraction of 17 macro-task crowdsourcing facilitation activities II) AI in macro-task crowdsourcing: Structured literature search on macro-task crowdsourcing and identification of 116 AI manifestations III) Initial AI affordances: Aggregation of results to an AI manifestation matrix and seven AI affordances in macro-task crowdsourcing facilitation

Goals of Stage
Performed Steps IV) Crowdsourcing initiatives: Observation of two observed macro-task crowdsourcing initiatives to validate macro-task crowdsourcing facilitation activities and to refine the AI manifestation matrix V) Interviews: Six semi-structured expert interviews including a final in-depth analyzation of the experts' feedback to refine the AI manifestation matrix and the seven AI affordances in macro-task crowdsourcing facilitation Iteratively refine AI affordances by receiving practical insights Triangulate lessons-learned from case studies Receive feedback from AI and facilitation experts Fig. 2 Overarching Two-Stage, Bottom-Up Approach determine whether facilitation was the core subject of each article. In doing so, we identified ten papers, plus one additional relevant paper from another outlet (Appendix A.1), whose full-text we further processed. We extracted 477 statements (i.e., excerpts) about activities or capabilities (i.e., repeatable patterns of action) relevant for facilitation. For each statement, we then decided whether the activity or capability was transferable to macro-task crowdsourcing facilitation. We excluded statements if the underlying activity did not necessarily need to be performed by a facilitator (e.g., recruitment of the worker) or if it neither contributed to fostering beneficial interactions among crowd workers or aligning joint actions with predefined goals (e.g., communication of the exercises' results or distributing rewards to workers). We categorized the 317 remaining statements into 17 broader macro-task crowdsourcing facilitation activities that iteratively emerged in the researcher team's discussions. These 17 activities served as comprehensive, foundational knowledge about macro-task crowdsourcing facilitation in the next two steps.
Step II) AI in macro-task crowdsourcing served to capture manifestations of AI in macro-task crowdsourcing. As highlighted above, the digital nature of crowdsourcing platforms means the application of AI in crowdsourcing is more widespread than in other situations where facilitation plays an essential role (e.g., face-to-face meetings). We conducted a systematic literature review on the topic of macro-task crowdsourcing (vom Brocke et al. 2015;Wolfswinkel et al. 2013). In keeping with our research goal of exploring "cues of potential uses" (Burlamaqui and Dong 2015, p. 305) of AI (i.e., affordances), we had identified 'information systems,' 'computer science,' and 'social science' as our fields of research. Hence, we selected four established databases (i.e., AIS eLibrary, ACM Digital Library, IEEE Explore Digital Library, and Web of Science) that covered this broad disciplinary spectrum. Our search query did not include specific AI terms since the literature includes various definitions and terms to refer to corresponding AI technologies (Bawack et al. 2019). Instead, we iteratively developed our search query and ended up with a more general tripartite version representing a process-driven perspective on online crowdsourcing:

('crowd*' OR 'collective intelligence') AND ('task' OR 'activity' OR 'action" OR 'process' OR 'capability' OR 'facilitat*') AND ('platform' OR 'information system' OR 'information technology' OR 'information and communications technology').
Applying this search query to the identified databases resulted in a total of 5,808 hits. To refine our sample of papers, we identified and removed 502 duplicates, which led to 5,306 distinct papers. In manually screening the papers, we applied the criteria listed in Table 3 to narrow our search results to macro-task crowdsourcing and ensure high levels of relevance and rigor.
Using these criteria, we narrowed the search results by sequentially analyzing title and abstract, which narrowed the total to 283 papers potentially relevant to macrotask crowdsourcing. We read these 283 papers in full text, finally identifying nine papers that name and describe AI in the context of macro-task crowdsourcing. We also included three papers, found elsewhere during our research process, that matched all of our defined criteria. We analyzed these 12 papers (Appendix A.2) in-depth to extract 116 statements about AI manifestations in macro-task crowdsourcing. In the next step, these statements were used together with the previous results to develop AI affordances.
In step III) Initial AI affordances, we developed an initial set of AI affordances by combining and aggregating the results of steps I) and II). Thereby, we assigned 116 manifestations (Appendix A.3) of AI in macro-task crowdsourcing to the 17 activities of macro-task crowdsourcing facilitation. To further distinguish and explain the role of AI in each manifestation, we used AI functions proposed by Hofmann et al. (2020). In doing so, we assigned each manifestation one specific AI function, describing how AI occurs or could occur as a solution in the selected manifestation (Hofmann et al. 2020). This two-dimensional matrix resulted in an AI manifestation mapping for macro-task crowdsourcing facilitation.
To create the initial set of affordances, the research team held discussions to identify archetypes within the manifestation mapping. We remained open-minded about whether an archetype would be created based on the functioning of AI (horizontal axis of the matrix) or the facilitation actions (vertical axis of the matrix). To support the development of AI affordances, we reached out to three scholars with expertise in affordance theory. They contributed valuable input regarding common pitfalls and best practices during the development stage. We recognized seven archetypes whose manifestations we then analyzed to identify affordances. Every affordance is described and classified in terms of AI functions and facilitation activity (see Sect. 4.1).

Validation & Refinement Stage
Although we rigorously identified our AI affordances based on scholarly knowledge, a practical validation was necessary to ascertain potential end-users' perceptions. To this end, we validated and refined our initial set of AI affordances from step III) with two observed macro-task crowdsourcing initiatives as well as six semi-structured interviews (Myers and Newman 2007).
In step IV) Crowdsourcing initiatives, we longitudinally observed two macro-task crowdsourcing initiatives, namely Trust CoLab (TCL) and Pandemic Supermind (PSM). Observing these initiatives not only helped us to validate our facilitation activities but also to gain rare practical insights on macro-task crowdsourcing facilitation and real-world AI manifestations. Table 4 describes the two macro-task crowdsourcing initiatives under consideration.
Our primary sources of data collection were documentation (e.g., mails, final reports, and meeting protocols) and participant observation (e.g., discussion within the facilitation team and analysis of AI tools used) that we gained from both crowdsourcing initiatives. Observation of the facilitators' actions in the macro-task crowdsourcing initiatives supported the set of 17 facilitation activities from step I). Each of the activities was observed, and no other major activities were found. Additionally, we could refine and enhance the manifestations within our AI manifestation mapping, which was created in step III), by analyzing the application of AI tools and the perceived demand for AI support within both initiatives. Nevertheless, the limited application of AI tools in both initiatives could not validate all affordances and suggested an additional validation and refinement step. Hence, in step V) Interviews, we conducted six semi-structured interviews, which we used to uncover potential affordances (Volkoff and Strong 2013). We selected experts from academia and practice with multiple years of experience in the AI or facilitation domain, as listed in Table 5 ( Myers and Newman 2007;Schultze and Avital 2011).
Interviews lasted between 37 and 72 min, were held in the native language of the interviewee, and were recorded with the consent of each interviewee. We informed the interviewees about the research topic and sent a detailed interview guide in advance to better allow the interviewees to prepare for the interview. The guide contained definitions and illustrations, the then-current set of affordances, and the intended structure of the interview. Appendix C.1 provides more details about the structure of the interview as well as the prepared questions.
The semi-structured interviews started with a short description of the research project and definitions of crowdsourcing and facilitation necessary to ensure a mutual understanding of crowdsourcing facilitation. After that, we encouraged the interviewees to share their experience of AI within an ideation section (i.e., a less structured and guided part of the interview). Next, we sought open-ended feedback on the affordances by asking questions regarding the completeness, comprehensiveness, meaningfulness, level of detail, and applicability of the criteria in relation to today's crowdsourcing initiatives (Sonnenberg and vom Brocke 2012). During the interviews, we took notes to highlight the experts' essential statements and better respond to the interviewee in the course of the conversation. We iteratively adapted and refined our affordances after each interview. The experts' feedback led us to overhaul one affordance entirely (i.e., workflow enrichment; previously: environment creation) and improve the descriptions of two other affordances (i.e., improvement triggering and worker profiling).
To align all of the practical and theoretical insights gained, we conducted a final reflective refinement after the interviews. Therein, we followed Schreier (2012) to carefully analyze all six experts' statements regarding our predefined criteria and enrich our AI manifestation, mapping with potential use cases of AI within macrotask crowdsourcing facilitation named by the experts. Appendix C.3 contains some exemplary expert quotes. Our concept-driven coding frame (Schreier 2012) comprised two categories: (1) feedback regarding artificial intelligence affordances and (2) potential use-cases of AI within macro-task crowdsourcing facilitation. While the feedback is structured in five subcategories according to our defined criteria, the potential use-cases encompass 17 subcategories representing the facilitation activities developed in step I). We extracted transcripts of all relevant statements from the interviewees and mapped these to our coding frame. Finally, we refined the AI affordances and AI manifestation mapping accordingly. The two validation and refinement steps yielded a validated and refined list of seven affordances and 44 manifestations. Appendix B contains a detailed description of the refinement and Appendix C.2 a description of the validation.

Macro-Task Crowdsourcing Facilitation
Based on our literature search, we identified 17 macro-task crowdsourcing facilitation activities. Table 6 comprises an exhaustive list of activities found in the current literature, from facilitation joining crowdsourcing's specific conditions. We argue that the distinction between a more straightforward administrative activity (e.g., sending invitation emails to the workers) and a more complex facilitation activity (e.g., writing a motivational text for the workers' invitation) can depend on each particular exercise. The borders of this distinction can also be fluid. Nevertheless, it is essential to clearly define the facilitator's role in each exercise to avoid misunderstandings between the facilitator and other stakeholders of the macro-task crowdsourcing initiative (e.g., the platform administrator or the requestor). In the validation & refinement stage, we observed two macro-task crowdsourcing initiatives (i.e., TCL and PSM). By carefully observing the facilitators within TCL and PSM, we were able to identify action patterns that matched the facilitation activities' descriptions. Thereby, we confirmed the existence of all 17 facilitation activities, although their scope varied within the two initiatives under consideration. Table 7 depicts exemplary actions in TCL and PSM, mainly performed by the facilitator, that matched the elaborated description of the 17 facilitation activities. Some of these facilitation activities were AI-augmented (i.e., the facilitator was supported by an AI tool), making both initiatives valuable subjects for further analysis regarding our AI affordances.
Throughout the macro-task crowdsourcing initiatives of TCL and PSM, dedicated teams used two different AI tools to support the facilitators in their work. In TCL, the facilitator was mainly supported in aggregating the workers' contributions between the four phases. All of the workers' contributions were first exported from the platform before a natural language processing Python script preprocessed the contributions (i.e., performing stemming and lemmatization). The script then created a detailed word cloud to provide the facilitator with a broad overview of the main concepts. Finally, the contributions were semantically clustered by the script using the Universal Sentence Encoder algorithm (Cer et al. 2018). We refer to Appendix D.1 for two interim results of the Python script. The results were discussed by the initiative's stakeholders and manually refined by the facilitator and the supporting team. In PSM, the two facilitators were supported by a web application written in R. The application could directly access the latest contribution data via an application programming interface provided by the crowdsourcing platform. Therefore, the facilitators could use the web application's algorithms during internal meetings to discuss the latest contribution data. In TCL, the AI tool only used the codified contributions made by the workers while the web application also used metadata such as comments or likes on the contributions. This metadata allowed a broad set of functionalities such as keyword extraction, topic modeling, word co-occurrences, network analysis, and word searches. We refer to Appendix D.2 for two screenshots of the web application.

Artificial Intelligence Affordances
Given the extensive knowledge base on macro-task crowdsourcing facilitation, we searched for AI manifestations by conducting a second literature search to create an initial AI manifestation mapping. By analyzing the macro-task crowdsourcing initiatives TCL and PSM regarding potential use-cases of AI-augmented facilitation, and gathering statements about potential use-cases for AI in facilitation from the six expert interviews, we were able to refine and extend our initial AI manifestation mapping. Therein, we searched for archetypes of manifestations that could lead to potential affordances. Table 8 lists the final seven affordances of AI for macro-task crowdsourcing facilitation and one example of how AI-augmented facilitation could be implemented in the case of each affordance.
In the following, we describe each of the affordances in detail. Thereby, we explain the relationship between the facilitator's goal and AI within macro-task crowdsourc-Activity Name Description Supporting Literature

Task Design
Decomposition of an overarching problem into small workable pieces that are bundled into tasks to be presented to the workers Antunes and Ho (2001) (2001)  Selection and design of a three-phase workflow (partially) supported by the platform to identify approaches for better pandemic resilience

Worker Motivation
Initial motivational mail that welcomes the workers and highlights the value of the workers' expected contributions to society User profile on the platform was prefilled with a short biography of the worker to value the workers' participation

Contribution Support
Video tutorials and FAQs were designed and made available Quick responses from the facilitators to questions that arose from the workers

Performance Monitoring
Bi-weekly manual report to track the current amount of workers' contributions Automated AI-augmented dashboard to monitor the contribution upload frequency, most used keywords, and topics arising

Tool Usage & Integration
Usage of one generic online crowdsourcing platform that has been customized to suit the scenario development process Integration of one AI tool to support the facilitation activities during and after each exercise Crowd Moderation Active participation by the supporting team in the discussions and contributions from the worker; reports to the facilitator Hosting of live virtual events to catalyze conversations about the topics within the ongoing task among the workers

Crowd Coordination
Continuous facilitator notes (notification sent to the crowd) regarding the current and future steps Creation of worker groups based on their professional background to coordinate parallel task execution in the first exercise; a merging of groups in the second exercise to support crossfertilization of ideas among workers

Participation Encouragement
Sending targeted emails to workers who were not active on the platform Weekly encouragement of the crowd via email to send feedback, which was regularly reflected and integrated by the facilitators

Contribution Evaluation
Iterative reviewing and selection of the contributions after each exercise; removal of duplicate contributions Weekly discussions between the facilitator and the requestor about recent contributions from the workers  ing. To further elaborate on the affordances, we highlight some AI manifestations found in the literature (step II), our two macro-task crowdsourcing initiatives (step IV), or our interviews (step V). These manifestations provide examples of what AI is perceived to afford within macro-task crowdsourcing facilitation. 1) Contribution Assessment. In bringing a macro-task crowdsourcing initiative to fruition, one of the biggest challenges facilitators face is dealing with the number of contributions made by workers (Blohm et al. 2013;Nagar et al. 2016) and " [understanding] all the results from a crowdsourcing exercise in a way that's empirical and meaningful" (Expert 5). AI affords the analysis of contributions such that the quality can be assessed and valuable ideas or relevant content can be extracted. Facilitators could use AI to analyze the content of a contribution via semantical natural language processing ) to determine its novelty or similarity compared to other contributions. (Semi-)automated contribution assessments could decide whether each contribution brings the initiative one step closer to the goal (Haas et al. 2015;Nagar et al. 2016;Rhyn and Blohm 2017). This could involve removing unnecessary information to allow a better assessment by the facilitator further downstream (Expert 4, 6) or to detect outliers by assessing each contribution's relevance to the topic at hand (Case PSM).
2) Improvement Triggering. In crowdsourcing, facilitators often face a 90-9-1 distribution, where only 1% of the workers create nearly all of the contributions (Troll et al. 2017). Since macro-task crowdsourcing heavily relies on active knowledge exchange and idea cross-fertilization between various workers , non-or low-active workers need to be triggered to contribute, thereby stimulating a better thematic discourse (Expert 2, 4). Yet, even if all workers contribute, their contributions may sometimes lack quality; ideas may lack originality or readability. AI affords recognition of individual contributions that are unoriginal or add no value (e.g., due to the existence of similar or identical contributions) (Hetmank 2013;Rhyn and Blohm 2017), and of workers who do not actively participate in the exercise (e.g., through lack of time or attention). Intelligent mechanisms such as personalized nudging (Expert 2, 4, 5) can improve behavior or quality (Chiu et al. 2014;Haas et al. 2015;Riedl and Woolley 2017). One approach would be to use natural language understanding to automatically notify workers during the creation of a contribution that theirs is similar to other available contributions or is not sufficiently comprehensive (Case PSM) -for example, by displaying a uniqueness score (Expert 3).
3) Operational Assistance. When creating a contribution, workers may experience technical difficulties or develop questions regarding idea formulation (Adla et al. 2011;Hosseini et al. 2015). Usually, workers will either stop working on their Activity Name Exemplary Action in TCL Exemplary Action in PSM  contributions or contact the facilitator, who then has to step in and solve the problem (Adla et al. 2011), consuming the worker's and the facilitator's precious time. AI could identify the cause of either process-or technical-related problems and offer assistance. Missing information, which could hinder the workflow, could be identified by AI and provided at the appropriate time (Chittilappilly et al. 2016;Seeber et al. 2016). Robotic process automation (also referred to as intelligent automation technologies) could assign workers appropriate tasks based on the worker's domain knowledge, previous crowdsourcing experience (Expert 4), or a lack of contributions in a specific task (Case PSM). Deep learning algorithms could help translate contributions or overcome language barriers (Expert 1). Pre-trained AI chat assistants could interactively explain the contribution creation process to the workers on a step-by-step basis and answer their questions accordingly (Tavanapour and Bittner 2018a). 4) Workflow Enrichment. To use the workers' time as effectively as possible, facilitators break down the goal of an exercise into smaller tasks (Vukovic and Bartolini 2010). They efficiently integrate these tasks into an effective workflow supported by a crowdsourcing platform (Hetmank 2013). This is usually accompanied by a reduction in special attention to the needs of individual workers. AI could suggest the facil- Semantical natural language processing to remove unnecessary information 2) Improvement Triggering AI affords identification and nudging of non or less-active workers towards higher participation and triggers improvement measures for inadequate contributions.
Nudging during contribution creation based on natural language understanding

3) Operational Assistance
AI affords support for workers through the whole process of contribution development, including the identification of relevant ideas, elaboration of (interim) results, and submission of the final contribution.
AI chat assistants to answer questions during the contribution creation process

4) Workflow Enrichment
AI affords the provision and integration of useful information and knowledge to a predefined workflow, enabling highly productive collaboration among workers.
Natural language understanding to identify mismatches between the facilitator's proposed task and the workers' contributions 5) Collaboration Guidance AI affords collective guidance for workers during their collaboration on the platform in such a way that they will focus on a predefined goal relating to the overarching problem.
Sentiment detection to generate semantic embeddings of the workers' contributions 6) Worker Profiling AI affords analysis of the network of workers to track the skills and activity of individuals as well as to monitor the quality of their created contributions.
(Social) Network algorithms to generate activity reports from the crowdsourcing platform data 7) Decisionmaking Preparation AI affords aggregate outcomes and synthesizes relevant contributions and, therefore, creates a valuable foundation for decision-makers.
Summary generation algorithms to synthesize the free-text contributions of the workers itator integrate additional or new information into the workflow (Chittilappilly et al. 2016;Riedl and Woolley 2017) or adjust the proposed next steps (Xiang et al. 2018). This could lead to a modified workflow or improved effectiveness. Natural language understanding could identify mismatches between the facilitator's proposed task and the workers' contributions, which may be the result of ambiguous task descriptions (Case PSM). Depending on the extent of the worker's domain knowledge, the description of the task could be paraphrased or extended using natural language generation (Expert 1, 5). If workers do not find appropriate resources supporting their idea, natural language processing could identify the topic and the facilitator could then refer the worker to relevant data or scientific sources (Expert 6). 5) Collaboration Guidance. Facilitation, in a narrow sense, involves fostering collaboration and interdisciplinary exchange of information (Expert 2). However, such thematic exchange can go astray and move away from the exercise's actual goal, despite facilitative support. Therefore, facilitators have to decide whether the existing discourse should be maintained or if the worker should be guided in a different direction (Xiang et al. 2018;Zheng et al. 2017). AI affords the evaluation of workers' moods and the direction of the discussion with reference to the content. This provides the facilitator with a better understanding of the current atmosphere among the workers and the thematic focus of their collaboration. On the one hand, automated text mining, like sentiment detection, could generate semantic embeddings of the contributions (Expert 4) (Nagar et al. 2016), which could help to assess the maturity of the collaboration Qiao et al. 2018). On the other hand, word-to-vec algorithms could focus the content of the discussion and uncover unprocessed areas (Expert 2, 3, 5) and frequently discussed topics (Case PSM), or help the facilitator to detect emerging topics (Case TCL).
6) Worker Profiling. Experienced facilitators mobilize the varied skills and expertise of the workers participating in an exercise (Tazzini et al. 2013). However, as the number of workers increases, getting to know one another becomes more difficult, particularly in an online crowdsourcing environment. Hence, facilitators may lack important information about workers, such as their previous experience in crowdsourcing or domain-specific skillsets. AI affords the use of interaction among workers (Dissanayake et al. 2014), as well as information on their backgrounds (Bozzon et al. 2013;Tazzini et al. 2013), to better assess the workers' activity and the characteristics of their collaboration . Natural language generation could be used to process information from the worker's publications or the worker's social media profile to create a summary of the individual's background (Expert 4). However, the rules of platform governance, as defined by the initiative stakeholders, must be upheld in any such investigations in order to avoid ethical concerns on the part of the workers (Alkharashi and Renaud 2018;Kocsis and Vreede 2016;Schlagwein et al. 2019). Alternatively, activity reports could be generated from the crowdsourcing platform via the use of (social) network algorithms (Case TCL) or natural language processing (Expert 1), leading to fully-automated dashboard generation for tracking the workers' activity (Expert 6). 7) Decision-making Preparation. After one or more exercises, the workers will have provided several contributions. Facilitators then have to aggregate and synthesize these contributions into a meaningful foundation for decision-makers, such as a final report (Chan et al. 2016;Gimpel et al. 2020). AI affords support in decision-making preparation and could provide a synthesis, such as a decision template or recommendation for action to the requestors (Hetmank 2013) (Expert 2). Neural networks that have been specifically trained using vocabulary from the exercise's domain could cluster the contributions and highlight unique ideas (Case TCL and PSM) (Expert 2, 5). Natural language understanding could be used to perform question answering based on the contributions, which could help a facilitator interact with the contributions and better understand the workers' ideas, even after the exercise and without contacting the workers (Expert 3). Furthermore, summary generation algorithms could comprehensively synthesize the clustered contributions in ways that are meaningful for the decision-maker (Case PSM) (Expert 1, 2).
Despite the immense potential of AI in macro-task crowdsourcing facilitation, as reflected by the seven affordances, the interviewed experts stressed that researchers and facilitators must carefully consider which facilitation activity should be enabled or performed by AI (Expert 1, 2, 3, 4, 5). AI is prone to biases (Expert 4) and could systematically discriminate against specific workers (e.g., a natural language processing contribution evaluation algorithm could systematically downrate contributions from workers with dyslexia). On top of that, the unreflected use of AI could lead facilitators to blindly believe in the underlying model and thereby reduce the overall level of goal achievement. Experts also argued that AI has limitations in understanding ethical and cultural factors and cannot fully imitate human interactions as facilitators (Expert 1, 5, 6). "I think it is really nice to have a name and a face to identify with a person who is communicating and asking you to do these things." (Expert 5). Furthermore, facilitators should also consider the effort and difficulties during the development: "The art of AI is often not to solve the task, but to explain and teach the AI the task." (Expert 3). Ultimately, the ill-considered use of AI in macro-task crowdsourcing could induce much bias in the outcome of an exercise (Expert 4) or decrease the workers' participation and performance (Expert 6).
Even after a full review, we could not establish a hierarchy among the affordances. Nonetheless, an interlocking of individual affordances cannot be ruled out. To better illustrate the interdependencies of the affordances, the individual facilitation activities, and AI functions, Table 9 shows the revised version of the AI manifestation mapping. This table records all AI manifestations (i.e., specific action possibilities) that occurred during our research process. Every AI manifestation therein was observed either in literature (L), our observed crowdsourcing initiatives (C), or our interviews (I) and describes a possible shape of the corresponding affordance (1)-7)) concerning a facilitation activity or AI function.
We observed that the way AI emerges in macro-task crowdsourcing facilitation is strongly dependent on the nature of the facilitation activity. In particular, we want to highlight two patterns in Table 9: Firstly, we did not find evidence for the AI function perceiving in any of the facilitation activities. We deem this to be due to the very nature of crowdsourcing initiatives since all data relevant to the facilitation of a crowdsourcing exercise has already been processed from the analog world (Hofmann et al. 2020). For instance, a conversation is not performed face-to-face but is stored in codified form on the crowdsourcing platform. Secondly, we argue that culture Table 9 Revised Artificial Intelligence Manifestation Mapping development is highly human-centered and requires empathy or social and emotional intelligence. Hence, there are very few cases wherein AI would have sufficient capabilities to perform this activity.

Theoretical Contribution
In this research, we have addressed two research questions on the intersection of macro-task crowdsourcing, facilitation, and AI. To answer our research questions, our results encompass three novel theoretical contributions for scholars: a more precise understanding of macro-task crowdsourcing, an extensive list of 17 macro-task crowdsourcing facilitation activities, and seven holistic AI affordances in macro-task crowdsourcing.
Our work advances the domain of macro-task crowdsourcing by distinguishing macro-task crowdsourcing from other crowdsourcing types such as micro-task crowdsourcing or flash organizations. In doing so, it highlights the unique features of macro-task crowdsourcing, such as the low level of the problem's decomposability and the nature of collaboration among workers that, together, form the demand for a facilitating instance. We further provide in-depth insights into two real-world macrotask crowdsourcing initiatives, including their particular AI tools, namely TCL and PSM. Both initiatives are dedicated to tackling wicked problems. Scholars can build upon this extensive understanding of macro-task crowdsourcing and better position their work in this area.
Our research contributes to the facilitation domain by deriving 17 macro-task crowdsourcing facilitation activities that holistically theorize facilitation as a suitable governance strategy for macro-task crowdsourcing. We also introduce a broad definition of macro-task crowdsourcing facilitation to merge the specific collaborative circumstances of macro-task crowdsourcing  with the current understanding of facilitation (Bostrom et al. 1993). This definition, along with the 17 activities, extends existing knowledge of facilitation and particular governance strategies for complex tasks in crowdsourcing, and may apply to other types of crowdsourcing or online collaboration. Our extensive understanding of facilitation in macro-task crowdsourcing differs from traditional knowledge of facilitation in that we consider the digital nature of crowdsourcing's collaborations. With a more vital link to the crowd and increased attention to collaboration levels, our facilitation activities also extend existing crowdsourcing governance concepts more focused on the platform or the initiative. Fellow researchers can, for example, harness these activities as a starting point for further investigation in the context of crowdsourcing, which may be expanded over time and with technological advancement.
Finally, we use affordance theory as a socio-technical lens to extend the body of knowledge on AI-augmented facilitation. Our research identifies seven perceived AI affordances in macro-task crowdsourcing and generalized manifestations triangulated from practice, literature, and expertise. These manifestations were structured on seven abstract AI functions (Hofmann et al. 2020). Even though these could be seen to contrast to other (more technical) conceptualizations of AI, they performed well in the socio-technical context of macro-task crowdsourcing, describing how AI (could) occur within the 17 facilitation activities. Our affordances further extend these insights and holistically describe how AI can be applied by the facilitator in macro-task crowdsourcing facilitation. Through the insights from two macro-task crowdsourcing initiatives and six expert interviews, it is clearly demonstrated that AI currently holds only supportive potential for crowdsourcing. Although AI is now delivering super-human performances in some specific tasks, and while the digital starting conditions provided by crowdsourcing are promising, (digital) collaboration as an environment for AI is proving particularly challenging due to the subtle nuances of human interaction. However, we are convinced that AI's potential will continue to increase as technologies evolve and will soon extend to collaboration and automation. Hence, we presume our AI affordances pave the way for AI scholars to undertake further research, for example, by helping scholars to structure future research projects or identify future research trends.

Practical Implications
From a practical perspective, we see two major stakeholder groups benefiting from our research findings: AI developers and facilitators.
Firstly, developers of AI-augmented facilitation systems or functionalities can use our seven affordances as a starting point to identify areas for action or improvement and to implement innovative systems, tools, or functionalities that support facilitation in crowdsourcing. On top of this, our two macro-task crowdsourcing initiatives revealed good practices in which AI functionalities could add value to crowdsourcing exercises. AI developers could pick these insights and integrate AI functionalities, for example, into the crowdsourcing platform accordingly. Developers could also use the AI manifestation mapping to illustrate the status-quo in AI opportunities.
Secondly, the fact that current research on crowdsourcing facilitation lacks insights about AI means practitioners also stand to benefit from our results. With our seven developed and validated AI affordances, we provide guidance on which functionalities could add value when integrated before, during, or after crowdsourcing exercises. Thereby, our observed macro-task crowdsourcing initiatives and interviews with experts point out possible ways to include and integrate AI, which could be highly relevant when setting up new crowdsourcing initiatives. Furthermore, the 44 manifestations within our AI manifestation mapping provide initial indications of which AI functionalities or use-cases have already been considered and help facilitators correctly assess AI's maturity. However, we recognize that the affordances are not equally relevant for all crowdsourcing exercises due to the complexity and variety of the latter. We argue that the list of affordances can also help communicate the usage and integration of AI tools or functionalities of existing crowdsourcing initiatives. This would also foster the exchange of knowledge in macro-task crowdsourcing, which is essential to find new approaches to tackling, for example, wicked problems in practice. Thereby, active and newly created crowdsourcing initiatives could increase their effectiveness and the efficiency of facilitation activities therein.

Limitations and Future Research
Despite the comprehensive nature of our results, grounded in two literature searches and subject to a two-way practical validation with crowdsourcing initiatives and expert interviews, our research features some limitations in both the development and validation of the macro-task crowdsourcing facilitation activities and AI affordances.
The development of our 17 macro-task crowdsourcing activities was based on a structured literature search solely within the journal 'Group Decision and Negotiation'. We deem its broad understanding of facilitation, developed over many years, to be sufficient for use in a crowdsourcing context. Nevertheless, we could have enhanced this literature search by including other journals or databases contributing to the crowdsourcing or facilitation domains.
Regarding the development of our AI research in macro-task crowdsourcing facilitation, we focused on creating perceived affordances. Due to the dynamic of the AI domain in crowdsourcing facilitation, we neither analyzed the actualization of their affordances nor elaborated on their existence (Ostern and Rosemann 2021;Volkoff and Strong 2017). Although we considered different forms of input (i.e., literature, crowdsourcing initiatives, expert interviews), we cannot formally claim that our affordances are complete. We argue that the fast-moving nature of research in AI could also impede efforts to compile an exhaustive list of AI affordances. Hence, increasing the scope of our literature searches could have resulted in a broader knowledge base. Furthermore, since we only analyzed macro-task crowdsourcing facilitation, we cannot verify the applicability or direct transferability to other types of crowdsourcing or collaboration. Testing whether our results can be generalized to other types of crowdsourcing and collaboration remains a task for future research.
Regarding the validation and refinement of our affordances, we only observed two macro-task crowdsourcing initiatives. We performed six interviews which, by their very nature, were both prone to bias (Yin 2018), such as response or selection bias. Even though we did not receive more insights at the end of our validation stage, we acknowledge that more initiatives or interviews could further improve or enhance our results. For instance, although we interviewed two experts with implementation knowledge, we can only make limited statements about the feasibility of implementing all affordances. Besides, even though experts have confirmed completeness, we still cannot guarantee that we have developed a complete list of affordances.
We hope that future research will address these limitations, offering multiple avenues for further investigation. Future research could specifically look into the practicability of the affordances. We anticipate that implementing AI-augmented facilitation prototypes for macro-task crowdsourcing based on our affordances will deliver valuable insights as to the feasibility of our affordances and their level of abstraction. Such prototypes could also enhance knowledge of actualized affordances or AI information systems in general and shed more light on how AI-augmented crowdsourcing could more efficiently tackle macro-tasks and their underlying complex problems. Furthermore, scholars could go beyond the descriptive nature of our results and (prescriptively) elaborate on how and why particular affordances enable macro-task crowdsourcing facilitation and how this could improve macro-task crowdsourcing initiatives. They also could use our affordances to extend design knowledge to design AI-augmented facilitation assistants (Maedche et al. 2019;Volkoff and Strong 2017). For example, researchers could derive design guidelines or principles that would aid facilitators in their burdensome amount of work ).

Conclusion
Our research was driven by opportunities that emerge from AI's widespread application to aid facilitators. We turned to affordance theory to analyze AI's potential applications in macro-task crowdsourcing facilitation. We answered our two research questions by defining macro-task crowdsourcing facilitation, constituting 17 activities, and introducing seven perceived AI affordances of macro-task crowdsourcing facilitation. We followed a two-stage, bottom-up approach consisting of an initial development stage comprising two literature searches, and a (second) validation & refinement stage, involving two macro-task crowdsourcing initiatives and six expert interviews. Our results could increase the efficiency of facilitation activities and the effectiveness of macro-task crowdsourcing, ultimately contributing to tackling wicked problems, such as the sustainable development goals.  Adla et al. (2011) A Proposal of Toolkit for GDSS Facilitators Antunes and Ho (2001) The Design of a GDSS Meeting Preparation Tool Azadegan and Kolfschoten (2014) An Assessment Framework for Practicing Facilitator Briggs et al. (2013) Facilitator-in-a-Box: Process Support Applications to Help Practitioners Realize the Potential of Collaboration Technology Franco and Nielsen (2018) Examining Group Facilitation in Situ: The Use of Formulations in Facilitation Practice Khalifa et al. (2002) The Effects of Process and Content Facilitation Restrictiveness on GSS-Mediated Collaborative Learning Kolfschoten et al. (2007) Issues in the Design of Facilitated Collaboration Processes Kolfschoten et al. (2011) Modifiers   Haas et al. (2015) Argonaut: Macrotask Crowdsourcing for Complex Data Processing Kittur et al. (2013) The Future of Crowd Work Nagar et al. (2016) Accelerating the Review of Complex Intellectual Artifacts in Crowdsourced Innovation Challenges Qiao et al. (2018) Feedback Based High-Quality Task Assignment in Collaborative Crowdsourcing Rhyn and Blohm (2017) Combining Collective and Artificial Intelligence: Towards a Design Theory for Decision Support in Crowdsourcing Rhyn et al. (2017) Understanding     Haas et al. (2015) Determining the quality of a task Predicting Task Design Kittur et al. (2013) Tasks may be structured through multi-stage workflows in which workers may collaborate either synchronously or asynchronously. As part of this, AI may guide (and be guided by) crowd workers.

Acting
Crowd Coordination Kittur et al. (2013) Decomposing tasks into subtasks Acting Task Design Kittur et al. (2013) Assignment of tasks in relation to individuals' abilities Decisionmaking Task Design Kittur et al. (2013) Guide workers to complete synchronous tasks Acting Contribution Support Kittur et al. (2013) Automatic assignment of group members to maximize collective intelligence

Decisionmaking
Task Design Kittur et al. (2013) Help make the crowd more efficient, skilled, and accurate Acting Crowd Coordination Facilitation Activity Kittur et al. (2013) Design machine learning algorithms that more deeply understand the human nature of these labels Reasoning Contribution Evaluation Kittur et al. (2013) Determine which work products may still be improved Recognizing Quality Control Kittur et al. (2013) Assign workers most likely to make such improvements Decisionmaking Task Design Kittur et al. (2013) Predict their expertise needs in advance, then train and adapt workers in an online fashion via automated tutoring or peer learning Predicting Task Design Kittur et al. (2013) Designing and integrating workflow, incentive, and instruction patterns  (2017) A DSS should remove irrelevant contributions from the data set.

Acting
Contribution Aggregation Rhyn and Blohm (2017) A DSS should aggregate redundant information for the decision-maker.

Reasoning
Contribution Aggregation Rhyn and Blohm (2017) A DSS should prioritize important information for the decision-maker.

Facilitation of the idea submission process
Acting Contribution Support

Tavanapour and
Bittner (2018) Reach each contributor on idea platforms in an initial idea submission process

Acting Task Communication
Tavanapour and Bittner (2018) Actively asks questions in one-to-one collaboration and encourages the contributor to think about missing details and add them before the idea is released to the filtering or voting process Acting Quality Control Tavanapour and Bittner (2018) Guide users in such a way that they can more thoroughly elaborate on their ideas

Appendix B -Refinement of the AI Manifestation Mapping
During the analysis of TCL, PSM, and their corresponding platforms, we derived 29 potential use-cases of AI-augmented facilitation, of which 10 have been actualized and 19 have been perceived as helpful to facilitate crowdsourcing initiatives. These 29 potential use-cases resulted in 18 distinct manifestations using the same matching approach as in step II). Three manifestations were not found within our literature search. Thus, we extended our AI manifestation mapping (i.e., recognizing, in crowd coordination; recognizing, in risk management; reasoning, in task communication). These three manifestations could all be assigned to the existing seven affordances. Hence, it was not necessary to change the affordances. Besides the potential use-cases and manifestations, we observed that the timing of the decision to use AI impacted the potential for integration. In TCL, the decision to include AI was made following our recognition that the workers had created many contributions during an exercise. Hence, considerations about the use of AI revolved around aggregating the contributions, which can be seen as the affordance 7) decision-making preparation. In PSM, a dedicated AI team had been set up during the workflow design, which specifically developed an extensive set of AI functionalities for facilitators, including use-cases of 6) worker profiling and 5) collaboration guidance. We conclude that, if facilitators are aware of AIs' potential before the start of the exercise, they can derive more benefits from it. In step V) Interviews, we conducted six expert interviews to validate and refine the AI manifestation mapping and the initial AI affordances. We integrated the input we received into the affordances after each interview and later analyzed the content in a structured manner in line with Schreier (2012). We gathered 72 statements about potential use-cases for AI in facilitation, which led to 30 distinct manifestations. Three manifestations were not identified by literature or the case studies (i.e., recognizing in culture development, decision-making in culture development, and recognizing in tool usage & integration). We were able to assign these three new manifestations to our affordances. When machines support the collaboration among humans: AI-enabled facilitation in crowdsourcing This is our research goal: We strongly believe that artificial intelligence can make our life more comfortable and help us in stressful situations. We all know collaboration can sometimes be very hard, even though we depend on it to fulfill tasks that we cannot do on our own (e.g., design and build a car). Hence, researchers are currently investigating opportunities of artificial intelligence in collaboration, for example, in crowdsourcing. Together with your help, we would like to identify which AI opportunities exist, helping us to improve or reinvent facilitation for crowdsourcing initiatives.

Definition of "crowdsourcing":
"Crowdsourcing is a type of participative online activity in which someone (e.g., an organization) proposes to a diverse group with varying knowledge a task that is difficult to tackle alone" Definition of "facilitation": "Facilitation in crowdsourcing comprises all observing and intervening activities to foster beneficial interactions among crowd workers aimed to make outcomes easier to achieve and ultimately to align joint actions with predefined goals."

Appendix C.2 -Validation of the Artificial Intelligence Affordances
We validated our affordances according to four defined criteria: completeness, comprehensiveness, meaningfulness, level of detail, and applicability in today's crowdsourcing initiatives (Sonnenberg and vom Brocke 2012). We could confirm four affordances without changes to their description or name (i.e., contribution assessment, operational assistance, collaboration guidance, and decision-making preparation). We could also confirm two further affordances with changes to their description (i.e., improvement triggering and worker profiling). For 2) improvement triggering experts highlighted that low activity does not necessarily mean that a worker should be considered "bad" (Expert 2, 6) and that it is essential to "understand why underperforming is happening" (Expert 5). Since the other experts see AI as capable of identifying and nudging inactive workers or triggering improvements (Expert 1, 3, 4), we enhanced the description of the affordance by including the identification of non-active workers. Experts (Expert 2, 4) specifically noted that including information about workers' backgrounds could be beneficial for monitoring the performance and moderating the crowd. Even though we agree with Expert 1 that facilitators have to consider ethical standards when profiling workers, we see benefits in considering the workers' skills (e.g., a better understanding of the workers' communication network or identifying echo chambers) and so changed the description of 6) worker profiling. Finally, we could confirm 4) workflow enrichment after substantially and successively revising the name (previously: environment creation) and the affordance description. Experts felt that the previous affordance description was not comprehensive enough (Expert 1, 2), too broad (Expert 4, 6), or not applicable to crowdsourcing initiatives (Expert 2, 3). We also picked up the feedback that the intended affordance is still mainly humandriven (Expert 2) and that AI could enrich the exercise with relevant information regarding the workflow or the discussed content (Expert 1, 3, 4). Hence, we shifted the focus of this affordance from creating and maintaining a highly productive collaboration and communication environment to providing and integrating useful information and knowledge into the predefined workflow, which was perceived as meaningful (Experts 5, 6) In addition to these suggestions for improvement, experts stated that there was no relevant aspect missing (Expert 2, 3, 4, 5, 6), from which we concluded that our affordances were complete. Five of the experts explicitly confirmed the comprehensiveness of the name and the description of the affordances (Expert 2, 3, 4, 5, 6). Expert 5 specifically stated, "I think it is the first time I had a research interview where there had been this comprehensible material for me to look at." According to all of the interviewed experts, our affordances are meaningful. Even though the affordances are not mutually exclusive and show some differences in the granularity of formulation, the experts did not see any problems regarding the level of detail (Expert 2). The experts also explicitly acknowledged the applicability of specific affordances in today's crowdsourcing initiatives (Expert 1, 2, 3, 4, 6)

1) Contribution Assessment
Expert 3*: "And it can be very useful to find similar discussion threads and similar ideas." Expert 5: "Understand all the results from a crowdsourcing exercise in a way that's empirical and meaningful." Expert 6: "There is also all of this other work [besides creating the contributions] that['s analyzation] would help with the facilitation." 2) Improvement Triggering Expert 2*: "So maybe also a little bit set a spark, a stimulus, or a nudge." Expert 4*: "What a moderator often does is triggering.
[…] So I try to inspire them [the worker] or give them some guidance. So that they can do it, and that's where I think you can use AI again." Expert 5: "Obviously, you could have automated message's that encourage people to participate if they haven't done something on the site."

3) Operational Assistance
Expert 1*: "The second thing that I would find exciting are intelligent translation programs.
[…] To use them in such a way that language barriers are broken down to some extent." Expert 3*: "For example, you could show a novelty score to the user as they type in their idea." Expert 5: "So one way I could see that happening: if people happen submitting ideas using the same syntax or verbiage, then a tool could be like 'Oh hey, maybe you would want to work with this other individual submitting similar ideas.'"

4) Workflow Enrichment
Expert 3*: "What has a lot of value for me personally with AI and data, which is probably not the case with facilitators, is 'how do we now, as the main facilitator, transfer these learnings into the next task?'." Expert 5: "Tool usage and integration, I think, could also be something that a tool could do. So, for example, it could say, 'You haven't done this yet. Maybe you want to check this feature out.' or something like that." Expert 5: "Communicating the task is something that I see being pre-generated, and humans go in and do checks and make sure that it is okay."

5) Collaboration Guidance
Expert 3* "In that respect, for crowd moderation or crowd coordination, Word-to-Vec would be great because you have a neutral algorithm that just shows what's happening without understanding whether that's good, bad, or whatever." Expert 4* "When it comes to, 'Tell me what's relevant here in this data set.' that I could just say, 'Look, could this be relevant here?' and I sort of let it [the AI] do the assessment after that." Expert 6: "I think something that could be helpful here, regarding contribution development, is if people find proposals that are similar or related to the ones that they are working on." 6) Worker Profiling Expert 1*: "And of course, what is based on numbers or is evaluable. In the sense that you say quality control should depend on the number of contributions or spelling. These are things that systems like NLP or machine learning can test well. Here you can set thresholds creating a warning: 'Attention [contribution] quality is decreasing'." Expert 2*: "One activity was redundancy detection. So, we wanted to avoid contributions are too similar." Expert 4*: Well, you could do profiling of people. You could do a priori profiling. You take their CVs, LinkedIn profiles, and maybe their publications or articles and profile them based on that.

7) Decisionmaking Preparation
Expert 1*: "And of course, everything goes in the direction of aggregating and interpreting results. Some technologies create natural text, that is, to create reports from the contributions. So clustering is somehow an exciting topic." Expert 2*: "I see the use of AI more in the sense of massive automation, simply getting information much faster and then ultimately providing the facilitator with a decision template, preferably also with recommendations for action."