The ongoing development of information technology (IT) enables organizations to introduce digital work as the new normal. Therefore, employees are facing new forms of work that might decrease personal interaction but increase interaction with IT. Nevertheless, these new ways of work entail that individuals cannot do their jobs with the same values and convictions as they are used to. There is a constant change that might impact selfbeliefs constituting professional identity at work, i.e., the perception of one’s role in the workplace. Experiencing a situation that contradicts one’s identity might lead to a loss of self-esteem and thus to a threat to identity (Petriglieri, 2011). This might enforce actions designed to protect self-esteem correlated with identity (Craig et al., 2019) as emerging technologies have changed the landscape and experiences of a variety of professions (Frick et al., 2021). The digitization of the workplace emphasizes the demand for digital work as the new normal in organizations (Baptista et al., 2020; Frick & Marx, 2021). One major factor driving this discussion is the ongoing development of artificial intelligence (AI), which that can be described as “the ability of a machine to perform cognitive functions that we associate with human minds, such as perceiving, reasoning, learning, interacting with the environment, problem solving, decision-making, and even demonstrating creativity” (Rai et al., 2019, p. iii). In digital work, AI is applied for various functions, such as (managerial) decision making (Haesevoets et al., 2021), data analysis and prediction work (Grønsund & Aanestad, 2020; Marabelli et al., 2021), or (human-AI) collaboration (Brachten et al., 2020; Mirbabaie et al., 2020, 2021). Therefore, AI will change workplaces and professions persistently, potentially threatening the livelihoods of individuals whose jobs are taken over (Aleksander, 2017; Haenlein & Kaplan, 2019). However, AI might lead to value co-destruction when discrepancies between users emerge (Camilleri & Neuhofer, 2017). Furthermore, the use of AI could also support the development of uncertainty and invasion of privacy (Cheng et al., 2021). This negative phenomenon is often referred to as the dark side of AI, referring to how AI presents risks for individuals, organizations, and society (Alt, 2018; Grundner & Neuhofer, 2021; Wirtz et al., 2020). Nonetheless, the utilization of AI in organizations could not only eliminate or change current jobs but also create new areas of work, for example, within engineering, programming, or even in social domains (Acemoglu & Restrepo, 2018). There is ongoing hype about AI and its economic impacts (Selz, 2020). Although the public discussion about AI has turned more optimistic in recent years, the fear of AI eliminating current jobs still outweighs the possible opportunities for human-AI collaboration (Acemoglu & Restrepo, 2018; Aleksander, 2017; Dwivedi et al., 2019; Wang & Siau, 2019; Zhang & Dafoe, 2019).

Human-AI interaction reveals that individuals’ perceptions of AI are based on different aspects. For example, salient cues, affordances, or collaborative interaction (Sundar, 2020; Wirtz et al., 2020) might affects individuals’emotions and, therefore, intentions toward AI (Shin, 2021). Employees establish an identity in relation to applied technology and their jobs. To adequately consider this phenomenon, we take on the perspective of Carter and Grover. (2015), who introduce the term IT identity and define it as “the extent to which a person views use of an IT as integral to his or her sense of self” (Carter and Grover, 2015, p. 938). Introducing AI in the workplace may contradict employees’ identification with their jobs and lead to resistance behavior such as algorithm aversion. Algorithm aversion describes a phenomenon in which, under the same conditions, employees prefer a human over algorithmic decision support (Dietvorst et al., 2015; Venkatesh, 2021). Craig et al., (2019, p. 269) describe this resistance as IT identity threat, defined as “the anticipation of harm to an individual’s self-beliefs, caused by the use of an IT, and the entity it applies to is the individual user of an IT.” Therefore, it is vital to understand the emergence of upcoming predictors influencing AI resistance based on IT identity threats as the application of AI is likely to alter jobs in organizations and thus might impact individuals’ identities.

To contribute to reducing resistance to AI, it is crucial to broaden the understanding of identity threats caused by AI as unique technology. To minimize AI related threats towards the identity of employees, it is essential to position AI as a benign and supportive collaboration technology. We argue that this is of great interest to information systems (IS) researchers and practitioners alike as the application of AI for generating business value will further increase, successively turning into key elements for enterprises (Dwivedi et al., 2019). Considering the social interactions between individuals and technology and its possible threat to identity, further investigations into IT identity threat are needed to understand this phenomenon (Craig et al., 2019). There is thus an urgent demand to conduct in-depth research on new forms of human-AI collaboration, especially on related consequences for employees in terms of attitudes and actions toward AI as well as psychological, emotional, and social aspects (Coombs et al., 2020). The continuing distribution of AI in electronic markets (Adam et al., 2020; Thiebes et al., 2020) therefore entails the pressing need for IS to investigate potential influences on the IT identity threat caused by AI. Based on these arguments, this study rests upon the following research question:

  • RQ: Which predictors influence the AI identity threat of employees in the workplace?

To answer this research question, we initially carried out a systematic literature review (SLR) following the latest methodological guidelines for literature reviews in IS (e.g., Okoli, 2015; Paré et al., 2015). We argue that the complexity as well as opportunities of AI and associated identities of employees in the workplace, as well as related threats, have not yet been adequately covered in the research. This descriptive procedure examines existing literature describing the current situation to synthesize research evidence based on scientific facts (Bear & Knobe, 2017; Bell, 1989; Grant & Booth, 2009). Building upon these findings, we derive a research model, including predictors for the AI identity threat of employees in the workplace, which was quantitatively evaluated using partial least square structural equation modeling (PLS-SEM) as a favorable approach for the assessment of multistage models (Fornell & Bookstein, 1982). As there are different interpretations of the appropriateness of PLS as a suitable SEM technique (e.g., Goodhue & Thompson, 2012; Hair et al., 2017; Marcoulides et al., 2009; McIntosh et al., 2014; Petter & Stafford, 2017), we followed the latest recommendations of Hair et al. (2019) to alleviate ascending concerns. We finally conducted post-hoc expert-interviews to evaluate the model and the identified predictors for the AI identity threat of employees in the workplace. Individuals thereby elaborated on the current state of the art and their experiences (e.g., descriptive procedure) (Bear & Knobe, 2017).

This paper contributes to theory and practice by extending our understanding of collaboration with AI in the workplace to drive future research in this field. Researchers will find the insights helpful in understanding the implications of employees’ identity when collaborating with AI in enterprises. By providing a theory-based framework, scholars will assess which predictors cause AI identity threat in the workplace. Practitioners will be able to comprehend which factors are particularly relevant when introducing AI for collaborative purposes. Readers will realize that employees perceive identity fears when AI is increasingly applied. Human-AI collaboration is more likely to be successful when possible threats are understood and overcome. We believe this study will be valuable to researchers, practitioners, and society for understanding and overcoming obstacles when collaborating with AI. Hence, this article extends the IS literature by updating our knowledge on identity and related threats in the context of AI in a collaborative working environment and how predictors are interconnected.

Theoretical background

The need for AI in the workplace

AI is increasingly being augmented in electronic markets, triggering a widespread acceptance in organizations across industrial boundaries and transforming the global economy (Adam et al., 2020; Guan et al., 2020; Thiebes et al., 2020). Therefore, AI is considered an integral part of business strategy and organizational decision making (Cheng et al., 2020a, b; Shrestha et al., 2019), thus making it a key element for generating business value (Dwivedi et al., 2019). As a ubiquitous concept (Siau & Wang, 2018), AI consists of multiple subfields and dimensions (i.e., think like a human, think rationally, act like a human, and act rationally) (Russel & Norvig, 2016), but without evolving into a consistent definition (Duan et al., 2019). AI is usually associated with human intelligence, but society is facing the question of how machines are able to reach this kind of intelligent behavior (Neuhofer et al., 2020), yielding a three-dimensional categorization: narrow AI, general AI, and superintelligence (Batin & Turchin, 2017). Narrow AI covers self-learning approaches that outperform humans on specific, narrow tasks. General AI explains self-learning comparable to the intelligence of humans. Superintelligence, hypothetically, is thought to exceed humans in all aspects. Most applied systems in organizations are considered narrow AI as they focus on particular work-related tasks (Batin & Turchin, 2017). AI is believed to fundamentally change the working environment and the way people work (Bednar & Welch, 2020), spreading across industries (Wang & Siau, 2019) with potential applications in almost every field (Barredo Arrieta et al., 2020). However, among the changes in working environments and models, AI possesses unique issues. Since AI is still implemented by humans, systems might contain certain biases as individual backgrounds and experiences (i.e., development of heuristics) are reflected. Some of these could lead to bigger or more frequent mistakes made by the AI (Wirtz et al., 2020). Moreover, not every AI model is explainable due to its sophisticated structure. It thus represents a black box for humans that prevents understanding certain decisions and leads to increased uncertainty (Venkatesh, 2021). In the context of generating predictions involving humans, this negatively affects the privacy perceptions of individuals (Cheng et al., 2021). While there are arguments that highlight the bright side of AI, such as improvements to efficiency or cost reduction, aspects of the dark side of AI may have a stronger effect on the perceptions of employees toward AI (Grundner & Neuhofer, 2021). Thus, a simple economic decision such as cost reduction based on decreasing the headcount in an organization might lead to the fundamental anxiety of being replaced by AI (Złotowski et al., 2017). The current generation of narrow AI in enterprise is still new and multi-faceted. There is widespread speculation about collaborative capabilities and relevance, and many employees only have a vague picture of collaboration with AI in their workplaces (Kühl et al., 2019). Research has demonstrated that human-AI interaction is able to shape processes more effectively and enhances the individual performance of employees. Employees are supported in the decision-making process (Dellermann et al., 2019; Metaxiotis, 2000), strategic decisions are facilitated (Aversa et al., 2018), and human-AI collaboration frees humans from repetitive tasks, helping them embrace strong economic potential (Yang & Siau, 2018). A recent study employed an AI-based system within the incident management process of an IT department collaborating with help desk employees on the hitherto manual categorization process (Frick et al., 2019). The study illustrated that over 90% of incidents were properly classified. This is not only superior to humans but also frees individuals from repetitive tasks. Another study proved that collaboration with AI in the form of virtual assistants (VAs), computer-based support systems, is capable of supporting employees during the solution of work-related tasks (Brachten et al., 2020). The findings indicate that, first, the collaborative execution of tasks is more efficient and, second, VAs decrease the load, illustrating that collaboration also reduces the perceived workload of employees in the workplace. Furthermore, Mirbabaie et al. (2020) examined social identity, i.e., the identification of employees with team members, and extended self, i.e., the incorporation of possessions into one’s sense of self, when humans work with VAs in teams. The authors derived a new concept, virtually extended identification, elaborating that individuals who identify themselves with (virtual) team members are also more likely to identify with the technology as a part of their extended selves and vice versa. This intertwining highlights that research needs to adjust its understanding of identification with AI and that different collaborative settings, such as identity and related threats, need to be examined.

On the one hand, AI is considered an engine for productivity and progress; on the other hand, AI entails societal upheaval (Yang & Siau, 2018). It is not only of particular interest to assess how to drive business success and value by applying AI; there are several theoretical and practical strategies (Adam et al., 2020; Frick et al., 2019, 2020; Selz, 2020). However, it is especially vital to examine whether employees establish a certain identity with AI (Mirbabaie et al., 2020) and which predictors influence AI identity threat. A successful application for collaborative purposes to generate business value is thus only achievable when employees somewhat identify themselves with AI and perceive AIs as equal collaborative partners and elemental parts of their work.

IT identity and related threats

This paper follows the perspective of concept identity as the “answer to the question ‘Who am I?’ in relation to a social category or object” (Carter and Grover, 2015, p. 933). A person always only holds one self-concept, while it is possible to hold multiple identities (Stets & Burke, 2000). Whereas the collective level of identity explains how identity is formed from a membership in social groups (Stets & Burke, 2000), the individual level describes how the connections and relationships of people influence their identity and, subsequently, behavior toward others (Burke & Stryker, 2016). In this context, role identity (e.g., work, sports, parents) is verified when people act according to the internalized expectations toward a specific role. Person identities are verified when people perform in ways that are aligned to values and norms that differentiate themselves as unique individuals. Last, material identities are verified when people are able to control or master a material object with which they are interacting (Carter and Grover, 2015). However, people do not only verify their identity based on their experience among role, person and material identity. Individuals also face contradictions to their self-believes that may result in experiencing potential harm to one’s identity (Petriglieri, 2011). Thus, “when an experience contradicts identity, individuals experience a loss of self-esteem and take action to preserve the self-esteem associated with identity” (Craig et al., 2019, p. 265). Those actions may yield into resistance behavior or a change of the current identity concept. Furthermore, the scholars describing the nature of the identity threat as harm to the meanings, value or enactment of an identity (Craig et al., 2019; Petriglieri, 2011). This potential harm to an individual’s identity may affect role, person or material identity. While the concepts of role and person identity focus on group processes and group identities, the material identity is more connected with individual behavior and thinking (Boudreau et al., 2014; Gong et al., 2020). Considering the perspective of identity in relation to the question “Who am I?”, we follow the conceptualization of Carter and Grover (2015), who understand IT identity as a new form of material identity. IT identity can be defined as “the extent to which a person views use of an IT as integral to his or her sense of self, where a strong IT identity represents identification—‘use of the [target IT] is integral to my sense of self (who I am)’—and a weak IT identity represents dis-identification—‘use of the [target IT] is completely unrelated to my sense of self (who I am)’ (Carter and Grover, 2015, p. 938). The authors state that in the context of IT identity, IT can be defined as any unit of technology that allows a user to consciously interact with it to produce, store, and communicate information and that is accessible at any time or place.

However, introducing and using a new technology might not yield acceptance. The identification with IT might be beneficial but entails certain drawbacks, separated into four categories of identity: IT identity, absent IT identity, anti-IT identity, and ambivalent IT identity (Carter et al., 2019). IS research already covers single factors of resistance to IT, e.g., the fear of losing human uniqueness (Stein et al., 2019), the deskilling of professionals (Boudreau et al., 2014), information interruption of emotional exhaustion (Cheng et al., 2020a, b), the protection of a certain threatening event (Sun et al., 2020), and the possibility of unemployment or the chance of losing safety (Złotowski et al., 2017). Considering the social interactions between individuals and technology and its possible threat to identity has mainly been neglected in IS research (Craig et al., 2019).

Therefore, scholars have introduced the concept of IT identity threat as “the anticipation of harm to an individual’s self-beliefs, caused by the use of an IT” (Craig et al., 2019, p. 269). The need for examining the threat of IT to individuals’ identities becomes apparent when observing the changing landscape of digital collaboration. Recent research has started to investigate IT identity by considering technology such as Excel or smartphones (Carter et al., 2020b). This step has been fundamental to strengthening the understanding of IT identity. However, regarding new forms of collaboration in the workplace, IT, or specific IT features such as Excel calculations, are often replaced by AI systems that support, for example, decision making or data analysis (Mirbabaie et al., 2020). Thus, analyzing IT identity by using traditional IT might not be sufficient to understand identification with technology such as AI in the workplace. Thus, research elaborating the role of AI as a technology that breaks the boundaries of historic IT is urgently needed. For example, Alahmad and Robert (2020) revealed that the identification with AI impacts job performance. Focusing on human-AI collaboration in organizations, we ask the question “Who am I as a professional when collaborating with AI in the workplace?” and define AI identity, leaning on existing explanations (Carter and Grover, 2015), such as “the extent to which individuals perceive the collaboration with AI in the workplace as an indispensable component of themselves.” While the identity threat has been researched in the context of different technologies (Craig et al., 2019; Stein et al., 2019), the unique and novel characteristics of AI in the workplace make it indispensable to investigate AI’s role in identity threat.

Research approach

Since the transfer of the concept of IT identity and related threats into the realm of AI has not been adequately covered by extant research, we performed an SLR to identify relevant literature and reveal interpretable patterns and theories (Paré et al., 2015). Existing conceptualizations and propositions (Paré et al., 2015) serve as a theoretical foundation to derive a research model predicting AI identity in the workplace. We further quantitively evaluated our research model using PLS-SEM, enabling us to explain causal relationships of identified predictors of AI identity threat (Hair et al., 2019). Finally, to examine whether the identified predictors are really relevant to AI identity threat, we conducted semi-structured interviews with experts who are familiar with AI in their workplaces. To better provide an overview of the applied methods, Fig. 1 shows the distinct steps of the procedure.

Fig. 1
figure 1

Procedures of the applied methods

Systematic literature review

An SLR is especially helpful for identifying existing knowledge about a topic, including related gaps, by searching for relevant articles using keywords in scientific databases (Fink, 2013; Webster & Watson, 2002). Since we use the SLR to emphasize our contribution to knowledge based on an interpretable pattern from existing literature (Fink, 2013; Webster & Watson, 2002), this is considered to be a descriptive process describing the current state of the research domain by relying on scientific facts (Bear & Knobe, 2017) and further transferring it to an emerging topic. We analyzed the literature according to evidence of predetermined qualitative themes (e.g., IT identity and related predictors), leading to deductive results (Bandara et al., 2015). Ali and Birley (1999) explain that within this procedure, “the scientist formulates a particular theoretical framework and then sets about testing it.” We used a priori coding themes from our studied phenomenon (Ali & Birley, 1999). This theoretical sensitivity is particularly helpful as codes can be derived from, for instance, scientific definitions or established constructs (Ryan & Bernard, 2003; Strauss & Corbin, 1990). Following the literature review process (vom Brocke et al., 2009, 2015), we defined our research scope based on the taxonomy of literature reviews (Cooper, 1988). We were interested in research focusing on the phenomenon of identity, IT identity, identity threats, and IT resistance in relation to the workplace.

The literature search was conducted using, an IT artifact with the aim of assisting IS researchers in retrieving relevant literature from major scientific sources (Boell & Wang, 2019). A search string was created with the ISSN numbers of the selected outlets, which was used in Scopus’s advanced search, making it possible to search across all indexed scientific sources (Boell & Wang, 2019). Litbaskets offers the possibility to set different filters and size restrictions to limit the search or exclude certain sources or publications. We deliberately chose the 154 essential IS journals (basket of eight and other high-ranked outlets, such as Information Systems Frontiers and Electronic Markets) as we wanted to focus on substantial and high-quality articles. Since litbaskets focuses on IS journals but omits conferences, we extended the literature search by selecting the elementary IS conferences (AMCIS, ECIS, ICIS, PACIS, HICSS, and Wirtschaftsinformatik) manually via Scopus. Our literature search acknowledged peer-reviewed articles. Less relevant sources, e.g., editorials or commentaries, were excluded. We carried out a full text and metadata search and deliberately did not limit it to metadata only as this might not contain the search term, and further narrowing down the possibility of overlooking relevant publications. The research procedure considered articles published up until May 2020. We used the following query for our fulltext search:

((“identity threat”) OR identity OR (“IT identity”) OR (“AI identity”) OR (“IT resistance”)) AND (AI OR (“artificial intelligence”) OR (robot*) OR (“disruptive technology”) OR (“digital transformation”) OR (“information system”) OR (“information technology”) OR (“digitali*ation”)) AND (task OR work* OR job OR profession* OR career OR firm OR organi*ation* OR company OR business OR office OR enterprise OR corporation* OR association)

To obtain thorough results and provide an overview on the current state of research, we carried out several exemplary literature searches prior to the execution of our SLR. This served us in finding the most promising results to identify predictors related to AI identity threat. Our final search string covered three major sections. First, we used terms that covered literature related to identity and possible threats. Combinations of IT and identity are frequently used in extant research. Second, phrases linked to AI and associated concepts and technologies were included. For example, we added digital transformation and digitali*ation to our search query as AI is considered as key factor in this advancement (Frick et al., 2021). Third, different appellations for organizations and workrelated contexts were added. Depending on each research focus and authors’ preferences, different terms such as organization, company, or enterprise were used. Parentheses nested clauses, Boolean expressions linked individual nomenclatures, quotation marks combined terms that must appear next to each other, and asterisks marked different spellings (such as in American and British English).

After the initial search was finished, we carefully read the title, abstract, and keywords of each publication to determine its relevance to our research questions. Our selection was guided by the following questions: How can IT identity be described in the workplace? Which negative effects for employees result from the application of IT? Which factors influence identity threat in organizations? Since we wanted to adapt theoretical insights from IT contexts to the area of AI, we focused on articles dealing with the application of IT in the workplace and associated threats. We excluded articles, for instance, if they only dealt with the technical implications of technology. To achieve a comprehensive review, we next performed a backward search. Further literature was identified by collecting every reference of the bibliographies of all the papers from the initial search. We included references to other journal or conference publications and excluded non-scientific sources, such as web pages or business reports. To determine the relevance according to our research questions, we conducted a similar approach as within the initial search: we read the title, abstract, and keywords, which was followed by categorization according to theoretical foundations. The last step of the literature review was a forward search to identify additional relevant literature. We acknowledged every paper that was identified during the previous initial and backward searches and that had been cited by other research after its initial publication.

The execution resulted in 49 relevant articles. We screened 5,649 articles, of which 21 (of 326) were found to be relevant during the initial search, 10 (of 2,057) via the backward search, and, finally, 18 (of 3,266) relevant publications were retrieved during the forward search. Table 1 outlines the number of search results per search type. No articles yielded by our search terms were published before 1998, and the number of publications has risen since. For example, in 2018, there were 5 (10.2%) publications, followed by 9 (18.4%) in 2019. The constant growth of research and publications reflects not only the pressing need for further studies but also represents the salience and legitimacy of this research area. Furthermore, from the total of 49 articles, 36 (73.5%) are in journals, and 13 (26.5%) are conference publications. The majority of articles were published at the International Conference on Information Systems (5, 10.2%) followed by Computers in Human Behavior, European Conference on Information Systems, European Journal of Information Systems, and Information and Organization, with three articles each.

Table 1 Number of search results per type

To understand how AI can be explained in the context of identity in the workplace (e.g., AI identity), we clustered the articles along the different identity types. Overall, we found 36 articles that dealt with identity, of which 15 (41.7%) dealt with identity in general, 16 (44.4%) researched professional identity, and the minority considered IT identity (5, 13.9%). There were no papers that dealt with the concept of AI identity. Current IS research on identity can be distinguished as follows (cf. Table 2): identity, which typically asks the question “Who am I?” (Alvesson et al., 2008; Lamb & Davidson, 2005; Park & Kaye, 2019), explains the social role and/or influence of individuals (Boudreau et al., 2014; D’Mello & Eriksen, 2010; Elbanna & Linderoth, 2015; Sime & Themelis, 2020). Professional identity asks, “Who am I as a professional” (Chreim et al., 2007), refers to the self-perception of an employee in the workplace (Jussupow, 2018; Jussupow et al., 2018), and is defined as “an individual’s selfdefinition as a member of a profession and is associated with the enactment of a professional role” (Chreim et al., 2007). IT identity specifies how humans identify with information technology and is interpreted as “the extent to which an individual views use of an IT as integral to his or her sense of self” (Carter and Grover, 2015).

Table 2 Articles assigned to different identity types

Model development

Besides obtaining an overview of current identity types, we analyzed the papers for possible identity threats and their type of influence. In total, we found 22 papers illustrating 24 predictors for identity threat. Since this research focuses on the AI identity threats experienced by employees, we were specifically interested in predictors that might cause individual identity shifts in the workplace. Professional self-image consists of having status and value (Alvarez, 2008) and is, besides technology, altered by certain factors (Jussupow et al., 2018). The identified predictors were thus analyzed regarding their potential strength of influence on the identity threat. If an individual faces threat to its identity, it may lead to identity protection responses that could lead to several effects (Jussupow et al., 2018). For instance, resistance behavior, refusing to engage with technology (Nach, 2015), and forming a certain anti-identity (Carroll & Levy, 2008) may arise as protection response behavior to reduce incoming identity threats by AI. Individuals might even emotionally regulate their threatened identities through, i.e., increasing work volume, meditating, or crying (da Cunha & Orlikowski, 2008). Thus, understanding identity threats caused by AI is important to resolve harmful effects that are rooted in the introduction of AI at the workplace. To this end, we found six predictors that might cause AI identity threats to employees in the workplace.

We further used the threat sources of the identity threat framework to cluster the final predictors (Craig et al., 2019). First, intergroup conflict explains the reputation and self-beliefs within a social group. Conflicts within, for example, a team in the workplace may result in the loss of resources, status, or prestige. Second, verification prevented describes the personal perception and inspection of self-beliefs initiated or influenced by technology. Third, meaning change is triggered by modifications of work-related tasks. Thus, changes in tasks alter the meaning of the job itself, which might not be consistent with the professional identity of an employee. Table 3 outlines the classification of the six predictors, along their threat sources, and presents explanations and exemplary influences.

Table 3 Possible predictors and hypotheses of AI identity threat

Loss of status position and threats to job security can be categorized as intergroup conflicts since both factors have a considerable impact on perception within social groups in the workplace. Introducing AI at the workplace may enhance the sense of loss of status and job security by overtaking certain tasks or responsibilities and therefore threaten one’s established identity. Złotowski et al. (2017) found that technology that is capable of solving employees’ tasks leads to an increased perception of threat to job security. Likewise, Craig et al. (2019) argue that IT that substitutes distinct parts of a job might lead to a reduction of power or prestige associated with that status position. In this context, Jussupow et al. (2018) argue that AI might act as an individual, direct threat to the social position of an employee. These threats might lead to reduction of power and prestige in the workplace. Therefore, we formulate the following hypotheses:

  • H1a: Threat to job security has a positive effect on AI identity threat in the workplace.

  • H1b: Loss of status position has a positive effect on the AI identity threat in the workplace.

Loss of skills/expertise and changes to work can be described with verification prevented as, due to the changes in work-related tasks, self-beliefs are validated. Jussupow et al. (2018) argue that the threat to skills/expertise is rooted in the usage of technology. This might lead to a decreased need for specialized skills and knowledge by employees (Hicks, 2014). Thus, one’s identity that is built on the personal skills and expertise may be vulnerable to new technology such as AI that need to quantitate certain tasks. Therefore, AI might be capable of changing the perception of one’s own skillset by exposing personal (performance) weaknesses. Thus, introducing a new technology such as AI is often linked to changes in the workplace (Craig et al., 2019). These new changes to work may evoke feelings such as restlessness or disengagement that lead to a lack of involvement and therefore to threat to one’s identity. Therefore, we derive H2a and H2b:

  • H2a: Loss of skills/expertise has a positive effect on AI identity threat in the workplace.

  • H2b: Changes to work have a positive effect on AI identity threat in the workplace.

Loss of autonomy and loss of controllability are covered by meaning changes as technology-induced changes might significantly alter professional responsibilities. This may lead to an increased threat to the identity as professional responsibilities might be strongly linked to one’s identity at the workplace. In this context, Schweitzer et al. (2019) identify distinct relationships with AI-based devices. The scholars reveal the fear of losing control over one’s digital self when interacting with a smart device. Furthermore, research indicates that technology may impact the control and autonomy of daily and work life as well. For example, Park and Kaye (2019) describe the renunciation of technology as a strategy to regain control of daily life. The scholars argue also that technology could be part of one’s self and therefore influences the individual identity either in a positive or negative way. Likewise, Prester et al. (2019) emphasize the close relationship between digital technologies, workplaces, and human identity in the context of professional autonomy and maintaining the self. Furthermore, Nach (2015) describes an upcoming threat by IT when an individual appraises a situation as uncontrollable. If an individual perceives the situation as controllable, they deploy strategies to neutralize the potential threat by an IT. However, the application of AI encompasses the capabilities of reducing the employee’s perceived autonomy and controllability in the workplace. This emphasizes the potential relationship between autonomy and controllability with an AI identity threat. Thus, the following hypotheses are derived:

  • H3a: Loss of autonomy has a positive effect on AI identity threat in the workplace.

  • H3b: Loss of controllability has a positive effect on AI identity threat in the workplace.

Besides the predictors identified by the conducted SLR, it can be assumed that the concept of AI identity, as previously defined, has a major impact on the AI identity threat itself. According to Carter and Grover (2015), individuals identify positively or negatively with a distinct IT. This assumption can be transferred to AI as the object of research. When employees perceive that AI is an integral part of themselves and are committed to collaborating with AI to accomplish specific work-related tasks, this may result in the technology not being regarded as a hazard (Grundner & Neuhofer, 2021; Złotowski et al., 2017). However, a negative identification with AI may support the perceived threat caused by AI in the workplace (Alahmad & Robert, 2020; Craig et al., 2019). This prevailing attitude of employees is crucial to consider when broadening the understanding of AI identity threat. Thus, we derive the following hypothesis:

  • H4: AI identity has a negative effect on AI identity threat in the workplace.

To put the hypothesis of the specific predictors into context, we conceptualize a research model (Fig. 2 and 3) that allows further analysis of the AI identity threat and the identified predictors.

Fig. 2
figure 2

Research model including constructs and hypotheses as developed from the literature

Fig. 3
figure 3

Research model with results (N = 303). *p ≤ 0.05; **p ≤ 0.01; ***p ≤ 0.001

Expert interviews

The SLR is suited for identifying predictors of AI identity threat in the workplace. We therefore adapted findings from existing research in which AI might not have been adequately covered. We conducted post-hoc interviews to validate whether the identified predictors are really relevant in our research context.

Expert interviews are sufficient for retrieving knowledge from qualified experts (Kvale & Brinkmann, 2009) and are especially fruitful for less explored research areas (Bogner et al., 2009). Experts possess specific domain expertise (Meuser & Nagel, 2009); in our case, these are individuals who have participated in or initiated an AIinduced change process in the workplace. To obtain a holistic picture, we chose participants working at the management level to broaden our understanding of the research field (Bogner et al., 2009). We followed the suggestions of Creswell and Creswell (2018) and included between three and ten individuals for our qualitative procedure. In total, we conducted six interviews with the male participants. Experts were between 24 and 38 years old (SD = 4.92; M = 33.66). The experts’ characteristics can be retrieved from the Appendix (Table 8). We followed an open interview technique, using a semi-structured approach. Experts were given the opportunity to elaborate on their subjective beliefs but were supported by a prefixed guideline with essential questions (Meuser & Nagel, 2009; Qu & Dumay, 2011). The interviews started with welcoming the participants and describing the interview process, including a short briefing of the study. In the first phase, we obtained the interviewees’ characteristics, which aided us in understanding their working environments. The next phase included questions on AI in the workplace; we provided a definition on AI to establish the same level of knowledge among all participants throughout the interviews. The third phase served to validate our derived research model and the identified predictors for AI identity threat in the workplace. In the last phase, experts had the opportunity to provide additional input or ask further questions. We concluded with a short debriefing. Since we were interested in the experts’ opinions on the model and predictors rather than their gestures or facial expressions, we used paraphrasing as suggested by Schilling (2006). We limited the data by deleting irrelevant words or formulations to form short and concise sentences that led to generalized explanations or interpretations. We further applied open coding using in-vivo or with simple phrases (second order themes) and summarized them under the identified predictors (first order themes) (Gioia et al., 2013).

Quantitative data collection

To determine which of the identified factors impact AI identity threat in the workplace, we used a quantitative approach for validating the derived hypotheses with a large population (Creswell & Creswell, 2018). Since we were not able to build upon theoretical foundations for every factor used in our research model, we needed to develop and validate constructs within a pre-study. The evaluated items, combined with measures adapted from existing literature, were used within the subsequent main study to test our developed research model.

Measurements and survey

We developed items for loss of status position and changes to work and evaluated them using a pre-study. Example items of experiences of loss of status position are “the introduction of AI will reduce my competences” and “by the introduction of AI, I am limited in my possibilities.” Changes to work included items such as “the introduction of AI increases my area of responsibility” and “by the introduction of AI, I am responsible for more tasks.” Extant research takes the standpoint that the introduction of new IT raises uncertainties, threats, and fears (Hirschheim & Newman, 1988; Laumer, 2011; Silva et al., 2016). Reducing and overcoming potential inhibitors is therefore especially crucial to avoid resistance behavior and an unfavorable climate for change (Bouckenooghe et al., 2009; Frick et al., 2021; Krovi, 1992). We argue that formulating the items using “the introduction of AI” is beneficial as anxieties (i.e., loss of status position and changes to work) are foremost observed prior to or during introduction processes (Joshi, 1991; Kim & Kankanhalli, 2009; Meissonier & Houzé, 2010).

We used a seven-point Likert scale ranging from “strongly disagree” to “strongly agree” to survey the items. In total, 91 participants started the questionnaire, and N = 68 respondents completed the survey. We screened the data manually for anomalies and suspicious responses (i.e., very short processing times and similar or identical answers) but did not need to exclude any further participants. The respondents were between 18 and 52 years old with a mean age of 32.3 (SD = 10.3), of which 22 (32.4%) were female and 46 (67.6%) were male. The results were tested for reliability using factor analysis. Loss of status position resulted in an overall Cronbach’s alpha value of 0.882 (M = 1.79; SD = 0.72), and changes to work yielded a Cronbach’s alpha of 0.864 (M = 3.16; SD = 0.90), indicating high reliability for the constructs. The evaluated scales were used within the following survey.

The main study was conducted to validate our research model. Besides demographic data, which were used as control variables, we adopted and modified constructs from previously validated instruments and applied the evaluated, self-developed items to ensure the high accuracy of the measures. The items are listed in Table 5 (see Appendix). All items were surveyed on a seven-point Likert scale ranging from “strongly disagree” to “strongly agree.” Threat to job security was adapted from Hellgren et al. (1999), who analyzed job insecurity as an influencing factor on employee attitudes and well-being. Loss of skills/expertise was derived from the Jussupow et al. (2018) conceptualization of professional identity threats in the workplace. Loss of autonomy as well as loss of controllability was based on Breaugh (1985), who developed instruments to capture different aspects of work autonomy in organizations. AI identity was measured using items from Carter (2012), who proposed measures for IT identity, including the items of dependence, emotional energy, and relatedness (i.e., feeling attached, having endurance, and being linked to IT). AI identity threat was adapted from Craig et al. (2019), who determined an operational measure for IT identity threat. AI identity and AI identity threat were originally specified as second order constructs; we adapted them as first order constructs. Carter et al. (2020a) observed that first- and second-order models of IT identity are statistically equivalent and provide nearly identical fit statistics. Furthermore, the evaluation of higher order reflective constructs tends to be challenging and might not be valid in terms of their prediction perspective, e.g. (Lee & Cadogan, 2013; Sarstedt et al., 2016). Furthermore, we were interested in developing first order constructs to serve as a basis for future studies combining all relevant aspects as a set of conceptually distinct variables (Carter et al., 2020a, b; Hair et al., 2019).

Participants were recruited using, a platform specifically designed to acquire participants for online surveys (Palan & Schitter, 2018). Participants were selected according to certain conditions: individuals needed to speak English fluently as the survey was presented in English. In addition, individuals needed to be part- or full-time employed and work in for-profit or non-profit organizations or in the local, state, or federal government. Furthermore, as we were concerned with AI, participants needed experience with AI in their private lives or in the workplace.

Descriptive statistics

Overall, 343 subjects participated in our survey, of which N = 303 finished the study properly. Again, we manually screened the data for anomalies regarding short processing times and suspicious responses with similar or identical answers. However, no answers needed to be removed from the dataset. Table 4 shows the ages of the respondents as being between 20 and 68 years old with a mean age of 39.1 (SD = 10.1). Gender was almost equally distributed. Two hundred and nine subjects lived in the UK (69.0%), followed by 15 from Portugal (5.0%), and 13 each (4.3%) from Italy and Spain. Participants worked in consumer goods (54, 17.8%), research and education (44, 14.5%), healthcare (36, 11.9%), industry (36, 11.9%), and IT (34, 11.2%). Most participants worked as regular employees (162, 53.5%) followed by group/team leaders (75, 24.8%) and heads of department/subdivision (45, 14.9%). The comprehensive descriptive statistics can be found in Table 5.

Table 4 Descriptive statistics (age)
Table 5 Descriptive statistics (gender, country, company, job)

PLS-SEM approach

The validation of our research model is based on PLS-SEM, which is particularly appropriate for multistage models containing diversified constructs, indicators, and relationships (Hair et al., 2011; Ringle et al., 2012; Sarstedt et al., 2016). To achieve significant levels of statistical power (Hair et al., 2011), PLS-SEM requires a minimum sample size, which is determined by multiplying the total number of paths directed at specific latent constructs by 10 (Hair et al., 2011; Marcoulides et al., 2009). We met this requirement with N = 303 participants. Since the construct indicators in our model are reflective measurements caused by latent variables (Churchill, 1979), we used consistent partial least square (PLSc) as a more robust method compared to traditional PLS approaches (Dijkstra & Henseler 2015). We calculated the PLS algorithm using a path weighting scheme with 300 iterations and 10–7 as the stop criterion. Bootstrapping was performed with a two-tailed, bias-corrected, and accelerated (BCa) confidence interval method with 4,999 subsamples (Henseler et al., 2016). Blindfolding was conducted with an omission distance of seven. SmartPLS (v. 3.3.2) (Ringle et al., 2020) was used for the PLSSEM analyses, and jamovi (v. 1.2.27. 0) was used for descriptive statistics.


In this section, we present the measurement and structural model analysis of our research model, followed by the calculation of group differences based on the latest considerations for the model development process by Hair et al. (2019).

Measurement model analysis

We initially examined the indicator loadings and removed items with values below the threshold of .708 (Hair et al., 2011) to achieve items with sufficient reliability (Fornell & Larcker, 1981; Hair et al., 2019). Internal consistency was ensured by measuring Cronbach’s alpha, composite reliability, and Rho_A, resulting in sufficient values above 0.7 (cf. Table 6) (Diamantopoulos et al., 2012; Dijkstra and Henseler 2015; Hair et al. 2019). Convergent validity was estimated by assessing the average variance extracted (AVE) with results higher than 0.50, i.e., a minimum of 50 percent of the variance of the construct’s items (Fornell & Larcker, 1981; Hair et al., 2019; Henseler et al., 2016).

Table 6 Reliability and validity measurements. CA = Cronbach’s alpha, RA = rho_A, CR = composite reliability, AVE = average variance extracted

The discriminant validity is based on the Fornell-Larcker criterion (Table 10 in the Appendix) (Fornell & Larcker, 1981) and the heterotrait-monotrait (HTMT) ratio (Table 11 in the Appendix) (Henseler et al., 2016). For the Fornell-Larcker criterion, validity can be assumed as the square root of AVE is greater than any inter-factor correlation (Fornell & Larcker, 1981). Regarding HTMT, validity was given to values below 0.90 (Franke & Sarstedt, 2019; Henseler et al., 2016). Harman’s one-factor test was conducted, alleviating concerns about common method bias. The variance inflation factors (VIFs) were used to evaluate the collinearity of the formative indicators. The values were below the threshold of 3.30 (Kock, 2015) (Table 12 in the Appendix). We finally observed the cross-loadings excluding incorrect assignment of indicators to constructs (Henseler et al., 2016) (Table 13 in the Appendix).

Structural model analysis

The structural model analysis includes the statistical significance and relevance of the path coefficients, the coefficient of determination, and the blindfolding-based, cross-validated redundancy (Hair et al., 2019). Statistical significance was ascertained by measuring the p values and t statistics (Greenland et al., 2016), where p ≤ 0.05 and t > 1.96 values are sufficient. Cohen’s f2 was calculated to explain statistical relevance. Weak, medium, or large effect sizes ranged from 0.02 to 0.15, 0.15 to 0.35, or greater than or equal to 0.35 (Benitez et al., 2020). R2 demonstrates explanatory power (Shmueli & Koppius, 2011) with values between 0 and 1: 0.25 is considered weak, 0.5 is moderate, and 0.75 or above is substantial (Hair et al., 2011; Reinartz et al., 2009).

To support explanatory significance, we calculated the Stone-Geisser measure Q2 to determine predictive relevance, i.e., explaining how well the data can be reproduced by the research model (Geisser, 1974; Stone, 1974). Satisfying results were above 0; values higher than 0, 0.25, and 0.5 were considered small, medium, and large effect sizes (Hair et al., 2019). The goodness of fit as a global validation index for the efficiency of a model (Henseler & Sarstedt, 2013; Tenenhaus et al., 2005) was assessed using AVE and R2 (adjusted). This resulted in 0.66, which is above the threshold of 0.36, indicating a valid research model (Wetzels et al., 2009). We controlled our model using the participants’ ages, genders, and countries, which we found to be insignificant (Table 7).

Table 7 Major effects measurements. O = original sample, M = sample mean, SD = standard deviation, t = t statistic, p = p value

Finally, we validated the moderating role of AI identity on the effects of threat source factors (e.g., loss of autonomy, changes at work, loss of controllability, loss of skills/expertise, loss of status position and threat to job security) on AI identity threat. However, there we no significant correlations (p values between 0.21 and 0.78).

Multi-group analysis

Group differences were calculated using the partial least squares multi-group analysis (PLS-MGA) and statistically tested (Reinartz et al., 2009) against the null hypothesis of no significant variations between the groups. Significant differences of path coefficients were indicated by p values below 0.05 or above 0.95 (Reinartz et al., 2009). We validated three group differences: employees who work with AI in their companies (N = 143; 47.3%) compared to individuals who do not use this technology (N =159, 65.7%), participants who have been using AI in the workplace for a short time (less than 2 years) (N = 87; 54.7%), and those who are familiar with handling a system (more than 2 years of experience) (N = 64; 45.3%). All hypotheses (H1a,b; H2a,b; H3a,b; H4) were evaluated for each group separately and tested for significant differences between the groups in their parameter estimates. The results indicate that the same set of hypotheses were significant in each subgroup and that there were no existing differences.

Qualitative evaluation

Initially, the experts elaborated on the uniqueness of AI within IT (see Table 14 in the appendix). Overall, the interviewees agreed that AI differs in specific aspects from IT. As E2 explained, “AI and IT are not substituting but complementary to each other.” E5 emphasized that the development of AI is based on algorithms and methods that are also found in IT. However, all experts agreed that AI exceeds the limits of IT by providing functionalities for analyzing data that is IT not capable of. For example, E3 mentioned that AI could lead to the creation of a “blackbox”. Humans may impact the results by manipulating parameters, but, unlike with IT, the employee loses full control of the procedure. E6 supported this perspective, explaining that individuals “cannot determine how the algorithm makes decisions. This is how AI differs from IT.” From a managerial perspective, E6 emphasized that “IT does not reduce headcounts—AI does.” Furthermore, E6 postulated that “AI and IT have to co-exist” but “IT computes the data as it is…AI is able [to get] insights from the data itself.” Based on the feedback of the experts, it can be argued that AI and IT might have the same origin, but AI provides functionalities that break the limits of IT. AI further entails challenges such as replacing human tasks or black- boxing the process of decision-making. These aspects may evoke a unique perception of employees toward AI compared to IT.

To evaluate the identified predictors, we presented the research model to the interviewees. To avoid bias regarding the statistical signifiers, the empirical results were initially disclosed. The experts agreed that the derived research model was comprehensive. The interviewees did not disagree with a single identified predictor, and E3 concluded that “no predictor seems extraneous or missing.” Thus, all predictors are perceived as important and crucial in the context of AI identity threat. We next asked the experts which predictors could be of particular relevance when the experts mentioned changes to work (E1, E5), job security (E1, E2, E5), loss of status position (E1), loss of skills/expertise (E2, E3), loss of controllability (E2, E3, E4, E5), loss of autonomy (E4, E6), and AI identity (E6) as the most important. All the experts agreed that the workplace changes when introducing AI, and E1 and E5 particularly highlighted the predictors of changes to work as a relevant influence on AI identity threat. Both perceived this as closely linked to the job. When introducing AI in the workplace, the experts emphasized the need for briefing the employees about the function, role, and capabilities of the specific AI. If this step is not done properly, threats related to job security and changes to work increase. Furthermore, E2 mentioned that the degree of perceived threat may differ between the levels of hierarchies: “There are employees who are still 1-2 hops away from being directly affected by AI. The focus is on indirectness that simple jobs are indirectly eliminated by AI. Here, a distinction must be made between management and positions such as mechanics.”

E1 highlighted that introducing AI might not only affect the perception of changes of work and job security but also significantly impacts perceived status positions in the workplace. E1 described this issue as follows: “It is sometimes unclear how good a person is at the job. However, AI requires a process to be measurable. As a result, it then becomes apparent that the human is not good at his job.” Furthermore, E1 distinguished between employee and manager perspectives. From an employee perspective, the human loses; from a managerial perspective, the company wins.

Considering the predictor loss of skills/expertise, E2 mentioned that a possible loss of skills may increase over time, thus potentially turning into a major threat. In contrast, E1 argued that there are use cases in which AI functions as a “preserver of knowledge.” However, E3 emphasized focusing on the human rather than the technology. Thus, AI might provide certain knowledge, but individual employees might lose abilities in the long run. E4 supported this argument by elaborating that “AI could provide knowledge and skills”; thus, the employee is not required to maintain knowledge.

Expertise with AI is also closely related to feelings of losing autonomy and controllability (E2). As AI is capable of performing certain tasks, this may lead to situations in which humans are not able to control certain conditions (E2, E3). Thus, working with AI also involves employees providing data regarding their own actions and work. E4 mentioned that “once the data is out there, the employee has no control over the data.” Likewise, E3 described that a loss of autonomy in the workplace indicates loss of controllability, as “if I am no longer autonomous, I lose a controllable environment. These factors are interrelated, they merge.” Regarding the feedback from E6, AI identity is one crucial factor that should be considered when broadening our understanding of AI identity threat: “human experience is very important, and decisions should not be accepted without supervision.” The employee should remain the central element in the workplace, and perceptions of and experiences with AI are fundamental for the process of introducing it to the workplace (E6).

The experts also provided valuable insights on strategies countering the AI identity threat by considering the identified predictors. One major factor that was mentioned is the human itself. For “shaping a better tomorrow” (E4) and reducing the negative impact of AI on humans, organizations need trustworthy, ethical, and responsible employees (E4, E6, E5). Currently, “AI is trained by humans with agenda and motivation” (E4). The output provided by AI is based on (input) data that sometimes lacks context; thus, a responsible employee needs to supervise the decision-making process (E6). However, employees need insights into the AI’s black box to understand its results. Thus, explainable AI is needed to better understand the underlying processes (E2, E3, E5, E6).


With the ongoing excitement of AI in electronic markets, human-AI interaction is increasingly being accepted by society, transforming enterprises across the globe (Adam et al., 2020; Guan et al., 2020; Thiebes et al., 2020). Organizations strive to understand the impact of AI on humans (Neuhofer et al., 2020), especially how AI can be introduced in the workplace to interact with employees. However, along with benefits for organizations, novel negative consequences for individuals arise. It is thus crucial to comprehend related drawbacks to reduce or eliminate the impacts of the dark sides of AI.

In this study, we identified predictors contributing to AI identity threat in the workplace, such as employees’ perception of AI as indispensable (Mirbabaie et al., 2020). To this end, the evaluated research model explains 56% of the variance of AI identity threat in the workplace. Three out of our seven constructs in the model are substantial, confirming hypotheses H1b, H2b, and H4. Therefore, we answer our research question by identifying loss of status position, changes to work, and AI identity as significant predictors influencing AI identity threat to employees with AI experience in the workplace. These findings are in accordance with the evaluation of the interviewed experts, who fully supported the derived research model.

First, the construct of loss of status position positively impacts AI identity threat. This is in line with the findings of Jussupow et al. (2018), who found that AI acts as a direct threat to the social position of an employee. Therefore, employees who fear a loss of their own competences or independence by AI tend to perceive a higher AI identity threat. For example, we asked the participants about their perception of statements such as “The introduction of AI will reduce my competences” or “By the introduction of AI, I am limited in my possibilities.” In accordance, Israeli (2019) described that employees seek to be perceived as accountable for their work and to enjoy their colleagues’ and clients’ appreciation. By introducing AI in a collaborative setting, employees might expect a loss of their competence and possibilities that lead to a threat to one’s identity. Thus, we are contributing to the threat source of intergroup conflict, by identifying the construct loss of status position as a relevant AI identity threat predictor (Craig et al., 2019). Identity threats classified in the domain of intergroup conflicts are, for example, connected to the reputation and self-beliefs within a team at work. Thus, the construct of loss of status position mirrors the impact on AI identity threat on a professional identity level (Jussupow et al., 2018). Understanding this relationship might be crucial for reducing harmful effects of AI on the employees.

Second, the construct of changes to work according to the threat source category verification prevented positively affects AI identity threat of employees in the workplace. This is in accordance with Craig et al. (2019), who report that introducing a new technology is often related to changes in the workplace. This becomes particularly evident from the results of the interviews, which emphasized the unique possibilities of AI in relation to IT. Thus, the findings suggest that, for example, employees who perceive more responsibilities but also the increased importance of their job positions tend to observe higher identity threat caused by AI. This could be due to the assumption that employees do not feel sufficiently prepared for the introduction of AI (Grundner & Neuhofer, 2021). These harmful influences on the identity of employees are based on personal perception and inspection of self-beliefs initiated or influenced by technology (Craig et al., 2019). In this context, employees need to reevaluate their traditional roles and responsibilities in light of upcoming changes in their workplaces (Craig et al., 2019; Hicks, 2014). This reevaluation takes place on the level of professional identity (Brooks et al., 2011; Jussupow et al., 2018).

Considering these two key findings reveals that the perceived change of work as well as the perceived loss of status position leads to higher AI identity threat. Subsequently, employees may perceive a loss of self-esteem in the domains of worth, competence, and authenticity (Craig et al., 2019).

As a third key finding, this study identified the construct of AI identity as a negative predictor for AI identity threat. However, the findings did not show a moderating effect of AI identity on the identified predictors of the AI identity threat (e.g., loss of status position, changes of work, or loss of autonomy). Thus, we can assume that AI identity is an independent predictor for further understanding and reducing the identity threat caused by AI. Nevertheless, the identification with AI does not directly impact the specific threat sources evoked by AI itself. From this we can conclude that employees who perceive interaction with AI as an indispensable component of their work perceive lower threat of AI to their identities. This in line with Craig and colleagues (Craig et al., 2019) describing the basic premise of the identity threat caused by technology as the anticipation of further harm leading to resistance behavior in the workplace. A positive association with AI could reduce resistance behavior and support the individual acceptance of AI. This supports the assumptions that humans and technology complement each other when individuals form their identities with technology as an integral component of themselves (Park & Kaye, 2019). Mirbabaie et al. (2021) reveal a connection between identification with AI as a teammate as well as a significant part of the employee’s identity. After forming this intertwinement (Carter and Grover, 2015), users tend to be more committed to working with AI conforming to their identities (Deci & Ryan, 2012; Reychav et al., 2019). A positive AI identity might stabilize individuals’ self-esteem related to AI and therefore reduce anticipated identity threats caused by AI (Carter and Grover, 2015; Craig et al., 2019; Petriglieri, 2011). Furthermore, establishing a positive AI identity in the workplace may strengthen one’s professional identity by reducing the harmful influence of the identity threat caused by AI. This interplay of the levels of identity, professional identity, and AI identity might be related to a positive development of self-identity (Carter and Grover, 2015; Craig et al., 2019; Park & Kaye, 2019). To better conceptualize the findings of this study, Figure 4 introduces the aggregated formation of predicting AI identity threat in the workplace.

Fig. 4
figure 4

Aggregated formation of predictors for AI identity threat in the workplace

In contrast, this study identified further possible predictors of the AI identity threat in the workplace that did not show significance but were highlighted by the experts. Within meaning change, autonomy illustrates the independent behavior of employees in the workplace, and controllability refers to the perceived control of employees. Recent research has identified these factors as major threats to identity (Bernardi & Exworthy, 2020; Craig et al., 2019) and extensions of professional identity in the workplace, further related to personal skills and efforts (Jussupow et al., 2018). However, the findings in this study did not provide significance. Thus, employees’ perception of losing of skills, autonomy, controllability, or their jobs as a result of the introduction of AI does not seem to influence a possible threat to AI identity in the context of AI-experienced employees. We interpret this to the mean that these constructs might be relevant when introducing AI in the workplace but are not relevant to this study’s focus (Carter and Grover, 2015; Craig et al., 2019). As literature as well as the qualitative findings hint at an essential role, we assume that the predictors are valid considering a different sample with divergent characteristics (i.e., no AI experience or expertise). In contrast to the findings of Mirbabaie et al. (2020), who describe a symbiotic association of identification with AI and colleagues, this study reveals how the predictors allocated to professional identity impact the identities of employees when introducing AI in the workplace. Looking closer at the construct of AI identity reveals that the initial measurement item of dependence was removed during the examination of indicator loadings to achieve a first-order construct with significant reliability (Fornell & Larcker, 1981; Hair et al., 2011, 2019). This needs to be mentioned in the context of the study by Carter (2012), from which the items were adapted, analyzed, and conceptualized referring to smartphones. We comprehend that AI, on the one hand, is not really a haptic object comparable to smartphones and, on the other, is currently not very widespread. It thus seems that dependence might not be a decisive factor (yet). AI is somewhat perceived as a component “floating around”, which clearly distinguishes it from traditional IT and explains why the original scale does not fit precisely.

Furthermore, the findings of this study do not support that a specific degree of experience with AI in the workplace alters the impact of the identified predictors. This contradicts the assumption that employees might not know what AI is really capable of and have a restricted perception of what collaboration with AI at their workplace would mean for them (Johnson & Verdicchio, 2017; Kühl et al., 2019). Thus, the attitude towards AI in the workplace might be detached from individual AI experience and anchored at a common AI setting.

This research is not free of limitations. The theoretical character of the identification of the predictors could be a limiting factor. Asking employees about potential influences may deliver further predictors that this study did not consider. In this context, two out of four items of the scale of loss of skills/expertise refer to the individual perception of employees in general that might influence the relationship with AI identity threat. However, this study is based on an extensive SLR that covers several studies, including qualitative and quantitative research. Therefore, knowledge about IT and AI identity threat is accumulated in this work. AI is a fast-developing technology that will impact the workplace over the coming decades. We thus suggest future research that evaluates predictors of AI identity threat. As this study is based on a sample with AI experience, we suggest that further research should focus on employees without AI experience. Additionally, we suggest that further research should consider the identified constructs when investigating human-AI collaboration in the workplace. As this research is in the IS discipline, the findings are mostly based in this research domain. As IS research is interdisciplinary, this study already covers insights from various scientific perspectives. However, further predictors might be revealed by extending this study to research fields that are not connected or are only minorly connected to the IS community.


In this research, we focused on AI identity threat, which is constructed by employees facing potential collaboration with AI. This study proposes a theoretical framework that identifies predictors of AI identity threat. Overall, our developed research model explains 56% of the variance of AI identity threat in the workplace. We found that changes to work, loss of status position, and AI identity are significant predictors. Furthermore, we indicated that experience with AI does not alter the perception of AI identity threat.

This study contributes to identity research by elaborating the role of AI as a unique technology in the context of identity threat in the workplace. Furthermore, this study reveals the urgent need for research focusing on the impact of AI on the individuum. To this end, this study introduces and defines the concepts of the AI identity threat and AI identity by conducting an SLR and expert interviews. This study broadens the understanding of the identity threat caused by AI by revealing loss of status position, changes of work, and AI identity as significant predictors for resolving the AI identity threat in the workplace. To this end, this study develops a research model to broaden our understanding of the AI identity threat. Regarding the identity theory's threat sources, loss of status position could be assigned to the domain of intergroup conflict and changes of work to verification prevented. This allocation gains importance when developing counteractions for reducing the AI identity threat. In summary, we reveal fundamental factors related to the negative impacts of AI on humans, the so-called dark side of AI.

This study further contributes to practice by revealing specific influencing factors that should be considered when introducing AI in the workplace. Practitioners may consider that individual experience with AI seems less relevant than expected by the literature. However, by first explaining and defining AI identity and further revealing influencing factors on employees’ identities in the workplace, practitioners should consider the special role of identity in the context of human-AI collaboration. Decision makers, such as managers and executives, may consider these insights to reduce upcoming resistance to AI in the workplace. Furthermore, it is vital to support a positive identification with AI in the workplace to reduce potential threats to employees’ identities. This might be achieved by reducing potential threats evoked by expected changes to work or personal status position. Based on the experts’ evaluation, we further suggest introducing explainable AI as well focusing on trustworthy, responsible, and ethical employees who supervise AI implementations and applications.

In terms of contributions to society, this study suggests a symbiotic relationship between humans and AI to support human self-esteem and well-being in the workplace. Through continuing education about AI’s capabilities in electronic markets, individuals and enterprises may use AI as a technology to extend their skills and competences in the workplace.