Introduction

Artificial Intelligence (AI) has been named as one of the most recent, fundamental developments of the convergence in electronic markets (Alt, 2021) and has become an increasingly relevant topic for information systems (IS) research (Abdel-Karim et al., 2021; Alt, 2018). While a large body of literature is concerned with designing AI to mimic and replace humans (Dunin-Barkowski, 2020; Fukuda et al., 2001), IS research in general, and decision support systems (DSS) research in particular, emphasize the support of humans with AI (Arnott & Pervan, 2005). Recent research in hybrid intelligence (HI) and human-AI collaboration offers a promising path in synthesizing AI research across different fields (Dellermann, 2019): The ultimate goal of HI is to leverage the individual advantages of both human and artificial intelligence to enable synergy effects (James & Paul, 2018) and to achieve complementarity (Hemmer et al., 2021).

However, in many cases in both research and practice, AI is simply equated with the concept of machine learning (ML)—negatively impacting terminological precision and effective communication. Ågerfalk (2020, p.2) emphasizes that differentiating between AI and ML is especially important for IS research: “Is it not our responsibility as IS scholars to bring clarity to the discourse rather than contributing to its decline? (…) It would mean to distinguish between different types of AI and not talk of AI as synonymous with ML, which in itself is far from a monolithic concept.”

The practical relevance of a clear understanding is underlined by observing confusion and misuse of the terms AI and ML: During Mark Zuckerberg’s U.S. senate hearing in April 2018, he stressed that Facebook had “AI tools to identify hate speech” as well as “terrorist propaganda” (The Washington Post, 2018). Researchers, however, would usually describe tasks identifying specific social media platform instances as classification tasks in the field of (supervised) ML (Waseem & Hovy, 2016). The increasing popularity of AI (Fujii & Managi, 2018) has led to the term often being used interchangeably with ML. This does not only hold true for the statement of Facebook’s CEO above, but also across various theoretical and application-oriented contributions in recent literature (Brink, 2017; ICO, 2017; Nawrocki et al., 2018). Camerer (2017) even mentions that he still uses AI as a synonym for ML despite knowing it is inaccurate.

As the remainder of this paper shows, both concepts are not identical—although in many cases both terms will appear in the same context. Such ambiguity might lead to multiple imprecisions in both research and practice when conversing about the relevant concepts, methods, and results. This is especially important in IS research—being interdisciplinary by nature (D’Atri et al., 2008). Ultimately, misuse can either lead to fundamental misunderstandings (Carnap, 1955) or to research that ought to be undertaken not being conducted (Davey & Cope, 2008; Lange, 2008). After all, misunderstandings can potentially lead to low perceived trustworthiness of AI (Thiebes et al., 2021).

It seems surprising that despite the frequent use of the terms, there is hardly any helpful academic delineation—apart from the notion that ML is a (not well-defined) subset of AI (Campesato, 2020), comparable to other possible subdisciplines of AI: Expert systems, robotics, natural language processing, machine vision, and speech recognition (Collins et al., 2021; Dejoux & Léon, 2018). Consequently, this paper aims to shed light on the relationship between the two concepts: We analyze the role of ML in AI and, more precisely, in intelligent agents, which are defined by their capability to sense and act in an environment (Schleiffer, 2005). We do so by taking an ML perspective on intelligent agents’ capabilities and their relevant implementation—with IS research in mind. To this end, we review the relevant literature for both terms and synthesize and conceptualize the results.

Our article’s contributions are twofold: First, we identify different contributions of ML to intelligent agents as specific AI instantiations. We base this on an expansion of the existing AI framework by Russell and Norvig (2020) — explicitly breaking down intelligent agents’ capabilities into separate “execution” and “learning” capabilities. Second, we develop a typology to provide a common terminology for AI-based information systems, where we conceptualize which systems employ ML—and which do not. The result should provide guidance when designing and analyzing systems.

Next, in Section “Terminology”, we review relevant literature in the fields of AI and ML. In Section “The role of rational agents in information systems”, we then analyze the capabilities of intelligent agents in more depth and examine the role of ML in them. Section “Towards a typology for machine learning in AI systems” develops a framework and typology to differentiate the terms AI and ML and to explain their relationship. In Section “Conclusion”, we conclude with a summary.

Terminology

Over the last decade, both terms, artificial intelligence (AI) and machine learning (ML), have enjoyed increasing popularity in information systems (IS) research. An analysis of the “AIS Senior Scholars’ Basket”Footnote 1 journals since 2000,Footnote 2 illustrates how the occurrences of both terms increased in titles, abstracts, and keywords (Fig. 1). While over the last 21 years, we observe a small but constant number of publications covering AI-related topics, ML only gained relevance in the literature after 2017: The late reflection of ML—despite of the earlier adoption and spread in industry (Brynjolfsson & Mcafee, 2017)—may raise questions about whether IS has picked up the topic early enough.

Fig. 1
figure 1

Appearance of the terms “artificial intelligence” and “machine learning” and in AIS Senior Scholars’ Basket journals

As the analysis demonstrates, the two terms do exist for quite some time, while their related subjects are highly and increasingly topical now. In this section, we will elaborate on the meaning of the terms.

Artificial intelligence

In 1956, a Dartmouth workshop, led by Minsky and McCarthy, coined the term “artificial intelligence” (McCarthy et al., 1956) —later taking in contributions from a variety of different research disciplines, such as computer science (K. He et al., 2016) and programming (Newell & Simon, 1961), neuroscience (Ullman, 2019), robotics (Brady, 1984), linguistics (Clark et al., 2010), philosophy (Witten et al., 2011), and futurology (Koza et al., 1996). While the terminology is not well defined across disciplines, even within the IS domain definitions do vary widely; Collins et al. (2021) provide a comprehensive overview. Recent AI definitions transfer the human intelligence concept to machines in its entirety as “the ability of a machine to perform cognitive functions that we associate with human minds, such as perceiving, reasoning, learning, interacting with the environment, problem solving, decision-making, and even demonstrating creativity” (Rai et al., 2019, p.1). Still, over the last decades various debates have been raging on the depth and objectives of AI. These two dimensions span the space for different AI research streams in computer science and IS that were categorized by Russell and Norvig (2020): On the one hand (depth dimension), it may target either the thought process or a concrete action (thinking vs. acting); on the other hand (objective dimension), it may try to either replicate human decision making or to provide an ideal, “most rational” decision (human-like vs. rational decision). The resulting research streams are depicted in Table 1.

Table 1 AI research streams; based on Russell and Norvig (2020).

According to the cognitive modeling (i.e., thinking humanly) stream, AI instantiations must be “machines with a mind” (Haugeland, 1989) that perform human thinking (Bellman, 1978). Not only should they arrive at the same output as a human when given the same input, but also apply the same reasoning steps leading to this conclusion (Newell & Simon, 1961). The laws of thought stream (i.e., thinking rationally) requires AI instantiations to arrive at a rational decision despite what a human might come up with. AI must therefore adhere to the laws of thought by using logic-based computational models (McDermott & Charniak, 1985). The Turing test stream (i.e., acting humanly) implies that AI must act intelligently when interacting with humans. To accomplish such tasks, AI instantiations must perform human tasks at least as well as humans (Rich & Knight, 1991), which can be tested via the Turing test (Turing, 1950). Finally, the rational agent stream considers AI as a rational (Russell & Norvig, 2020) or intelligent (Poole et al., 1998) agent.Footnote 3 This agent does not only act autonomously, but also with the objective of achieving the rationally ideal outcome.

Machine learning

Many researchers perceive ML as an (exclusive) part of AI (Collins et al., 2021; Copeland, 2016; Ongsulee, 2017). In general, learning is a key facet of human cognition (Neisser, 1967). Humans process a vast amount of information by utilizing abstract knowledge that helps them to better understand incoming input. Owing to their adaptive nature, ML models can mimic a human being’s cognitive abilities (Janiesch et al., 2021): ML describes a set of methods commonly used to solve a variety of real-world problems with the help of computer systems, which can learn to solve a problem instead of being explicitly programmed to do so (Koza et al., 1996). For instance, instead of explicitly telling a computer system which words within an tweet would indicate it to contain a customer need, the system (given a sufficient set of training samples) learns the typical patterns of words and their combination which results in a need classification (Kühl et al., 2020).

In general, we differentiate between unsupervised, supervised, and reinforcement ML. Unsupervised ML comprises methods that reveal previously unknown patterns in data. Consequently, unsupervised learning tasks do not necessarily have a “correct” solution, as there is no ground truth (Wang et al., 2009).

Supervised ML refers to methods that allow the building of knowledge about a given task from a series of examples representing “past experience” (Mitchell, 1997). In the learning process, no manual adjustment or programming of rules or strategies to solve a problem is required, i.e., the model is capable to learn “by itself”. In more detail, supervised ML methods always aim to build a model by applying an algorithm to a set of known data points to gain insight into an unknown set of data (Hastie et al., 2017): Known data points are semantically labeled to create a target for the ML model. So-called semi-supervised learning combines elements from supervised and unsupervised ML by jointly using labeled and unlabeled data (Zhu, 2005).

Reinforcement learning refers to methods that are concerned with teaching intelligent agents to take those kinds of actions that increase their cumulative reward (Kaelbling et al., 1996). It differs from supervised learning in that no correctly matched features and targets are required for training. Instead, rewards and penalties allow the model to continuously learn over time. The focus is on a trade-off between the exploration of the uncharted environment and the exploitation of the existing knowledge base.

The role of rational agents in information systems

To further elaborate on the role of ML within AI, we need to take a clear perspective on the different definitions of AI to be beneficial to IS research. IS traditionally utilizes ML in predictive analytics tasks within (intelligent) decision support systems (DSS) (Arnott & Pervan, 2005; Müller et al., 2016) where the goal is to generate the best possible outcome (Arnott, 2006; Hunke et al., 2022; Power et al., 2019). As Phillips-Wren et al. (2019, p.63) emphasize, DSS “should help the decision-maker think rationally”. The perspective of rationality is also endorsed by other researchers in the field (Bakos & Treacy, 1986; Dellermann, 2019; Kloör et al., 2018; Power et al., 2019; Schuetz & Venkatesh, 2020). Thus, in the following we will explore the relationship between ML and AI in IS from the lens of the rational agent stream as discussed above. Furthermore, we will focus on supervised ML as it is the most common type of ML (Jordan & Mitchell, 2015). In the remainder of this section, we will first distinguish different types of (rational) agents and then use the insights to differentiate between the necessary layers when designing them as part of information systems.

Types of rational agents

According to the selected research stream, intelligence manifests itself in how rational agents act. Five features characterize agents in general: they “operate autonomously, perceive their environment, persist over a prolonged time period, adapt to change, and create and pursue goals” (Russell & Norvig, 2020, p.4). An agent defines its action, not for itself, but within the environment it operates and interacts with. It recognizes the environment through its sensors, relies on an agent program to handle and digest input data, and performs actions via actuators. A rational agent targets to achieve the highest expected outcome according to one or multiple objective performance measures—which are based on current and past knowledge of the environment and possible actions. For example, a rational agent within a medical diagnosis system aims to maximize the health of a patient measured via blood pressure, heart rate, and blood oxygen (potentially while minimizing the financial costs of a treatment as a secondary condition) (Grosu, 2022).

The agent’s conceptualization and surroundings are summarized in the agent-environment framework. It consists of three components: an agent, an environment, and a goal. Intelligence is the measurement of the “agent’s ability to achieve goals in a wide range of environments” (Legg & Hutter, 2007, p. 12). The agent obtains input through perceptions that the environment generates. Observations of the environment are one type of perception, while others are reward signals that indicate how well the agents’ goals have been achieved. Based on these input signals, the agent decides to perform actions, which are subsequently communicated back as signals to the environment.

Rational agents in information system architectures

As we investigate the role of ML in AI for IS research, we also need — apart from the theoretical and definitory aspects of agents — to consider how the functionality of a rational agent is reflected in an IS architecture. The implementation of agents is a key step to embed their functionality into practical, real-world (intelligent) information systems in general or into DSS specifically (Gao & Xu, 2009; Zhai et al., 2020). Any rational agent needs to be capable of at two least two tasks: cognition (Lieto et al., 2018) and (inter)action with the environment (Russell & Norvig, 2020). If we map these capabilities to system design terms, then acting capabilities are the ones built into a frontend, while the cognitive capabilities are embedded in a backend.

The frontend as the interface to the environment may take various forms; it may be designed as a very abstract, machine-readable web interface (Kühl et al., 2020), a human-readable application (Engel et al., 2022; Hirt et al., 2019), or even a humanoid template with elaborated expression capabilities (Guizzo, 2014). For the frontend to interact with the environment, two technical components are required: sensors and actuators. Sensors detect events or changes in the environment and forward the information via the frontend to the backend. They can, for instance, read the signals within an industrial process network (Hein et al., 2019), read visuals of an interaction with a human (Geller, 2014), but also perceive a keystroke input (Russell & Norvig, 2020). Actuators, on the other hand, are components responsible for moving, controlling, or displaying content. While sensors merely process information, actuators act, for instance, by automatically making bookings (Neuhofer et al., 2021) or changing a humanoid’s facial expressions (Berns & Hirth, 2006). One could argue that the Turing test (Turing, 1950) takes place at the environment’s interaction with the frontend, or, more precisely, when sensors and actuators are combined in a way to test the agent’s AI for acting humanly.

The backend provides the required functionalities to depict an intelligent agent’s cognitive capabilities. More precisely, this executing backend allows the agent to draw on its built-in knowledge. The backend translates signals from the frontend and transforms them into signals sent back to the frontend as a response by executing actions. In some cases, there is an additional component modifying this response function over time, and thus modifying the execution part of the backend. We call this the learning part of the backend as depicted in Fig. 2. Within the next subsections, we will further elaborate this framework and its components.

Fig. 2
figure 2

Conceptual framework describing the general architecture for intelligent agents in AI-based information systems

The role of machine learning in rational agents

In terms of supervised ML, we need to further differentiate between the process task of building (training) adequate ML models (Witten et al., 2011) and the one of executing the deployed models (Chapman et al., 2000). To further understand ML’s role in intelligent agents, we partition the agent’s cognition layer into a learning sublayer (model building) and an executing sublayer (model execution).Footnote 4 We, therefore, regard the implementation required by the learning sublayer as the learning backend, while the executing backend denotes the executing sublayer.

The learning backend first dictates if the intelligent agent is able to learn, and, second, how it does sowith respect to the algorithms it actually uses, the type of data processing it applies, and the handling of concept drift (Gama et al., 2014). Using the terminology of Russell and Norvig (2020), we distinguish two different types of intelligent agents: simple-reflex agents and learning agents. This differentiation holds explicitly in terms of a ML perspective on AI because it considers whether the underlying models in the cognition layer are trained just once and after that never touched (simple-reflex), or whether they are continuously updated to be adaptive (learning). Related work provides suitable examples of both. Kitts and Leblanc (2004) build a bidding agent for digital auctions as a simple-reflex agent: While building and testing the model for the agent may show convincing results, the system’s adaptive learning after deployment could be critical. Other examples of agents with models trained just once are common in different areas, for example, in terms pneumonia warning for hospitals (Oroszi & Ruhland, 2010), the (re)identification of pedestrians (Z. Zheng et al., 2017), and object annotation (Jorge et al., 2014). On the other hand, recent literature also provides examples of learning agents. Mitchell et al. (2015) present the concept of “never-ending learning” agents that strongly focus on continuously building and updating models in agents. Neuhofer et al. (2015) suggest an agent capable of personalization through a continuous learning processes of guest information for digital platforms, which an example of such an agent. Other examples include agents capable of making recommendations on music platforms (Liebman et al., 2015), regulating heat pump thermostats (Ruelens et al., 2015), acquiring collective knowledge across different tasks (Rostami et al., 2017), and learning the meanings of words (Yu et al., 2017). The choice of the learning type in agents (simple-reflex vs. learning agent) influences the agent’s general overall design and the contribution of ML.

As a result from the layers of agents and types of learning, our conceptual framework combining both is shown in Fig. 2. Regarding the previously mentioned ML methods, supervised ML can be the basis for either simple-reflex or learning agents, depending on whether the learning backend exists and on its feedback to the agent’s knowledge base. In terms of reinforcement learning, the agent, by definition, is a learning agent. However, there are also examples of where an agent functions without the utilization of ML—because the execution is based on rules (H. Wang et al., 2005), formulas (Billings et al., 2002) or other methods (Abasolo & Gomez, 2000). From this perspective, this means there can be AI without ML.

Towards a typology for machine learning in AI systems

Based on the differentiation between simple-reflex and learning agents, we can now derive a typology for IS research. We refer to IS systems as static AI-based systems if they employ simple reflex agents that may be based on a model trained with ML. Adaptive AI-based systems, though, use learning agents, i.e., do have a learning backend— that may be based on ML, but alternatively also could be based, e.g., on rule-based knowledge representation. We, thus, propose the typology (as depicted in Table 2) for AI-based IS along the two dimensions: the existence of an ML-base for the executing backend and the existence of a learning backend.

Table 2 Typology of AI-based information systems

We illustrate these findings in concrete IS research examples: Static AI systems are characterized by an executing backend which is based on algorithms not classified as ML and they lack a learning backend, i.e. they have a fixed response model (Chuang & Yadav, 1997). The executing backend of such systems is based on rules (like nested if-else statements), formulas (like mathematic equations describing a phenomena) or algorithms (like individual formal solution descriptions for specific problems). As an example for such systems, Hegazy et al. (2005) build a static AI system based on a self-developed algorithm and evaluate its performance within a cybersecurity context by simulating multiple attacks. Another example is provided by Ritchie (1990) who has developed an architecture and an instantiation of a static AI system for a traffic management platform.

In contrast, a static ML-based AI system has an executing backend which is based on ML. An example is provided in S. He et al. (2018). The authors develop an artifact to classify marketing on Twitter in either defensive or offensive marketing and show convincing prediction results. While their work did not aim at designing a productive artifact and is rather focused on showing the general feasibility of the approach, they choose a static ML-based AI system—which, however, might not be sufficient for permanent use: After the release of the article in 2018, Twitter changed its tweet size from 140 to 280 characters, thus changing the environment. It would be interesting to see how the developed model would need to adapt to this change. As another example, Samtani et al. (2017) build a model to identify harmful code snippets, typically utilized by hackers. They show how to design an artifact that can detect these code assets accurately for a proactive cyber threat intelligence. However, also in this case the environment and the assets of the hackers could and will change over time.

Adaptive AI systems, which are not based on ML, do comprise an executing backend with the flexibility to dynamically adapt the model to changing environments. This type of system is oftentimes enabled through the interaction between humans and AI systems. Most of the times, the system provides means and triggers for updates, while the human provides “manually encoded” knowledge updates. For example, Zhou et al. (2009) implement an adaptive AI system for pipeline leak detection which is based on a rule-based expert system and offers means to update the system online. In another example, Hatzilygeroudis and Prentzas (2004) develop an adaptive AI system to support the teaching process which has a specific component for knowledge updates. Both examples are inherently knowledge-based, but are explicitly designed to allow and force updates—although not on the basis of ML.

Finally, adaptive ML-based AI-systems implement learning in both sublayers of the cognition layer. For example, Q. Zheng et al. (2013) design a reinforcement-learning-based artifact to obtain information from hidden parts (“deep web”) of the internet. As their developed system perceives its current state and selects an action to submit to the environment (the deep web), the system continuously learns and builds up experience. In another example, Ghavamipoor and Hashemi Golpayegani (2020) build an adaptive ML-based AI system to predict the necessary service quality level and adapt an e-commerce system accordingly. As their system is continuously learning, their results show the total profits improve through effective cost reduction and revenue enhancement.

Conclusion

In this article, we clarify the relationship of machine learning (ML) in artificial intelligence (AI), particularly in intelligent agents, for the field of information systems research. Based on a rational agent view, we differentiate between AI agents capable of continuously improving as well as those who are static. Within these agents as instantiations of artificial intelligence, (supervised) ML can serve to support in different ways: either to contribute a once-trained model to define a static response pattern or to provide an adaptive model to realize dynamic behavior. As we point out, both could also be realized without the application of ML. Thus, “ML” and “AI” are not terms that should be used interchangeably—but as a conscious choice. Without question, ML is an important driver of AI, and the majority of modern AI cases will utilize ML. However, as we illustrate, there can be cases of AI without ML (e.g., based on rules or formulas).

This distinction enables our proposed framework to apply an intelligent agent’s perspective on AI-based information systems, enabling researchers to differentiate the existence and function of ML in them. Interestingly, as of today, many AI-based information systems remain static, i.e. employ once-trained ML models (Kühl et al., 2021). With increasing focus on deployment and life cycle management, we will see more adaptive AI-based systems that sense changes in the environment and use ML to learn continuously (Baier et al., 2019). Our framework and the resulting typology should allow IS researchers and practitioners to be more precise when referring to ML and AI, as it highlights the importance of not using the terms interchangeably but clarifying the role ML plays in AI’s system design.