1 Introduction

Interactions with conversational AI (CAI) systems become more and more widespread in everyday life. Virtual voice assistants offer hands-free communication in people’s homes and cars and are increasingly deployed in public spaces such as in health institutions, accommodation places, or professional environments. Moreover, text-based systems are common to assist people in their online activities, e.g., in e-commerce or finance. To provide assistance, these systems ask people to disclose various personal information or request access to a wide range of personal data. In addition, inferences can be drawn from users’ voice recordings or transcripts revealing sensitive information about themselves (Singh, 2019; Welch et al., 2019). To protect themselves from self-disclosure, users are faced with an increasing number of privacy and security decisions. They do not only need to decide whether to share information with a system but also in which way their information can be used and processed. However, given the complexity of the conversational AI ecosystem, it can be difficult for users to understand data-processing flows and possible implications to their privacy—a condition that is described as information asymmetry (Abdi et al., 2019; Acquisti et al., 2018). Moreover, information disclosure can be amplified as CAI systems aim to interact with users in a natural and human-like way and create an enjoyable and frictionless interaction (Seaborn et al., 2022). A positive mood can thereby influence people to underestimate privacy and security threats, increase the likelihood of disclosure, and serve as a mental shortcut (Alashoor et al., 2018; Dinev et al., 2015). In addition, peoples’ privacy decision-making can suffer from varying systematic deviations in judgements, i.e., cognitive and behavioural biases (Acquisti et al., 2018).

While recognizing the importance of evaluation and mitigation of algorithmic biases for CAI systems (Beattie et al., 2022; Orphanou et al., 2022), in this paper, we focus on cognitive biases and their impact on human decision-making concerning privacy and security throughout the development and usage of CAI systems. Moreover, a large body of research has discussed privacy and security attacks, privacy risks for conversational AI systems and mitigation strategies from a technical point of view (Alepis & Patsakis, 2017; Bispham et al., 2022; Bispham et al., 2020; Pal et al., 2020). While technical safeguards are essential to ensuring privacy and security, it is equally important to support people in overcoming their biases to enable the implementation and usage of privacy-preserving techniques and non-regrettable privacy decision-making. Consequently, we take on a human-centric approach by focusing on the people involved in development, deployment, and usage of these systems and present strategies that support their privacy decision-making.

To understand and assist users’ privacy and security decision-making, a growing body of research has applied behavioral economics (Acquisti et al., 2015; Ioannou et al., 2021). Design strategies that build on behavioral economic research aim to mitigate cognitive biases and improve users’ privacy choices. One stream of behavioral economic research has explored nudging strategies to nudge users towards “better” decisions without restricting their options (Thaler & Sunstein, 2021). While nudging strategies for privacy decision-making have been successfully deployed and tested in the context of mobile applications, e-commerce, and social media (Acquisti et al., 2018; Almuhimedi et al., 2015; Ioannou et al., 2021; Wang et al., 2013), to the best of our knowledge, nudging strategies for CAI have been researched more generally but without specific focus on privacy choices (Zargham et al., 2022). However, as every design decision can influence users’ choices for better or worse (Thaler & Sunstein, 2021), system providers and conversation designers have the potential to create a more private and secure experience by understanding users’ cognitive and behavioral biases and by applying nudging techniques. In addition, cognitive flaws can prevent system providers and developers from designing, implementing, and deploying secure and private systems in the first place. Therefore, strategies based on behavioral economics can assist all actors in the ecosystem to make better judgements regarding privacy and security.

Nudging strategies are only one way to mitigate cognitive biases and support people’s decision-making. In social science research, they are largely described as modifications to the environment (Soll et al., 2015). Yet, debiasing strategies can also focus on modifying a person’s cognitive process. The medical field is especially rich in strategies that aim to mitigate biases and support decision-making through educational and cognitive strategies (Croskerry et al., 2013; Lambe et al., 2016). Therefore, we draw from previous categorizations of debiasing strategies to establish a novel categorization framework for debiasing techniques in the CAI context, adapt existing privacy debiasing techniques to conversational AI systems, and assign them to the relevant actors of the ecosystem. Therefore, in this position paper, we first provide an overview of actors of the CAI ecosystem in Section 4.1. While an extensive overview of biases and heuristics is out of the scope, we introduce the main sources of poor privacy and security decision-making in Section 4.2. We then establish a novel categorization framework, cluster debiasing strategies, and discuss their adaptation to CAI in Section 4.2.1. Based on our differentiation of actors in the conversational AI ecosystem, we introduce complementary strategies that can benefit various actors in the CAI ecosystem. Lastly, we discuss limitations and future research challenges in Section 4.2.3.

2 Conversational AI Ecosystem and Actors

Conversational AI can encompass a multitude of systems as it generally refers to technologies that allow natural interactions between machines and humans by leveraging AI-enabled speech and text processing (McTear, 2021). Thereby, conversational AI refers to text-based as well as voice-enabled systems. While voice-enabled systems require access to microphones and speakers, text-based systems rely on graphical user interfaces to allow input and output of messages. Moreover, multimodal applications are possible, e.g., voice assistants may be accompanied by a screen which allows displaying of complementary information. Both text as well as voice-based systems can be deployed on different physical instances such as on smart speakers, computers, or smart devices. In this position paper, we adopt a broader perspective by focusing on CAI systems as a whole considering their widespread application, multimodality, and growing relevance (e.g., OpenAI (2023)). All CAI systems, text-, and voice-based as well as multimodal systems share their conversational nature, i.e., users interact in natural language with them. By developing a framework for this broader scope that can be applied to all CAI systems, we hope to address challenges of debiasing human decision-making in CAI comprehensively. While it is possible that different types of CAI systems (e.g., text- vs. voice-based) may differ in the strategies that work best with them, a more general framework can be applied nonetheless, e.g., as guide to systematically test and compare strategies.

In their guidelines on virtual voice assistants (VVA), the European Data Protection Board (EDPB) (2021) identified relevant actors in the ecosystem. They differentiate between the VVA provider or designer, application developer, integrator, owner, and user. Figure 1 shows different actors and their individual tasks. We believe that their differentiation provides a suitable starting point to address relevant actors in the CAI ecosystem as the actor’s task description is also applicable to the development of text-based CAI systems. Moreover, the number of defined roles is manageable and suitable for addressing the roles’ hindrances in decision-making and recommending suitable debiasing strategies. Finally, the roles are not closely connected to legal definitions, e.g., data collector or processor under GDPR. Instead, due to the complexity of the CAI ecosystem, actors can take on different responsibilities or share data controlling (European Data Protection Board (EDPB), 2021; van Mil & Quintais, 2022). As a legal assessment is out of scope of this article, we will rely only on the actors’ tasks as shown in Fig. 1 to map suitable debiasing strategies. In general, debiasing strategies can be applied independently of rights and obligations under legal regulations contributing to the overall principle of privacy by design and human’s right to privacy (Cavoukian, 2009; United Nations, 1948). Moreover, individuals and organizations might not only inhabit one role but multiple roles. This is emphasized by Fig. 1 as it shows the close relationship between the provider, integrator, and application developer to successfully create a CAI system. For instance, businesses can act as providers and integrators by developing the system’s main functionalities and manufacturing necessary components, e.g., a smart speaker. In addition, businesses can provide platforms that allow application developers to enhance the default functionality of the system, e.g., Google Actions or Alexa Skills. Even if inhabiting more than one role, people and organisations can profit from an overview that shows which debiasing strategies can be applied to these individual roles.

Fig. 1
figure 1

Differentiation of actors in the CAI ecosystem as described by (European Data Protection Board (EDPB), 2021)

Importantly, the owner role can differ from the user role as CAIs are established in accommodation places or professional environments such as office spaces or schools. For example, when a system is deployed at a workplace, the company may be the owner of the system while its users are employees. Similarly, when the system is used in a school, the school may be the owner while teachers and students are its users. However, teachers may also inherit the role of an owner as they might be in charge of a single classroom where the conversational AI system is provided to students.

So far, nudges for privacy and security decision-making have largely focused on the relationship between online services and their end-users (Ioannou et al., 2021; Kitkowska et al., 2020). However, given the complexity of conversational AI ecosystems and the different actors involved in creating these experiences, we believe that nudges and debiasing strategies in general can provide helpful tools for fostering privacy and security solutions throughout the ecosystem.

3 Sources of Poor Privacy and Security Decision-Making

Different actors might suffer from varying hindrances and biases to make desirable privacy and security decisions. While insufficient access to information can be one of the major factors for users to engage in adverse privacy decisions, it might be less of a concern for application developers. They are involved in data usage, processing, and storage and therefore have access to more information than users. While a complete analysis of the varying hindrances of optimal decision-making for the individual stakeholders is out of scope, we identify some major hurdles in the following sections. This will allow us to recommend and suggest potentially helpful interventions for the different groups.

3.1 Information Asymmetry

In the field of privacy and security, it is common for users to be subject to incomplete or asymmetric information as data collectors usually have more or better information about the collection, processing, and storage of data (Acquisti et al., 2020). Therefore, users may be unable to make informed decisions about their privacy, e.g., whether to disclose information to a CAI system. Informed decision-making can be particularly difficult as users might be unaware of the existence of different actors and their data collection and processing, e.g., third parties (Abdi et al., 2019). Moreover, due to the power imbalance, users can become subject to persuasive conversations with possible consequences to their privacy and security (Murtarelli et al., 2021).

Professionals involved in building or establishing CAI systems form a large group (European Data Protection Board (EDPB), 2021). They can include owners, integrators, application developers, and providers—and all of them may be more knowledgeable than users about CAI systems’ capabilities. However, technical expertise may vary within a role. For example, people with different backgrounds may be owners of CAI systems. While a teacher may have little technical background and may experience similar power imbalances as users, a manager with engineering background who decides to roll out CAI in the workplace may have a better understanding of the CAI ecosystem and its accompanied privacy and security risks. While app developers can be considered data collectors or processors when accessing or requesting certain personal attributes from the users, they might have limited knowledge about data processing done by conversational AI system providers. For instance, Amazon Alexa voice recordings are not shared with skill developers (Amazon Inc., 2019). In contrast, providers can have incomplete information about how data is handled by the application developers and whether personal attributes are directly requested (Lentzsch et al., 2021). This suggests that information asymmetry matters for all actors in the conversational AI ecosystem even though its impact might differ.

3.2 Heuristics and Biases

Privacy and security decision-making is often subject to uncertainty as long-term risks can be unknown or difficult to grasp and decisions are constrained by time and available information (Acquisti et al., 2015; Leschanowsky et al., 2021). Moreover, assessing the probability of possible malicious events and privacy breaches can be effortful and exhausting. The fact that human decision-making is subject to limited resources was first discussed by Simon (1990) under the concept of “bounded rationality.” He pointed out that decision-makers may rely on heuristics or mental shortcuts to simplify the process. Later, Kahneman (2011) extended this idea by developing the dual-process model of cognition—the distinction between system 1 and system 2. System 1 refers to intuitive, fast, and effortless thinking which can, however, result in biased and suboptimal choices. In contrast, system 2 describes a slower, more conscious, and controlled thinking process which is likely to be beneficial for making complex privacy choices.

For an extensive overview of heuristics and biases that can impact privacy decisions, we refer readers to Acquisti et al., (2018, 2020). While research on cognitive biases that impact privacy decision-making has merely focused on online environments, a majority of biases is applicable to the context of conversational AI systems and their actors. For example, anthropomorphism, i.e., the perceived level of human-like characteristics, and other salient cues can influence users’ intentions, disclosure, and privacy concerns (Cai et al., 2022; Ha et al., 2021; Ischen et al., 2020). Thereby, self-disclosure of CAI systems could effectively encourage users to share personal information (Rao et al., 2022). Moreover, mass media portrays and brand loyalty can influence privacy perceptions and usage of CAI systems (Maroufkhani et al., 2022; Sin et al., 2021). As a detailed analysis of specific biases of CAI systems is out of scope, we will rely on biases that are known to influence privacy and security decision-making for the remainder of this paper. Nevertheless, we emphasize the need to further investigate cognitive and behavioral biases that are specific or amplified in conversational AI systems, e.g., machine heuristic (Sundar & Kim, 2019).

3.3 Decision Readiness

Lastly, we want to focus on decision readiness that can negatively influence privacy decision-making in CAI. Decision readiness refers to the fact that system 2—the slow and more controlled thinking—is ready to monitor and if necessary intervene in intuitive thinking (Soll et al., 2015). However, this capability can be impeded by factors like fatigue, distraction, visceral influences, and individual differences. As CAI systems may be particularly useful when hands-free interaction is necessary or multiple tasks are carried out, e.g., in a car, factors like distraction constitute a crucial source of biased decision-making. As previously discussed, visceral influences might include visual or auditory cues of conversational AI systems which can lead to increased self-disclosure or suboptimal privacy decisions (Ischen et al., 2020). Lastly, individual differences such as differences in training, cognitive ability, or self-reflection can impact privacy decision-making. Therefore, debiasing strategies might show different levels of effectiveness depending on an individual’s characteristics and role in the CAI ecosystem.

4 Categorizations of Debiasing Strategies and Their Limitations

An increased understanding of cognitive biases has spurred the development of bias mitigation strategies across various domains, including healthcare, finance and privacy, and security. Additionally, efforts were taken to categorize individual debiasing strategies into high-level classes, although categorizations differ within and between disciplines. For instance, in the medical field, debiasing strategies have been grouped into cognitive, technological, motivational, and affective strategies (Broussard & Wulfert, 2019; Larrick, 2004; Ludolph & Schulz, 2018). In contrast, other studies have distinguished between educational and workplace strategies as well as forcing functions (Croskerry et al., 2013; Lambe et al., 2016; Neal & Brodsky, 2016). The former categorization is based on assumptions of different strategies on how to approximate ideal decision-making outcomes (Larrick, 2004), while the latter distinguishes based on the temporal appearance of a debiasing effect with educational strategies influencing future decision-making and workplace strategies helping to overcome bias at the time of decision-making without necessarily changing the individual (Croskerry et al., 2013). Interestingly, medical research has predominantly focused on strategies that modify the person, e.g., through cognitive training (Lambe et al., 2016; Ludolph & Schulz, 2018), whereas research on debiasing in the privacy and security domain has primarily concentrated on the implementation and evaluation of nudges, e.g., nudging with information and presentation, defaults, or incentives (Acquisti et al., 2018; Ioannou et al., 2021; Kitkowska et al., 2020). In general, the term “nudge” stems from behavioral economics and describes “any aspect of the choice architecture that alters people’s behavior predictably without forbidding any options or significantly changing their economic incentives” (Thaler & Sunstein, 2021). It can be used by “choice architects” to influence decision-making by modifying the environment (Thaler & Sunstein, 2021). The term “nudges” can also be utilized as an acronym to cluster different interventions, i.e., iNcentives, Understand mappings, Defaults, Give feedback, Expect errors, Saliency (Acquisti et al., 2018). However, due to the nature of CAI systems, their seamless way of interacting with users via natural language, and their complex ecosystems, there is a need to move beyond the application of nudges for privacy and security decision-making, to draw from other disciplines, and to establish a more comprehensive categorization framework. While a universally accepted taxonomy for debiasing strategies is yet to be established, Soll et al. (2015) proposed a more broadly applicable categorization distinguishing strategies that modify the person to those that modify the environment. Their framework has been applied in diverse research fields such as geoscience education and management decision-making (Muntwiler, 2021; Wilson et al., 2019). In particular, Muntwiler (2021) illustrate how debiasing strategies can be theoretically grouped into a two-level categorization framework following Larrick (2004) and Soll et al. (2015) with modifications to the person and the environment building higher-level categories.

4.1 Towards a Categorization Framework for Debiasing Strategies for Conversational AI Systems

Given previous research on classification logics, we identify two high-level categorization frameworks to cluster debiasing strategies for CAI systems, i.e., categorization depending on the temporal appearance of the debiasing effect (Croskerry et al., 2013) and categorization depending on the type of modification (Soll et al., 2015). Due to their simplicity and clarity, they can help to navigate the landscape of debiasing strategies for CAI and provide a starting point for more detailed and nuanced frameworks. While each framework individually offers valuable insights, we recognize their complementary perspectives and their potential to capture the underlying principles and patterns of debiasing strategies for CAI. Being derived from the medical field with a strong focus on cognitive interventions, the distinction between educational and workplace strategies might overlook the impact of environmental modifications on future or real-time decision-making. Likewise, distinguishing only between modifications to the individual and the environment does not adequately address the benefits of CAI systems in seamlessly interacting with users and their potential to debias human privacy and security decisions. For instance, CAI systems can proactively function as guides, mentors, or teachers on privacy, security, and protective mechanisms and thereby modifying both the environment and the individual. Combining these two frameworks enables a comprehensive approach, integrating complementary perspectives while maintaining clarity and simplicity in the classification of debiasing strategies for CAI systems.

We propose a two-dimensional categorization framework for debiasing strategies in the context of privacy and security for CAI. Thereby, Fig. 2 facilitates a better understanding of the relationships between the frameworks, their compatibility, and usefulness for consolidating approaches from various disciplines. By establishing a graph visualization, we do not consider debiasing classes to be mutually exclusive. Thus, we acknowledge that certain debiasing strategies, i.e., cognitive strategies, may lie in the area between modifications to the person and the environment as well as affecting peoples’ decision-making now and in the future.

Fig. 2
figure 2

Schematic representation of the categorization framework for debiasing strategies for conversational AI. We rely on previous categorization logics and research to define the placement of classes (Acquisti et al., 2018; Croskerry et al., 2013; Soll et al., 2015). By using dotted lines, we emphasize that the expansion of the classes is based on our assessment and might vary depending on the debiasing strategies considered for a certain context

For the remainder of this paper, we will particularly focus on five high-level classes, i.e., education, incentives, cognitive strategies, assisted decision-making, defaults, and nudging with information and presentation. While previous work has introduced the class of automated decision-making or technological strategies (Larrick, 2004; Soll et al., 2015), we will refer to this class as assisted decision-making to emphasize the role of humans in the decision-making process. The classes were chosen as they are reoccuringly discussed in previous research in various fields (Acquisti et al., 2018; Croskerry et al., 2013; Soll et al., 2015) and their benefits of complementing each other. For example, while education focuses on modifying the person and their future decision-making, information and presentation provision refers to a change in the environment that can influence people at the time of decision-making. On the other hand, incentives can be considered modifications to the environment with mid-term to long-term influence on peoples’ decision-making, while assisted decision-making such as privacy assistants modify the person and their decision in a specific moment. Our framework emphasizes that the chosen strategies need to be seen as complementary to each other, as all these interventions come with strengths and weaknesses.

4.2 Debiasing Strategies

We will now extend on the previously introduced classes by presenting corresponding debiasing strategies adaptable to CAI and their limitations. For each class, we provide an overview of discussed strategies through the use of tables (see Tables 1, 2, 3, 4, 5, and 6). Whenever applicable, we point to references that discuss the proposed strategies in the context of CAI or related technologies, e.g., app development, and show for which actors they have been applied. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas. Thereby, we aim to draw attention to promising opportunities for further investigation.

Lastly, we want to emphasize that we present a first attempt of mapping existing debiasing strategies to CAI and that we do not provide a comprehensive overview of available debiasing strategies. Instead, we bring together strategies that have been applied for privacy and security decision-making in online environments and effective strategies from other disciplines. Thereby, we focus on expanding the range of possible debiasing strategies for CAI while keeping them applicable to the various actors of the ecosystem.

4.2.1 Education

Education for Debiasing Decision-Making

One way to improve decision-making is to educate individuals and increase their competencies in privacy and data protection over time. Studies have shown that a higher level of online privacy literacy can significantly increase the usage of protective strategies (Masur et al., 2017; Park, 2013). Thereby, most educational strategies aim at modifying the person and impacting their future decision-making capabilities. Privacy literacy can be divided into factual and procedural knowledge (Masur, 2019). While factual knowledge refers to expertise on certain technical or legal aspects regarding data protection and processing, procedural knowledge is concerned with the ability to use protective privacy strategies (Masur, 2019). While both aspects are essential for all actors in the conversational AI ecosystem, there might be certain priorities set for different stakeholders. For example, educational strategies for users might emphasize procedural knowledge while strategies that focus on providers and app developers need to stress knowledge of technical and legal aspects. Moreover, users might benefit the most from procedural knowledge that allows them to apply strategies for individual protection, while providers and app developers have to be knowledgeable about strategies for users’ privacy protection.

As privacy and security become an essential part of technological systems, the need for education and training is growing as well as the diversity of resources for privacy education. An interview study by Subramaniam et al. (2019) revealed that people rely on different sources to educate themselves on the topic of privacy. These include school and educational lessons and job training, experiences, and knowledge from family members and friends and so-called privacy experts, i.e., bank employees or librarians as well as popular organizations. Moreover, experiences with privacy issues and system-programmed privacy measures, e.g., regular prompts to change passwords, played a crucial role in engaging in protective behavior (Subramaniam et al., 2019). On the other hand, studies have shown that certain groups might suffer from lower levels of privacy literacy and may be less likely to apply protective strategies. For example, sociodemographic factors, e.g., income, education, age, and gender, can influence the level of privacy literacy (Park, 2013). Children form a particularly vulnerable group as they are not yet literate in privacy risks but are exposed to CAI systems through children’s toys and smart home technologies (Mann et al., 2022). Children’s usage of IoT devices and smart speakers and their protection largely depends on their guardians who come with significant differences in their level of privacy literacy and concerns, e.g., due to gender, racial, and socioeconomic differences (Garg & Sengupta, 2019; Mann et al., 2022). Nevertheless, when it comes to voice recordings, people are in general largely unaware of inferences that can be drawn from their voices (Kröger et al., 2022). This shows a need for easily useable and adaptive educational programs to raise awareness of privacy risks of CAI systems for various user groups.

Educational Debiasing Strategies in CAI

While educational strategies have traditionally focused on modifying the person through training courses, tutoring, or similar interventions (Croskerry et al., 2013), due to their human-like capabilities, CAI systems can proactively trigger educational interventions taking into account individual differences and context. Proactive educational approaches can be purposefully designed and therefore constitute not only to modifications to the person but also to the environment. Thereby, CAI systems can act as guides and mentors to people to raise awareness and promote privacy literacy (Leschanowsky et al., 2022). In previous work, we showed how the concept of Guided Reflection—a strategy that has been successfully applied in the medical context to increase diagnostic accuracy through mentoring and self-reflection—can be leveraged by CAI systems (Leschanowsky et al., 2022). In fact, conversational agents have been successfully used to support workers in their self-reflection and self-learning, e.g., by prompting workers to journal activities (Kocielnik et al., 2018). As children form a particularly vulnerable group, providing suitable education and mentoring on privacy aspects through CAI can significantly increase the young generation’s privacy literacy. Thereby, design recommendations on learning applications for children in CAI (Garg & Sengupta, 2020) can inform the development of educational privacy tools (Table 1).

Table 1 Overview of debiasing strategies for the class of “education” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors’ columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

While most of the discussion above focused on user’s privacy literacy and its limitations, boosting provider's, app developer's, and owner's privacy literacy (see Fig. 1) is key to creating private and secure CAI systems. However, recent studies found that privacy is not considered thoughtfully throughout development. Edu et al. (2022) investigated voice applications on the Alexa Marketplace and found that 36% of skills follow bad privacy practices such as broken traceability, i.e., the privacy policy does not cover data practices. Moreover, Liao et al. (2020) could show that current privacy policies of voice applications are often non-existent, incorrect, or inaccessible.

To counter bad practices among mobile app developers, Hatamian (2020) designed a Technical Guidelines Catalog by mapping legal principles of the General Data Protection Regulation (GDPR) to technical privacy and security solutions. Such a catalog could also assist CAI application developers in integrating privacy during application design and development. Moreover, Privacy-by-Design offers a more extensive approach to integrate privacy into a system throughout the entire development lifecycle (Cavoukian, 2009). Therefore, privacy patterns provide concrete tools for common privacy problems to ensure privacy-friendly systems (UC Berkeley School of Information, 2019). Technical privacy and security solutions derived from legal principles may overlap with privacy patterns, but privacy patterns can be more diverse and do not necessarily map to certain legal requirements. While privacy patterns can benefit providers, developers, and integrators of conversational AI systems, pattern catalogs often lack consistency and are scattered among various platforms (Colesky et al., 2018). Moreover, while many of these patterns are applicable in the context of conversational AI systems, patterns matching the unique context of CAI are missing. For instance, human–computer interaction (HCI) privacy patterns rely mostly on visual cues and graphical interfaces (UC Berkeley School of Information, 2019). Yet, work by Murad et al. (2021) shows how grounding design guidelines for CAI on existing GUI heuristics can benefit adoption and how to transition between them. Future research could explore similar approaches for privacy patterns.

Limitations

While education is a necessary and undeniable strategy to support people in making better privacy decisions, it is insufficient to fully mitigate biases. One reason is that it remains unclear how much of a difference privacy literacy makes and whether people can apply their skills in concrete situations. Fernandes et al. (2014) found that the efficacy of financial literacy training is modest while training effects were larger for students who were trained over longer periods. However, as the acquired financial literacy declined over time, the authors conclude that the most effective form of training is education that is provided at times when needed. Similarly, users who had additional training on privacy literacy might not experience a long-lasting effect. Therefore, complementary debiasing strategies that modify the environment by informing people at the time of decision-making are crucial and will be discussed in detail in Section 4.2.6.

While most of the proposed educational methods have focused on users or software engineers, little is known about educational methods for owners of conversational AI systems. While users interact with the system on a one-to-one basis, an owner is in charge of the physical space and does not necessarily need to interact with the system directly. Nevertheless, the owner needs to be aware of possible interconnectivity of the system and its data accessibility. For instance, when a system is deployed in an accommodation space, safeguards need to be taken such that guests cannot access data that has been provided by former guests. Moreover, in educational settings, teachers might take on an owner role if they are in charge of a classroom where the system is deployed. As children are considered a vulnerable group and their recordings might be sensitive, teachers need to undergo specific training to ensure that appropriate security and privacy measures are taken and the systems are used appropriately (Terzopoulos & Satratzemi, 2020). As these systems become widespread in accommodation, professional, and educational places, educating owners on privacy and data protection is crucial to ensure users’ privacy and their acceptance of the technology.

Takeaway

Educational strategies can support all actors in the CAI ecosystem. CAI systems can leverage guided reflection to make users reflect and learn about their privacy and security decisions. Providers, application developers, and integrators can benefit from technical guideline catalogues and privacy patterns tailored for conversational AI systems. Lastly, there is a need to explore educational methods for owners of CAI systems.

4.2.2 Cognitive Strategies

Cognitive Strategies for Debiasing Decision-Making

While education is concerned with boosting privacy literacy over time, cognitive strategies aim to impact people’s cognitive abilities at the time of decision-making. Yet, cognitive strategies have also been used in the medical field to enhance decision-making over time and have the potential to create long-lasting effects on improved decision-making (Croskerry, 2003). While a variety of cognitive strategies has been tested in various fields, e.g., medical field, AI overreliance, privacy, and security (Bucinca et al., 2021; Croskerry, 2003; Wang et al., 2013), they are yet to be investigated in the field of CAI. Therefore, in this position paper, we focus on cognitive strategies that have been frequently discussed in previous research on debiasing (Croskerry et al., 2013; Larrick, 2004; Soll et al., 2015). Cognitive strategies can ask people to identify situations in which decision-making errors are likely to occur and deliberately apply strategies to avoid decision errors (Croskerry, 2003). Others aim at directly inducing reflection and asking individuals to engage their system 2 thinking capabilities through interruptions or specific ways of presenting choices. While these strategies can be triggered by a CAI system and therefore constitute to modifications to the environment, they are likely to modify the person by altering their thinking process (see Fig. 2 for the classification of cognitive strategies into the framework).

Cognitive Debiasing Strategies in CAI

Generating alternatives and evaluating them based on established decision criteria is crucial for making rational decisions. However, due to cognitive biases, people are unlikely to engage in rational thinking and might have difficulties in generating alternatives (Soll et al., 2015). “Consider the Opposite” can be seen as a related strategy and has been proven helpful in clinical AI support (Bach et al., 2023). Due to their unique possibility of interacting with users naturally, CAI could support them in generating alternatives based on their own decision objectives. As privacy objectives can be highly subjective and dependent on people’s attitudes and values, system designers and developers might have difficulties in sensibly curating alternatives for the users. Therefore, CAI systems can foster a rational decision-making process by having users list their decision criteria, e.g., privacy concerns or interest in using the service, and weigh them according to their importance. Moreover, CAI could assist users in generating alternatives and finding their optimal choice. While research has shown that generating alternatives is most problematic and difficult for humans (Nutt, 2004), CAI systems might be capable of generating alternatives in a fast, effective, and comprehensive way. This can make it easy for users to choose among a few alternatives that have been found to fit their decision criteria best. Engaging in a rational decision process can be especially useful in situations where decisions are complex and only need to be made once, e.g., deciding whether to use voice authentication or when setting up the system for the first time.

In addition, cognitive forcing strategies can support people in their decision-making process and have been applied in medical research and research on the overreliance of AI (Bucinca et al., 2021; Croskerry, 2003). They have been described as a “specific debiasing technique that introduces self-monitoring of decisionmaking [sic!]” (Croskerry et al., 2013). We previously adapted cognitive forcing strategies to CAI to make people consider alternatives and reconsider disclosure (Leschanowsky et al., 2022). Depending on the number of possible alternatives, the consideration or generation of alternatives could result in an active choice condition (Table 2).

Table 2 Overview of debiasing strategies for the class of “cognitive strategies” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors’ columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

Active choice can help to induce reflection, to avoid mindless acceptance of default options, and to overcome decision avoidance (Keller et al., 2011). Choice architects might have difficulties in coming up with sensible default options as privacy preferences are subjective and heterogeneous. Choice paradigms have been predominantly explored for graphical user interfaces, particularly in the context of cookie consent notices, leaving the need to apply these insights to CAI systems (Habib et al., 2022; Utz et al., 2019). As long as choices are simple, e.g., asking users whether they like to have their data deleted or stored in a certain use case, alternatives can be presented directly to the users and CAI systems can require users to actively choose among them. However, active choice imposes a high cognitive load on individuals and therefore should not be used excessively but applied sensibly (Thaler & Sunstein, 2021). Similar to generating alternatives and cognitive forcing, active choice can be seen as a tool that can effectively support users’ decision-making. Other actors in the CAI ecosystem are less likely to benefit from active choice as privacy requirements play a more crucial role than their individual privacy preferences.

As users’ privacy decisions may suffer from an underestimation of risks, instructions that make people think of opposite outcomes as initially expected and prospective hindsight can counter optimistic privacy choices. By utilizing prospective hindsight, people are asked to imagine their future selves and to experience bad outcomes of their earlier judgements (Mitchell et al., 1989). For example, a conversational AI system might proactively ask users to imagine their future selves 2 years from now and to question why their personal information has been shared with company X and used for profiling. Such a strategy can prevent people from being overly optimistic that privacy breaches will not affect them and therefore trigger the usage of privacy protective strategies. Moreover, prospective hindsight can be triggered by dialog editors that may be used by providers, developers, and integrators to create CAI. This could make them consider the impact of privacy breaches on their reputation, revenue, and employment and help to design CAI with protective strategies in mind. Lastly, owners can similarly benefit by regularly utilising prospective hindsight.

To increase the accuracy of the judgement, multiple judgements by others or the same person at different times or with mental focus can be beneficial (Herzog & Hertwig, 2009; Larrick, 2004). As decisions are based on only a subset of accessible information, subsets can vary once people are asked to rely on different decision strategies, e.g., making one intuitive and one thoughtful decision (Lambe et al., 2016). In the privacy context, the accuracy of judgements could relate to how well users’ decisions match their attitudes and values. Consequently, users’ regrets and frustration about privacy decisions might decrease while their overall satisfaction might increase. Therefore, users’ satisfaction with their privacy decision-making could benefit from offering them the option to reconsider their decision or asking them to decide twice following different instructions—one based on intuitive thinking and one based on in-depth and analytical thinking (Leschanowsky et al., 2022).

Similarly, planned interruptions or forced breaks provide another way to introduce reflection and increase the accuracy of judgements. Similar techniques such as diagnostic time-outs or slowing-down decision-making have been successfully applied in the medical field (Lambe et al., 2016). Moreover, Wang et al. (2013) investigated a timer nudge as one of several privacy nudges on social media. The timer nudge would delay the Facebook post and allow users to reflect and possibly cancel their actions. Overall, their timer nudge was perceived positively as it provided the chance to correct typos, post better quality content, or cancel unnecessary posts. In a recent chatbot experiment, we investigated the impact of a timer nudge on users’ behavior but found that the additional delay did not significantly impact their decisionmaking (Leschanowsky et al., 2023). While the timer nudge did not negatively affect the usability in our study, conversation designers might be unlikely to apply forced breaks as they generate friction and let the conversation appear less natural. However, planned interruptions can support providers, integrators, and application developers to reflect on their usage of users’ personal information and reduce unnecessary permission requests.

Finally, planning prompts ask people to specify “when, where, and how” a goal is achieved (Wust & Beck, 2018). These concrete plans help to translate goals into actions and become a commitment which individuals are less likely to break. Therefore, planning prompts provide a simple and effective nudge for goal achievement. Cuadra et al. (2021) explored planning prompts for virtual voice assistants and found that their voice application was perceived as helpful and improved planning behavior. In the privacy context, planning prompts could support both developers as well as users of conversational AI systems. System providers, developers, integrators, and owners could use planning prompts to make specific plans for incorporating privacy into their system and application or to delete unused and old data. Moreover, users can benefit from conversational AI systems that proactively encourage them to make plans for checking their privacy settings.

Limitations

A variety of cognitive strategies could support actors in the CAI ecosystem to make better decisions about their privacy. However, based on our assessment, not every cognitive strategy is suitable for all actors. Future research should investigate various cognitive strategies and their suitability to different actors in the CAI system. While only few cognitive strategies have been applied and evaluated in privacy scenarios (Wang et al., 2013), most strategies are yet to be investigated in the context of CAI. Therefore, evaluation measures are needed to test the strategies’ effectiveness on people’s decision-making. While medical research assesses their effectiveness by evaluating error rates in diagnostic reasoning (Lambe et al., 2016), error rates are not easily accessible in the privacy context. In particular, when investigating cognitive strategies for users of CAI systems who come with highly subjective privacy preferences and attitudes, the optimal outcome of a privacy decision usually remains unknown. We will further discuss the need for a comprehensive evaluation in Section 4.2.3.

Takeaway

Cognitive strategies aim to mitigate bias in human judgements at the time of decision-making and to trigger a more rational thinking process. CAI systems can assist users in generating alternative choices or make them consider alternatives and reconsider decisions by applying cognitive forcing strategies. As long as choices are easily understandable, they should be directly presented through active choice mechanisms. Prospective hindsight and planning prompts are two of the cognitive strategies that can support all actors in the CAI ecosystem to improve their judgements. Instead, planned interruptions might be most suitable to providers, developers, and integrators as they can cause considerable friction to the dialog.

4.2.3 Assisted Decision-Making

Assisted Decision-Making for Debiasing Decision-Making

Another way to prevent people from biased decision-making is to replace human judgements with automated or assisted decisions, e.g., by applying linear models or using decision support systems (Larrick, 2004). While these models can incorporate human judgement and subjective ratings, they mostly rely on historical data and objective ratings. However, as mentioned earlier, an individual’s privacy preferences are highly subjective and contextdependent (Nissenbaum, 2010). Therefore, the selection of suitable attributes that need to be included in a predictive model is challenging and can again be prone to cognitive biases (Soll et al., 2015).

Assisted Decision-Making in CAI

In the privacy context, privacy assistants have been investigated which can offer varying levels of automation. While some privacy assistants only inform users and ask them to make decisions, others automatically decide for the users (Colnago et al., 2020) (Table 3).

Table 3 Overview of debiasing strategies for the class of “assisted decision-making” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors; columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

In addition, checklists can be seen as a tool for assisted decision-making as they provide an efficient, systematic, and consistent way of carrying out tasks (Gawande, 2009). They are especially useful in situations of low decision readiness or when certain tasks are likely to be overlooked and left out (Gawande, 2009). In previous work, we showed how checklists can be adapted to privacy in conversational AI systems, e.g., by setting up a privacy checklist and confirming user-specific privacy requirements before installing a new application (Leschanowsky et al., 2022). Moreover, checklists can provide a helpful tool for all actors in the conversational AI ecosystem. They can ensure that app developers follow certain steps that are necessary to protect users’ privacy, e.g., checking whether the data asked for is truly relevant. Similarly, providers can benefit from a privacy-related checklist to ensure that data flows on the platform are appropriate and purposeful. For example, the non-profit organization Open Voice Network has realized the need for privacy checklists and released ethical guidelines that can be interpreted as privacy checklists for voice interfaces (Open Voice Network, 2022, 2023). Lastly, checklists can be distributed to owners to ensure that privacy guidelines are followed when setting up the devices.

Limitations

In their study on privacy assistants for IoT, Colnago et al. (2020) found little consensus among users regarding the level of automation and possible control options. Therefore, allowing users to adjust and configure privacy assistants to their needs is highly recommended. Yet, this raises questions on the effectiveness of privacy assistants as a debiasing strategy as it would require additional mechanisms to make users engage with the tool and its control options. Moreover, models that learn peoples’ privacy preferences based on historic data can present a privacy threat themselves and need to be implemented in secure and privacy-friendly ways.

Takeaway

Assisted decision-making replaces human judgements altogether and can therefore prevent biased decision-making. Users can benefit from privacy assistants deployed on conversational AI systems, but because of varying preferences, suitable control options are indispensable. In addition, checklists can be considered a tool for assisted decision-making and can support all actors in the ecosystem to make efficient and consistent choices.

4.2.4 Incentives

Incentives for Debiasing Decision-Making

Monetary as well as non-financial incentives such as badges or peer pressure have been proven beneficial for people to make better decisions (Acquisti et al., 2018; Lindbeck, 1997). Thereby, incentives can serve as motivators to transition between system 1 and system 2, i.e., between fast and slow thinking, and can be especially useful when undesired choices stem from insufficient attention or a lack of effort (Larrick, 2004). Privacy costs are often difficult to assess as they require an estimate of long-term consequences. Therefore, individuals may have a clear understanding of benefits while the costs remain elusive and hard to grasp (Leschanowsky et al., 2021). Thus, providing the right incentives can make costs understandable and help people in considering long-term consequences. Incentives can be either rewarding, e.g., rewarding individuals for privacy-preserving decisions or considering privacy costs, or punishing, e.g., increasing the costs to choose non-privacy-preserving options or disclosing information about costs and negative consequences of insecure behavior (Lindbeck, 1997). Incentives always present modifications to the environment (see Fig. 2 for their classification into our framework). Yet, while incentives such as organizational measures or regulations focus on influencing decision-making in the future, incentives that are directly rewarding or punishing can also influence peoples’ choices at the time of decision-making.

Incentives in CAI

System providers, developers, integrators, and owners can profit largely by introducing the right incentives themselves as well as by being exposed to them. For instance, virtual badges or app reviews can act as strong incentives for app developers to offer privacy-preserving applications (Acquisti et al., 2018). Similarly, badges and reviews can motivate owners of CAI systems in accommodation places to protect users’ privacy and security. Moreover, being accountable and liable for security failures, e.g., having to pay increased fines or to pass additional training, has been shown to create a moral hazard (Acquisti et al., 2018; Anderson, 2001). Being held accountable increases the cost of failure and consequently the effort of making a desired decision (Larrick, 2004; Lerner & Tetlock, 1999) (Table 4).

Table 4 Overview of debiasing strategies for the class of “incentives” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors’ columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

Herath and Rao (2009) investigated factors that influence employees’ intentions to comply with security policies. They found that intrinsic, i.e., perceived effectiveness, and extrinsic motivators, i.e., penalties and social pressure, influence employee behavior. However, while the certainty that possible security breaches are detected positively influenced employees’ behavior, the severity of penalties did not. Thus, CAI system providers and integrators should make use of efficient and visible organizational measures to detect privacy breaches without needing to severely punish detected breaches. This might add to a positive error culture. Moreover, Herath and Rao (2009) found that social pressure and normative beliefs can strongly impact security behavior. This is particularly interesting from the privacy perspective as privacy breaches are often a result of inappropriate internal information flows rather than security issues. Therefore, system providers and integrators should emphasize privacy practices and expectations throughout the company and communicate to employees the importance of their individual privacy practices.

However, we need to be clear that market forces alone are not sufficient for system providers to push for privacy-preserving solutions and innovations (Stucke & Ezrachi, 2017). Therefore, additional incentives such as regulations need to be set by legislators and policymakers to ensure that privacy measures are integrated. Sætra (2020) shows that if privacy is seen as an “aggregate public good,” governmental interventions and regulations are beneficial and necessary. In addition to regulations and incentives based on penalties for those who fail to comply, policymakers can include rewarding incentives in their portfolio. Kosseff (2016) refers to “positive cybersecurity law” where companies are encouraged to protect themselves from cybersecurity attacks. While they focus solely on cybersecurity protection, policies like a “safe harbor from data security lawsuits” or tax incentives could be extended towards privacy.

Limitations

One limitation of incentives is characterized by the nature of incentives themselves. As most of them focus on enhancing decision-making in the future by introducing new regulations or measures, their adoption takes time and might need to be accentuated by additional guidelines, trainings, and similar interventions (Dalela et al., 2022).

Takeaway

Incentives can be either rewarding or punishing and can be of a financial as well as non-financial nature. Actors designing, implementing, and deploying CAI systems can benefit from virtual badges, reviews, organizational measures, and regulations. These incentives can be either created by themselves or by a higher-level body, e.g., by policymakers.

4.2.5 Defaults

Defaults for Debiasing Decision-Making

As humans often stick to default options due to the status quo bias and the difficulty in overcoming inertia, defaults are powerful tools for “choice architects” (Thaler & Sunstein, 2021). It has been shown that defaults can have a significant impact on people’s decision-making in fields such as retirement savings, food consumption, and health care (Acquisti et al., 2018; Thaler & Sunstein, 2021). Due to their power and robustness, default options are particularly important to ensure privacy in conversational AI systems.

Defaults in CAI

In fact, researchers have argued for implementations of privacy-friendly default settings (Table 5), e.g., storage of voice commands and their usage for the system’s improvement should be disabled by default (Hernández Acosta & Reinhardt, 2022; Lau et al., 2018). However, current conversational AI systems rarely follow these recommendations and come with varying default settings. For instance, Amazon Alexa default options include an unlimited retention period of voice recordings while Apple’s Siri does not retain audio recordings by default (Amazon Inc., 2022a; Apple Inc., 2022).

Table 5 Overview of debiasing strategies for the class of “defaults” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors’ columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

So far, we have only touched on privacy defaults for users of conversational AI systems, but similar privacy-friendly defaults need to be investigated for system providers, developers, integrators, and owners. For example, Amazon does not share any voice recordings with third-party skill developers, and skills can be configured to request permissions (Amazon Inc., 2022b). However, Lentzsch et al. (2021) found that instead of making permission requests through the API, skills can access users’ personal information by asking them directly in a conversation. While making the permission request the default way to access personal information might seem sufficient from a technical point of view, practical implementations prove it wrong. This urges the need for system providers to re-design tools that are used for building CAI and to make privacy-friendly defaults more sticky and less likely to be circumvented by developers. For instance, dialog editors could detect whenever developers ask for personal information in their application and display a prompt that allows them with one click to request the desired information through the API. This would shorten dialogs and has the potential to improve user experience as well as data collection transparency.

Limitations

While default options can lead to more privacy-preserving systems, they might not serve all users equally well. Defaults should be deployed for individual and public welfare especially when they are set in place by policymakers. However, problems may arise as defaults that benefit a majority could be suboptimal for some people (Smith et al., 2013). Several studies on CAI systems have shown that while there is some agreement on certain privacy aspects, e.g., implementation of shorter retention periods for voice recordings (Malkin et al., 2019), people’s privacy preferences can differ largely (Lau et al., 2018). Therefore, more research is needed on people’s privacy preferences in conversational AI to curate sensible default options. In addition, studies in other fields have shown that if easily adjustable controls are offered, people overcome inertia and change defaults if they dislike the outcome (Thaler & Sunstein, 2021). Such controls should be investigated for conversational AI to accompany defaults.

Smith et al. (2013) discuss defaults as “hidden persuadors” and their potential to erode consumers’ autonomy. To counter this, they suggest the usage of “smart defaults” which are based on consumer information and adapted to optimally fit a specific consumer. While these smart defaults work well for a variety of contexts, e.g., “Advanced Air Bag System” (Smith et al., 2013), they cause problems in the context of privacy. First, they require access to personal information to adapt to individual preferences, a procedure that comes with varying privacy risks. Second, individuals’ valuations of privacy are inconsistent and sensitive to non-normative factors (Acquisti et al., 2013). By trading away privacy for convenience or economic benefits, privacy protection may be led by individual interest rather than social welfare. Therefore, smart defaults (and similar models such as privacy assistants as discussed in Section 4.2.3) can only provide suitable protection once a certain level of privacy is ensured by design. Lastly, defaults can significantly lose impact once companies and consumer interests are not aligned. Based on the example of tracking, Willis (2014) provide an extensive argument why defaults are likely to fail as long as companies can push back and leverage similar biases to make default options more or less sticky. Drawing on information-cost theory, Bar-Gill and Ben-Shahar (2021) show how current legal regulations, such as the GDPR and California Consumer Privacy Act (CCPA), attempt to make privacy-preserving defaults more sticky by asking for explicit consent while at the same time reducing the cost for people to become informed by requiring easily understandable notices. Lowering the information costs can thus result in more users acting upon their attitudes. However, it has yet to be investigated how privacy notices could be made easily understandable in CAI. As they are based on natural language, written notices may need to be translated into dialogues for text- and audio-based interactions.

Takeaway

Nudging with defaults can be used for all actors in the CAI ecosystem by curating sensible defaults for application settings and tools used to design, develop, and deploy these systems. To design acceptable defaults, preferences need to be understood and control options to easily change defaults need to be available. Control options should be available both in dialogs and graphical interfaces to make them easily accessible and more sticky. In addition, information costs and other biases have to be considered as they may influence the effectiveness of defaults.

4.2.6 Nudging with Information and Presentation

Information and Presentation for Debiasing Decision-Making

While general education on privacy literacy can support users in overcoming their biases and making more informed decisions, nudges that inform users at the time of decision-making can additionally prove beneficial. Especially, in situations where decision readiness might be low, additional nudges that disclose information warn or remind people can lead to improved decision-making. Shaping information in a way that is intuitive to understand and evokes interest can encourage people to make better decisions (Acquisti et al., 2018; Thaler & Sunstein, 2021).

Nudging with Information and Presentation in CAI

Privacy and security labels are one option to support people during purchase or download of applications. For instance, Emami-Naeini et al. (2020) developed a two-layered privacy and security label for IoT devices based on expert and user studies. Along the same lines, Johansen et al. (2022) discuss privacy labels and their potential from a multidisciplinary perspective. They also show how privacy labels can not only have an educating effect on users but how these labels can benefit programmers in integrating privacy into their development.

Privacy labels disclose privacy and security information in a more understandable and easily readable format and can support people in their purchasing decisions. However, it might still be difficult to compare privacy labels across multiple products. In their book, Thaler and Sunstein (2021) favor so-called choice engines that can help consumers to decide between many alternatives. For instance, travel websites allow users to search among many different options based on their preference selection. Once disclosures about privacy and security attributes are machine-readable, choice engines can allow users to easily compare between varying products and filter for privacy and security options. First attempts to this can be seen by Mozilla (2022) who created a guide to help shopping secure products with the option to choose between varying categories, e.g., smart home and health care applications, and to filter for products where “privacy is and is not included.” Moreover, Tamò et al. (2021) propose a right to customization where companies are asked to offer multiple variants with different data processing options and trade-offs between privacy and utility. To support users in their decision-making, choice engines can play a crucial role in comparing applications and products across companies as well as within companies (Table 6).

Table 6 Overview of debiasing strategies for the class of “nudging with information and presentation” for ensuring privacy and security in conversational AI systems. Whenever applicable, we point to references that discuss these strategies in the context of CAI or related technologies. We show them in the actors’ columns that are directly addressed by those references or closely connected (see Section 2 for a detailed description of the actors). For greater clarity, we have restricted the choice of references to one and will discuss further examples in the associated sections. We deliberately leave cells blank where the proposed strategy has—to the best of our knowledge—not yet been studied in-depth for a specific actor in CAI or related ecosystems, emphasizing the need for future research in these areas

While presentation nudges for privacy and security have mostly focused on graphical user interfaces (Acquisti et al., 2018; Kitkowska et al., 2020), it is unclear how information and presentation nudges can be applied in conversational AI and in particular voice-enabled systems. For instance, Pearman et al. (2022) refined a consent flow for the US Health Insurance Portability and Accountability Act (HIPAA) authorization in a text-based chatbot. Their iterative process of redesign aimed to make the consent form shorter, clearer, and easier to understand. They found that while their redesigns improved understandability, it was not sufficient to ensure informed consent and recommend conversational privacy to tackle shortcomings. Harkous et al. (2016) proposed “Conversational Privacy Bots (PriBots)” which can present privacy policies and enable changing of privacy settings in natural language. Brüggemeier and Lalone (2022) explored conversational privacy in a chatbot by allowing users to control their data or ask for privacy-related information in natural language. They found differences in perception between an offer to delete data and an offer to delete sensitive data. Only the option to delete data was perceived as significantly more private and secure indicating the need to explore possible framing and priming effects and their impact on privacy decision-making. Moreover, while research on anthropomorphism has shown to significantly impact peoples’ perceptions of conversational AI systems (Cai et al., 2022; Ha et al., 2021; Ischen et al., 2020), it is unknown how changes to the visual appearance or voice can effect users in their privacy perceptions and behavior. In addition to conversational approaches, other modalities for information and presentation nudges need to be explored for voice-enabled CAI systems and could be developed by providers, integrators, and owners. Yeasmin et al. (2020) investigated modalities for privacy notifications in varying contexts and user preferences. They distinguished between visual and audio notifications and notifications via SMS, email, or app. While user preferences varied depending on the context, a majority preferred audio and visual notifications and notifications via app.

Lastly, warnings and reminders can serve as nudges to support developers in integrating security and privacy into their workflow. Thereby, it is important to note that developers are mostly aware of necessary security measures but lack knowledge of privacy practices (Balebako & Cranor, 2014). Nevertheless, in both cases, nudges that provide information can benefit developers and consequently their users. On the security side, security advice integrated into cryptographic APIs has been shown to significantly reduce insecure code (Gorski et al., 2018). On the privacy side, Peddinti et al. (2019) tested a nudge to inform mobile app developers of unnecessary permission requests. Thereby, they included information about permission requests of similar applications to incentivize developers to minimize personal data usage. They found that nudges were effective in reducing permission requests across a broad range of mobile application categories. Moreover, Tahaei et al. (2021) investigated framing nudges on developers with respect to mobile advertising networks. Among other conditions, they presented application developers with a privacy-focused framing that explained the impact of personalized ads on user privacy. Developers exposed to these options were significantly more likely to choose non-personalized ads over personalized ads, and most of them expressed the need to protect users’ privacy.

Limitations

Previous research has largely focused on nudging strategies for graphical user interfaces (Acquisti et al., 2018; Ioannou et al., 2021; Kitkowska et al., 2020). Yet, CAI systems can come without screens and require adaptable ways of presenting information to users. While research has started investigating new information and presentation nudges (Brüggemeier & Lalone, 2022; Harkous et al., 2016; Pearman et al., 2022; Yeasmin et al., 2020), there remain many open challenges. Future research could investigate the interplay between anthropomorphic features and conversational privacy as well as the influence of context on conversational privacy. Moreover, efforts need to be taken to allow comparability among CAI systems with respect to their privacy and security. While first evaluation frameworks are available to combat unethical design in CAI (Mildner et al., 2022), further research on standardized design guidelines and evaluation measures is required to ensure lawful and ethical design.

Takeaway

While previous attempts on nudging with information and presentation have been successful to influence privacy and security decision-making and can be adapted to CAI systems, CAI poses unique challenges to informing owners and users. Privacy labels and choice engines can be adjusted to fit the context of CAI and help owners and users to make more informed choices. In addition, new information and presentation nudges that resemble CAI’s modalities need to be explored. Conversational privacy can leverage CAI’s unique capabilities to communicate in natural language to inform owners and users about the system’s privacy and security. Lastly, warnings and reminders can nudge all actors in the ecosystem towards privacy-preserving choices.

5 Discussion and Future Work

In this position paper, we make two main contributions about applying debiasing strategies in the context of conversational AI. First, we establish a categorization framework for debiasing strategies based on previous research (Croskerry et al., 2013; Soll et al., 2015) and adapt existing privacy debiasing strategies to the context of CAI (see Fig. 2). Second, we assign those strategies to the relevant stakeholders of the CAI ecosystem as defined by European Data Protection Board (EDPB) (2021). Our proposed debiasing framework can serve as a suitable starting point to further investigate debiasing strategies for CAI but does not come without limitations and future research challenges.

First, we did not include a detailed overview of cognitive biases and heuristics in CAI systems as our focus was primarily on debiasing strategies and their application to CAI. However, due to the human-like nature of CAI systems and the complexity of the CAI ecosystem, CAI-specific biases might arise that should be explored in future work. Moreover, as described in Section 4.2, biases can also be specific to individual actors. Consequently, a comprehensive mapping of biases to actors could inform the design of novel debiasing strategies. Moreover, people or organizations can take on combinations of roles as shown in Section 4.1. Our discussion of debiasing strategies to the actors can help to identify useful strategies for such combined roles.

Second, our discussion has focused on mitigating biases in individual judgements rather than focusing on decision-making in groups. However, as conversational AI systems are often designed, developed, and deployed by a team of engineers and developers, mitigating bias on an individual level might not be sufficient. Importantly, strategies that are suitable for individuals might even introduce new biases on the group level (Kerr & Tindale, 2004). Therefore, future work should explore available approaches for groups to foster unbiased privacy and security decision-making.

Third, we focused on well-known debiasing strategies and their adaptation to CAI. Yet, social sciences and the medical field is especially rich in various debiasing strategies we have not addressed in this position paper, e.g., strategies based on pre-commitement (Ariely & Wertenbroch, 2002; Lambe et al., 2016). Future research could investigate additional debiasing strategies for privacy and security decision-making in CAI. Our proposed framework can thereby help to classify new strategies and set them in relation to existing ones. As all these interventions come with strengths and weaknesses, they should be seen as complementary to each other. Thus, our framework can support the development of holistic solutions by applying combinations of debiasing strategies to support people in their privacy decision-making.

Fourth, as a starting point, we focused on actors defined by the European Data Protection Board (EDPB) (2021). Yet, due to the complexity of the CAI ecosystem, other actors are likely to play a role in the design, implementation, and deployment of the systems. Moreover, we have only slightly touched on the role of policymakers and regulators in incentivizing privacy and security. As policymakers have relied on nudges to assist decision-making in many fields (Thaler & Sunstein, 2021), they form an influential group and should be considered in more detail in future work.

While we aimed at providing an overview of debiasing strategies from different fields, i.e., social science, medical field, privacy, and security, we acknowledge that there are research directions that we have only slightly touched upon or have not considered. For example, we included virtual badges and app reviews in Section 4.2.4—a debiasing strategy that falls into the area of gamification. Gamification focuses on triggering intrinsic motivation through the adoption of game elements and has developed independently from nudging and behavioral economics (De Troyer, 2021). Nevertheless, gamification can support people in their privacy and security decision-making, e.g., by incentivizing privacy and security design or by applying them to educational settings to enhance individuals’ privacy literacy. Recent years have seen combined research on gamification and nudging to support sustainability behavior, engagement in mental health applications, or mitigation of cognitive biases (Auf et al., 2021; Dunbar et al., 2014; Luger-Bazinger & HornungPrähauser, 2021). Therefore, future research should explore gamification as a meta-strategy and its effect on debiasing privacy and security decision-making in CAI.

Moreover, we focused on conversational AI systems in general including text-based as well as voice-based systems. Yet, these differences in modality can influence peoples’ perceptions (Cho, 2019). Future research could investigate debiasing strategies for various modalities and explore their differences and commonalities. Here, our categorization framework can help to compare debiasing strategies across modalities.

Finally, we want to urge the need for a comprehensive evaluation framework for debiasing strategies for privacy and security decision-making. So far, only a few studies have discussed potential evaluation measures and guidelines for ethical nudge design (Acquisti et al., 2018; Barev et al., 2021; Renaud & Zimmermann, 2018). Yet, they do not provide quantitative measures to evaluate the effectiveness of debiasing strategies. Instead, the privacy field could benefit from drawing on evaluation measures used in the medical field or on studies on cognitive control and rational decision-making (Kahneman, 2011; Lambe et al., 2016; Mushtaq et al., 2011). In more recent work, Habib and Cranor (2022) present an evaluation framework for privacy choice mechanisms. Thereby, they include the aspect of neutrality to evaluate privacy choice mechanisms to address nudging patterns and in particular dark patterns that nudge users away from privacy-protective options. Yet, such a framework does not take into account bright patterns and nudging strategies towards privacy-preserving behavior. Moreover, only a few have focused on evaluating debiasing strategies in light of current legal regulations and their legitimacy (Barev et al., 2021, 2022). Therefore, interdisciplinary research is necessary to pave the way for effective and legitimate debiasing strategies for privacy decision-making in CAI.