Introduction

Almost every person holds information about others. This means we can share information about each other, which, in turn, decreases the degree to which we can self-protect our privacy. Privacy interdependence is a substantial privacy threat (for some potential consequences, see Chaudhry et al., 2015; Humbert et al., 2019), and many large online businesses, like Meta, harvest this third-party data for profit (Nunan & Di Domenico, 2013; Sarigol et al., 2014). People who share others’ data may sometimes consider the benefits for the other person when they make the decision to share data about others, for example when recommending others in a referral program. However, in many interdependent privacy situations the direct benefits to others are limited or unclear. This also holds for app-driven privacy infringements. In an age in which everyone has others’ data on their mobile phones (e.g., contacts, photos and messages) and apps regularly request access to these data, this is an important ethical phenomenon. To illustrate, consider the Cambridge Analytica scandal: although only 270,000 users had installed the app-based quiz on Facebook, data were eventually harvested from 87 million users (Lapovsky, 2018).

Notably, giving away the data of others without their knowledge or permission is ethically questionable. In the analogue world, giving away the personal data of very many others is considered morally wrong (Petronio, 2015) and people commonly obtain permission before agreeing to share the contact details of a single person, let alone all their contacts. How can these same people then apparently justify the giving away of others’ data to an app? This is particularly relevant, given that the scale of app-driven privacy infringements means consumers are likely to be both victims and offenders. While a lack of awareness has been highlighted as an issue (Kamleitner & Mitchell, 2019), 70% of a student sample and 65% of a UK Prolific sample decided to continue using a social media app even after they had been made aware (Kamleitner et al., 2018). This implies awareness on its own may have limited impact on reducing the behavior.

Addressing the question of how people justify giving away the data of others in an app, our research draws on the theoretical lens of the theory of neutralization (Kaptein & van Helvoort, 2019). This theory captures nuanced justifications that people use to neutralize, any guilt and blame after having engaged in deviant behavior. The way people justify themselves holds information about the degree to which they perceive a behavior to need justifying, i.e., how much this behavior is a norm violation. This offers a novel application of the theory of neutralization, namely, that people’s justifications of an act can be used to investigate its normative acceptability. In situations where norms are ambiguous or evolving (such as the sharing of others’ data through apps) and where individuals may feel uncertain about what is right or wrong, the way they justify a behavior can provide valuable insights into the emerging norms. In addition, responding to recent comments that the victim–offender overlap has not received sufficient attention in examining the explanation and control of misdemeanors (Berg & Schreck, 2022, p. 278), we empirically explore offenders’ justifications, and the justifications victims expect them to adopt. This novel juxtaposition of perspectives allows a unique window into the interplay of norms surrounding the giving away of others’ data in apps. Since justifications mirror actions, the dynamic interaction between the rationales offered by offenders and the perceptions of victims also suggests the direction the formation of norms is likely to take.

This research contributes to the literature in several ways. To the interdependent privacy literature (e.g., Demmers et al., 2022; Kamleitner & Mitchell, 2019; Pu & Grossklags, 2016; Woods & Böhme, 2022), we provide novel insights into the justifications and moral considerations behind the widespread and problematic behavior of app-driven privacy infringements. To the literature on neutralizations (Kaptein & van Helvoort, 2019; Sykes & Matza, 1957; Willison & Warkentin, 2013), we contribute in three ways. First, we add another important context for elucidating context-specific neutralizations. Second, we introduce a novel application showing how neutralizations can provide insights on norms and their formation. Third, we show how this theory can be applied to explore and contrast perspectives other than that of the offender, in our case the victim perspective (Berg & Mulford, 2020; Gottfredson & Hirschi, 1990). This novel twist allows for a better understanding of how neutralizations operate in circumstances where offenders are also victims at different times. We also show how this novel twist provides insights into the moral dynamics of the behavior and the formation of norms around it. Finally, and based on our qualitative insights, we also add to neutralization theory by allowing for the possibility that people may decide not to neutralize their privacy-infringing behavior.

The paper is structured as follows: We first briefly outline the key phenomenon of interest and why sharing others’ data with an app ought to be considered wrong thus requiring justification. We next conceptually integrate the categories of neutralization and literature on norms and the victim–offender perspective before exploring the use of neutralizations empirically. An initial qualitative pilot study supports the merit of perspective taking and suggests the addition of a “responsibility acceptance” (non-neutralization) category. Three subsequent quantitative studies systematically vary victim–offender perspectives and details of the app-driven privacy infringement allowing us to explore the moral dynamics of the behavior. We conclude by outlining the theoretical and practical implications of our insights and by highlighting future research opportunities.

Theoretical Background

Interdependent Privacy Infringements

In general, interdependent privacy refers to situations in which the “actions of some individuals affect the privacy of others” (Humbert et al., 2019, p. 3). Digitalization has enormously boosted the potential of interdependent privacy infringements and an increasing number of scholars (for a review, see Humbert et al., 2019) are starting to address the topic (e.g., Biczók & Chia, 2013; Pu & Grossklags, 2016; Sarigol et al., 2014; Woods & Böhme, 2022) including its potential for harm (Harkous & Aberer, 2017; Olteanu et al., 2017).

In the context of apps, consumers regularly click accept to requests for data about others (Pu & Grossklags, 2016) and consequently, peer-privacy protection frequently fails (Symeonidis et al., 2016; Woods & Böhme, 2022). Although technically avoidable, this is true for almost all social media apps and happens for various other applications such as TripAdvisor, where the likelihood that the information of other people gets collected can be greater than 80% (Symeonidis et al., 2018).

Existing research has started to examine the user side of the phenomenon. The empirical evidence to date suggests that people tend to value their own information more highly than the information of others (e.g., Pu & Grossklags, 2016) and largely volunteer others’ data without realizing they are doing so (Franz & Benlian, 2022; Kamleitner & Mitchell, 2019). Failure to realize that data, and any data transfer, are also about others explains a large part of the problem. For example, in a social media app download simulation, 95% of students and 71% of the general public were unable to correctly recall what data they volunteered to an app and 42% of students and 49% of the general public did not realize that permissions could involve contacts or pictures that could infringe on the data rights of people other than themselves (Kamleitner et al., 2018).

Clearly, awareness is an important part of the issue, but people have also been observed to continue giving away others’ data even once they have been made aware. For example, Kamleitner et al. (2018) found that 65% of the general population and 70% of a student sample said they would keep an app despite knowing that it infringed on others’ data. In a recent experimental simulation (Franz & Benlian, 2022), a salience nudge requesting to actively confirm others’ consent (taken from Kamleitner & Mitchell, 2019) reduced the probability that people would disclose others’ data by 62%. This is an impressive effect, but it still implies that 38% were prepared to wrongly claim (in a simulation with no consequences for themselves) that they had obtained the consent of all of their contacts (for other instances of the conscious giving away of others’ data, see Demmers et al., 2022).

Despite the behavior of accepting an app’s request to others’ data being widespread, it is arguably morally wrong. At least that is what it would be if judged against the social norms of the offline world. Offline, implicit norms have been negotiated about “what, why, and to whom information is shared within specific relationships” (Martin, 2016, p. 551). When asked in person, people do not commonly pass on their friends’ contact details without first obtaining their consent (Kamleitner & Mitchell, 2019; Petronio, 2000). This reigning social norm exists for a good reason. Costs or difficulties arising from the sharing of others’ data in an app download context tend to be borne by others, which is why it has been linked to the economic concept of negative externalities (Biczók & Chia, 2013). In fact, the potential for harm from app-driven privacy infringements is massive as illustrated by the Cambridge Analytica case. Online, more data can be shared more quickly and there is more potential for third party use or abuse of others’ data. In view of its long-term potential for harm, one might expect that people would condemn the practice of giving away others’ data via apps.

One might also expect that it might be prohibited from a legal perspective. However, privacy and data protection laws vary from country to country. To illustrate the complexities, consider the key regulation about privacy in the European Union, the General Data Protection Regulation. Article 2 (2) (c) of the GDPR specifically excludes the processing of personal data by an individual solely for the purpose of personal or family-related activities (Bergauer, 2020). However, when controllers or processors facilitate personal data processing for such activities (for instance, providing software or tools), the GDPR still applies.Footnote 1 Additionally, the exception does not apply if the processing of personal data is driven by an underlying economic motive (Zukic, 2019). The first point to be determined is thus whether GDPR applies at all. If applicable in a particular case, the next question is whether one of the lawful grounds for data processing, i.e., consent by the data subject (Article 6) or legitimate interest, is present. If consent applies, then Article 7 specifies that informed consent requires notification of the original data owner together with easy withdrawal of consent. GDPR emphasizes explicit consent, which Faden and Beauchamp (1986) define as consent given by a subject with a substantial understanding of the transaction, in substantial absence of coercion by others, intentionally, and authorizing a certain course of action. Evidently, people do not provide explicit consent when others share their data without them being notified. That leaves the issue of implied consent, which occurs when individuals take an action that presupposes consent to specific privacy practices, policies, or other agreements without formally providing verbal or written permission (Veatch, 2007). However, when providing a colleague, friend or acquaintance with personal information, the lack of realization documented in prior research (Franz & Benlian, 2022; Kamleitner & Mitchell, 2019) combined with reigning social norms (Petronio, 2000) suggests that people do not commonly presuppose that their data will be handed on to third parties. In addition, it may not always be clear who the data owner is. For example, some personal data, such as conversations and pictures, may have joint ownership. Given these complexities, there is a potential legal threat to the individual private offender and companies who collect third-party data (Kamleitner et al., 2018). In this research, we aim to explore how people justify this morally and legally questionable behavior by drawing on the theory of neutralization and its ability to provide a glimpse of the presence of norms surrounding a behavior.

Theory of Neutralization

Neutralizations are explanations that people use to help “argue away, fully or partly, someone’s responsibility” for a harmful behavior so that an offender experiences no or less guilt (Kaptein & van Helvoort, 2019, p. 4) and averts or reduces repercussions. Neutralizations have become one of the most commonly used concepts to understand and explain morally questionable behavior and people’s propensity to repeat such behaviors (Serviere-Munoz & Mallin, 2013). Developing the original five techniques first identified by Sykes and Matza (1957), other work has applied and adapted these to numerous contexts such as employees’ unauthorized use of IT devices (Silic et al., 2017), app developers’ ethical conceptions (Shilton & Greene, 2019) or the illegal sharing and downloading of files in a peer-to-peer context (Cohn & Vaccaro, 2006; Harris & Dumas, 2009; Odou & Bonnin, 2014). In an effort to deepen and consolidate the various applications of the theory of neutralization, Kaptein and van Helvoort (2019) have developed a comprehensive theoretical framework which organizes 60 specific techniques into four generic categories of neutralization. Figure 1 illustrates them.

Fig. 1
figure 1

Note Distribution of predominant categoriesDistribution of predominant categories of Adapted from Kaptein and van Helvoort (2019). The fifth category has been added in response to the results of our pilot study and in response to our novel application of neutralization theory, i.e., using justifications to infer the presence of (strict) norms

An extended list of categories of neutralization along a responsibility allocation continuum.

These categories are formulated at a higher level of abstraction and can be applied across different contexts and types of transgressions including behaviors that may not even be considered wrong by many. Notably, categories differ in terms of how much they enable offenders to shift responsibility and blame away from themselves. Kaptein and van Helvoort (2019) consider these four categories of neutralizations from (1) to (4) as increasingly socially ‘unsafe’, meaning the closer one comes to admitting responsibility. The first two categories, (1) distorting the facts (it’s not true) and (2) negating the norm (it’s not decisive), depict the denying of a behavior being deviant or people’s genuine belief that a behavior is morally acceptable.

The next two categories, (3) blaming the circumstances (it’s beyond my control) and (4) hiding behind oneself (it’s a lack of self-control), depict denying full responsibility while admitting to moral wrongdoing, i.e., to a norm violation.

Because the categories of neutralization are tied to the ease with which offenders can avoid taking responsibility (see Fig. 1), the Kaptein and van Helvoort (2019) study also provides insights about how much people feel that the behavior needs justifying. The categories thus also deliver insights into whether people feel that a behavior is morally wrong and violates a norm. In our research, we aim to make use of this little regarded implication of the framework.

Generally, literature on neutralizations has been interested in understanding how people manage to justify and even distance themselves from a norm violation. Kaptein and van Helvoort's (2019) framework, however, highlights that people’s neutralizations are more than individual excuses. The categories of neutralization carry different moral weights and reflect the degree to which a behavior feels morally wrong, i.e., violates a norm. The categories of (1) distorting the facts and (2) negating the norm essentially correspond to the denial of moral wrongdoing. Individuals drawing on these categories deny the existence of a norm violation, either because they genuinely do not feel a norm has been violated or because they consider the norm is lax, debatable or not socially agreed upon. In contrast, categories (3) blaming the circumstances and (4) hiding behind oneself correspond to the admittance of a norm violation and simply differ on how a person tries to avert the blame. The categories of neutralization highlight the entanglement between justifications and norms. Importantly, the violation of norms gives rise to neutralizations, but neutralizations also provide a window on norms and their formation. In particular when norms are unclear or evolving and when uncertainty blurs moral consensus (such as in a novel moral problem), individuals’ justifications provide insights on a behavior’s moral acceptability that are prone to be more reliable than asking about norms directly. In other words, the process of neutralization itself can be indicative of the normative beliefs forming within a society.Footnote 2 A key merit of assessing justifications is that they hold more concrete details and are less abstract than so called norms. They are easier to grasp and thus report on concrete rather than abstract notions (Trope et al., 2007). Knowing how people justify a behavior is also useful in predicting individuals’ future actions, e.g., ignore, repeat or (self-)punish the behavior, and in understanding how people will talk about the behavior from a moral perspective; after all much normative conversation centers on whether and how people justify their own or others’ behaviors. In short, neutralizations allow insights on how the dialog about a (potentially) problematic behavior, its prevalence and its moral acceptability are prone to evolve over time. The moral acceptability of a behavior is, however, multi-faceted.

Moral Acceptability, Norms and the Sharing of Others’ Data via Apps

Norms guide people’s behaviors and inform them about what is or feels right or wrong. Those norms can be personal or social in nature and together they determine the moral acceptability of a behavior. Social norms are “a person’s beliefs about the common and accepted behaviors of other people” (Schultz, 2022, p. 2) and essential enablers of human societies (Rossano, 2012). Social norms can either be descriptive or injunctive. Descriptive social norms specify what is done, i.e., provide guidance by highlighting what most others do, while injunctive norms constitute “rules or beliefs as to what constitutes morally approved and disapproved conduct” (Cialdini et al., 1990, p. 1015). Most people are intrinsically motivated to comply with social norms (Amiot et al., 2013) and noncompliance with injunctive (as opposed to descriptive) norms tends to be socially sanctioned (Lapinski & Rimal, 2005). In contrast to social norms, personal norms are internalized beliefs and values that guide an individual’s behavior and emerge at the level of the individual (Schwartz, 1977). Although personal norms often reflect reigning social norms, compliance with personal norms is not based on “fear of social sanctions but the anticipation of having broken her/his personal norms” (Bamberg et al., 2007, p. 191). All norms tend to be context specific (Pedersen & LaBrie, 2008) and they can shift (Burke et al., 2010). In particular, social norms are constantly negotiated and dynamically evolve as people develop a joint understanding about new contexts (Yoshida et al., 2012).

Norms guide all human behavior including data and information sharing (Acquisti et al., 2011; Barth et al., 2006; Hoyle et al., 2020; Nissenbaum, 2004). However, in many respects app-driven privacy infringement is a new context. For one, it happens largely automatically and invisibly (Kamleitner & Mitchell, 2018). This makes the behavior hard to observe and comprehend. Yet, directly observing what others do and how others react to a behavior is an important basis for the establishment of social norms (Kenrick et al., 2003). Lack of observability also hampers blame attribution in case of harm. While sharing others’ data may be legally questionable, it is often hard to prove that a specific harm has been brought about by a specific act of data sharing by a specific individual. People mostly cannot observe what others do. Upon reflection, however, most are bound to realize that they allowed apps to access others’ data and other people will have provided access to their data (Kamleitner & Mitchell, 2019). The fact that people are both victims and offenders makes the establishment of norms and the corresponding forgiving and blaming complicated (Worthington Jr, 2009). It may either increase the uncertainty around its moral acceptability or reduce the sense of norm violation (Cohen & Felson, 1979; Osgood et al., 1996).

Categories of Neutralization and Victim–Offender Perspectives

Remember that we suggest that the categories of neutralization people draw on are likely to provide information about whether they consider a behavior as morally defensible or violating a norm. Individuals’ usage of categories (3) and (4) implies the admittance of a norm violation but it does not tell us what kind of norm(s) have been violated. Offenders are prone to experience guilt and justification pressure when violating either personal or social norms. Although personal norms carry more weight for an individual’s immediate reaction, social norms inform about a behavior’s acceptability, about “whether it is beneficial or easy to perform” (Bamberg & Möser, 2007, p. 17) and about whether offenders experience external blame on top of potential self-blame. Giving away others’ data might be at odds with personal norms that respect others’ privacy, yet those norms might relax over time if they are not supported by equally prohibitive social norms or legal sanction. The fact that most people are sharing others’ data suggests that prohibitive injunctive social norms might be lacking. In fact, when we asked directly, 90% of a US sample said that there were no very clear norms around the behavior, yet over 80% saw it as morally wrong (see Figs. 4 and 5 in Appendix B).

But what does this presumable lack of clear social norms mean in practice? How do people translate this uncertainty into an overall moral assessment of the behavior? Do people fill the void in social norms with their personal norms, or do they interpret it to indicate moral acceptability? Answering these questions matters because it provides a window into the future prevalence of the behavior. If a lack of clear social norms is equated with normative permissiveness, there will be no social sanction and blame, even if offenders feel guilty for violating their own personal norms. The hurdle to reengage in the behavior is lower and eventually even the prohibitive personal norms may start to relax.

Personal and social norms combine to pave the way for a behavior’s future prevalence. Notably, this future is shaped by how both offenders and victims think about and react to it. In the case of interdependent privacy infringements, the importance of offender and victim perspectives is particularly pronounced. After all, most people have had their data given away by others and have given away others’ data themselves. Cultural theories have long taken the view that the victim–offender overlap is a product of conduct norms that endorse crime and violence (see Lauritsen & Laub, 2007). Looking only at the offender perspective provides a limited view on the behavior’s moral acceptability. Consequently, we introduce perspectives (offender vs. victim) to the neutralization framework and build on the fact that people can justify their own behavior as well second-guess another’s justifications. For example, when Susan learns that Fred shared her number, she may second-guess whether and how Fred would justify himself, e.g., by claiming that it wasn’t ideal, but he only did what others do. This second-guess is likely to affect Susan’s own behavior in the future. If she thinks that Fred didn’t think it a big deal, she is likely to think it less of a big deal herself in the future (Gino & Galinsky, 2012). Both Susan’s second-guess and Fred’s actual justification reflect a blend of their personal and perceived social norms. However, the relative composition of this blend is likely to differ. Past literature has shown a gap between people’s own attitudes and those they ascribe to others (Alicke et al., 1995; Anderson & Godfrey, 1987; Weiss et al., 2018). Research also suggests that victims and offenders do not draw on the same normative basis (McCarthy et al., 2021), with offenders being particularly likely to draw on their personal norms, which are a key source of guilt (Bamberg et al., 2007, Schwartz, 1977). By and large, offenders' neutralizations are prone to unveil both personal and social norms, but if justifications are voiced to oneself, personal norms are prone to be more decisive. In contrast, when victims second-guess offenders’ neutralizations, they are prone to be based on social norms which is the common ground between victims and offenders.

Proposition 1

There will be a difference between offenders’ use of categories of neutralization and victims’ second-guesses of their use. This difference provides insights on the overall social acceptability of the giving away of others’ data.

This dynamic interplay significantly shapes the development and evolution of emerging social norms related to interdependent privacy. In essence, we suggest that the ongoing interaction between offenders’ justifications and victims’ perceptions actively contributes to the negotiation and establishment of new societal standards in the realms of privacy and data sharing. The app-based sharing of others’ data happens across a large range of apps and it arises in various interpersonal relationships. Given that social norms are often context specific (Pedersen & LaBrie, 2008) and that privacy behavior is a highly contextual phenomenon (Morando et al., 2014), these differences are likely to matter for the behavior’s moral acceptability. If they do, such variations provide us with further insights on the origins and dynamics of the moral acceptability of interdependent privacy infringements. We explore two such variations: the type of app and whether the “other” is specified or not.

Some apps, such as messenger apps, are about connecting people. Other apps, such as information apps, are not. This difference plausibly affects how justifiable the sharing of others’ data is deemed. Consumers may well assume that apps that connect people need others’ data to properly function and this could render the sharing of others’ data more acceptable. In fact, prior evidence suggests that people are particularly likely to devalue others’ data when they believe that this data will improve the apps’ functionality (Pu & Grossklags, 2017). We thus expect that sharing others’ data may be considered less of a norm violation for communication apps than for other apps (e.g., information apps).

Proposition 2

App type will affect the moral acceptability of sharing others’ data with an app. This influence will become evident in the categories of neutralization that victims and offenders use.

Interdependent privacy issues arise because people are socially related and there is reason to believe that knowing who exactly the other is makes a difference. Both the identification of victims and the social proximity between people affect moral judgements and decisions (Lee & Holyoak, 2020; Sah & Loewenstein, 2012). In app-based privacy infringements, specifying the other could be important in affecting the moral acceptability of the behavior for both victim and offender perspectives. Theoretically, different directions are plausible. On the one hand, findings in criminology show that putting a human face on a victim or meeting a victim helped offenders realize the extent of the consequences of their actions (Choi et al., 2011). In other words, offenders might feel more responsible toward specific others and this may translate into the use of higher categories of neutralization. On the other hand, offenders might feel more entitled to share information about specific others—in particular if they are close to them (Lerner & Mikula, 2013). It may thus be easier to justify the behavior when affecting close others. This would be equivalent to what has been found from a victim perspective. People were observed to allot less responsibility to offenders who were close to them (Hofmann et al., 2018) and to argue away moral blame when becoming the victim of the actions of a close other (Gino & Galinsky, 2012). We thus propose:

Proposition 3

Specifying the other versus not specifying the other will affect offenders’ use of categories of neutralization and victims’ second-guesses of their use.

Overview of Empirical Exploration

In four studies, we explore how offenders (are assumed to) neutralize the sharing of others’ data once they learn that this is what they did when agreeing that an app can access the contacts on their phones. Our starting point for all studies is the offender’s self-perspective, i.e., we query what category of neutralization offenders primarily draw on to explain their behavior. In all studies, we also explore victims’ second-guesses about offenders’ explanations. This contrast of perspective allows us to explore the overall moral acceptability of the behavior. To ensure the nuance of our insights, we explore the prevalence of categories of neutralizations across different situations and samples. Note that our interest in the prevalence of neutralization categories across different perspectives invites quantitative exploration. While the design we adopt closely resembles classical hypothesis testing, the spirit is not. We introduce variations to contexts as a way to probe how people currently make moral sense of contextual variations and firmly expect that specific results will change as norms evolve. Our exploratory insights thus serve as a map of the behaviors’ current normative portrayal and as a benchmark for its future change.

We begin with a qualitative pilot study that reaffirms the fact that sharing others’ data is morally in question. In addition, it highlights the complexity of neutralizations and the need to add a fifth “no neutralization” category to our framework. Further, it unearths the tendency of victims to take the perspective of the offender. Studies 1 and 2 next adopt a quantitative paradigm and examine what neutralization categories, including a no neutralization category, dominate and whether their relative prevalence varies across perspectives (proposition 1). Addressing proposition 2, study 1 also varies the type of app. Addressing proposition 3, study 2 additionally varies the salience of a specific “other”. Study 3 finally adds additional information in response to questions arising from the prior studies. Specifically, it further enhances the realism of the study setting and it explores a potential angle to curb the behavior by making people aware that the act of sharing others’ data is preventable by the press of a single button. Adding theoretical refinement, study 3 also assesses the complementary perspectives of victims’ own moral judgements and offenders’ second-guesses about these. Table 1 gives an overview of all three quantitative studies. Given our theoretical focus and the limitations of journal space, we only report on the categories of neutralization in our results. The methodological appendices A and B comprise the full set of items and additional analyses.

Table 1 Overview of samples, variations and purpose for quantitative studies 1, 2, and 3

Pilot Study: A Qualitative Exploration of Responses to App-Driven Privacy Infringements

In this pilot study (reported in Appendix B), we explored 115 Austrian university students’ spontaneous written reactions to a situation in which an app had been given access to another person’s data. The goals of this initial study were (a) to explore whether participants would actually neutralize the behavior and consider it morally questionable and (b) to ensure that the generic descriptions of the categories of neutralization made sense against the backdrop of a natural discourse around the behavior. To this end, we varied whether participants adopted the role of an offender or a victim. In a lab environment, participants were randomly assigned to one of the two roles (offender vs. victim) and asked to write down their thoughts and feelings on a situation in which either they themselves had permitted a game app to access their friends’ data (contact details, pictures and call logs) (offender) or in which a friend had permitted a game app to access their own data (victim). Two independent coders coded responses in terms of neutralization categories (see Fig. 1) and the presence of perspective taking.

Even though participants had not been directly asked to justify the behavior, there was evidence for the usage of all four neutralization categories, often within the same person, from both the victim and offender perspectives (for detailed results see Appendix B). The wording used to justify the behavior was often also remarkably close to the descriptions of the categories of neutralization. For example, “At first it doesn’t seem ‘that bad’—but on reflection you get an uncomfortable feeling” (female offender, aged 28) or “Feels a bit like a loss of control. Friend is not to blame, because I also always just press ‘install’ without reading through the conditions” (female victim, aged 25). Indicating a behavior with moral question marks, participants appeared to use signals from themselves and others to infer what appears acceptable (Lin & McFerran, 2016; Wenzel, 2005). Notably, multiple participants in the victim condition spontaneously put themselves in the shoes of the offender. They second-guessed offenders’ thoughts and drew on what they appeared to assume was the social norm, e.g. “I can’t blame my friend, though, since everyone has done this before” (female victim, aged 20).

The pilot study affirmed the ecological validity of our approach. It confirmed (a) that the app-driven sharing of others’ data is considered morally questionable and spontaneously neutralized, (b) that people draw on multiple neutralizations and use wordings that resemble the neutralization categories, and offered the additional insight, (c) that people put into the position of a victim were prone to second-guess neutralizations on behalf of the offender. Going beyond our expectations, victims sometimes put themselves in the shoes of the offender to explicitly derive what they believe to be universally accepted norms. Finally, we also observed that the four categories of neutralization were unable to capture all the morally relevant responses. Sometimes offenders and victims did not resort to neutralization, but were willing to assume full responsibility. In our subsequent studies and in line with our novel application of the theory of neutralization as a window on a behavior’s moral acceptability, we thus add a ‘no neutralization’ category (see Fig. 1).

Study 1: Categories of Neutralization Across Perspectives and App Types

In study 1, we examine the prevalence of neutralization categories. Specifically, we aim to assess category use by offenders and to learn about victims’ corresponding second-guesses following proposition 1. To get a fuller picture of the overall moral acceptability of the behavior, we also varied the app type and examined whether the pattern of justifications used is sensitive to app type following proposition 2.

Sample and Procedure

We recruited 293 members of a representative sample of the Austrian working population via a professional online panel (53% female, 47% male, 0.3% diverse, mean age = 42.56, SD = 13.52). Study 1 employs a 2 (perspective: offender—self vs. victim—second-guessing offender) × 2 (app type: communication vs. information) between-subject design. Participants read about a scenario of interdependent privacy infringement before getting an opportunity to react to this situation (see Appendix A for study materials). In both conditions, an app (either a communication or information app) was said to have been granted permission to access data such as pictures and contacts. As a result, the app provider was able to access and use the data of others. Participants in the offender condition read that they gave away the data of a person close to them when installing an app. Participants in the victim condition read that a person close to them had given away their data when installing an app.

Following the scenario, participants were asked to indicate which statement best described the offender’s thoughts (their own thoughts in the offender condition vs. their close other’s thoughts in the victim condition) about giving away the victim’s data. Participants could choose one of the following five options that represent the different categories of neutralization.

  1. 1.

    I do not think that anything happened at all.

  2. 2.

    Yes, something happened, but it really is not bad.

  3. 3.

    Yes, something happened that should not have happened, but it is not my fault as I did not have any real control over it.

  4. 4.

    Yes, something happened that should not have happened, but it really would have been difficult for me to prevent it.

  5. 5.

    Yes, something happened that should not have happened. I take full responsibility for it.

Additional exploratory variables that go beyond our key research question are featured in Appendix B (Table 3).

Results

Figure 2 depicts the distribution of neutralization categories across all exploratory conditions of studies 1 and 2. Darker shades indicate neutralization categories suggestive of the absence of a norm violation, while lighter shades suggest the presence of a norm violation. The top half of the figure features offenders’ explanations, the bottom half features victims’ second-guesses about these explanations. The boxed in area in the middle contains a summary of both studies split into perspectives.

Fig. 2
figure 2

Note To facilitate a grasp of the bigger picture, results are sorted according to perspective rather than studies. The boxed area in the middle contains average responses per perspective across studies 1 and 2. Note Please note that in some cases percentages do not add up to 100 percent exactly due to rounding

Distribution of predominant categories of neutralization for each cell of studies 1 and 2.

Across conditions of study 1, we find that all categories of neutralization were used by at least part of the sample. Approximately half of the participants denied that any norm violation had occurred in response to the offender giving away the victim’s data when downloading the app (53%). Of these, the majority denied that anything had happened at all (37% distorting the facts vs. 16% negating the norm). The other half of participants acknowledged the presence of a norm violation (47%) and the dominant way of doing so was to fully embrace responsibility and (self-)blame the offender (27%).

Comparing the pattern of answers across conditions puts these overall results into perspective. Overall, and speaking to the absence of a clear social norm violation, participants adopting the victim perspective tended to use the lower categories of neutralization, indicating they thought offenders were holding themselves less responsible than they actually were. This finding supports the reasoning behind proposition 1. In line with proposition 2, the imbalance between offender and victim perspectives persists across types of apps but becomes less pronounced in the case of a communication app, which reduces the propensity of responsibility taking by the offenders (see Fig. 2). Notably, while participants in the offender condition were sensitive to the context, participants in the victim condition were not. They were equally likely to excuse more (information app) and less (communication app) avoidable privacy infringements.

To determine the effects of both perspective and type of app, we conducted a multinomial logistic regression (see Table 2 for detailed results) with the no neutralization category serving as a reference category. In support of our overall observation of an effect of perspective, this factor approximated significance for the choice of all neutralization categories. Participants in the victim condition were significantly more likely to distort the facts (category 1) by indicating that the offender would think that nothing had happened (victim: 47% vs. offender: 29%) and to blame the circumstances (category 3) by maintaining that the offender had no real control over it (victim: 10% vs. offender: 3%). In contrast, participants in the offender condition indicated significantly more often they would take full responsibility for the transgression (category 5; offender: 37% vs. victim: 14%, χ2(1) = 18.93, p < 0.001). Interestingly, neither an effect of app type nor an interaction effect between perspective and app type emerged. Having said that, Fig. 2 shows that this lack of significance is due to our theoretically motivated choice of reference category. In the case of the communication app, 28% of offenders were willing to take full responsibility, while in the case of the information app this percentage increased to 44%. In essence, offenders were more likely to take responsibility when the app could be plausibly believed to fully function without accessing others’ data.

Table 2 Multinomial logistic regression of experimental variations on usage of neutralization categories in studies 1, 2, and 3

Study 2: Categories of Neutralization Across Perspectives and (Un)specified Others

In study 1, we assessed what happens at the generic level of the problem description. We abstractly described the giving away of others’ data to an app. We neither specified the exact app, nor the specific other or the exact permissions. In study 2, we deepened the exploration and went concrete. We used a mock-up news aggregator app and were clear about the permissions granted. In addition, we addressed proposition 3 and explored whether naming a concrete other affects people’s use of neutralization categories.

Sample and Procedure

We recruited 348 American citizens via Prolific Academic. All participants owned a smartphone, passed attention checks, and completed the online questionnaire (49% female, 50% male, 1% other, mean age = 38.14, SD = 14.35). Participants were randomly assigned to one of the four experimental conditions of our 2 (perspective: offender—self vs. victim—second-guessing offender) × 2 (other: specified vs. unspecified) between-subject experimental design. The scenario was based on study 1 but instead of featuring an abstract reference to an app, participants read about and were shown a concrete news app called Spotlight News, which offers a personalized news alert service. A realistic screenshot detailed the permissions participants needed to grant.

Participants in the victim condition read that a person close to them (specified) or many other people (unspecified) had given away their data when installing Spotlight News and participants in the offender condition read that they gave away the data of a person close to them (specified) or many other people (unspecified) when installing Spotlight News. Prior to this scenario, participants in the specified other conditions were asked to name a concrete other person. To stress the manipulation, we inserted this specific name throughout the study via the piped text feature offered by Qualtrics. All key variables were assessed as in study 1 (see Appendix A, Table 4).

Results

Across conditions, 38% of the participants denied that any norm violation had occurred when the offender gave away the victim’s data by downloading the app (category 1: 22% distorted the facts, category 2: 16% negated the norm) compared to 62% who acknowledged the presence of a norm violation. As in study 1, the dominant category (category 5: 37%) was to fully embrace responsibility and (self-)blame the offender followed by hiding behind oneself (category 4: 16%) and blaming the circumstances (category 3: 9%).

We regressed the categories of neutralization on the perspective, other specificity and their interaction, and again used category 5 as a reference category. Results reconfirmed the main effect of perspective and reasoning behind proposition 1 qualified by a significant interaction effect (see Table 2 for detailed results). Essentially, this interaction effect is partly in line with proposition 3. Offenders were twice as likely (4% vs. 8%) to adopt category 1 when the victim was specified compared to when the victim was not specified but there was no difference in offenders’ use of categories 2, 3 and 4. In contrast, victims guessed that specified offenders as opposed to unspecified offenders would be less likely to draw on categories 2, 3 and 4. The most pronounced difference, however, emerged in victims’ assumption of the use of the reference category 5. Some victims guessed that a specified offender they well knew would take full responsibility (18%) but no victim assumed that unspecified offenders would do so (0%; χ2 (1) = 12.00, p = 0.001).

Key Insights of Studies 1 and 2

Figure 2 highlights three overarching insights. First, all categories of neutralization were used by at least part of the sample. The sharing of others’ data was neither universally judged as wrong nor universally considered free from responsibility. In fact, the two opposite extremes were chosen most often: category 1 as in “nothing happened that anyone could be blamed for” and category 5 as in “something bad happened and there is no excuse”. Perhaps surprisingly, participants least frequently chose category 3, blaming the circumstances as in “it was out of my control”. Nearly no participants in the offender perspective resorted to this justification. Participants did not seem to see the app provider as the main culprit but to rather see this as a consumer-to-consumer issue.

Second, the perspective matters. The apparent disparity in category use across offender and victim perspectives hints at a normative disagreement between victims and offenders. Overall, offenders were more prepared to admit to wrongdoing and to take responsibility than victims second-guessed. This finding is remarkably robust across samples (Austria and USA) and contexts (types of apps and specificity of offenders). In line with our theorizing, this suggests that victims and offenders do not draw on the same normative basis (McCarthy et al., 2021). Victims substantially underestimated the extent to which offenders were willing to admit to wrongdoing. Arguably, most victims felt that offenders would not perceive the app-driven sharing of others’ data as a norm violation and thus rendered the behavior socially acceptable. At the same time, sharing others’ data was considered a violation of norms by most offenders.

Third, our contextual variations primarily affected participants adopting the offender’s perspective. As the rather straight dotted line in the lower half of Fig. 2 illustrates, victims’ second-guesses appeared insensitive to the pronounced contextual variations across and within studies 1 and 2. People likely feel uncertain about what the social norms are and infer that the app-driven sharing of others’ data at least cannot be socially condemned.

Contrasting results across studies 1 and 2 yield additional contextual insights beyond app type and other specificity. Recall that another key difference between those studies was the concreteness of the scenario with study 2 providing a concrete app and an illustration of app permissions. The neutralization categories indicating the presence of a norm violation (3–5) were used more often than in study 1, in particular among offenders. While this may be a reflection of the different samples, it may also signal that a more concrete grasp of app-driven privacy infringements helps to highlight its moral doubtfulness.

Study 3: Examining Offenders’ and Victims’ Explanations of Each Other

Study 3 deepened the exploration in the following ways. First, it follows up on the possibility that a concrete and practical understanding of how the sharing happens may reduce its moral acceptability. We thus explored a situation in which participants understand that the sharing of others’ data results from a single avoidable click. In study 3, all participants saw a pop-up window asking them to accept or refute an app’s access to one’s contacts.

Second, study 3 addressed another potential shortcoming of studies 1 and 2. Thus far, offenders were asked to assume that they had just given away others’ data. Given the prevalence of the issue, this is an ecologically sound assumption. Yet, participants in the offender condition may not have considered this assumption to be an equally sound assumption when it comes to them personally. Offenders’ preparedness to take responsibility could have been confounded with their belief that they would not have actually engaged in the behavior. The design of study 3 ensures that offenders actually agree to share others’ data before being offered a chance to neutralize this behavior.

Finally, study 3 revisits the perhaps most puzzling finding of studies 1 and 2: victims’ assumption that offenders would declare the behavior as normatively acceptable. This finding is concerning because it suggests that victims may fail to hold offenders responsible, allowing the problem to continue unabated. We directly address the substance of this implication by adding two additional perspectives, namely the victim’s personal viewpoint, i.e., their propensity to hold the offender responsible, and the offender’s second-guesses of the victim’s viewpoint, i.e., offenders’ expectations of potential repercussions.

Sample and Procedure

Overall, 267 participants were recruited from the same representative panel and following the same criteria as in study 1 (51% female, 49% male, 0.4% diverse, age 18 to 77 years, M = 42.43, SD = 12.81). Study 3 employs a 2 (role: offender vs. victim) × 2 (perspective: self vs. second-guessing other) between-subject design. All participants were asked to name a specific other before reading about a fictitious new messenger app called “ChitChat”. Participants learned about the download process and saw a realistic pop-up message asking them to allow or deny the messenger app (ok vs. not allow) access to their contacts (see Appendix A for full materials). Participants in the offender condition had to click on one of these options. Participants in the victim condition were told that their specified close other (we inserted the specific name) had clicked “ok”. We oversampled the offender condition to ensure a sufficiently large sample. Participants who clicked “not allow” (34%) were redirected to the victim condition and told that their close other had clicked “ok”.Footnote 3 Next all participants were told that clicking “ok” means that either their data (victim) or their close others’ data (offender) had been shared with the app. All participants next saw the same five statements as in study 1. Depending on condition, they either picked the statement that best reflected their own view of what had happened (self) or the statement that best reflected their guess of their close others’ view (second-guessing). The wording was adjusted to fit to make it sound natural for all roles and perspectives.

Results and Discussion

Figure 3 summarizes results across conditions. The box in the middle summarizes responses (own neutralizations and second-guesses) of offenders and victims, respectively. The box at the bottom contrasts people’s own responses with their second-guesses (regardless of role). To capture the relative importance of roles and perspectives, we again ran a multinomial logistic regression predicting the categories of neutralization used (see Table 2) and followed up with chi2-tests comparing specific distributions. Results replicate an earlier key observation. While a clear majority of participants (78%) considered the behavior a norm violation themselves, this verdict was significantly less frequent during second-guessing (58%; χ2 (1) = 10.91, p = 0.001). This finding is in line with the theorizing behind proposition 1 and shows that the difference in perspective emerges for both offenders and victims.

Fig. 3
figure 3

Note Please note that in some cases percentages do not add up to 100 percent exactly due to rounding

Distribution of categories of neutralization per cell for study 3.

Another aim of study 3 was to examine the match between victims’ second-guesses of offender justifications and their propensity to blame offenders. A chi2-test contrasting the distributions of categories of neutralization across these conditions yielded no significant difference (χ2 (4) = 5.37, p = 0.252). However, if we only look at the fault line of norm violation, we find that participants in the victim-self condition were more likely to see the behavior as a norm violation than they second-guessed offenders would do so (categories 3–5: 83% self vs. 69% second-guessing, χ2 (1) = 4.19, p = 0.041). As Fig. 3 indicates, this difference was driven by neutralization category 4. Perspective made no difference to victims’ propensity to hold the offender fully responsible (23% self vs. 22% second-guessing).

Study 3 thus reconfirms that participants in the offender-self condition are more likely to accept full responsibility than participants in any other condition. Given that all offenders in study 3 had freely decided to commit the offence, this finding suggests that the app-driven sharing of others’ data feels personally wrong for many offenders. In addition, results show that victims consider the behavior overall less acceptable than their second-guesses about others might have indicated.

Study 3 also aimed to explore what happens when victims and offenders realize that the preventable click of a single button can make the difference to infringing privacy or not. Except for highlighting that it all hinges on a single click, study 3 had been equivalent to the communication app condition of study 1. Comparing these two studies suggests the following: realizing the power of a single click seems to increase offenders’ propensity to accept full blame (42% vs. 28% in study 1) while reducing their propensity to claim that nothing happened (12% vs. 31%). We also see differences across studies in the victim second-guessing condition. Compared to study 1, victims were less likely to assume that offenders would think that nothing had happened (18% vs. 43%) and somewhat more likely to second-guess that offenders would take responsibility (22% vs. 14%). Realizing that the offender had chosen to click “ok” to share data rather than “not allow” seems to make a difference. This practical understanding seems to curb victims’ willingness to excuse the behavior and may thus help the development of more prohibitive social norms.

There was, however, one condition that mapped exceedingly well onto victims’ second-guesses in studies 1 and 2: offenders’ second-guesses about victims’ views, i.e., the only condition in which people may not have paid attention to the role of this single click. The distribution of categories of neutralization in this condition is virtually identical to the distribution of the overall victim condition in the prior studies (see Fig. 2). When second-guessing, offenders seem to have drawn on the same abstract (rather than practical) understanding of the sharing of others’ data as “normal” that victims in studies 1 and 2 appear to have drawn on.

Discussion and Implications

Allowing apps access to contacts or photos amounts to sharing others’ data and this potentially harmful behavior is morally and legally questionable. Prior research suggests that people infringe on others’ information privacy even if they are aware of this implication (Kamleitner & Mitchell, 2019). However, despite the behavior being widespread, offenders are hard, if not impossible, to identify. This makes it difficult to determine the behavior’s overall moral acceptability. In addition, most offenders are prone to be both offenders and victims and their moral verdict may differ across these roles. We therefore drew on offenders’ neutralizations and victims’ second-guesses of these neutralizations in order to capture the nuanced moral acceptability and evolving norms guiding and enabling the behavior. Building on this novel theoretical backbone, we conducted four exploratory studies (one qualitative and three quantitative) and reached the following key insights.

  1. a.

    The app-driven sharing of others’ data is morally ambiguous. The two opposite extremes ‘nothing happened’ and ‘something bad and inexcusable happened’ dominated the responses.

  2. b.

    These opposing answers result from the juxtaposition of offenders’ own perspective with victims’ second-guesses about offenders’ perspective. Offenders tended to take more responsibility than victims either allotted to them (study 3) or expected them to take (studies 1 and 2). One way to interpret these findings is to suggest an absence of clearly established social norms in light of more prohibitive personal norms.

  3. c.

    Neutralizations vary across contexts, but these variations are remarkably small and mostly arose in the offender perspective. We deduce from this that people’s personal norms are sensitive to contextual variations while their social norms are largely robust to context. While cultures could be reasonably expected to differ (Licht, 2008), the pattern of results is remarkably similar across our samples in Austria and US.

  4. d.

    A practical understanding of the ease with which others’ data can be protected as well as given away seems to increase the propensity to consider the behavior as a norm violation, in particular among victims.

Contributions to Theory

First, our insights align with prior literatures suggesting that people are struggling to understand online privacy (Kokolakis, 2017; Oetzel & Gonja, 2011), that digitalization is shifting norms (Bauman, 2013; Rosa, 2013), and that the online environment often negatively affects ethical judgements (Freestone & Mitchell, 2004) by creating an unreal environment free from offline norms (Runions & Bak, 2015). We find that these observations also extend to app-driven interdependent privacy. Prior literature suggests that when agreed social norms become violated, victims allot more responsibility to offenders than these offenders tend to be willing to take themselves (McCarthy et al., 2021). Our exploratory findings (see also Fig. 4 in Appendix B) suggest the absence of clear social norms for the app-based sharing of others’ data. In this situation, we find the opposite: victims are less accusatory than offenders. A lack of clear social norms may explain this, but so does the fact that most victims are also offenders. Both considerations have scarcely been examined in current theorizing.

Second, we add a new theoretical lens, neutralizations, to existing thinking in the area of interdependent privacy. To the best of our knowledge, we are the first to focus on the moral dimension of the phenomenon. Traditionally, the study of neutralizations has centered on understanding individuals’ justifications for deviating from these norms rather than delving into the norms themselves. We show how the very process of neutralization could serve as an indicator of the evolving social norms. Our results also shine a novel light on how much people respect others’ data; something which Kamleitner and Mitchell (2019) consider the third and last step in a series of steps (the 3Rs; Realize, Recognize, Respect) that can lead to the sharing of others’ data. Our research confirms their speculation that the disrespect of others’ rights may represent a lack of inhibiting social norms.

Third, we expand on the literatures on neutralizations (Kaptein & van Helvoort, 2019; Sykes & Matza, 1957; Willison & Warkentin, 2013) and social norms (Acquisti et al., 2011; Barth et al., 2006; Hoyle et al., 2020; Nissenbaum, 2004). We find that people may be willing to admit to wrongdoing without making excuses and suggest the addition of a “no neutralization” category. This ensures that there is a fitting answer option for all situations in which a social or personal norm gets infringed. By assessing neutralizations at the level of categories, we also highlight that it may be possible to compare neutralizations across behaviors.

Finally, and perhaps most novel, we contribute to the neutralization literature by introducing perspectives. We draw attention to the possibility of victim–offender entanglement and highlight the merit of assessing second-guesses. Although this needs further investigation, we assume that the contrast between offenders’ own justifications and victims’ second-guesses about these justifications highlights the extent to which prohibitive personal and social norms exist and align. Whenever social norms are stricter than offenders’ personal norms, we would expect victims to second-guess the use of higher categories of neutralization than offenders actually do. Whenever social norms are more relaxed than offenders’ personal norms, we would expect the opposite. The interactive process of offenders’ and victims’ perceptions and the ongoing dialog between the two parties play a pivotal role in shaping and solidifying new societal norms surrounding interdependent privacy concerns.

Practical Implications

Our empirical findings pave the way for different courses of action and recommendations. First, they suggest a lack of moral consensus about the behavior. This makes it unlikely that the public will be prepared to press policy makers and app providers into bringing about change. Although the privacy-infringing behavior (currently) violates many individuals’ own moral standards, it is likely to continue. This implies the need for clearer regulation and actual legal sanctions. Greater legal clarity about the lawfulness of this behavior and its legal ramifications for individuals and companies would assist in the development of norms and norm-breaking acknowledgments.

In the absence of clearer, stricter, and factually executed regulations, social norms are a key lever. These could be changed via some form of mass public education campaign such as has been successful at establishing preventative social norms around drinking and driving. Government is the primary stakeholder able to launch such a campaign together with information commissioners, which could also be supplemented by privacy NGOs.

While descriptive norms are often stronger to guide intention and behavior than injunctive norms (e.g., Elgaaied-Gambier et al., 2018), this is not helpful when thinking about the sharing of others’ data which is widespread. Making people aware of its prevalence could serve to increase the behavior. Our results also suggest the absence of injunctive social norms. This leaves a focus on prohibitive personal norms that can matter more for actual behavior (Hornsey et al., 2007) than social norms (de Groot et al., 2021). We therefore propose interventions that stress personal norms and, for example, invoke the parallel to offline privacy.

Another approach could be the use of so-called dynamic norms which can include perceptions about expected future developments in relation to the behavior (Loschelder et al., 2019), e.g., “more and more people are protecting their friends’ personal data”. Yet another approach to combat a “nothing happened” mentality is fear of repercussions, although prior literature suggests that these appeals need to be particularly well crafted to work (for specific recommendations see Tannenbaum et al., 2015).

There are also implications for businesses. Not collecting data shared via others avoids any potential prosecution under GDPR and can be a business opportunity. Recent work suggests that promoting privacy friendliness can lead to robust positive changes in market shares and revenues (Eggers et al., 2022). In particular, our results suggest that businesses could get a competitive advantage by helping offenders avoid breaking their own norms, e.g., via pop-up messages on app permission reminding users of their responsibilities.

Notably, most people are both offenders and victims. Traditionally interventions focus on offenders. Our results suggest that the victim perspective gives policy makers, app providers and privacy activist groups another angle for action. Because of the high degree of victim–offender overlap, interventions could target the same person with different appeals. When targeting the victim perspective, messages could challenge the assumption that offenders will think nothing happened. Also, from study 3, we see that a practical understanding of the ease with which others’ data can be protected as well as given away seems to increase the propensity to consider the behavior as a norm violation, particularly among victims. This could be made salient.

Although neutralizations vary across app contexts, these variations are remarkably small and mostly arose in the offender perspective. Nonetheless, the small differences could still have implications. Specifically, offenders were more likely to take responsibility when the app could plausibly be believed to fully function without accessing others’ data. In particular for the communication app, messages might try to undermine offenders’ defence that nothing happened.

Our results also suggest that making the other salient might play a role. Paradoxically, offenders were twice as likely (4% vs. 8%) to claim nothing happened when the victim was specified compared to when they were not specified. The most pronounced difference emerged in victims’ use of category 5 (take full responsibility). Some victims thought a specified offender they knew well would take full responsibility (18%), but none thought that unspecified offenders would. Specifying the other after it happened thus appears to be of limited use and might even backfire.

Conclusions and Further Research

App-driven privacy infringements not only differ from sharing one’s own data, but also from sharing others’ data offline, which is much more conscious, traceable, punishable, and far less frequent. We argue that a deeper understanding of the moral acceptability of app-driven privacy infringements is an important piece in the puzzle. Yet, our results are a mere first step in exploring a facet of interdependent privacy and we consider some future directions that appear to show promise. One way forward is to extend theorizing and benefit from other theories that can address the learning of norms such as differential association (Matsueda, 2001) or social learning theory (Bandura & Walters, 1977).

Looking more broadly at the phenomenon, one of its most resounding characteristics is its embeddedness in layers of ambiguity. Since personal data are retained even when passed on (Kamleitner & Mitchell, 2018), and multiple copies may exist without a person knowing, there is uncertainty around the true culprit in case of subsequent harm, and it lowers potential levels of detection and reduces punishment for potential offenders (Freestone & Mitchell, 2004). A focus on variations in ambiguity of data ownership and potential punishments or consequences (Demmers et al., 2022; Kamleitner & Mitchell, 2018) is a promising avenue for future research.

This brings us to a final noteworthy observation, the fact that so many offenders indicated they would take responsibility. While this sounds positive, it does invite more questions. For example, it is possible that taking responsibility was considered an easy option with no repercussions. In light of this, future studies could also consider two independent axes when examining the sharing of others’ data, namely the degree of responsibility taken and the degree of infringement.Footnote 4

Finally, it is important to acknowledge that interdependent privacy breaches are a broader set of activities. The app-based sharing of others’ data is but one of multiple instances of interdependent privacy issues. There are issues of social cost and benefit considerations which mostly manifest when people actively and visibly share others’ data, e.g., when sharing others’ pictures online (Litt & Hargittai, 2014). In addition, there are issues of co-created data, such as joint conversations or pictures taken together to consider. There are also considerations of (economic) personal benefits. Stretching the scope to consider the many facets of interdependent privacy paves the way for rich additional considerations. For example, referral programs are a great example of the active, conscious, limited and targeted giving away of others’ data (typically an email address), which differs greatly from the passive mass transfer of all contacts data that describes app-based privacy infringements. Our insights are relevant to many contexts but there is much more to be learned about interdependent privacy.