In the past two decades, the growing emphasis on the behavior of autonomous vehicles (AVs) has had an important influence on moral judgment investigation. Interest in this technology escalated quickly with the vision of the upcoming revolution in transportation, which will most likely induce a gradual reduction of human decision-making abilities in driving (Society of Automotive Engineers [SAE], 2021). Researchers have made arguments regarding the advantages and challenges of autonomous transportation (Elliott et al., 2019; Fagnant & Kockelman, 2015; Martinez-Díaz & Soriguera, 2018; Maurer et al., 2016; Rahwan et al., 2019), and several studies have focused on individuals’ attitudes and motivation regarding this new technology (Haboucha et al., 2017; Jobin et al., 2019; Othman, 2021). Among these factors, the perceived risk of losing control of driving operations is considered a significant psychological roadblock to the adoption of this technology (Shariff et al., 2017), especially considering how AVs will manage possible harm in dangerous situations. This obstruction to the full endorsement of AVs has been treated as a moral issue that overcomes autonomous transportation’s reported advantages (Meyer et al., 2017), and researchers have explored the ethical concerns regarding human-independent decisions (Hidalgo, 2021).

Based on its suitability, the sacrificial dilemma has been widely considered a flexible tool in the investigation of moral judgments of AVs’ behavior (Martí-Vilar et al., 2021; Unger, 1996). Traditionally, moral dilemmas have been used to compare two opposing moral doctrines: utilitarianism, aimed at minimizing harmful consequences (Bentham, 1781), and deontologism, aimed at adhering to categorical norms and duties (e.g., “The ends never justify the means”; Kant 1785). The trolley problem is the most popular example of a sacrificial dilemma (Foot, 1978), depicting a brakeless running trolley that is about to run over five track workers. The only way to save the workers is by pulling a lever that will divert the trolley onto a secondary track, where only one worker will be sacrificed. Trolley-like problems are described as incidental (Lotto et al., 2014), where harm is permissible only as a foreseen but unintended side effect to save the largest number of people, on the basis of Aquinas’s (1952) doctrine of the double effect. On the contrary, when harm is intentionally caused in pursuit of a greater good, the dilemma is defined as instrumental, an example of which is the footbridge problem (Thomson, 1985). Here, the utilitarian moral code is respected only by physically pushing a large man off an overpass to stop the trolley with his body. In this context, the unwillingness to endorse the utilitarian behavior is traditionally interpreted as non-utilitarian choice, reflecting the prohibition of a personal moral violation despite its utilitarian value (Gleichgerrcht & Young, 2013; Greene et al., 2004; Cushman et al., 2012). In the dual-process model framework (Greene et al., 2001, 2008), intentional and personal harm typically elicit non-utilitarian judgments as a consequence of a stronger emotional response (e.g., the footbridge problem).

Incidental and instrumental moral dilemmas can be structured as sacrificial dilemmas, in which at least one life must be sacrificed to fulfill the selected moral code (Kahane 2015). In time, sacrificial dilemmas have been applied to a wide variety of barely realistic circumstances (e.g., Greene et al., 2001; Moore et al., 2008), leading to questioning their ecological validity (Bauman et al., 2014; Gold et al., 2014; Kahane et al., 2018). Additionally, traditional dilemma sets have been developed assuming moral judgment to follow a ‘structure-based’ interpretation rule (e.g., incidental vs. instrumental), which results in downplaying the importance of contextualization (Schein, 2020). Although a ‘structure-based’ moral reasoning is widely endorsed in the relevant literature (Greene et al., 2001; Lotto et al., 2014; Moore et al., 2008), several studies demonstrated that more plausible and lifelike storylines may actually prompt utilitarian moral reasoning (Bruno et al., 2022; Körner et al., 2019).

Sacrificial dilemmas’ reliability has been intensively debated in recent years (Bartels et al., 2014, Bauman et al., 2014), highlighting scenarios in which characters’ lives are in jeopardy. Traditionally, when a sacrificial act saves the life of the moral agent or those of other people (self-involvement dilemmas), the preference for self-protection has always been assumed as part of the utilitarian option (e.g., “Should you kill this man to save yourself and the other five people?”; The burning building dilemma; see Moore et al., 2008).

Utilitarian judgment of AVs’ moral behavior

The experimental deployment of sacrificial dilemmas has grown exponentially in recent years because of the intensive investigation of the cognitive and emotional basis of human morality (Awad et al., 2020; Cushman et al., 2006; Greene et al., 2001, 2008; Sarlo et al., 2012). In this framework, applying the sacrificial and self-involving version of the trolley problem in the driving context seems straightforward (Bruno et al., 2022), and in numerous studies, researchers have opted to adopt its basic structure—specifically readapted—following different approaches. Indeed, this version of a moral dilemma has been used to investigate the association between morality and AVs’ behavior in immersive driving simulators (Frison et al., 2016; Samuel et al., 2020) in virtual reality (VR) settings (Faulhaber et al., 2019; Kallioinen et al., 2019; Riegler et al., 2021; Sütfeld et al., 2019, 2017) and—mainly—responding to image- or text-based moral scenarios (Bonnefon et al., 2016; Huang et al., 2019; (Martin et al., 2021a). Among some variations (e.g., number of characters involved, decision maker’s perspective, and risk level), the autonomous version of the trolley problem typically proposes an AV that is driving n passenger(s) and unexpectedly encounters m pedestrians who are crossing the road (n < m). The only way to avoid the accident—following the utilitarian moral code—is for the AV to steer off the street suddenly, sacrificing its own passenger(s). Traditionally, the alternative solution—always interpreted as non-utilitarian—is for the AV to continue straight, sacrificing the pedestrians and protecting its passenger(s). Usually, in the autonomous-driving version of the sacrificial dilemma, the sacrifice of the moral agent as the AV’s passenger occurs upon enactment of the proactive utilitarian option (e.g., swerving off to the side of the road, where the car will impact a barrier, killing the passenger/s but leaving the pedestrians unharmed; Bonnefon et al., 2016), which differs from traditional sacrificial dilemmas, in which the agent’s life is usually jeopardized in the non-utilitarian counterpart (Greene et al., 2001, 2004; Lotto et al., 2014; Moore et al., 2008).

In the development of AVs’ moral investigation through text-based dilemmas, the seminal work of Bonnefon et al. (2016) lit the fuse. The authors observed that the support of utilitarian AVs for the minimization of overall harm had decreased at the individual level, where participants preferred self-protective AVs for themselves. Importantly, the Moral Machine project (Awad et al., 2018) massively investigated moral preferences regarding AVs from a global and cross-cultural perspective, confirming the preference for sparing the largest possible number of lives and assessing the moderating role of other important factors (e.g., age of lives, whether lives are human, and whether road users’ behavior is lawful). De Melo et al. (2021) demonstrated the mediation effect of total perceived risk and other drivers’ behavior on the likelihood of utilitarian AV maneuvers, confirming the role of probabilistic outcomes in the resolution of moral dilemmas (Bazerman & Greene, 2010). Moreover, a limited number of studies have confirmed that decision-makers’ perspectives effectively influence moral judgment (Huang et al., 2019; Kallionen et al., 2019, Mayer et al., 2021). Interestingly, following the veil of ignorance (Rawls, 2009) reasoning seems to favor utilitarian resolutions, decreasing the inconsistency between moral judgment and willingness to buy utilitarian AVs (Martin et al., 2021a, b).

Despite the relevant attention directed toward the investigation of moral judgments related to autonomous technology, very few studies have focused on traditional human driving or the comparison with its autonomous counterpart. Bruno et al. (2022) detected relative ease in the endorsement of utilitarian behavior in human-driving dilemmas when compared to traditional sacrificial scenarios (e.g., Lotto et al., 2014), interpreting the plausibility of the on-road context as a facilitator in moral judgments. Li et al. (2016) recognized the utilitarian moral code as the default norm for human and autonomous driving vehicles, with a stronger expectation of utilitarianism in autonomous agents (Malle et al., 2015) but ascribing less responsibility to them in the case of mistakes. This utilitarian expectation of AV behavior has also been recently confirmed in VR settings (Kallionen et al., 2019). In terms of attribution of moral responsibility, McManus and Rutchick (2019) detected a positive relationship between agency and blame: In the case of negative consequences, AVs are considered less blameworthy than human drivers—because they cannot act deliberatively (Pizarro et al., 2003). Nonetheless, evidence on the allocation of moral responsibility to Artificial Intelligence (AI) systems still appears somewhat contrasting. Hong et al. (2020) observed a higher level of blaming towards AI drivers than human drivers, proportionate to the severity of the damage. Also, Bennett et al. (2020) found an inverse relationship between driving automation and human blaming in case of a road accident (at the expense of AV manufacturers), even though the attribution of final responsibility being mainly addressed to human drivers. Gill (2021) also investigated this topic, claiming a reduction in direct human responsibility when aboard an AV. Data confirmed this hypothesis, showing a reduction of frequency and moral permissibility of self-sacrifice solutions in a one-to-one AV dilemma, compared to human-driving dilemmas. Despite these preliminary results, the roles of moral responsibility and agency in autonomous and human-driving sacrificial dilemmas require further investigation, considering prior knowledge about the intentionality of harm (Cushman et al., 2006; Greene et al., 2004).

Emotions and moral judgment

Altogether, several individual and contextual factors are involved in moral judgment. Among them, emotion is probably one of the most widely discussed factors (Haidt, 2001; McHugh et al., 2022), despite its role in the mechanisms of moral cognition is still debated (Byrd & Conway, 2019; Crockett, 2013; Cushman, 2013; Greene et al., 2016) and remains somewhat unclear (Huebner et al., 2009; Landy & Goodwin, 2015). In the evaluation of a moral issue, individual moral behavior is insidiously influenced by moral emotions, which serve as mediators between individual moral principles (i.e., norms and conventions) and moral decisions (Tangney et al., 2007). Moral emotions arise because of daily events that motivate people to engage in—or avoid—righteous or wrong moral actions (Kroll & Egan, 2004), linked to the interest of one or more individuals other than themselves (Greenbaum et al., 2020; Haidt, 2003). Importantly, moral emotions can affect a moral agent before the actuation of the decision—during the evaluation of the potential alternatives (Tangney et al., 2007).

Moral emotions can be divided into two main categories: self-conscious emotions (e.g., shame and guilt) and other-condemning emotions (e.g., anger and disgust; Haidt 2003). In the first case, negative feelings are directed to the self in the form of self-evaluation in the violation of moral standards (Gehm & Scherer, 1988; Tangney & Dearing, 2002). On the one hand, shame is conceived as a public-oriented emotion (“What others will think of me?”; Buss 1980) and involves a global negative evaluation of the inner self and the event as objects of disapproval. On the other hand, guilt is perceived as a private-oriented emotion (“I did a bad thing”) condemning the negative behavior only and not the self as a whole (Lewis, 1971). Shame seems to be a more powerful moral experience than guilt (Behrendt & Ben-Ari, 2012), and empirical evidence suggests that although shame leads to defensiveness and distance, guilt promotes constructive responses (Tangney et al., 2007). In contrast, anger and disgust are the two main other-condemning moral emotions, depicted as negative feelings in response to a third party’s moral violation. Anger takes the form of indignation regarding mistreatments of and injustices affecting the self or others, spawning an immediate response against the immoral event (Hutcherson & Gross, 2011). In contrast, disgust represents a repulsion that arises against filthy moral conduct and seemingly correlates with the moral judgment’s severity (Huebner et al., 2009; Rozin et al., 1999; Schnall et al., 2008). These two emotions are elicited by different cues in the moral context (Russell & Giner-Sorolla, 2011), and like self-conscious emotions, they are experienced when the moral decision is made and when it is observed (Haidt, 2003). Referring to self-referred and other-referred moral emotions may be useful in disentangling moral judgments of AVs’ behavior, assuming the intrinsic difference between a traditional human-driven (hands-on-the-wheel) vehicle or an autonomous (hands-off-the-wheel) vehicle and when considering how the decrease in agency induces less responsibility attribution to and blaming of the AV’s harmful actions (Malle et al., 2014; McManus & Rutchick, 2019).

The present study

In the field of moral psychology, the investigation of the moral perception of AV technology is typically investigated through sacrificial moral dilemmas. In a typical sacrificial AV dilemma, when the moral agent’s life is at stake, the endorsement of the utilitarian option corresponds to the acceptance of the moral agent’s own sacrifice (i.e., “I die, but many survive”). This is an important structural difference from traditional self-involved sacrificial dilemmas (e.g., Greene 2001; Lotto et al., 2014), in which the endorsement of the utilitarian behavior matches the self-protective choice (i.e., “I live, and many survive”). In this context, we investigated the potential role of the agent’s sacrifice in the acceptance of the AV’s utilitarian behavior, by testing the following hypothesis:

H1: When the utilitarian moral behavior leads to the moral agent’s self-sacrifice (i.e., “I die, and many survive”), we expect a lower endorsement of the utilitarian behavior and a reduction of its moral acceptability, as compared to life-saving sacrificial scenarios (i.e., “I live, and many survive”).

Furthermore, the comparison between traditional human driving and autonomous driving has been often overlooked in the literature, especially when using text-based moral dilemmas. Several behavioral studies have investigated general attitudes, emotional activation, and moral perception in the evaluation of sacrificial AV dilemmas (e.g., Bonnefon et al., 2016), but no evidence has been collected in terms of differences from traditional manual driving. In this study, we aimed to compare moral judgments in moral dilemmas applied to these two fundamentally different modes of transportation, with the formulation of the following hypothesis:

H2: In line with the role played by lifelike storylines in the enhancement of utilitarian moral reasoning (Bruno et al., 2022; Körner et al., 2019), we predict a higher endorsement of utilitarian behavior in the case of human-driving moral dilemmas. At the same time, since the driving dilemmas share the same incidental structure – and considering the evidence in favor of a ‘structure-based’ interpretation of moral dilemmas - we expect the endorsement of the utilitarian moral code to be considerably high in both autonomous- and human-driving scenarios.

Finally, for the first time in the context of autonomous and non-autonomous transportation, we investigated the role of two self-referred (shame and guilt) and two other-referred (anger and disgust) moral emotions. In this context, we formulated the following hypotheses:

H3: A higher intensity of self-referred emotions will be reported after moral decisions concerning traditional human-driven vehicles than after those involving autonomous vehicles.

H4: A higher intensity of other-referred emotions will be reported after moral decisions concerning autonomous vehicles than after those involving traditional human-driven vehicles.

H5: Regardless of the level of automation, a higher intensity of self-referred emotions will be reported after the endorsement of self-protective behaviors, and a higher intensity of other-referred emotions will be reported after the endorsement of self-sacrificial behaviors.

Methods

Participants

Before analyzing any data, we tested a baseline equation, assuming a small effect size (Cohen’s d = 0.10) and a correlation of 0.50 among repeated measures, with a bidirectional hypothesis and an alpha error probability of 0.05 with 0.90 power, calculated with the G*Power statistical software (Faul & Erdfelder, 1992). The analysis suggested a minimum of 140 subjects, and we recruited 183 participants (94 women). Their mean age was 27.82 years (SD = 10.55, range: 18–66), and their mean education duration was 16.7 years of formal schooling (SD = 2.03, range: 11–24). Of the participants, 65.22% (n = 120) were enrolled in university courses, with 50.55% (n = 93) enrolled in human sciences degree programs (e.g., psychology or sociology). Most of the participants (90.21%, n = 166) held driver’s licenses, and most of them (75.54%, n = 139) drove a maximum of 15,000 km per year. Half of the sample (48.91%, n = 90) had been involved in at least one car accident in their lives, and only 4.35% (n = 8) had had a collision in the past 12 months. Of the participants, 87.5% (n = 161) stated that they had heard about AVs. Additionally, we administered the Positive and Negative Affect Schedule scale (PANAS) to assess the participants’ positive affect and negative affect during the 7 days leading up to the survey date (Terracciano et al., 2003; Watson et al., 1988). The participants had a mean positive-affect score of 31 (SD = 7.20) and a mean negative-affect score of 22.80 (SD = 8.03), below the normative thresholds and with no differences between men and women.

Stimuli

Following the structure of the validated set of Lotto et al. (2014), we designed 12 self-involvement sacrificial moral scenarios specifically for this study, adapted from the sacrificial human-driving set by Bruno et al. (2022). Coherently with the previous studies and based on the doctrine of double effect (Aquinas, 1952), we structured our dilemmas as incidental, interpreting the sacrifice as a predicted — albeit undesired — consequence that is mandatory to protect the highest number of lives. Specifically, the dilemma set comprised 12 moral scenarios (see Table 1 for samples) and is included in the supplementary material (https://osf.io/pb3xc/?view_only=4ae203cc39e24d68859da3f6b67591a5). The study was structured as a two-by-two repeated measures factorial design. In six of the scenarios, the participant was requested to identify with the actual driver of a traditional human-driven vehicle whereas in the remaining six, they had to identify with a passenger of a completely autonomous and self-driving car (Level 5 of automated driving; SAE, 2021). In the traditional car storyline, the participant alone was in charge of the driving decision, whereas in the AV scenarios they had to go with the vehicle’s decision.

In each dilemma, the driver or AV faced a particular traffic situation (e.g., overtaking a slow vehicle) with the participant and one other passenger on board. Suddenly, a critical problem arose from an unpredicted event, forcing the driver or AV to choose between the passengers’ safety and that of a larger number of pedestrians. We presented each moral dilemma as a textual description of the situation and two possible solutions: a utilitarian behavior and a non-utilitarian behavior. Consistently with the utilitarian doctrine, opting for the utilitarian maneuver always resulted in an active action (i.e., steering) aiming at safeguarding the highest number of characters and so minimizing the overall harmful consequences. Oppositely, the endorsement of the non-utilitarian option resulted in holding the current track, rejecting harm as an unintended side effect of collective welfare (Bruno et al., 2022; Cushman et al., 2012; Lotto et al., 2014).

To investigate the life-threatening factor’s effect on the moral agents’ endorsement of the utilitarian code, we further framed the dilemma set for this factor. With this aim, three scenarios per driving typology (n = 6) depicted the moral agent’s sacrifice in the utilitarian outcome (“I die, but many survive”; utilitarian sacrifice framing [USF]) whereas in the remaining six dilemmas, the non-utilitarian act involved the moral agent’s sacrifice (“I die, and many die”; non-utilitarian sacrifice framing [NSF]).

We controlled all 12 of the driving dilemmas for a number of factors that may play a role in the moral decision-making process (Awad et al., 2018; Bruno et al., 2022). We maintained a 1:2 ratio between lives saved and lives sacrificed, and we provided no additional information about or characterization of the other road users. To avoid a mediated allocation of responsibility, we avoided mentioning traffic rule violations and using leading language. Furthermore, we adopted scrupulous control of the number of words used to avoid intraindividual differences in reading times. The mean reading time was 31.31 s (SD = 63.75), and the mean decision time was 9.45 s (SD = 16.51).

Table 1 Sample Autonomous and Human-Driving Dilemmas (Text Translated from Italian)

Experimental procedure

Each participant signed an informed consent form before participation, which was voluntary and unremunerated. The local ethics committee approved the study (ID number 3514). We programmed and distributed the task as an online survey via Qualtrics software (Qualtrics, Provo, UT). To avoid device compatibility issues (Krebs & Höhne, 2021), we required the participants to complete the experiment using a laptop or computer.

At the beginning of the task, the participant had the opportunity to read and sign the informed consent form and then read the experimental procedure and instructions. Following the administration of the PANAS, the participants received a careful explanation of the dilemma presentation mode. At this point, the 12 moral scenarios were randomly administered to the participants. For each of them, the textual description remained on the screen for the entire time the participant needed to understand the situation. Then, the first alternative, randomly selected between the two, appeared on the screen. After 7 s, the second moral option appeared on the screen. Finally, after 7 more seconds, the participant selected their morally preferred outcome (see Bruno et al., 2022). After each moral dilemma, the participant had to evaluate the moral acceptability of the two proposed moral outcomes on an 8-point Likert scale (0 = completely immoral, 7 = completely moral). Subsequently, the participant had to rate the perceived intensity of the four moral emotions, two of which were self-referred (shame and guilt) and two of which were other-referred (anger and disgust). The evaluation referred to the present time, after the moral decision, and was expressed on an 8-point Likert scale (0 = no intensity, 7 = maximum intensity) (Fig. 1).

Fig. 1
figure 1

The Experimental Procedure of the Present Study. The Sequence Was Repeated 12 Times, One per Each Administered Dilemma. The Presentation of the Two Outcomes (A and B) was Randomized Between-Subject

Statistical analysis

We conducted the statistical analysis in the R environment (version 4.1.1). As a first step, we preliminarily tested data distributions for the variables of interest (fitdistrplus package in R; Delignette-Muller & Dutang 2015). To test the experimental hypothesis, we set the moral decision choices, the moral acceptability (for the utilitarian and non-utilitarian options), and the four moral emotions (shame, guilt, anger, and disgust) as dependent variables for the statistical investigation. We coherently fit one generalized mixed-effect linear model—for the dichotomous moral decision—and six linear mixed models to the data, setting the participants as random intercepts (lme4 package in R; Bates et al., 2015). We selected the models after a specific onward stepwise model selection procedure, considering in each case all the predictors of interest (moral decision, sacrifice framing, driving style, experimental order, gender, and car accidents experienced in the past) and their interactions, in the described order. Final models were selected accordingly to the Akaike Weights comparison procedure (Wagenmakers & Farrell, 2004). When needed, we conducted post hoc pairwise comparisons to investigate interlevel differences (emmeans package in R; Length, 2020). Tables 2 and 3 present the statistical analysis’s descriptive results. In the Results section and for each of the seven models, all the selected predictors have been listed coherently with their order of implementation. We include full, detailed predictors based on the seven model selection procedures as well as the complete data set and further detailed information on our statistical approach in the supplementary materials.

Results

The binomial distribution was set for implementing the generalized mixed linear model m1 on moral decisions (utilitarian, non-utilitarian). Following the model comparison, we set driving style (autonomous, manual), sacrifice framing (USF, NSF), and gender (female, male) as fixed effects, together with the interaction between driving style and sacrifice framing. Consistently with H1, when we framed the agent’s sacrifice in the non-utilitarian option (NSF), we observed a greater frequency of utilitarian behavior (χ21 = 88.97, p < .001). As expected in H2, we observed a slightly higher endorsement of the utilitarian behavior in the human-driving condition (78.71%) than with the autonomous version (74.59%; χ21 = 6.78, p = .009). Descriptively, in the USF, if an autonomous vehicle performed it, the utilitarian maneuver was selected less often (Fig. 2). Nonetheless, we observed no significant interaction between the two factors (p = .14).

Fig. 2
figure 2

Bar Chart of Moral Decision Percentage Frequencies, Divided by Driving Style (AV: Autonomous-Driving; Human: Human-Driving) and Sacrifice Framing (NSF: Non-utilitarian Sacrifice Framing, USF: Utilitarian Sacrifice Framing)

Focusing on the evaluated moral acceptability of the two proposed options (utilitarian and non-utilitarian), we fitted two linear mixed models to the data setting the moral acceptability of the utilitarian option (m2) and the moral acceptability of the non-utilitarian option (m3) as dependent variables. In both cases, the resulting final models considered the fixed effects of driving style, moral decision, sacrifice framing, experimental order, and gender, together with the interaction between moral decision and driving style. Participants rated the moral acceptability of both alternatives as low (Table 2), with lower scores for the non-utilitarian option. As expected, in the case of endorsement of the utilitarian behavior, we found the non-utilitarian option was less acceptable than its counterpart (χ21 = 162.48, p < .001; x̄ non−utilitarian = 2.02; x̄ utilitarian = 2.45). We observed no statistical effects of the driving style (p = .10) or the interaction with the moral decision (p = .78), but we detected a significant reduction in moral acceptability during the course of the experiment (χ211 = 43.53, p < .001). Interestingly, the participants rated the utilitarian behavior as more acceptable when it was coupled with the agent’s self-sacrifice (x̄ = 2.74, Fig. 3) than the self-protective utilitarian option (x̄ = 1.96; χ21 = 296.98, p < .001). We detected no significant effects when we set the non-utilitarian behavior as the dependent variable (m3).

Subsequently, we fitted four linear mixed models setting the two self-referred (shame and guilt) and the two other-referred moral emotions (anger and disgust) as dependent variables. In both the self-referred emotions cases (shame m4 and guilt m5), the model comparison procedure indicated to consider the models including moral decision, driving style, sacrifice framing, experimental order, and gender as fixed effects, as well as the interactions between (i) moral decision and driving style and (ii) moral decision and sacrifice framing. Interestingly, the results showed a similar trend between the two self-referred emotions after the moral decision, with overall consistently lower scores in the case of shame (see Tables 2 and 3). As expected in H3 on the intensity of self-referred moral emotions between levels of automation, shame and guilt were perceived as less intense after decisions in AV dilemmas (m4: χ21 = 23.43, p < .001; m5: χ21 = 44.84, p < .001), in USF dilemmas (m4: χ21 = 106.85, p < .001; m5: χ21 = 113.67, p < .001), and following the utilitarian moral decision (m4: χ21 = 52.91, p < .001; m5: χ21 = 36.05, p < .001). With reference to the hypothesized difference of intensity of self-referred moral emotions in presence of the endorsement of self-protective behaviors (H5), the moral decision showed a significant interaction with the sacrifice framing factor (USF, NSF), highlighting greater intensities of shame and guilt when one pursues self-protection, compared to the self-sacrificial option (m4: χ21 = 91.99, p < .001; m5: χ21 = 104.67, p < .001; Fig. 3). Additionally, men reported significantly lower emotion intensities than women in terms of shame (men: x̄ = 1.78; women: x̄ = 2.84) and guilt (men: x̄ = 2.65; women: x̄ = 4.00).

Finally, in both the other-referred moral emotion cases (anger m6 and disgust m7), the model comparison procedure indicated to consider the models including moral decision, driving style, sacrifice framing, and experimental order as fixed effects, as well as the interactions between (i) moral decision and driving style and (ii) moral decision and sacrifice framing.

We observed no effects in the anger m6 model, with the exception of a significant interaction between sacrifice framing and moral decision (χ21 = 13.65, p < .001), showing greater anger intensity when the moral decision corresponded to self-sacrifice (Fig. 3) and consistently with H5. Model m7 partially endorsed H4 on the intensity of other-referred moral emotions between levels of automation, showing only a significant effect of driving style on disgust intensity (χ211 = 7.30, p = .006), with lower scores in the case of human-driving dilemmas (human driving: x̄ = 2.41; AV: x̄ = 2.55). Furthermore, anger and disgust intensities increased during the experimental procedure (m6: χ211 = 37.20, p < .001; m7: χ211 = 11.66, p < .001).

Fig. 3
figure 3

Error Bars Plot Representing Means and Standard Errors of the Intensity of Other- and Self-Referred Emotions (Anger and Disgust, Shame and Guilt), Divided by Preferred Outcome (Self-Protection, Self-Sacrifice), Despite the Sacrifice Framing

Table 2 Mean and Standard Deviation (between brackets) of the Considered Dependent Variables, Divided by Driving Style (Human, Autonomous) and Sacrifice Framing (NSF: Non-utilitarian Sacrifice Framing, USF: Utilitarian Sacrifice Framing). For Each Dilemma Category, the Total Percentage of Utilitarian Decisions is Reported in the First Row
Table 3 Mean and Standard Deviation (between brackets) of the Considered Dependent Variables, Divided by Driving Style (Human, Autonomous) and Sacrifice Framing (NSF: Non-utilitarian Sacrifice Framing, USF: Utilitarian Sacrifice Framing). For Each Dilemma Category, the Total Percentage of Utilitarian Decisions is Reported in the First Row

Discussion

Sacrificial moral dilemmas are widely considered a widespread, flexible, and powerful instrument to investigate morality in autonomous transportation (Bonnefon et al., 2016; Huang et al., 2019; (Martin et al., 2021a). When facing life-threatening dilemmatic situations involving an AV’s passenger, the moral agent needs to face a non-utilitarian resolution that leads to their protection regardless of the number of consequential casualties, as well as a utilitarian resolution that allows for the preservation of the largest possible number of characters (e.g., pedestrians), but possibly at the expense of their own life (i.e., “I die, but many survive”). This tradeoff appears fundamentally different from the moral juxtaposition typically depicted in the traditional versions of sacrificial self-involvement dilemmas (e.g., Greene et al., 2001, 2004; Lotto et al., 2014; Moore et al., 2008). Indeed, in nondriving dilemmas, the moral agent can typically rely on a utilitarian resolution that provides the opportunity to protect the largest number of people and themselves (i.e., “I live, and many survive”). As compared to the AV dilemmas, such utilitarian outcome may mitigate the moral issue at stake, allowing for a resolution that is suitable at both the individual (i.e., “I can protect myself”) and collective (i.e., “I can also protect the largest possible number of people”) levels. We aimed to investigate this structural difference, focusing on the potential role of self-sacrifice in affecting the endorsement of utilitarian behavior when facing autonomous- and human-driving moral scenarios.

As expected (H1), when self-sacrifice was framed in the utilitarian options (USF), we observed a clear decrease in the endorsement of the utilitarian moral code, as compared to the non-utilitarian framing (NSF). From a merely rational or evolutionary perspective, this result seems intuitive (i.e., “I prefer to live rather than to die”; e.g., Petrinovich et al., 1993) but reveals the importance of accounting for the self-sacrifice framing in the moral investigation (Huebner & Hauser, 2011; Thomson, 2008). In this context, the role of self-protection in human morality has been slightly elaborated in the literature. Despite Haidt’s (2007) claims regarding the suppression of self-interest in moral reasoning, there is evidence that the moral agent places greater value on their own life than on a stranger’s life (Huebner & Hauser, 2011; Moore et al., 2008). Nonetheless, Sachdeva et al. (2015) showed that the endorsement of self-sacrifice in moral dilemmas is perceived as more morally praiseworthy than sacrificing a third person. Overall, the evidence we collected appears to fit this conclusion, in that participants clearly endorsed the utilitarian behavior mainly when it allowed to protect themselves but judged this outcome as more immoral than the self-sacrificial choice. Interestingly, albeit only descriptively, when self-sacrifice was framed in the utilitarian option (USF), more people endorsed this behavior in the traditional human-driven vehicle than in the AV. This trend deserves further investigation, deepening the role of direct vs. indirect agency in the evaluation of moral behavior (Gill, 2021) and user preferences for AVs (i.e., Awad et al., 2018; Bonnefon et al., 2016; Haboucha et al., 2017) when the moral agent’s life is at stake.

Additionally, considering the few studies focusing on moral judgment when human beings are at the wheel of a traditional human-driven vehicle (cf. Bruno et al., 2022), we checked for substantial differences in the endorsement of humans and autonomous drivers’ moral behavior. In support of our hypothesis (H2), we observed a higher endorsement of the utilitarian behavior in human-driving scenarios, compared to those with their autonomous counterparts. This result is in line with the evidence Bruno et al. (2022) provided on the advantage of lifelike moral situations in bringing out the utilitarian moral code and with the potential distortion of moral judgment caused by the description of highly implausible events (Körner et al., 2019). Here, an AV hitting the road and opting between two maneuvers may seem a clear example of an unlikely event. The AV technology is surely developing quickly, but its rise and implementation seem still distant, especially in the general public’s perspective (e.g., Guo et al., 2021). Indeed, the definition of AVs’ public acceptance and public perception remains a work in progress for experts and policymakers (e.g., Hilgarter & Granig 2020; Othman, 2021). In contrast to what observed in terms of moral decisions, no differences were obtained between levels of automation in moral acceptability of driving maneuvers, suggesting that controlling for the level of automation may have a role in endorsing a certain moral behavior, but not in evaluating its moral acceptability. This result seems coherent with the finding of low acceptability of utilitarian AVs reported in previous studies (Bonnefon et al., 2016; (Martin et al., 2021a), allowing for a preliminary generalization of this tendency to both autonomous and non-autonomous vehicles. Clearly, these results deserve further testing in order to gain additional insights into the differences underlying moral perception of autonomous and human driving. Nonetheless, the observed predominance of the utilitarian moral code in both driving conditions seems in line with a ‘structure-based’ interpretation of moral dilemmas, claiming for the independence of moral choice from the specific contextualization.

In the present study, morality in driving-based sacrificial and self-involvement dilemmas was also examined from an emotional perspective, by measuring the intensity of four moral emotions, two self-referred (shame and guilt) and two other-referred (anger and disgust), experienced after decision-making. Our findings suggest that the consideration of these moral emotions is a valuable integration of previous investigations on moral judgment. Indeed, the results confirmed our third hypothesis (H3), as self-referred moral emotions (shame and guilt) were reported as more involved in traditional human-driving dilemmas, when the moral agent is in charge of the vehicle maneuvers, physically controlling the wheel. Specifically, a higher intensity of guilt in human-driving dilemmas was predictable, as this moral emotion aims for constructive responses (Tangney et al., 2007) and is more specifically linked to personal transgressions in the moral realm (e.g., Sabini & Silver 1997; Smith et al., 2002). As expected, the participants perceived disgust - as a negative feeling in response to a third party’s moral violation - as more intense in response to critical moral events depicting AVs, partially confirming our hypothesis (H4) that other-referred moral emotions would be experienced with higher intensity in the AV- than human-driving dilemmas, and coherently with AI attribution of blame described by Hong et al. (2020). The inconsistency observed between disgust and anger is not surprising, as these moral emotions are known to be elicited by different cues in a moral situation (Gutierrez & Giner-Sirolla, 2007; Russell & Giner-Sirolla, 2011), with anger evoked by contextual cues of harm and intentionality, which are typical features of sacrificial and incidental dilemmas.

Interestingly, the decision to sacrifice or protect the self in the endorsement of the utilitarian behavior differentially affected the intensity of self-referred and other-referred moral emotions experienced after decision-making (H5). Indeed, the intensity of shame and guilt was higher when a self-protective decision was taken (i.e., “I live, and many survive”; NSF), as compared to a self-sacrificial decision (i.e., “I die, but many survive”; USF). Additionally, participants reported a higher intensity of anger when opting for self-sacrificial behavior (USF). This outcome is reasonable, since shame and guilt may be elicited by public and private selfish behaviors (Buss, 1980; Dillenberger & Sadowski, 2012; Gehm & Scherer, 1988), and injustice toward the self - or others - is able to trigger anger towards unspecified third-party in response to an immoral event (Haidt, 2003; Hutcherson & Gross, 2011). Overall, we can infer that people prefer to pursue utilitarianism while protecting the self, but this solution is perceived as more morally unacceptable, shameful, and blameworthy than the more praiseworthy self-sacrifice for a greater utilitarian goal (Sachdeva et al., 2015). This result seems consistent for manual and autonomous-driving vehicles, suggesting that the level of automation does not affect the evaluation of moral acceptability or the perceived intensity of moral emotions in sacrificial self-involvement dilemmas.

We acknowledge some limitations of the present study. Despite the experimental flexibility of text-based moral dilemmas, their application to on-road situations seems limited, especially in describing and deploying complex driving decisions and intricate traffic dynamics. Nevertheless, Sütfeld et al. (2019) compared VR-based and text-based AVs’ moral scenarios, confirming the reliability of abstract contextualization in comparison with the more ecological - and more expensive - VR assessment. In this sense, combining several approaches is advisable to enhance the reliability of results with the help of immersive and realistic traffic environments. In addition, in the present study only two levels of automation were investigated, Level 0 (no automation) and Level 5 (full automation), similar to the opposite poles of the SAE’s classification (2021). Our goal was to compare these two opposite means of transportation, in which the human has a completely different role in the selection of moral behavior. Nevertheless, autonomous driving features still require the driver’s active involvement, and future applications on AV’s morality may focus on more actual, intermediate levels of automation (e.g., Level 3), at which the human takeover (i.e., control of the vehicle being transferred from human driver to AV or vice versa) plays a key role in the allocation of driving responsibilities (Bellet et al., 2019). Finally, two features of moral alternatives’ structure are worthy of attention. First, to maintain a constant and realistic ratio between lives saved and lives sacrificed, we added to the scenarios an additional passenger who shared the moral agent’s fate. In several studies, research focused on the role, features, and number of characters involved in AV dilemmas (Awad et al., 2018; Bonnefon et al., 2016; Faulhaber et al., 2019), suggesting slight variations in terms of agreement with the utilitarian AV behavior or in the willingness to purchase this technology when more people are involved. Secondly, the non-utilitarian outcome in the USF dilemmas offered only partial accessibility, making explicit pedestrians’ but not passengers’ fate. Accessibility has been investigated in AV dilemmas leveraging on philosophical theories (Huang et al., 2019; (Martin et al., 2021a) and on the role of context in moral judgment (Schein, 2020). Kusev et al. (2016) revealed the role of full accessibility to the consequences of moral actions in boosting the endorsement of utilitarian behaviors, albeit only in the context of other-involvement scenarios. Further studies may confirm these results by manipulating the number of characters involved or controlling for the explicit availability of all the resulting consequences.

Overall, the present study has focused on self-sacrifice framing as a structural characteristic of sacrificial self-involvement dilemmas. We demonstrated that even if the pursuit of utilitarianism and self-protection is still conceived as the best possible outcome, obtaining a personal benefit from the utilitarian option is perceived as more immoral and shameful than the more praiseworthy self-sacrifice in the vision of the protection of the most. We collected this evidence by applying the moral dilemma tool to human vs. autonomous driving, detecting a preference for utilitarianism when the moral agent is in direct control of a traditional non-autonomous vehicle. Regarding the emotional impact of moral decision-making, take a stand in human-driving dilemmas seem to strongly elicit self-referred moral emotions (shame and guilt), indicating different emotional reactions to human- and autonomous-driving moral behaviors.

In conclusion, these results suggest taking into careful consideration the self-sacrifice factor in the development of sacrificial self-involvement dilemmas, as it seems to shape moral judgment and moral reactions to the utilitarian moral code. This finding is particularly important in the investigation of moral perception of driving behaviors, where, until today, the utilitarian behavior and the self-sacrifice factor have been combined differently than in typical sacrificial self-involvement dilemmas.