1 Introduction

Savage’s (1972) subjective expected utility (SEU) theory is one of the most important achievements in decision theory. It brilliantly combines a subjective interpretation of probability with von Neumann and Morgenstern’s (1944) expected utility theory laying a solid foundation for manifold theoretical developments in information economics, principal-agent theory, and portfolio and capital market theory. The basic idea is that preferences for decision acts are based on probabilities for decision-relevant events. From which considerations, if any, these probabilities follow is beyond the scope of SEU. For the decision-maker, however, the extent to which probability estimates are supported by event-specific knowledge can make a difference. Following this idea, Knight (1921) argued that confidence in probabilities plays a role in decision-making. Ellsberg’s (1961) famous thought experiments supported this view by suggesting that bets on events with well-specified probabilities are preferred over ambiguous bets where event-specific knowledge is not available. Numerous empirical studies have confirmed the relevance of ambiguity aversion for decision-making.

Ambiguity aversion implies a violation of Savage’s Sure Thing Principle, which is the element of SEU that ensures the additivity of the probability measure. Consequently, the Sure Thing Principle was substituted by weaker assumptions leading to Choquet’s expected utility (CEU) theory which applies nonadditive probabilities (capacities). For its axiomatization, Schmeidler (1989) used the Anscombe and Aumann (1963) approach assuming the existence of lotteries with exogenously determined probabilities. Gilboa (1987) applied Savage’s setup and thus established CEU on a purely subjective basis. Schmeidler (1986) showed that CEU conforms to the minimization of expected utility across probability measures in the core of the capacity when nonadditive probabilities are convex, which can be interpreted as a special kind of ambiguity aversion. However, a separation of ambiguity perception and ambiguity attitude is hardly possible, since CEU differs from SEU only in the nonadditivity of probabilities. More recent approaches attempt to overcome this problem.

As a close relative to CEU, Gilboa and Schmeidler (1989) axiomatized maxmin expected utility (MEU) theory, which postulates the existence of a set of probability measures from which the most unfavorable one is chosen. A refinement is the \(\alpha\)-maxmin expected utility theory (\(\alpha\)-MEU), introduced by Ghirardato et al. (2004), which uses a weighted mean of expected utilities based on the most favorable and the most unfavorable subjective probability measure. The weight can be interpreted as a coefficient measuring the decision-maker’s ambiguity aversion. An even better prerequisite for the separation of ambiguity perception and ambiguity attitude is provided by approaches that use second-order probabilities and second-order utility functions. Klibanoff et al. (2005) developed a representation where second-order expected utilities based on second-order probabilities (Smooth Ambiguity Preferences) are used to account for ambiguity. Nau (2006), Ergin and Gul (2009), Seo (2009), and Denti and Pomatto (2022) presented alternative axiomatizations. An issue is that second-order probabilities and second-order utilities are difficult to determine empirically.Footnote 1

This paper contributes to the literature by presenting a new representation of ambiguity-related preferences without nonadditive or second-order probabilities. The representation is computationally simple, entails SEU as a special case, explains all common types of Ellsberg phenomena, and offers a variety of possibilities for separating and measuring ambiguity perception and ambiguity attitude. Moreover, methods for the elicitation of the applied constructs are straightforward. The basic idea is that ambiguity results from a lack of knowledge that makes event-specific probability estimates impossible and requires the application of Laplace’s (1812) principle of indifference. At the core of the representation is an additive decomposition of probabilities \(P(A) = Q(A, A) + Q(A, A^{C})\) indicating to what extent the indifference principle is used. If P(A) is based entirely on event-specific knowledge, event A is perfectly unambiguous and \(Q(A, A^C)\) vanishes. If, on the other hand, there is no event-specific knowledge and P(A) is only based on the indifference principle, event A is perfectly ambiguous and P(A) is distributed proportionally between Q(AA) and \(Q(A, A^C)\). Complementary to the probability decomposition, the utility function of SEU is extended. u(xx) is the utility of outcome x when the underlying event is unambiguous. u(xz) refers to two alternatively possible outcomes x and z of one decision act and modifies the utility of these outcomes when there is ambiguity concerning the corresponding events. Combining the utility function with the probability decomposition results in the preference functional \(\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) Q(A_{i}, A_{j})\). With two possible outcomes x and z associated with perfectly unambiguous events A and \(A^C\), the preference value is \(u(x, x) Q(A, A) + u(z, z) Q(A^C, A^C) = u(x, x) P(A) + u(z, z) P(A^C)\). In the case of an ambiguous event and an ambiguity-averse decision-maker, the preference value is \(u(x, x) Q(A, A) + u(x, z) Q(A, A^C) + u(z, z) Q(A^C, A^C) + u(z, x) Q(A^C, A) < u(x, x) P(A) + u(z, z) P(A^C).\)

The approach is based on Savage’s framework, i.e., acts are modeled as mappings from states of the world to arbitrary outcomes. Hence, it differs conceptionally from approaches that use the Anscombe and Aumann (1963) framework or place restrictions on the outcome space. All probabilities are derived from preferences over acts. Thus, the approach also differs from those that use Savage’s setup but rely on exogenously given probabilities. Klibanoff et al. (2005), for example, use Savage’s setup but assume the existence of a subset of acts, called lotteries, that are measurable with respect to a partition of the state space over which probabilities are objectively given. Regarding the representation of preferences, the proposed approach uses different constructs (the probability decomposition and the extended utility function) and a different preference functional compared to CEU, MEU, \(\alpha\)-MEU, or Smooth Ambiguity Preferences. In general, it is a generalization of SEU but neither a generalization nor a specification of these approaches. Except for SEU-compliant preferences, it overlaps with CEU, for example, only under narrow conditions.Footnote 2 The distinction between perfectly unambiguous and perfectly ambiguous events finds some parallel in the approaches of Nau (2006) and Ergin and Gul (2009) who model the state space as the Cartesian product of two sets and use the resulting structure to distinguish different types of events. Here, in contrast, the state space remains unstructured. Differing from Sarin and Wakker (1992) and Epstein (1999), who use exogenously specified unambiguous events, all definitions needed for the characterization of perfectly unambiguous and perfectly ambiguous events refer to the preferences of the decision-maker. In this respect, parts of the approach are comparable to Epstein and Zhang (2001) and Zhang (2002), who also use preference-based definitions of unambiguous events. Preference-based definitions are introduced in this paper, however, for two classes of events, perfectly unambiguous and perfectly ambiguous events. Gul and Pesendorfer (2014) also use a utility function with two arguments, called interval utility, in their Expected Uncertain Utility Theory (EUU). In this theory, outcomes are modeled as prize intervals. The arguments of the utility function are the endpoints of these intervals. In contrast, the outcome space here is arbitrary and the utility function refers to two alternatively possible outcomes.

The proof of the central representation theorem is largely based on Savage’s proof of SEU. After setting the model framework, a preference order is established and perfectly unambiguous events (S-events) are defined as events for which the Sure Thing Principle applies. Next, partitions of the state of the world are considered for which the Sure Thing Principle does not apply but to which probabilities can nevertheless be assigned using the indifference principle. On this basis, perfectly ambiguous events (L-events) are introduced. Additional axioms ensure that S-events and L-events do not overlap except for null and universal events. This makes it possible to separate perfectly ambiguous acts (L-acts) and perfectly unambiguous acts (S-acts), and to replace outcomes with L-acts in Savage’s framework. Further parts of Savage’s axiomatic system are adapted, and on this basis, a representation theorem for simple S-acts with L-acts as consequences is stated (Theorem 1). Subsequently, a link between perfectly ambiguous and perfectly unambiguous acts is established, enabling L-acts with an arbitrary number of possible outcomes to be traced back to a risky mixture of L-acts with two possible outcomes. This forms the basis for a representation theorem on L-acts that is based on the extended utility function (Theorem 2). By merging Theorem 1 and Theorem 2, the final theorem (Theorem 3) is developed. It establishes the intended representation based on the extended utility function and the probability decomposition for all simple acts.

Concerning the separation and measurement of ambiguity perception and ambiguity attitude, the presented approach opens up various new possibilities. First, based on the probability decomposition, a measure of event ambiguity is suggested. It is shown that this measure satisfies intuitively plausible requirements for measures of event ambiguity proposed in the literature. To derive a measure of act ambiguity, it is useful to interpret the probability decomposition as a two-dimensional probability distribution (\(SL\)-distribution) and to utilize concepts from probability theory. Following this idea, an ambiguity measure for decision acts based on the covariance of the corresponding \(SL\)-distributions is introduced. Finally, a measure of ambiguity aversion is presented that is linked to the characteristics of the utility function in a way that is similar to Arrow’s (1971) and Pratt’s (1964) measure of risk aversion. Altogether, the proposed approach offers possibilities for separating and measuring ambiguity perception and ambiguity attitude that go beyond those in CEU, MEU, and \(\alpha\)-MEU and are comparable to those in approaches based on second-order concepts (Smooth Ambiguity Preferences). In contrast to these approaches, the proposed measures are based on “first-order concepts”, in particular on the probability decomposition and the extended utility function. Hence, the problems associated with the use of second-order probabilities and second-order utilities should not occur. Finally, it is also noteworthy that the proposed approach is the only purely subjective approach without restrictions on the state space and the outcome space that separates ambiguity perception and ambiguity attitude and allows for different degrees of ambiguity aversion.

The paper is based on a specific perspective on decision-making under ambiguity. While ambiguity in more recent approaches is mostly associated with the existence of second-order probabilities, ambiguity is related here to different possibilities of assigning probabilities. The idea is that in the case of perfectly ambiguous events, the decision-maker has no event-specific knowledge upon which to determine probabilities. In the absence of other possibilities, he or she must apply the indifference principle. The resulting probability is less reliable, however. It appears uncertain due to the lack of event-specific knowledge. Ambiguity is thus defined as a lack of event-specific knowledge leading to uncertainty about probabilities that can only be determined by the indifference principle. Note that the indifference principle is referred to as a method for calculating probabilities when no event-specific knowledge is available. Central to the notion of ambiguity is the absence of event-specific knowledge that forces the application of the indifference principle, rather than the application of the indifference principle itself.Footnote 3

The paper is organized as follows. In the next section, the probability decomposition, the utility function, and the representation of preferences are introduced. Subsequently, axioms are presented and representation theorems are derived. All relevant proofs are given in the appendix. The last part of the paper elaborates and illustrates ways to measure ambiguity and ambiguity attitudes. The paper concludes with remarks on special issues and future research.

2 Representation of Beliefs and Preferences

Let \((\Omega , {\mathscr {A}}, P)\) be a probability space with state space \(\Omega\), \({\mathscr {A}}\) the \(\sigma\)-algebra of decision-relevant events, and P a countably additive probability measure. Suppose \({\mathscr {A}}\) is generated by two \(\sigma\)-algebras \({\mathscr {S}}\) and \({\mathscr {L}}\) which contain perfectly unambiguous and perfectly ambiguous events, respectively. Then \(Q: {\mathscr {A}} \times {\mathscr {A}} \rightarrow [0, 1]\) is an SL-decomposition of P if the following conditions hold:

  1. (1)

    Adaptation:      \(Q(A, \Omega ) = P(A)\) for all \(A \in {\mathscr {A}}\)

  2. (2)

    Additivity: \(Q(\bigcup _{i=1}^{\infty } A_i, B) = \sum _{i=1}^{\infty } Q(A_i, B)\) for all pairwise disjoint \(A_i \in {\mathscr {A}}\) with \(i \in \mathbb {N}\) and all \(B \in {\mathscr {A}}\)

  3. (3)

    Symmetry: \(Q(A, B) = Q(B, A)\) for all \(A, B \in {\mathscr {A}}\)

  4. (4)

    Separation: \(Q(A, A^C) = 0\) for all \(A \in {\mathscr {S}}\)

  5. (5)

    Dispersion: \(Q(A, A^C) = Q(A, \Omega )\, Q(A^C, \Omega )\) for all \(A \in {\mathscr {L}}\)

  6. (6)

    Independence: \(Q(A \cap B, C \cap D) = Q(A, C)\, Q(B, D)\) for all \(A, C \in {\mathscr {S}}\) and \(B, D \in {\mathscr {L}}\)

It can be shown that an SL-decomposition always exists if \(P(A \cap B) = P(A) P(B)\) holds for all \(A \in {\mathscr {S}}\) and \(B \in {\mathscr {L}}\).Footnote 4 If an SL-decomposition exists, it is uniquely determined.

Based on Q, the probability of an event can be decomposed into additively combined parts that show the extent to which it is associated with event classes \({\mathscr {S}}\) and \({\mathscr {L}}\), respectively. For example, for an arbitrary event A with \(P(A)>0\), adaptation, additivity, and symmetry imply \(P(A) = Q(A, \Omega ) = Q(A, A) + Q(A, A^C)\). Using additivity, Q(AA) and \(Q(A, A^C)\) can be further decomposed in any desired way, leading to finer and finer decompositions of P(A). To see how Q informs about the ambiguity of event A, suppose \(A \in {\mathscr {S}}\). Separation yields \(P(A) = Q(A, A)\). Events with this property are called S-events because the existence of a subjective probability measure is implied by Savage’s axioms. For the characterization of S-events as perfectly unambiguous, the Sure Thing Principle is crucial. It allows the evaluation of one part of an act independently of the other, which requires a strict separation of the corresponding events based on event-specific knowledge. Without such knowledge, changing outcomes in one part of the act can cast a different light on the outcomes of the other part.Footnote 5 As a consequence, the Sure Thing Principle implies that the decision-maker considers the corresponding S-events to be perfectly unambiguous. Since \({\mathscr {S}}\) is an algebra, \(\Omega\) and \(\emptyset\) are S-events. If all events are S-events, i.e., \({\mathscr {A}} = {\mathscr {S}}\), there is no ambiguity.

By separation, \(Q(A, A^C) = 0\) holds for all \(A \in {\mathscr {S}}\). If, on the other hand, \(Q(A, A^C) > 0\), A cannot be an S-event and must be at least partially ambiguous. Ambiguity comes into play with event class \({\mathscr {L}}\). The elements of \({\mathscr {L}}\) are called L-events since, in the absence of any event-specific knowledge, Laplace’s indifference principle can be used to determine probabilities. Due to the lack of event-specific knowledge, L-events are considered to be perfectly ambiguous. Since \({\mathscr {L}}\) is an algebra, \(\Omega\) and \(\emptyset\) are L-events. If all events are L-events, i.e., \({\mathscr {A}} = {\mathscr {L}}\), ambiguity is perfect. Note that adaptation, separation, and dispersion jointly exclude the possibility of an event A with \(0< P(A) < 1\) that belongs to both, \({\mathscr {S}}\) and \({\mathscr {L}}\). Moreover, adaptation and independence imply that L-events are not informative about S-events and vice versa, i. e. \(P(A \cap B) = P(A) P(B)\) for all \(A \in {\mathscr {S}}\) and \(B \in {\mathscr {L}}\).Footnote 6

Considering the decomposition \(P(A) = Q(A, A) + Q(A, A^C)\) with \(P(A) > 0\) and \(Q(A, A^C) > 0\), event ambiguity is perfect for \(A \in {\mathscr {L}}\). In this case, adaptation and dispersion yield \(Q(A, A^C) = P(A) (1-P(A))\), indicating a uniform distribution of probability mass according to the indifference principle. If A is neither an L-event nor an S-event, \(Q(A, A^C) < P(A) (1-P(A))\). For example, if \(A = B \cap C\) with \(B \in {\mathscr {S}}\) and \(C \in {\mathscr {L}}\), the properties of the SL-decomposition imply \(Q(A, A^C) = Q(A, B {\setminus } A)\) with \(Q(A, B {\setminus } A) < P(A) (1-P(A))\).Footnote 7 All in all, Q(AA) reflects the unambiguous and \(Q(A, A^C)\) the ambiguous part of P(A). For an ambiguity-neutral decision-maker, the decomposition of P(A) into Q(AA) and \(Q(A, A^C)\) is irrelevant. An ambiguity-averse decision-maker, however, evaluates outcomes differently depending on the degree to which the underlying events are subject to ambiguity.

The SL-decomposition Q is complemented by a utility function \(u: X \times X \rightarrow \mathbb {R}\) where X denotes the set of outcomes, u is symmetric, and \(u(x, x) \ge u(x, z) \ge u(z, z)\) holds for all \(x, z \in X\) with \(u(x, x) \ge u(z, z)\). The arguments of u refer to alternatively possible outcomes of one decision act. The stand-alone utility of an outcome x is measured by u(xx). With different outcomes x and z, u(xz) captures additional preferential aspects measuring the decision-maker’s ambiguity attitude. As the main result of this paper shows, preferences concerning simple acts can be represented by the preference functional \(\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) Q(A_{i}, A_{j})\) where \(x_{i}\) denotes the outcome associated with event \(A_{i}\). According to the axioms presented in the next section, Q is uniquely determined, and u is unique up to positive linear transformations. The decision-maker follows SEU with utility function \(U(x_i) = u(x_i, x_i)\) for all \(x_i \in X\) if he or she perceives no ambiguity, i.e., \(Q(A_{i}, A_{i}) = P(A_{i})\) for all i and \(Q(A_i, A_j) = 0\) for all ij with \(i \ne j\) and, consequently, \(\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) Q(A_{i}, A_{j}) = \sum _{i=1}^{n} U(x_{i}) P(A_{i})\). An inspection of the preference functional shows that the same holds for \(u(x_i, x_j) = 0.5 u(x_i, x_i) + 0.5 u(x_j, x_j)\)Footnote 8 which reflects a neutral attitude toward ambiguity. In all other cases, the preference functional indicates deviations from SEU according to the ambiguity of the decision-relevant events and the ambiguity attitude of the decision-maker.

As the representation of beliefs and preferences is intended to model ambiguity-related decision-making, it is illustrated by considering the classical Ellsberg experiments.

Example 1

(Ellsberg, two urns) The decision-maker must choose between two urns, each containing 100 red or black balls. He or she knows from urn S that it contains 50 red and 50 black balls. The composition of balls in urn L is unknown. When a red ball is drawn, the decision-maker wins $1. The draw of a red ball from urn S is regarded as an S-event, while the draw of a red ball from urn L is interpreted as an L-event. Thus, \({\mathscr {S}} = \{\emptyset , \Omega , R^{S}, B^{S}\}\) and \({\mathscr {L}} = \{\emptyset , \Omega , R^{L}, B^{L} \}\).Footnote 9 The complete event algebra is \({\mathscr {A}} = \{\emptyset , \Omega , R^{S}, B^{S}, R^{L}, B^{L}, R^{S} \cap R^{L}, R^{S} \cap B^{L}, B^{S} \cap R^{L}, B^{S} \cap B^{L}, R^{S} \cup R^{L}, R^{S} \cup B^{L}, B^{S} \cup R^{L}, B^{S} \cup B^{L} \}\). Given the known distribution of red and black balls, the decision-maker’s subjective probability of drawing a red ball from urn S is 0.5. The subjective probability for a draw of a red ball from urn L is also 0.5. It is based, however, on the indifference principle. Using these subjective probabilities, the properties of the SL-decomposition imply the entries in Table 1. For example, \(Q(R^S \cap R^L, R^S \cap B^L) = Q(R^S, R^S) Q(R^L, B^L) = P(R^S) P(R^L) P(B^L) = 0.125\), where the first identity is implied by independence, \(Q(R^S, R^S) = P(R^S)\) follows from adaptation, additivity, and separation, and \(Q(R^L, B^L) = P(R^L) P(B^L)\) follows from adaptation and dispersion. For \(u(1,1) = 1\), \(u(0,0) = 0\), and \(u(1,0) = u(0,1) = 0.4\), the preference values are 0.5 for urn S and 0.45 for urn L. The decision-maker prefers urn S because he or she is ambiguity-averse and the draw of a red ball from urn S is a less ambiguous event. For \(u(1,0) = u(0,1) = 0.5\) the decision-maker is ambiguity-neutral and thus indifferent to urn S and urn L.

Table 1 SL-decomposition in Ellsberg’s experiment with two urns

Example 2

(Ellsberg, one urn) There is one urn containing 30 red balls and 60 black and yellow balls, the latter in unknown proportion. The decision-maker can bet on a draw of a red or a black ball. When a ball of the chosen color is drawn, he or she wins $1. According to the known number of red balls, the subjective probability of drawing a red ball is \(0.\overline{333}\). The probability of drawing a black ball is the same. However, since the number of black balls is not known, the indifference principle is needed for the determination of this probability. Table 2 displays the assumed SL-decomposition. For \(u(1,1) = 1\), \(u(0,0) = 0\) and \(u(1,0) = u(0,1) = 0.4\), the preference values for R and B are 0.\(\overline{33}\) and 0.30, respectively. The ambiguity-averse decision-maker prefers the bet on R since it is less ambiguous than the bet on B. For \(u(1,0) = u(0,1) = 0.5\), the decision-maker is ambiguity-neutral and thus indifferent to a bet on R and a bet on B.

Table 2 SL-decomposition in Ellsberg’s experiment with one urn

In this example, B cannot be interpreted as an L-event if R is assumed to be an S-event.Footnote 10 Hence, in order to calculate the SL-decomposition as in Example 1, suitable S-events and L-events must first be constructed. To this end, the situation is interpreted as follows: The considered urn contains 30 red and 60 non-red balls. If a red ball is drawn, event R occurs. Otherwise, the ball is replaced by a ball drawn from another urn with an unknown composition of 60 black and yellow balls. For this procedure, Table 2 can be calculated in the same way as Table 1. It should be noted, however, that the entries in Table 2 are basically not subject to calculations. The SL-decomposition is a subjective quantity derived from preferences. The above considerations only serve to clarify the concept of S-events and L-events. The decision-maker might interpret the events in question differently, or not worry about an interpretation at all.

3 Axioms and Theorems

The decision situation is modeled as in Savage (1972). \(\Omega\) is the state space with states of nature s, t. Events are subsets of \(\Omega\) and are denoted by A, B, or C. The decision-relevant events form the \(\sigma\)-algebra \({\mathscr {A}}\).Footnote 11X is the outcome space with outcomes x, y, z. Acts f, g, h are modeled as \({\mathscr {A}}\)-measurable functions. F is the set of all acts. f is a constant act if an outcome \(x \in X\) exists such that \(f(s) = x\) for all \(s \in \Omega\). Hence, every \(x \in X\) can be identified with a constant act. An act is simple if it has finite support. Throughout this paper, only simple acts are considered. \(f\,A\,g\,B\,h\) is the act that leads to the outcome of act f (g) when event A (B) occurs and to the outcome of act h otherwise. It is presupposed that A and B are disjoint. \((f\,A\,g)\,B\,h\) denotes \(f\,A \cap B \,g\,A^{C} \cap B\,h\). When this notation is used, there are no restrictions on A or B. \(F_{i=1}^{n} f_{i}\,A_{i}\) abbreviates \(f_{1}\,A_{1}\,f_{2}\,A_{2}\,\dots \,f_{n-1}\,A_{n-1}\,f_{n}\). Preferences are modeled by a binary preference relation \(\preceq\) on F.

The first axiom states that \(\preceq\) on F establishes a weak order:

Axiom 1

(Ordering) \(\preceq\) on F is complete and transitive.

Preference relations \(\succeq\), \(\prec\), \(\succ\) and \(\sim\) are defined in the usual way. \(A \in {\mathscr {A}}\) is a null event if \(f\,A\,h \preceq g\,A\,h\) for all \(f, g, h \in F\), as defined in Savage (1972). Axiom 1 implies \(\emptyset\) is a null event. The complement of a null event is a universal event, hence, \(\Omega\) is a universal event. \({\mathscr {N}}\) is the set of null events, \({\mathscr {U}}\) is the set of universal events. f is a certain act if an outcome \(x \in X\) and a universal event A exist such that \(f(s) = x\) for all \(s \in A\).

For its particular importance, the following definition is specially labeled:

Definition 1

(S-events) \(A \in {\mathscr {A}}\) is an S-event if \(h\,A\,f \preceq h\,A\,g \Rightarrow h'\,A\,f \preceq h'\,A\,g\) for all \(f, g, h, h' \in F\).

A partition of the state space into S-events is an S-partition. \({\mathscr {S}}\) is the set of all S-events. An \({\mathscr {S}}\)-measurable act is an S-act. The set of S-acts is denoted by \(F_{S} \subseteq F\). The definition of null events implies universal events are S events. Hence, certain acts are S-acts.

The definition of S-events reflects that outcomes associated with S-events do not affect the evaluation of outcomes that may otherwise occur. It is an adapted version of Epstein and Zhang’s (2001) definition of linearly unambiguous eventsFootnote 12 resp. Zhang’s (2002) definition of unambiguous eventsFootnote 13 but differs from that definition in two ways: First, it refers only to an event A, not to its complement \(A^{C}\). Here, a restricted form of Savage’s Sure Thing Principle ensures that \(A^{C}\) is an S-event whenever A is an S-event. Second, Epstein’s and Zhang’s definition refers to outcomes x and \(x'\) rather than acts h and \(h'\). The restricted Sure Thing Principle, however, is formulated in terms of acts, and consequently the definition of S-events is based on acts as well. Before discussing further aspects of Definition 1, the addressed axiom is introduced:

Axiom 2

(Restricted Sure Thing Principle) \(f\,A\,h \preceq g\,A\,h \Rightarrow f\,A\,h' \preceq g\,A\,h'\) for all \(f, g, h, h'\in F\) and all \(A \in {\mathscr {S}}\).

The only difference between Axiom 2 and Savage’s Sure Thing Principle concerns the event A which must be an S-event. Note, however, that S-events are not exogenously determined but introduced by definition and that f, g, h, and \(h'\) need not be S-acts as in Sarin and Wakker (1992).

Given an S-event A, Axiom 2 implies \(A^{C}\) is also an S-event. Moreover, Axiom 2 is sufficient to establish \({\mathscr {S}}\) as an algebra. However, Savage’s proof is based on a \(\sigma\)-algebra of events. To utilize this proof, it must be ensured that \({\mathscr {S}}\) contains countable unions. In addition, the countable additivity of the subjective probability measure is required for the proofs of Theorem 2 and Theorem 3 below. Both requirements are fulfilled by the following axiom:

Axiom 3

(Monotone Continuity) Let \(\{ A_{i} \}\) be an event sequence with \(A_{i} \in {\mathscr {S}}\) and \(A_{i} \subseteq A_{i+1}\) for \(i \in \mathbb {N}\). Then \(\bigcup _{i=1}^{\infty } A_{i}\) is an S-event and \(x\,A_{i}\,z \preceq x\,B\,z\) for \(i \in \mathbb {N}\) \(\Rightarrow\) \(x\,\bigcup _{i=1}^{\infty } A_{i}\,z \preceq x\,B\,z\) for all \(x, z\in X\) and \(B \in {\mathscr {S}}\).

Note that \({\mathscr {S}}\) is the set of S-events according to Definition 1 and not a primitive of the axiomatization. Therefore, Axiom 3 does not interfere with the setting specified in the prefix. Alternatively, the requirement that a countable union of S-events is an S-event could have been included in Axiom 2. It is stated here as part of Axiom 3 to separate behavioral and technical conditions. The second part of Axiom 3 establishes the monotone continuity of the probability relation and thus ensures the countable additivity of the probability measure on \({\mathscr {S}}\).Footnote 14 The use of technical conditions that ensure countable additivity has become common in decision theory since Arrow’s axiomatization of SEU.Footnote 15

The first part of Axiom 3 guarantees that \({\mathscr {S}}\) is not just an algebra but a \(\sigma\)-algebra, which is stated in Lemma 1 below.

Lemma 1

\({\mathscr {S}}\) is a \(\sigma\)-algebra.

The idea that \({\mathscr {S}}\) is a \(\sigma\)-algebra is of particular importance for the interpretation of S-events. Zhang (1999), Zhang (2002) and Epstein and Zhang (2001) argue that unambiguous events cannot form an algebra as it is not plausible to assume that the intersection of two unambiguous events is unambiguous in general. They illustrate this reasoning by an Ellsberg urn with a total of 100 balls of different colors R, B, Y, and W, where \(R + B = B + Y = 50\) is known. The authors term it intuitive that \(R\,\cup \,B\) and \(B\,\cup \,Y\) are unambiguous, whereas B is ambiguous, and, consequently, introduce the set of unambiguous events as Dynkin System. As every \(\sigma\)-algebra is a Dynkin System, the concept of S-events is more restrictive than Epstein and Zhang’s concept of unambiguous events when the respective context is taken into account. For a clear distinction, S-events are referred to as perfectly unambiguous events.

The difference is due to different research objectives. One objective of Epstein and Zhang (2001) is to find a basis for probability sophisticated preferences.Footnote 16 To this end, the idea of unambiguous events must be as inclusive as possible to provide the broadest possible basis for decision theories that assume given probabilities. A more restrictive concept would face the problem that effects are attributed to ambiguity for which other explanations have been found, e. g., nonlinear preference functionals.Footnote 17 The present paper, in contrast, aims at an explanation for ambiguity-related preferences by introducing a new representation formalism. Ambiguity is ascribed to a particular class of events, which are referred to as perfectly ambiguous. S-events, on the other hand, are completely unaffected by perfectly ambiguous events. Thus, they serve as an extreme limiting the range of possible ambiguity from below. For this purpose, Epstein and Zhang’s concept of unambiguous events is too less restrictive. To see this, consider a second urn containing 100 balls of types R, B, Y, and W. The decision-maker knows that there are 25 balls of each type. According to the reasoning of Epstein and Zhang, the draw of \(R\,\cup \,B\) from this second urn is an unambiguous event. However, the decision-maker has more event-specific knowledge in this situation and may therefore perceive a draw from the second urn as less ambiguous than a draw from the first urn. A preference for betting on \(R\,\cup \,B\) from the second urn, however, can not be explained using Epstein and Zhang’s concept of unambiguous events, since there is no way to distinguish between the two events in question.

Example 3

(Epstein and Zhang) With respect to the situation outlined above (first urn), the SL-decomposition is shown in Table 3. Regardless of the utility function, the ambiguity-averse decision-maker is indifferent to bets on R, B, Y, or W. All these events are equally ambiguous. For \(u(1,1) = 1\), \(u(0,0) = 0\) and \(u(1,0) = u(0,1) = 0.4\), the preference value for a bet on \(R_1\,\cup \,B_1\) is 0.475, while the preference value for a bet on \(R_1\,\cup \,Y_1\) is 0.45. \(R_1\,\cup \,B_1\) is more unambiguous compared to \(R_1\,\cup \,Y_1\) but not perfectly unambiguous. Hence, the decision-maker prefers to bet on \(R_1\,\cup \,B_1\). In contrast, the SL-decomposition for the second urn is given in Table 4. The events \(R_2\,\cup \,B_2\) and \(R_2\,\cup \,Y_2\) are perfectly unambiguous. The preference value for a bet on these events is 0.5 in both cases. The decision-maker prefers a bet on \(R_2\,\cup \,B_2\) to a bet on \(R_1\,\cup \,B_1\).

Table 3 Epstein and Zhang example (first urn)
Table 4 Modified Epstein and Zhang example (second urn)

The counterpart of S-events in terms of ambiguity are L-events, whose definition is based on L-partitions:

Definition 2

(L-partitions) A partition \(\{ C_{i} \}\) of \(\Omega\) with \(i = 1, \ldots , n\) and \(n > 1\) is an L-partition if \(f\,C_{i}\,g\,C_{j}\,h \sim g\,C_{i}\,f\,C_{j}\,h\) for all \(C_{i}, C_{j} \in \{ C_{i} \}\), all \(f, g \in F_{S}\) and all \(h \in F\).

Note that, by definition, all L-partitions are finite. This rules out the paradoxes associated with applying the indifference principle to an infinite number of events, e.g., the Bertrand paradox. The \(\sigma\)-algebra generated by L-partitions is denoted by \({\mathscr {L}}\). Elements of \({\mathscr {L}}\) are L-events, and elements of an L-partition with two events are called \(R_{L}\)-events. Null events and universal events are L-events as a modification of an L-partition with respect to a null event leads to another L-partition. An L-act is an \({\mathscr {L}}\)-measurable act, hence, certain acts are L-acts. The set of L-acts is \(F_{L} \subseteq F\). For the interpretation of the approach, it is important to note that L-events and L-acts, like S-events and S-acts, are not exogenously given but derived from the preferences of the decision-maker.

The definition of L-partitions is based on the idea that there is a class of decision-relevant events that are not included in \({\mathscr {S}}\) but to which probabilities can nevertheless be assigned. There is no event-specific knowledge concerning the elements of an L partition, however, the decision-maker is indifferent to permuting the consequences of these events. This is plausible only if he or she has no reason to believe that one event is more probable than another (the principle of insufficient reason). Consequently, the decision-maker considers the elements of an L-partition to be equally probable.

The next two axioms cover further properties of L-acts and L-events.

Axiom 4

(Outcome Monotonicity for L-acts) \(x \preceq z \Leftrightarrow x\,C\,h \preceq z\,C\,h\) for all \(x, z \in X\), all \(h \in F_{L}\), and all \(C \in {\mathscr {L}} {\setminus } {\mathscr {N}}\).

Axiom 5

(Decomposability of L-events) There is an L-partition \(\{ C_{i} \}\) with \(i=1,..., n\) such that \(x\,A\, \cup \,C_{i}\,z \sim x\,C_{i}\,z\) or \(x\,A\,\cup \,C_{i}\,z \sim z\) for all \(x, z \in X\) and all \(A \in {\mathscr {L}}\).

Axiom 4 corresponds to Savage’s P3, related to L-acts, and appears largely uncontroversial. An immediate consequence is:

Lemma 2

L-partitions do not include null events.

The main function of Axiom 5 is guaranteeing the existence of L-partitions on which the construction of a subjective probability measure on \({\mathscr {L}}\) can be based. To begin with, Axiom 5 is used to show that all L-events except null events and universal events can be decomposed into the elements of an L-partition.

Lemma 3

For all \(A \in {\mathscr {L}} {\setminus } ( {\mathscr {N}} \cup {\mathscr {U}} )\) there is an L-partition \(\{ C_{i} \}\) with \(i=1,..., n\) such that \(A \cap C_{i} = C_{i}\) or \(A \cap C_{i} = \emptyset\) for all i.

Taken together, Lemma 2 and Lemma 3 imply that S-events and L-events do not intersect except for null events and universal events:

Lemma 4

\({\mathscr {S}}\,\cap \,{\mathscr {L}} = {\mathscr {N}}\,\cup \,{\mathscr {U}}\)

Consequently, S-acts and L-acts do not overlap except for certain acts. This opens the possibility to treat L-acts as consequences of S-acts and to adapt Savage’s P3 to P6.

Axiom 6

(L-act Monotonicity for S-acts) \(f \preceq g \Leftrightarrow f\,A\,h \preceq g\,A\,h\) for all \(f, g \in F_{L}\), all \(h \in F\), and all \(A \in {\mathscr {S}} {\setminus } {\mathscr {N}}\).

Axiom 7

(Weak Comparative Probability) \(f\,A\,g \preceq f\,B\,g \Leftrightarrow f'\,A\,g' \preceq f'\,B\,g'\) for all \(A, B \in {\mathscr {S}}\), and all \(f, g, f', g' \in F_{L}\) such that \(f \succ g\) and \(f' \succ g'\).

Axiom 8

(Nontriviality) There exist \(x, z \in X\) such that \(x \succ z\).

Axiom 9

(Small Event Continuity) For all \(f, g \in F\) such that \(f \succ g\) and all \(h \in F_{L}\), there exists an S-partition \(\{ E_{i} \}\) with \(i = 1, \ldots , n\) such that \(f \succ h\,E_{i}\,g\) and \(h\,E_{i}\,f \succ g\) for all \(i = 1, \ldots , n\).

Axioms 6 to 9 resemble Savage’s P3 to P6, with outcomes replaced by L-acts. Note that the interpretation of consequences as acts is common in decision theory. The approach of Anscombe and Aumann (1963), for example, is characterized by the interpretation of consequences as lotteries. Even Savage (1972) considered the possibility of modeling so-called small world consequences as grand world acts, suggesting the idea of identifying consequences with acts.Footnote 18 Since constant acts are L-acts, Axiom 6 implies Savage’s P3 if the considered event is an S-event. Together with the other axioms, Axioms 5 and 6 imply monotonicity with respect to all decision acts. Except for the replacement of outcomes by L-acts, Axioms 7 to 9 are well known and require no further comments. Designations come from Machina and Schmeidler (1992).

Given Axioms 1 to 9, Savage’s representation theorem can be adapted:

Theorem 1

(Subjective expected utility theory with L-acts as consequences) Suppose that Axioms 1 to 9 hold. Then there is a countably additive probability measure \(P_S: {\mathscr {S}}\rightarrow [0, 1]\) and a function \(U: F_{L} \rightarrow \mathbb {R}\) such that

$$\begin{aligned} F_{i=1}^{n} f_{i}\,A_{i} \preceq F_{i=1}^{n} g_{i}\,A_{i} \Leftrightarrow \sum _{i=1}^{n} U(f_{i}) P_S(A_{i}) \le \sum _{i=1}^{n} U(g_{i}) P_S(A_{i}) \end{aligned}$$

holds for all \(f_{i}, g_{i} \in F_{L}\) and all \(A_{i} \in {\mathscr {S}}\) with \(i = 1, \ldots , n\). \(P_S\) is uniquely determined, and U is unique up to positive linear transformations.

To prove Theorem 1, notice that Axioms 1, 2, 6, 7, and 9 imply P1 to P4 and P6 in Savage (1972), related to L-acts as consequences. The analogon of Savage’s P5 concerning L-acts follows from Axiom 8 since constant acts are L-acts. Hence, the proof of Theorem 1 with respect to a finitely additive probability measure is given in Savage (1972). The countable additivity of \(P_S\) follows from Axiom 3, as proven in Villegas (1964).Footnote 19

Theorem 1 may give the impression that it concerns the evaluation of a two-stage decision act: L-acts are determined in the first stage, and outcomes in the second. This reminds of the approach in Segal (1987), where ambiguous decision acts are modeled as two-stage lotteries and ambiguity is interpreted as uncertainty about probability distributions. Crucial for Segal’s approach is the assumption that the decision-maker does not apply the reduction of compound lotteries axiom.Footnote 20 Consequently, in his approach, two-stage decision acts are judged differently than single-stage acts, even when the probability distribution of outcomes is identical. In contrast, the approach presented here does not rely on this assumption. For illustrative purposes, it may be useful to imagine a stepwise resolution of uncertainty, but it is not part of the modeling of the decision situation. The representation of decision acts in Theorem 1 does not result from a temporal order of S-events and L-events but from their independence according to the definition of L-partitions and Lemma 4. Consequently, the temporal resolution of uncertainty does not matter for the validity of Theorem 1.

Theorem 1 can serve as a basis for decision theories that determine the utility of L-acts in different ways. Given Lemma 4, one possibility is to consider so-called horse lotteries with roulette lotteries in terms of S-acts as consequences, thus providing a subjective foundation for the modeling framework of Anscombe and Aumann (1963). The approach taken here involves relating L-acts and S-acts, thus establishing a link between risk and ambiguity. To this end, some additional concepts must be introduced: An S-partition \({\{ D_{i} \}}\) with i = 1, ..., n is uniform if \(f\,D_{i}\,g \sim f\,D_{j}\,g\) holds for all \(f, g \in F_{L}\) and all i, j. Hence, all events forming a uniform S-partition have the same probability. The elements of a uniform S-partition with two elements are called \(R_{S}\)-events. Two uniform S-partitions \({ \{ D_{i} \}}\) with \(i = 1, \ldots , n\) and \({\{ E_{j}\}}\) with \(j = 1, \ldots , m\) are independent if the events \(D_{i}\,\cap \,E_{j}\) form a uniform S-partition. Using these concepts, Axiom 10 is stated as follows:

Axiom 10

(SL-Isometry) Let A be an \(R_{L}\)-event, B an \(R_{S}\)-event, \(\{ D_{i} \}\), \(\{ E_{i} \}\) with \(i = 1,\dots , n\) independent uniform S-partitions, and \(\{ C_{i} \}\) with \(i = 1, \dots , n\) an L-partition. \(\{ E_{i} \}\) and \(\{ B, B^{C} \}\) are independent. Then for all \(x_{i}, x_{j} \in X\)

$$\begin{aligned} (F_{i=1}^{n} x_{i}\,D_{i} )\,A\,( F_{i=1}^{n} x_{i}\,E_{i} ) \sim (F_{i=1}^{n} x_{i}\,C_{i} )\,B\,( F_{i=1}^{n} x_{i}\,E_{i} ). \end{aligned}$$

For the interpretation of Axiom 10, note that the indifference principle assigns the same probability to all elements of an L-partition. Hence, \(F_{i=1}^{n} x_{i}\,C_{i}\), \(F_{i=1}^{n} x_{i}\,D_{i}\), and \(F_{i=1}^{n} x_{i}\,E_{i}\) produce identical outcomes with identical probabilities. Acts with this property will be called risk-equivalent. With this in mind, Axiom 10 states that the decision-maker is indifferent between an ambiguous mixture of two independent risk-equivalent S-acts and an unambiguous mixture of one of these S-acts and a risk-equivalent L-act. Consequently, Axiom 10 makes it possible to trace the effects of ambiguity back to a systematic juxtaposition of two outcomes and the corresponding events.

Axiom 10 refers to uniform and independent S-partitions as devices for the mixture of L-acts and S-acts. The existence of these partitions is ensured by the following lemma, which essentially goes back to Savage:

Lemma 5

For all S-events A and all \(r \in [0, 1]\) there is an S-event \(B \subseteq A\) such that \(P_S(B) = r P_S(A)\). It follows that

  • there is an \(R_{S}\)-event.

  • there is a uniform S-partition \(\{ D_{i} \}\) with \(i = 1, \ldots , n\) for all \(n \in \mathbb {N}\).

  • for all uniform S-partitions \(\{ D_{i} \}\) with \(i = 1, \ldots , n\) there is an S-partition \(\{ E_{j} \}\) with \(j = 1, \ldots , m\) that is independent from \(\{ D_{i} \}\).

For the first statement see Savage (1972).Footnote 21 The other statements are immediate implications. Next, the existence of at least one \(R_{L}\)-event must be guaranteed:

Axiom 11

(Existence of \(R_L\)-events) There is an \(R_{L}\)-event.

Axiom 11 is a departure from SEU since, according to Lemma 4, an \(R_{L}\)-event is not an S-event. From a technical point of view, an \(R_{L}\)-event is needed for a constructive proof of the intended representation theorem using Axiom 10. Since the indifference principle implies equal probabilities for an \(R_{L}\)-event and its complement, Axiom 11 reminds of Ramsey’s (1931) postulate of the existence of an ethically neutral event. Here, the required “neutrality” includes the idea that the event carries no decision-relevant information about S-events, which follows from the definition of L-partitions and Lemma 4.

With the introduction of Axiom 11, the prerequisites for the utilization of Axiom 10 have been completed. As a first consequence, Axiom 10 implies that preferences for L-acts depend only on the possible outcomes and the cardinality of the relevant L-partition:

Lemma 6

Let \(\{ C_{i} \}\) and \(\{ D_{i} \}\) with \(i = 1,..., n\) be arbitrary L-partitions. Then \(F_{i=1}^{n} x_{i}\,C_{i} \sim F_{i=1}^{n} x_{i}\, D_{i}\) holds for all \(x_{i} \in X\).

Lemma 6 ensures that preferences for outcomes of L-acts can be separated from beliefs concerning L-events.

Lemma 7

For all L-acts \(h = x\,A\,z\) with A an arbitrary \(R_{L}\)-event and outcomes \(x \succ z\), \(\frac{3}{4}\, U(x)+\frac{1}{4} \,U(z) \ge U(h) \ge \frac{1}{4}\, U(x)+ \frac{3}{4}\, U(z)\).

Lemma 7 is an implication of Axiom 4 and Axiom 10 and shows that the utility of an L-act according to Theorem 1 is constrained by the utilities of possible outcomes, which excludes violations of the monotonicity principle.

In the next step, a utility function \(u: X \times X \rightarrow \mathbb {R}\) is introduced that will be used for the representation of preferences for L-acts. u is symmetric if \(u(x, z) = u(z, x)\) for all \(x, z \in X\), and u-monotonic if \(u(x, x) \ge u(x, z) \ge u(z, z)\) holds for all \(x, z \in X\) with \(u(x, x) \ge u(z, z)\).

Lemma 8

There is a symmetric and u-monotonic function \(u: X \times X \rightarrow \mathbb {R}\) such that

$$\begin{aligned} U(F_{i=1}^{n} x_{i} C_{i})=\dfrac{1}{n^{2}} \sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) \end{aligned}$$

for all \(x_{i}, y_{i} \in X\) and all L-partitions \({ C_{i} }\) with \(i = 1, \ldots , n\).

As the proof of Lemma 8 shows, u can be constructed starting from the utility function U in Theorem 1 by setting \(u(x, x) = U(x)\) for all \(x \in X\). Since u has two arguments, however, it allows representing differing preferences for S-acts and L-acts. To establish a preference representation for all simple L-acts, a further implication of Axiom 4 is needed:

Lemma 9

Each partition \(\{C_{k}\}\) of \(\Omega\) with \(C_{k} \in {\mathscr {L}}\) for all k has a finite number of nonnull events.

Lemma 9 is used, among other things, for the construction of a countably additive probability measure on \({\mathscr {L}}\). As a consequence of Lemma 8 and Lemma 9, the following result concerning preferences for simple L-acts can be derived:

Theorem 2

(Representation of preferences for L-acts) Suppose Axioms 1 to 11 hold. Then there is a countably additive probability measure \(P_L: {\mathscr {L}} \rightarrow [0, 1]\) and a symmetric and u-monotonic function \(u: X \times X \rightarrow \mathbb {R}\) such that

\(F_{k=1}^{r} x_{k} C_{k} \preceq F_{k=1}^{r} y_{k} C_{k} \Leftrightarrow\)

$$\begin{aligned} \sum _{k=1}^{r} \sum _{k'=1}^{r} u(x_{k}, x_{k'}) P_L(C_{k}) P_L(C_{k'}) \le \sum _{k=1}^{r} \sum _{k'=1}^{r} u(y_{k}, y_{k'}) P_L(C_{k}) P_L(C_{k'}) \end{aligned}$$

for all \(x_{k}, y_{k} \in X\) and all \(C_{k} \in {\mathscr {L}}\) with \(k = 1, \ldots , r\). \(P_L\) is uniquely determined, and u is unique up to positive linear transformations.

For \(u(x, z) = 0.5 u(x, x) + 0.5 u(z, z)\), the preference functional for L-acts resulting from Theorem 2 is identical to that for S-acts in Theorem 1. Hence, L-acts and S-acts are treated the same way, indicating that the decision-maker has a neutral attitude toward ambiguity. For \(u(x, z) < 0.5 u(x, x) + 0.5 u(z, z)\), the decision-maker is ambiguity-averse. Substitution of \(v(x_k)=\sum _{k'=1}^{r} u(x_{k}, x_{k'}) P_L(C_{k'})\) and \(v(y_k)=\sum _{k'=1}^{r} u(y_{k}, y_{k'}) P_L(C_{k'})\) shows that he or she still maximizes expected utility. The utility function, however, is modified in such a way that it effectively overweights less preferred outcomes. Moreover, the utility function is act-specific due to the inapplicability of the Sure Thing principle. The quadratic form of the preference functional in Theorem 2 was studied in detail in Chew et al. (1991) in the context of relaxations of the independence axiom. For \(u(x, z) = {\text {min}}\{ u(x, x), u(z, z) \}\), it reflects a special case of the rank-dependent utility theory of Quiggin (1982) and others.Footnote 22 The relationship is comparable to that between the approach developed here and CEU discussed in the introduction.

The final idea in this section is that all decision-relevant events can be interpreted as mixtures of S-events and L-events. To this end, \({\mathscr {K}}\) is defined as the \(\sigma\)-algebra generated by \({\mathscr {S}}\) and \({\mathscr {L}}\). The following axiom ensures that \({\mathscr {K}}\) encompasses all decision-relevant events:

Axiom 12

(Decision-relevant events) For all \(A \in {\mathscr {A}}\) there is an event \(B \subseteq A\) with \(B \in {\mathscr {K}}\) such that \(f\,A {\setminus } B\,h \sim g\,A {\setminus } B\,h\) for all \(f, g, h \in F\).

\(A \setminus B\) in Axiom 12 is a null event. As null events are S-events, \({\mathscr {A}}\) is the \(\sigma\)-algebra generated by \({\mathscr {S}}\) and \({\mathscr {L}}\), i.e., \({\mathscr {A}} = {\mathscr {K}}\).

By merging Theorem 1 and Theorem 2, the central result can now be stated as follows:

Theorem 3

(SL expected utility theory) Suppose Axioms 1 to 12 hold. Then there are two \(\sigma\)-algebras \({\mathscr {S}}\) and \({\mathscr {L}}\) with \({\mathscr {S}} \cap {\mathscr {L}} = {\mathscr {N}} \cup {\mathscr {U}}\) that jointly generate the \(\sigma\)-algebra \({\mathscr {A}}\), a countably additive probability measure \(P: {\mathscr {A}} \rightarrow [0, 1]\), an SL-decomposition \(Q: {\mathscr {A}} \times {\mathscr {A}} \rightarrow [0, 1]\) of P, and a symmetric and u-monotonic function \(u: X \times X \rightarrow \mathbb {R}\) such that

\(F_{i=1}^{n} x_{i} A_{i} \preceq F_{i=1}^{n} y_{i} A_{i} \Leftrightarrow\)

$$\begin{aligned}\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) Q(A_{i}, A_{j}) \le \sum _{i=1}^{n} \sum _{j=1}^{n} u(y_{i}, y_{j}) Q(A_{i}, A_{j}) \end{aligned}$$

for all \(x_{i}, y_{i} \in X\) and all \(A_{i} \in {\mathscr {A}}\) with \(i = 1, \ldots , n\). P and Q are uniquely determined, and u is unique up to positive linear transformations.

Theorem 3 uses the SL-decomposition Q defined in the last section and generalizes Theorem 1 and Theorem 2. If \(Q(A, A) = P(A)\) for all \(A \in {\mathscr {A}}\) or \(u(x, z) = 0.5 u(x, x) + 0.5 u(z, z)\) for all \(x, z \in X\), the preference functional corresponds to SEU, i.e., Theorem 1. If \(Q(A, B) = P(A) P(B)\) for all \(A, B \in {\mathscr {A}}\), the preference functional for L-acts according to Theorem 2 applies. Taken together, the diverging ambiguity of events and decision acts as well as different attitudes toward ambiguity can be represented, as discussed in the next section.

4 Measurement of ambiguity and ambiguity aversion

Ambiguity can be measured with respect to events and to decision acts. The SL-decomposition and its interpretation suggest the following measure of event ambiguity:

Definition 3

(Event ambiguity) The ambiguity of an event A is measured by \(\alpha (A) = Q(A, A^C)\).

Definition 3 is intuitively plausible since \(Q(A, A^C)\) measures the part of the probability of event A that is not based on event-specific knowledge. In a more formal way, \(\alpha\) can be introduced by first defining a qualitative ambiguity relation and then showing that \(\alpha\) represents this relation. To this end, let \(x, z \in X\) be arbitrary outcomes with \(x \succ z\). Further, for all events \(A \in {\mathscr {A}}\), construct S-events \(A_S\) and \(A_S^C\) with \(x\,A\,z \sim x\,A_S\,z\) and \(x\,A^C\,z \sim x\,A_S^C\,z\), respectively, applying Theorem 3 and Lemma 5. An ambiguity-averse decision-maker prefers to bet on unambiguous events, hence \(A_S\,\cap \,A_S^C = \emptyset\) can be assumed without loss of generality.Footnote 23 A qualitative ambiguity relation \(\preceq\) is then defined as follows: Event A is (weakly) less ambiguous than event B, \(A \preceq B\), if \(x\,A_S\, \cup \, A_S^C\,z \succeq x\,B_S \, \cup \, B_S^C\,z\) holds for an ambiguity-averse decision-maker. Theorem 3 implies \(x\,A_S\, \cup \,A_S^C \,z \succeq x\,B_S\, \cup \, B_S^C\,z\) iff \(Q(A, A^C) \ge Q(B, B^C)\) for \(u(x, z) < 0.5 u(x, x) + 0.5 u(z, z)\). Consequently, \(\alpha\) represents the ambiguity relation \(\preceq\). In view of Theorem 3, it is evident that \(\preceq\) is independent of the defining outcomes and forms a weak order.

Another approach to introduce a measure of event ambiguity is to first establish properties that this measure should satisfy and then to develop an appropriate axiomatic system. Fishburn (1993) lists the following propertiesFootnote 24:

  1. 1.

    \(A \preceq B \Rightarrow \alpha (A) \le \alpha (B)\) and \(A \prec B \Rightarrow \alpha (A) < \alpha (B)\)

  2. 2.

    \(\alpha (\emptyset ) = 0\) and \(\alpha (A) \ge 0\)

  3. 3.

    \(\alpha (A)=\alpha (A^C)\)

  4. 4.

    \(\alpha (A \cup B) + \alpha (A \cap B) \le \alpha (A) + \alpha (B)\)

The proposed measure satisfies all these properties,which further underlines its plausibility. Property 1 requires \(\alpha\) to preserve the qualitative ambiguity relation, which follows from \(A \preceq B \Leftrightarrow \alpha (A) \le \alpha (B)\) derived above. Properties 2 to 4 can be readily deduced from the properties of the SL-decomposition listed in Section 2.Footnote 25 Properties 3 and 4, complementary equality and submodularity, respectively, are considered to be central to the distinction between event ambiguity and event probability. Complementary equality reflects the intuitively plausible idea that the ambiguity of an event is equal to that of its complement. Submodularity implies that the ambiguity of the union of two disjoint events is less than the sum of the stand-alone ambiguities.

To define a measure of act ambiguity, it is supposed that the SL-decomposition Q and the utility function u are given and that all outcomes are measurable on a cardinal scale. \(q: \mathbb {R}^{2} \rightarrow [0, 1]\) with \(q_{ij} = q(x_{i}, x_{j}) = Q(A_{i}, A_{j})\) for \(A_{i} = \{ s \in \Omega : f(s) = x_{i} \}\) and \(A_{j} = \{ s \in \Omega : f(s) = x_{j} \}\) is referred to as \(SL\)-distribution of decision act f. As a consequence of the properties of the SL-decomposition, q can be interpreted as a symmetric two-dimensional probability distribution with marginal distribution p where \(p_{i} = p(x_{i}) = P(A_{i})\) holds for \(A_{i} = \{ s \in \Omega : f(s) = x_{i}\}\). Based on p, the expected value \(\mu\) and the variance \(\sigma ^{2}\) of outcomes can be calculated as usual. Two \(SL\)-distributions with identical marginal distributions are called risk-equivalent.

Based on the \(SL\)-distribution q, a convenient way to measure act ambiguity is as follows:

Definition 4

(Act ambiguity) The ambiguity of an \(SL\)-distribution q with marginal distribution p is measured by \(\varphi = \sigma ^2 - c\), where \(\sigma ^2\) denotes the variance of p and c the covariance of q.

\(0 \le c \le \sigma ^{2}\) and, consequently, \(0 \le \varphi \le \sigma ^{2}\) result from the characteristics of the SL-decomposition. \(\varphi = 0\) reflects a perfectly unambiguous outcome distribution where all relevant events are S-events. \(\varphi = \sigma ^{2}\) results when act ambiguity is perfect, i.e., probabilities for relevant events are not based on event-specific knowledge but on the indifference principle. In all other cases, the decision act is more or less ambiguous since the outcomes depend on the occurrence of S-events and L-events.

Example 4

(Measurement of Act ambiguity) The monetary outcome x of an investment depends on a perfectly unambiguous factor \(z_S\) and a perfectly ambiguous factor \(z_L\), \(x = z_S + z_L\). Possible realizations of \(z_S\) are 0, 100, 200, or 300 with underlying S-events \(B_1\), \(B_2\), \(B_3\), and \(B_4\), and probabilities 0.2, 0.3, 0.4, and 0.1, respectively. \(z_L\) results in \(-100\), 0, or 100 with underlying L-events \(C_1\), \(C_2\), and \(C_3\), and probabilities 0.2, 0.6, and 0.2, respectively. Overall, the expected outcome is \(\mu = \mu _S + \mu _L = 140 + 0 = 140\). From the properties of S-events and L-events, \(z_S\) and \(z_L\) are necessarily uncorrelated, hence the outcome variance is \(\sigma ^2 = \sigma _S^2 + \sigma _L^2 = 8,400 + 4,000 = 12,400\). To determine the relevant \(SL\)-distribution, note that \(z_S\) is perfectly unambiguous, i.e., \(q(z_S^i, z_S^{i'}) = p(z_S^i)\) for \(i = i'\) and \(q(z_S^i, z_S^{i'}) = 0\) otherwise, while \(z_L\) is perfectly ambiguous, hence \(q(z_L^j, z_L^{j'}) = p(z_L^j) p(z_L^{j'})\) for all \(j, j'\). The resulting \(SL\)-distribution is shown in Table 5. For example, the properties of SL-decompositions imply \(q(200, 300) = Q((B_1 \cap C_3) \cup (B_2 \cap C_2) \cup (B_3 \cap C_1)\), \((B_2 \cap C_3) \cup (B_3 \cap C_2) \cup (B_4 \cap C_1)) = Q(B_2 \cap C_2, B_2 \cap C_3) + Q(B_3 \cap C_1, B_3 \cap C_2) = P(B_2) P(C_1) P(C_3) + P(B_3) P(C_1) P(C_2) = 0.06\). The covariance of q is \(c = 8,400\). Consequently, according to Definition 4, \(\varphi = 4,000\) is calculated for the ambiguity of the investment. It results solely from the perfectly ambiguous factor \(z_L\) and equals its ambiguity \(\varphi _L\).

Table 5 SL-distribution of investment outcome

As a starting point for the consideration of ambiguity aversion, note that risk aversion can be defined and measured with regard to the utility function \(U(x) = u(x, x)\) in the usual way. To define a measure of ambiguity aversion remind that an ambiguity-neutral decision-maker acts as if all decision acts were S-acts. This is the case if and only if \(\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) q_{ij} = \sum _{i}^{n} u(x_{i}, x_{i}) p_{i}\) holds for all \(SL\)-distributions q with marginal distributions p. Accordingly, a decision-maker is (strictly) ambiguity-averse if \(\sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) q_{ij} < \sum _{i}^{n} u(x_{i}, x_{i}) p_{i}\) for all \(f \in F {\setminus } F_{S}\). Thus, an ambiguity-averse decision-maker strictly prefers an S-act against all other decision acts that have the same probability distribution of outcomes but are not S-acts. Monotony and differentiability presupposed, a necessary and sufficient condition for this to hold is \(u_{11}(x, z) > 0\) for all \(x, z \in X\). Consequently, a measure of ambiguity aversion can be defined as follows:

Definition 5

(Ambiguity aversion) The local ambiguity aversion of a decision-maker is measured by \(A(x,z) = \dfrac{u_{11}(x,z) }{u_{10}(x,z)+u_{01}(x,z)}\)

If A(xz) has a negative sign, the decision-maker is ambiguity-loving. The measure A(xz) remains, like the Arrow/Pratt measure, unaffected by positive linear transformations of the utility function. Note that ambiguity aversion is in general independent of risk aversion, i.e., a risk-averse decision-maker can be ambiguity-neutral and a risk-neutral decision-maker can be ambiguity-averse. The measurement of ambiguity aversion is illustrated by example 5.

Example 5

(Measurement of Ambiguity aversion)

Suppose \(u(x, z) = 0.5 (x + z)-0.5 (a + b) (x^{2} + z^{2}) + b x z\). The domain of u is restricted to outcomes in the interval [0, 1]. To ensure that u exhibits weak risk aversion and weak ambiguity aversion, \(a, b \ge 0\) and \(a+b<0.5\) is presupposed. The Arrow/Pratt measure is \(R(x)= \frac{2 a}{1-2 a x}\). For \(a>0\), u indicates increasing absolute risk aversion as typical for the quadratic utility function in SEU. For \(b > 0\), the measure of ambiguity aversion \(A(x,z) = \frac{b}{1-a(x+z)}\) reflects absolute ambiguity aversion increasing in both arguments. In analogy to the \(\mu\)-\(\sigma ^{2}\)-principle, the following result for all simple \(SL\)-distributions q with support [0, 1] can be derived:

$$\begin{aligned} \sum _{i=1}^{n} \sum _{j=1}^{n} u(x_{i}, x_{j}) q_{ij} = \mu - a(\mu ^{2} + \sigma ^{2}) - b \varphi \end{aligned}$$

The last term on the right-hand side refers to the measure of act ambiguity according to Definition 4. Given \(\mu\) and \(\sigma ^{2}\), preference increases with decreasing ambiguity when the decision-maker is ambiguity-averse, i.e., \(b > 0\). For \(b = 0\), the decision-maker is ambiguity-neutral and follows SEU with \(U(x)=x - a x^{2}\). If risk aversion is held constant, the ambiguity-neutral decision-maker achieves the same utility level as the ambiguity-averse decision-maker with a risk-equivalent but perfectly unambiguous act.

5 Concluding remarks

A central feature of the proposed theory is the possibility to separate ambiguity and ambiguity attitudes in a relatively simple way. Together with the additivity of the SL-decomposition, this allows analyzing the consequences of ambiguity and ambiguity aversion in model-theoretic and empirical studies in quite the same way as the consequences of risk and risk aversion. In doing so, all results concerning risk remain valid as the theory presented is an extension of SEU. From a technical point of view, it is particularly advantageous that the \(SL\)-distributions used to model decision acts can be interpreted as two-dimensional probability distributions. This makes it possible to apply concepts from probability theory when modeling and measuring ambiguity. In this respect, the proposed theory is similar to approaches that use second-order probabilities to model ambiguity.

As for the axiomatic basis of the proposed theory, the number of axioms could be an issue. The presented axiomatic system includes twelve axioms, while Savage’s and Gilboa’s axiomatizations of SEU and CEU, respectively, each consist of six axioms. Reducing the number of axioms by merging axioms (e.g., Axioms 4 and 6) is possible, but only to a limited extent and at the cost of more complicated and less plausible axioms. This is also a consequence of the model framework which is subject to comparatively few restrictions. In general, the number of axioms decreases with increasing structural specifications. Axiomatic systems based on Savage’s approach, for example, usually include more axioms than those based on Anscombe and Aumann’s framework. With this in mind and given the higher complexity resulting from the distinction between S-events and L-events, it seems questionable whether the axiomatic basis of the proposed theory can be reduced to a significantly smaller number of axioms without imposing additional structural specifications.

Considering the axioms in detail, Axiom 10 and Axiom 12 are of particular interest. Axiom 10 connects preferences for S-acts and L-acts thus establishing a link between risk and ambiguity. The central question here is whether Axiom 10 leads to reasonable and empirically tenable hypotheses about ambiguity-related decision-making. Axiom 12 raises some interesting epistemic questions, for example: Are there possibilities of assigning probabilities to events beyond knowledge-based separation and indifference-based dispersion? Or, more closely to the topic of this paper: Are there sources of ambiguity that cannot be related to the application of the indifference principle? If so, Axiom 12 would have to be abandoned. Consequently, either the scope of Theorem 3 would have to be restricted to K-acts or additional axioms would have to be introduced that govern preferences for the corresponding decision acts.

With regard to the latter point, it could also be argued that the presented approach basically considers only perfectly unambiguous and perfectly ambiguous events and that there are ambiguous events that cannot be traced back to such events. The idea is that if necessary, the \(\sigma\)-algebra of decision-relevant events can always be extended in such a way that suitable L-events and S-events can be identified. If this is the case, the argument leads to the debate as to whether axioms in decision theory may only refer to actually existing events or whether fictitious events may also be used for axiomatization. Savage’s P6, for example, was criticized for its use of infinitely divisible events. Today, most decision theorists seem to agree that referring to fictitious events is acceptable.

From the constructive proof of the theorems in the last section, it is easy to derive a procedure for eliciting the SL decomposition Q and the utility function u of an individual decision-maker. An issue here is that in this context events are needed that are S-events and/or L-events from the decision-maker’s perspective. The situation is comparable to elicitation techniques for SEU that use certain events. As with certain events, different decision-makers may interpret different events as S-events or L-events. For a purely subjective theory, this is not a conceptual problem. It requires, however, special care in the conceptualization of elicitation procedures.

Further research questions are manifold. Of particular interest is an ongoing investigation of the relationship between ambiguity and information. In the context of the presented theory, it seems plausible to assume that subjective probabilities should be updated according to Bayes’ rule. However, it is by no means self-evident how the SL-decomposition must be adjusted. This may depend on the extent to which the information concerns S-events or L-events. An investigation of this issue could contribute to a better understanding of the relationship between ambiguity and information and also raise new questions concerning ambiguity-related behavior.