Keywords

1 Introduction: Decision Theory and Deontic Logic

Consider the following two decision problems.

Example 1

(Gambler ): An agent, \(\upalpha \), is offered a gamble. If she accepts, she pays $5. A coin is then tossed: on Heads she wins $10; on Tails she wins nothing. If she declines the gamble, she simply keeps her $5.

Example 2

( Matching Pennies ): Two agents, \(\upalpha \) and \(\upbeta \), simultaneously choose whether to display a penny Heads up or Tails up. If the displayed sides of the two pennies match, then \(\upalpha \) wins \(\$1\) from \(\upbeta \). If the two sides do not match, then \(\upbeta \) wins \(\$1\) from \(\upalpha \).

Assuming that these agents value money positively and that there are no relevant external considerations, what should \(\upalpha \) do in these two scenarios? To find answers, we might look to two distinct normative frameworks: decision theoryFootnote 1 and deontic logic. Decision theory, despite its many paradoxes and controversies, provides our most successful formal account of rational choice. Deontic logic, with its own paradoxes and controversies, offers an alternative way to think about what \(\upalpha \) ought to do.

What is the relationship between decision theory and deontic logic? We might think of them as directed towards answering different questions. Decision theory rests on sharp assumptions about preferences and belief, but also upon less clear assumptions about causation, choice and counterfactuals. Deontic logic has no place for probabilities or probabilistic reasoning and does not pretend to offer a comprehensive theory of rational choice. Yet both theories can be applied to examples like Gambler and Matching Pennies. This suggests that they might, in some way, be rivals.

There is a third possibility. Rather than see them as unrelated or as rivals, we might regard deontic logic as a kind of proto-decision theory. Conceived in this way, deontic logic would play three roles. First, it would provide a rigorous logical framework for decision theory, a framework that clarifies foundational assumptions about causation, choice, counterfactuals and other relevant concepts. Second, stronger and weaker systems of deontic logic would be definable in this common framework. Third and finally, these systems of deontic logic would also be rudimentary decision theories: they would provide norms for choices by agents that are compatible with basic principles of decision theory.

When deontic logic is viewed in this way, the approach developed by Horty (2001) is exemplary. Horty’s deontic logic (in contrast to many earlier approaches) is prescriptive: it is about choices by agents. It proposes semantics for what a particular agent ought to do at a particular moment in time. Horty’s framework is built on top of Belnap’s modal logic of agency, stit theory, a clear and rigorous logical and metaphysical account of agents making choices in indeterministic branching time.Footnote 2 Stit theory already takes us part way to causal decision theory because it incorporates causal notions: agents, branching time and a formulation of causal independence. Horty takes us further by incorporating utilities and dominance reasoning into his account, although he does not introduce probabilistic concepts. Still, his deontic logic does provide a weak decision theory: under reasonable assumptions, the set of obligations for an agent on the Horty semantics is a subset of the set of obligations that the agent has according to any sound principle of decision theory.

The main thesis of this chapter is that Horty’s approach can be fruitfully enriched, first by a slight generalization of his account of causal independence and second by adding a ‘thin’ layer of probabilistic concepts.Footnote 3 The result is a framework in which deontic logic is even better suited to serve as proto-decision theory by playing the three roles mentioned above. There may be good reasons not to introduce full-blooded probability into the stit universe.Footnote 4 But it is plausible to introduce probabilities for mixed strategies by agents, and more generally for chance mechanisms (such as coin tosses and dice rolls). Horty’s deontic logic can then be expanded fruitfully towards different branches of decision theory.

The chapter proceeds as follows. Section 2 reviews the basic ideas of stit (seeing-to-it-that). Section 3 proposes a generalization of Horty’s account of causal independence. Horty’s Dominance Ought is reviewed in Sect. 4, along with a slight modification corresponding to the generalization of Sect. 3. The remaining sections explore expansions of Horty’s deontic framework to decisions under ignorance (Sect. 5), decisions under risk (Sect. 6) and elementary game theory (Sect. 7). While the focus of the chapter is on endowing deontic logic with resources from decision theory, I conclude with a brief discussion of return benefits for decision theory.

2 Seeing to it That (stit)

In a series of articles, Belnap and others have proposed semantics for the modal construction, “\(\upalpha \) sees to it that A,” or \({[}\upalpha \) stit \({:}\;A{]}\) for short. To keep things brief, I pass over the philosophical motivation and simply review concepts that are crucial for this chapter. The best single source of information on stit is the volume of articles (Belnap et al. 2001).

2.1 Semantics for cstit with One Agent

There are three accounts of \({[}\upalpha \) stit \({:}\;A{]}\) on offer: the Belnap “achievement stit” (astit), the Horty/von Kutschera “deliberative stit” (dstit) and the “Chellas stit” (cstit). Since the latter is employed by Horty in his deontic logic, this section presents, in cursory form, only the semantics for the Chellas stit, following notation that borrows from both (Belnap et al. 2001) and (Horty 2001).Footnote 5 The fundamental idea of \({[}\upalpha \) cstit \({:}\;A{]}\) is that A is guaranteed by a present choice of agent \(\upalpha \) (\({[}\upalpha \) dstit \({:}\;A{]}\) is more complex because it requires this same positive condition, together with the negative condition that A is not ‘settled-true’ in the sense of (3) below).

The framework begins with an indeterministic branching time structure \(<Tree,\, < \quad >\), where Tree is a non-empty set of moments, m, and \(<\) is a treelike partial ordering of those moments. A history, h, in Tree is a maximal chain of moments, i.e., a complete temporal evolution of the world. If m is a moment, write \(H_{m} = \{h / m \in h\}\) for the set of all histories containing (passing through) m. The situation is illustrated in Fig. 1, where the upward direction represents later moments.

Fig. 1
figure 1

Histories

In the picture, \(H_{m} = \{h_{1}, \ldots , h_{6}\}\), while \(H_{m2} = \{h_{1}, h_{2}, h_{3}\}\). The histories \(h_{1}\) and \(h_{2}\) are undivided at m because they share a later moment \((m_{2})\) in common; the histories \(h_{1}\) and \(h_{4}\) are divided at m.

Sentences are constructed from propositional variables A, B, ... using the following operators:

  1. (i)

    Truth-functional operators: \(\sim , \vee \) (with abbreviations \(\wedge , \supset , \equiv )\)

  2. (ii)

    Necessity operators: Universally:, Settled:

  3. (iii)

    Tense operators: Will: and Was:

  4. (vi)

    Agentive operator: \({[}\upalpha \) cstit \({:}\;\_\_ {]}\) where \(\upalpha \) denotes an individual agent

The truth of sentences is evaluated relative to a moment-history pair m/h, where m is a moment belonging to history h. A model M pairs the tree structure with an interpretation that maps each propositional variable A to a set of m/h pairs where A is true:

  1. (1)

    \(M, m/h \models A\, \text {iff}\, A\,\text {is true at}\, m/h\).

In Fig. 1, for example, \(M, m_{2}/h_{1} \models A\,\text {while}\, M, m_{3}/h_{4}\,\nvDash A\). The clauses for the truth-functional operators are standard. For the modal operators, Universally: represents truth throughout Tree, while Settled: represents truth throughout a moment. The relevant clauses are as follows:

  1. (2)

    \(M, m/h \models Universally\!\!: A\,\qquad \text {iff}\,\qquad M, m^{\prime }/h^{\prime } \models A\, \text {for all}\, m^{\prime }/h^{\prime }\, \text {in}\, Tree\)

  2. (3)

    \(M, m/h \models Settled\!\!: A\, \,\qquad \qquad \,\,\,\text {iff}\,\qquad M, m/h^{\prime }\, \models A\,\text {for all}\, h^{\prime }\, \text {with}\, m \in h^{\prime }\)

For the tense operators, we have

  1. (4)

    \(M, m/h \models Will\!\!: A\qquad \qquad \quad \quad \text {iff}\qquad M, m^{\prime }/h \models A\,\text {for some}\, m^{\prime }\, \text {in}\,h\, \text {with}\, m < m^{\prime }\)

  2. (5)

    \(M, m/h \models Was\!\!: A \!\qquad \quad \qquad \qquad \text {iff}\qquad M, m^{\prime }/h \models A\,\text {for some}\, m^{\prime } \, < m.\)

In Fig. 1, Will: A is true at \(m/h_{1}\) but false at \(m/h_{4}\). Settled: A is true at \(m_{2}/h_{1}\) but false at \(m_{3}/h_{5}\).

Finally, we come to \({[}\upalpha \) cstit \({:}\;A{]}\). This requires enriching the branching time framework of \(< Tree, < \quad >\) with a nonempty set AGENT of agents (denoted \(\upalpha , \upbeta \), and so forth) and a function Choice that represents choices by agents. The most important idea is that of a choice set for \(\upalpha \) at moment m, which is a partition of the histories passing through m into choice cells (or simply choices) for \(\upalpha \). \(\upalpha \)’s power of choice consists in “constraining the course of events to lie within some definite subset of the possible histories still available”. (Belnap et al. 2001, 33). That is, choice is identified with the selection of one cluster of histories. The agent picks the cluster, but cannot select a unique history within the choice cell. All of this is formalized in the following definitions.

  1. (6)

    stit frames.

    A \({\textit{stit frame}}\, <{Tree,}\, <, \,{AGENT},\, {Choice}>\) is a structure with Tree and \(<\) as above, AGENT a nonempty set of agents, and Choice a function mapping agent \(\upalpha \) and moment m into a partition of \(H_{m}\) characterized as follows:

    • \({Choice}^{m}_{\upalpha }\) is a partition of \({H}_{m}\) into mutually exclusive and exhaustive sets.

    • Each member of \({Choice}^{m}_{\upalpha }\) is called a choice cell (or choice) for \(\upalpha \) at m

    • h and \(h^{\prime }\) are \({choice}\)-\({equivalent\;for}\;\upalpha \; at\; m (\text {written}\, h^{\prime }\equiv ^{\upalpha }_{m}h)\) if they belong to the same choice cell for \(\upalpha \) at m (no choice that \(\upalpha \) can make at m tells them apart).

Choice is subject to the following condition (and one further condition, Weak Independence of Agents, to be described shortly):

  1. (7)

    No Choice between Undivided Histories.

    If h and \(h^{\prime }\) are undivided at m, then h and \(h^{\prime }\) must belong to the same choice cell for \(\upalpha \) at m.

A choice by some agent is the most obvious means by which histories are divided. Histories also divide as the result of chance processes in Nature. A coin toss serves as a paradigm example.

With this apparatus on board, we can define what it is for \({[}\upalpha \) cstit \({:}\;A{]}\) to hold at m/h:

  1. (8)

    \(M, m/h \models [\upalpha {cstit\!\!:}\;\mathrm{A}]\,\text {iff}\, A\) is true at \((m, h^{\prime })\) for all \(h^{\prime }\) with \(h^{\prime }\equiv ^{\upalpha }_{m}h\). (By choosing the cell containing \(h, \upalpha \) guarantees that A is true as A holds on all histories consistent with \(\upalpha \)’s choice.)

Figure 2 provides the basic picture. In the picture, m is a moment in Tree that has been blown up to reveal the choice structure. The three boxes represent choice cells for \(\upalpha \) at m; each history in \(H_{m}\) belongs to exactly one box. The truth-value of A is shown for each moment–history pair. Here, \({[}\upalpha \) cstit \({:}\;{A}{]}\) holds just at \({m}/{h}_{1}\) and \({m}/{h}_{2}\) (Note that [\(\upalpha \,{cstit}{:}\; \sim \) A] is false throughout m; there is no law of excluded middle for seeing-to-it-that).

Fig. 2
figure 2

\({[}\upalpha \) cstit \({:}\; A{]}\)

2.2 Multiple Agents: Independence and Joint Agency

The concept of cstit generalizes to groups of agents. Most of the ideas can be made clear by considering just two agents, \(\upalpha \) and \(\upbeta \), making simultaneous choices. The simplest way to represent this is once again with a blown-up picture of the moment, this time two-dimensional, with the choices of \(\upalpha \) represented on the horizontal axis and the choices of \(\upbeta \) on the vertical axis, as in Fig. 3.

Fig. 3
figure 3

Multiple agents

Here, \(\upalpha \) and \(\upbeta \) face non-trivial choices, formally specified by choice sets (partitions) \(Choice^{m}_{\upalpha }\) and \({Choice}^{m}_{\upbeta }\). In the picture, the possible choices are \({\varvec{a}}_{1}\) and \({\varvec{a}}_{2}\) for \(\upalpha \), and \({\varvec{b}}_{1}\) and \({\varvec{b}}_{2}\) for \(\upbeta \); thus, the choice sets are \(\{{\varvec{a}}_{1}, {\varvec{a}}_{2}\}\) and \(\{{\varvec{b}}_{1}, {\varvec{b}}_{2}\}\).

The key assumption made by Belnap and Perloff is that every combination of choices by \(\upalpha \) and \(\upbeta \) is possible:

  1. (9)

    [Weak] Independence of Agents.

    For each moment and for each way of selecting one choice for every agent (in the set AGENT) from among that agent’s set of possible choices at that moment, the intersection of all the choices selected must contain at least one history.

To formalize this condition, we follow Horty in defining a selection function s at moment m to be a mapping from AGENT into \(H_{m}\) that selects one action for each agent: \(s(\upalpha ) \in {Choice}^{m}_{\upalpha }\) for each \(\upalpha \). Let \({Select}_{m}\) be the set of all such functions. We re-state (9) as follows:

For each moment m and each selection function s in \({Select}_{m}\),

$$ \bigcap _{\upalpha \in AGENT} {s(\upalpha )\ne \phi .} $$

Belnap comments (Belnap et al. 2001, p. 218) that while Independence is a “fierce” constraint (implying, for example, that no two agents can have the same possible choices at the same moment), it is also fairly weak (“banal” is Belnap’s term): it would be strange indeed if without causal priority, one agent’s choices could constrain what the other agent may choose.

The other important idea is joint agency. Again, the basic idea can be explained with just two agents. Let \(\Gamma = \{\upalpha , \upbeta \}\), where \(\upalpha \) and \(\upbeta \) are distinct agents. We want to define truth conditions for \([\Gamma \, {cstit}{:}\;A]\). The concept is illustrated by referring once again to Fig. 3. With the assignments given by our model M, neither \(\upalpha \) nor \(\upbeta \) can see to it that A on any history in m. However, M, \(m/h_{1} \models [\Gamma \,{\textit{stit}}{:}\;A]\) because A holds at every history \(h^{\prime }\) that belongs to the choice cell containing \(h_{1}\) that is determined jointly by \(\upalpha \) and \(\upbeta \). The formal definition is the same as (8) except that the condition invokes equivalence within the choice cell determined jointly by the agents in \(\Gamma \).

3 Causal Independence

Both causal decision theory and Horty’s deontic logic depend upon the concept of causal independence. Before examining how causal independence can be characterized in the stit framework, it is helpful to review its role in causal decision theory. I focus on the formulation due to Skyrms (1980).Footnote 6

Skyrms’s formulation requires the identification of a set of independent causal factors that provide the background for an agent’s choices. Each independent causal factor is represented as a random variable \(X_{i}\). For simplicity’s sake, suppose that the set of possible outcomes \(O_{1}, {\ldots }, O_{n}\) of interest to the agent is finite, that the set \(X_{1},\ldots , X_{N}\) of independent factors relevant to these possible outcomes is also finite, and that each variable \(X_{i}\) can take on finitely many values. Then the set \({\varvec{S}}\) consisting of all possible combinations of assignments to these variables is also finite. This set constitutes a partition of the set of possible worlds into causal background contexts or \({states}\, S_{1}, \ldots , S_{M}\): each state \(S_\mathrm{i}\) is obtained by specifying one possible value for each of \(X_{1},\ldots , X_{N}\). Suppose that the agent has a finite set \(\{K_{1},{\ldots }, K_{m}\}\) of available alternative acts. The crucial idea is that the causal factors in conjunction with these alternative acts determine relevant conditional chances of the possible outcomes: the conditional chances \(P(O_{k}~/~K_{i}~\wedge ~S_{j})\) are constant within each \(K_\mathrm{i} \wedge S_{j}\). Finally, the agent assigns a utility \(u(O_{k}~\wedge ~K_{i}~\wedge ~S_{j})\) to each outcome-act-state combination.

For a standard example, let \(K_{i}\) be the selection of a ball from one of m urns, let \(O_{1}, {\ldots }, O_{n}\) represent n different colours of ball that may be drawn, and let \(S_{1}, {\ldots }, S_{M}\) stand for M possible initial arrangements of coloured balls in the m urns. Then the probability \(P(O_{k}~/~K_{i} \wedge S_{j})\) is the conditional chance of drawing a ball of colour \(O_{k}\), given arrangement \(S_{j}\) and the act \(K_{i}\) of drawing from urn i. The utility \(u(O_{k}~\wedge ~K_{i}~\wedge ~S_{j})\) depends upon the desirability of each combination (perhaps the agent has placed a bet in advance).

To allow for cases in which the agent is uncertain about the background context, Skyrms introduces a subjective probability distribution \({prob}(S_{j})\) over all of the states. In the urn example, this represents your initial credence about the likelihoods of the different possible arrangements. The expected utility of act \({K}_{i}\) is then given by the formula

  1. (10)

    Expected utility.

    $$\begin{aligned} {\fancyscript{U}}(K_{i})=\Sigma _{j} prob(S_{j}) \Sigma _{k} P(O_{k} / K_{i} \wedge S_{j}) \cdot u(O_{k} \wedge K_{i} \wedge S_{j}). \end{aligned}$$

The thesis of causal decision theory is that a rational agent maximizes expected utility as given by (10). The equation highlights the importance of independent causal factors in the theory; the outer summation is over all possible states.Footnote 7

Horty’s deontic logic has a similar, but much weaker, guiding principle. Without conditional chances or credences, his analysis (outlined in the next section) is based solely upon the concept of dominance. Yet dominance reasoning shares with causal decision theory the need for a set of independent causal factors and a corresponding set of causal background contexts. Stit frames help to make these things precise by providing a concrete interpretation of possible worlds and a plausible way to identify some of the independent causal factors. I first review Horty’s account and then propose a slight generalization.

Horty begins with an informal explication of causal independence:

...the basic intuition ... is that a proposition is supposed to be causally independent of the actions available to a particular agent whenever its truth or falsity is guaranteed by a source of causality other than the actions of that agent. (2001, 82)

Recall postulate (9), Weak Independence of Agents: if agents in a group make simultaneous choices, then the intersection of the relevant choice sets is non-empty. Horty strengthens this in two ways. First, he assumes that all choices at m by agents other than \(\upalpha \) are independent causal factors for \(\upalpha \)’s choice at m. This strengthened assumption, stated in counterfactual terms, applies to all moment–history pairs \(m/h\) and all agents \(\upalpha \).

  1. (11)

    Strong Independence (of Agents).

    Let S represent the intersection of all actual choices (i.e., choice cells) of all agents other than \(\upalpha \) at m/h. If \(\upalpha \) were to make a different choice than the one made at m/h, the other agents would still (collectively) choose S.

Second, Horty adopts a provisional simplifying assumption that I shall refer to as Causal Completeness of AGENT.

  1. (12)

    Causal Completeness (of AGENT).

    Choices by agents in \({AGENT}\backslash \{\upalpha \}\) (i.e., agents other than \(\upalpha )\) are the only independent causal factors relevant to \(\upalpha \)’s choice.

Taken together, the two assumptions imply that the independent causal factors for \(\upalpha \)’s choice are precisely choices by other agents.

To illustrate these ideas, consider Fig. 3 again. Suppose that the picture represents choices by two agents, \(\upalpha \) and \(\upbeta \), to cooperate in moving a heavy box at moment m. At \({m}/{h}_{1}\) and \({m}/{h}_{2}\), the box is successfully moved (represented by the proposition A). Here, \(\upalpha \) chooses \({\varvec{a}}_{1}\) and \(\upbeta \) chooses \({\varvec{b}}_{1}\). The choice by \(\upbeta \) is causally independent of the choice by \(\upalpha \) (by Strong Independence): if \(\upalpha \) were to choose \({\varvec{a}}_{2}\) (don’t cooperate) at m, then \(\upbeta \) would still choose \({\varvec{b}}_{1}\). Further, this is the only relevant independent causal factor for \(\upalpha \)’s choice (by Causal Completeness).

What defense can we give for assumptions (11) and (12)? For the first, the argument is that from Weak Independence of Agents and the simultaneity of choices by other agents, it is reasonable to infer Strong Independence. Simultaneous choices by agents must be causally independent of each other and of \(\upalpha \)’s choices at m.Footnote 8 By contrast, Causal Completeness is offered merely as a useful “initial approximation” for Horty’s deontic logic. Horty explicitly identifies two sources of independent causal influence that are not reflected in his account: “nonagentive sources” (Nature) and later choices by agents other than \(\upalpha \) (Horty 2001, 89–95). While I agree with Horty that fully to incorporate these influences into the analysis would be a “substantial research task”, I believe that important special cases can be accommodated without great difficulty.

Consider first the case of Nature. Figure 4 illustrates a version of Example 1 (Gambler), described at the start of this chapter.Footnote 9 At moment \(m, \upalpha \) has a choice of gambling (G) or not. The gamble costs $5. A fair coin toss is to be performed at m whether or not \(\upalpha \) gambles. If the coin comes up \({Heads} (H), \upalpha \) leaves with $10, but on Tails (T) she gets nothing. If \(\upalpha \) declines the gamble, she keeps her $5. The outcomes are as shown, where W signifies a win, L a loss and N the status quo where \(\upalpha \) neither wins nor loses.

Fig. 4
figure 4

Gambler (I)

There are no other agents besides \(\upalpha \) in this picture.Footnote 10 Yet the coin toss has the characteristics of an independent causal factor. It occurs simultaneously with \(\upalpha \)’s choice. It satisfies an analogue of Weak Independence: any choice by \(\upalpha \) is compatible with either result, Heads or Tails. Finally, it is reasonable to regard the coin toss as satisfying Strong Independence. Consider

  1. (13)

    If \(\upalpha \) had gambled, he would have won.

We endorse the truth of (13) at \(m/h_{3}\) and \(m/h_{4}\), and its falsity at \(m/h_{7}\) and \(m/h_{8}\). To summarize these observations, we introduce a random variable Toss with values \(\{{Heads}, {\textit{Tails}}\}\). With respect to both Weak Independence and Strong Independence, Toss has characteristics analogous to those of an agent with choice set \(\{{Heads}, {\textit{Tails}}\}\).

The generalization proposed here is to extend Horty’s account of causal independence to include not just agents but also chance mechanisms operating simultaneously with \(\upalpha \)’s choice. By chance mechanisms, I mean well-understood processes such as those employed in games of chance: coin tosses, dice rolls, card drawings and so forth. These are singled out for two reasons. First, such processes have outcomes with well-defined and unproblematic probabilities.Footnote 11 Second, these processes are crucial in defining mixed strategies, which will be important later in this chapter. Each such mechanism can be modeled as a random variable X that may take different possible values \({X} = {x}_{i}\) at the moment m, i.e., at distinct moment history pairs \({m}/{h}\) and \({m}/{h}^{\prime }\). Let VAR be the set of independent variables representing chance processes.Footnote 12 We shall make assumptions about VAR that are entirely parallel to those for AGENT.

First, each random variable X must satisfy an analogue of the No Choice Between Undivided Histories condition, representing the fact that the chance process operates at moment m (rather than at a later moment). A pair of definitions makes this clear.

  1. (14)

    \({Rng}^{m}(X)\).

    By analogy with \({Choice}^{m}_{\upalpha }\), if X is a random variable, define \({rng}^{m}(X)\) as the partition of \({H}_{m}\) corresponding to the possible values \({X} = {x}_\mathrm{i}\) at m. (Two histories \({h}_{1}\) and \({h}_{2}\) belong to the same element of \({rng}^{m}(X)\) if for some \({x}_{i}\), both \({M}, {m/h}_{1} \models {X} = {x}_{i}\) and \({M}, {m/h}_{2} \models {X} = {x}_{i}\).) For simplicity, we shall assume that these values \(x_\mathrm{i}\) are always real numbers.Footnote 13

  2. (15)

    No Separation of Undivided Histories.

    Whenever \(h_{1}\) and \(h_{2}\) are undivided at \(m, X\) has the same value \(X = x_{i}\) at both \(m/h_{1}\) and \(m/h_{2}\). That is, \(h_{1}\) and \(h_{2}\) must belong to the same element of \({rng}^{m}(X)\).

Condition (15) rules out random variables that partition \({H}_{m}\) based on future processes.

Next, we need analogues for (9) Weak Independence and (10) Strong Independence. We want these analogues to apply to agents and chance processes taken together, which motivates the following definitions.

  1. (16)

    FACTOR.

    FACTOR is the union of VAR and the set of random variables representing choices by agents:

    $$\begin{aligned} {\textit{FACTOR}} = AGENT\, \cup \, {\textit{VAR}} \end{aligned}$$
  2. (17)

    Extended Selection Function.

    An extended selection function s at moment m is a mapping from FACTOR into \(H_{m}\) that selects a choice in \({Choice}^{m}_{\upalpha }\) for each agent \(\upalpha \) in AGENT, and an element of \({rng}^{m}(X)\) for each variable X in VAR.

As before, we use the notation \({Select}_{m}\) for the set of all such functions.

It is sometimes convenient to regard the agents in AGENT as random variables, and to represent \({Choice}^{m}_{\upalpha }\) as \({rng}^{m}(\upalpha )\). This allows us to state a compact analogue of (9):

  1. (18)

    Weak Independence of FACTOR.

    For each moment m and each extended selection function s in \({Select}_{m}\),

    $$ \bigcap _{X\in {F\!ACT\!O\!R}} {s(X)\ne \phi .} $$

We also have an obvious formulation of Strong Independence:

  1. (19)

    Strong Independence of FACTOR.

    Let S represent the background state for \(X = x_{i}\) at m/h, i.e.,

    $$ S=\bigcap _{Y\in {F\!ACT\!O\!R}\backslash \{X\}} {s(Y)} , $$

    where \(s(Y)\) is the element of \({rng}^{m}(Y)\) selected for Y at \(m/h\). If a different value \({X} = {x}_{j}\) were selected at m, the other variable values and hence the background state S would remain the same.

    [In particular, if any agent \(\upalpha \) were to make a different choice, all choices by other agents and all variable values would remain the same].Footnote 14

On this account, causal independence extends to chance mechanisms that operate independently of each other and of agents, such as the coin toss in Gambler (I). We acknowledge this modification by extending our earlier definition of stit frames.

  1. (20)

    Extended stit frames.

    An extended stit frame is a structure \(<{Tree}, <,\;\) FACTOR, \(Rng >\) that satisfies all earlier assumptions as well as (15), (18) and (19).

We are not quite done. A separate approach is needed to represent chance mechanisms initiated by agents. Consider a variation of Gambler: \(\upalpha \) tosses the coin if and only if she accepts the gamble; if she declines, there is no coin toss (Fig. 5).

Fig. 5
figure 5

Gambler (II)

In this case, it is inappropriate to model Toss as an independent causal factor. There is no independent partition of \({H}_{m}\); the toss does not happen if \(\upalpha \) declines to gamble. So both Weak Independence and Strong Independence fail.

The difficulty is that while Gambler (II) involves a well-understood chance mechanism that (to paraphrase Horty) represents a source of causality other than the actions of \(\upalpha \), the mechanism does not operate independently of \(\upalpha \) and cannot be modelled as a random variable in VAR. An alternative approach, following Skyrms, is to represent the operation of such mechanisms via conditional chances for outcomes within each causal background context. This approach will be developed below in Sect. 6.

4 Horty’s Dominance Ought

Consider Gambler (I), as illustrated in Fig. 4. Substitute numerical values 10 in place of W (a winning gamble), 0 in place of L (a loss), and 5 in place of N (no gamble). These values represent the money that agent \(\upalpha \) possesses when the dust settles. We can also think of them as utilities that represent \(\upalpha \)’s preferences.

Adding utilities to a stit frame gives us a utilitarian stit frame, defined by Horty as a structure of the form

$$\begin{aligned} <Tree, \,<, AGENT, \,Choice, \,Value>, \end{aligned}$$

where \({Tree}, <, {AGENT}\) and Choice are as in Sect. 2, and Value is a function that assigns a real number \({Value}(h)\) to each history. A utilitarian stit model combines a utilitarian stit frame with an assignment of truth values to propositions. Figure 6 illustrates Gambler (I) in a utilitarian stit model.

Fig. 6
figure 6

Gambler (I) (utilitarian stit model)

Horty provides a semantics for statements of the form

\(\odot [\upalpha \,{ cstit}{:}\; A]\qquad \qquad \qquad (\upalpha \;\mathrm{ought\, to\, see\, to\, it\, that}\; A).\) Footnote 15

The basic idea of his “dominance ought” is that \(\upalpha \) ought to see to it that A iff A is guaranteed by each optimal (non-dominated) choice. It takes care to make this precise and to handle cases where there are no optimal choices. The elements of Horty’s account include background states, a value ordering on propositions at m (subsets of \({H}_{m}\)), and the dominance relation between possible choices for an agent.

  1. (21)

    Dominance ordering on choices.

    • \({State}^\mathrm{m}_{\upalpha }\): the partition of histories through m into background causal contexts S for \(\upalpha \)’s choice. For Horty, as we have seen, these background contexts are simply joint choices by all members of AGENT other than \(\upalpha \).

    • Ordering relations \((\le \, \text {and}\, <)\) on propositions at moment m: If X and Y are two subsets of \({H}_{m}\), then (1) \({X} \le Y\) if \({Value}(h) \le Value(h^{\prime })\) for each \({h} \in X\) and \(h^{\prime } \in \mathrm{Y}\), and (2) \({X} < Y\) if \({X} \le Y\) and in addition, \({Value}(h) < {Value}({h}^{\prime })\) for some \({h} \in X\) and \({h}^{\prime } \in Y\).

    • Dominance relations (\(\preccurlyeq \!\!\cdot \) and \(\prec \!\!\cdot )\) on \({Choice}^\mathrm{m}_{\upalpha }\): If K and \({K}^{\prime }\) are members of \({Choice}^\mathrm{m}_{\upalpha }\) (i.e., possible choices for \(\upalpha \) at m), then (1) \({K} \preccurlyeq \!\!\cdot \,\, {K}^{\prime }\) (\(K^{\prime }\) weakly dominates K) if \({K} \cap S \le K^{\prime } \cap S\) for each state S in \({State}^\mathrm{m}_{\upalpha }\), and (2) \({K} \prec \!\!\cdot \, K^{\prime }\) (\({K}^{\prime }\) strongly dominates K) if \({K} \preccurlyeq \!\!\cdot \, K^{\prime }\) and, in addition, \({K} \cap S < K^{\prime } \cap S\) for some state S in \({State}^\mathrm{m}_{\upalpha }\).

    • Optimal acts. If \({K} \in \,{Choice}^\mathrm{m}_{\upalpha }\) is a possible act for \(\upalpha \) at m, and there is no \({K}^{\prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that \({K} \prec \!\!\cdot \, K^{\prime }\), then K is an optimal act for \(\upalpha \) at m.

To illustrate these ideas, imagine that in Fig. 6, the result Heads or Tails is determined by another agent, \(\upbeta \). In this case, \({ State}_{\upalpha }^\mathrm{m}\) is \(\{{ Heads}, { Tails}\}\), and \( (G\, \& \, { Tails})\) \( < {\sim }{G} < (G \, \& \,{ Heads})\) on the propositional ordering. Neither \([\upalpha \,{ cstit}{:}\; {G}]\) nor \([\upalpha \,{ cstit}{:}\; {\sim }{G}]\) is a dominated act for \(\upalpha \), so both are optimal.

In simple cases where \(\upalpha \) has only finitely many possible choices, Horty’s account of obligation is as follows:

  1. (22)

    Horty Dominance Ought (Finite Choice case).

    \(M, m/h \models \odot [\upalpha \,{ cstit}{:}\,A]\) iff \(M, m/h^{\prime }\models {A}\) for all \({h}^{\prime }\) belonging to any choice K that is optimal at m.

That is, \(\upalpha \) ought to see to it that A iff every optimal choice guarantees A. In the case of Gambler, the Horty account tells us that neither \(\odot [\upalpha { cstit}{:}\;{G}]\) nor \(\odot [\upalpha { cstit}{:}\;{\sim }{G}]\) is true. In the absence of probabilistic information, gambling and not gambling are both permitted.

There may be situations where \(\upalpha \) has infinitely many options, none of which is optimal. For instance, if \(\upalpha \) and \(\upbeta \) are playing the greatest integer game, where the person who names the largest integer wins, then there is no optimal choice. In such cases, there are still dominated choices and hence there are still obligations—for instance, the obligation to choose an integer greater than 1,000. To accommodate such cases, Horty provides a more general evaluation rule.

  1. (23)

    Horty Dominance Ought (general case).

    \({M}, {m}/{h} \models \odot [\upalpha ~{cstit}{:}~{A}]\) iff for each choice \({K} \in {Choice}^\mathrm{m}_{\upalpha }\) that does not guarantee A, there is a choice \({K}^{\prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that (1) \({K} \prec \!\!\cdot \,\, {K}^{\prime }\) (\(K^{\prime }\) strongly dominates K), (2) \(M, m/h^{\prime } \models A\) for all \({h}^{\prime }\) belonging to \({K}^{\prime }\), and (3) for every choice \({K}^{\prime \prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that \({K}^{\prime } \preccurlyeq \!\!\cdot \,\, K^{\prime \prime }, M, m/h^{\prime \prime }\models A\) for all \(h^{\prime \prime }\) belonging to \(K^{\prime \prime }\).

The requirement for \(\odot [\upalpha ~{cstit}{:}\;{A}]\) is that any action K that does not guarantee A is dominated by an action \({K}^{\prime }\) that does guarantee A and is either optimal or dominated only by other actions that guarantee A.

Suppose we modify Gambler so that the values are as shown in Fig. 7. Once again, we imagine that the result Heads or Tails is determined by the choice of another agent, \(\upbeta \), so that the background states for \(\upalpha \)’s choice are \(\{{\textit{Heads}}, {\textit{Tails}}\}\).

As before, gambling is not always better than not gambling: the value of \({h}_{3}\) and \({h}_{4}\) exceeds that of \(h_{5}\) and \(h_{6}\). But this time, the act of gambling dominates the act of not gambling, so that \(\odot [\upalpha ~{cstit}{:}\;{G}]\) is true at m.Footnote 16

As a slight modification of Horty’s account, let us bring in the generalization of causal independence introduced in the extended stit frames of Sect. 3. Suppose that, in Fig. 7, the result Heads or Tails is determined by a chance mechanism (a coin toss) rather than by another agent’s choice. Since Horty’s causal background contexts only take into account choices by other agents, his account gives us no partition between Heads and Tails, no dominance of \([\upalpha {cstit}{:}\;{G}]\) over \([\upalpha ~{cstit}{:}\;{\sim }{G}]\), and hence no obligation to gamble. The extension of AGENT to FACTOR, as explained in Sect. 3, remedies this problem by counting chance mechanisms such as coin tosses as independent causal factors on par with choices. If we allow this extension, then \(State^\mathrm{m}_{\upalpha } = \{{Heads}, {{{Tails}}}\}\) in Fig. 7, which restores the dominance reasoning that leads to \(\odot [\upalpha ~{cstit}{:}\;{G}]\). Henceforth, \(\preccurlyeq \!\!\cdot \) and \(\odot \) will be understood as incorporating this extended concept of factors and background states. Other than the change to \({State}^\mathrm{m}_{\upalpha },\) there is no formal modification required for definitions (21), (22) and (23).Footnote 17

The distinctive feature of Horty’s approach, in contrast to a great deal of earlier work in deontic logic, is that his semantics for obligation is based on an ordering on choices, rather than an ordering on histories or worlds. Horty’s deontic logic gives us a weak decision theory, namely, the part of decision theory that corresponds to dominance reasoning. Let us say that an ordering \(\preccurlyeq \) on choices K in \({Choice}^\mathrm{m}_{\upalpha }\) is admissible if it extends Horty’s dominance ordering: \({K} \preccurlyeq {K}^{\prime }\) whenever \({K} \preccurlyeq \!\!\cdot \,\, K^{\prime }\). Any decision principle based upon an admissible ordering on choices will preserve obligations that hold according to the dominance ought. But the reverse is not true: stronger decision principles justify assertions of obligation that fail on the Horty semantics. The remainder of this chapter shows how three of these stronger decision principles, and the corresponding notions of obligation, can be modeled by extensions within Horty’s framework.

Fig. 7
figure 7

Gambler III (utilitarian stit frame)

5 Decisions Under Ignorance: The Maximin Ought

In this section, I show how Horty’s account might be extended to incorporate a principle that is sometimes used for making decisions under ignorance: maximin. The maximin rule tells the agent to compare minimum utilities possible for each available act, and to choose the act with the maximal minimum utility. The rationale behind this rule is conservatism: by following maximin, the agent guarantees the least bad outcome. In Gambler (I), for instance, the choice “Don’t Gamble” guarantees a utility of 5, while “Gamble” allows possible outcomes with utilities of 0 and 10. Maximin thus prescribes the choice of not gambling. By contrast, as we have just seen, Horty’s dominance ought prescribes nothing, since both gambling and not gambling are optimal choices.

There is an important ambiguity in the phrase “decisions under ignorance.” Commonly, such decisions are characterized as those made “when it makes no sense to assign probabilities to the outcomes emanating from one or more of the acts” (Resnik 1987, 14). We can distinguish between cases of total ignorance, where the agent has no probabilistic information at all (not even knowledge of independence), and ignorance of probabilities, where the agent has no quantitative probabilistic information but does possess knowledge of causal independence. The latter case, in which the agent has exactly the same information as required for the dominance ought, is our focus. The objective is to provide a semantics for maximin ought-to-do, \(O_{m}\), that strengthens Horty’s dominance ought in the following sense:

  1. (24)

    \(\odot [\upalpha ~{cstit}{:}\;{A}] \models O_{m}[\upalpha ~{cstit}{:}\;{A}]\).

Both operators are formulated within utilitarian stit frameworks. The meaning of (24) is that for any utilitarian stit model M and for any \({m},{h}\) pair, if \({M}, {m}/{h} \models \odot [\upalpha ~{cstit}{:}\;{A}]\) then \({M},~{m}/{h} \models O_{m}[\upalpha ~{cstit}{:}\;{A}]\).

Unfortunately, our preliminary statement of the maximin rule is inconsistent with Horty’s dominance ought. To see this, consider a kid-friendly version of Gambler that rewards a decision to gamble with $10 if Heads, $5 if Tails; a decision not to gamble yields $5 regardless of outcome.Footnote 18 Gambling is plainly the dominant act. Yet the simple maximin rule regards gambling and not gambling as equally good because the worst outcome on either choice is $5. This violation of dominance can be avoided by moving to a lexical version of maximin,Footnote 19 but an alternative approach will be offered below.

Another weakness of maximin as stated is its inability to handle a situation of infinite choices, such as the greatest integer game discussed in the preceding section. Even though no available act attains a maximal minimum value, it seems clear that maximin reasoning should license many of the same conclusions as dominance reasoning—for instance, that one ought to select an integer greater than 1,000.

We proceed in stages, starting with a new ordering on choices that combines maximin with the dominance ordering \(\preccurlyeq \!\!\cdot \) defined in (21). The idea is to apply maximin only to pairwise comparisons where neither choice dominates the other.

  1. (25)

    Non-dominance. If K and \(K^{\prime }\) are members of \({Choice}^\mathrm{m}_{\upalpha }\) (i.e., possible choices for \(\upalpha \) at m), write \(K\;{|\!\!\!\!\!\lessgtr }\;K^{\prime }\) if neither \(K\,{\preccurlyeq \!\!\cdot }\,K^{\prime }\) nor \(K^{\prime }\,{\preccurlyeq \!\!\cdot }\,K\).

  2. (26)

    Maximin ordering (\(\preccurlyeq _{m}\) and \(\prec _{m}\)) on \({Choice}^\mathrm{m}_{\upalpha }\):

    If K and \({K}^{\prime }\) are members of \({Choice}^\mathrm{m}_{\upalpha }\) (i.e., possible choices for \(\upalpha \) at m), then

    1. (1)

      \({K} \preccurlyeq _{m}~K^{\prime }\) if (i) \(K\,{\preccurlyeq \!\!\cdot }\,K^{\prime }\) or (ii) \(K\;{|\!\!\!\!\!\lessgtr }\;K^{\prime }\) and \({{{inf}}}\{{Val}(h)/{h}\in K\} \le {{inf}}\{Val(h^{\prime }) / {h}^{\prime }\in {K}^{\prime }\}\); and

    2. (2)

      \({K} \prec _{m} K^{\prime }\) if (i) \({K} \prec \!\!\cdot K^{\prime }\) or (ii) \(K\;{|\!\!\!\!\!\lessgtr }\;K^{\prime }\) and \({inf}\{{Val}({h})/{h}\in {K}\} < {{inf}}\{{Val}({h}^{\prime })/ {h}^{\prime }\in {K}^{\prime }\}\).Footnote 20

  3. (27)

    Maximin ought, \({O}_{m}\).

    \(M, m/h \models O_{m}[\upalpha ~{cstit}{:}\;{A}]\) iff for each choice \({K} \in {Choice}^\mathrm{m}_{\upalpha }\) that does not guarantee A, there is a choice \({K}^{\prime } \in Choice^\mathrm{m}_{\upalpha }\) such that (1) \({K} \prec _{m} {K}^{\prime }\), (2) \({M}, {m}/{h}^{\prime }\models A\) for all \({h}^{\prime }\) belonging to \({K}^{\prime }\), and (3) for every choice \({K}^{\prime \prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that \({K}^{\prime } \preccurlyeq _{m} K^{\prime \prime }, M,~m/{h}^{\prime \prime }\models {A}\) for all \({h}^{\prime \prime }\) belonging to \(K^{\prime \prime }\).

The relationship (24), that \(\odot [\upalpha ~{cstit}{:}\;{A}]\) entails \({O}_{m}[\upalpha ~{cstit}{:}\;{A}]\), is clear because dominance is built into the definition of \({O}_{m}\). By way of example: in the kid-friendly version of Gambler, there is an obligation to gamble because gambling dominates not gambling. In the original version of Gambler (illustrated in Fig. 6) there is no dominant choice, but not gambling is superior to gambling on the maximin ordering; as a consequence, the agent has an obligation not to gamble \((O_{m}[\upalpha ~{cstit}{:}\;{\sim }{G}])\). The same result holds when the result of Heads or Tails is achieved through placement of the coin by an independent agent. Finally, consider an infinite choice situation such as the Greatest Integer Game, where each possible choice of an integer is dominated by any choice of a larger integer. By (27), it is still true that one ought to choose an integer larger than 1,000.

The formulation of maximin ought in (27) has some advantages over the traditional formulation of maximin in decision theory. The first is its compatibility with dominance reasoning. The standard version of maximin, as noted earlier, does not always exclude dominated choices; the same problem applies to some forms of lexical maximin reasoning.Footnote 21 Other versions of lexical maximin, which respect dominance, are defined only for finite choice situations. By contrast, (27) is defined for arbitrary choice situations and is always compatible with dominance reasoning. A second advantage of the present formulation, indeed, is its ability to accommodate infinite choice situations, as noted in the preceding paragraph. In infinite choice situations where no individual choice is rational, we can still identify obligations. This highlights a general advantage of locating decision principles within deontic logic: whereas decision theory is focused specifically on rational acts, deontic logic provides truth conditions for all sentences of the form \({O}_{m}[\upalpha ~{cstit}{:}\;{A}]\).

The point of this discussion is not to endorse the maximin ought over Horty’s dominance ought. The weaknesses of maximin reasoning are well known.Footnote 22 There are two motives for developing \({O}_{m}\). The first is simply to flesh out the claim that the Horty semantics can be strengthened to yield a stronger decision theory. The second is that maximin reasoning plays an important role in game theory (Sect. 7).

6 Decision Under Risk: Probabilistic Utilitarian stit Frames

This section extends Horty’s account in a different direction by incorporating a simple type of probabilistic information – that which is related to chance mechanisms — into the semantics of obligation. This results in a strengthening of the dominance ought that is incompatible with maximin (just as expected utility reasoning is incompatible with maximin reasoning in decision theory). For simplicity, we ignore other agents; we have a single agent, \(\upalpha \), making choices. We assume that \(\upalpha \) has finitely many choices and that there are only finitely many relevant independent causal factors, so both \({Choice}^\mathrm{m}_{\upalpha }\) and \({State}^\mathrm{m}_{\upalpha }\) are finite.

Let us begin with Gambler (I) as depicted in Figs. 4 (without utilities) and 6 (with utilities). Suppose that we have a coin toss with known probabilities 0.5 for Heads and Tails. In the theory of decision under risk, a straightforward application of expected utility reasoning yields a tie: gambling and not gambling have equal expected utility (EU \(=\) 5). A similar analysis yields a tie for Gambler (II) as depicted in Fig. 5, where the coin toss only occurs if \(\upalpha \) decides to gamble. But it is clear in these examples that slight changes to the utilities would tip the decision one way or the other. To extend Horty’s account to such cases, we need to add some concepts to utilitarian stit frames. It suffices to add two additional concepts: outcomes and conditional chances.

We shall assume a finite set \({O}_{1}, {\ldots }, O_{n}\) of outcomes of interest. These are propositions at moment m (subsets of \({H}_{m}\)) that constitute a partition of \({H}_{m}\) and which, in conjunction with the background contexts and the agent’s choices, influence the assignment of conditional chance and utility. In particular, they allow us to represent probabilistic information about chance processes initiated by agents; such processes cannot be treated as independent causal factors, as explained at the end of Sect. 3. In Gambler, the outcomes may be described as \(\{Win, {Lose}, {Neither}\}\).

For the conditional chance function on \({H}_{m}\), the simplest approach is to take the underlying algebraFootnote 23 of subsets of \({H}_{m}\) to consist of all finite unions of sets \({K}_{i} \wedge {S}_{j} \wedge O_{k}\), where \(K_{1},{\ldots }, K_{m}\) are available acts, \(S_{1},{\ldots }, S_{M}\) are the background contexts, and \(O_{1}, {\ldots }, O_{n}\) are the outcomes. Probabilistic information about chance mechanisms is given by a conditional chance function P, assigning values \(P(O_{k} / K_{i} \wedge S_{j})\) that we take as primitive. P must satisfy the standard axioms of the probability calculus. Since the algebra is finite, P need only be finitely additive.Footnote 24

The following two assumptions would allow easy extension of Horty’s framework to handle probabilistic choices:

  1. (a)

    Uniform conditional chances. \({P}({O}_{k} / {K}_{i} \, \wedge {S}_{j})\) is constant for each relevant outcome \({O}_{1},{\ldots }, {O}_{n}\) within each \({K}_{i} \wedge {S}_{j}\).

  2. (b)

    Uniform utilities. \({Val}(h) = {Val}(h^{\prime })\) for all \({h},\;{h}^{\prime }\,\in \,{K}_{i} \wedge {S}_{j} \wedge {O}_{k}\), for all \({i}, {j}, {k}\).

Given these assumptions, we can “import” decision theory into the stit framework. We can use expected utility maximization as the criterion for what agent \(\upalpha \) ought to do at m, in simple cases such as Gambler (I) and (II).

But these assumptions need not always hold. First, assumption (a) might fail. Some of the objective chances used in decision theory are not conditional chances associated with chance mechanisms. For example, they may be derive from observed frequencies. So there may be information about objective conditional chances that is not represented in the stit framework. Second, assumption (b) might fail. Within a single choice-state combination, we might find histories with different utilities based (for example) on future choices by agents or future events. In general: if it is impossible to find a set of outcomes satisfying (a) and (b), then it is impossible to apply straightforward expected utility maximization. To keep matters simple, however, I shall assume that condition (a) is satisfied but that (b) may fail.

This leads us to a definition of probabilistic utilitarian stit frames.

  1. (28)

    A probabilistic utilitarian stit frame is a structure of the form.

    $$\begin{aligned} <Tree, \,<, {\textit{FACTOR}},\, Rng, \,Value,\, Outcome,\, P>, \end{aligned}$$

    where \({Tree}, <, {{\textit{FACTOR}}}, {Rng}\) and Value are as in Sects. 23, and 4, Outcome is a function that assigns to each moment m a partition \(\{O_{1}, {\ldots }, O_{n}\}\) of \(H_{m}\) and \({P}( \cdot / \cdot )\) is a conditional probability function that assigns a value \(P(O_{k} / K_{i} \wedge S_{j})\) for each outcome \(O_{k}\), choice \(K_{i}\) and state \(S_{j}\).Footnote 25

    A probabilistic utilitarian stit model combines a probabilistic utilitarian stit frame with an assignment of truth values to propositions.

In order to formulate the concept of obligation in probabilistic utilitarian stit frames, consider the following case (Fig. 8). In this example, the coin toss is an independent factor and, as usual, Heads and Tails have fixed conditional chances of 0.5 regardless of whether \(\upalpha \) gambles. The outcomes are Win, Lose and Neither, but this time the utilities of Win and Lose are not fixed (i.e., assumption (b) fails). So there is no sharp value for the expected utility of gambling. Still, we can see that the expected utility of gambling is at least (0.5)(9) \(+\) (0.5)(2) \(=\) 5.5, which exceeds the expected utility of not gambling. Thus, we ought to gamble.

This motivates the following account of obligation, replacing dominance with dominating expectation in the Horty semantics. We continue to assume that both Choice \(^\mathrm{m}_{\upalpha }\) and State \(^\mathrm{m}_{\upalpha }\) are finite.

  1. (29)

    Dominating expectation ordering (\(\preccurlyeq _{d}\) and \(\prec _{ d})\) on \({Choice}^\mathrm{m}_{\upalpha }\).

    If K and \({K}^{\prime }\) are members of \({Choice}^\mathrm{m}_{\upalpha }\) (i.e., possible choices for \(\upalpha \) at m), then

    1. (1)

      \({K} \preccurlyeq _{d} {K}^{\prime }\) if for all choices \({h_\mathrm{{jk}}}, {h_\mathrm{{jk}}}^{\prime }\) of histories in \({K} \cap {S}_{j} \cap O_{k}\) and \(K^{\prime } \cap S_{j} \cap O_{k}\) respectively, \(\Sigma _\mathrm{{j,k}} {Val}(h_{jk}) P(O_{k} / {K}\wedge S_{j}) \le \Sigma _\mathrm{{j,k}}{Val}({{h}_{jk}}^{\prime }){P}({O}_{k} / {K}^{\prime } \wedge {S}_{j})\); and

    2. (2)

      \({K} \prec _{d} {K}^{\prime }\) if \({K} \preccurlyeq _{d} {K}^{\prime }\) but not \({K}^{\prime } \preccurlyeq _{d} K\).

This principle says that act \({K}^{\prime }\) is better than act K if the expectation of \({K}^{\prime }\) dominates the expectation of K.

Corresponding to this new ordering on choices, we have a new concept of obligation defined analogously to the earlier definitions (23) and (27). It states, roughly, that \({O}_{d}[\upalpha {cstit}{:}\; {A}]\) if A is guaranteed by all choices whose expectation is not dominated.

  1. (30)

    Dominating expectation ought, O\(_{d}\).

    \({M}, {m}/{h}\, \models {O}_{d}[\upalpha ~{cstit}{:}\;{A}]\) iff for each choice \({K} \in {Choice}^\mathrm{m}_{\upalpha }\) that does not guarantee A, there is a choice \({K}^{\prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that (1) \({K} \prec _{d} {K}^{\prime }\), (2) \({M}, {m}/{h}^{\prime }\models {A}\) for all \({h}^{\prime }\) belonging to \({K}^{\prime }\), and (3) for every choice \({K}^{\prime \prime } \in {Choice}^\mathrm{m}_{\upalpha }\) such that \({K}^{\prime } \preccurlyeq _{d} {K}^{\prime \prime }, M,~{m}/{h}^{\prime \prime }\models A\) for all \(h^{\prime \prime }\) belonging to \(K^{\prime \prime }\).

It should be clear that \(O_{d}\) is a strengthening of Horty’s \(\odot \), since \(K \preccurlyeq _{d} K^{\prime }\) whenever \(K \preccurlyeq \!\!\cdot \,\, K^{\prime }\).Footnote 26 On the other hand, just as we would expect, \({O}_{d}\) diverges from the maximin ought \({O}_{m}\). In the version of Gambler depicted by Fig. 8, we have \({O}_{d}[\upalpha {cstit}{:}\; {G}]\) while \({O}_{m}[\upalpha {cstit}{:}\;{\sim }{G}]\).

Fig. 8
figure 8

Gambler (IV)

7 Game Theory and Mixed Strategies

Finally, I consider briefly how Horty’s account might be extended to handle obligations in the setting of game theory. To do this in general would introduce many complications, including the need for separate Value functions to keep track of each agent’s utilities.Footnote 27 My interest here lies mainly in showing how we might use the probabilistic utilitarian stit frames of the preceding section to make sense of mixed strategies in game theory. For this reason, I limit the discussion to two-person zero-sum games. The Value function represents \(\upalpha \)’s utilities, while utilities for the other agent, \(\upbeta \), are exactly the negative of \(\upalpha \)’s utilities. We initially assume that finitely many choices—pure strategies in game theory—are available to both agents, and that there are no additional independent causal factors. Thus, the background contexts in \({State}^{m}_{\upalpha }\) are simply \(\upbeta \)’s possible choices.

By way of motivation, notice that a very simple game, Matching Pennies (introduced at the outset of this chapter), generates difficulties for Horty’s account. In this game, both \(\upalpha \) and \(\upbeta \) simultaneously display a coin with one side up. If the two displayed sides match (both Heads or both Tails), then \(\upbeta \) pays $1 to \(\upalpha \); if the sides do not match, then \(\upalpha \) pays $1 to \(\upbeta \). The situation is illustrated in Fig. 9, with the stit and game-theoretic representations side by side.

Fig. 9
figure 9

Matching pennies (a) stit picture; (b) game theory

In this situation, neither choice by \(\upalpha \) is optimal. Consequently, on Horty’s account, we then have neither \(\odot [\upalpha {cstit}: {Head}\mathrm{s}]\) nor \(\odot [\upalpha {cstit}{:}\;{{{Tails}}}]\). That is reasonable if the only choices available are the pure strategies [Display] Heads or [Display]Tails. But Horty’s conclusion is not plausible if we allow mixed strategies of the form

Display Heads with probability p and Display Tails with probability 1 – p, abbreviated as

$$\begin{aligned}{}[p \,Heads, (1 - p) \,{\textit{Tails}}] \end{aligned}$$

or more simply as

$$\begin{aligned} p\, Heads. \end{aligned}$$

In game theory, this type of problem is solved by finding a Nash equilibrium: a pair of choices such that neither player can do better by unilaterally changing his or her choice. In Matching Pennies, there is a unique Nash equilibrium where both agents adopt the mixed strategy: 1/2 Heads. This is the unique rational choice on the assumption that each player has full knowledge of the game and adopts the best possible strategy. In the remainder of this section, we suggest one way in which Horty’s account can be expanded to accommodate mixed strategies, and then propose a semantics that yields the obligation to adopt an equilibrium strategy.

But first let’s consider a preliminary question. How well does Horty’s account fare if we limit ourselves to two-person zero-sum games with only pure strategies? Consider the following game (Fig. 10), with utilities for \(\upalpha \) shown.

Fig. 10
figure 10

A zero-sum game

Here, \(\upalpha \) chooses between the left \((A_{1})\) and right \((A_{2})\) columns, while \(\upbeta \) chooses between the top \((B_{1})\) and bottom \((B_{2})\) rows. From \(\upalpha \)’s point of view, neither choice is dominant: \({A}_{1}\) does better if \(\upbeta \) chooses \({B}_{1}\), while \({A}_{2}\) does better if \(\upbeta \) chooses \({B}_{2}\). So Horty’s “dominance ought” yields no obligation: neither \(\odot [\upalpha {cstit}{:}\;{A}_{1}]\) nor \(\odot [\upalpha {cstit}{:}\;{A}_{2}]\) is true. However, both players will recognize that the top row is the dominant choice for \(\upbeta \) (whose utilities are the negative of those in Fig. 10). Given that \(\upbeta \) will choose \({B}_{1}, \upalpha \) ought to choose \(A_{1}\) (Indeed, \(A_{1}\) and \(B_{1}\) constitute the unique Nash equilibrium for this game). This example shows that even without mixed strategies, Horty’s dominance ought is inadequate for game theory.

One promising possibility might be the maximin ought \((O_{m})\) of Sect. 5, which combines dominance with maximin reasoning. Applied to Fig. 10, \(O_{m}\) gives the correct result: \(A_{1}\) guarantees the maximal minimum, so \(O_{m}[\upalpha {cstit}{:}\;{A}_{1}]\). Further encouragement comes from a standard result of game theory (Resnik 1987, p. 130):

Minimax equilibrium test. In a two-person zero-sum game, a necessary and sufficient condition for a pair of (pure) strategies to be in (Nash) equilibrium is that the payoff determined by them equal the minimal value of its (column) and the maximal value of its (row).Footnote 28 The values for all such equilibrium pairs are the same.

It is easy to establish the following proposition:

  1. (31)

    Proposition: If there is a pure-strategy equilibrium pair in a two-person zero-sum game and if A is true at all non-dominated choices \(K\) for \(\upalpha \) that belong to such an equilibrium pair, then \(O_{m}[\upalpha {cstit}{:}\;{A}]\).

    Proof:

    For each such \({K}, {K}^{\prime }\,\preccurlyeq _{m}\,{K}\) for all choices \({K}^{\prime }\); thus, these K are optimal with respect to the ordering \(\preccurlyeq _{m}\) and we have \({O}_{m}[\upalpha {cstit}{:}\;{A}]\).

(31) shows that whenever there is a pure Nash equilibrium, the maximin ought correctly prescribes choices that are part of such an equilibrium.

Despite this success, maximin ought appears to overshoot the mark. It prescribes choices even when there is no pure equilibrium. Consider the following example (Fig.11). Here there are no dominant choices for either player and no equilibrium. Nevertheless, maximin ought prescribes \({A}_{2}\) for \(\upalpha \) and \(B_{1}\) for \(\upbeta \), since these choices maximize minimal utility.

Fig. 11
figure 11

Another zero-sum game

It might be interesting to consider whether there is any merit to these prescriptions. It might also be worth investigating whether there is a notion of obligation, intermediate in strength between \(\odot \) and \({O}_{m}\), that corresponds precisely to acts that comprise a Nash equilibrium. I pass over such investigations for the following reason: once we allow mixed strategies, the problem of capturing Nash equilibria with a Horty-style account of obligation is solved through an interesting combination of maximin reasoning and the weak concept of expected utility introduced in Sect. 6.

The first task is to give an analysis of mixed strategies. In game theory, a mixed strategy is commonly characterized as the use of a chance mechanism to select a pure strategy, followed by acting on the selected strategy. The details may not matter much in game theory, but they matter a great deal in the stit framework. If the chance mechanism operates at a moment prior to the choice of the pure strategy, then the analysis of a mixed strategy will involve both the prior moment when the mechanism operates and alternative later moments at which the agent chooses a pure strategy. To make things worse, the stit picture for the later moment will be identical to the original ‘pure strategy’ picture. If we evaluate the obligation at that later moment, it is unclear how the earlier operation of a chance mechanism can make any difference.

Perhaps the simplest approach, and the one which will be adopted here, is to represent each available mixed strategy as a separate choice existing at the same moment as the pure strategies. It is the choice of a chance mechanism whose possible outcomes are identical in structure with the pure strategies. Strictly speaking, this analysis requires that we modify Choice \(^\mathrm{m}_{\upalpha }\) by adding one additional choice for each available chance distribution over the (finitely many) pure strategies. In practice, it usually suffices to represent all of the pure strategies plus a single choice that stands for an arbitrary mixed strategy (incorporating probabilistic parameters) or, on occasion, for a particular mixed strategy. Figure 12 illustrates Matching Pennies with mixed strategies. Dotted lines are used to separate outcomes for the case of choices that involve chance mechanisms.

Fig. 12
figure 12

Matching pennies with mixed strategies

None of the concepts of obligation described above gives the correct result here, namely, the obligation to choose 1/2 Heads. According to the dominance ought, there are no obligations. The same is true for the maximin ought, since each choice has the same worst case. The dominating expectation ought is not even defined for settings involving multiple agents.

A helpful way to obtain the right choice ordering and the right concept of obligation is to exploit a well-known result from game theory (Resnik 1987, p. 136):

Maximin Theorem for two-person zero-sum games.

For every two-person zero-sum game there is at least one strategy (mixed or pure) for Row and at least one strategy for Col that form an equilibrium pair. If there is more than one such pair, their expected utilities are equal.

The expected utility for the equilibrium pair is referred to as the security level for both players because, by playing the equilibrium strategy, each player maximizes his or her minimum expected utility. The security level in Matching Pennies is 1/2, which can be guaranteed by playing 1/2 Heads. In contrast to our earlier discussion of zero-sum games with pure strategies, the inclusion of mixed strategies ensures the existence of an equilibrium.

The right ordering, then, is that one mixed strategy is preferable to another if its minimal expected utility exceeds that of the other. This ordering can be defined within the setting of probabilistic utilitarian stit frames. Write \(P_{\upalpha }\) and \(P_{\upbeta }\) for \(\upalpha \)’s and \(\upbeta \)’s choice of mixed strategy. \({P}_{\upalpha }\) and \({P}_{\upbeta }\) are probability distributions over the choices available to \(\upalpha \) and \(\upbeta \), respectively. That is, if \({K}_{1},{\ldots }, {K}_{m}\) are the available pure strategies for \(\upalpha \), i.e., the members of Choice \(^\mathrm{m}_{\upalpha }\), then \({P}_{\upalpha }(K_{i}) = {p}_{i}\) with \(\Sigma {p}_{i} = 1\); similarly, \(P_{\upbeta }(B_{j}) = {q}_{j}\) for pure strategies \(B_{1},{\ldots }, B_{n}\). The choice of a pure strategy \(K_{i}\) or \(B_{j}\) is just the special case where \(p_{i} = 1\) or \({q}_{j} = 1\). Let \(^{*}\) Choice \(^\mathrm{m}_{\upalpha }\) be the set of mixed strategies \({P}_{\upalpha }\) based on the pure strategies in Choice \(^\mathrm{m}_{\upalpha }\).

  1. (32)

    Equilibrium ordering (\(\preccurlyeq _{e}\) and \(\prec _{e})\) on \(^{*}\) Choice \(^\mathrm{m}_{\upalpha }\).

    If \({P}_{\upalpha }\) and \({P}_{\upalpha }^{\prime }\) are members of \(^{*}\) Choice \(^\mathrm{m}_{\upalpha }\) (i.e., mixed strategies for \(\upalpha \) at m), then

    1. (1)

      \({P}_{\upalpha } \preccurlyeq _{e} {P}_{\upalpha }^{\prime }\) if \(\text {inf}\{\Sigma _\mathrm{{i,j}}{Val}({h}_{ij}){P}_{\upalpha }({K}_{i}) {P}_{\upbeta }({B}_{j}) / {h}_{ij} \in {K}_{i} \cap B_{j}\) and \(P_{\upbeta }\) a mixed strategy for \(\upbeta \}\) \(\le \,\text {inf}\{\Sigma _\mathrm{{i,j}}{Val}({h}_{ij}){P}_{\upalpha }^{\prime }({K}_{i}) {P}_{\upbeta }({B}_{j}) / {h}_{ij} \in {K}_{i} \cap {B}_{j}\) and \({P}_{\upbeta }\) a mixed strategy for \(\upbeta \}\);

      and

    2. (2)

      \({P}_{\upalpha } \prec _{e} {P}_{\upalpha }^{\prime }\) if \({P}_{\upalpha } \preccurlyeq _{e} P_{\upalpha }^{\prime }\) but not \(P_{\upalpha }^{\prime } \preccurlyeq _{e} P_{\upalpha }\).

The mixed strategy \(P_{\upalpha }^{\prime }\) is better than \(P_{\upalpha }\) if it has greater minimal expected utility. The ordering \(\preccurlyeq _{e}\) is admissible in the following special sense: if \({P}_{\upalpha }\,{\preccurlyeq \!\!\cdot }\,{P}_{\upalpha }^{\prime }\) where both \(P_{\upalpha }\) and \(P_{\upalpha }^{\prime }\) are pure strategies, then \(P_{\upalpha }~\preccurlyeq _{e}~{P}_{\upalpha }^{\prime }\).Footnote 29

  1. (33)

    Equilibrium ought, \(O_{e}\).

    \(M, m/h \, \models O_{e}[\upalpha {cstit}{:}\;{A}]\) iff for each \({P}_{\upalpha } \in {^{*}{Choice}}^\mathrm{m}_{\upalpha }\) that does not guarantee A, there is a \({P}_{\upalpha } ^{\prime } \in {^{*}{Choice}}^\mathrm{m}_{\upalpha }\) such that (1) \({P}_{\upalpha } \prec _{e} {P}_{\upalpha }^{\prime }\), (2) M, \({m}/{h}^{\prime }\models {A}\) for all \({h}^{\prime }\) belonging to \({P}_{\upalpha }^{\prime }\), and (3) for every \({P}_{\upalpha }^{\prime \prime } \in {^{*}{Choice}}^\mathrm{m}_{\upalpha }\) such that \(P_{\upalpha } ^{\prime } \preccurlyeq _{e} P_{\upalpha } ^{\prime \prime }, M, m/{h}^{\prime \prime }\models {A}\) for all \({h}^{\prime \prime }\) belonging to \({P}_{\upalpha }^{\prime \prime }\).

In two-person zero-sum games where an equilibrium exists, (33) states that \({O}_{e}[\upalpha {cstit}{:}\;{A}]\) if A is guaranteed by all equilibrium mixed strategies. \({O}_{e}\) is a strengthening of Horty’s \(\odot \), and it gives the right answer in the case of Matching Pennies: \({O}_{e}[\upalpha {cstit}{:}\;{P}_{1/2}]\).

Summarizing: mixed strategies can be defined in Horty’s framework, and we can give an ordering on mixed strategies that yields the correct account of what agents ought to do in two-person zero-sum games. Extending these ideas to games involving more than two agents and to cooperative games may or may not be feasible.

8 Conclusion

Horty observes that his account of obligation “closes the gap” between deontic logic and act utilitarianism. That gap existed so long as deontic logic was viewed as an account of classifying states of affairs as right or wrong, while utilitarianism was concerned with classifying actions. Horty’s dominance ought clearly goes a long way towards closing another gap as well: the one between deontic logic and decision theory.

Because of the weakness of dominance reasoning, however, Horty’s account seems of limited value as a theory of choice. This chapter suggests how, with modest extensions, Horty’s framework can move beyond dominance into the three main branches of the theory of decision: decisions under ignorance, decisions under risk and game theory. This leads to a motivational question: what is the point of trying to bring deontic logic “up to speed” if we already have a successful decision theory? I close by suggesting two main ways in which deontic logic provides return benefits for decision theory.

The first, noted at the outset of this chapter, is by offering rigorous analysis of foundational notions: causation, choice, counterfactuals and background states. That such analyses matter should be clear to anyone who has followed the history of decision theory as formulated by Savage, modified by Jeffrey, and re-formulated by causal decision theorists. For example, we claimed here that the states of decision theory are causal background contexts and provided an analysis of causal independence and background contexts within stit models. By contrast, (Joyce (1999), 61) writes that states include all “aspects of the world that lie outside the decision maker’s control”. He tells us that future choices and events, if relevant to our present decision problem, must be incorporated into the background states for that decision. Now it is harmless to incorporate future choices and events into the background states if they are causally independent of the agent’s present choice, but not so harmless if their future occurrence is contingent upon present choices. Stit frames take care of this automatically: histories belonging to distinct states at m must be divided at m. This rules out treating future choices or processes as constituents of states at m. Future chance processes must be incorporated into decisions via conditional chances for outcomes (as described in Sect. 6). To handle sequential choices in the stit framework requires something like Horty’s strategic ought (2001, Chap. 7), which takes us beyond the present discussion.

As a second benefit, deontic logic offers a model for thinking about problems where decision theory and game theory cannot offer clear, uncontroversial solutions. One source of such problems is infinite decision theory, comprising decision problems in which an agent has to deal with infinite utilities, an infinity of possible acts, or both.Footnote 30 Some of these problems are genuinely paradoxical and have no clear solution. In other cases, however, there are clear prescriptions, yet decision theory is silent because there is no optimal act. Because deontic logic is concerned with the truth of obligation sentences \(O[\upalpha {cstit}{:}\;{A}]\) even where A does not describe an act, it has the resources to offer advice in such cases.

One example of this kind, noted earlier, is the greatest integer game. Decision theory cannot recommend the choice of any particular integer, but our deontic logics tell us that \(\odot [\upalpha {cstit}{:}\;{A}_{n}]\) and \({O}_{m}[\upalpha {cstit}{:}\;{A}_{n}]\) where \({A}_{n}\) is the proposition “\(\upalpha \) chooses an integer larger than n”. As a similar example, imagine that \(\upalpha \) is a perfectionist attempting to finish a journal article. Suppose that \(\upalpha \) represents his position to himself as an infinite choice situation where \({A}_{0}\) stands for not submitting the chapter at all, \({A}_{1}\) for submitting the current version as is, and \({A}_{2}, A_{3},{\ldots }\) for producing and submitting polished versions, each \(A_{n+1}\) slightly better than \(A_{n}\). Suppose that all of the relevant utilities are bounded above by some fixed limit. In such a case, no act \(A_{n}\) is optimal. But our deontic logics still give us \(\odot [\upalpha {cstit}{:}\;{\sim }{A}_{0}]\) and \({O}_{m}[\upalpha {cstit}{:}\;{\sim }{A}_{0}]\), representing the obligation to submit the chapter.

Decision theory need not always be concerned about the metaphysical details of choice, or the precise characterization of the acts and background states needed to specify a decision problem. But at the boundaries of decision theory, where those details matter, stit-based deontic logic, made possible thanks to Belnap’s rigorous analysis of agency, provides a wonderful resource.