# What is conditionalization, and why should we do it?

## Abstract

Conditionalization is one of the central norms of Bayesian epistemology. But there are a number of competing formulations, and a number of arguments that purport to establish it. In this paper, I explore which formulations of the norm are supported by which arguments. In their standard formulations, each of the arguments I consider here depends on the same assumption, which I call Deterministic Updating. I will investigate whether it is possible to amend these arguments so that they no longer depend on it. As I show, whether this is possible depends on the formulation of the norm under consideration.

One of the central tenets of traditional Bayesian epistemology is Conditionalization. There are various formulations of this norm, but they all agree that it concerns the way your credences should change in response to evidence. I spell out the three formulations that I’ll consider below. On the first, Conditionalization concerns how you *actually* update your credences when you receive a piece of evidence; on the second, it concerns how you are *disposed* to update when you receive evidence; and on the third, it concerns how you *plan* to update. In this paper, I am concerned not so much with which formulation of the norm is *correct*—after all, they are not incompatible with each other, and some are independent of each other. Rather, I am concerned with which formulation is *justified* by the existing arguments.

I consider three versions of Conditionalization, and four arguments in their favour. For each combination, I’ll ask whether the argument can support the norm when it is formulated in that way. In each case, I note that the standard version of the argument relies on a particular assumption, which I call *Deterministic Updating* and which I formulate precisely below. I’ll ask whether the argument really does rely on this assumption, or whether it can be amended to support the norm without that assumption.

This is important because Deterministic Updating says that your updating plan or disposition should specify, for any piece of evidence you might receive, a unique way to update on it. But this seems unmotivated, at least from the Bayesian point of view. After all, subjective Bayesianism is a very permissive theory when it comes to your initial credence function, that is, the one you have at the beginning of your epistemic life before you’ve gathered any evidence. But, once that initial credence function is chosen from the wide array that Bayesianism deems permissible, the theory is very restrictive about how you should update your credences upon receipt of new evidence. We tolerate this discrepancy because the same sorts of argument seem to give us both the permissiveness of Probabilism and the restrictiveness of Conditionalization. But if it turns out that these arguments only give the latter when we make an unmotivated assumption, this spells trouble for Bayesianism.^{1}

I don’t claim that the four arguments I consider here exhaust the putative justifications of Conditionalization. Besides these, there are decision-theoretic arguments by Savage (1954, Section 3.5), arguments from symmetry considerations by van Fraassen (1989, Section 13.2) and Grove and Halpern (1998), and arguments from a principle of minimal change due to Diaconis and Zabell (1982, Section 5.1) and Dietrich et al. (2016). I focus on the four described here in the interests of space and because these four arguments are naturally grouped together. We might call them the teleological arguments for Conditionalization, for they seek to establish that norm by pointing to ways that updating in the way it demands optimises different aspects of the goodness of your credences, whether that is their pragmatic utility or their epistemic utility. I leave the question of how the alternative, non-teleological justifications of Conditionalization relate to the assumption of Deterministic Updating to future work.^{2}

I start in Sect. 1 by presenting the various formulations of the norm precisely. Then I introduce the four arguments informally. Then, in Sect. 2, I introduce some of the formal machinery required to state the arguments. Sections 3, 4, 5 and 6 contain the central results of the paper. In those sections, I work through each of the four arguments in turn, provide its standard presentation, which assumes Deterministic Updating, and then ask whether we can do without that assumption. As we’ll see, for one of the arguments, we cannot do without Deterministic Updating; for the other three, if we drop Deterministic Updating, we face a choice—if we go one way, we can justify the three formulations of Conditionalization without assuming Deterministic Updating; if we go the other way, we cannot. In Sect. 7, I ask what lessons we can learn from these results.

## 1 Three formulations and four arguments

According to the second, Plan Conditionalization, the norm governs the updating behaviour you would endorse in all possible evidential situations you might face.

Actual Conditionalization (AC)If

- (i)

cis your credence function att(I’ll often refer to this as yourprior);- (ii)
the total evidence you receive between

tand \(t'\) comes in the form of a propositionElearned with certainty;- (iii)
\(c(E) > 0\);

- (iv)
\(c'\) is your credence function at the later time \(t'\) (I’ll often refer to this as your

posterior);then it should be the case that \( c'(-) = c(-|E) = \frac{c(-\ \& \ E)}{c(E)}\).

According to the third formulation, Dispositional Conditionalization, the norm governs the updating behaviour you are disposed to exhibit.

Plan Conditionalization (PC)If

- (i)

cis your credence function att;- (ii)
the total evidence you receive between

tand \(t'\) will come in the form of a proposition learned with certainty, and that proposition will come from the partition \({\mathcal {E}}= \{E_1, \ldots , E_n\}\);^{3}- (iii)

Ris the plan you endorse for how to update in response to each possible piece of total evidence,then it should be the case that, if you were to receive evidence \(E_i\) and if \(c(E_i) > 0\), then

Rwould exhort you to adopt credence function \( c_i(-) = c(-|E_i) = \frac{c(-\ \& \ E_i)}{c(E_i)}\).

Next, let’s meet the four arguments. Since it will take some work to formulate them precisely, I will give only an informal gloss here. There will be plenty of time to see them in high-definition in what follows.

Dispositional Conditionalization (DC)If

- (i)

cis your credence function att;- (ii)
the total evidence you receive between

tand \(t'\) will come in the form of a proposition learned with certainty, and that proposition will come from the partition \({\mathcal {E}}= \{E_1, \ldots , E_n\}\);- (iii)

Ris the plan you are disposed to follow in response to each possible piece of total evidence,then it should be the case that, if you were to receive evidence \(E_i\) and if \(c(E_i) > 0\), then

Rwould exhort you to adopt credence function \( c_i(-) = c(-|E_i) = \frac{c(-\ \& \ E_i)}{c(E_i)}\).

Diachronic Dutch book or Dutch strategy argument (DSA)This purports to show that, if you violate Conditionalization, there is a pair of decisions you might face, one before and one after you receive your evidence, such that your prior and posterior credences lead you to choose options when faced with those decisions that are guaranteed to be worse by your own lights than some alternative options (Lewis 1999).

Expected pragmatic utility argument (EPUA)This purports to show that, if you will face a decision after learning your evidence, then your prior credences will expect your updated posterior credences to do the best job of making that decision only if they are obtained by conditionalizing on your priors (Savage 1954; Good 1967; Brown 1976).

Expected epistemic utility argument (EEUA)This purports to show that your prior credences will expect your posterior credences to be best epistemically speaking only if they are obtained by conditionalizing on your priors (Greaves and Wallace 2006).

Epistemic utility dominance argument (EUDA)This purports to show that, if you violate Conditionalization, then there will be alternative priors and posteriors that are guaranteed to be better epistemically speaking, when considered together, than your priors and posteriors (Briggs and Pettigrew 2018).

## 2 The framework

In the following sections, I will consider each of the arguments listed above. As we will see, these arguments are concerned directly with updating plans or dispositions, rather than actual updating behaviour. That is, the targets of these arguments—the items that they assess for rationality or irrationality—don’t just specify how you in fact update in response to the particular piece of evidence you actually receive. Rather, they assume that your evidence between the earlier and later time will come in the form of a proposition learned with certainty (Certain Evidence); they assume the possible propositions that you might learn with certainty by the later time form a partition (Evidential Partition); and they assume that each of the propositions you might learn with certainty is one about which you had a prior opinion (Evidential Availability); and then they specify, for each of the possible pieces of evidence in your evidential partition, how you might update if you were to receive it.

Some philosophers, like Lewis (1999), assume that all three of these assumptions—Certain Evidence, Evidential Partition, and Evidential Availability—hold in all learning situations. Others deny one or more. For instance, Jeffrey (1992) denies Certain Evidence and Evidential Availability; Konek (2019) denies Evidential Availability but not Certain Evidence; van Fraassen (1999), Schoenfield (2017), and Weisberg (2007) deny Evidential Partition. But all agree, I think, that there are certain important situations when all three assumptions are true; there are certain situations where there is a set of propositions that forms a partition and about each member of which you have a prior opinion, and the possible evidence you might receive at the later time comes in the form of one of these propositions learned with certainty. Examples might include: when you are about to discover the outcome of a scientific experiment, perhaps by taking a reading from a measuring device with unambiguous outputs; when you’ve asked an expert a yes/no question; when you step on the digital scales in your bathroom or check your bank balance or count the number of spots on the back of the ladybird that just landed on your hand. So, if you disagree with Lewis, simply restrict your attention to these cases in what follows.

As we will see, we can piggyback on conclusions about plans and dispositions to produce arguments about actual behaviour in certain situations. But in the first instance, I will take the arguments to address plans and dispositions defined on evidential partitions primarily, and actual behaviour only secondarily. Thus, to state these arguments, I need a clear way to represent updating plans or dispositions. I will talk neutrally here of an updating rule. If you think Conditionalization governs your updating dispositions, then you take it to govern the updating rule that matches those dispositions; if you think it governs your updating intentions, then you take it to govern the updating rule you intend to follow.

I’ll introduce a slew of terminology here. You needn’t take it all in at the moment, but it’s worth keeping it all in one place for ease of reference.

*Agenda* I will assume that your prior and posterior credence functions are defined on the same set of propositions \({\mathcal {F}}\), and I’ll assume that \({\mathcal {F}}\) is finite and \({\mathcal {F}}\) is an algebra. We say that \({\mathcal {F}}\) is your *agenda*.

*Possible worlds* Given an agenda \({\mathcal {F}}\), the set of possible worlds relative to \({\mathcal {F}}\) is the set of classically consistent assignments of truth values to the propositions in \({\mathcal {F}}\). I’ll abuse notation throughout and write *w* for (i) a truth value assignment to the propositions in \({\mathcal {F}}\), (ii) the proposition in \({\mathcal {F}}\) that is true at that truth value assignment and only at that truth value assignment, and (iii) what we might call the omniscient credence function relative to that truth value assignment, which is the credence function that assigns maximal credence (i.e. 1) to all propositions that are true on it and minimal credence (i.e. 0) to all propositions that are false on it.

*Updating rules*An

*updating rule*has two components:

- (i)
a set of propositions, \({\mathcal {E}}= \{E_1, \ldots , E_n\}\)

this contains the propositions that you might learn with certainty at the later time \(t'\); each \(E_i\) is in \({\mathcal {F}}\), so \({\mathcal {E}}\subseteq {\mathcal {F}}\); \({\mathcal {E}}\) forms a partition;

- (ii)
a set of finite sets of credence functions, \({\mathcal {C}}= \{C_1, \ldots , C_n\}\)

for each \(E_i\), \(C_i\) is the set of possible ways that the rule allows you to respond to evidence \(E_i\); that is, it is the set of possible posteriors that the rule permits when you learn \(E_i\); each \(c'\) in \(C_i\) in \({\mathcal {C}}\) is defined on \({\mathcal {F}}\).

^{4}

*Deterministic updating rule*An updating rule \(R = ({\mathcal {E}}, {\mathcal {C}})\) is

*deterministic*if each \(C_i\) is a singleton set \(\{c_i\}\). That is, for each piece of evidence there is exactly one possible response to it that the rule allows.

*Stochastic updating rule*A

*stochastic updating rule*is an updating rule \(R = ({\mathcal {E}}, {\mathcal {C}})\) equipped with a probability function

*P*.

*P*records, for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), how likely it is that I will adopt \(c'\) in response to learning \(E_i\). I write this \(P(R^i_{c'} | E_i)\), where \(R^i_{c'}\) is the proposition that says that you adopt posterior \(c'\) in response to evidence \(E_i\).

I assume \(P(R^i_{c'} | E_i) > 0\) for all \(c'\) in \(C_i\). If the probability that you will adopt \(c'\) in response to \(E_i\) is zero, then \(c'\) does not count as a response to \(E_i\) that the rule allows.

Note that every deterministic updating rule is a stochastic updating rule for which \(P(R^i_{c'} | E_i) = 1\) for each \(c'\) in \(C_i\). If \(R = ({\mathcal {E}}, {\mathcal {C}})\) is deterministic, then, for each \(E_i\), \(C_i = \{c_i\}\). So let \(P(R^i_{c_i} | E_i) = 1\).

*Conditionalizing updating rule*Suppose \(R = ({\mathcal {E}}, {\mathcal {C}})\) is an updating rule. Then

*R*is a weak conditionalizing rule for a prior

*c*if, whenever \(c(E_i) > 0\), \(C_i = \{c_i\}\) and \(c_i(-) = c(-|E_i)\). And

*R*is a strong conditionalizing rule for c if, for all \(c(E_i)\), \(c(E_i) > 0\), \(C_i = \{c_i\}\) and \(c_i(-) = c(-|E_i)\)

^{5}

*Conditionalizing pairs * A pair \(\langle c, R \rangle \) of a prior and an updating rule is a *weak (strong) conditionalizing pair* if *R* is a weak (strong) conditionalizing rule for *c*.

*Super-conditionalizing updating rule*Suppose \(R = ({\mathcal {E}}, {\mathcal {C}})\) is an updating rule. Then let \({\mathcal {F}}^*\) be the smallest algebra that contains all of \({\mathcal {F}}\) and also \(R^i_{c'}\) for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\). (As above \(R^i_{c'}\) is the proposition that says that you adopt posterior \(c'\) in response to evidence \(E_i\).) Then

- (a)
*R**is a weak super-conditionalizing rule for**c*if there is an extension \(c^*\) of*c*such that, for all \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), if \(c^*(R^i_{c'}) > 0\), then \(c'(-) = c^*(-|R^i_{c'})\).That is, each posterior to which you assign positive prior credence is the result of conditionalizing the extended prior \(c^*\) on the evidence to which it is a response and the fact that it was your response to this evidence.

- (b)
*R**is a strong super-conditionalizing rule for**c*if there is an extension \(c^*\) of*c*such that, for all \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), \(c^*(R^i_{c'}) > 0\) and \(c'(-) = c^*(-|R^i_{c'})\).That is, you assign positive prior credence to each posterior and each posterior is the result of conditionalizing the extended prior \(c^*\) on the evidence to which it is a response and the fact that it was your response to this evidence.

*Super-conditionalizing pair*A pair \(\langle c, R \rangle \) of a prior and an updating rule is a

*weak (strong) super-conditionalizing pair*if

*R*is a weak (strong) super-conditionalizing rule for

*c*.

*F*—which says that it made landfall in Florida, that is, in Pensacola or Panama City—or \({\overline{F}}\)—which says it hit elsewhere, that is, in Mobile or Biloxi. Her prior is

*c*:

Panama City | Pensacola | Mobile | Biloxi | |
---|---|---|---|---|

| 60% | 20% | 15% | 5% |

Panama City | Pensacola | Mobile | Biloxi | |
---|---|---|---|---|

\(c'_F\) | 75% | 25% | 0% | 0% |

\(c'_{{\overline{F}}}\) | 0% | 0% | 75% | 25% |

\(c^\circ _F\) | 70% | 30% | 0% | 0% |

\(c^\circ _{{\overline{F}}}\) | 0% | 0% | 70% | 30% |

\(c^\dag _F\) | 80% | 20% | 0% | 0% |

\(c^\dag _{{\overline{F}}}\) | 0% | 0% | 80% | 20% |

| \({\overline{F}}\) | Det. | Cond. | W./S. super-cond. | |
---|---|---|---|---|---|

\(R_1\) | \(\{c'_F\}\) | \(\{c'_{{\overline{F}}}\}\) | \(\checkmark \) | \(\checkmark \) | \(\checkmark \) |

\(R_2\) | \(\{c^\circ _F\}\) | \(\{c^\circ _{{\overline{F}}}\}\) | \(\checkmark \) | \(\times \) | \(\times \) |

\(R_3\) | \(\{c^\circ _F, c^\dag _F\}\) | \(\{c^\circ _{{\overline{F}}}, c^\dag _{{\overline{F}}}\}\) | \(\times \) | \(\times \) | \(\checkmark \) |

\(R_4\) | \(\{c^\circ _F\}\) | \(\{c^\circ _{{\overline{F}}}, c^\dag _{{\overline{F}}}\}\) | \(\times \) | \(\times \) | \(\times \) |

We’ll see in detail below why \(R_3\) is a strong super-conditionalizing rule for *c*, but roughly speaking the reason is that it has two properties that are jointly sufficient for being such a rule, as Lemma 2 shows: first, each posterior that \(R_3\) permits assigns maximum credence to the evidence to which it is a response; second, *c* is a weighted average of those permitted posteriors in which the weights are all positive.

In what follows, I will present each argument in its standard formulation, which assumes Deterministic Updating. Then I will explore what happens when we remove that assumption.

Deterministic updating (DU)Your updating rule should be deterministic.

## 3 The Dutch strategy argument (DSA)

The DSA and EPUA both evaluate updating rules by considering their pragmatic consequences. That is, they look to the choices that your priors and/or your possible posteriors lead you to make and they conclude that they are optimal only if your updating rule is a conditionalizing rule for your prior.

### 3.1 DSA with deterministic updating

Let’s look at the DSA first. In what follows, I’ll take a decision problem to be a set of options that are available to an agent: e.g. accept a particular bet or refuse it; buy a particular lottery ticket or don’t; take an umbrella when you go outside, take a raincoat, or take neither; and so on. The idea behind the DSA is this. One of the roles of credences is to help us make choices when faced with decision problems. They play that role badly if they lead us to make one series of choices when another series is guaranteed to serve our ends better. The DSA turns on the claim that, unless we update in line with Conditionalization, our credences will lead us to make such a series of choices when faced with a particular series of decision problems.

Here, I restrict attention to a particular class of decision problems you might face. They are the decision problems in which, for each available option, its outcome at a given possible world obtains for you a certain amount of a particular quantity, such as money or chocolate or pure pleasure, and your utility is linear in that quantity—that is, obtaining some amount of that quantity increases your utility by the same amount regardless of how much of the quantity you already have. The quantity is typically taken to be money, and I’ll continue to talk like that in what follows. But it’s really a placeholder for some quantity with this property. I restrict attention to such decision problems because, in the argument, I need to combine the outcome of one decision, made at the earlier time, with the outcome of another decision, made at the later time. So I need to ensure that the utility of a combination of outcomes is the sum of the utilities of the individual outcomes.

Now, suppose *c* is our prior and \(R = ({\mathcal {E}}= \{E_1, \ldots , E_n\}, {\mathcal {C}}= \{C_1, \ldots , C_n\})\) is our updating rule. As I do throughout, I assume that *c* is a probability function, and so is each \(c'\) in \(C_i\) in \({\mathcal {C}}\). And I will assume further that, when your credences are probabilistic, and you face a decision problem, then you should choose from the available options one that maximises expected utility relative to your credences.

Then the Dutch strategy argument is based on the following mathematical fact (de Finetti 1974):Strong Dutch strategies\(\langle c, R \rangle \) is vulnerable to astrong Dutch strategyif there are two decision problems, \({\mathbf {d}}\) and \({\mathbf {d}}'\), such that

- (i)

crequires you to choose optionAfrom the possible options available in \({\mathbf {d}}\);- (ii)
for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), \(c'\) requires you to choose

Xfrom \({\mathbf {d}}'\);- (iii)
there are alternative options,

Bin \({\mathbf {d}}\) andYin \({\mathbf {d}}'\), such that, at every possible world, you’ll receive more utility from choosingBandYthan you receive from choosingAandX. In the language of decision theory, \(B + Y\) strongly dominates \(A + X\).Weak Dutch strategies\(\langle c, R \rangle \) is vulnerable to aweak Dutch strategyif there is a decision problem \({\mathbf {d}}\) and, for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), a further decision problem \({\mathbf {d}}^i_{c'}\) such that

- (i)

crequires you to chooseAfrom \({\mathbf {d}}\);- (ii)
for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), \(c'\) requires you to choose \(X^i_{c'}\) from \({\mathbf {d}}^i_{c'}\);

- (iii)
there is an alternative option,

Bin \({\mathbf {d}}\), and, for each \(E_i\) in \({\mathcal {E}}\) and \(c'\) in \(C_i\), there is an alternative option, \(Y^i_{c'}\) in \({\mathbf {d}}^i_{c'}\), such that (a) for each \(E_i\), each world in \(E_i\), and each \(c'\) in \(C_i\), you’ll receive at least as much utility at that world from choosingBand \(Y^i_{c'}\) as you’ll receive from choosingAand \(X^i_{c'}\), and (b) for some \(E_i\), some world in \(E_i\), and some \(c'\) in \(C_i\), you’ll receive strictly more utility at that world fromBand \(Y^i_{c'}\) than you’ll receive fromAand \(X^i_{c'}\).

### Theorem 1

*Suppose*

*R*

*is a deterministic updating rule. Then:*

- (i)
*If**R**is not a weak or strong conditionalizing pair for**c**, then*\(\langle c, R \rangle \)*is vulnerable to a strong Dutch strategy;* - (ii)
*If**R**is a conditionalizing rule for**c**, then*\(\langle c, R \rangle \)*is not vulnerable even to a weak Dutch strategy.* - (iii)
*If**R**is a weak conditionalizing rule for**c**, then*\(\langle c, R \rangle \)*is not vulnerable to a strong Dutch strategy.*

That is, if your updating rule is not a conditionalizing rule for your prior, then your credences will lead you to choose a strongly dominated pair of options when faced with a particular pair of decision problems; if it is, that can’t happen.

Now that we have seen how the argument works, let’s see whether it supports the three versions of Conditionalization that we met above: Actual (AC), Plan (PC), and Dispositional (DC) Conditionalization. Since they speak directly of rules, let’s begin with PC and DC.

The DSA shows that, if you endorse a deterministic rule that isn’t a conditionalizing rule for your prior, then there is pair of decision problems, one that you’ll face at the earlier time and the other at the later time, where your credences at the earlier time and your planned credences at the later time will require you to choose a dominated pair of options. And it seems reasonable to say that it is irrational to endorse a plan when you will be rendered vulnerable to a Dutch strategy if you follow through on it. So, for those who endorse deterministic rules, DSA plausibly supports Plan Conditionalization.

The same is true of Dispositional Conditionalization. Just as it is irrational to *plan* to update in a way that would render you vulnerable to a Dutch strategy if you were to stick to the plan, it is surely irrational to be *disposed* to update in a way that will render you vulnerable in this way. So, for those whose updating dispositions are deterministic, DSA plausibly supports Dispositional Conditionalization.

- 1.
I might assume: (i) between an earlier and a later time, there is always a partition such that you know that the strongest evidence you will receive between those times is a proposition from that partition learned with certainty; (ii) if you know you’ll receive evidence from some partition, you are rationally required to plan how you will update on each possible piece of evidence before you receive it; and (iii) if you plan how to respond to evidence before you receive it, you are rationally required to follow through on that plan once you have received it. Together with PC, these give AC.

This is the most common route to AC, and has therefore received the most attention. Schoenfield (2017), Weisberg (2007), van Fraassen (1999), and Bronfman (2014) deny (i); van Fraassen (1989) denies (ii); and Pettigrew (2016) denies (iii).

- 2.
I might assume: (i) you have updating dispositions. So, if you actually update other than by Conditionalization, then it must be a manifestation of a disposition other than conditionalizing. Together with DC, this gives AC.

- 3.
I might assume: (i) that you are rationally required to update in any way that can be represented as the result of updating on a plan that you were rationally permitted to endorse or as the result of dispositions that you were rationally permitted to have, even if you did not in fact endorse any plan prior to receiving the evidence nor have any updating dispositions. Again, together with PC + DU or DC, this gives AC.

### 3.2 DSA without deterministic updating

*F*, rule \(R_3\) allows her to update from her prior

*c*to posterior \(c^\circ _F\) or posterior \(c^\dag _F\). And if she receives \({\overline{F}}\), it allows her to update to \(c^\circ _{{\overline{F}}}\) or to \(c^\dag _{{\overline{F}}}\). Notice that \(\langle c, R_3 \rangle \) violates Conditionalization thoroughly: it is not deterministic; and, moreover, as well as not mandating the posteriors that Conditionalization demands, it does not even permit them. The posterior \(c(-|F)\) does not appear in \(C_F\) and \(c(-|{\overline{F}})\) does not appear in \(C_{{\overline{F}}}\). Can we adapt the DSA to show that \(R_3\) is irrational for someone with prior

*c*? As we’ll see, the answer is no. The reason is that \(\langle c, R_3 \rangle \) is not vulnerable to a Dutch strategy.

To see this, I first note that, while \(R_3\) is not deterministic and not a conditionalizing rule for *c*, it is a super-conditionalizing rule for *c*. And to see that, it helps to state the following representation theorem for super-conditionalizing rules, which we mentioned informally above:

### Lemma 2

- (i)
*R**is a weak super-conditionalizing rule for**c**iff there is, for each*\(E_i\)*in*\({\mathcal {E}}\)*and*\(c'\)*in*\(C_i, 0 \le \lambda ^i_{c'} \le 1\)*with*\(\sum _{E_i \in {\mathcal {E}}} \sum _{c' \in C_i} \lambda ^i_{c'} =1\)*such that*- (a)
*for all*\(E_i\)*in*\({\mathcal {E}}\)*and*\(c'\)*in*\(C_i\)*, if*\(\lambda ^i_{c'} > 0\)*, then*\(c'(E_i) = 1\)*, and* - (b)
\(c(-) = \sum _{E_i \in {\mathcal {E}}} \sum _{c' \in C_i} \lambda ^i_{c'} c'(-)\).

- (a)
- (ii)
*R**is a strong super-conditionalizing rule for**c**iff there is, for each*\(E_i\)*in*\({\mathcal {E}}\)*and*\(c'\)*in*\(C_i, 0< \lambda ^i_{c'} < 1\)*with*\(\sum _{E_i \in {\mathcal {E}}} \sum _{c' \in C_i} \lambda ^i_{c'} =1\)*such that*- (a)
*for all*\(E_i\)*in*\({\mathcal {E}}\)*and*\(c'\)*in*\(C_i, c'(E_i) = 1\)*; and* - (b)
\(c(-) = \sum _{E_i \in {\mathcal {E}}} \sum _{c' \in C_i} \lambda ^i_{c'} c'(-)\).

- (a)

- (a)
\(c^\circ _F(F) = 1 = c^\dag _F(F)\) and \(c^\circ _{{\overline{F}}}({\overline{F}}) = 1 = c^\dag _{{\overline{F}}}({\overline{F}})\)

- (b)
\(c(-) = 0.4c^\circ _F(-) + 0.4c^\dag _F(-) + 0.1c^\circ _{{\overline{F}}}(-) + 0.1 c^\dag _{{\overline{F}}}(-)\)

*c*. What’s more:

### Theorem 3

- (i)
*If**R**is not a weak or strong super-conditionalizing rule for**c**, then*\(\langle c, R \rangle \)*is vulnerable at least to a weak Dutch strategy, and possibly also a strong Dutch strategy.* - (ii)
*If**R**is a strong super-conditionalizing rule for**c**, then*\(\langle c, R \rangle \)*is not vulnerable to a weak Dutch strategy.* - (iii)
*If**R**is a weak super-conditionalizing rule for**c**, then*\(\langle c, R \rangle \)*is not vulnerable to a strong Dutch strategy.*

Thus, by Theorem 3(ii), \(\langle c, R_3 \rangle \) is not vulnerable even to a weak Dutch strategy. The DSA, then, cannot say what is irrational about Condi if she begins with prior *c* and either endorses \(R_3\) as an updating plan or is disposed to update in line with it. Thus, the DSA cannot justify Deterministic Updating. And without DU, it cannot support PC or DC either. After all, \(R_3\) violates each of those, but it is not vulnerable even to a weak Dutch strategy. And moreover, each of the three arguments for AC break down because they depend on PC or DC. The problem is that, if Condi updates from *c* to \(c^\circ _F\) upon learning *F*, she violates AC; but there is an updating rule—namely, \(R_3\)—that allows \(c^\circ _F\) as a response to learning *F*, and, for all DSA tells us, she might have rationally endorsed \(R_3\) before learning *F* or she might rationally have been disposed to follow it. Indeed, the only restriction that DSA can place on your actual updating behaviour is that you should become certain of the evidence that you learned. After all:

### Theorem 4

*Suppose*

*c*

*is your prior and*\(c'\)

*is your posterior. Then, providing*\(c'(E_i) = 1\)

*, there is a rule*

*R*

*such that:*

- (i)
\(c'\)

*is in*\(C_i\)*, and* - (ii)
*R**is a strong super-conditionalizing rule for**c*.

Thus, at the end of this section, we can conclude that, whatever is irrational about planning to update using non-deterministic updating rules that are nonetheless strong super-conditionalizing rules for your prior, it cannot be that following through on those plans leaves you vulnerable to a Dutch strategy, for it does not. And similarly, whatever is irrational about being disposed to update in those ways, it cannot be that those dispositions will equip you with credences that lead you to choose dominated options, for they do not. With PC and DC thus blocked, our route to AC is therefore also blocked.

## 4 The expected pragmatic utility argument (EPUA)

Let’s look at EPUA next. Again, I will consider how our credences guide our actions when we face decision problems. In this case, there is no need to restrict attention to monetary decision problems. I will only consider a single decision problem, which we face at the later time, after we’ve received the evidence, so I won’t have to combine the outcomes of multiple options as I did in the DSA. The idea is this. Suppose you will make a decision after you receive whatever evidence it is that you receive at the later time. And suppose that you will use your later updated credence function to make that choice—indeed, you’ll choose from the available options by maximising expected utility from the point of view of your new updated credences. Which updating rules does your prior expect will lead you to make the choice best?

### 4.1 EPUA with deterministic updating

*R*. Then, if

*w*is a possible world and \(E_i\) is the element of the evidential partition \({\mathcal {E}}\) that is true at

*w*, the idea is that we take the pragmatic utility of

*R*relative to \({\mathbf {d}}\) at

*w*to be the utility at

*w*of whatever option from \({\mathbf {d}}\) we should choose if our posterior credence function were \(c_i\), as

*R*requires it to be at

*w*. But of course, for many decision problems, this isn’t well defined because there isn’t a

*unique*option in \({\mathbf {d}}\) that maximises expected utility by the lights of \(c_i\); rather there are sometimes many such options, and they might have different utilities at

*w*. Thus, we need not only \(c_i\) but also a selection function, which picks a single option from any set of options. If

*f*is such a selection function, then let \(A^{\mathbf {d}}_{c_i, f}\) be the option that

*f*selects from the set of options in \({\mathbf {d}}\) that maximise expected utility by the lights of \(c_i\). And let

### Theorem 5

*Suppose*

*R*

*and*\(R^\star \)

*are both deterministic updating rules. Then:*

- (i)
*If**R**and*\(R^\star \)*are both conditionalizing rules for**c**, and**f*,*g**are selection functions, then for all decision problems*\({\mathbf {d}}\),$$\begin{aligned} \sum _{w \in W} c(w) u_{{\mathbf {d}}, f}(R, w) = \sum _{w \in W} c(w) u_{{\mathbf {d}}, g}(R^\star , w) \end{aligned}$$ - (ii)
*If**R**is a conditionalizing rule for**c**, and*\(R^\star \)*is not, and**f*,*g**are selection functions, then for all decision problems*\({\mathbf {d}}\),$$\begin{aligned} \sum _{w \in W} c(w) u_{{\mathbf {d}}, f}(R, w) \ge \sum _{w \in W} c(w) u_{{\mathbf {d}}, g}(R^\star , w) \end{aligned}$$*with strict inequality for some decision problems*\({\mathbf {d}}\).

That is, a deterministic updating rule maximises expected pragmatic utility by the lights of your prior just in case it is a conditionalizing rule for your prior.

As in the case of the DSA above, then, if we assume Deterministic Updating (DU), we can establish PC and DC. On the back of those, we can establish AC as well, using one of the arguments from the end of Sect. 3.1. After all, it is surely irrational to plan to update in one way when you expect another way to guide your actions better in the future; and it is surely irrational to be disposed to update in one way when you expect another to guide you better. And as before there are the same three arguments for AC on the back of PC and DC.

### 4.2 EPUA without deterministic updating

*f*, the pragmatic utility of rule

*R*at world

*w*is the utility of the option that you would choose when faced with \({\mathbf {d}}\) using the credence function that

*R*mandates at

*w*and

*f*: that is, if \(E_i\) is true at

*w*, then

*R*is not deterministic, there might be no single credence function that it mandates at

*w*. If \(E_i\) is the piece of evidence you’ll learn at

*w*and

*R*permits more than one credence function in response to \(E_i\), then there might be a range of different options in \({\mathbf {d}}\), each of which maximises expected utility relative to a different credence function in \(C_i\). So what are we to do?

There are (at least) two possibilities: *the fine-graining response* and *the coarse-graining response*. On the former, we cannot establish PC or DC without assuming DU; on the latter, we can.

*R*mandates at world

*w*, a natural response is to say that we should specify the worlds in more detail, so that they determine not only the truth or falsity of the propositions in \({\mathcal {F}}\), but also which credence function you in fact adopt from those that

*R*permits. In fact, given that we will be comparing the expected pragmatic utility of two different updating rules,

*R*and \(R^\star \), we need worlds that specify not only a credence function that someone following

*R*adopts but also a credence function that someone following \(R^\star \) adopts. If

*w*is in \(E_i\), \(c'\) is in \(C_i\), and \(c^{\star \prime }\) is in \(C^\star _i\), then let \( w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }}\) be the world at which the propositions in \({\mathcal {F}}\) that are true at

*w*are true, the propositions in \({\mathcal {F}}\) that are false at

*w*are false, the person with rule

*R*adopts \(c'\) in response to receiving evidence \(E_i\) and the person with \(R^\star \) adopts \(c^{\star \prime }\) in response to that evidence. And define the pragmatic utilities of

*R*and \(R^\star \) at this world relative to a decision problem \({\mathbf {d}}\) and a selection function

*f*in the natural way:

\( u_{{\mathbf {d}}, f}(R, w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }} ) = u(A^{{\mathbf {d}}}_{c', f}, w)\)

\( u_{{\mathbf {d}}, f}(R^\star , w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }} ) = u(A^{{\mathbf {d}}}_{c^{\star \prime }, f}, w)\)

*w*in

*W*, there is no guarantee that it assigns one to each \( w\ \& \ R^i_{c'}\ \& \ R^{\star \prime }_{c^{\star \prime }}\). So what’s to be done?

A natural proposal is this: an updating rule *R* is rationally permissible from the point of view of a prior *c* just in case there is some way to extend *c* to \(c^*\) such that *R* maximises expected pragmatic utility by the lights of the extended prior, \(c^*\). However, it is straightforward to see that any super-conditionalizing rule for a prior is rationally permissible by this standard. After all, if *R* is a weak or strong super-conditionalizing rule for *c*, then there is an extension of \(c^*\) such that *R* is a conditionalizing rule for \(c^*\), and then we can piggyback on Theorem 5.

### Theorem 6

*Suppose*

*R*

*is an updating rule. Then, if*

*R*

*is a weak or strong super-conditionalizing rule for*

*c*

*, then there is an extension*\(c^*\)

*of*

*c*

*such that, for all updating rules*\(R^\star \)

*, for all selection functions*

*f*,

*g*

*, and all decision problems*\({\mathbf {d}}\)

So, if we opt for the fine-graining response to the problem of defining the pragmatic utility of a non-deterministic rule at a world, then we cannot establish either PC or DC without assuming DU and restricting the set of permissible updating rules to include only the deterministic ones.

But we might instead adopt the coarse-graining response. On this response, we retain the original possible worlds *w* in *W*, and we define the pragmatic utility of a rule at a world as either the expectation or the average of its pragmatic utility, depending on whether we are thinking of the rule as representing our dispositions or our plans, and thus whether we aim to establish DC or PC.

*w*, we might take the pragmatic utility of

*R*relative to \({\mathbf {d}}\) and

*f*at

*w*to be the expectation of the utility of the options that each permitted response to \(E_i\) (and selection function

*f*) would lead us to choose:

^{6}

### Theorem 7

*Suppose*

*R*

*and*\(R^\star \)

*are both updating rules. Then:*

- (i)
*If**R**and*\(R^\star \)*are both conditionalizing rules for**c**, and**f*,*g**are selection functions, then for all decision problems*\({\mathbf {d}}\),$$\begin{aligned} \sum _{w \in W} c(w) u_{{\mathbf {d}}, f}(R, w) = \sum _{w \in W} c(w) u_{{\mathbf {d}}, g}(R^\star , w) \end{aligned}$$ - (ii)
*R**is a conditionalizing rule for**c**, and*\(R^\star \)*is a stochastic but not conditionalizing rule, and**f*,*g**are selection functions, then for all decision problems*\({\mathbf {d}}\),$$\begin{aligned} \sum _{w \in W} c(w) u_{{\mathbf {d}}, f}(R, w) \ge \sum _{w \in W} c(w) u_{{\mathbf {d}}, g}(R^\star , w) \end{aligned}$$*with strictly inequality for some decision problems*\({\mathbf {d}}\).

This shows the first difference between the DSA and EPUA. The latter, but not the former, provides a route to establishing Dispositional Conditionalization (DC). If we adopt the coarse-graining response to the problem of defining the pragmatic utility of an updating rule, we can establish DC. If we assume that your dispositions are governed by a chance function, and we use that chance function to calculate expectations, then we can show that your prior will expect your posteriors to do worse as a guide to action unless you are disposed to update by conditionalizing on the evidence you receive.

Next, suppose we are interested in Plan Conditionalization (PC). In this case, we might try to appeal again to Theorem 7. To do that, we must assume that, while there are non-deterministic updating rules that we might endorse, they are all at least stochastic updating rules; that is, they all come equipped with a probability function that determines how likely it is that I will adopt a particular permitted response to the evidence I receive. That is, we might say that the updating rules that we might endorse are either deterministic or non-deterministic-but-stochastic. In the language of game theory, we might say that the updating strategies between which we choose are either pure or mixed. And then Theorem 7 will show that we should adopt a deterministic-and-conditionalizing rule, rather than any deterministic-but-non-conditionalizing or non-deterministic-but-stochastic rule. The problem with this proposal is that it seems just as arbitrary to restrict to deterministic and non-deterministic-but-stochastic rules as it was to restrict to deterministic rules in the first place. Why should we not be able to endorse a non-deterministic and non-stochastic rule—that is, a rule that says, for at least one possible piece of evidence \(E_i\) in \({\mathcal {E}}\), there are two or more posteriors that the rule permits as responses, but does not endorse any chance mechanism by which we’ll choose between them? But if we permit these rules, how are we to define their pragmatic utility relative to a decision problem and at a possible world?

*w*. And suppose \({\mathbf {d}}\) is a decision problem and

*f*is a selection rule. Then we might take the pragmatic utility of

*R*relative to \({\mathbf {d}}\) and

*f*and at

*w*to be the average (specifically, the mean) utility of the options that each permissible response to \(E_i\) and

*f*would choose when faced with \({\mathbf {d}}\). That is,

*R*permits.

^{7}If that’s the case, then we have the following:

### Theorem 8

*Suppose*

*R*

*and*\(R^\star \)

*are updating rules. Then if*

*R*

*is a conditionalizing rule for*

*c*

*, and*\(R^\star \)

*is not deterministic, not stochastic, and not a conditionalizing rule for*

*c*

*, and*

*f*,

*g*

*are selection functions, then for all decision problems*\({\mathbf {d}}\),

*with strictly inequality for some decision problems*\({\mathbf {d}}\).

Put together with Theorems 5 and 7, this shows that our prior expects us to do better by endorsing a conditionalizing rule than by endorsing any other sort of rule, whether that is a deterministic and non-conditionalizing rule, a non-deterministic but stochastic rule, or a non-deterministic and non-stochastic rule.

So, again, we see a difference between DSA and EPUA. Just as the latter, but not the former, provides a route to establishing DC without assuming Deterministic Updating, so the latter but not the former provides a route to establishing PC without DU. And from both of those, we have the usual three routes to AC enumerated at the end of Sect. 3.1. This means that, if we respond to the problem of defining pragmatic utility by taking the coarse-graining appoach, the EPUA could explain what’s irrational about endorsing a non-deterministic updating rule, or having dispositions that match one. If you do, there’s some alternative updating rule that your prior expects to do better as a guide to future action.

## 5 Expected epistemic utility argument (EEUA)

The previous two arguments criticized non-conditionalizing updating rules from the standpoint of pragmatic utility. The EEUA and EUDA both criticize such rules from the standpoint of epistemic utility. The idea is this: just as credences play a pragmatic role in guiding our actions, so they play other roles as well—they represent the world; they respond to evidence; they might combine more or less coherently. These roles are purely epistemic, and so just as I defined the *pragmatic* utility of a credence function at a world when faced with a decision problem, so we can also define the *epistemic* utility of a credence function at a world—it is a measure of how valuable it is to have that credence function from a purely epistemic point of view.

### 5.1 EEUA with deterministic updating

I won’t give an explicit definition of the epistemic utility of a credence function at a world. Rather, I’ll simply state two properties that I’ll take measures of such epistemic utility to have. These are widely assumed in the literature on epistemic utility theory and accuracy-first epistemology, and I’ll defer to the arguments in favour of them that are outlined there (Joyce 2009; Pettigrew 2016; Horowitz 2019).

*s*that takes a single credence and a truth value—either true (1) or false (0)—and returns the epistemic value of having that credence in a proposition with that truth value. Thus,

*s*(1,

*p*) is the epistemic value of having credence

*p*in a truth, while

*s*(0,

*p*) is the epistemic value of having credence

*p*in a falsehood. A global epistemic utility function is a function

*EU*that takes an entire credence function defined on \({\mathcal {F}}\) and a possible world and returns the epistemic value of having that credence function when the propositions in \({\mathcal {F}}\) have the truth values they have in that world.

We can then define the epistemic utility of a deterministic updating ruleStrict proprietyA local epistemic utility functionsisstrictly properif

- (i)

s(1,x) ands(0,x) are continuous functions ofx;- (ii)
each credence expects itself and only itself to have the greatest epistemic utility. That is, for all \(0 \le p \le 1\),is uniquely maximised, as a function of$$\begin{aligned} ps(1, x) + (1-p) s(0, x) \end{aligned}$$x, at \(x = p\).^{8}AdditivityA global epistemic utility function isadditiveif, for each propositionXin \({\mathcal {F}}\), there is a local epistemic utility function \(s_X\) such that the epistemic utility of a credence functioncat a possible world is the sum of the epistemic utilities at that world of the credences it assigns. Ifwis a possible world and we writew(X) for the truth value (0 or 1) of propositionXatw, this says:$$\begin{aligned} EU(c, w) = \sum _{X \in {\mathcal {F}}} s_X(w(X), c(X)) \end{aligned}$$

*R*in the same way we defined its pragmatic utility above: if \(E_i\) is true at

*w*, and \(C_i = \{c_i\}\), then

### Theorem 9

*Suppose*

*R*

*and*\(R^\star \)

*are deterministic updating rules. Then:*

- (i)
*If**R**and*\(R^\star \)*are both conditionalizing rules for**c**, then*$$\begin{aligned} \sum _{w \in W} c(w) EU(R, w) = \sum _{w \in W} c(w) EU(R^\star , w) \end{aligned}$$ - (ii)
*If**R**is a conditionalizing rule for**c**and*\(R^\star \)*is not, then*$$\begin{aligned} \sum _{w \in W} c(w) EU(R, w) > \sum _{w \in W} c(w) EU(R^\star , w) \end{aligned}$$

That is, a deterministic updating rule maximises expected epistemic utility by the lights of your prior just in case it is a conditionalizing rule for your prior.

So, as for DSA and EPUA, if we assume Deterministic Updating, we obtain an argument for PC and DC, and indirectly arguments for AC too.

### 5.2 EEUA without deterministic updating

If we don’t assume Deterministic Updating, the situation here is very similar to the one we encountered above when we considered EPUA. Suppose *R* is a non-deterministic updating rule. Then again we have two choices: the fine- and the coarse-graining response.

*R*at a fine-grained world is

### Theorem 10

*Suppose*

*R*

*and*\(R^\star \)

*are both updating rules. Then, if*

*R*

*is a weak or strong super-conditionalizing rule for*

*c*

*, then there is an extension*\(c^*\)

*of*

*c*

*such that*

So it seems that super-conditionalizing updating rules are rationally permissible, at least by the lights of expected epistemic utility.

*R*is non-deterministic but stochastic. Then we let its epistemic utility at a coarse-grained world be the expectation of the epistemic utility that the various possible posteriors permitted by

*R*take at that world. That is, if \(E_i\) is the proposition in \({\mathcal {E}}\) that is true at

*w*, then

### Theorem 11

*Suppose*

*R*

*and*\(R^\star \)

*are updating rules. Then if*

*R*

*is a conditionalizing rule for*

*c*

*, and*\(R^\star \)

*is stochastic but not a conditionalizing rule for*

*c*

*, then*

*R*is a non-deterministic but also a non-stochastic rule. Then we let its epistemic utility at a world be the average epistemic utility that the various possible posteriors permitted by

*R*take at that world. That is, if \(E_i\) is the proposition in \({\mathcal {F}}\) that is true at

*w*, then

### Theorem 12

*Suppose*

*R*

*and*\(R^\star \)

*are updating rules. Then if*

*R*

*is a conditionalizing rule for*

*c*

*, and*\(R^\star \)

*is not deterministic, not stochastic, and not a conditionalizing rule for*

*c*

*. Then:*

So the situation is the same as for EPUA. If we take the coarse-graining approach, whether we assess a rule by looking at how well the posteriors it produces guide our future actions or how good they are from a purely epistemic point of view, our prior will expect a conditionalizing rule for itself to be better than any non-conditionalizing rule. And thus we obtain PC and DC, and indirectly AC as well.

## 6 Epistemic utility dominance argument (EUDA)

Finally, I turn to the EUDA. In EPUA and EEUA, we assess the pragmatic or epistemic utility of the updating rule from the viewpoint of the prior. In DSA, we assess the prior and updating rule together, and from no particular point of view; and, unlike the EPUA and EEUA, we do not assign utilities, either pragmatic or epistemic, to the prior and the rule. In EUDA, like in DSA and unlike EPUA and EEUA, we assess the prior and updating rule together, and again from no particular point of view; but, unlike in DSA and like in EPUA and EEUA, we assign utilities to them—in particular, epistemic utilities—and assess them with reference to those.

### 6.1 EUDA with deterministic updating

*R*is a deterministic updating rule. Then, as before, if \(E_i\) is true at

*w*, let the epistemic utility of

*R*be the epistemic utility of the credence function \(c_i\) that it mandates at

*w*: that is,

### Theorem 13

*Suppose EU is an additive, strictly proper epistemic utility function. And suppose*

*R*

*and*\(R^\star \)

*are deterministic updating rules. Then:*

- (i)
*If*\(\langle c, R \rangle \)*is not conditionalizing, there is*\(\langle c^\star , R^\star \rangle \)*such that, for all**w*,$$\begin{aligned} EU(\langle c, R \rangle , w) < EU(\langle c^\star , R^\star \rangle , w)) \end{aligned}$$ - (ii)
*If*\(\langle c, R \rangle \)*is conditionalizing, there is no*\(\langle c^\star , R^\star \rangle \)*such that, for all**w*,$$\begin{aligned} EU(\langle c, R \rangle , w) < EU(\langle c^\star , R^\star \rangle , w)) \end{aligned}$$

That is, if *R* is not a conditionalizing rule for *c*, then together they are *EU*-dominated; if it is a conditionalizing rule, they are not. Thus, like EPUA, EEUA, and DSA, if we assume Deterministic Updating, EUDA gives PC, DC, and indirectly AC.

### 6.2 EUDA without deterministic updating

Now suppose we permit non-deterministic updating rules as well as deterministic ones. As before, there are two approaches: the fine- and coarse-graining approaches. Here is the relevant result for the fine-graining approach:

### Theorem 14

*Suppose*

*EU*

*is an additive, strictly proper epistemic utility function. Then:*

- (i)
*If**R**is a weak or strong super-conditionalizing rule for**c**, there is no*\(\langle c^\star , R^\star \rangle \)*such that, for all*\(E_i\)*in*\({\mathcal {E}}, w\)*in*\(E_i, c'\)*in*\(C_i\)*and*\(c^{\star \prime }\)*in*\(C^\star _i\)$$ \begin{aligned} EU(\langle c, R \rangle , w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }}) < EU(\langle c^\star , R^\star \rangle , w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }}) \end{aligned}$$ - (ii)
*There are rules**R**that are not weak or strong super-conditionalizing rules for**c**such that there is no*\(\langle c^\star , R^\star \rangle \)*such that, for all*\(E_i\)*in*\({\mathcal {E}}, w\)*in*\(E_i, c'\)*in*\(C_i\)*and*\(c^{\star \prime }\)*in*\(C^\star _i\)$$ \begin{aligned} EU(\langle c, R \rangle , w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }}) < EU(\langle c^\star , R^\star \rangle , w\ \& \ R^i_{c'}\ \& \ R^{\star i}_{c^{\star \prime }}) \end{aligned}$$

Interpreted in this way, then, and without the assumption of Deterministic Updating, EUDA is the weakest of all the arguments. Whereas DSA at least establishes that your updating rule should be a weak or strong super-conditionalizing for your prior, even if it does not establish that it should be conditionalizing, EUDA does not establish even that.

And here is the relevant result for the coarse-graining approach:

### Theorem 15

*Suppose*

*EU*

*is an additive, strictly proper epistemic utility function. Then, if*\(\langle c, R \rangle \)

*is not a conditionalizing pair, there is an alternative pair*\(\langle c^\star , R^\star \rangle \)

*such that, for all*

*w*,

This therefore supports an argument for PC and DC and indirectly AC as well.

## 7 Conclusion

One upshot of this investigation is that, so long as we assume Deterministic Updating (DU), all four arguments support the same conclusions, namely, Plan (PC) and Dispositional (DC) Conditionalization, and also Actual Conditionalization (AC). But once we drop DU, that agreement vanishes.

Without DU, DSA shows only that, if we plan to update using a particular rule, it should be a super-conditionalizating rule for our prior; and similarly for our dispositions. As a result, it cannot support AC. Indeed, it can support only the weakest restrictions on our actual updating behaviour, since nearly any such behaviour can be seen as an implementation of a super-conditionalizing rule—as long as we become certain of the evidence we receive after we receive it, we can be represented as having followed a strong super-conditionalizing rule.

EPUA, EEUA, and EUDA are more hopeful, at least if we adopt the coarse-graining response to the question of how to define the pragmatic or epistemic utility of a non-deterministic updating rule at a world. Let’s consider our updating dispositions first. It seems natural to assume that, even if these are not deterministic, they are at least governed by objective chances. If so, and if we define the pragmatic or epistemic utility of the rule that represents those dispositions to be the expectation of the pragmatic or epistemic utility of the credence functions it produces, then we obtain DC without assuming DU. That is, we can justify DU by appealing to pragmatic or epistemic utility. However, if we use the fine-graining response, this isn’t possible. And similarly when we consider our updating plans. Here, if we use the coarse-graining response and define the pragmatic or epistemic utility of the rule that represents our plan to be the average pragmatic or epistemic utility of the credence functions it produces, then we obtain PC without assuming DU. And again, if we use the fine-graining response, this isn’t possible. Indeed, if we use the fine-graining approach, EPUA and EEUA establish only that you should plan to update using a weak or strong super-conditionalizing rule. And EUDA doesn’t even establish that.

So, at least if we look just to our existing arguments for Conditionalization, the fates of DC and PC seem to turn on making one of two responses. First, we might simply assume Deterministic Updating. That is, to establish PC, we might simply assume that we are rationally required to plan to update in a deterministic way; and, to establish DC, we might assume that we are rationally required to have deterministic updating dispositions. This doesn’t seem promising. Typically, those philosophers who offer one of the four arguments for Conditionalization studied here do so precisely because we aren’t content to make brute normative assumptions about how it is rational to update; we wish to know what is so good about updating in the way that Conditionalization describes and what is so bad about updating in some other way. Part of that is a desire to know what is so good about updating deterministically and what is so bad about updating non-deterministically. As I mentioned at the beginning, the four arguments considered here are teleological, so they specifically tell you the goods that Conditionalization and deterministic updating obtain for you. To simply assume DU is to leave the latter a mystery.

Second, we might take the coarse-graining response to the problem of defining pragmatic and epistemic utility, and then appeal to EPUA, EEUA, or EUDA. This seems more promising. That response certainly seems reasonable—that is, it produces a reasonable way to define the pragmatic or epistemic utility of an updating rule at a coarse-grained world. The problem is that we need to do more than that if we are to establish DC and PC. It is not sufficient to show that the coarse-graining response is reasonable and therefore *permissible*. We have to show that it is *mandatory*. After all, if the fine- and the coarse-graining responses are both permissible, then there is a permissible way of defining pragmatic and epistemic utility on which non-conditionalizing updating rules are permissible. And that suggests that those rules are themselves permissible. And that conflicts with DC and PC. What is needed is an argument that the fine-graining response is not legitimate. For myself, I don’t see what that might be.

## Footnotes

- 1.
Thanks to an anonymous referee for suggesting that I mention this motivation.

- 2.
Thanks to an anonymous referee for urging me to clarify the scope of the present paper.

- 3.
A partition is a set of exhaustive and mutually exclusive propositions. That is, the disjunction of the propositions is a tautology, and the conjunction of any two propositions is a contradiction.

- 4.
For ease of exposition, I’ll assume throughout that each \(C_i\) contains only finitely many credence functions. Similar results hold if we lift this restriction, but their proofs are more involved and these stronger results aren’t needed to make our central point here.

- 5.
Note that a conditionalizing rule for a prior need not be a deterministic updating rule. It need only be deterministic for those possible pieces of evidence to which the prior assigns positive credence.

- 6.
Recall: we assumed that each \(C_i\) is finite, so this is well-defined.

- 7.
Again, recall that each \(C_i\) is finite, so this is well-defined.

- 8.That is, if \(p\ne q\), then$$\begin{aligned} ps(1, p) + (1-p)s(0, p) > ps(1, q) + (1-p) s(0, q) \end{aligned}$$
- 9.
The convex hull of a set of points is the smallest convex set that contains it as a subset. A set is convex if, for any two points in it, any convex combination of those two points also lies in the set.

## Notes

### Acknowledgements

I am extremely grateful to Catrin Campbell-Moore, Kenny Easwaran, Jason Konek, and Ben Levinstein as well as two anonymous referees for this journal for helpful comments on earlier versions of the material.

## References

- Briggs, R. A., & Pettigrew, R. (2018). An accuracy-dominance argument for conditionalization.
*Noûs*. https://doi.org/10.1111/nous.12258 CrossRefGoogle Scholar - Bronfman, A. (2014). Conditionalization and not knowing that one knows.
*Erkenntnis*,*79*(4), 871–892.CrossRefGoogle Scholar - Brown, P. M. (1976). Conditionalization and expected utility.
*Philosophy of Science*,*43*(3), 415–419.CrossRefGoogle Scholar - de Finetti, B. (1974).
*Theory of probability*(Vol. I). New York: Wiley.Google Scholar - Diaconis, P., & Zabell, S. L. (1982). Updating subjective probability.
*Journal of the American Statistical Association*,*77*(380), 822–830.CrossRefGoogle Scholar - Dietrich, F., List, C., & Bradley, R. (2016). Belief revision generalized: A joint characterization of Bayes’s and Jeffrey’s rules.
*Journal of Economic Theory*,*162*, 352–371.CrossRefGoogle Scholar - Good, I. J. (1967). On the principle of total evidence.
*The British Journal for the Philosophy of Science*,*17*, 319–322.CrossRefGoogle Scholar - Greaves, H., & Wallace, D. (2006). Justifying conditionalization: Conditionalization maximizes expected epistemic utility.
*Mind*,*115*(459), 607–632.CrossRefGoogle Scholar - Grove, A. J., & Halpern, J. Y. (1998). Updating sets of probabilities. In
*Proceedings of the 14th conference on uncertainty in AI*(pp. 173–182). San Francisco, CA: Morgan Kaufman.Google Scholar - Horowitz, S. (2019). Accuracy and educated guesses. In T. S. Gendler & J. Hawthorne (Eds.),
*Oxford studies in epistemology*(Vol. 6). Oxford: Oxford University Press.Google Scholar - Jeffrey, R. (1992).
*Probability and the art of judgment*. New York: Cambridge University Press.CrossRefGoogle Scholar - Joyce, J. M. (2009). Accuracy and coherence: Prospects for an alethic epistemology of partial belief. In F. Huber & C. Schmidt-Petri (Eds.),
*Degrees of belief*. Berlin: Springer.Google Scholar - Konek, J. (2019). The art of learning. In T. S. Gendler & J. Hawthorne (Eds.),
*Oxford studies in epistemology*(Vol. 7). Oxford: Oxford University Press.Google Scholar - Lewis, D. (1999). Why conditionalize?
*Papers in metaphysics and epistemology*(pp. 403–407). Cambridge: Cambridge University Press.CrossRefGoogle Scholar - Pettigrew, R. (2016).
*Accuracy and the laws of credence*. Oxford: Oxford University Press.CrossRefGoogle Scholar - Predd, J., Seiringer, R., Lieb, E. H., Osherson, D., Poor, V., & Kulkarni, S. (2009). Probabilistic coherence and proper scoring rules.
*IEEE Transactions of Information Theory*,*55*(10), 4786–4792.CrossRefGoogle Scholar - Savage, L. J. (1954).
*The foundations of statistics*. Hoboken: Wiley.Google Scholar - Schoenfield, M. (2017). Conditionalization does not (in general) maximize expected accuracy.
*Mind*,*126*(504), 1155–87.CrossRefGoogle Scholar - van Fraassen, B. C. (1989).
*Laws and symmetry*. Oxford: Oxford University Press.CrossRefGoogle Scholar - van Fraassen, B. C. (1999). Conditionalization a new argument for.
*Topoi*,*18*(2), 93–96.CrossRefGoogle Scholar - Weisberg, J. (2007). Conditionalization, reflection, and self-knowledge.
*Philosophical Studies*,*135*(2), 179–97.CrossRefGoogle Scholar

## Copyright information

**Open Access**This article is distributed under the terms of the Creative Commons Attribution 4.0 International License (http://creativecommons.org/licenses/by/4.0/), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made.