, Volume 30, Issue 1, pp 3–15 | Cite as

Triviality Pursuit

  • Alan Hájek


The thesis that probabilities of conditionals are conditional probabilities has putatively been refuted many times by so-called ‘triviality results’, although it has also enjoyed a number of resurrections. In this paper I assault it yet again with a new such result. I begin by motivating the thesis and discussing some of the philosophical ramifications of its fluctuating fortunes. I will canvas various reasons, old and new, why the thesis seems plausible, and why we should care about its fate. I will look at some objections to Lewis’s famous triviality results, and thus some reasons for the pursuit of further triviality results. I will generalize Lewis’s results in ways that meet the objections. I will conclude with some reflections on the demise of the thesis—or otherwise.


Probabilities of conditionals Conditional probabilities Stalnaker’s thesis Triviality results Conditionalization Imaging Blurred imaging Maximum entropy Minimum cross entropy Boldness Moderation Revision rules 

1 Setting the Scene

Like Jason of the Friday the 13th franchise, the thesis that probabilities of conditionals are conditional probabilities has enjoyed a number of resurrections. Lewis (1976) appeared to have killed it in his famous triviality results; but van Fraassen (1976) resuscitated it. Stalnaker (1976) attacked it some more; but Rehder (1982) revived it. Further triviality results appeared to deliver fatal blows to it (e.g., Lewis 1986a, b; Hájek 1989, 1994; Hall 1994; Milne 2003); yet it lives on still, albeit transmogrified, in the writings of Edgington (1995) and Bennett (2003). In this paper I will assault it yet again; but it will doubtless survive in some form in further sequels.

While I think that the various negative and positive results against the thesis are interesting in their own right—and I will offer a negative result of my own—I don’t want this paper to be merely an exercise in theorem-proving. I will begin by motivating the thesis and discussing some of the philosophical ramifications of its fluctuating fortunes. I will canvas various reasons, old and new, why the thesis seems plausible, and why we should care about its fate. There are good reasons why it keeps making comebacks. I will look at some objections to Lewis’s famous results, and thus some reasons for the pursuit of further triviality results. I will generalize Lewis’s results in ways that meet the objections. I will conclude with some reflections on the demise of the thesis—or otherwise.

2 The Thesis—or Theses

“The probability that Collingwood will win if they are ahead at half time is high.” This seems to say that the probability of a certain conditional is high. And it seems to say exactly the same thing as: “The probability that Collingwood will win given that they are ahead at half time is high.” Moreover, the probability of the conditional seems to rise or fall in lockstep with the conditional probability. Symbolically:
$$ P\left( {{\text{ahead}} \to {\text{win}}} \right) = P({\text{win}}\left| {\text{ahead}} \right.), $$
where ‘→’ represents the conditional ‘if … then’, and P(win | ahead) is given by the usual ratio formula for conditional probability, P(win & ahead)/P(ahead). But this equation seems to generalize to all sentences for which the conditional probability is defined:
$$ \left( {\text{PCCP}} \right)\quad P\left( {A \to B} \right) = P(B\left| A \right.) $$
for all A and B in the domain of P, provided P(A) > 0.

And offhand, it seems to generalize to all probability functions:

For all probability functions P, PCCP.

However, this thesis is easily refuted. Consider a 3-ticket lottery, and let L i = ticket i wins, i = 1, 2, 3. Model it with the obvious 3-world probability space, with P(L i) = 1/3 for all i. Then the conditional probability P(L 1 | L 1 v L 2) is obviously ½. But all unconditional probabilities are a multiple of 1/3, so in particular the unconditional probability P((L 1 v L 2) → L 1) is a multiple of 1/3. And ½ is not a multiple of 1/3.

Clearly, then, we need to restrict suitably the probability functions for which PCCP is supposed to hold. For each such restriction, we get a corresponding thesis that probabilities (so restricted) of conditionals are conditional probabilities (so restricted). For example, Stalnaker (1970) thought that PCCP holds for all probability functions that could represent a rational agent’s systems of beliefs—all rational credence functions. This has come to be known in the literature as Stalnaker’s Thesis. I adopt that moniker with some misgivings since, to be fair to Stalnaker, some years later he went on to renounce the Thesis, and he even proved a striking triviality result against it (in Stalnaker 1976). And he was the original architect of causal decision theory as a rival to Jeffrey’s evidential decision theory, with probabilities of conditionals replacing Jeffrey’s conditional probabilities (more on this shortly). In order to be a rival to, rather than merely a restatement of, Jeffrey’s theory, probabilities of conditionals had better come apart from conditional probabilities for rational agents at least sometimes. So this later important project of Stalnaker’s requires Stalnaker’s Thesis to be false! That said, he deserves much credit for inspiring the fertile literature that has grown out of his original Thesis, and at least the moniker serves to remind us of that.

We could also come to the restriction on the probability functions for which PCCP allegedly holds from the other side, so to speak. I have just provided a negative result—a very modest one—that some restriction is required. Further negative results will impose further restrictions—the stronger the restrictions, the more interesting the result. We can imagine various theses of various strengths, which might occupy intermediate ground: weaker than Stalnaker’s Thesis, and indeed weak enough to evade all the triviality results, but arguably strong enough still to be of interest. Hájek and Hall (1994) survey some candidates for such theses.

3 Why Care About the Theses?

3.1 The Semantics of Conditionals

The larger the class of probability functions for which PCCP holds, the better for allies of Stalnaker’s Thesis. Stalnaker argued that PCCP’s holding for all rational credence functions had ramifications for the semantics of conditionals. Reading the equation from left to right, we can regard it as constraining the assignment of a probability to a conditional. This, in turn, constrains the logic of the conditional. As Stalnaker (1970) observes,

although the interpretation of probability is controversial, the abstract calculus is a relatively well defined and well established mathematical theory. In contrast to this, there is little agreement about the logic of conditional sentences… Probability theory could be a source of insight into the formal structure of conditional sentences” (p. 1071).

In particular, PCCP implies (a probabilistic version of) conditional excluded middle. This was the locus of disagreement between Stalnaker and Lewis on the logic of counterfactuals, the former advocating it, the latter rejecting it. The intuitiveness of PCCP seemed to provide Stalnaker with a weighty argument in his favor. No surprise, then, that Lewis was motivated to expose PCCP’s limited tenability.

3.2 Probabilistic Validity

Meanwhile, Adams (1975) was developing an important approach to conditionals of his own. He believed that conditionals do not have truth conditions, and thus that the traditional ‘truth-preservation’ account of validity of arguments was inadequate for ones in which conditionals appears. But he was happy to speak of probabilities attaching to conditionals, and he offered a probabilistic surrogate for validity. Define the uncertainty of X to be 1−P(X). A probabilistically valid argument is one in which necessarily the sum of the uncertainties of the premises is at least as great as that of the conclusion. Adams invoked PCCP to govern the assignment of probabilities to indicative conditionals, arguing that the resulting scheme respected intuitions about which inferences were reasonable, and which not.

While Adams appealed to the same thesis (at least superficially) as Stalnaker did, unlike Stalnaker he did not regard the probability of a conditional as probability of its truth. Further, Adams’ ‘probabilities’ of conditionals do not conform to the usual probability calculus. In particular, according to him conditionals are not fit to enter into Boolean combinations with other sentences, and so the domain of a probability function that allows conditionals to be among its arguments is not a field. Thus, Lewis (1976) suggested that they be called “assertabilities” instead, a practice that has been widely adopted subsequently. So the left hand side of PCCP is perhaps best read as “the assertability of B, if A” on the Adams view. This conditional assertability then goes by P(B | A). This is the famous Adams’ Thesis, a touchstone of much of the literature on conditionals.

3.3 Implicit Definition of the Conditional

Still reading PCCP from left to right, we might even regard it as providing an implicit definition of the conditional: it is whatever connective obeys the equation. Now, it might seem odd that a logical operator should be defined probabilistically. It seems less odd if we regard it as encoding a rational agent’s updating dispositions, an idea that we find in Ramsey (1965) and developed further by Levi (1980).

3.4 Iterated Conditionals

PCCP would also illuminate right-nested iterated conditionals. It implies:
$$ P\left( {A \to \left( {B \to C} \right)} \right) = P\left( {B \to C|A} \right) $$
(I assume here that the iterated conditional is in the domain of P.) Now define P A (_) = P(_|A), the result of conditioning P on A. Then we may continue:
$$ \begin{aligned} P\left( {B \to C|A} \right) & = P_{A} \left( {B \to C} \right) \\ & = P_{A} \left( {C|B} \right),\quad {\text{by PCCP applied to}}\,P_{A} . \\ \end{aligned} $$
The eye-glazing probability of the iterated conditional is elegantly simplified to a perfectly manageable conditional probability.

3.5 The Interpretation of Conditional Probability

Reading PCCP from right to left, we enhance our understanding of conditional probability. De Finetti (1972) complained that the usual ratio gives the formula, but not the meaning, of conditional probability. PCCP would characterize it: it is simply the probability of a conditional! So the rather technical notion of conditional probability would be reduced to unconditional probability, which is arguably a simpler notion, and ‘if … then’, which is part of ordinary English.

3.6 Conditional Probability as a Special Case of Unconditional Probability

Pursuing this thought further, Kolmogorov already offered a reduction of conditional probability to unconditional probabilities—namely, to a ratio of them—with the formula for P(B | A) that is now standard.

3.7 The Judy Benjamin Problem

Again exploiting the right-to-left reading of PCCP, it would allow us to solve an outstanding problem for Bayesian epistemology, what van Fraassen (1989) calls “the Judy Benjamin problem”.2 The general problem for probability kinematics is: given a prior probability function P, and the imposition of some constraint on the posterior probability function, what should this posterior be? This problem has a unique solution for certain constraints—for example:
  1. 1.

    Assign probability 1 to some proposition3 E, while preserving the odds of all propositions that imply E. Solution: conditionalize P on E.

  2. 2.

    Assign probabilities p 1, …, p n to the cells of the partition {E 1, …, E n}, while preserving the odds of all propositions within each cell. Solution: Jeffrey conditionalize P on this partition, according to the specification.

    But consider the constraint:

  3. 3.

    Assign conditional probability p to B, given A.

    The Judy Benjamin problem is that of finding a rule for transforming a prior, subject to this third constraint.

    Van Fraassen provides arguments for three distinct such rules, and suggests that such uniqueness results “will not extend to more broadly applicable rules in general probability kinematics. In that case rationality will not dictate epistemic procedure even when we decide that it shall be rule governed” (1989, p. 343). But if Stalnaker’s Thesis were true, a particularly simple solution would present itself. After all, constraint 3 would then be equivalent to:

  4. 3'.

    Assign probability p to A → B,

    and this is uniquely met by a simple Jeffrey conditioning on the partition {A → B, ¬(A → B)}, assuming that the odds of propositions within each cell are to remain the same.


3.8 A New Constraint on Unconditional Probabilities

Now let’s read PCCP symmetrically, not privileging either direction. We can regard it simply as a new constraint on three unconditional probabilities: P(A → B), P(A & B), and P(A). Understood this way, it plays a role akin to that of the additivity axiom—it is part of probability theory. The Bayesian, in turn, would then presumably want to appropriate it as a coherence constraint, one obeyed by any ideally rational agent.

Such is the bounty that PCCP promises. If Stalnaker’s Thesis were tenable, logic, probability theory, and Bayesian epistemology would all be enriched. But its untenability is also philosophically fecund.

3.9 Decision Theory

Here is a quick recap of the lore regarding decision theory. In the beginning there were formulations of Bayesian decision theory by Ramsey (1931/1990) and Savage (1954); these were refined and presented in a philosopher-friendly manner by Jeffrey (1966), and his evidential decision theory became the philosophical orthodoxy.

Then along came Newcomb’s problem. Evidential decision theory apparently recommended one-boxing, while many regarded this as irrational. So causal decision theory, which recommended two-boxing, was born. Indeed, a number of versions of it were developed—first by Stalnaker (1976), whose theory was elaborated by Gibbard and Harper (1978); then by authors such as Skyrms (1980), Lewis (1981), Sobel (1994), and Joyce (1999). But—this is essential to the lore—these versions are merely stylistic variants of one another. Lewis (1981) writes: “We causal decision theorists share one common idea, and differ mainly on matters of emphasis and formulation” (p. 5). Or Harper and Skyrms (1988): “It can be argued … that the various forms of causal decision theory are equivalent—that an adequate version of any one of these approaches will be interdefinable with adequate versions of the others.” (Introduction, p. x). Joyce (1999) concurs: “I think this is basically right; … the causal decision theorist can adopt an attitude of benign indifference …” (pp. 171–172). We can thus speak of ‘causal decision theory’ in the singular.

Finally, and crucially, the lore insists that evidential decision theory and causal decision theory typically agree. In garden-variety decision problems, such as whether you should carry your umbrella to work, or whether you should bring white or red wine to the dinner party, the two theories give the same advice. It is only in rather recherché Newcomb-like cases that the theories come apart.

I have long been puzzled by the lore. For starters, I am not convinced that all the various versions of causal decision theory are equivalent. Indeed, surely some pairs of them are not equivalent. For example, while Gibbard and Harper (1978) weight their utilities with probabilities of counterfactuals, Skyrms (1980) weights his with expectations of conditional chances. Gibbard and Harper’s counterfactuals invoke other possible worlds, while Skyrms’s chances are completely anchored in the actual world. When the counterfactuals concern suppositions about the chances being different from what they actually are, presumably these two versions will come apart. Moreover, where Gibbard and Harper’s formulation involves unconditional probabilities (of counterfactuals), Skyrms’s involves conditional probabilities (conditional chances). Here we enter triviality result territory, where the different behaviors of these quantities are brought into relief—more on that shortly. And an agent’s credences may outrun her expectations of conditional chances, since she may assign credences to propositions that she thinks lack chances. For example, de Finetti (1972) had plenty of credences, but he thought that chance statements were nonsense, so his expectations of them were either 0 (nonsense being known not to be true) or undefined (nonsense being unfit for credence assignments at all). Or someone can have plenty of credences, while thinking like Cartwright (1999) that only rather special physical systems have chances, and thus expectations of conditional chances. And even those of us who think there are lots of chances out there, may still think there are some chance gaps, which may nevertheless be proper objects of credences; again we will find credences where there are no corresponding expectations of conditional chances. If such gaps are among the options of a decision problem for an agent, the Gibbard-Harper formulation and the Skyrms formulation will presumably come apart, contra the lore.

This, in turn, makes me suspicious of the other centerpiece of the lore: that causal decision theory and evidential decision theory typically agree. Certainly, if the causal decision theories don’t all agree with each other, even on garden variety decision problems, then they don’t all agree with evidential decision theory on those problems! Take Gibbard and Harper’s version as representative of causal decision theory—according to the lore, it should do as well as any other version. It replaces evidential decision theory’s conditional probabilities of the form P(state | act) with probabilities of conditionals of the form P(act □→ state). But apparently the triviality results collectively teach us that probabilities of conditionals come apart from conditional probabilities easily, often, and by a lot.4 (My result in this paper will only drive this point home further.) And they do so for ‘umbrella’ and ‘wine’ propositions as much as for ‘Newcomb’ propositions. To be sure, expected utilities are sums of products of such probabilities and utilities, so they might somehow compensate for the differences between the probabilities of conditionals and the corresponding conditional probabilities taken individually. Well, fingers crossed! It would seem almost miraculous if all the differences between the two kinds of quantities washed out like this. And this article of faith is surely not something that we should be left to intuit—it requires proof. The foundations of decision theory deserve no less.

Turning this thought on its head: causal decision theorists should hope that the various formulations are not equivalent. If they were, we should have no confidence that causal decision theory delivers correct results in garden-variety decision problems—for we are confident that evidential decision theory does. After all, if the Gibbard-Harper formulation comes apart from evidential decision theory even for those problems, then it is in trouble; and if it is a good representative of all the causal decision theories, then they all are in trouble. And yet ironically, the causal decision theorist presumably wants the triviality results to have some purchase. After all, if probabilities of conditionals really are conditional probabilities, then the Gibbard-Harper theory is equivalent to Jeffrey’s; but since the latter is simpler, then the former is otiose, as are all its allegedly equivalent reformulations. Thirty-five years of literature on causal decision theory would then appear to be a big red herring. So the causal decision theorist who upholds the lore should hope that the triviality results work—but not too well!

4 Why Believe the Theses?

So much for reasons for caring about Stalnaker’s Thesis and other theses with various restrictions on the probability functions that conform to PCCP. Why should we believe them? Consider Stalnaker’s Thesis in particular—hereafter, ‘the Thesis’.

4.1 It Sounds Right

We have already seen what is surely the main reason: it sounds right (see van Fraassen 1976; Bennett 2003). The slogan that ‘probabilities of conditionals are conditional probabilities’ rolls easily off the tongue. And case-by-case evidence seems to support it—remember the ‘Collingwood’ example.

I think this is a rather weak reason to believe the Thesis. The slogan that ‘probabilities of limits are limits of probabilities’ rolls off the tongue with comparable ease, and we could marshal much case-by-case evidence to support it. But that thesis is controversial—it is guaranteed to hold only for countably additive probability functions, and friends of merely finite additivity such as de Finetti will insist that it does not generalize. The methodology of case-by-case checking is unreliable in any case. I surmise that we typically test mentally a handful of simple cases like the ‘Collingwood’ one, give them a tick, and then hastily generalize our positive appraisal of them to the infinitely many remaining cases. But we know that this methodology over-generalizes. Exhibit A: the claim that in the probability space of the 3-ticket lottery, the probability of ‘if one of the first two tickets wins, then the first one wins’ equals the conditional probability of ‘the first one wins, given one of the first two tickets wins’ sounds right. But we know that it is wrong, as we saw above. More generally, we apparently know that probabilities of conditionals come apart from conditional probabilities easily, often, and by a lot. But they keep sounding like they should agree.

4.2 The Two Sides are Structurally Alike

A better approach is to argue more systematically why the two sides of PCCP should coincide. After all, they seem structurally and semantically alike. On both sides, the probability takes wide scope. Both sides involve a restriction to the cases in which A is true; the probability is to measure, among those cases, how likely are the B cases.

4.3 The Ramsey Test

In the most famous passage ever written on conditionals, Ramsey (1965) writes: “If two people are arguing ‘If p will q?’ and are both in doubt as to p, they are hypothetically adding p to their stock of knowledge and arguing on that basis about q.” Applied to your own evaluation of ‘if A, then B’, the Ramsey test bids you first to hypothetically add A to your system of beliefs, minimally revising what you currently believe in order to do so; second, to evaluate B on the basis of your revised body of beliefs. P(A → B) measures how well the conditional performs on Ramsey’s test. But apparently P(B | A) does too. For conditioning on A prima facie seems to capture the notion of ‘minimally revising what you currently believe in order to accommodate A’; and your evaluation of B in your new belief state P(_ | A) is just P(B | A).

Better still, Ramsey offers a version of the test based on the corresponding conditional probability. He goes on to say: “We can say that they are fixing their degrees of belief in q, given p.” Now the identification of the probability of the conditional with the corresponding conditional probability is even more explicit. The Ramsey test has much currency even today—so much so that it is often taken to be the starting point of theorizing about conditionals.

Ramsey’s imaginary protagonists are arguing about an indicative conditional, and plausibly the Ramsey test should be restricted to such conditionals. But this is all to the good as far as the Thesis is concerned. For conditional probability has a similarly ‘indicative’, rather than ‘counterfactual’, nature. Correspondingly, conditioning is meant to capture updating rather than supposing. We assume that A is compatible with your belief state, and the indicative conditional A → B has you reason about B under the assumption that A is the case. But so does the conditional probability, P(B | A).

4.4 Adams’ Thesis

Assertability is said to go by subjective probability of truth. (Or at least it usually does, though not always: for example, the assertability of ‘A but B’ differs from that of ‘A and B’, yet the two have the same subjective probability, namely P(A & B).) At first sight, the indicative conditional appears to provide a counterexample to this dictum. After all, according to the popular understanding of Adams’ Thesis, the assertability of A → B equals P(B | A)—and even someone who, unlike Adams, believes that the indicative conditional has truth conditions, may find Adams’ Thesis about its assertability compelling. (Of course, someone like Adams thinks that the indicative conditional provides a counterexample to the dictum, since according to him, there is no such thing as the probability of its truth.) It would be nice for proponents of the truth-conditional view of the indicative conditional if it conformed to the dictum. If Stalnaker’s Thesis were true, it would then explain Adams’ Thesis admirably: the assertability of A → B goes by P(A → B) (as per the dictum), which in turn equals P(B | A). We infer the truth of Stalnaker’s Thesis from Adams’ Thesis since the Thesis provides a good—perhaps the best—explanation of it. This is the motivation for the Thesis discussed in Lewis (1976) and (1986a).

So much for reasons to want Stalnaker’s Thesis to be true, and to believe that it is true. However, there are much stronger reasons to believe that it is false. Lewis initiated the industry of providing them in his classic (1976) paper, with his so-called ‘triviality results’. To these I now turn.

5 Lewis’s Triviality Results

While Lewis presents two triviality results in this paper, logically speaking there is only one: the second one entails the first, and their proofs are almost identical—so much so that it is perhaps a little surprising that Lewis distinguishes them. So let us confine our attention to the stronger result. A number of authors misstate it. For example, Adams (1975) claims that in his (second) triviality result, “Lewis assumes that probabilities change by conditionalization” (8). But Lewis’ result makes no such assumption. To be sure, this assumption plays an important role in motivating the interest of the result; but that’s another matter (and even the interest of the result is not so-confined, as I will soon argue). I mention this because, as we will see, Lewis’s result is not exactly intuitive, and it is rather complicated to state precisely. It is surprisingly hard both to paraphrase it and to get it exactly right. (Exercise for those who think they already know the result: try doing so yourself!)

Some notation and terminology: P X (_) is the function that one gets from P(_) by conditioning on X—that is, for all A and X in the domain of P,
$$ P_{X} \left( A \right) = P\left( {A|X} \right),\quad {\text{provided}}\,P\left( X \right) > 0. $$

If PCCP holds for the pair <→, P > , we will say that → is a probability conditional for P. If → is a probability conditional for every probability function in some class of probability functions, we will call → a probability conditional for the class. Say that P is a trivial probability function if there are not two propositions A and B such that P(A) < 1, P(A & B) > 0 and P(A & ¬B) > 0.

Lewis assumes the following, but he only sketches its proof. I will give the proof in more detail.

Import–Export Lemma

If PCCP holds throughout a class of probability functions that is closed under conditioning, then the following also holds throughout the class:
$$ P(A \to B\left| C \right.) = P(B\left| A \right.\& C),\quad {\text{if}}\,P\left( {A\& C} \right) > 0. $$


Suppose that PCCP holds throughout a class Open image in new window of probability functions that is closed under conditioning, let P belong to Open image in new window , and suppose that P(A & C) > 0.
$$ \begin{aligned} P(A \to B\left| C \right.) & = P_{C} (A \to B),\quad {\text{by definition of}}\,P_{C} \\ & = P_{C}(B\left| A \right.),\quad {\text{by PCCP applied to}}\,P_{C} , \\ {\text{since\,}}P\left( {A\,\& \,C} \right) > 0{\text{ and hence}}P_{C} \left( A \right) > 0 \\ & = P_{C} (A\,\& \,B)/P_{C} (A) \\ {\text{by the formula for conditional probability}},{\text{ since}}P_{C} \left( A \right) > 0 \\ & = P\left( {A\,\& \,B\left| C \right.} \right)/P\left( {A|C} \right),{\text{ by definition of}}\,P_{C} \\ & = P\left( {\left( {A\,\& \,B} \right)\,\& \,C} \right)/P\left( {A\,\& \,C} \right),{\text{ by the formula for conditional probability}},\,{\text{since}}\,P\left( {A\& C} \right) > 0{\text{ and hence}}\,P\left( C \right) > 0 \\ & = P\left( {B|A\,\& \,C} \right),{\text{ by the formula for conditional probability}}. \\ \end{aligned} $$

So what exactly is Lewis’s result? Here it is.


If a class of probability functions is closed under conditioning, then there is no probability conditional for that class unless the class consists entirely of trivial probability functions.


Suppose for reductio that PCCP holds throughout a class of probability functions that is closed under conditioning. Take any probability function P in the class and any sentences A and B such that P(A & B) > 0 and P(A & ¬B) > 0.
$$ \begin{aligned} P(A \to B) & = P(A \to B\left| B \right.)P\left( B \right) + P(A \to B\left| {\neg B} \right.)P(\neg B) \\ {\text{by \, the \, law \, of \, total \, probability}}, \, {\text{ since}} \, P\left ( {A\& B} \right){\text{ and \, hence}} \, P\left ( B \right) > 0,{\text{ and \, since}} \, P\left ( {A\& \, \neg B} \right){\text{ and \, hence}} \, P\left ( {\neg B} \right) > 0. \\ & = P (B\left| A \right.\& B)P(B) + P(B\left| A \right.\& \neg B)P(\neg B),\quad {\text{by \, the \, Import}}-{\text{Export \, Lemma}} \\ & = P\left( B \right) + 0 \\ & = P\left( B \right). \\ \end{aligned} $$

6 Questioning Lewis’s Results

6.1 Against Conditioning

Conditioning has a central place in Bayesian orthodoxy. However, for some purposes it seems inadequate—for example, for updating credences on sentences involving indexicals (especially temporal ones), or for modeling certainty loss (see Titelbaum 2008). Indeed, some authors have questioned the rationality of updating by conditioning quite generally (see Bacchus et al. 1990; Hild 1998). Let us grant for the sake of argument, pace Lewis, that conditioning is not a rational updating method. This is not sufficient to undermine the importance of Lewis’s result, however, for it does not impugn the rationality of two different agents whose probability functions are related by conditioning (note that according to Adams’ presentation of the result, which I quoted above, conditioning not being a rational updating method is sufficient to undermine the importance of the result—this shows how getting the details right matters). Still, it would be nice not to make any assumption at all about → being a probability conditional for functions related by conditioning.

While van Fraassen is not opposed to conditioning—his permissive epistemology is hard to offend—he is hardly its most ardent supporter. After all, he believes that rationality does not require conditioning (see his 1989). Indeed, by his lights rationality does not require rule-following at all, permitting wholesale discarding of one’s priors, and epistemic leaps that are not grounded solely in evidence. While not undermining Lewis’s result itself, this view threatens to undercut its importance a little. Conditioning no longer enjoys pride of place. So we should welcome a result that obviates the need to appeal to it.

6.2 The Conditional as Probability-Sensitive

Lewis assumes that the conditional is fixed, etched in stone irrespective of the probability function in whose scope it appears. In particular, he assumes that the same ‘→’ appears in the arguments of P, P B , and P ¬B . But van Fraassen (1976) demurs: “Would it not seem rather, that our probabilities are inextricably involved in the way we represent the possibilities, and nearness relations among them, to ourselves… if our ideas about the one change, will we not revise our modelling of the other?” (p. 274). He suggests that the conditional may be radically sensitive to the probability function—every change in the latter brings about a change in the former. On this view, we do not have a single conditional at all, but rather one conditional that we could denote ‘→ P ’ when it appears in the scope of P, another conditional ‘→ P B ’ when it appears in the scope of P B , and yet another conditional ‘→ P B ’ when it appears in the scope of P ¬B .5

7 Generalizing Lewis’s Triviality Result

Now I will generalize Lewis’s triviality result. Doing so will go a long way towards answering these objections, and it will have other benefits besides.

For our purposes, a probability revision rule takes as input an initial probability distribution and a proposition, and yields as output a (new) probability distribution. We will regard an agent’s system of beliefs as being represented by a probability function, and a change in the belief system to accommodate certainty about the truth of some proposition as taking place according to some revision rule, such that the proposition gets probability 1 after the revision.

As we have seen, Lewis focuses on the revision rule of conditioning. However, there are other well-known rules—here are four.
  • Imaging on some proposition E is the rule that moves all the probability from each ¬E world to its nearest neighbor inside E—‘nearest’ as determined by the relevant similarity relation (this assumes, as Stalnaker (1968) does, that there is always a unique such neighbor).

  • A blurred imaging on E removes all the probability from a ¬E world, but spreads it over more than one ‘nearest’ E world. Various ways of doing this have been proposed—see Gärdenfors (1982) and Lewis (1986a, b) for more details and references.

  • Maxent is the rule that takes an initial probability function and a constraint, and yields the probability function of maximum entropy that satisfies the constraint. It will suffice for my purposes to define ‘entropy’ for discrete cases: if countably many worlds i = 1, 2, … receive positive probabilities P(i) then the entropy of this distribution is −∑ i P(i) log P(i). If the constraint on possible final probability measures is ‘the probability of E equals 1’, then we have the case of accommodating E by maxent; the distribution that is produced is uniform over E.

  • Minxent, or Minimum Cross Entropy, revises a probability function P to the nearest probability function Q that meets a given constraint, where ‘nearness’ is measured by the so-called Kullback–Leibler distance, which for discrete distributions P and Q is given by:

$$ D_{\text{KL}} \left( {P||Q} \right) = \sum_{i} P\left( i \right)\log {\frac{P(i)}{Q(i)}} $$

I will show that similar results to Lewis’ are available for imaging, various blurred imagings, maxent, minxent, and other revision rules. Lewis’s second result will fall out as a corollary, but my result will be more general than it is. Like his result, and unlike those of Stalnaker (1976) and Hall (1994), mine makes no assumption about the logic of the conditional.

8 Boldness, Conservativeness, Moderation

Suppose we have an initial probability distribution P, and we want to revise it in order to accommodate a proposition E. We use some revision rule to derive a distribution P E (here I use more generally the notation that previously I used just for the revision rule of conditioning). Call the rule bold if for any P and for any E such that P(E) > 0, P E (E) = 1. ‘Bold’, for two reasons. Firstly, the rule is prepared to take as input any initial probability function and proposition. Secondly, the function produced as output by the rule is not afraid to commit itself fully, giving probability 1 to the proposition (as it must, in order for the proposition to be genuinely accommodated). Bold revision rules take us from some initial probability distribution to a new distribution that is fully concentrated on any proposition that we specify. Conditioning, imaging, the varieties of blurred imaging, maxent and minxent are all bold.

Furthermore, some revision rules are what we may call conservative. Such a rule takes a function P and a proposition E to a function P’ with the following property: for any A that implies E, P′(A) ≥ P(A). A conservative rule never decreases the probabilities of propositions that imply the proposition that is accommodated. Conditioning is conservative; so is imaging and various blurred imagings, assuming that the similarity relation obeys centering (each world is closer to itself than any other world is to it). Still more rules are what I will call moderate: for any A that implies E, if P(A) > 0, then P′(A) > 0. Moderate rules can decrease the probabilities of propositions that imply the proposition that is accommodated, but never all the way to 0.

Moderation is surely a desideratum of a revision rule: your belief system would be highly fragile if you could suddenly fully disbelieve some proposition A that previously you gave some credence, when you learned something implied by A. By the lights of your original credence function, A is confirmed (assuming the usual Bayesian account of confirmation); but by the lights of your new credence function, it is maximally disconfirmed. A little more precisely: if P(E) < 1, A entails E, and P(A) > 0, then A and E are initially correlated:
$$ P(E\left| A \right.) > P\left( E \right). $$
But if P is non-trivial and is revised by a rule that it is not moderate, then there are some A and E such that all these conditions are met, E is the strongest proposition that is learned, and yet A and E are maximally anti-correlated after the revision:
$$ P(E\left| A \right.) = 0. $$

It is hard to see how anything that bears so radically on the correlation between E and A has been learned. The radical revision in the assessment of A’s evidential bearing on E seems entirely gratuitous, and certainly not justified by the evidence E itself.

All conservative rules are moderate. Moreover, so is maxent and minxent (at least for the finite cases that I have defined above). So too are various blurred imagings, even when the similarity relation obeys merely weak centering: any world is one of the nearest worlds to itself.

Even if, pace Lewis, probability functions in a rational epistemic history are not related by conditioning, a negative result analogous to his holds, provided that they are related by some particular rule that is bold and moderate; and similarly if the probability functions of two different rational agents are related by such a rule.

9 The Main Theorem

Here’s the idea. Suppose that → is a probability conditional for a particular non-trivial probability function. Then we can easily find another probability function for which it is not—indeed, we can exhibit a specific pair of propositions A and B for which PCCP fails for this function. We use a bold and moderate rule to shift all probability to a particular proposition that depends on A, B, and A → B. Once shifted, the two sides of PCCP are cleaved apart—one side zero, the other positive. The algebraic proof is a bit fiddly because we have to ensure that the relevant conditional probabilities are always defined (though the proofs of Lewis’s result and the Import–Export Lemma are even more fiddly in this regard). But the underlying idea is simple.

Main theorem

Let P be a non-trivial probability function, let  be a probability conditional for P, and let ℜ be a bold6and moderate revision rule. Then there exists a proposition X such that  is not a probability conditional for the function that results from P by accommodating X using ℜ.


Let P be non-trivial. Then there are two propositions A and B such that P(A) < 1, P(A & B) > 0 and P(A & ¬B) > 0. Let be a bold and moderate rule. There are two cases:

Case 1. (A & B) − (A  B) is non-empty7

Let X = (A & B) − (A → B). Use to revise P to accommodate X, (as we know we can, by ’s boldness) thus producing P X . Clearly P X (A → B) = 0, since we have accommodated X, and X is incompatible with A → B. But P X (B | A) = 1, since P X (X) = 1, and X implies A & B, so that P X (A & B) = 1, and thus P X (A) = 1, ensuring that this conditional probability is defined. So → is not a probability conditional for P X .

Case 2. (A & B) − (A  B) is empty

¬(A & B) is hatched with lines oriented from bottom left to top right, A → B is hatched with lines oriented from top left to bottom right; they intersect in the double-hatched blob that hangs below A & B. (The blob may also invade A & ¬B, although in that case → violates modus ponens. I want to allow this since I don’t want this result to rest on any assumptions about the logic of → ).

The blob has positive probability:
$$ P\left( {\left[ {A \to B} \right] - \left( {A\,\& \,B} \right)} \right) > 0 $$
for otherwise P(A → B) = P(A & B), but then by PCCP, P(B | A) = P(A & B), contradicting our assumption that 0 < P(A & B) and P(A) < 1.
Let X = ¬(A & B). Use to revise P to accommodate X (as we know we can, by ’s boldness), thus producing P X . (1) implies that
$$ P_{X} \left( {[A \to B] - \left( {A\,\& \,B} \right)} \right) > 0 $$
since [A → B]−(A & B) entails X, and is moderate. The blob retains positive probability after the revision.
(2) implies that
$$ P_{X} \left( {A \to B} \right) > 0 $$
since probability is non-decreasing through entailment. (The blob is part of A → B.)
$$ P_{X} \left( {A\& B} \right) = 0, $$
since P X (X) = P X (¬(A & B)) = 1, while
$$ P_{X} \left( A \right) > 0 $$
since P X (A & ¬B) > 0, by being moderate and A & ¬B entailing X, and since probability is non-decreasing through entailment. By (4) and (5),
$$ P_{X} \left( {B|A} \right) = 0 $$

((5) was needed to ensure that this conditional probability is defined.) (3) and (6) imply that → is not a probability conditional for P X .□

You may have wondered why I didn’t use much the same trick in case 2 as I did in case 1: simply use to move all probability onto (A → B) − (A & B) (instead of onto ¬(A & B)), so that the probability of A → B becomes 1 while the conditional probability of B, given A, becomes 0. In fact, couldn’t I have appealed simply to ’s boldness, dispensing with its moderateness and strengthening my result accordingly? No—for we do not want the conditional probability to go undefined, as it would if the probability of A became 0. We have good reason to think that (A → B) − (A & B) is not compatible with A, for their compatibility would require a failure of modus ponens. At a couple of points I needed to be moderate to ensure that the relevant conditional probabilities are defined.


If a class of probability functions is closed under a bold and moderate rule, then there is no probability conditional for that class unless the class consists entirely of trivial probability functions.


Suppose for reductio that there is a probability conditional → for a class Open image in new window consisting of at least one non-trivial probability function, and that the class is closed under a bold and moderate rule ℜ. Choose a non-trivial probability function P from Open image in new window , and an A and B as in the proof of the main theorem. We may use to revise P in order to accommodate either X = (A & B) − (A → B) (case (i)), or X = ¬(A & B) (case (ii)), yielding a probability function P X for which → is not a probability conditional. But by Open image in new window ’s closure under , P X belongs to Open image in new window . This contradicts our assumption that → is a probability condition for the (entire) class Open image in new window . □

Lewis’s second triviality result is a corollary of this corollary, since conditioning is a bold and moderate revision rule. But this corollary is stronger, since there are other such rules. The main theorem is stronger still. And it dispenses with Lewis’s somewhat complicated notion of a class of probability functions being closed under a revision rule. Just one (non-trivial) probability function and one application of the rule to that function suffices to produce a violation of PCCP—and we can identify a specific A → B for which the violation occurs.

10 For Aficionados: Further Generalizations

Lewis (1986a) went on to prove a third and fourth triviality result, extending his earlier results to cover different closure assumptions, respectively, closure under conditioning on the members of a single finite partition, and closure under (2-celled) Jeffrey conditioning. As Bas van Fraassen has pointed out to me, we could modify the proof of the main theorem to get a result that has Lewis’s third triviality result as a corollary, provided we could assume that contained within (A & B) − (A → B) in Case 1, and within ¬(A & B) in Case 2, there is some member of the finite partition, upon which we can conditionalize.

And Aidan Lyon has suggested to me that we could also modify the proof to get a result that has Lewis’s fourth triviality result as a corollary. So far we have only considered revision rules that revise probability functions to fully accommodate particular propositions. But we could also consider rules that revise probability functions to partially accommodate these propositions, yielding functions that assign them specified probabilities that fall short of 1. Jeffrey conditioning is such a rule, but we could countenance various others. Instead of using a bold revision rule to move all probability to designated propositions (which I labeled ‘X’), as I did in the original proof, we could use a confident but not-quite-bold rule to move sufficiently large amounts of it, so as still to cleave apart the two sides of PCCP. In the version of the new proof that I have come up with—following Lyon’s leads—the moderation of the revision rule needs to be strengthened to an analogue of conservativeness (appropriate for partial accommodation). But the broad outline of the proof is as before, with Case 1 and Case 2 and the designated propositions as in the original proof.

I have decided not to reproduce the new proof here, as it is somewhat tedious, and I have a sense that it is overkill at this stage; this paper is long enough as it stands! I think that the main theorem is already telling against Stalnaker’s Thesis. The next section explains why.

11 Vale Stalnaker’s Thesis: and Weaker Theses?

The main theorem shows how precarious PCCP is: while it may hold for a single probability function (for all that the theorem tells us), it is easily torn asunder. If you are a Bayesian agent who seeks to conform to PCCP at all times, you are apparently unable to revise boldly and moderately your opinions regarding certain propositions. These propositions then have a curious status for you: you give them positive credence, but you can never learn them—where learning is modeled by a bold and moderate rule. Borrowing terminology from van Fraassen (1984), they are ‘Moore propositions’—propositions that you cannot learn without violating a structural constraint that is imposed on you (in this case, the upholding of PCCP). There is something Moore paradoxical about your saying, or thinking: “p has positive probability, but it is impossible for me to learn it”, where p may be easily accessible to your enquiry. After all, the propositions—labeled ‘X’—that are accommodated in the proof of the main theorem are quite generic.

Similarly, the Bayesian agent cannot suppose them, where supposing is modeled by a bold and moderate rule. Collins (1991) and Joyce (1999) argue that supposing should be modeled by imaging; Gärdenfors (1982) generalizes it to blurred imaging. Jaynes (2003) advocates maxent as a revision rule. And Jaeger (1995) favors minxent as a revision rule.

Here, a defender of some version of the Thesis might appeal to interpretivism about credences, an influential position in the philosophy of mind: your mental state is the set of probability and utility functions that rationalize your behavioral dispositions as well as possible. We may think of this state as the one attributed to you by an ideal interpreter, whose task is to represent you as charitably as possible (see Lewis 1974). If rationality constrains your credences to obey PCCP, then perhaps an ideal interpreter will always attribute to you PCCP-obeying credence functions, in which case your credences do in fact always obey PCCP. The trouble is that rationality then apparently may overconstrain you, for at the same time it may require you to accommodate boldly and moderately a proposition of the kind I designated ‘X’. The ideal interpreter is then unable to represent you as entirely rational—either you violate PCCP, or you are diachronically irrational. Lewis already made this ideal interpreter’s task an unenviable one with his result about conditioning. To the extent that there are other rational ways of revising one’s credences boldly and moderately on ‘X’ propositions, I have only made the interpreter’s job harder.

In any case, echoing a remark I made earlier about Lewis’s results, the probability functions related by bold and moderate revisions need not be credences of a single agent at different times; they may be those of two different agents at a single time. The upholder of Stalnaker’s Thesis, or even various weaker theses, must insist that at least one of the agents is irrational. This sounds like big news for our theory of rationality.

Since the main theorem extends Lewis’s result to a number of other important revision rules, objections merely to conditioning do nothing to undercut my results—the objections would need to carry over to the other rules as well. One objection does: recall van Fraassen’s insistence that rationality does not require rule-following at all, so by his lights it does not require imaging, or blurred imaging, or revision by maxent, minxent, or any other bold and moderate rule. But while I found it convenient to state the theorem in terms of a revision rule, inspection of the proof shows that what matters is the revision, not the rule. Wholesale discarding of one’s priors and epistemic leaps of faith are just as problematic for the upholding of PCCP as long as they have the effect of shifting all probability onto one of the propositions I designated ‘X’. They may take place even though they are not rule-governed. Surely nothing in van Fraassen’s permissive epistemology prevents this. On the contrary, its very permissiveness only makes such shifts all the more permissible!

My result also goes some way to mitigating van Fraassen’s concern about the fixed interpretation of the conditional throughout changes in probability functions. It does not go all the way—to be sure, some changes in the identity of a given conditional may sustain PCCP through the sorts of probability revisions that I have envisaged. However, some do not; in fact, some may make matters worse, driving the two sides of PCCP even further apart. For example, where P X (A → B) was positive in my proof, it could be even greater if A → B grows through the probability revision, expanding to include all the worlds it previously included and more besides. This would drive P X (A → B) even further away from 0 = P X (B | A).

Like Lewis’s result, mine assumes nothing about the logic of the conditional beyond its being a 2-place connective. You may replace the ‘→’ throughout with some less evocative symbol throughout—say, Open image in new window . There is no binary operator Open image in new window , conditional-like or otherwise, such that the equation P(A Open image in new window B) = P(B | A) survives all revisions by a given bold and moderate revision rule. Conditional probability is not a special kind of unconditional probability, however one might construe that kind. It’s the probability that drives my result rather than the logic.

But in fact my result needs virtually no probability theory either. For example, it does not assume that probabilities are additive, nor the law of total probability. Of course, to the extent that the Thesis is one concerning the probabilities of conditionals, we should not begrudge these assumptions. Still, it is all the better if we can do without them, as I have.

12 Transmogrifying Stalnaker’s Thesis into Adams’ Thesis

The last observation brings to mind Adams’ Thesis, the variant of Stalnaker’s Thesis, applied to indicative conditionals that lack truth values. Lewis (1976) concedes that Adams need not be troubled by his triviality results:

Adams does say that conditional probabilities are probabilities of conditionals. Nevertheless he does not mean by this that the indicative conditional is what I have here called a probability conditional; for he does not claim that the so-called “probabilities” of conditionals are probabilities of truth, and neither does he claim that they obey the standard laws of probability. They are probabilities only in name. Adams’s position is therefore invulnerable to my triviality results, which were proved by applying standard laws of probability to the probabilities of conditionals…But if it be granted that the “probabilities” of conditionals do not obey the standard laws, I do not see what is to be gained by insisting on calling them “probabilities”. It seems to me that a position like Adams’s might best be expressed by saying that indicative conditionals have neither truth values nor probabilities, and by introducing some neutral term such as “assertability” or “value” which denotes the probability of truth in the case of nonconditionals and the appropriate conditional probability in the case of indicative conditionals (pp. 303–305).

It might look like my result can claim another victim, scotching any interesting version of Adams’ Thesis. With my assuming so little about ‘P’, perhaps it can just as well be interpreted as representing “assertability”, in which case is it still answerable to the main theorem?

Alas, no. For recall that the reason that Adams-style assertabilities do not obey the probability calculus is that their domain is not a field. In particular, Adams assumes that indicative conditionals cannot be negated or conjoined with other sentences. So Adams would baulk at Boolean combinations such as (A & B) − (A → B) or (A → B) − (A & B), the centrepieces of my proof.

In my opinion, far from saving Adams’s position from trivialization, this only reveals its incompleteness. For English takes such combinations in its stride. “If Oswald didn’t shoot Kennedy then somebody else did, but it’s not the case that both those things happened—so Oswald must have shot Kennedy” seems perfectly felicitous. More generally, Adams must face the familiar ‘Frege-Geach’-style objection that indicative conditionals compound truth-functionally with other sentences. That said, various ‘no-truth-value’ theorists in the Adams tradition, such as McGee (1989), Edgington (1995) and Bennett (2003) face up to this problem, and this is still an area of lively debate. Indeed, the very triviality results against Stalnaker’s Thesis motivate their no-truth-value accounts. The upshot is that my result here does not scathe Adams’ Thesis.

I have another triviality result that I believe does. It is based on my (1989) theorem, strengthened in Hájek and Hall (1994), that PCCP cannot hold for any finite-ranged probability function. I show that for any such function, there are more distinct conditional probability values than there are distinct unconditional probability values, so there is no way of matching all the former with all the latter. A fortiori, there is no way of matching all the distinct conditional probability values with probability-of-conditional values—there will always be some unmatched conditional probability values. If Adams’ Thesis about assertability is to hold, then assertability must be more nuanced than unconditional probability: conditional probability values outrun unconditional probability values, so if assertability values are to keep pace, they too must outrun unconditional probability values. I doubt that this is the case: if anything, the latter outrun the former. But this is just a promo for a coming attraction; for the full story, see this paper’s sequel (Hájek forthcoming).

13 Final Scene

For now, I am content to have chased down and cornered Stalnaker’s Thesis some more. I have closed off more of its escape routes. It appears to be moribund. But as fans of the Friday the 13th series know, you should never be surprised if Jason suddenly reappears in a new guise—and the same is true of Stalnaker’s Thesis.


  1. 1.

    Page references are to Harper et al. (1981) for all articles that are cited as appearing there.

  2. 2.

    I owe this point to Ned Hall.

  3. 3.

    I will speak interchangeably of ‘propositions’ and ‘sentences’, and I will use both set-theoretic operations and sentential connectives, in a way that I hope is perspicuous.

  4. 4.

    At the end of the paper I will briefly discuss ‘no truth value’ accounts of indicative conditionals, which seemingly evade the triviality results, suggesting that even these accounts may not evade them altogether after all.

  5. 5.

    One wonders what the conditional becomes when it appears in the scope of two or more probability functions—when, for example, P assigns probabilities to various hypotheses about another function P*’s assignments! (Shades here of the contextualist and relativist literature on reported speech and eavesdropping cases.).

  6. 6.

    I don't actually need the full strength of the assumption of boldness in the proof. All I need is that, for any non-trivial probability function P, and propositions A and B as described in the proof, the rule can revise P so as to give probability 1 to (A & B) − (A□ → □B), or so as to give probability one to (A & B). That is certainly weaker than boldness, but I don't know a neat way of characterizing it.

  7. 7.

    As I hope is obvious, P − Q is shorthand for P & ¬Q.



I am indebted to Ned Hall, Richard Jeffrey, David Lewis, Bas van Fraassen, and Lyle Zynda for very helpful comments on a short, early precursor to this paper. I am grateful to participants in a reading group in Leuven for discussion of a resuscitated version of it—especially Jake Chandler, Igor Douven, and David Etlin. I have recently substantially revised and expanded it. For further valuable comments on this transmogrified version, I thank especially Rachael Briggs, John Cusbert, Daniel Greco, Aidan Lyon, Daniel Nolan, Wolfgang Schwarz, and Dan Singer. Thanks also to Brett Calcott and Ralph Miles for their help.


  1. Adams E (1975) The logic of conditionals. Reidel, DordrechtGoogle Scholar
  2. Bacchus F, Kyburg HE, Thalos M (1990) Against conditionalization. Synthese 85:475–506CrossRefGoogle Scholar
  3. Bennett J (2003) Conditionals. Oxford University Press, OxfordGoogle Scholar
  4. Cartwright N (1999) The dappled world: a study of the boundaries of science. Cambridge University Press, CambridgeGoogle Scholar
  5. Collins J (1991) Belief revision. Ph.D. dissertation, Princeton UniversityGoogle Scholar
  6. de Finetti B (1972) Probability, induction and statistics. Wiley, New YorkGoogle Scholar
  7. Edgington D (1995) On conditionals. Mind 104(414):235–329CrossRefGoogle Scholar
  8. Eells E, Skyrms B (eds) (1994) Probability and conditionals. Cambridge University Press, CambridgeGoogle Scholar
  9. Gärdenfors P (1982) Imaging and conditionalization. Journal of Philosophy 79:747–760CrossRefGoogle Scholar
  10. Gibbard A, Harper W (1978) Counterfactuals and two kinds of expected utility. In: Hooker CA, Leach JJ, McClennen EF (eds) Foundations and applications of decision theory, vol 1. Reidel, DordrechtGoogle Scholar
  11. Hájek A (1989) Probabilities of conditionals—revisited. J Philos Logic 18:423–428CrossRefGoogle Scholar
  12. Hájek A (1994) Triviality on the cheap? In: E Eells, B Skyrms (eds)Google Scholar
  13. Hájek A (forthcoming) The fall of Adams’ thesis? J Lang Logic InfGoogle Scholar
  14. Hájek A, Hall N (1994) The hypothesis of the conditional construal of conditional probability. In: E Eells, B Skyrms (eds)Google Scholar
  15. Hall N (1994) Back in the (CCCP). In: E Eells, B Skyrms (eds)Google Scholar
  16. Harper WL, Skyrms B (1988) Introduction to causation, chance, and credence. Kluwer, DordrechtGoogle Scholar
  17. Harper WL, Stalnaker R, Pearce G (eds) (1981) Ifs. Reidel, DordrechtGoogle Scholar
  18. Hild M (1998) The coherence argument against conditionalization. Synthese 115:229–258CrossRefGoogle Scholar
  19. Jaeger M (1995) “Minimum cross entropy reasoning” International joint conference on artificial intelligence. In: Proceedings of the 14th international joint conference on artificial intelligence, Vol 2, 1847–1852Google Scholar
  20. Jaynes ET (2003) Probability theory: the logic of science. Cambridge University Press, CambridgeCrossRefGoogle Scholar
  21. Jeffrey R (1966) The logic of decision. University of Chicago Press, Chicago, 2nd edn, 1983Google Scholar
  22. Joyce JM (1999) The foundations of causal decision theory. Cambridge University Press, CambridgeCrossRefGoogle Scholar
  23. Levi I (1980) The enterprise of knowledge. MIT Press, Cambridge, MAGoogle Scholar
  24. Lewis D (1973) Counterfactuals. Blackwell and Harvard University PressGoogle Scholar
  25. Lewis D (1974) Radical interpretation. Synthese 27: 331–344; reprinted in Philosophical Papers, Vol 1. Oxford University Press, New York and Oxford, 1983Google Scholar
  26. Lewis D (1976) Probabilities of conditionals and conditional probabilities. Philos Rev 85: 297–315; reprinted in Harper et al.Google Scholar
  27. Lewis D (1981) Causal decision theory. Aust J Philos 59: 5–30; reprinted in Lewis (1986a)Google Scholar
  28. Lewis D (1986a) Probabilities of conditionals and conditional probabilities II. Philos Rev 95:581–589CrossRefGoogle Scholar
  29. Lewis D (1986b) Philosophical papers, vol II. Oxford University Press, OxfordGoogle Scholar
  30. McGee Vann (1989) Conditional probabilities and compounds of conditionals. Philos Rev 98(4):485–541CrossRefGoogle Scholar
  31. Milne P (2003) The simplest Lewis-style triviality proof yet? Analysis 63.4(October): 300–303Google Scholar
  32. Ramsey FP (1931/1990) Truth and probability. In: RB Braithwaite (ed) Foundations of mathematics and other essays. Routledge & P. Kegan, London; reprinted in DH Mellor (ed) (1990) Philosophical Papers. Cambridge University Press, CambridgeGoogle Scholar
  33. Ramsey FP (1965) The foundations of mathematics (and other logical essays). Routledge and Kegan Paul, LondonGoogle Scholar
  34. Rehder W (1982) Conditions for probabilities of conditionals to be conditional probabilities. Synthese 53:439–443CrossRefGoogle Scholar
  35. Savage LJ (1954) The foundations of statistics. Wiley, New YorkGoogle Scholar
  36. Skyrms B (1980) Causal necessity. Yale University Press, New HavenGoogle Scholar
  37. Skyrms B (1984) Pragmatics and empiricism. Yale University Press, New HavenGoogle Scholar
  38. Sobel JH (1994) Taking chances: essays on rational choice. Cambridge University Press, New YorkGoogle Scholar
  39. Stalnaker R (1968) A theory of conditionals. Studies in Logical Theory. American Philosophical Quarterly Monograph Series, No. 2. Blackwell, OxfordGoogle Scholar
  40. Stalnaker R (1970) Probability and conditionals. Philos Sci 37: 64–80; reprinted in Harper et al.Google Scholar
  41. Stalnaker R (1976) Letter to van Fraassen. In: Harper WL, Hooker CA (eds) Foundations of probability theory, statistical inference and statistical theories of science, vol I. Reidel, Dordrecht, pp 302–306Google Scholar
  42. Titelbaum MG (2008) The relevance of self-locating beliefs. Philos Rev 117:555–606CrossRefGoogle Scholar
  43. van Fraassen B (1976) Probabilities of conditionals. In: Harper WL, Hooker CA (eds) Foundations of probability theory, statistical inference and statistical theories of science, vol I. Reidel, Dordrecht, pp 261–301Google Scholar
  44. van Fraassen B (1984) Belief and the Will. J Philos 81:235–256CrossRefGoogle Scholar
  45. van Fraassen B (1989) Laws and symmetry. Clarendon Press, OxfordCrossRefGoogle Scholar

Copyright information

© Springer Science+Business Media B.V. 2011

Authors and Affiliations

  1. 1.Research School of Social SciencesAustralian National UniversityCanberraAustralia

Personalised recommendations