1 Introduction

The precautionary principle (PP) has played an important role in both environmental law and chemical regulation over the last decades.Footnote 1 More recently, the principle has been appealed to—implicitly and explicitly—when deciding how to respond to the Covid-19 pandemic. For instance, both the imposition of restrictions on movement between EU countries and rules about mandatory quarantines have been interpreted as applications of the principle (see, respectively, Goldner Lang (2021) and Raposo (2022)).

Although widely cited and purportedly applied, there is no agreement on the precise formulation of the precautionary principle.Footnote 2 But one influential formulation of the principle can be found in the Rio Declaration, principle 15 of which states that:

In order to protect the environment, the precautionary approach shall be widely applied by States according to their capabilities. Where there are threats of serious or irreversible damage, lack of full scientific certainty shall not be used as a reason for postponing cost-effective measures to prevent environmental degradation.

The above is however far from being a clear and action guiding decision rule. In fact, it has been argued (Peterson, 2006; Stefánsson, 2019) that there is no coherent formulation of PP as a decision rule. For instance, suppose that one accepts a seemingly weak Archimedean (trade-off) condition:

Weak Archimedes

Some (nonnegligible) increase in the relative likelihood of a catastrophic outcome can be at least counterbalanced by the relative likelihood of a non-catastrophic outcome being increased in relation to a strictly worse non-catastrophic outcome.

Then PP understood as a decision-rule either violates transitivity,Footnote 3 or gives conflicting recommendations depending on whether we apply it ‘globally’ or ‘locally’,Footnote 4 argues Stefánsson (2019) (based on work by Peterson (2006)).

Bartha and DesRoches (2021) and Steel and Bartha (2023) have recently responded to such arguments by suggesting that we should understand PP as the injunction to maximise lexical expected utility.Footnote 5 They defend the lexical utility model of the precautionary principle as recommending non-contradictory choice(s) in any decision scenario—thus avoiding the charge of incoherence—and, relatedly, as satisfying all the standard axioms of expected utility theory except for continuity.Footnote 6\(^,\)Footnote 7 By allowing for violations of continuity specifically when it comes to gambles that may result in catastrophic outcomes, the lexical utility model does not satisfy the above Archimedean condition.

The key concept for the lexical utility decision rule—and, in fact, for the precautionary principle too—is a harm threshold, which intuitively distinguishes catastrophic outcomes from non-catastrophic outcomes. Beyond that distinction, no fixed interpretation is given to this threshold (Bartha and DesRoches 2021, 8704, 8720). In the simple two dimensional lexical utility model on which Bartha, DesRoches, and Steel mostly focus, the harm threshold is accounted for by using two independent utility functions, \(u_1\) and \(u_2\). The first function takes the value \(-\,1\) if the outcome to which it is applied is catastrophic, but otherwise it takes the value 0. The latter is a standard utility function, summarising the outcome’s more ‘ordinary’ (i.e., non-catastrophic) costs and benefits.

The recommendation of the lexical utility decision rule is always to maximise lexical expected utility. To state this more precisely, let’s suppose that there are n states of the world, \(s_1, \ldots , s_n\), representing the uncertainty of the decision situation.Footnote 8 Let \(p(s_i\mid A)\) be the probability of state \(s_i\) when alternative A is chosen while \(u_j(A \& s_i)\) is the utility (according to function j) of choosing alternative A when state \(s_i\) obtains. The lexical utility decision rule is based on the expected utilities (EUs) that are generated by both \(u_1\) and \(u_2\):

$$\begin{aligned}{} & {} EU_1(A)=\sum _{i=1}^n u_1(A \& s_i)p(s_i\mid A)\\{} & {} EU_2(A)=\sum _{i=1}^n u_2(A \& s_i)p(s_i\mid A) \end{aligned}$$

Finally, the lexical utility decision rule says that, for any alternatives A and B, the latter should be chosen over the former just in case either

$$\begin{aligned}{} & {} EU_1(A)<EU_1(B), \\{} & {} \quad \ \ \text {or}\\{} & {} EU_1(A)=EU_1(B) \ \ \text {and} \ \ EU_2(A)<EU_2(B) \end{aligned}$$

Note that, since \(u_1\) takes the value of either 0 or \(-\,1\)—the catastrophe either occurs or notFootnote 9\(EU_1(A)\) is simply the negative of the probability that A results in a catastrophe. So, informally, the lexical utility decision rule says that for any two alternatives A and B, you should choose B over A in case either:

  • B offers a lower probability of the catastrophe, or

  • the probability of the catastrophe is the same given either alternative, but B offers a higher expected utility according to \(u_2\).

The lexical utility decision rule thus gives absolute priority to minimising the long-run (Steel and Bartha 2023: 265–266) risk of catastrophe. If B offers an all-things-considered lower probability of catastrophe than A, then B should be chosen rather than A irrespective of how the two alternatives otherwise compare.

Is this type of absolute (lexical) priority plausible? For instance, is it plausible that, say, no aggregate welfare benefits can ever outweigh any chance of a public health catastrophe? I find it very hard to believe that the true (or correct) prospect axiology could have such a lexical structure. However, Bartha and DesRoches (2021) offer a pragmatic defence of the lexical utility model (based on remarks by Christiansen (2019)), which may show that the use of the lexical utility decision rule could be justified even if the true axiology does not have this lexical structure.

First: we may face a high-stakes decision where the relevant utilities and probabilities are not sufficiently determinate to settle the issue of continuity, and not sufficiently determinate to allow a decision using standard decision theory. Second: it may be that our preferences are locally discontinuous, given the realistic range of utilities and probabilities. [...] In both sorts of case, the need for action can justify a model that represents our preferences as discontinuous and, in these cases, lexical utility maximization provides a framework for articulating a simple and straightforward rationale for precautionary measures. (Bartha and DesRoches 2021, 8728)

Similarly, Steel and Bartha (2023) point out that a lexical utility model may—even if theoretically not quite right—provide a “sufficiently good approximation and likely to produce better results in practice than a model wherein utilities are continuous and the probabilities must be specified with precision” (262).

Although perhaps pragmatically appealing, the lexical utility model faces some theoretical difficulties (as Bartha, DesRoches, and Steel are of course awareFootnote 10). In this paper I raise a new theoretical challenge for any attempt at using a lexical utility model to formalise the precautionary principle. The basic problem is that such a version of the precautionary principle faces a trilemma—in fact, it faces two versions of this same trilemma for two independent reasons:

  • Marginal value differences (in fact, the smallest value difference possible, given the context) result in extreme axiological differences,Footnote 11 or

  • a vague threshold is used in a way that either implies a choice cycle, or

  • leads to paralysis, that is, a situation where no choice is permissible.

Here is an illustration of the first horn of the trilemma. Suppose we think that not imposing ‘lockdown’ in response to the spread of a respiratory virus could result in at most n patients needing intensive care within a particular time period. Then some economic gain would justify avoiding lockdown, according to the lexical rule. However, if we then come to think that there is any chance, no matter how small, that not locking down could result in \(n+1\) patients needing intensive care, then no economic gain could justify avoiding lockdown (because \(n+1\) patients needing intensive care would be catastrophic).

In the lockdown example, the second horn of the dilemma arises if we decide that the threshold for what counts as a catastrophic number of patients in intensive care, which no economic gain could offset, is vague. As we shall see in the next section, this can lead to choice cycles. That seems particularly problematic given, first, that the lexical utility model of the precautionary principle was partly formulated as a response to the charge that any decision-rule version of the principle is incoherent. It is hard to see that a decision rule is coherent if it implies that one should choose A over B, B over C, and C over A. Second, the lexical utility model of the precautionary principle was advertised as a decision rule that satisfies transitivity,Footnote 12 which means that it should not permit cyclical choices.

An instance of the third horn of the dilemma would be that both imposing and not imposing lockdown is impermissible. Similarly to the second horn, this horn seems particularly problematic in light of the fact that Bartha and DesRoches (2021: 8731–8732) argue that their lexical utility model of PP does not lead to the kind of paralysis that for instance Sunstein (2005) criticises PP for. However, if Bartha and DesRoches’ version of the principle is to avoid the first two horns of the trilemma, then it would seem they have to insist that there may be situations where their rule implies that none of the available options can permissibly be chosen.

The lexical formulation of the precautionary principle faces two versions of the above trilemma. In the next section I formulate and discuss the first version, in the section after I formulate and discuss the second version.

Before considering the trilemma in more detail, it may however be worth saying a few words about the scope—or, rather, the limit—of my analysis. As previously mentioned, my aim is to identify a new theoretical problem for the lexical utility model. In practice, someone applying the lexical utility model might nevertheless do better, even in some extremely important decision-contexts, than someone who applies (or tries to apply) the expected utility model or standard cost-benefit analysis. As Bartha and DesRoches (2021) point out (e.g., pp. 8728–8729), there are some decision-problems—in particular, problems involving potential catastrophes and limited information about the relevant probabilities—where trying to figure out the exact probabilities and utilities, in order to calculate expected utility, would be futile. Similarly, “computational convenience” (Bartha and DesRoches 2021: 8723) in many cases speaks in favour of the lexical model over the expected utility model. So, despite the theoretical problems I identify, we may well be justified in often using the lexical utility model. Still, this paper highlights some potential costs of doing so; and, similarly, identifies situations in which one must proceed carefully with the lexical utility model.

Moreover, it is possible that the lexical utility model is a faithful formalisation of PP despite the theoretical problems I identify. Perhaps the above mentioned trilemma simply is an inherent feature of the principle. And, in fact, “descriptive accuracy” is the main defence that Bartha and DesRoches (2021) offer in favour of their lexical utility model: “the lexical approach faithfully models precautionary reasoning in its handling of actual examples” (8704). If that is so, then my paper can be read as highlighting the theoretical costs of PP and as identifying situations in which one must proceed carefully when applying the principle.

2 A threshold for graded harms

In some applications of the precautionary principle, the harm threshold may reflect something that is truly binary, say, human extinction, a climatic tipping point, or an ecosystem collapse. In those cases it may be that a miss is as good as a mile. For instance, the difference between, on the one hand, humanity going extinct, and, on the other hand, humanity almost going extinct may not be a difference in degree; it may be a difference in kind. So, it may make sense to treat human extinction as being categorically different from “almost extinction”.Footnote 13

The above is clearly not true of all applications of the precautionary principle. Sometimes the harm threshold reflects some cut-off point in a variable that, if not continuous, at least comes in degree. An example is the application of the principle when evaluating a public health policy, such as a strategy against the coronavirus (which Steel and Bartha (2023) in fact use to illustrate their model). In such applications, the harm threshold will typically be something like: “x number of patients need intensive care” (within some particular time period), or perhaps “x people die from the coronavirus disease”. I shall assume the former in what follows.

Any such cut-off raises the question: why this number? Why not \(x+1\)? Any choice may seem arbitrary.Footnote 14 That is not, however, the problem I shall be concerned with. Instead, the problem I shall focus on is that, if one takes the lexical utility model seriously, we face the problem that marginal value differences result in extreme axiological differences.Footnote 15 Recall the lockdown example from the introduction; let me spell it out a little.

Suppose that, early in the Covid-19 pandemic, the government has gathered their economic and health policy advisors to try to decide if it should lock down the economy. To simplify greatly, suppose that all the economic advisors agree that the expected economic cost of locking down—or, to put it differently, the expected economic gain from not locking down—is X billion dollars. All the public health experts in the room agree that not locking down would at most result in \(x-1\) additional patients needing intensive care within the relevant period . As the government is about to decide that, in that case, the gain from not locking down is worth the cost, a fringe health expert enters the room, late as usual. He tends to be controversial, just for the sake of it, and the government knows this. The problem is that he expresses the opinion that the result of not locking down will be that x patients need intensive care within the relevant period, which happens to be the number deemed catastrophic. Everyone agrees that this maverick is the least trustworthy of all the health experts. Still, there is some chance that he is right, so the government feels it must take his opinion into account. But then the lexical utility model implies that no possible economic gain could justify not locking down.

The reason why I think the above implication is implausibly extreme, is that the only way to make sense of the shift that happens when the fringe expert expresses his opinion, would be to think of the difference between x patients needing intensive care and \(x-1\) patients needing intensive care as being extremely important.Footnote 16 If there is some economic benefit that could offset risking \(x-1\) patients needing intensive care, but no economic benefit that could offset risking x patients needing intensive care, then that seems to suggest that the difference between x patients needing intensive care and \(x-1\) patients needing intensive care is extremely important. How could we otherwise justify treating these two outcomes so differently, for instance, being willing to risk \(x-1\) patients needing intensive care if the potential economic benefit is sufficiently hight while not being willing to risk x patients needing intensive care no matter the potential economic benefit? But, from a social point of view—which surely is the relevant point of view when evaluating public health policies—the evaluative difference between x patients needing intensive care and \(x-1\) patients needing intensive care is trivial. So, the lexical utility formulation of the precautionary principle seems to have implausible implications, when applied to public health.Footnote 17

In response, the defender of the lexical utility model might argue that, actually, the evaluative difference between x patients needing intensive care and \(x-1\) patients needing intensive care is not trivial, even from the social point of view. After all, it is the difference between a catastrophic outcome and a non-catastrophic one!Footnote 18 Although this is certainly a possible response, I suspect that defenders of the lexical utility model would find it too counterintuitive. It would at least require a convincing explanation of why x is so important.Footnote 19

The only plausible reason that I can think of for why x could be so important, is if \(x-1\) is the number of patients that can be treated by society’s intensive care units (within the relevant period). Then the difference between \(x-1\) and x patients needing intensive care may not be quite trivial, even from a social point of view. After all, that means that some very difficult (and tragic) choices will have to be made about to whom to provide intensive care. Still, it is hard to see that, even in this case, the difference between \(x-1\) and x patients needing intensive care is sufficiently important to justify sometimes being willing to risk the former while never being willing to risk the latter. After all, even if the intensive care capacity is \(x-1\) patients, the worse outcome only means that a single patient will not be granted needed intensive care. And although that may be tragic, it is hard to see that it could justify treating these gambles so extremely differently. From the point of view of each patient, the difference between the two outcomes is that in the latter outcome, each patient faces an average \(\frac{1}{x}\) probability of not getting the needed intensive care. But recall that \(x-1\) is the number of patients that can receive intensive care (within some time period). For any modern society,Footnote 20\(\frac{1}{x}\) is a very small probability (no matter how small the aforementioned time period). So, we still find that the lexical utility model of the precautionary principle seems to have implausible implications, when applied to public health.

Alternatively, the defender of the lexical utility model might respond that if indeed the evaluative difference between x patients needing intensive care and \(x-1\) patients needing intensive care is trivial, from a social point of view, then if \(x-1\) patients needing intensive care is not a catastrophe, we were simply mistaken in treating x as a catastrophe.Footnote 21 The problem, of course, is that to avoid the above result—without claiming that each patient in intensive care is a catastrophe—they would have to say in addition that for any integer y, y patients needing intensive care is not a catastrophe if \(y-1\) patients needing intensive care is not a catastrophe. And that assumption also leads to trouble, as we shall see in a moment.

Suppose that despite the above problem, we want to maintain that the lexical utility framework is a good model for the precautionary principle in cases where the relevant harm threshold concerns a variable that comes in degree. (In the next section I consider what might be truly “binary” catastrophes.) What are then our options, if we still want to avoid extreme results like that above? I think that our only feasible option is to say that the harm threshold is vague.Footnote 22

Let’s assume, then, that the threshold is some vague or fuzzy range that includes x. In that case, the fact that the maverick health expert expresses his opinion that the number of patients needing intensive care will be x rather than \(x-1\), need not make such an extreme a difference. If the harm threshold is vague, and if \(x-1\) patients fall below the threshold, then plausibly x patients will not be (determinately) above the threshold.

Unfortunately we are not free from trouble yet. Suppose there is a second fringe health expert, who also arrives late and expresses her opinion that, if the government doesn’t lock down, then that will result in \(x+1\) patients needing intensive care. And so on, imagining a third, fourth, etc., fringe health experts, respectively expressing their opinion that not locking down will result in the number of patients needing intensive care being \(x+2\), \(x+3\), etc. To avoid the above extreme result, we might be tempted to make the assumption that for any integer y, if there is some potential economic gain Y that can compensate for y patients needing intensive care, then there is some potential economic gain (greater than Y) that can compensate for \(y+1\) patients needing intensive care. More generally, if y patients needing intensive care is (determinately) not a catastrophe, then \(y+1\) patients needing intensive care is also not a catastrophe. And that may seem very natural, given the stipulation that the threshold is vague. (Compare: if the threshold for being bald is vague, then it would be natural to say that, for any integer y, if having y hairs means that a person is not bald, then having \(y+1\) hairs means that a person is not bald.)Footnote 23

But it is not hard to see that a problem is looming. Even if the threshold is vague, there must surely be some integer z such that, even though there is some potential economic gain that can compensate for y patients needing intensive care, no potential economic gain can compensate for \(y+z\) patients needing intensive care. Otherwise the lexical utility model would be trivialised, and we would be back to standard cost-benefit analysis, which the precautionary principle (and the lexical utility model) is meant to be a genuine alternative to.

The above means that we face a Sorites paradox. For any number y of patients needing intensive care that does not constitute a catastrophic outcome, no single additional patient in intensive care makes the outcome catastrophic; still, there is some number of patients in intensive care that constitutes a catastrophe. While puzzling, these types of paradoxes are very familiar, and not what I shall be focusing on. Instead, I shall focus on how this puzzling feature of vague thresholds can result in choice cycles, when choices are guided by the lexical utility model.

As an illustration of how such a cycle can arise, consider the following (admittedly somewhat unrealistic) example. Suppose that policy 0 is predicted to result in at most y patients needing intensive care, which is determinately non-catastrophic, and is also expected to result in economic gain of magnitude Y. Policy 1 however is expected to result in additional economic gain of magnitude \(Y^1\) (so, gain \(Y+Y^1\) in total) and is predicted to result in at most one additional patient needing intensive care (so, \(y+1\) patients in total). By previous assumptions, \(y+1\) patients needing intensive care is not catastrophic. So, there must be some \(Y^1\) for which policy 1 should chosen over policy 0.

We can continue this reasoning. Ultimately we then reach a policy z, the implementation of which means that the total expected economic gain is \(Y+Y^1+\cdots +Y^z\) and the number of patients needing intensive care is predicted to be at most \(y+z\). The assumption that the harm/catastrophe threshold is vague seems to imply that for each i, policy \(i+1\) should be chosen over policy i for some additional economic gain \(Y^{i+1}\). But suppose that \(y+z\) patients needing intensive care is a determinately catastrophic outcome. In contrast, we assumed that y patients in intensive care was determinately not a catastrophic outcome. Then there is no economic gain that offsets risking \(y+z\) patients needing intensive care rather than y patients. So, policy 0 should be chosen over policy z, for any economic gain \(Y+Y^1+\cdots +Y^z\). In other words, we have a choice cycle.

There may be ways of avoiding such choice cycles, without ending up back at the first horn of the trilemma. Perhaps there is an integer j such that, although \(y+j\) patients needing intensive care is determinately not catastrophic, it is indeterminate whether \(y+j+1\) patients needing intensive care is catastrophic. And if this seems too close to the first horn of the trilemma, we can appeal to second order vagueness and say that it is indeterminate whether this shift happens at integer j rather than some other integer in a (fuzzy) interval around j (cf. Andersson (2022)). The implication would then be that for some integer(s) j, it is indeterminate whether some economic gain justifies choosing policy \(j+1\) over policy j (in the sequence from the last paragraph). Moreover, where this indeterminacy starts is indeterminate.

Now, whether the above move prevents choice cycles depends on what one ought to do when what one ought to choose is indeterminate. But it seems we should say one of the following.Footnote 24 First, we might say that if it is indeterminate whether option A should be chosen over B, then it is permissible to choose either option when only these are available. It is not hard to see that this would permit choice cycles (e.g., by the argument from the last section).Footnote 25 Second, we could say that if it is indeterminate whether option A should be chosen over B, then it is indeterminate whether it is permissible to choose either option when only these are available. But that just raises the question of how one should choose when what is permissible is indeterminate, to which analogous answers to those considered here would seem to be the only plausible options. Third, we might say that if it is indeterminate whether option A should be chosen over B, then one is permitted to choose neither. But that would constitute a particularly unwelcome instance of incompleteness,Footnote 26 and in fact paralysis, when A and B are the only options.

In other words, we can construct a sequence of options such that the lexical utility model, with a vague harm threshold, implies one of the following:

  1. 1.

    The model permits cyclical choices, or

  2. 2.

    the model leads to paralysis.

Without a vague threshold, however, the lexical utility model implies that trivial value differences result in extreme axiological differences. Thus, the trilemma.

3 A probabilistic threshold

I now turn to a second reason for why the lexical utility decision rule leads to a trilemma structurally very similar to the one raised in the last section. It might be worth acknowledging right away that the trilemma discussed in this section may be less troubling—and, perhaps, less interesting—than the trilemma discussed in the last section. The trilemma discussed in the last section arises because, first, lexical utilities were meant to model the precautionary principle, and, second, the principle is often applied in contexts where the harm threshold is a cut-off point in some graded variable. In contrast, the trilemma discussed below derives directly from the lexical utility model, irrespective of the uses to which it is put. Therefore, it should not come as a surprise to those who are attracted to the lexical utility model. Nevertheless, since the trilemma in this section is structurally almost identical to the previous trilemma, I hope it is worth discussing.

Let us now focus on catastrophes, and corresponding harm thresholds, that are what Steel and Bartha (2023: 264) call terminal, that is, an outcome such that “once [it] occurs, it makes no difference what happens afterward”. Maybe human extinction is the most plausible example of such a catastrophe. So, it may be quite right—not just in practice but even in theory—to treat human extinction as being categorically different from any outcome that is close to but not quite extinction.

The lexical utility decision rule then implies that if we are evaluating two alternatives, and one has any chance—no matter how small—of resulting in human extinction whereas the other has no chance of resulting in human extinction, then we should choose the latter, irrespective of other costs and benefits. But note that this is a version of the first horn of the trilemma discussed in the last section: marginal value differenceFootnote 27 results in extreme axiological difference.

To take an example, recall that before CERN’s test run of the Large Hadron Collider in 2008, concerns were raised that this might create a black hole that could mean the end of humanity.Footnote 28 Law suits were even filed, including to the European Court of Human Rights, to stop the project. Courts dismissed the claims, and CERN’s own safety review concluded:

[Large Hadron Collider] collisions present no danger and [...] there are no reasons for concern. Whatever the [Large Hadron Collider] will do, Nature has already done many times over during the lifetime of the Earth and other astronomical bodies.Footnote 29

CERN and the courts were no doubt right in concluding that the experiment was justifiable. But could they say with absolute certainty that there was no chance that critics, such as the German biochemistry professor Otto Rössler—one of the more vocal critics of the experiment—were mistaken? That is, were they justifiably 100% certain that the experiment would not create a black hole? I think they were not. In fact, it seems to me that the arguments by Otto Rössler and others, while almost certainly mistaken, should have had (and maybe did have) some, but perhaps very small, positive impact on the degree to which it was believed possible that the experiment would create a black hole. But, if we take the lexical utility decision rule seriously, it then follows that the experiment should have been abandoned, irrespective of how small impact the argument did or should have on the relevant agents’ degrees of belief (i.e., on their subjective probabilities) and irrespective of how beneficial they predicted the experiment would be. So, we get the first horn of the trilemma again: marginal value difference results in extreme axiological difference.

Now, the courts and CERN may well have been justifiably confident that the catastrophic risk in question was what in legal contexts is often called de minimis (Adler, 2007; Peterson, 2002), or small enough to be ignored. The lexical utility model does not contain such a de minimis threshold,Footnote 30 but it of course could, in which case we would not get the troubling results from the previous example. However, the exact same problem would re-appear. Suppose that a second dissenting professor had come to the same conclusion as Otto Rössler. Further suppose that the testimony of this (or some subsequent, individiual) dissenting scientist moved CERN’s and the judges’ degrees of belief that the experiment would result in a black hole very slightly above the de minimis threshold. For instance, let’s say a catastrophic risk is de minimis when the probability of the catastrophe is no greater than \(\delta\), and suppose that, before the second professor made their argument, the agents in question believed to degree \(\gamma \le \delta\) that the experiment would result in a black hole, while after the professor had made their argument, the agents in question believed to degree \(\epsilon >\delta\) that the experiment would result in a black hole. Then if the agents applied a lexical utility model, they would have had to abandon the experiment, irrespective of how beneficial they predicted the experiment would be and no matter how close \(\epsilon\) is to \(\gamma\). So, yet again, we get the first horn of the trilemma: Marginal value difference results in extreme axiological difference.Footnote 31

Perhaps Bartha, DesRoches, and Steel would simply bite the bullet and accept the above implication of their model (in fact, I suspect they would, as I alluded to at the beginning of this section). However, if we want to avoid this problem, while maintaining the lexical utility model, then the obvious solution would seem to be to assume a vague de minimis threshold. In other words, we might stipulate some vague or fuzzy probability interval such that if the probability that some alternative results in a catastrophe determinately falls below that interval, then that catastrophic risk is de minimis and may be ignored,Footnote 32 but not if the probability is determinately above the interval.

The problem is that while this move avoids the first horn of the trilemma, it either leads to paralysis or permits choice cycles. Imagine that a sequence of seemingly independent fringe scientists come forward with support of professor Rössler’s argument, one after the other. For each scientist, CERN and the courts revise upwards their degrees of belief that the experiment could result in a black hole by some tiny magnitude \(r^-\). To avoid the first horn of the trilemma, we might want to say that if a risk is de minimis when the relevant probability is \(\zeta\), then so is a risk when the probability is \(\zeta +r^-\). So, then if the courts and CERN though that the risk of the experiment turning out catastrophic was de minimis when the \(n{\text {th}}\) professor expressed their agreement with Rössler’s argument, then they should also think that the risk was de minimis when the \((n+1){\text {th}}\) professor expressed their agreement with Rössler’s argument. But, obviously, given a long enough sequence of such professors, we eventually reach a sufficiently high probability such that the risk is determinately not de minimis. So, again we face a Sorites paradox.

The explanation for why we in addition either face choice cycles or paralysis is structurally very similar to the argument in the last section. If a risk that an alternative results in a catastrophe is de minimis, then the alternative should—according to the lexical utility decision rule with a de minimis threshold—be judged on its non-catastrophic costs and benefits. Let a risk with probability \(\gamma\) be determinately de minimis. Now imagine that God makes the government the following offer: she will raise the GDP by $X, but the cost is that the chance of human extinction will increase to \(\gamma\). Using the lexical utility model with a de minimis threshold, the government accepts the offer. Now God offers to instead raise the GDP by $\(X^+>X\), which comes with additional \(\gamma +r^-\) chance of human extinction. Using the lexical utility model, the government finds this option to be even better than God’s previous offer.

The problem, of course, is that God can continue this sequence, repeatedly offering the government options that seem better to them than the previous option, but where eventually the risk of human extinction is determinately not de minimis. So, if we are to avoid the first horn of the trilemma—i.e., to avoid marginal differences having extreme axiological differences—when it comes to any pair of adjacent offers by God, then it seems we will (by reasoning analogous to that in the last section) have to accept either a choice cycle or that no choice is permissible between some adjacent options in the sequence of options offered by God. Thus, yet again, the trilemma:

  • Marginal value difference results in extreme axiological difference, or

  • a vague threshold is used in a way that either implies a choice cycle, or

  • leads to paralysis, that is, a situation where no choice is permissible.

Now, as alluded to in the introduction, it might well be that (something like) the above trilemma is really what we should expect when applying the precautionary principle to such an extreme example as the Large Hadron Collider. In that case, Bartha and his colleagues are correct when they claim that the lexical utility model is a faithful representation of the principle. And this section then illustrates problems that can arise when applying PP, but that are not unique to the lexical utility model of the principle.

Before concluding this section, I should address the objection that all options could result in human extinction; hence, the fact that one option has some chance of resulting in extinction need not imply that it should be ruled out, according to the lexical utility model. However, the problem is that the lexical utility model implies that, if the two available options can both result in whatever is designated as a (terminal) catastrophe, then the option that has a lower chance of resulting in that catastrophe should be chosen.Footnote 33 This means that we are back in the same trilemma. Either a tiny difference in the risk of a catastrophe has extreme axiological effect; or we impose a vague de minimis condition—applied to either differences in probabilities or absolute probabilities—in a way that leads to choice cycles or paralysis.

4 Concluding remarks

The lexical utility decision rule appears to face a trilemma. Either it is extremely sensitive to trivial value differences, or else it leads to choice cycles or paralysis. But, to conclude, I would like to revisit the point I made in the introduction about the problems raised in this paper being merely theoretical. To derive the problematic results, I had to make some presumably unrealistic assumptions, for instance, about the exactness of the predicted value of the variable underlying the harm threshold, and about the types of alternatives with which a policy maker can be faced. So, despite the arguments in this paper, Bartha, DesRoches, and Steel may well be correct that the lexical utility model leads to better results than its alternatives in many practically important situations, in particular given the resource- and time-constraints of ordinary decision-makers. As previously mentioned, the standard expected utility model may for instance require information that is not available and computations that are infeasible. Still, the results in this paper highlight some theoretical costs of the lexical utility model, and identifies some situations where its use requires extra caution.