1 Introduction

In many real-life situations, communication relies on hard evidence. For example, a jury’s verdict is based on hard evidence presented in the court, rather than exchanges of empty claims. Evidence games study such situations. There is a sender (e.g., a prosecutor), and a receiver (e.g., a jury). The sender has private hard evidence about an unknown state of the world (e.g., whether a defendant is guilty) that she can selectively present to the receiver, and the receiver takes an action (e.g., conviction or acquittal) that is payoff relevant to both players. Full revelation of evidence is often impossible in the presence of conflict of interest between the sender and the receiver–the receiver wants to learn the payoff relevant state and act accordingly, whereas the sender merely wants to induce her preferred receiver action (e.g., convicting the defendant). Therefore, the sender has an incentive to persuade the receiver that a certain state is more likely by partially revealing evidence.

Formally, verifiability of hard evidence is modeled by assuming that the sender’s feasible set of disclosure depends on her type (i.e., her evidence), and in this paper, we assume that the sender’s type space is ordered.Footnote 1 That is, some sender types have more evidence than others, and the sender can disclose less evidence than she has, hence the feasible set of disclosure is the lower contour set of her type under the “more evidence” order. Moreover, the sender’s payoff depends only on the receiver’s action and not her type or the state of the world. For example, the prosecutor’s objective is to convict the defendant. This is not affected by what evidence she has. In equilibrium, her chance of convicting the defendant may depend on the evidence she has, because when she has more evidence, there are more ways to present evidence in the court, and thereby she can better persuade the jury.

Evidence games often have multiple (Nash) equilibria. For instance, there is a trivial equilibrium where, regardless of her evidence, the prosecutor presents no evidence to the court, and the jury acquits the defendant regardless of what is presented (this must be optimal on the equilibrium path for the jury if the presumption of innocence is practiced). This is undoubtedly not a sensible prediction of what happens in courtrooms. However, this equilibrium is both perfect (Selten 1975) and sequential (Kreps and Wilson 1982) under mild assumptions.Footnote 2 Consider a perturbation to the prosecutor’s strategy that assigns higher probability on disclosing evidence that supports conviction than on disclosing evidence that supports acquittal, and a perturbation to the jury’s strategy such that the probability of convicting the defendant after seeing any evidence is no larger than that after seeing no evidence. As both perturbations diminish, this gives a convergent sequence of \(\varepsilon\)-constrained equilibria in completely mixed strategies. Therefore, the trivial equilibrium is perfect. Similarly, given proper perturbations in the prosecutor’s strategy, it is consistent for the jury to hold the belief that, after seeing any evidence, the actual evidence possessed by the prosecutor favors acquitting the defendant. Therefore, this trivial equilibrium is also a sequential equilibrium.

There are also extensive discussions on the value of commitment power in evidence games. That is, whether the receiver can achieve a higher payoff by committing ex ante to a mapping from the sender’s disclosure to a distribution over his actions. Glazer and Rubinstein (2006) show that there is no value of commitment when the receiver’s action is a binary one; Sher (2011) shows the same result when the receiver’s payoff is concave in his action, and the receiver’s actions can be either finite or continuous.

Hart et al. (2017) (henceforth HKP) generalize the condition of concavity in Sher (2011) for the case of continuous receiver actions. They focus on the receiver’s commitment to a deterministic mechanism. That is, the receiver commits to an action for each possible disclosure, and he cannot randomize over his actions.Footnote 3 They show that committing to a deterministic mechanism has no value if the receiver chooses an action on the real line, the receiver’s expected payoff is a single-peaked function of his action given any distribution of the state, and the sender strictly prefers higher receiver action. Moreover, HKP propose the following equilibrium refinement in evidence games. A truth-leaning equilibrium is a perfect Bayesian equilibrium such thatFootnote 4

(Truth-leaning):

Given the receiver’s strategy, the sender discloses her evidence truthfully if doing so is optimal;

(Off-path beliefs):

The receiver takes any off-path disclosure at face value (i.e., he believes that the sender discloses truthfully).

As is argued in HKP, these conditions follow the straightforward intuition that there is a “slight inherent advantage” for the sender to tell the whole truth, and “there must be good reasons for not telling it.” Under the assumption that the receiver takes a continuous action, HKP show that a truth-leaning equilibrium exists and is receiver optimal. That is, it gives the receiver the same ex ante payoffs as the optimal deterministic mechanism.

However, in many applications of evidence games, the receiver takes a discrete action. For example, juries choose between conviction and acquittal, banks decide whether or not to grant a loan, and rating agencies rate financial assets into finitely many grades. A part of this paper is to answer the following question: is truth-leaning equilibrium a “good” solution concept when the receiver’s action set is finite?

The short answer is “no,” and one reason is that a truth-leaning equilibrium may fail to exist. Loosely speaking, nonexistence arises because the sender lacks a strict incentive to persuade the receiver.Footnote 5 Truth-leaning equilibrium also ignores that players often face small payoff uncertainties in evidence games. It is an idealization to assume, for example, that a prosecutor knows perfectly a jury’s criteria when making their verdict. Solution concepts that ignore this may lead to unrealistic predictions.

To address these problems, we propose the following solution concept by introducing a small uncertainty (i.e., disturbance) to the receiver’s payoff à la (Harsanyi 1973).Footnote 6 Suppose that the receiver receives a random private payoff shock associated with each of his actions. In this disturbed game, the sender has a strict incentive to persuade the receiver, and a truth-leaning equilibrium exists.Footnote 7 We define a purifiable truthful equilibrium as the limit of a sequence of truth-leaning equilibria in the disturbed games as the disturbances converge to zero. That is, a purifiable truthful equilibrium is a truth-leaning equilibrium of an infinitesimally disturbed game. A purifiable truthful equilibrium always exists and is a perfect Bayesian equilibrium.

Purifiable truthful equilibria are also receiver optimal. That is, purifiable truthful equilibria maximize the receiver’s ex ante payoff among all perfect Bayesian equilibria. And the receiver’s purifiable truthful equilibrium payoff is the same as his payoff in the optimal deterministic mechanism. The receiver, however, may achieve a higher payoff than his purifiable truthful equilibrium payoff by committing to a stochastic mechanism. This is akin to the results in HKP.

Another problem of truth-leaning equilibrium in finite evidence games is that it may not follow the intuition that the sender is slightly more advantageous if she discloses truthfully. To formalize this intuition, we revisit the perturbed game in HKP, where the sender receives a small reward if she discloses truthfully, and the sender must disclose truthfully with at least some small probability. We define a weakly truth-leaning equilibrium as the limit of a sequence of perfect Bayesian equilibria of the perturbed games as the perturbations converge to zero.Footnote 8 HKP show that truth-leaning equilibrium is equivalent to weakly truth-leaning equilibrium. When the receiver’s action space is finite, however, this equivalence is no longer true. It turns out that purifiability is the missing connection: a weakly truth-leaning equilibrium that is also purifiable truthful is a truth-leaning equilibrium; a truth-leaning equilibrium that is also purifiable truthful is a weakly truth-leaning equilibrium in “almost all” (in a precise sense, see Proposition 7) evidence games.

The paper proceeds as follows. Section 2 presents a simple example where truth-leaning refinement leads to nonexistence of equilibrium and discusses some other limitations of the existing refinements. We construct the purifiable truthful equilibrium of this example. Section 3 models evidence games. Section 4 studies purifiable truthful equilibrium and compares various equilibrium refinements of evidence games. The last section concludes. Proofs are in the Appendix.

2 A simple example

Every new aircraft design has to be certified by the Federal Aviation Administration (FAA) before any aircraft built according to this design can enter service. Like other innovations, altering the design of an aircraft often entails high level of risks. The FAA often has to rely on information and test results provided by airplane manufacturers, yet airplane manufactures’ disclosure is far from complete.Footnote 9

Imagine an airplane manufacturer (the sender) seeking to get a new aircraft design certified by the FAA (the receiver). The design can be good or bad with equal likelihood. If the design is bad, the aircraft manufacturer has some bad evidence (e.g., mechanical failures during test flights) with probability \(\frac{2}{3}\). Otherwise, the aircraft manufacturer has no evidence. The FAA does not know the quality of the design and chooses to Approve (\(a=1\)) or Reject (\(a=0\)) the aircraft design based on evidence disclosed by the sender. The disclosure of bad evidence is voluntary and verifiable. That is, disclosing no evidence is always possible, but the airplane manufacturer can disclose bad evidence only if it has bad evidence. The airplane manufacturer’s payoff depends only on the FAA’s action: it receives 1 if the design is approved and 0 if the design is rejected. The FAA, on the other hand, gains from approving a good design and loses from approving a bad design. Its payoff is 0 if it rejects the design, 1 if it approves a good design, and -2 if it approves a bad design. Hence, the FAA has a cutoff decision rule. If, after observing the disclosed evidence, its posterior belief that the design is good exceeds \(\frac{2}{3}\), its optimal action is Approve; if its posterior belief is less than \(\frac{2}{3}\), its optimal action is Reject; if its posterior belief is exactly \(\frac{2}{3}\), either action as well as any randomization between the two actions is optimal.

A strategy of the sender describes how it discloses bad evidence. Let \(p\) be the probability that the sender discloses no evidence if it has bad evidence. Since bad evidence fully reveals that the design is bad, the receiver always chooses Reject (thus the sender gets 0) after seeing bad evidence. Let \(q\) be the probability that the receiver chooses Approve after seeing no evidence. Let \(\mu\) be the receiver’s posterior belief that the design is good after seeing no evidence. Since no evidence is disclosed with positive probability, Bayes’ rule requires that \(\mu =\frac{3}{4+2p}\).

2.1 Truth-leaning equilibrium

It is easy to verify that the game has a continuum of perfect Bayesian equilibria–any \(p\ge \frac{1}{4}\), \(q=0\), and \(\mu =\frac{3}{4+2p}\le \frac{2}{3}\) constitute an equilibrium. That is, the sender with bad evidence discloses no evidence with at least probability \(\frac{1}{4}\), and the receiver always rejects the new design.

However, there is no truth-leaning equilibrium. Given the receiver’s strategy, the sender with bad evidence is indifferent between disclosing no evidence and disclosing truthfully, since both actions yield zero payoff. Truth-leaning therefore requires the sender to disclose bad evidence truthfully (i.e., \(p=0\)), which is not satisfied by any perfect Bayesian equilibrium.

2.2 Purifiable truthful equilibrium

Suppose that the receiver receives a payoff shock \(\zeta\) for choosing Approve, where \(\zeta\) is normally distributed according to \(\mathcal {N}(0,\varepsilon ^2)\) and is private information of the receiver (hence the receiver’s type). That is, the receiver’s payoff from approving a good design is \(1+\zeta\), and that from approving a bad design is \(\zeta -2\). Given any posterior belief \(\mu\), almost all receiver types have a unique optimal action after seeing no evidence, which is Approve if \(\mu >\frac{2-\zeta }{3}\) (equivalently, \(\zeta >2-3\mu\)) and Reject if \(\mu <\frac{2-\zeta }{3}\) (equivalently, \(\zeta <2-3\mu\)). Hence, in any perfect Bayesian equilibrium of the disturbed game, the design is approved with probability \(\Phi \left( \frac{3\mu -2}{\varepsilon }\right)\) if the sender discloses no evidence, where \(\Phi\) is the cdf of the standard normal distribution. Since this probability is strictly positive for all \(\mu\), the sender strictly prefers disclosing no evidence to disclosing truthfully.

To summarize, let \(q(\zeta )\) denote the probability that the type \(\zeta\) receiver approves the design after observing no evidence. The perfect Bayesian equilibrium of the disturbed game is unique (except for the strategy of a single receiver type), where \(p=1\), \(\mu =\frac{1}{2}\), and \(q(\zeta )=0\) if \(\zeta <\frac{1}{2}\), \(q(\zeta )=1\) if \(\zeta >\frac{1}{2}\). Since the sender strictly prefers disclosing no evidence, this equilibrium is also truth-leaning. In this equilibrium, Approve is chosen with probability \(\Phi (-\frac{1}{2\varepsilon })\) after the receiver observes no evidence. That is, the disturbed game has a unique truth-leaning equilibrium outcome: the sender discloses no evidence, and after seeing no evidence, the receiver chooses Approve with probability \(\Phi (-\frac{1}{2\varepsilon })\) and believes that the design is good with \(\frac{1}{2}\) probability.

As the disturbance diminishes (i.e., as \(\varepsilon \downarrow 0\)), the unique equilibrium outcome of the disturbed game converges to a perfect Bayesian equilibrium of the original evidence game where the sender discloses no evidence, the receiver always chooses Reject, and the receiver’s posterior belief on the good design is \(\frac{1}{2}\) after seeing no evidence (i.e., \(p=1\), \(q=0\), \(\mu =\frac{1}{2}\)).

2.3 Weakly truth-leaning equilibrium

Consider the following perturbed game. Let \(\varepsilon _1\) and \(\varepsilon _2\) be small positive reals that are common knowledge to the sender and the receiver. The sender receives a reward \(\varepsilon _1\) if it discloses (bad evidence) truthfully, and the sender must disclose truthfully with at least probability \(\varepsilon _2\).

If its posterior belief \(\mu >\frac{2}{3}\), then the receiver has a unique optimal action Approve after observing no evidence. Then, for \(\varepsilon _1<1\), the sender strictly prefers disclosing no evidence, so the Bayesian consistent belief is \(\mu =\frac{1}{2}<\frac{2}{3}\). If \(\mu <\frac{2}{3}\), the receiver’s unique optimal action is Reject after observing no evidence. With the reward for truth-telling, the sender strictly prefers disclosing truthfully, so the Bayesian consistent belief is \(\mu =\frac{3}{4}>\frac{2}{3}\). Hence, the receiver’s posterior belief \(\mu =\frac{2}{3}\) in any perfect Bayesian equilibrium of the perturbed game. Indeed, for \(\varepsilon _1<1\) and \(\varepsilon _2\le \frac{3}{4}\), the perturbed game has a unique perfect Bayesian equilibrium, where \(p=\frac{1}{4}\), \(q=\varepsilon _1\), \(\mu =\frac{2}{3}\).

As \(\varepsilon _1,\varepsilon _2\downarrow 0\), the perfect Bayesian equilibrium of the perturbed game converges to a perfect Bayesian equilibrium of the original game, where \(p=\frac{1}{4}\), \(q=0\), \(\mu =\frac{2}{3}\).

2.4 Discussion

Figure 1 illustrates the equilibria of the game. There is a continuum of perfect Bayesian equilibria which differ in the sender’s strategy. Among them, the weakly truth-leaning equilibrium maximizes the probability that the sender discloses truthfully. The purifiable truthful equilibrium maximizes the receiver’s posterior belief on the good design.

Fig. 1
figure 1

The probability that the sender discloses no evidence when having bad evidence (\(p\))

The fact that this simple game does not possess a truth-leaning equilibrium suggests that truth-leaning equilibrium is not an appropriate solution concept for finite evidence games. A more fundamental problem of truth-leaning equilibrium is the discrepancy between the refinement and the intuition behind it. The requirement that the sender weakly prefers disclosing truthfully seemingly stems from the sender having an infinitesimal reward for truth-telling, but in the example, the weakly truth-leaning equilibrium constructed by adding an infinitesimal reward for truth-telling is not the same as imposing the truth-leaning refinement on perfect Bayesian equilibria.Footnote 10 The following proposition summarizes these observations. The negative result motivates the study of purifiable truthful equilibrium.

Proposition 1

In a finite evidence game, a truth-leaning equilibrium may fail to exist, and a weakly truth-leaning equilibrium may not be truth-leaning.

Weakly truth-leaning equilibrium exists in finite evidence games, but it also has several shortcomings. A distinctive feature of the weakly truth-leaning equilibrium in the above example is that the receiver is indifferent between choosing Approve and Reject after seeing no evidence, but it is prescribed to choose only Reject. This feature is prevalent and not specific to this example, and it leads to several problems. First, the equilibrium may fail to be perfect. In the example, given any mixed strategy of the receiver, disclosing no evidence is a strictly better response for the sender than disclosing bad evidence truthfully. Hence, the weakly truth-leaning equilibrium where the sender plays a mixed strategy is not a perfect equilibrium.Footnote 11

Second, weakly truth-leaning equilibrium may not be robust to incomplete receiver payoff information. As is shown above, the sender strictly prefers disclosing no evidence once we introduce a small uncertainty to the receiver’s payoff. In defense of weakly truth-leaning equilibrium, the perfect Bayesian equilibrium in every perturbed game where the sender receives a small reward for truth-telling (i.e., \(p=\frac{1}{4}\), \(q=\varepsilon _1\), \(\mu =\frac{2}{3}\)) is robust to incomplete receiver payoff information in our example,Footnote 12 but this is not a generic result. In general, a weakly truth-leaning equilibrium may fail to be the limit point of a sequence of equilibria of perturbed games that are robust to incomplete receiver payoff information.

Third, different sequences of perturbations may select different weakly truth-leaning equilibria, and not all sequences of perturbed games have a convergent sequence of perfect Bayesian equilibria as the perturbation goes to zero.

Consider a slight variant to our example, where the sender’s bad evidence is either type 1 or type 2 (think about software failures and hardware failures). If the design is bad, the sender has type 1 bad evidence, type 2 bad evidence, and no evidence each with \(\frac{1}{3}\) probability; if the design is good, the sender has no evidence. The sender with a certain type of bad evidence can disclose truthfully or no evidence but cannot disclose the other type of bad evidence. Let \(p_i\) denote the probability that the sender with type \(i\) bad evidence discloses no evidence, \(q\) the probability that the receiver chooses Approve after seeing no evidence, and \(\mu\) the receiver’s belief that the design is good after seeing no evidence. The game has a continuum of perfect Bayesian equilibria, where \(p_1+p_2\ge \frac{1}{2}\), \(q=0\), and \(\mu =\frac{3}{4+p_1+p_2}\).

Now, let us consider the following perturbed game. Given small positive reals \(\varepsilon _1,\varepsilon _2<1\) and \(\varepsilon _{1|1},\varepsilon _{2|2}\le \frac{1}{2}\), the sender receives a reward \(\varepsilon _i\) if it truthfully discloses type \(i\) bad evidence, and the sender with type \(i\) bad evidence must disclose truthfully with at least probability \(\varepsilon _{i|i}\). If \(\varepsilon _i<\varepsilon _j\), the unique perfect Bayesian equilibrium is \(p_i=\frac{1}{2}\), \(p_j=0\), \(q=\varepsilon _i\), \(\mu =\frac{2}{3}\). That is, the receiver randomizes between Approve and Reject after seeing no evidence in order to match the lower reward \(\varepsilon _i\); the sender with type \(i\) bad evidence is indifferent and randomizes between disclosing no evidence and disclosing truthfully, while the sender with type \(j\) bad evidence strictly prefers disclosing truthfully because of the higher reward \(\varepsilon _j\). If \(\varepsilon _1=\varepsilon _2\), there is a continuum of perfect Bayesian equilibria, where \(p_1+p_2=\frac{1}{2}\), \(q=\varepsilon _1=\varepsilon _2\), \(\mu =\frac{2}{3}\). Hence, as \((\varepsilon _1,\varepsilon _{1|1},\varepsilon _2,\varepsilon _{2|2})\rightarrow 0\), whether there exists a convergent sequence of perfect Bayesian equilibria depends on the rates of convergence of \(\varepsilon _1\) and \(\varepsilon _2\). If \(\varepsilon _1=\varepsilon _2\) almost always, then any perfect Bayesian equilibrium of the unperturbed game such that \(p_1+p_2=\frac{1}{2}\) is the limit point of a sequence of perfect Bayesian equilibria of the perturbed games. If \(\varepsilon _i\le \varepsilon _j\) almost always and \(\varepsilon _i<\varepsilon _j\) infinitely often, then the unique weakly truth-leaning equilibrium is \(p_i=\frac{1}{2}\), \(p_j=0\), \(q=0\), and \(\mu =\frac{2}{3}\). If neither case happens, there is no convergent sequence of perfect Bayesian equilibria of the perturbed game. In conclusion, the unperturbed game has a continuum of weakly truth-leaning equilibria, where \(p_1+p_2=\frac{1}{2}\), \(q=0\), \(\mu =\frac{2}{3}\), and different weakly truth-leaning equilibria may be selected by different sets of infinitesimal perturbations.

Purifiable truthful equilibrium is spared from similar problems. For almost all evidence games, purifiable truthful equilibria do not involve the receiver’s “borderline” beliefs, and any purifiable truthful equilibrium is infinitesimally close to a truth-leaning equilibrium of any infinitesimally disturbed game. That is, purifiability does not depend on the selection of disturbances. The normality of the receiver’s payoff shock in our example is dispensable. Moreover, the set of purifiable truthful equilibria has a simple structure, and we give a characterization of the set of purifiable truthful equilibria of any evidence game.

3 The evidence game

There are two stages. Two players, a sender (she) and a receiver (he), move sequentially. At the outset of the game, a state of the world \(\omega \in \{G,B\}\) is realized with probability \(\pi _0\in (0,1)\) on \(\omega =G\). Neither player observes the realized state \(\omega\),Footnote 13 and the prior \(\pi _0\) is common knowledge. In the first stage, the sender observes a piece of hard evidence \(e\in E\) and discloses \(m\in E\) to the receiver, where \(E\) is a finite set of evidence. In the second stage, the receiver observes the disclosed evidence \(m\) and chooses an action \(a\in A\), where \(A=\{a_1<a_2<\dots <a_K\}\) is a finite subset of the real line with \(K\ge 2\).

3.1 Evidence and disclosure

Let \(F_G\) and \(F_B\) be two distributions over the set of evidence \(E\). The sender’s evidence \(e\) is a random draw from either \(F_G\) or \(F_B\), depending on the realized state. If \(\omega =G\), \(e\) is drawn from distribution \(F_G\); if \(\omega =B\), it is drawn from distribution \(F_B\).

Disclosure is verifiable. That is, the set of evidence that the sender can feasibly disclose depends on the evidence she has (in contrast, in a signaling game, the sender chooses from the same set of signals regardless of her type). Throughout the paper, we maintain the following assumptions that are standard in the literature:

(Reflexivity):

The sender can always truthfully disclose her evidence \(e\);

(Transitivity):

If the sender can disclose \(e'\) when she has evidence \(e\), and she can disclose \(e''\) when she has evidence \(e'\), then she can disclose \(e''\) if she has evidence \(e\).

Under these assumptions, we can represent the “disclosure rule” as a preorder \(\precsim\) on \(E\). Disclosing \(m\) is feasible given evidence \(e\) if and only if \(m\precsim e\), and the feasible set of disclosure given a piece of evidence \(e\) is its lower contour set \(\{m\in E:m\precsim e\}\), denoted \(LC(e)\).

3.2 Payoffs

The receiver’s payoff \(u_R(a,\omega )\) depends on both his action and the realized state of the world (but not the true evidence or the disclosed evidence), and the receiver maximizes his expected payoff.Footnote 14 We assume that the receiver’s payoff function satisfies the following assumptionFootnote 15

(Increasing differences):

\(u_R(a,G)-u_R(a,B)\) is strictly increasing in \(a\).

Under this assumption, the receiver wants to match the state of the world. That is, his optimal action is weakly increasing in his posterior belief that the state is good. More precisely, given \(\mu \in [0,1]\), the solution to the receiver’s maximization problem

$$\begin{aligned} \phi (\mu ) = \mathop {\textrm{argmax}}\limits _{a\in A} \mu u_R(a,G) + (1-\mu )u_R(a,B) \end{aligned}$$

is upper hemicontinuous and weakly increasing in \(\mu\).Footnote 16

The sender’s payoff equals the receiver’s action, i.e., \(u_S(a,\omega )=a\). Given the assumption on the receiver’s payoff, the sender has a weak incentive to persuade the receiver that the state is good. Notice that the evidence \(e\), the disclosed evidence \(m\), and the realized state \(\omega\) are payoff irrelevant to the sender.

An evidence game is a tuple \(\mathcal {G}=\langle \pi _0,(E,\precsim ),F_G,F_B,A,u_R\rangle\).

3.3 Strategies and perfect Bayesian equilibrium

A strategy of the sender is \(\sigma :E\rightarrow \Delta (E)\) such that \(supp(\sigma (\cdot |e))\subset LC(e)\), a strategy of the receiver is \(\rho :E\rightarrow \Delta (A)\), and a system of beliefs of the receiver is \(\mu :E\rightarrow [0,1]\), where \(\mu (m)\) denotes the receiver’s posterior belief that the state is good after observing \(m\).

A perfect Bayesian equilibrium of \(\mathcal {G}\) is a collection of the sender’s strategy, the receiver’s strategy, and the receiver’s system of belief \((\sigma ,\rho ,\mu )\) such that:

(Sender optimality):

Given \(\rho\),

$$\begin{aligned} supp(\sigma (\cdot |e)) \subset \mathop {\textrm{argmax}}\limits _{m\precsim e}\sum _{a\in A}a\cdot \rho (a|m) \end{aligned}$$

for all \(e\in E\);

(Receiver optimality):

Given \(\mu\),

$$\begin{aligned} supp(\rho (\cdot |m)) \subset \phi (\mu (m)) \end{aligned}$$

for all \(m\in E\);

(Bayesian consistency):

For all on-path disclosure \(m\in \bigcup _{e\in E}supp(\sigma (\cdot |e))\),

$$\begin{aligned} \mu (m)=\frac{\sum _{e\in E}\sigma (m|e)F_G(e)\pi _0}{\sum _{e\in E}\sigma (m|e)[F_G(e)\pi _0+F_B(e)(1-\pi _0)]}. \end{aligned}$$

4 Refinements of perfect Bayesian equilibrium

Sections 4.1 through 4.3 study truth-leaning equilibrium, purifiable truthful equilibrium, and weakly truth-leaning equilibrium. Section 4.4 shows the relationship between these refinements. Section 4.5 shows that purifiable truthful equilibria are receiver optimal, and that there is no value of committing to a deterministic mechanism.

4.1 Truth-leaning equilibrium

A truth-leaning equilibrium of \(\mathcal {G}\) is a perfect Bayesian equilibrium \((\sigma ,\rho ,\mu )\) such that:

(Truth-leaning):

Given \(\rho\),

$$\begin{aligned} e \in \mathop {\textrm{argmax}}\limits _{m\precsim e}\sum _{a\in A}a\cdot \rho (a|m) \Rightarrow \sigma (e|e) = 1; \end{aligned}$$
(Off-path beliefs):

For all off-path disclosure \(m\), \(\mu (m)=\nu (m)\), where

$$\begin{aligned} \nu (m) = \frac{F_G(m)\pi _0}{F_G(m)\pi _0+F_B(m)(1-\pi _0)}. \end{aligned}$$

As the example in Sect. 2 shows, a truth-leaning equilibrium may not exist. The following proposition shows that nonexistence happens extensively. Fix an evidence structure and vary only the receiver’s payoffs. Unless “more evidence” implies “better evidence” (i.e., \(\nu\) is weakly increasing), there is a positive measure of evidence games in which a truth-leaning equilibrium does not exist.

Proposition 2

Fix \(\pi _0,(E,\precsim ),F_G,F_B\), and \(A\). Let \(\mathscr {G}\) be the set of all evidence games with prior \(\pi _0\), evidence space \((E,\precsim )\), distributions of evidence \(F_G\) and \(F_B\), and receiver action space \(A\). Identify \(\mathscr {G}\) with a subset of \(\mathbb {R}^{2K}\) by the bijection

$$\begin{aligned} \langle \pi _0,(E,\precsim ),F_G,F_B,A,u_R\rangle \mapsto \{u_R(a,G),u_R(a,B)\}_{a\in A}. \end{aligned}$$

If \(\nu :(E,\precsim )\rightarrow [0,1]\) is weakly increasing, then every evidence game in \(\mathscr {G}\) has a truth-leaning equilibrium, and in all truth-leaning equilibria, the sender discloses truthfully. If \(\nu\) is not weakly increasing, a truth-leaning equilibrium does not exist in a positive measure of evidence games in \(\mathscr {G}\).

4.2 Disturbed games and purifiable truthful equilibrium

A disturbed game is where the receiver has a private payoff shock (i.e., his type) \(\zeta :A\rightarrow \mathbb {R}\). The receiver has type dependent payoff \(v_R(a,\omega |\zeta )=u_R(a,\omega )+\zeta (a)\). We identify the set of the receiver’s types with \(\mathbb {R}^K\), where \(K=|A|\) is the number of available receiver actions. Let \(\eta\) be a distribution over \(\mathbb {R}^K\) that has full support and is absolutely continuous with respect to the Lebesgue measure.Footnote 17 Denote by \(\mathcal {G}_R(\eta )\) the disturbed game where the receiver’s type is distributed according to \(\eta\).

In the disturbed game, a strategy of the sender is \(\sigma :E\rightarrow \Delta (E)\) such that \(supp(\sigma (\cdot |e))\subset LC(e)\), a strategy of the receiver in \(\mathcal {G}_R(\eta )\) is \(r:E\times \mathbb {R}^K\rightarrow \Delta (A)\), and a system of beliefs of the receiver is \(\mu :E\rightarrow [0,1]\), where \(\mu (m)\) is the receiver’s posterior belief that the state is good after observing \(m\).Footnote 18 Given any strategy of the receiver \(r\), let \(\rho :E\rightarrow \Delta (A)\) be the induced distributions over the receiver’s actions. That is,

$$\begin{aligned} \rho (a|m) = \int r(a|m,\zeta )\eta (d\zeta ) \end{aligned}$$

is the probability that the receiver takes action \(a\) after \(m\) is disclosed. We shall also use the shorthand notation and write this as \(\rho =\langle r,\eta \rangle\).

A truth-leaning equilibrium of \(\mathcal {G}_R(\eta )\) is a tuple \((\sigma ,r,\mu )\) such that:

  • (Receiver optimality in disturbed games) Given \(\mu\),

    $$\begin{aligned} supp(r(\cdot |m,\zeta )) \subset \tau (\mu (m),\zeta ) \end{aligned}$$

    for all \(m\in E\) and \(\zeta \in \mathbb {R}^K\), where \(\tau (\bar{\mu },\zeta )\subset A\) is the solution to the type \(\zeta\) receiver’s problem given posterior belief \(\bar{\mu }\in [0,1]\) on the good state, i.e.,

    $$\begin{aligned} \tau (\bar{\mu },\zeta ) = \mathop {\textrm{argmax}}\limits _{a\in A} \bar{\mu } u_R(a,G)+(1-\bar{\mu })u_R(a,B)+\zeta (a); \end{aligned}$$
  • (Sender optimality), (Bayesian consistency), (Truth-leaning), and (Off-path beliefs), as are defined above for the original game \(\mathcal {G}\).

If \((\sigma ,r,\mu )\) is a truth-leaning equilibrium, we say \((\sigma ,\rho ,\mu )\) is a truth-leaning equilibrium outcome of \(\mathcal {G}_R(\eta )\).

In any disturbed game, the sender has a strict incentive to persuade the receiver. That is, from the sender’s perspective, the expected value of the receiver’s optimal action is strictly increasing in his posterior belief. Therefore, a truth-leaning equilibrium exists in any disturbed game, and it is equivalent to a weakly truth-leaning equilibrium. Moreover, a truth-leaning equilibrium is essentially unique, and the receiver’s equilibrium system of beliefs depends only on the evidence space \((E,\precsim )\) and the distributions \(F_G\) and \(F_B\). Specifically, it is independent of the disturbance \(\eta\).Footnote 19 Hence, the receiver’s equilibrium system of beliefs is the same across all truth-leaning equilibria of all disturbed games. In fact, the set of truth-leaning equilibria is the same in all disturbed games.

Lemma 3

A truth-leaning equilibrium exists in all disturbed games. Moreover, there exist a closed and convex set \(\Sigma ^\star \subset \Delta (E)^E\) and a system of beliefs of the receiver \(\mu ^\star\) such that for all disturbed games \(\mathcal {G}_R(\eta )\), \((\sigma ,r,\mu )\) is a truth-leaning equilibrium of \(\mathcal {G}_R(\eta )\) if and only if \(\sigma \in \Sigma ^\star\), \(\mu =\mu ^\star\), and \(supp(r(\cdot |m,\zeta ))\subset \tau (\mu (m),\zeta )\) for all \(m\in E\) and \(\zeta \in \mathbb {R}^K\).

We define a purifiable truthful equilibrium as the limit point of a sequence of truth-leaning equilibrium outcomes of the disturbed games as the payoff uncertainty goes to zero. Formally, a purifiable truthful equilibrium of \(\mathcal {G}\) is a tuple \((\sigma ,\rho ,\mu )\) such that there exists a sequence of disturbances \(\{\eta ^n\}_{n=1}^\infty\), and for each \(\eta ^n\), a truth-leaning equilibrium outcome \((\sigma ^n,\rho ^n,\mu ^n)\) of \(\mathcal {G}_R(\eta ^n)\) such that \(\eta ^n\) converges weakly to the point mass at \(0\), denoted \(\eta ^n\xrightarrow {w}\delta _0\), and \((\sigma ^n,\rho ^n,\mu ^n)\rightarrow (\sigma ,\rho ,\mu )\).

By Lemma 3, it is easy to see that a purifiable truthful equilibrium exists, and in any purifiable truthful equilibrium, \(\sigma \in \Sigma ^\star\) and \(\mu =\mu ^\star\). Since the receiver’s problem in any disturbed game depends only on his type and his posterior belief, the receiver’s action after seeing the disclosed evidence in any purifiable truthful equilibrium should depend only on his posterior belief. That is, if two pieces of evidence \(m\) and \(m'\) are such that \(\mu ^\star (m)=\mu ^\star (m')\), then \(\rho (\cdot |m)=\rho (\cdot |m')\) in any purifiable truthful equilibrium. Conversely, any perfect Bayesian equilibrium satisfying these conditions is a purifiable truthful equilibrium.

Theorem 4

A purifiable truthful equilibrium exists and is a perfect Bayesian equilibrium. Moreover, \((\sigma ,\rho ,\mu )\) is a purifiable truthful equilibrium if and only if \(\sigma \in \Sigma ^\star\), \(\mu =\mu ^\star\), \(supp(\rho (\cdot |m))\subset \phi (\mu (m))\) for all \(m\in E\), and \(\mu (m)=\mu (m')\Rightarrow \rho (\cdot |m)=\rho (\cdot |m')\).

Theorem 4 shows that the set of purifiable truthful equilibria is a connected set in all evidence games. If \(\phi (\mu ^\star (m))\) is a singleton for all \(m\in E\), then the receiver’s purifiable truthful equilibrium strategy is unique and is a pure strategy (i.e., \(\rho (a|m)=\textbf{1}_{a\in \phi (\mu ^\star (m))}\)), and all purifiable truthful equilibria differ only on the sender’s strategies. Since the sender’s action is payoff irrelevant, all purifiable truthful equilibria have the same payoff relevant outcome in the sense that the joint distribution of \((a,\omega )\) is the same. Moreover, any purifiable truthful equilibrium can be approached using arbitrary disturbances (see the remarks in Appendix A.3). That is, for all purifiable truthful equilibria \((\sigma ,\rho ,\mu )\) and all sequences of disturbances \(\eta ^n\xrightarrow {w}\delta _0\), a sequence of truth-leaning equilibrium outcomes \((\sigma ^n,\rho ^n,\mu ^n)\) of the disturbed games \(\mathcal {G}_R(\eta ^n)\) converges to \((\sigma ,\rho ,\mu )\). In Appendix A.6, we show that this is generic: for almost all evidence games, the receiver’s purifiable truthful equilibrium strategy is unique and in pure strategies; consequently, any purifiable truthful equilibrium is infinitesimally close to a truth-leaning equilibrium of any infinitesimally disturbed game.

Because the disclosed evidence is payoff irrelevant, these results do not follow the standard results for “generic” extensive form games. Since the sender can have a continuum of equilibrium strategies, and different strategies correspond to different joint distributions of \((e,m)\), there is a continuum of purifiable truthful equilibrium outcomes defined as distributions over terminal nodes identified by \((a,e,m,\omega )\) in evidence games. In contrast, for almost all finite extensive form games, the set of Nash equilibrium outcomes is finite (Kreps and Wilson 1982). It is also worth mentioning that purifiable truthful equilibria may not be regular (consider the associated normal form game and apply the definition of regularity by van Damme (1996)) even in generic evidence games. For example, in the purifiable truthful equilibrium of the example in Sect. 2, given that the receiver always chooses Reject, the sender is indifferent between any value of \(p\). The equilibrium \(p=1\), \(q=0\) is therefore irregular.

For nongeneric evidence games, i.e., where \(\phi (\mu ^\star (m))\) is not a singleton for some \(m\in E\), there is a continuum of the receiver’s purifiable truthful equilibrium strategies. A given purifiable truthful equilibrium may be the limit point of truth-leaning equilibrium outcomes only for some sequences of disturbed games, and not all sequences of disturbed games have a convergent sequence of truth-leaning equilibrium outcomes. For example, consider a slight variant of the example in Sect. 2 where the receiver’s payoff from approving a bad design is -1 (instead of -2). As a result, the receiver’s belief threshold is \(\frac{1}{2}\). There exists a continuum of purifiable truthful equilibria, where \(p=1\), \(q\in [0,1]\), \(\mu =\frac{1}{2}\). Specifically, there exists a purifiable truthful equilibrium in which the receiver chooses Approve (\(a=1\)) and Reject (\(a=0\)) with equal probability after seeing no evidence, i.e., \(q=\frac{1}{2}\). But in order to approach this equilibrium using truth-leaning equilibrium outcomes of the disturbed games, the sequence of disturbances \(\{\eta ^n\}_{n=1}^\infty\) must satisfy \(\eta ^n(\{\zeta (1)>\zeta (0)\})\rightarrow \frac{1}{2}\). That is, the probability that the receiver has a strict incentive to choose Approve at belief \(\frac{1}{2}\) must converge to \(\frac{1}{2}\), equating the probability that the receiver chooses Approve in the intended purifiable truthful equilibrium.

4.3 Perturbed games and weakly truth-leaning equilibrium

Let \(\varepsilon =\{\varepsilon _e,\varepsilon _{e|e}\}_{e\in E}\) be a collection of positive real numbers. The perturbed game \(\mathcal {G}_S(\varepsilon )\), as is defined in HKP, is an evidence game where the sender who has evidence \(e\) receives an extra payoff \(\varepsilon _e\) if she discloses truthfully, and she must disclose truthfully with at least probability \(\varepsilon _{e|e}\). That is, the sender’s payoff is \(v_S(a,e,m)=a+\varepsilon _e\textbf{1}_{e=m}\), and a strategy of the sender is \(\sigma :E\rightarrow \Delta (E)\) such that \(supp(\rho (\cdot |e))\subset LC(e)\) and \(\sigma (e|e)\ge \varepsilon _{e|e}\) for all \(e\).

A perfect Bayesian equilibrium of \(\mathcal {G}_S(\varepsilon )\) is a collection of the sender’s strategy, the receiver’s strategy, and the receiver’s system of beliefs \((\sigma ,\rho ,\mu )\) such that:

  • (Sender optimality) Given \(\rho\),

    $$\begin{aligned} \sigma (m|e) > 0 \Rightarrow m\in \mathop {\textrm{argmax}}\limits _{m\precsim e} \sum _{a\in A} v_S(a,e,m)\cdot \rho (a|m) \end{aligned}$$

    for all \(e\) and \(m\ne e\);

  • (Receiver optimality) and (Bayesian consistency), as are defined for \(\mathcal {G}\).

A weakly truth-leaning equilibrium of \(\mathcal {G}\) is a tuple \((\sigma ,\rho ,\mu )\) such that there exists a sequence of perturbations \(\{\varepsilon ^n\}_{n=1}^\infty\) and for each \(\varepsilon ^n\), a PBE \((\sigma ^n,\rho ^n,\mu ^n)\) of \(\mathcal {G}_S(\varepsilon ^n)\) such that \(\varepsilon ^n\rightarrow 0\), and \((\sigma ^n,\rho ^n,\mu ^n)\rightarrow (\sigma ,\rho ,\mu )\).

Proposition 5

A weakly truth-leaning equilibrium exists and is a perfect Bayesian equilibrium.

Unlike purifiable truthful equilibria, weakly truth-leaning equilibria often involve “borderline” receiver beliefs where the receiver is indifferent between two actions (see the remarks in Appendix A.5). Consequently, they may not be proper, and different sequences of perturbations may select different weakly truth-leaning equilibria, as is the case in the example in Sect. 2.

4.4 Relationship between truth-leaning, weakly truth-leaning, and purifiable truthful equilibria

HKP show that truth-leaning equilibrium and weakly truth-leaning equilibrium are equivalent in a class of evidence games where the receiver continuously chooses an action on the real line. The leading example in Sect. 2 shows that this is not true for finite evidence games. It turns out that the equivalence can be restored for purifiable truthful equilibria. On the one hand, if a purifiable truthful equilibrium is weakly truth-leaning, it is also a truth-leaning equilibrium. On the other hand, in almost all evidence games, a purifiable truthful equilibrium that is also truth-leaning is a weakly truth-leaning equilibrium.

Proposition 6

If a purifiable truthful equilibrium is weakly truth-leaning, it is also a truth-leaning equilibrium.

Proposition 7

Fix \(\pi _0,(E,\precsim ),F_G,F_B\), and \(A\). Let \(\mathscr {G}\subset \mathbb {R}^{2K}\) be the set of all evidence games with prior \(\pi _0\), evidence space \((E,\precsim )\), distributions of evidence \(F_G\) and \(F_B\), and receiver action space \(A\). Let \(\mathscr {N}\subset \mathscr {G}\) be the set of evidence games that have a purifiable truthful equilibrium that is truth-leaning but not weakly truth-leaning. \(\mathscr {N}\) has Lebesgue measure zero.

For nongeneric games, a purifiable truthful equilibrium that is truth-leaning equilibrium need not be weakly truth-leaning. Consider again the example presented in Sect. 4.2 where the receiver’s belief threshold is \(\frac{1}{2}\). There exists a continuum of truth-leaning equilibria, where \(p=1\), \(q>0\), and \(\mu =\frac{1}{2}\). That is, the sender always discloses no evidence, and the receiver approves the design with positive probability so that the sender’s incentive to disclose no evidence is strict. All truth-leaning equilibria are purifiable truthful. However, there is a unique weakly truth-leaning equilibrium, where \(p=1\), \(q=1\), and \(\mu =\frac{1}{2}\). That is, the receiver must choose Approve after seeing no evidence. This is because, in every perturbed game, the sender cannot report no evidence with probability one, the receiver’s posterior belief is therefore strictly higher than the belief threshold \(\frac{1}{2}\), and he strictly prefers choosing Approve after seeing no evidence. This example is not generic, since the receiver is indifferent between Approve and Reject after seeing no evidence.

The example in Sect. 2 also shows that purifiable truthful equilibrium and weakly truth-leaning equilibrium do not imply each other, and neither implies truth-leaning equilibrium. To complete this part, we now show, using a variant of the example in Sect. 2, that an equilibrium that is both truth-leaning and weakly truth-leaning can fail to be purifiable truthful. Suppose that we alter the distribution of the sender’s evidence when the design is good such that the sender has bad evidence and no evidence with equal probability. The distribution when the design is bad remains unchanged. The game has a unique truth-leaning equilibrium, where \(p=0\), \(q=0\), \(\mu =\frac{3}{5}\).Footnote 20 That is, the sender discloses truthfully, the receiver always rejects the design, and the receiver’s belief on the good design is \(\frac{3}{5}\) after seeing no evidence. Notice that this is also the unique weakly truth-leaning equilibrium of the game. However, it is not a purifiable truthful equilibrium. In the unique purifiable truthful equilibrium, the sender always discloses no evidence, the receiver always rejects the project, and its belief on the good design is \(\frac{1}{2}\) after seeing no evidence (i.e., \(p=1,q=0,\mu =\frac{1}{2}\)).

4.5 Receiver optimality and value of commitment

A justification for truth-leaning equilibria in HKP is that they are receiver optimal. Moreover, HKP show that optimal deterministic mechanisms give the receiver the same expected payoff as truth-leaning equilibria.Footnote 21 Nonexistence of truth-leaning equilibrium prevents us from claiming the same in finite evidence games. The following proposition shows that purifiable truthful equilibria have these desired properties.

Proposition 8

Purifiable truthful equilibria are receiver optimal. The receiver’s purifiable truthful equilibrium payoff equals his payoff in the optimal deterministic mechanism.

To better illustrate this result, consider modifying the example in Sect. 2 as follows. The receiver’s belief threshold is \(\frac{1}{3}\), and the sender has conclusive good evidence with probability \(\frac{2}{3}\) if the state is good. That is, the sender can have (conclusive) good evidence, (conclusive) bad evidence, or no evidence. In the unique purifiable truthful equilibrium, the sender truthfully discloses good evidence and never discloses bad evidence, and the receiver chooses Accept if and only if good evidence is disclosed. Besides this purifiable truthful equilibrium, there exists a continuum of perfect Bayesian equilibria, where the sender type with bad evidence discloses no evidence, the sender type with good evidence discloses no evidence with at least \(\frac{1}{4}\) probability, and the receiver always chooses Accept on the equilibrium path. Clearly, the purifiable truthful equilibrium gives the receiver a strictly higher expected payoff than any other perfect Bayesian equilibrium.

Sher (2011) shows that if the receiver’s payoff is a concave function in his action, even stochastic mechanisms cannot give the receiver a higher payoff than his payoff in the receiver optimal equilibrium. Our assumption of increasing differences is weaker, and the receiver may achieve a higher payoff than his purifiable truthful equilibrium payoff by committing to a stochastic mechanism. As an example, consider adding a third receiver action, Use at own risk (\(a=\frac{1}{2}\)), to the example in Sect. 2. If this action is chosen, the sender’s payoff is \(\frac{1}{2}\), and the receiver’s payoff is \(x\) if the design is good and \(-x\) if the design is bad, where \(0<x<\frac{1}{4}\). Notice that the receiver’s payoff function is not concave in his action in the good state. Let \(q(0),q(\frac{1}{2})\), and \(q(1)\) denote the receiver’s probability of choosing Reject, Use at own risk, and Approve after seeing no evidence, respectively. And again, \(p\) is the sender’s probability of disclosing no evidence when having bad evidence, and \(\mu\) is the receiver’s posterior belief that the design is good after seeing no evidence. This game has a continuum of perfect Bayesian equilibria, where \(p=1\), \(q(0)+q(\frac{1}{2})=1\), and \(\mu =\frac{1}{2}\). That is, the sender always discloses no evidence, the receiver chooses Reject, Use at own risk, or randomizes between these two actions after seeing no evidence, and the receiver’s posterior belief on the good design is \(\frac{1}{2}\) after seeing no evidence.Footnote 22 In every equilibrium, the receiver’s ex ante expected payoff is zero. The receiver can achieve a positive payoff by committing to a stochastic mechanism. Suppose that it commits to choosing Use at own risk after seeing bad evidence and randomizing over Reject and Approve with equal probability after seeing no evidence. It is then optimal for the sender to disclose bad evidence truthfully, since it receives \(\frac{1}{2}\) regardless. As the sender discloses evidence truthfully, the receiver’s ex ante expected payoff is \(\frac{1}{12}-\frac{1}{3}x>0\).

5 Conclusion

HKP propose truth-leaning equilibrium as a solution concept in evidence games. The intuition is that the sender may find it slightly more advantageous to disclose evidence truthfully when being indifferent. This paper points out two problems of applying this solution concept to finite evidence games. First, it may fail to exist. Second, it may not agree with the intuition that the sender receives an infinitesimal reward for truth-telling. That is, truth-leaning equilibrium is not equivalent to weakly truth-leaning equilibrium in finite evidence games.

We propose a simple solution to restore existence by adding a small payoff uncertainty to the receiver. In the disturbed game, the sender is as if she faces a single receiver whom she has strict incentive to persuade, and therefore, a truth-leaning equilibrium exists. A purifiable truthful equilibrium is a truth-leaning equilibrium in an infinitesimally disturbed game. We show that a purifiable truthful equilibrium always exists and characterize the set of purifiable truthful equilibria.

We also show the equivalence between truth-leaning and weakly truth-leaning for purifiable truthful equilibria. If a purifiable truthful equilibrium is weakly truth-leaning, it is also a truth-leaning equilibrium. Conversely, in almost all finite evidence games, a purifiable truthful equilibrium that is also truth-leaning is a weakly truth-leaning equilibrium.

Finally, we show that purifiable truthful equilibria are receiver optimal, and the receiver cannot achieve a higher payoff by committing to a deterministic mechanism.