Abstract
This paper studies deceptions conducted by agents in the presence of a regulator. The regulator is supposed to detect deviations from the “rightful” behavior through costly monitoring; thus she may not choose to be diligent in her job because of the associated costs. The goal is to understand the occurrence of deceptions when the interaction of the parties is not contractible, their behavior is not observable and the regulator has reputation concern for being perceived as diligent in a repeated incomplete-information setting. It is found that when the regulator faces a sequence of myopic agents, her payoff at any Nash equilibrium converges to the maximum payoff as the discount factor approaches to one for any prior belief on the regulator’s type. This suggests that, contrary to the well-known disappearance of reputation results in the literature, the reputation of the regulator for being diligent persists in the long-run in any equilibrium. These findings imply that socially undesirable behavior of the agents could be prevented through reputation concerns in this repeated setting.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsNotes
- 1.
- 2.
For instance, Bernard Madoff, a prominent investment manager, was arrested and found guilty to several offenses including securities fraud and making false statements with the Securities and Exchange Commission (SEC). He began the Ponzi scheme in the early 1990s, yet he was arrested in late 2008 even though the SEC had previously conducted several investigations since 1992. SEC has been criticized for failure to act on Madoff fraud. The SEC inspector confessed that “Despite several examinations and investigations being conducted, a through and competent investigation or examination was never performed” (see “SEC criticized for failure to act on Madoff” at http://business.timesonline.co.uk by Seib and “Madoff Explains How He Concealed the Fraud” at www.cbsnews.com). Another investment fraud charge was against Robert Allen Stanford in 2009. A report of investigation by the SEC Office of the Inspector General shows that the agency has been following Stanford’s companies for much longer and reveals lack of diligence in the SEC enforcement (see the Report of Investigation at http://www.sec.gov/news/studies/2010/oig-526.pdf). The negligence of regulation may also have fatal consequences. For instance, there was a mine accident that took place in Soma, Turkey, which caused loss of 301 lives on May 13, 2014. In the response to a parliamentary question in the aftermath of the accident, it is understood that The General Directorate of Mining Affairs of Turkey—GDMA (that is connected to Ministry of Energy and Natural Resources), who is in charge of reviewing the conditions of mine fields, could only afford to audit less than one fourth all the mine fields annually. Yet, GDMA claimed that this particular mine had been reviewed many times. Although minor fees had been charged for infringement of some rules, an extensive audit of the field and mandatory safety measures had never been done and fatal mistakes went unnoticed according to the reports.
- 3.
An alternative formulation would be to have an agent who has the option to choose a proper action that generates a good signal from a set of actions and a regulator who can monitor the agent to check if she has chosen the proper action or not. There will be a bad signal only when the agent has chosen an improper action and the regulator monitors the agent. If the agent has chosen the proper action, the public signal generated is going to be always good regardless of the monitoring strategy chosen by the regulator.
- 4.
For an extensive overview of the reputation literature, we refer to [18].
- 5.
These findings are likely to change when the regulator faces a long-lived agent having future objectives. Some other structures (such as stochastic replacement of the regulator or reform in the regulation system) would be needed to obtain recurrent reputation then. Cripps et al. [6] suggest that to obtain non-transient reputations other mechanisms should be incorporated into the model. One string of literature attains recurrent reputations by assuming that the type of the player is governed by a stochastic process through time. The reader is referred to [8, 12, 15, 17, 21, 23].
- 6.
- 7.
These studies include: [22] (for games of conflicting interests with asymmetric discount factors); [4] and [1] (for games with imperfect monitoring and asymmetric discount factors); [7] (for games of strictly conflicting interests with equal discount factors); [3] and [2] (for games of locally nonconflicting or strictly conflicting interests with equal discount factors); [5] (for games with equal discount factors where the commitment action is a dominant action).
- 8.
The results of [11] and [6] differ because [11] fixes the prior belief of being the commitment type and selects a threshold discount factor depending on this prior above which the player is sufficiently patient for their results to hold; whereas [6] fixes the discount factor while allowing the posterior belief to vary which eventually becomes so low that makes the required threshold discount factor (for [11]’s result to hold) to exceed way above the fixed discount factor.
- 9.
The proof is along the lines with [9]. However, in their model, the long-lived agent’s concern for differentiating himself from his bad counterpart results in the loss of all surplus and market collapse since the short-lived players choose not to participate the game.
- 10.
Take any Nash equilibrium in which the agents have been truthful with probability one until date s > t. In these periods, the reputation of the regulator will stay the same regardless of her behavior. Thus, it is a best response for the regulator to be lazy during these periods, thus her payoff is zero. Then the continuation play starting from date s is a Nash equilibrium with the same prior γ whose payoff can be no more than the original game.
- 11.
When the agent is truthful with probability σ A, t (h t), choosing lazy is superior to diligent. Thus, the constraint involves an inequality rather than an equality.
References
Aoyagi, M.: Reputation and dynamic Stackelberg leadership in infinitely repeated games. J. Econ. Theory 71, 378–393 (1996)
Atakan, A.E., Ekmekçi, M.: Reputation in long-run relationships. Rev. Econ. Stud. 79 (2), 451–480 (2011)
Atakan, A.E., Ekmekçi, M.: Reputation in the long-run with imperfect monitoring. Working paper (2011)
Celentani, M., Fudenberg, D., Levine, D.K., Pesendorfer, W.: Maintaining a reputation against a long-lived opponent. Econometrica 64 (3), 691–704 (1996)
Chan, J.: On the non-existence of reputation effects in two-person infinitely repeated games. Working paper (2000)
Cripps, M.W., Mailath, G.J., Samuelson, L.: Imperfect monitoring and impermanent reputations. Econometrica 72 (2), 407–432 (2004)
Cripps, M.W., Dekel, E., Pesendorfer, W.: Reputation with equal discounting in repeated games with strictly conflicting interests. J. Econ. Theory 121, 259–272 (2005)
Ekmekçi, M., Gossner, O., Wilson, A.: Impermanent types and permanent reputations. J. Econ. Theory 147, 162–178 (2012)
Ely, J.C., Välimäki, J.: Bad reputation. Q. J. Econ. 118 (3), 785–814 (2003)
Fudenberg, D., Levine, D.K.: Reputation and equilibrium selection in games with a patient player. Econometrica 57 (4), 759–778 (1989)
Fudenberg, D., Levine, D.K.: Maintaining a reputation when strategies are imperfectly observed. Rev. Econ. Stud. 59, 561–579 (1992)
Gale, D., Rosenthal, R.: Price and quality cycles for experience goods. RAND J. Econ. 25, 590–607 (1994)
Gössner, O.: Simple bounds on the value of reputation. Econometrica 79 (5), 1627–1641 (2011)
Harsanyi, J.C.: Games of incomplete information played by Bayesian players, Parts I–III. Manag. Sci. 14 (3), 159–182 (1967–68)
Holmström, B.: Managerial incentive problems: a dynamic perspective. Rev. Econ. Stud. 66, 169–182 (1999)
Kreps, D., Wilson, R.: Sequential equilibria. Econometrica 50, 863–894 (1982)
Mailath, G.J., Samuelson, L.: Who wants a good reputation? Rev. Econ. Stud. 68, 415–441 (2001)
Mailath, G.J., Samuelson, L.: Repeated Games and Reputations: Long-Run Relationships. Oxford University Press, Oxford (2006)
Milgrom, P., Roberts, J.: Predation, reputation and entry deterrence. J. Econ. Theory 27, 280–312 (1982)
Ozdogan, A.: Disappearance of reputations in two-sided incomplete-information games. Games Econ. Behav. 88, 211–220 (2014)
Phelan, C.: Public trust and government betrayal. J. Econ. Theory 130, 27–43 (2006)
Schmidt, K.M.: Commitment through incomplete information in a simple repeated bargaining game. J. Econ. Theory 60, 114–139 (1993)
Wiseman, T.: Reputation and impermanent types. Games Econ. Behav. 62 (1), 190–210 (2008)
Acknowledgements
The author gratefully acknowledges financial support from the Scientific and Technological Research Council of Turkey (TUBITAK) Project No.115K567.
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Appendices
Appendix
A.1 The Proof of Lemma 1
The utility of agent by being truthful is u A (T, σ R ) = 0. And, her expected utility by being untruthful is
The agent’s best response correspondence against the strategic regulator’s strategy (which implies the cutoff detection probability) is given by:
From this, we can deduce the cutoff prior beliefs. The strategy of the regulator that makes the agent indifferent between being truthful and untruthful, \(\sigma _{R} = \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)}\), is greater than 0 if \(\gamma <\gamma ^{{\ast}} = \frac{g} {\beta (g+l)}\) and equals to 0 if γ = γ ∗. If γ > γ ∗, then BR A (σ R ) = 1 for any value of σ R , i.e., even if the strategic regulator is lazy for sure.
The expected utility of the regulator by choosing to be diligent is u R (σ A , D) = (1 −σ A )[β d − (1 −β)f] − c, whereas his expected utility by choosing to be lazy is u R (σ A , L) = −(1 −σ A )f. Thus, regulator’s best response is given by:
The strategy of agent that makes regulator indifferent between choosing to be diligent and lazy, \(\sigma _{A} = 1 - \frac{c} {\beta (d+f)}\), is greater than 0 if \(\beta > \frac{c} {f+d}\) (which holds by assumption).
- Case 1:
-
γ > γ ∗: In this case, BR A (σ R ) = 1 for any σ R . The unique fixed point of the best response correspondences is σ A = 1 and σ R = 0.
- Case 2:
-
γ = γ ∗: The strategy that makes the agent indifferent between telling the truth and lying is σ R = 0. For σ R > 0, BR A (σ R ) = 1. But, against σ A = 1, σ R > 0 cannot be a best response. Thus, the equilibrium strategies are \(\sigma _{A} \in [1 - \frac{c} {\beta (d+f)},1]\) and σ R (D) = 0. As we have assumed that the agent is truthful for sure when she is indifferent σ A = 1 and σ R = 0 in this case.
- Case 3:
-
γ < γ ∗: The unique intersection of the best response correspondences in this case is when \(\sigma _{A} = 1 - \frac{c} {\beta (d+f)}\) and \(\sigma _{R} = \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)}\).
A.2 The Proofs of Lemmas 2–5
Proof (Proof of Lemma 2).
Given that agent chooses σ A, t (h t) = 1 at h t, choosing diligent or lazy generates the same distribution of public signals so that the continuation payoff v(h t, i d ) will be the same. Since the one-period utility u R (T, L) = 0 > u R (T, L) = −c, we conclude that σ R, t (h t) = 0 complying with the one-shot deviation principle.
Proof (Proof of Lemma 3).
Suppose that there exists a Nash equilibrium with a positive probability history h t at which σ R, t (h t) = 1. If agent is truthful with probability one at h t, then the regulator would have one-shot deviation gain by switching to lazy since the distribution of public signals, the posterior belief, and thus the continuation payoff would not change regardless of regulator’s action choice when the agent is truthful. And, the agent chooses to be untruthful with some probability on a Nash equilibrium only when the belief γ t−1(h t) at t − 1 is less than γ ∗ and the regulator chooses to be diligent by less than \(\bar{\sigma }_{R} = \frac{\pi ^{{\ast}}-\beta \gamma } {\beta (1-\gamma )} < 1\).
Proof (Proof of Lemma 4).
Consider an arbitrary public history h t that is reached with positive probability with respect to some Nash equilibrium. Suppose for a contradiction that for every τ ≥ t and every h τ that comes after h t, σ R, τ (h τ) = 0. Given γ t−1(h t) < γ ∗ and there will not be any detections after h t for every τ ≥ t and history since σ R, τ (h τ) = 0 by hypothesis, and thus the expected probability of detection is going to be less than π ∗ for every τ ≥ t and h τ. Thus, all the myopic agents are untruthful at every date and history starting h t. Thus, the regulator’s expected continuation payoff becomes v(h t) = −f which is less than the minmax payoff − e, providing us the desired contradiction.
Proof (Proof of Lemma 5).
First, note that in order to observe a detection at time t, the agent must have been untruthful. For the agent to have chosen untruthfulness with some positive probability at h t, the expected probability of detection must be lower than π ∗. Given that the belief at the beginning of time t at h t is γ ≡ γ t−1 (h t) < γ ∗, this requires that \(\sigma _{R,t}(h^{t}) \leq \frac{\pi ^{{\ast}}-\beta \gamma } {\beta (1-\gamma )}\), which is derived from (2). And, σ R, t (h t) must be greater than zero because otherwise there would not be any detections. Lastly, it is easy to see from expression (3) that the smallest posterior is obtained when \(\sigma _{R,t}(h^{t}) =\bar{\sigma } _{R}\) and equals to \(\varGamma (\gamma ) = \frac{\gamma \beta } {\pi ^{{\ast}}}\).
Rights and permissions
Copyright information
© 2016 Springer International Publishing Switzerland
About this chapter
Cite this chapter
Özdog̃an, A. (2016). Occurrence of Deception Under the Oversight of a Regulator Having Reputation Concerns. In: Petrosyan, L., Mazalov, V. (eds) Recent Advances in Game Theory and Applications. Static & Dynamic Game Theory: Foundations & Applications. Birkhäuser, Cham. https://doi.org/10.1007/978-3-319-43838-2_10
Download citation
DOI: https://doi.org/10.1007/978-3-319-43838-2_10
Published:
Publisher Name: Birkhäuser, Cham
Print ISBN: 978-3-319-43837-5
Online ISBN: 978-3-319-43838-2
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)