Skip to main content

Occurrence of Deception Under the Oversight of a Regulator Having Reputation Concerns

  • Chapter
  • First Online:

Part of the book series: Static & Dynamic Game Theory: Foundations & Applications ((SDGTFA))

Abstract

This paper studies deceptions conducted by agents in the presence of a regulator. The regulator is supposed to detect deviations from the “rightful” behavior through costly monitoring; thus she may not choose to be diligent in her job because of the associated costs. The goal is to understand the occurrence of deceptions when the interaction of the parties is not contractible, their behavior is not observable and the regulator has reputation concern for being perceived as diligent in a repeated incomplete-information setting. It is found that when the regulator faces a sequence of myopic agents, her payoff at any Nash equilibrium converges to the maximum payoff as the discount factor approaches to one for any prior belief on the regulator’s type. This suggests that, contrary to the well-known disappearance of reputation results in the literature, the reputation of the regulator for being diligent persists in the long-run in any equilibrium. These findings imply that socially undesirable behavior of the agents could be prevented through reputation concerns in this repeated setting.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info
Hardcover Book
USD   54.99
Price excludes VAT (USA)
  • Durable hardcover edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Notes

  1. 1.

    The pioneer work that shows the disappearance of noncredible reputations in a class of games that encompasses the same payoff structure of the current paper with different signalling technology is [6] (see Sect. 1.1 for further discussion).

  2. 2.

    For instance, Bernard Madoff, a prominent investment manager, was arrested and found guilty to several offenses including securities fraud and making false statements with the Securities and Exchange Commission (SEC). He began the Ponzi scheme in the early 1990s, yet he was arrested in late 2008 even though the SEC had previously conducted several investigations since 1992. SEC has been criticized for failure to act on Madoff fraud. The SEC inspector confessed that “Despite several examinations and investigations being conducted, a through and competent investigation or examination was never performed” (see “SEC criticized for failure to act on Madoff” at http://business.timesonline.co.uk by Seib and “Madoff Explains How He Concealed the Fraud” at www.cbsnews.com). Another investment fraud charge was against Robert Allen Stanford in 2009. A report of investigation by the SEC Office of the Inspector General shows that the agency has been following Stanford’s companies for much longer and reveals lack of diligence in the SEC enforcement (see the Report of Investigation at http://www.sec.gov/news/studies/2010/oig-526.pdf). The negligence of regulation may also have fatal consequences. For instance, there was a mine accident that took place in Soma, Turkey, which caused loss of 301 lives on May 13, 2014. In the response to a parliamentary question in the aftermath of the accident, it is understood that The General Directorate of Mining Affairs of Turkey—GDMA (that is connected to Ministry of Energy and Natural Resources), who is in charge of reviewing the conditions of mine fields, could only afford to audit less than one fourth all the mine fields annually. Yet, GDMA claimed that this particular mine had been reviewed many times. Although minor fees had been charged for infringement of some rules, an extensive audit of the field and mandatory safety measures had never been done and fatal mistakes went unnoticed according to the reports.

  3. 3.

    An alternative formulation would be to have an agent who has the option to choose a proper action that generates a good signal from a set of actions and a regulator who can monitor the agent to check if she has chosen the proper action or not. There will be a bad signal only when the agent has chosen an improper action and the regulator monitors the agent. If the agent has chosen the proper action, the public signal generated is going to be always good regardless of the monitoring strategy chosen by the regulator.

  4. 4.

    For an extensive overview of the reputation literature, we refer to [18].

  5. 5.

    These findings are likely to change when the regulator faces a long-lived agent having future objectives. Some other structures (such as stochastic replacement of the regulator or reform in the regulation system) would be needed to obtain recurrent reputation then. Cripps et al. [6] suggest that to obtain non-transient reputations other mechanisms should be incorporated into the model. One string of literature attains recurrent reputations by assuming that the type of the player is governed by a stochastic process through time. The reader is referred to [8, 12, 15, 17, 21, 23].

  6. 6.

    See [10] for games with perfect monitoring, [11] for games with imperfect public monitoring, and [13] for games with imperfect private monitoring.

  7. 7.

    These studies include: [22] (for games of conflicting interests with asymmetric discount factors); [4] and [1] (for games with imperfect monitoring and asymmetric discount factors); [7] (for games of strictly conflicting interests with equal discount factors); [3] and [2] (for games of locally nonconflicting or strictly conflicting interests with equal discount factors); [5] (for games with equal discount factors where the commitment action is a dominant action).

  8. 8.

    The results of [11] and [6] differ because [11] fixes the prior belief of being the commitment type and selects a threshold discount factor depending on this prior above which the player is sufficiently patient for their results to hold; whereas [6] fixes the discount factor while allowing the posterior belief to vary which eventually becomes so low that makes the required threshold discount factor (for [11]’s result to hold) to exceed way above the fixed discount factor.

  9. 9.

    The proof is along the lines with [9]. However, in their model, the long-lived agent’s concern for differentiating himself from his bad counterpart results in the loss of all surplus and market collapse since the short-lived players choose not to participate the game.

  10. 10.

    Take any Nash equilibrium in which the agents have been truthful with probability one until date s > t. In these periods, the reputation of the regulator will stay the same regardless of her behavior. Thus, it is a best response for the regulator to be lazy during these periods, thus her payoff is zero. Then the continuation play starting from date s is a Nash equilibrium with the same prior γ whose payoff can be no more than the original game.

  11. 11.

    When the agent is truthful with probability σ A, t (h t), choosing lazy is superior to diligent. Thus, the constraint involves an inequality rather than an equality.

References

  1. Aoyagi, M.: Reputation and dynamic Stackelberg leadership in infinitely repeated games. J. Econ. Theory 71, 378–393 (1996)

    Article  MathSciNet  MATH  Google Scholar 

  2. Atakan, A.E., Ekmekçi, M.: Reputation in long-run relationships. Rev. Econ. Stud. 79 (2), 451–480 (2011)

    Article  MathSciNet  Google Scholar 

  3. Atakan, A.E., Ekmekçi, M.: Reputation in the long-run with imperfect monitoring. Working paper (2011)

    Google Scholar 

  4. Celentani, M., Fudenberg, D., Levine, D.K., Pesendorfer, W.: Maintaining a reputation against a long-lived opponent. Econometrica 64 (3), 691–704 (1996)

    Article  MathSciNet  MATH  Google Scholar 

  5. Chan, J.: On the non-existence of reputation effects in two-person infinitely repeated games. Working paper (2000)

    Google Scholar 

  6. Cripps, M.W., Mailath, G.J., Samuelson, L.: Imperfect monitoring and impermanent reputations. Econometrica 72 (2), 407–432 (2004)

    Article  MathSciNet  MATH  Google Scholar 

  7. Cripps, M.W., Dekel, E., Pesendorfer, W.: Reputation with equal discounting in repeated games with strictly conflicting interests. J. Econ. Theory 121, 259–272 (2005)

    Article  MathSciNet  MATH  Google Scholar 

  8. Ekmekçi, M., Gossner, O., Wilson, A.: Impermanent types and permanent reputations. J. Econ. Theory 147, 162–178 (2012)

    Article  MathSciNet  MATH  Google Scholar 

  9. Ely, J.C., Välimäki, J.: Bad reputation. Q. J. Econ. 118 (3), 785–814 (2003)

    Article  MATH  Google Scholar 

  10. Fudenberg, D., Levine, D.K.: Reputation and equilibrium selection in games with a patient player. Econometrica 57 (4), 759–778 (1989)

    Article  MathSciNet  MATH  Google Scholar 

  11. Fudenberg, D., Levine, D.K.: Maintaining a reputation when strategies are imperfectly observed. Rev. Econ. Stud. 59, 561–579 (1992)

    Article  MathSciNet  MATH  Google Scholar 

  12. Gale, D., Rosenthal, R.: Price and quality cycles for experience goods. RAND J. Econ. 25, 590–607 (1994)

    Article  Google Scholar 

  13. Gössner, O.: Simple bounds on the value of reputation. Econometrica 79 (5), 1627–1641 (2011)

    Article  MathSciNet  MATH  Google Scholar 

  14. Harsanyi, J.C.: Games of incomplete information played by Bayesian players, Parts I–III. Manag. Sci. 14 (3), 159–182 (1967–68)

    Google Scholar 

  15. Holmström, B.: Managerial incentive problems: a dynamic perspective. Rev. Econ. Stud. 66, 169–182 (1999)

    Article  MATH  Google Scholar 

  16. Kreps, D., Wilson, R.: Sequential equilibria. Econometrica 50, 863–894 (1982)

    Article  MathSciNet  MATH  Google Scholar 

  17. Mailath, G.J., Samuelson, L.: Who wants a good reputation? Rev. Econ. Stud. 68, 415–441 (2001)

    Article  MathSciNet  MATH  Google Scholar 

  18. Mailath, G.J., Samuelson, L.: Repeated Games and Reputations: Long-Run Relationships. Oxford University Press, Oxford (2006)

    Book  Google Scholar 

  19. Milgrom, P., Roberts, J.: Predation, reputation and entry deterrence. J. Econ. Theory 27, 280–312 (1982)

    Article  MathSciNet  MATH  Google Scholar 

  20. Ozdogan, A.: Disappearance of reputations in two-sided incomplete-information games. Games Econ. Behav. 88, 211–220 (2014)

    Article  MathSciNet  MATH  Google Scholar 

  21. Phelan, C.: Public trust and government betrayal. J. Econ. Theory 130, 27–43 (2006)

    Article  MathSciNet  MATH  Google Scholar 

  22. Schmidt, K.M.: Commitment through incomplete information in a simple repeated bargaining game. J. Econ. Theory 60, 114–139 (1993)

    Article  MathSciNet  MATH  Google Scholar 

  23. Wiseman, T.: Reputation and impermanent types. Games Econ. Behav. 62 (1), 190–210 (2008)

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

The author gratefully acknowledges financial support from the Scientific and Technological Research Council of Turkey (TUBITAK) Project No.115K567.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ayça Özdog̃an .

Editor information

Editors and Affiliations

Appendices

Appendix

A.1 The Proof of Lemma 1

The utility of agent by being truthful is u A (T, σ R ) = 0. And, her expected utility by being untruthful is

$$\displaystyle{ u_{A}(U,\sigma _{R}) =\gamma [(1-\beta )g -\beta l] + (1-\gamma )\sigma _{R}[(1-\beta )g -\beta l] + (1-\gamma )(1 -\sigma _{R})g. }$$

The agent’s best response correspondence against the strategic regulator’s strategy (which implies the cutoff detection probability) is given by:

$$ \displaystyle\begin{array}{rcl} \sigma _{A} \equiv BR_{A}(\sigma _{R}) = \left \{\begin{array}{@{}l@{\quad }l@{}} 1 \quad &\quad \text{if}\quad \sigma _{R} > \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)} \\{} [0,1]\quad &\quad \text{if}\quad \sigma _{R} = \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)} \\ 0 \quad &\quad \text{if}\quad \sigma _{R} < \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)}. \end{array} \right.& &{}\end{array}$$
(A.1)

From this, we can deduce the cutoff prior beliefs. The strategy of the regulator that makes the agent indifferent between being truthful and untruthful, \(\sigma _{R} = \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)}\), is greater than 0 if \(\gamma <\gamma ^{{\ast}} = \frac{g} {\beta (g+l)}\) and equals to 0 if γ = γ . If γ > γ , then BR A (σ R ) = 1 for any value of σ R , i.e., even if the strategic regulator is lazy for sure.

The expected utility of the regulator by choosing to be diligent is u R (σ A , D) = (1 −σ A )[β d − (1 −β)f] − c, whereas his expected utility by choosing to be lazy is u R (σ A , L) = −(1 −σ A )f. Thus, regulator’s best response is given by:

$$\displaystyle\begin{array}{rcl} \sigma _{R} \equiv BR_{R}(\sigma _{A}) = \left \{\begin{array}{@{}l@{\quad }l@{}} 1 \quad &\quad \text{if}\quad \sigma _{A} < 1 - \frac{c} {\beta (d+f)} \\{} [0,1]\quad &\quad \text{if}\quad \sigma _{A} = 1 - \frac{c} {\beta (d+f)} \\ 0 \quad &\quad \text{if}\quad \sigma _{A} > 1 - \frac{c} {\beta (d+f)}. \end{array} \right.& &{}\end{array}$$
(A.2)

The strategy of agent that makes regulator indifferent between choosing to be diligent and lazy, \(\sigma _{A} = 1 - \frac{c} {\beta (d+f)}\), is greater than 0 if \(\beta > \frac{c} {f+d}\) (which holds by assumption).

Case 1:

γ > γ : In this case, BR A (σ R ) = 1 for any σ R . The unique fixed point of the best response correspondences is σ A  = 1 and σ R  = 0.

Case 2:

γ = γ : The strategy that makes the agent indifferent between telling the truth and lying is σ R  = 0. For σ R  > 0, BR A (σ R ) = 1. But, against σ A  = 1, σ R  > 0 cannot be a best response. Thus, the equilibrium strategies are \(\sigma _{A} \in [1 - \frac{c} {\beta (d+f)},1]\) and σ R (D) = 0. As we have assumed that the agent is truthful for sure when she is indifferent σ A  = 1 and σ R  = 0 in this case.

Case 3:

γ < γ : The unique intersection of the best response correspondences in this case is when \(\sigma _{A} = 1 - \frac{c} {\beta (d+f)}\) and \(\sigma _{R} = \frac{g-\gamma \beta (g+l)} {(1-\gamma )\beta (g+l)}\).

A.2 The Proofs of Lemmas 2–5

Proof (Proof of Lemma 2).

Given that agent chooses σ A, t (h t) = 1 at h t, choosing diligent or lazy generates the same distribution of public signals so that the continuation payoff v(h t, i d ) will be the same. Since the one-period utility u R (T, L) = 0 > u R (T, L) = −c, we conclude that σ R, t (h t) = 0 complying with the one-shot deviation principle.

Proof (Proof of Lemma 3).

Suppose that there exists a Nash equilibrium with a positive probability history h t at which σ R, t (h t) = 1. If agent is truthful with probability one at h t, then the regulator would have one-shot deviation gain by switching to lazy since the distribution of public signals, the posterior belief, and thus the continuation payoff would not change regardless of regulator’s action choice when the agent is truthful. And, the agent chooses to be untruthful with some probability on a Nash equilibrium only when the belief γ t−1(h t) at t − 1 is less than γ and the regulator chooses to be diligent by less than \(\bar{\sigma }_{R} = \frac{\pi ^{{\ast}}-\beta \gamma } {\beta (1-\gamma )} < 1\).

Proof (Proof of Lemma 4).

Consider an arbitrary public history h t that is reached with positive probability with respect to some Nash equilibrium. Suppose for a contradiction that for every τ ≥ t and every h τ that comes after h t, σ R, τ (h τ) = 0. Given γ t−1(h t) < γ and there will not be any detections after h t for every τ ≥ t and history since σ R, τ (h τ) = 0 by hypothesis, and thus the expected probability of detection is going to be less than π for every τ ≥ t and h τ. Thus, all the myopic agents are untruthful at every date and history starting h t. Thus, the regulator’s expected continuation payoff becomes v(h t) = −f which is less than the minmax payoff − e, providing us the desired contradiction.

Proof (Proof of Lemma 5).

First, note that in order to observe a detection at time t, the agent must have been untruthful. For the agent to have chosen untruthfulness with some positive probability at h t, the expected probability of detection must be lower than π . Given that the belief at the beginning of time t at h t is γ ≡ γ t−1 (h t) < γ , this requires that \(\sigma _{R,t}(h^{t}) \leq \frac{\pi ^{{\ast}}-\beta \gamma } {\beta (1-\gamma )}\), which is derived from (2). And, σ R, t (h t) must be greater than zero because otherwise there would not be any detections. Lastly, it is easy to see from expression (3) that the smallest posterior is obtained when \(\sigma _{R,t}(h^{t}) =\bar{\sigma } _{R}\) and equals to \(\varGamma (\gamma ) = \frac{\gamma \beta } {\pi ^{{\ast}}}\).

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this chapter

Cite this chapter

Özdog̃an, A. (2016). Occurrence of Deception Under the Oversight of a Regulator Having Reputation Concerns. In: Petrosyan, L., Mazalov, V. (eds) Recent Advances in Game Theory and Applications. Static & Dynamic Game Theory: Foundations & Applications. Birkhäuser, Cham. https://doi.org/10.1007/978-3-319-43838-2_10

Download citation

Publish with us

Policies and ethics