Skip to main content
Log in

Payment schemes in infinite-horizon experimental games

  • Published:
Experimental Economics Aims and scope Submit manuscript

Abstract

We consider payment schemes in experiments that model infinite-horizon games by using random termination. We compare paying subjects cumulatively for all periods of the game; with paying subjects for the last period only; with paying for one of the periods, chosen randomly. Theoretically, assuming expected utility maximization and risk neutrality, both the cumulative and the last period payment schemes induce preferences that are equivalent to maximizing the discounted sum of utilities. The last period payment is also robust under different attitudes toward risk. In comparison, paying subjects for one of the periods chosen randomly creates a present-period bias. We further provide experimental evidence from infinitely repeated prisoners’ dilemma games that supports the above theoretical predictions.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Notes

  1. In inter-generational experiments, subjects are often motivated by an induced inter-generational discount rate, which is the fraction of the next generation’s payoff to be added to the current player’s payoff; see, e.g., Schotter and Sopher (2003). With random continuation, however, such a procedure would result in double discounting and may distort players’ dynamic incentives.

  2. Holt (1986) shows that the random selection method may be used if subjects behave in accordance with the independence axiom of expected utility theory. Azrieli et al. (2011) demonstrate that in a multi-decision setting, paying for one decision problem, chosen randomly, is the only mechanism that elicits subjects’ choice behavior across various decision problems in an incentive compatible way. Several carefully designed experiments give reassuring evidence for using the random selection method in individual choice experiments (Starmer and Sugden 1991; Cubitt et al. 1998; Hey and Lee 2005). We are unaware of experimental studies that test the validity of the random selection method in game theory experiments.

  3. Charness and Genicot (2009) and Chandrasekhar et al. (2012) use the random payment method in infinite-horizon experiments. Charness and Genicot (2009) note that the discount factor would increase from earlier to later periods of a repeated game under random payment but claim that this effect is rather small. We demonstrate that behavioral implications may be quite substantial. Our earlier working paper (Sherstyuk et al. 2011) compares payment schemes in a different experimental setting. Chandrasekhar and Xandri (2011) confirm our theoretical findings but do not test the findings experimentally.

  4. Fudenberg and Tirole (1991, p. 148) note that the discount factor in an infinitely repeated game can represent pure time preference or the possibility that the game may terminate at the end of each period.

  5. Assume a strictly concave, increasing, and (without loss of generality) nonnegative-valued utility function u. Then u is subadditive, u(π 1+π 2+⋯+π t )<u(π 1)+u(π 2)+⋯+u(π t ) for all t. Hence, u(π 1+π 2)=u(π 1)+α 2 u(π 2) for some 0<α 2<1. Similarly, we have u(π 1+π 2+π 3)=u(π 1)+α 2 u(π 2)+α 3 u(π 3) for some 0<α 3<1, and so on. Therefore

    $$E\mathit{Pay}^{Cum} = u(\pi_1) + \sum_{t=2}^\infty p^{t-1} \alpha_t u(\pi_t), \quad 0< \alpha_t<1 \ \mbox{for all}\ t=2,\ldots. $$

    Clearly, the weight placed on the utility in period 1 (relative to the utilities in the subsequent periods) is larger under the cumulative payment scheme than in (1).

  6. Previous experimental evidence (Dal Bo and Frechette 2011) indicates that within a repeated game, cooperation rates are the highest in period 1 and then decrease in later periods. This suggests that incentives to cooperate are the most critical in period 1.

  7. Cooperation remains an SPNE under cumulative pay even if the players are risk averse to the degree commonly observed in laboratory experiments. Assuming a constant relative risk-aversion utility function u(π)=π 1−r, cooperation is supportable as an SPNE for the whole range of r>0 as estimated in Holt and Laury (2002), Table 3, and is further supportable as a risk-dominant equilibrium for r≤0.2. An affine transformation can be applied to this functional form to guarantee that it is increasing and non-negative valued for all relevant payoff levels. A similar point is made in footnote 5 in Dal Bo and Frechette (2011).

  8. It is possible to come up with parameter values such that, under cumulative pay, cooperation is both supportable as an SPNE, and a risk-dominant action, whereas under random pay, it is not supportable as a SPNE; e.g., a=52, b=96, c=27, d=0, and p=3/4. However, gains from cooperation relative to defection are smaller under such parameter values, and the basin of attraction of the AD strategy is larger. Previous studies (Dal Bo and Frechette 2011; Blonski et al. 2011) indicate that cooperation may prevail only when gains from cooperation far outweigh gains from defection. We therefore choose a setting that is very pro-cooperative under cumulative pay, and borderline cooperative, and not risk-dominant, under random pay.

  9. This matching protocol is the same as that reported in Duffy and Ochs (2009) and Aoyagi and Frechette (2009); in comparison, Dal Bo and Frechette (2011) use random rematching across repeated games.

  10. The existing literature on indefinitely repeated games shows that past relationship length has a positive and significant effect on subjects’ behavior, with longer past games leading to more cooperative decisions; see Engle-Warnick and Slonim (2006b) and also Dal Bo and Frechette (2011). Studies that use the same pre-drawn sequences of game lengths in multiple sessions include Engle-Warnick and Slonim (2006a) and Fudenberg et al. (2012).

  11. The second half of the session is considered to start with the first repeated game that starts after 50 decision periods have passed.

  12. With four sessions per treatment, p=0.0625 is the lowest p-value obtainable in the signed ranks test.

  13. We checked the robustness of the results by excluding each of the twelve sessions, one at a time, from the regressions. The findings are robust to these modifications. In particular, the treatment effects persist if we exclude the most cooperative session (cumulative pay session, Draw 1: cooperation rate=0.75 %), or the least cooperative session (random pay session, Draw 2: cooperation rate=0.17 %), from the analysis.

  14. The Trigger-once-forgiving strategy is closely related to Grim2 strategy in Fudenberg et al. (2012). Both strategies prescribe switching from cooperation to defection after the second, not the first, observed defection of the other player. The only difference is that Grim2 waits for two consecutive defections, whereas the Trigger-once-forgiving strategy looks at the cumulative number of observed defections in all previous rounds of the game.

  15. Trigger-with-Reversion would be equivalent to Trigger in the absence of player errors, but it is different if players make mistakes or if players’ actions are not perfectly implemented, as in Fudenberg et al. (2012). Fudenberg et al. do not consider this strategy in their strategy set. See Table S1 in the Supplementary materials for details.

  16. Table S2, included in the Supplementary materials, reports the percentages of all actions that can be explained by each of the strategies listed above. Interestingly, the strategy that explains the highest percentage of actions overall is Trigger-with-Reversion, correctly predicting between 72 and 78 % of all actions in each of the three treatments. TFT closely follows, explaining between 69 and 76 % of all actions.

  17. Participants’ responses to the post-experiment questionnaire indicate other possible misconceptions about game durations. Some participants believed that the probability of a repeated game ending increased once the game continued beyond the expected four rounds. For example, Subject 8 in Session 3 explained his choice between A and B as follows: “I chose A in the beginning, then chose it until either it was the 5th round where I chose B or until the other person chose B.”

  18. Further, if we normalize the payoffs so that the payoff from joint cooperation is one and the payoff from joint defection is zero, then our setting is somewhat similar to the treatment in Dal Bo and Frechette (2011) with p=3/4 and a=40, b=50, c=25, d=12 (R=40 treatment in their notation). Again, the overall cooperation rate of 55 % that we observe is close to that of 58.71 % reported in Dal Bo and Frechette.

  19. We observe that, overall, cooperation rates in the first round were higher than in all rounds by 5 % under cumulative pay (60 % versus 55 %), by 7 % under random pay (43 % versus 36 %), and by 14 % under last period pay (67 % versus 53 %); see Table 3.

References

  • Aoyagi, M., & Frechette, G. (2009). Collusion as public monitoring becomes noisy: experimental evidence. Journal of Economic Theory, 144, 1135–1165.

    Article  Google Scholar 

  • Azrieli, Y., Chambers, C., & Healy, P. J. (2011). Incentive compatibility across decision problems. Mimeo, Ohio State University. November.

  • Blonski, M., & Spagnolo, G. (2001). Prisoners’ other dilemma. Mimeo.

  • Blonski, M., Ockenfels, P., & Spagnolo, G. (2011). Equilibrium selection in the repeated prisoner’s dilemma: axiomatic approach and experimental evidence. American Economic Journal: Microeconomics, 3, 164–192.

    Article  Google Scholar 

  • Camerer, C., & Weigelt, K. (1993). Convergence in experimental double auctions for stochastically lived assets. In D. Friedman & J. Rust (Eds.), The double auction market: theories, institutions and experimental evaluations (pp. 355–396). Redwood City: Addison-Wesley.

    Google Scholar 

  • Charness, G., & Genicot, G. (2009). Informal risk-sharing in an infinite-horizon experiment. Economic Journal, 119, 796–825.

    Article  Google Scholar 

  • Chandrasekhar, A., & Xandri, J. P. (2011). A note on payments in experiments of infinitely repeated games with discounting. Mimeo, Massachusetts Institute of Technology. December.

  • Chandrasekhar, A., Kinnan, C., & Larreguy, H. (2012). Informal insurance, social networks, and saving access: evidence from a framed field experiment. Mimeo, Northwestern University. April.

  • Cox, J. C. (2010). Some issues of methods, theories, and experimental designs. Journal of Economic Behavior and Organization, 73, 24–28.

    Article  Google Scholar 

  • Cubitt, R., Starmer, C., & Sugden, R. (1998). On the validity of the random lottery incentive system. Experimental Economics, 1, 115–131.

    Google Scholar 

  • Dal Bo, P. (2005). Cooperation under the shadow of the future: experimental evidence from infinitely repeated games. American Economic Review, 95(5), 1591–1604.

    Article  Google Scholar 

  • Dal Bo, P., & Frechette, G. (2011). The evolution of cooperation in infinitely repeated games: experimental evidence. American Economic Review, 101, 411–429.

    Article  Google Scholar 

  • Davis, D. D., & Holt, C. A. (1993). Experimental economics. Princeton: Princeton University Press.

    Google Scholar 

  • Duffy, J., & Ochs, J. (2009). Cooperative behavior and the frequency of social interactions. Games and Economic Behavior, 66, 785–812.

    Article  Google Scholar 

  • Engle-Warnick, J., & Slonim, R. (2006a). Inferring repeated-game strategies from actions: evidence from trust game experiments. Economic Theory, 28, 603–632.

    Article  Google Scholar 

  • Engle-Warnick, J., & Slonim, R. (2006b). Learning to trust in indefinitely repeated games. Games and Economic Behavior, 54, 95–114.

    Article  Google Scholar 

  • Fischbacher, U. (2007). z-Tree: Zurich toolbox for ready-made economic experiments. Experimental Economics, 10, 171–178.

    Article  Google Scholar 

  • Fudenberg, D., Rand, D., & Dreber, A. (2012). Slow to anger and fast to forgive: cooperation in an uncertain world. American Economic Review, 102(2), 720–749.

    Article  Google Scholar 

  • Fudenberg, D., & Tirole, J. (1991). Game theory. Cambridge: MIT Press.

    Google Scholar 

  • Hey, J., & Lee, J. (2005). Do subjects separate (or are they sophisticated)? Experimental Economics, 8, 233–265.

    Article  Google Scholar 

  • Holt, C. A. (1986). Preference reversals and the independence axiom. American Economic Review, 76(3), 508–515.

    Google Scholar 

  • Holt, C. A., & Laury, S. K. (2002). Risk aversion and incentive effects. American Economic Review, 92(5), 1644–1655.

    Article  Google Scholar 

  • Lei, V., & Noussair, C. (2002). An experimental test of an optimal growth model. American Economic Review, 92(3), 549–570.

    Article  Google Scholar 

  • Murnighan, J. K., & Roth, A. (1983). Expecting continued play in prisoner’s dilemma games. Journal of Conflict Resolution, 27(2), 279–300.

    Article  Google Scholar 

  • Nowak, M. A., & Sigmund, K. (1993). A strategy of win-stay, lose-shift that outperforms tit-for-tat in prisoner’s dilemma. Nature, 364, 56–58.

    Article  Google Scholar 

  • Offerman, T., Potters, J., & Verbon, H. A. A. (2001). Cooperation in an overlapping generations experiment. Games and Economic Behavior, 36(2), 264–275.

    Article  Google Scholar 

  • Roth, A., & Murnighan, J. K. (1978). Equilibrium behavior and repeated play of the prisoner’s dilemma. Journal of Mathematical Psychology, 17, 189–198.

    Article  Google Scholar 

  • Schotter, A., & Sopher, B. (2003). Social learning and convention creation in inter-generational games: an experimental study. Journal of Political Economy, 111(3), 498–529.

    Article  Google Scholar 

  • Sherstyuk, K., Tarui, N., Ravago, M., & Saijo, T. (2009). Games with dynamic externalities and climate change. Mimeo, University of Hawaii at Manoa. http://www2.hawaii.edu/~katyas/pdf/climate-change_draft050209.pdf.

  • Sherstyuk, K., Tarui, N., Ravago, M., & Saijo, T. (2011). Payment schemes in random-termination experimental games. University of Hawaii Working Paper 11-2. http://www.economics.hawaii.edu/research/workingpapers/WP_11-2.pdf.

  • Starmer, C., & Sugden, R. (1991). Does the random-lottery incentive system elicit true preferences? An experimental investigation. American Economic Review, 81(4), 971–978.

    Google Scholar 

Download references

Acknowledgements

The research was supported by a research grant from the University of Hawaii College of Social Sciences and a Grant-in-Aid for Scientific Research on Priority Areas from the Ministry of Education, Science and Culture of Japan. Special thank you goes to Andrew Schotter for a motivating discussion. We are grateful to Jacob Goeree and two anonymous referees for many helpful suggestions, to P.J. Healy, David Rand and seminar participants at the University of Hawaii at Manoa for their valuable comments, and to Jay Viloria, Joshua Jensen and Chaning Jang for research assistance.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Katerina Sherstyuk.

Electronic Supplementary Material

Below is the link to the electronic supplementary material.

Sample Instructions (PDF 836 kB)

Appendices

Appendix A: Discount factors in random continuation games in periods beyond period 1

Consider how the relative weights between the current and the future change under random pay as the game progresses beyond period 1. In period 2, using manipulations similar to those for (3), we obtain that the expected payoff is:

(9)

Here, \(\delta^{rt}_{\tau}\) denotes the weight put on period τ under random pay when the game is in period t. Observe that \(\delta^{r2}_{1}=\delta^{r2}_{2}\) in period 2, whereas we had, from (3), \(\delta^{r1}_{1}>\delta^{r1}_{2}\) in period 1.

In general, assume the game has progressed to period t≥2. We obtain that the expected payoff in period t is:

(10)

where \(\delta^{rt}_{s}=\frac{1-p}{p^{t}}\{-\log(1-p)-\sum_{\tau=2}^{t} \frac{p^{\tau-1}}{\tau-1}\}\) is the weight put on each past period, s<t, and on the current period, s=t; and \(\delta^{rt}_{s}=\frac{1-p}{p^{t}}\{-\log(1-p)-\sum_{\tau=2}^{s} \frac{p^{\tau-1}}{\tau-1}\}\) is the payoff weight put on a future period s>t. This implies that the relative weights put on the current period t and the future periods s>t, \(\delta^{rt}_{t}/ (\sum_{s=t+1}^{\infty} \delta^{rt}_{s})\), change as t increases.

Appendix B: Incentives to cooperate in later periods

Do relative incentives to cooperate and defect change as the game progresses beyond the first period? As noted before, under cumulative pay, the gains and losses from defection do not change in periods beyond t=1. Under random pay, the relative gains and losses from cooperation and defection may change in later periods due to the changes in relative weights put on the present and the future. Comparing gains from cooperation and defection under random pay in period t≥2, from (10), the weight put on the current period t is \(\delta^{rt}_{t}\), and the future payoff weights are \(\sum_{\tau =t+1}^{\infty} \delta^{rt}_{\tau}\). Hence, under the Nash reversion, the players in the PD game will have incentives to cooperate in period t>1 if

$$\delta^{rt}_t (b-a) \leq (a-c) \sum _{\tau =t+1}^{\infty} \delta^{rt}_\tau , $$

or

$$ \frac{\delta^{rt}_t}{\sum_{\tau =t+1}^{\infty} \delta^{rt}_\tau } \leq \frac{a-c}{b-a} , $$
(11)

a condition less demanding than the requirement for cooperation in period t=1. Figure 4 presents a numerical simulation of the current to future payoff weight ratios under continuation probability p=3/4.

Fig. 4
figure 4

Ratios of the current payoff weight to the future payoff weights, p=3/4

The figure indicates that, for periods t>1, the random payment scheme continues to induce the present-period bias in behavior compared to the cumulative payment scheme, as \(\frac{\delta^{rt}_{t}}{\sum_{\tau =t+1}^{\infty} \delta^{rt}_{\tau}} > \frac{1-p}{p}\). However, this bias decreases and \(\frac{\delta^{rt}_{t}}{\sum_{\tau =t+1}^{\infty} \delta^{rt}_{\tau}}\) approaches \(\frac{1-p}{p}\) from above as t grows. This suggests that incentives to cooperate increase as the game progresses under random pay, but they are never as strong as the incentives to cooperate under cumulative pay.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Sherstyuk, K., Tarui, N. & Saijo, T. Payment schemes in infinite-horizon experimental games. Exp Econ 16, 125–153 (2013). https://doi.org/10.1007/s10683-012-9323-y

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s10683-012-9323-y

Keywords

JEL Classification

Navigation