Abstract
We study discrete-time dynamic games with effectively identical players who possess private states that evolve randomly. Players in these games are concerned with their undiscounted sums of period-wise payoffs in the finite-horizon case and discounted sums of stationary period-wise payoffs in the infinite-horizon case. In the general semi-anonymous setting, the other players influence a particular player’s payoffs and state evolutions through the joint state–action distributions that they form. When dealing with large finite games, we find it profitable to exploit symmetric mixed equilibria of a reduced feedback type for the corresponding nonatomic games (NGs). These equilibria, when continuous in a certain probabilistic sense, can be used to achieve near-equilibrium performances when there are a large but finite number of players. We focus on the case where independently generated shocks drive random actions and state transitions. The NG equilibria we consider are random state-to-action maps that pay no attention to players’ external environments. Yet, they can be adopted for a variety of real situations where the knowledge about the other players can be incomplete. Results concerning finite horizons also form the basis of a link between an NG’s stationary equilibrium and good stationary profiles for large finite games.
Similar content being viewed by others
References
Adlakha S, Johari R (2013) Mean field equilibrium in dynamic games with complementarities. Oper Res 61:971–989
Al-Najjar NI, Smorodinsky R (2001) Large nonanonymous repeated games. Games Econ Behav 37:26–39
Al-Najjar NI (2008) Large games and the law of large numbers. Games Econ Behav 64:1–34
Aumann RJ (1964a) Markets with a continuum of traders. Econometrica 32:39–50
Aumann RJ (1964b) Mixed and behavior strategies in infinite extensive games. In: Dresher M, Shapley L, Tucker A (eds) Annals of mathematics studies, 52, Advances in game theory. Princeton University Press, Princeton, pp 627–650
Balbus L, Dziewulski P, Reffett K, Woźny L (2017) A qualitative theory of large games with strategic complementarities. Econ Theory (forthcoming)
Balder EJ (1995) A unifying approach to existence of nash equilibria. Int J Game Theory 24:79–94
Balder EJ (2002) A unifying pair of Cournot–Nash equilibrium existence results. J Econ Theory 102:437–470
Banerjee A, Moll B (2010) Why does misallocation persist? Am Econ J Macroecon 2:189–206
Bensoussan A, Sung KCJ, Yam SCP, Yung SP (2014) Linear-quadratic mean field games. 2016. Linear-quadratic mean field games. J Optim Theory Appl 169:496–529
Bergin J, Bernhardt D (1995) Anonymous sequential games: existence and characterization of equilibria. Econ Theor 5:461–489
Carmona G (2004) Nash equilibria of games with a continuum of players. Working Paper, Universidade Nova de Lisboa
Carmona R, Lacker D (2015) A probabilistic weak formulation of mean field games and applications. Ann Appl Probab 25:1189–1231
Ethier SN, Kurtz TG (1986) Markov processes: characterization and convergence. Wiley, New York
Fudenberg D, Levine DK (1988) Open-loop and closed-loop equilibria in dynamic games with many players. J Econ Theory 44:1–18
Gabaix X, Lasry J-M, Lions P-L, Moll B (2016) The dynamics of inequality. Econometrica 84:2071–2111
Green EJ (1980) Non-cooperative price taking in large dynamic markets. J Econ Theory 22:155–182
Green EJ (1984) Continuum and finite-player noncooperative models of competition. Econometrica 52:975–993
Hart S, Hildenbrand W, Kohlberg E (1974) On equilibrium allocation as distributions on the commodity space. J Math Econ 1:159–167
Hildenbrand W (1974) Core and equilibria of a large economy. Princeton University Press, Princeton
Hopenhayn HA (1992) Entry, exit, and firm dynamics in long run equilibrium. Econometrica 60:1127–1150
Housman D (1988) Infinite player noncooperative games and the continuity of the Nash equilibrium correspondence. Math Oper Res 13:488–496
Huang M, Caines PE, Malham PR (2003) Individual and Mass Behaviour in Large Population Stochastic Wireless Power Control Problems: Centralized and Nash equilibrium solutions. In: Proceedings of the 42nd IEEE conference on decision and control, Maui, Hawaii, pp 98–103
Huang M, Malhame RP, Caines PE (2006) Large population stochastic dynamic games: closed-loop McKean–Vlasov systems and the Nash certainty equivalence principle. Commun Inf Syst 6:221–252
Jovanovic B, Rosenthal RW (1988) Anonymous sequential games. J Math Econ 17:77–88
Kalai E (2004) Large robust games. Econometrica 72:1631–1665
Khan MA (1985) On Extensions of the Cournot-Nash Theorem. In: Aliprantis CD, Burkinshaw O, Rothman NJ (eds) Advances in equilibrium theory. Springer lecture notes in economics and mathematical systems, pp 79–106
Khan MA, Rath KP, Sun Y (1997) On the existence of pure strategy equilibrium in games with a continuum of players. J Econ Theory 76:13–46
Lasry J-M, Lions P-L (2007) Mean field games. Jpn J Math 2:229–260
Light B, Weintraub GY (2019) Mean field equilibrium: uniqueness, existence, and comparative statics. Working Paper, Stanford University, CA
Mas-Colell A (1984) On a theorem of Schmeidler. J Math Econ 13:201–206
Mertens JF, Parthasarathy T (1987) Equilibria for discounted stochastic games. CORE Discussion Paper No. 8750
Parthasarathy KR (2005) Probability measures on metric spaces. AMS Chelsea Publishing, Providence
Podczeck K (2009) On purification of measure-valued maps. Econ Theor 38:399–418
Podczeck K (2010) On existence of rich fubini extensions. Econ Theor 45:1–22
Rath KP (1992) A direct proof of the existence of pure strategy equilibria in games with a continuum of players. Econ Theor 2:427–433
Rath KP (1996) Existence and upper hemicontinuity of equilibrium distributions of anonymous games with discontinuous payoffs. J Math Econ 26:305–324
Reny PJ, Perry M (2006) Toward a strategic foundation for rational expectations equilibrium. Econometrica 74:1231–1269
Sabourian H (1990) Anonymous repeated games with a large number of players and random outcomes. J Econ Theory 51:92–110
Schmeidler D (1973) Equilibrium points of nonatomic games. J Stat Phys 7:295–300
Shapley LS (1953) Stochastic games. Proc Nat Acad Sci 39:1095–1100
Solan E (1998) Discounted stochastic games. Math Oper Res 23:1010–1021
Sun Y (2006) The exact law of large numbers via Fubini extension and characterization of insurable risks. J Econ Theory 126:31–69
Weintraub GY, Benkard CL, van Roy B (2008) Markov perfect industry dynamics with many firms. Econometrica 76:1375–1411
Weintraub GY, Benkard CL, van Roy B (2011) Industry dynamics: foundations for models with an infinite number of firms. J Econ Theory 146:1965–1994
Wiszniewska-Matyszkiel A (2000) Existence of pure equilibria in games with nonatomic space of players. Topol Methods Nonlinear Anal 16:339–349
Wiszniewska-Matyszkiel A (2002) Discrete time dynamic games with continuum of players, I: decomposable games. Int Game Theory Rev 4:331–342
Wiszniewska-Matyszkiel A (2003) Discrete time dynamic games with continuum of players, II: semi-decomposable games. Int Game Theory Rev 5:27–40
Wiszniewska-Matyszkiel A (2005) A dynamic game with continuum of players and its counterpart with finitely many players. In: Nowak AS, Szajowski K (eds) Advances in dynamic games. Annals of the international society of dynamic games, vol 7. Birkhäuser, Boston, pp 455–469
Wiszniewska-Matyszkiel A (2008) Common resources, optimality and taxes in dynamic games with increasing number of players. J Math Anal Appl 337:840–841
Wiszniewska-Matyszkiel A (2014) Open and Closed Loop Nash Equilibria in Games with a Continuum of Players. J Optim Theory Appl 160:280–301
Yang J (2017) A link between semi-anonymous sequential games and their large finite counterparts. International Journal of Game Theory 46:383–433
Yang J (2020) Supplementary Material to the Paper “Analysis of Markovian Competitive Situations using Nonatomic Games”. Working Paper, Rutgers University
Acknowledgements
The author wishes to thank Professor Thomas G. Kurtz for referring him to Ethier and Kurtz [14], a source instrumental to the proof of Lemma 3. He would also like to thank an anonymous referee for pointing out critical errors concerning the definition of \(\mathcal{K}(A,B,\pi _B,C)\), as well as the proofs of Lemma 4 and Proposition 1. In addition, the author wishes to acknowledge the support of the National Science Foundation of China Grant 11871362.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Electronic supplementary material
Below is the link to the electronic supplementary material.
Appendices
Technical Developments in Sect. 4
The Prohorov Metric: The Prohorov metric \(\rho _A\) for a metric space A is such that, for any distributions \(\pi ,\pi '\in \mathcal{P}(A)\),
where
The metric \(\rho _A\) is known to generate the weak topology for \(\mathcal{P}(A)\).
According to Parthasarathy [33] (Theorem II.7.1), the strong LLN applies to the empirical distribution under the weak topology, and hence under the Prohorov metric. In the following, we state its weak version.
Lemma 1
Given separable metric spaces A and B, suppose distribution \(\pi _A\in \mathcal{P}(A)\) and measurable mapping \(y\in \mathcal{M}(A,B)\). Then, for any \(\epsilon >0\), as long as n is large enough,
For a separable metric space A, a point \(a\in A\), and an \((n-1)\)-point empirical distribution space \(\pi \in \mathcal{P}_{n-1}(A)\), we use \((a,\pi )_n\) to represent the member of \(\mathcal{P}_n(A)\) that has an additional 1/n weight on the point a, but with probability masses in \(\pi \) being reduced to \((n-1)/n\) times of their original values. For \(a\in A^n\) and \(m=1,\ldots ,n\), we have \((a_m,\varepsilon (a_{-m}))_n=\varepsilon (a)\). Concerning the Prohorov metric, we have also a simple but useful observation.
Lemma 2
Let A be a separable metric space. Then, for any \(n=2,3,\ldots \), \(a\in A\), and \(\pi \in \mathcal{P}_{n-1}(A)\),
Proof
Let \(A'\in {\mathscr {B}}(A)\) be chosen. If \(a\notin A'\), then
if \(a\in A'\), then
Hence, it is always true that
In view of (A.1) and (A.2), we have
We have thus completed the proof. \(\square \)
The following result is important for showing the near-trajectory evolution of aggregate environments in large multi-period games. Among other things, it relies on Lemma 1.
Lemma 3
Given a separable metric space A and complete separable metric spaces B and C, suppose \(y_n\in \mathcal{M}(A^n,B^n)\) for every \(n\in {\mathbb {N}}\), \(\pi _A\in \mathcal{P}(A)\), \(\pi _B\in \mathcal{P}(B)\), and \(\pi _C\in \mathcal{P}(C)\). If
for any \(\epsilon >0\) and any n large enough, then
for any \(\epsilon >0\) and any n large enough.
Proof
Suppose sequence \(\{\pi '_{B1},\pi '_{B2},\ldots \}\) weakly converges to the given probability measure \(\pi _B\), and sequence \(\{\pi '_{C1},\pi '_{C2},\ldots \}\) weakly converges to the given probability measure \(\pi _C\). We are to show that the sequence \(\{\pi '_{B1}\otimes \pi '_{C1},\pi '_{B2}\otimes \pi '_{C2},\ldots \}\) weakly converges to \(\pi _B\otimes \pi _C\).
Let F(B) denote the family of uniformly continuous real-valued functions on B with bounded support. Let F(C) be similarly defined for C. We certainly have
Define F so that
where \(\mathbf{1}\) stands for the function whose value is 1 everywhere. By (A.7) and (A.8),
According to Ethier and Kurtz [14] (Proposition III.4.4), F(B) and F(C) happen to be \(\mathcal{P}(B)\) and \(\mathcal{P}(C)\)’s convergence determining families, respectively. As B and C are complete, Ethier and Kurtz ([14], Proposition III.4.6, whose proof involves Prohorov’s Theorem, i.e., the equivalence between tightness and relative compactness of a collection of probability measures defined for complete separable metric spaces) further states that F as defined through (A.8) is convergence determining for \(\mathcal{P}(B\times C)\). Therefore, we have the desired weak convergence by (A.9).
Let \(\epsilon >0\) be given. In view of the above product-measure convergence and the equivalence between the weak topology and that induced by the Prohorov metric, there must be \(\delta _B>0\) and \(\delta _C>0\), such that \(\rho _B(\pi '_B,\pi _B)<\delta _B\) and \(\rho _C(\pi '_C,\pi _C)<\delta _C\) will imply
By (A.1) and the given hypothesis, there is \({\bar{n}}^1\in {\mathbb {N}}\), so that for \(n={\bar{n}}^1,{\bar{n}}^1+1,\ldots \),
where \({\tilde{A}}_n\) contains all \(a\in A^n\) such that
By (A.1) and Lemma 1, on the other hand, there is \({\bar{n}}^2\in {\mathbb {N}}\), so that for \(n={\bar{n}}^2,\bar{n}^2+1,\ldots \),
where \({\tilde{C}}_n\) contains all \(c\in C^n\) such that
For any \(n={\bar{n}}^1\vee {\bar{n}}^2,{\bar{n}}^1\vee {\bar{n}}^2+1,\ldots \), let (a, c) be an arbitrary member of \({\tilde{A}}_n\times {\tilde{C}}_n\). We have from (A.10), (A.12), and (A.14) that,
Noting the facilitating (a, c) is but an arbitrary member of \({\tilde{A}}_n\times {\tilde{C}}_n\), we see that
which by (A.11) and (A.13), is greater than \(1-\epsilon \). \(\square \)
Because the equivalence between tightness and relative compactness of a collection of probability measures is indirectly related to the proof of Lemma 3, we require B and C to be complete separable metric spaces.
Lemma 4
Given separable metric spaces A, B, C, and D, as well as distributions \(\pi _A\in \mathcal{P}(A)\), \(\pi _B\in \mathcal{P}(B)\), and \(\pi _C\in \mathcal{P}(C)\), suppose \(y_n\in \mathcal{M}(A^n,B^n)\) for every \(n\in {\mathbb {N}}\) and \(z\in \mathcal{K}(B,C,\pi _C,D)\). If
for any \(\epsilon >0\) and any n large enough, then
for any \(\epsilon >0\) and any n large enough.
Proof
Let \(\epsilon >0\) be given. Since \(z\in \mathcal{K}(B,C,\pi _C,D)\), there exist \(C'\in {\mathscr {B}}(C)\) satisfying
as well as
such that for any \(b,b'\in B\) and \(c,c'\in C'\) satisfying \(d_{B\times C}((b,c),(b',c'))<\delta \),
For any subset \(D'\) in \({\mathscr {B}}(D)\), we therefore have
This leads to \((z^{-1}(D'))^\delta \setminus (B\times (C \setminus C'))\subseteq z^{-1}((D')^\epsilon )\), and hence due to (A.17),
On the other hand, by the hypothesis, we know for n large enough,
where
By (A.23), for any \((a,c)\in E'_n\) and \(F'\in {\mathscr {B}}(B\times C)\),
Combining the above, we have, for any \((a,c)\in E'_n\) and \(D'\in {\mathscr {B}}(D)\),
where the first inequality is due to (A.21), the second inequality is due to (A.24), and the third inequality is due to (A.18). That is, we have
In view of (A.18) and (A.22), we have the desired result. \(\square \)
With Lemmas 1 to 4 ready, we can now prove Proposition 1 and then Theorem 1.
Proof of Proposition 1:
Let \(t=1,\ldots ,{\bar{t}}-1\) and \(x\in \mathcal{K}(S,G,\gamma ,X)\) be given. Define a map \(z\in \mathcal{M}(S\times G\times I,S)\), such that
where \(M(\sigma ,x)\) is given in (7). In view of (10) and (A.27), we have, for any \(S'\in {\mathscr {B}}(S)\),
For \(n\in {\mathbb {N}}\), \(g\equiv (g_m)_{m=1,\ldots ,n}\in G^n\), and \(i\equiv (i_m)_{m=1,\ldots ,n}\in I^n\), also define an operator \(T'_n(g,i)\) on \(\mathcal{P}_n(S)\) so that \(T'_n(g,i)\circ \varepsilon (s)=\varepsilon (s')\), where for \(m=1,2,\ldots ,n\),
It is worth noting that (A.29) is different from the earlier (15). In view of (A.27) and (A.29), we have, for \(S'\in {\mathscr {B}}(S)\), that \([T'_n(g,i)\circ \varepsilon (s)](S')\) equals
Combining (A.28) and (A.30), we arrive to a key observation that
In the rest of the proof, we first show the asymptotic closeness between \(T_t(x)\circ \sigma \) and \(T'_n(g,i)\circ \varepsilon (s_n(a))\), and then that between the latter and \(T_{nt}(x,g,i)\circ \varepsilon (s_n(a))\).
First, due to the hypothesis on the convergence of \(\varepsilon (s_n(a))\) to \(\sigma \), the completeness of the spaces S, G, and I and hence also the completeness of \(G\times I\), as well as Lemma 3,
for any \(\epsilon '>0\) and any n large enough. We then show z as defined in (A.27) is a member of \(\mathcal{K}(S,G\times I,\gamma \otimes \iota ,S)\) according to the definition around (17) using (S1) and \(x\in \mathcal{K}(S,G,\gamma ,X)\).
Fix any \(\epsilon >0\). By (S1), there exist \(\delta >0\) and \(I'\in {\mathscr {B}}(I)\) with \(\iota (I')>1-\epsilon /2\) such that
for any \((s,y),(s',y')\in S\times X\) and \(i,i'\in I'\) satisfying \(d_{S\times X\times I}((s,y,i),(s',y',i'))<\delta \). Since \(x\in \mathcal{K}(S,G,\gamma ,X)\), there exist \(\delta '\in (0,\delta /2]\) and \(G'\in {\mathscr {B}}(G)\) with \(\gamma (G')>1-\epsilon /2\) such that
for any \(s,s'\in S\) and \(g,g'\in G'\) satisfying \(d_{S\times G}((s,g),(s',g'))<\delta '\).
Now suppose \(s,s'\in S\) and \((g,i),(g',i')\in G'\times I'\) satisfy
By the first inequality of (3), \(d_{S\times G}((s,g),(s',g'))<\delta '\). This and the fact that \(g,g'\in G'\) would result in (A.34). Due to the first inequality of (3), another consequence of (A.35) is
With the second inequality of (3), we can conclude from (A.34) and (A.36) that
As \(i,i'\in I'\), we can see from (A.33) that
In addition, the measures of \(G'\) and \(I'\) would lead to
Since \(\epsilon >0\) is arbitrary, (17), (A.35), (A.38), and (A.39) would together mean that z as defined through (A.27) is a member of \(\mathcal{K}(S,G\times I,\gamma \otimes \iota ,S)\).
By Lemma 4, this fact along with (A.32) will lead to the strict dominance of \(1-\epsilon '\) by
for any \(\epsilon '>0\) and any n large enough. By (A.31), this is equivalent to that, given \(\epsilon >0\), there exists \({\bar{n}}^1\in {\mathbb {N}}\) so that for any \(n={\bar{n}}^1,{\bar{n}}^1+1,\ldots \),
where \({\tilde{A}}_n(\epsilon )\in {\mathscr {B}}^n(A\times G\times I)\) is equal to
Next, note that the only difference between \(T_{nt}(x,g,i)\circ \varepsilon (s_n(a))\) and \(T'_n(g,i)\circ \varepsilon (s_n(a))\) lies in that \(\varepsilon (s_{n,-m}(a),g_{-m})\) is used in the former as in (15), whereas \(\sigma \otimes \gamma \) is used in the latter as in (A.29). Here, \(s_{n,-m}(a)\) refers to the vector \((s_{n1}(a),\ldots ,s_{n,m-1}(a),s_{n,m+1}(a),\ldots ,s_{nn}(a))\). By (S2), there is \(\delta \in (0,\epsilon /4]\) and \(I'\in {\mathscr {B}}(I)\) with
so that for any \((s,g,i)\in S\times G\times I'\) and any \(\mu '\in \mathcal{P}(S\times X)\) satisfying \(\rho _{S\times X}(M(\sigma ,x),\mu ')<\delta \),
For each \(n\in {\mathbb {N}}\), define \(I'_n\) so that
Also important is that by (A.44) and (A.45), for any \(S'\in {\mathscr {B}}(S)\) and \(i\equiv (i_m)_{m=1,\ldots ,n}\in I'_n\),
whenever
It can be shown that \(I'_n\) will occupy a big chunk of \(I^n\) as measured by \(\iota ^n\) when n is large. Define a map q from I to \(\{0,1\}\) so that \(q(i)=1\) or 0 depending on whether or not \(i\in I'\). By (A.43), \(\iota \cdot q^{-1}\) is a Bernoulli distribution with \((\iota \cdot q^{-1})(\{1\})>1-\epsilon /4\). So by (A.45), \(I'_n\) contains all \(i\equiv (i_m)_{m=1,\ldots ,n}\in I^n\) that satisfy
Therefore, by Lemma 1, there exits \({\bar{n}}^2\in {\mathbb {N}}\), so that for \(n={\bar{n}}^2,{\bar{n}}^2+1,\ldots \),
We can also demonstrate that (A.47) will be highly likely when n is large. By Lemma 3 and the hypothesis on the convergence of \(\varepsilon (s_n(a))\) to \(\sigma \), we know \(\varepsilon (s_n(a),g)\) will converge to \(\sigma \otimes \gamma \) in probability. Due to Lemma 2, this conclusion applies to the sequence \(\varepsilon (s_{n,-m}(a),g_{-m})\) as well. The fact that \(x\in \mathcal{K}(S,G,\gamma ,X)\) certainly leads to \((\text{ prj }^{S\times G}_S,x)\in \mathcal{K}(S,G,\gamma ,S\times X)\). So by Lemma 4, there is \({\bar{n}}^3\in {\mathbb {N}}\), so that for \(n={\bar{n}}^3,{\bar{n}}^3+1,\ldots \),
where
Consider arbitrary \(n={\bar{n}}^1\vee {\bar{n}}^2\vee {\bar{n}}^3,\bar{n}^1\vee {\bar{n}}^2\vee {\bar{n}}^3+1,\ldots \), \((a,g,i)\in \tilde{A}_n(\epsilon )\cap ({\tilde{B}}_n(\delta )\times I'_n)\), and \(S'\in {\mathscr {B}}(S)\). By (A.1) and (A.42), we see that
Combining this with (A.46), (A.47), and (A.51), we obtain
According to (A.1), this means
Therefore, for \(n\ge {\bar{n}}^1\vee {\bar{n}}^2\vee {\bar{n}}^3\),
whereas the latter is, in view of (A.41), (A.49), and (A.50), greater than \(1-\epsilon \). \(\square \)
Proof of Theorem 1:
We use induction to show that, for each \(\tau =0,1,\ldots ,{\bar{t}}-t+1\),
for any \(\epsilon >0\) and n large enough, where \(\tilde{A}_{n\tau }(\epsilon )\in {\mathscr {B}}^n(S\times G^\tau \times I^\tau )\) is such that, for any \((s_t,g_{[t,t+\tau -1]},i_{[t,t+\tau -1]})\in {\tilde{A}}_{n\tau }(\epsilon )\),
Once the above is achieved, we can then define \({\tilde{A}}_n(\epsilon )\) required in the theorem by
By this and (A.56), we have \(\left( \sigma _t\otimes \gamma ^{{\bar{t}}-t+1}\otimes \iota ^{{\bar{t}}-t+1}\right) ^n\left( \tilde{A}_n(\epsilon )\right) \) greater than
for any \(\epsilon >0\) and n large enough.
Now we proceed with the induction process. First, note that \(T_{n,[t,t-1]}\circ \varepsilon (s_t)\) is merely \(\varepsilon (s_t)\) itself and \(T_{[t,t-1]}\circ \sigma _t\) is merely \(\sigma _t\) itself. Hence, we will have (A.56) for \(\tau =0\) for any \(\epsilon >0\) and n large enough just by Lemma 1. Then, for some \(\tau =1,2,\ldots ,{\bar{t}}-t+1\), suppose
for any \(\epsilon >0\) and n large enough. We may apply Proposition 1 to the above, while at the same time identifying \(S\times G^{\tau -1}\times I^{\tau -1}\) with A, \(\sigma _t\otimes \gamma ^{\tau -1}\otimes \iota ^{\tau -1}\) with \(\pi \), \(x_{t+\tau -1}\) with x, \(T_{n,[t,t+\tau -2]}(x_{[t,t+\tau -2]},g_{[t,t+\tau -2]},i_{[t,t+\tau -2]})\circ \varepsilon (s_t)\) with \(\varepsilon (s_n(a))\), and \(T_{[t,t+\tau -2]}(x_{[t,t+\tau -2]})\circ \sigma _t\) with \(\sigma \). This way, we will verify (A.56) for any \(\epsilon >0\) and n large enough. Therefore, the induction process can be completed. \(\square \)
Technical Developments in Sect. 5
Proof of Proposition 2:
Because payoffs are bounded, the value functions are bounded too. We then prove by induction on t. By (20), we know the result is true for \(t={\bar{t}}+1\). Suppose for some \(t={\bar{t}},{\bar{t}}-1,\ldots ,2\), we have the continuity of \(v_{t+1}(s_{t+1},\sigma _{t+1},x_{[t+1,{\bar{t}}]},x_{t+1})\) in \(s_{t+1}\). By this induction hypothesis, the distribution-wise uniform continuity of \(x_t\), (S1), (F1), and the boundedness of the value functions, we see the continuity of the right-hand side of (21) in \(s_t\). So, \(v_t(s_t,\sigma _t,x_{[t{\bar{t}}]},x_t)\) is continuous in \(s_t\), and we have completed our induction process. \(\square \)
Proof of Proposition 3:
We prove by induction on t. By (20) and (24), we know the result is true for \(t={\bar{t}}+1\). Suppose for some \(t={\bar{t}},{\bar{t}}-1,\ldots ,2\), we have the convergence of \(v_{n,t+1}(s_{t+1,1},\varepsilon (s^n_{t+1,-1}),x_{[t+1,{\bar{t}}]},x_{t+1})\) to \(v_{t+1}(s_{t+1,1},\sigma _{t+1},x_{[t+1,{\bar{t}}]},x_{t+1})\) at an \(s_{t+1,1}\)-independent rate when \(s_{t+1,-1}\equiv (s_{t+1,2},s_{t+1,3},\ldots )\) is sampled from \(\sigma _{t+1}\). Now, suppose \(s_{t,-1}\equiv (s_{t2},s_{t3},\ldots )\) is sampled from \(\sigma _t\). Let also \(g\equiv (g_1,g_2,\ldots )\) be generated through sampling on \((G,{\mathscr {B}}(G),\gamma )\) and \(i\equiv (i_1,i_2,\ldots )\) be generated through sampling on \((I,{\mathscr {B}}(I),\iota )\). In the remainder of the proof, let \(s^n_t\equiv (s_{t1},s_{t2},\ldots ,s_{tn})\) for any arbitrary \(s_{t1}\in S\), \(g^n\equiv (g_1,\ldots ,g_n)\) and \(i^n\equiv (i_1,\ldots ,i_n)\).
Due to Lemma 1, \(\varepsilon (s^n_{t,-1})\) will converge to \(\sigma _t\). By Lemma 2, \(\varepsilon (s^n_t)\) will converge to \(\sigma _t\) at an \(s_{t1}\)-independent rate. By Proposition 1, we know that \(T_{nt}(x_t,g^n,i^n)\circ \varepsilon (s^n_t)\) will converge to \(T_t(x_t)\circ \sigma _t\) in probability at an \(s_{t1}\)-independent rate, and by Lemma 2 again, so will \([T_{nt}(x_t,g^n,i^n)\circ \varepsilon (s^n_t)]_{-1}\) to \(T_t(x_t)\circ \sigma _t\). Now Lemma 3 will lead to the convergence in probability of \(\varepsilon (s^n_{t,-1},g^n_{-1})\) to \(\sigma _t\otimes \gamma \). Due to \(x_t\)’s distribution-wise uniform continuity, Lemma 4 will lead to the convergence in probability of \(M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1})\) to \(M(\sigma _t,x_t)\). Thus,
1. \(\psi _t(s_{t1},x_t(s_{t1},g_1),M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1}))\) will converge to \(\psi _t(s_{t1},x_t(s_{t1},g_1), M(\sigma _t,x_t))\) in probability at an \(s_{t1}\)-independent rate due to (F2);
2. \(v_{n,t+1}(\theta _t(s_{t1},x_t(s_{t1},g_1),M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1}),i_1),[T_{nt}(x_t,g^n,i^n)\circ \varepsilon (s^n_t)]_{-1}, x_{[t+1,{\bar{t}}]},x_{t+1})\) will converge to \(v_{t+1}(\theta _t(s_{t1},x_t(s_{t1},g_1),M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1})),i_1),T_t(x_t)\circ \sigma _t,x_{[t+1,{\bar{t}}]},x_{t+1})\) in probability at an \(s_{t1}\)-independent rate due to the induction hypothesis; the latter will in turn converge to \(v_{t+1}(\theta _t(s_{t1},x_t(s_{t1},g_1),M(\sigma _t,x_t),i_1),T(x_t)\circ \sigma _t,x_{[t+1,{\bar{t}}]},x_{t+1})\) in probability at an \(s_{t1}\)-independent rate due to (S2) and Proposition 2.
As per-period payoffs are bounded, all value functions are bounded. The above convergences will then lead to the convergence of the right-hand side of (25) to the right-hand side of (21) at an \(s_{t1}\)-independent rate. That is, \(v_{nt}(s_{t1},\varepsilon (s^n_{t,-1}),x_{[t{\bar{t}}]},x_t)\) will converge to \(v_t(s_{t1},\sigma _t,x_{[t{\bar{t}}]},x_t)\) at a rate independent of \(s_{t1}\). We have completed the induction process. \(\square \)
Proof of Theorem 2:
Let us consider subgames starting with some time \(t=1,2,\ldots ,{\bar{t}}\). For convenience, we let \(\sigma _t=T_{[1,t-1]}(x^*_{[1,t-1]})\circ \sigma _1\). Now let \(s_t\equiv (s_{t1},s_{t2},\ldots )\) be generated through sampling on \((S,{\mathscr {B}}(S),\sigma _t)\), \(g\equiv (g_1,g_2,\ldots )\) be generated through sampling on \((G,{\mathscr {B}}(G),\gamma )\), and \(i\equiv (i_1,i_2,\ldots )\) be generated through sampling on \((I,{\mathscr {B}}(I),\iota )\). In the remainder of the proof, we let \(s^n_t\equiv (s_{t1},\ldots ,s_{tn})\), \(s^n_{t,-1}\equiv (s_{t2},\ldots ,s_{tn})\), \(g^n\equiv (g_1,\ldots ,g_n)\), and \(i^n\equiv (i_1,\ldots ,i_n)\).
By Lemma 1 and Proposition 1, we know that \(\varepsilon (s^n_t)=\varepsilon (s_{t1},\ldots ,s_{tn})\) converges to \(\sigma _t\) in probability, and also that \(T_{nt}(x^*_t,g^n,i^n)\circ \varepsilon (s^n_t)\) converges to \(T_t(x^*_t)\circ \sigma _t\) in probability. Due to Lemma 2, \(\varepsilon (s^n_{t,-1})\) and \([T_{nt}(x^*_t,g^n,i^n)\circ \varepsilon (s^n_t)]_{-1}\) will have the same respective convergences. Also, Lemma 3 will lead to the convergence in probability of \(\varepsilon (s^n_{t,-1},g^n_{-1})\) to \(\sigma _t\otimes \gamma \). Due to \(x_t\)’s distribution-wise uniform continuity, Lemma 4 will lead to the convergence in probability of \(M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1})\) to \(M(\sigma _t,x_t)\). Then,
1. \(\psi _t(s_{t1},y(s_{t1},g_1),M_n(\varepsilon (s^n_{t,-1}),x_t,g_{-1}))\) will converge to \(\psi _t(s_{t1},y(s_{t1},g_1), M(\sigma _t,x_t))\) in probability at a y-independent rate due to (F2);
2. \(v_{n,t+1}(\theta _t(s_{t1},y(s_{t1},g_1), M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1}),i_1),[T_{nt}(x^*_t,g^n,i^n)\circ \varepsilon (s^n_t)]_{-1}, x^*_{[t+1,{\bar{t}}]},x^*_{t+1})\) will converge to \(v_{t+1}(\theta _t(s_{t1},y(s_{t1},g_1),M_n(\varepsilon (s^n_{t,-1}),x_t,g^n_{-1}),i_1),T_t(x^*_t)\circ \sigma _t, x^*_{[t+1,{\bar{t}}]}, x^*_{t+1})\) in probability at a y-independent rate due to Proposition 3, which, due to (S2) and Proposition 2, will converge to \(v_{t+1}(\theta _t(s_{t1},y(s_{t1},g_1),M(\sigma _t,x_t),i_1),T_t(x^*_t)\circ \sigma _t,x^*_{[t+1,{\bar{t}}]},x^*_{t+1})\) in probability at a y-independent rate.
As per-period payoffs are bounded, all value functions are bounded. By (21) and (25), the above convergences will then lead to the convergence of the left-hand side of (31) to the left-hand side of (27). At the same time, the right-hand side of (31) plus \(\epsilon \) will converge to the right-hand side of (27) due to the convergence of \(\varepsilon (s^n_{t,-1})\) to \(\sigma _t\), Proposition 3, and the uniform boundedness of the value functions. By (27), as long as n is large enough, (31) will be true for any \(\epsilon >0\) and \(y\in \mathcal{M}(S\times G,X)\). This would then lead to (32) due to Theorem 1 and the boundedness of payoff functions. \(\square \)
Technical Developments in Sect. 6
Value Functions for the Stationary Case: For \(t=0,1,\ldots \), we define \(v_t(s,\sigma ,x,y)\) as the total expected payoff a player can make from period 1 to t, when he starts period 1 with a state \(s\in S\) and a state-variable profile \(\sigma \), while all players keep on using the strategy x from period 1 to t with the exception of the current player in the very beginning, who deviates to \(y\in \mathcal{M}(S\times G,X)\) then. As a terminal condition, we have
Due to the stationarity of the setting, we have, for \(t=1,2,\ldots \),
This is much like (21); but with (39) being true, the last term in (C.2) actually appears simpler than its counterpart. Using (C.1) and (C.2), we can inductively show that
The sequence \(\{v_t(s,\sigma ,x,y)\}_{t=0,1,\ldots }\) is thus Cauchy with a limit point \(v_\infty (s,\sigma ,x,y)\). This \(v_\infty (s,\sigma ,x,y)\) can be understood as the infinite-horizon total discounted expected payoff a player can obtain by starting with state s and environment \(\sigma \), while all players adhere to the action plan x except for the current player in the beginning, who deviates to y then.
Now we move on to the n-player game \(\Gamma _n\) with the same stationary features provided by \(\psi \), \(\theta \), and \(\alpha \). The in-action environment experienced by player m will be \(M_n(\varepsilon (s_{-m}),x,g_{-m})\), as defined in (14), when the other players start with the state vector \(s_{-m}\equiv (s_l)_{l\ne m}\), all act according to some x strategy, and experience the pre-action shock vector \(g_{-m}\equiv (g_l)_{l\ne m}\). Given any strategy \(x\in \mathcal{M}(S\times G,X)\), pre-action shock vector \(g\equiv (g_m)_{m=1,\ldots ,n}\in G^n\), and post-action shock vector \(i\equiv (i_m)_{m=1,\ldots ,n}\in I^n\), we define \(T_n(x,g,i)\) as the operator on \(\mathcal{P}_n(S)\) that converts a period’s state-variable profile into that of the next period. Following the transient version (15), \(\varepsilon (s')=T_n(x,g,i)\circ \varepsilon (s)\) is such that
Let \(v_{nt}(s_1,\varepsilon (s_{-1}),x,y)\) be the total expected payoff player 1 can make from period 1 to t, when the player’s starting state is \(s_1\in S\), the other players’ initial states are given by the vector \(s_{-1}\equiv (s_m)_{m\ne 1}\), and all players adopt the strategy \(x\in \mathcal{M}(S\times G,X)\) with the exception of player 1, who adopts the strategy \(y\in \mathcal{M}(S\times G,X)\) in the beginning. We have
For \(t=1,2,\ldots \), similarly to (25), \(v_{nt}(s_1,\varepsilon (s_{-1}),x,y)\) is equal to
In (C.6), \([T_n(x,g,i)\circ \varepsilon (s)]_{-1}\) stands for \(\varepsilon (s'_{-1})\), while \(s'\) comes from \(\varepsilon (s')=T_n(x,g,i)\circ \varepsilon (s)\). Using (C.5) and (C.6), we can inductively show that
Thus, the sequence \(\{v_{nt}(s_1,\varepsilon (s_{-1}),x,y)\}_{t=0,1,\ldots }\) is Cauchy with limit \(v_{n\infty }(s_1,\varepsilon (s_{-1}),x,y)\).
Proof of Theorem 3:
Let \(\epsilon >0\) be fixed. For \(t=1,2,\ldots \) satisfying \(t\ge \ln (6{{\bar{\psi }}}/(\epsilon \cdot (1-\alpha )))/\ln (1/\alpha )+1\), we have from (C.6) and (C.7),
Therefore, we need merely to select such a large t and show that, when n is large enough,
For \(t=1,2,\ldots \), since \((x^*,\sigma ^*)\) forms an equilibrium for \(\Gamma \), we know (42) is true. This, as well as (C.2) and (C.3), lead to
for \(\tau =1,2,\ldots ,t\), \(g\in G\), \(s\in S\), and \(y\in \mathcal{M}(S\times G,X)\).
We associate entities here with those defined in Sect. 5 when \({\bar{t}}\) there is fixed at the t here. To signify the difference in the two notational systems, we add superscript “K” to symbols defined in the previous section. For instance, we write \(v^K_\tau \) for the \(v_\tau \) defined in that section, which has a different meaning than the \(v_\tau \) here. Now, our \(\alpha ^{t-\tau }\cdot v_\tau (s,\sigma ^*,x^*,y)\) can be understood as \(v^K_{t+1-\tau }(s,\sigma ^*,x',y)\), with \(x'\equiv (x'_{t+1-\tau },\ldots ,x'_t)\in (\mathcal{M}(S\times G,X))^\tau \) being such that \(x'_{t'}=x^*\) for \(t'=t+1-\tau ,\ldots ,t\). Due to the consistency of \(\sigma ^*\) with \(x^*\) through the definition (39), we can understand \(\sigma ^*\) as \(T^K_{[1,\tau -1]}(x'_{[1,\tau -1]})\circ \sigma ^K_1\), where \(x'_{[1,\tau -1]}\equiv (x'_1,\ldots ,x'_{\tau -1})\in (\mathcal{M}(S\times G,X))^{\tau -1}\) is such that \(x'_{t'}=x^*\) for \(t'=1,2,\ldots ,\tau -1\).
With these correspondences, (C.10) can be translated into something akin to (27), with the only difference being that \(-\epsilon /3\) should be added to all the right-hand sides. That is, we now know that the current \((x^*,\sigma ^*)\) offers an \((\epsilon /3)\)-Markov equilibrium for the nonatomic game \(\Gamma ^{K}(\sigma ^*)\) with \({\bar{t}}=t\), \(\theta ^K_\tau =\theta \), and \(\psi ^K_\tau =\alpha ^{\tau -1}\cdot \psi \). Even though Theorem 2 is nominally about going from a 0-equilibrium for the nonatomic game to \(\epsilon \)-equilibria for finite games, we can follow exactly the same logic used to prove it to go from an \((\epsilon /3)\)-equilibrium for the nonatomic game to \((2\epsilon /3)\)-equilibria for finite games.
Thus, from one of the theorem’s claims, we can conclude that, for n large enough and any \(y\in \mathcal{M}(S\times G,X)\),
where \(x'_{[1t]}\) is again to be understood as the strategy that takes action \(x^*(s,g)\) whenever the most immediate state–shock pair is (s, g). But this translates into (C.9). \(\square \)
Rights and permissions
About this article
Cite this article
Yang, J. Analysis of Markovian Competitive Situations Using Nonatomic Games. Dyn Games Appl 11, 184–216 (2021). https://doi.org/10.1007/s13235-020-00356-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13235-020-00356-x