Abstract
We study asymptotic properties of the following Markov system of \(N \ge 3\) points in [0, 1]. At each time step, the point farthest from the current centre of mass, multiplied by a constant \(p>0\), is removed and replaced by an independent \(\zeta \)-distributed point; the problem, inspired by variants of the Bak–Sneppen model of evolution and called a p-contest, was posed in Grinfeld et al. (J Stat Phys 146, 378–407, 2012). We obtain various criteria for the convergences of the system, both for \(p<1\) and \(p>1\). In particular, when \(p<1\) and \(\zeta \sim U[0,1]\), we show that the limiting configuration converges to zero. When \(p>1\), we show that the configuration must converge to either zero or one, and we present an example where both outcomes are possible. Finally, when \(p>1\), \(N=3\) and \(\zeta \) satisfies certain mild conditions (e.g. \(\zeta \sim U[0,1]\)), we prove that the configuration converges to one a.s. Our paper substantially extends the results of Grinfeld et al. (Adv Appl Probab 47:57–82, 2015) and Kennerberg and Volkov (Adv Appl Probab 50:414–439, 2018) where it was assumed that \(p=1\). Unlike the previous models, one can no longer use the Lyapunov function based just on the radius of gyration; when \(0<p<1\) one has to find a more finely tuned function which turns out to be a supermartingale; the proof of this fact constitutes an unwieldy, albeit necessary, part of the paper.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
This paper extends the results of [3, 4] on the so-called Keynesian beauty contest, or, as it was called in [4], Jante’s law process. Following [3], we recall that in the Keynesian beauty contest, we have N players guessing a number, and the person who guesses closest to the mean of all the N guesses wins; see [5, Ch. 12, Sect. V]. The formal version, suggested by Moulin [6, p. 72], assumes that this game is played by choosing numbers on the interval [0, 1], the “p-beauty contest”, in which the target is the mean value, multiplied by a constant \(p>0\). For the applications of the p-contest in the game theory, we refer the reader to e.g. [7]; see also [1, 3] and references therein for further applications and other relevant papers.
The version of the p-contest with \(p\equiv 1\) was studied in [3, 4]. In [3] it was shown that in the model where at each unit of time the point farthest from the center of mass is replaced by a point chosen uniformly on [0, 1], then eventually all (but one) points converge almost surely to some random limit the support of which is the whole interval [0, 1]; many of the results were extended for the version of the model on \({\mathbb {R}}^d\), \(d\ge 2\). The results of [3] were further generalized in [4], by removing the assumption that a new point is chosen uniformly on [0, 1], as well as by removing more than one point at once, these points being chosen in such a way that the moment of inertia of the resulting configuration is minimized. However, the case \(p\ne 1\) was not addressed in either of these two papers.
Let us now formally define the model; the notation will be similar to those in [3, 4]. Let \({{\mathcal {X}}}= \{ x_1, x_2, \ldots , x_N\}\in {\mathbb {R}}^N\) be an unordered N-tuple of points in \({\mathbb {R}}\), and \( (x_{(1)} , x_{(2)} , \dots , x_{(N)})\) be these points put in non-decreasing order, that is, \(x_{(1)}\le x_{(2)}\le \dots \le x_{(N)}\). As in [3, 4] let us define the barycentre of the configuration as
Fix some \(p>0\) and also define the p-centre of mass as \(p\mu _N(x_1,\dots ,x_N)\).
The point, farthest from the p-centre of mass, is called the extreme point of \({{\mathcal {X}}}\), and it can be either \(x_{(1)}\) or \(x_{(N)}\) (with possibility of a tie), and the core of \({{\mathcal {X}}}\), denoted by \({{\mathcal {X}}}'\), is constructed from \({{\mathcal {X}}}\) by removing the extreme point; in case of a tie between the left-most and the right-most point, we choose either of them with equal probability (same as in [3, 4]). Throughout the rest of the paper, \(x_{(1)}(t),\dots ,x_{(N-1)}(t)\) shall denote the points of the coreFootnote 1\({{\mathcal {X}}}'(t)\) put into non-decreasing order.
Our process runs as follows. Let \({{\mathcal {X}}}(t)=\{X_1(t), \dots , X_N(t)\}\) be an unordered N-tuple of points in \({\mathbb {R}}\) at time \(t=0,1,2,\dots \). Given \({{\mathcal {X}}}(t)\), let \({{\mathcal {X}}}'(t)\) be the core of \({{\mathcal {X}}}(t)\) and replace \({{\mathcal {X}}}(t)\setminus {{\mathcal {X}}}'(t) \) by a \(\zeta \)-distributed random variable so that
where \(\zeta _{t}\), \(t=1,2,\dots \), are i.i.d. random variables with a common distribution \(\zeta \).
Finally, to finish the specification of our process, we allow the initial configuration \({{\mathcal {X}}}(0)\) to be arbitrary or random, with the only requirement being that all the points of \({{\mathcal {X}}}(0)\) must lie in the support of \(\zeta \).
Throughout the paper we will use the notation \(A\underset{\text {a.s.}}{\Longrightarrow }B\) for two events A and B, whenever \({\mathbb {P}}(A\cap B^c)=0\), that is, when \(A\subseteq B\) up to a set of measure 0. We will also write, with some abuse of notations, that \(\lim _{t\rightarrow \infty }{{\mathcal {X}}}'(t)= a \in {\mathbb {R}}\) or equivalently \({{\mathcal {X}}}'(t)\rightarrow a\) as \(t\rightarrow \infty \) if \({{\mathcal {X}}}'(t)\rightarrow (a,a,\dots ,a)\in {\mathbb {R}}^{N-1}\), i.e. \(\lim _{t\rightarrow \infty }x_{(i)}(t)=a\) for all \(i=1,2,\dots ,N-1\). Similarly, for an interval (a, b) we will write \({{\mathcal {X}}}'(t)\in (a,b)\) whenever all\(x_{(1)}(t),\dots ,x_{(N-1)}(t)\in (a,b)\). Finally, we will assume that \(\inf \emptyset =+\infty \), and use the notation \(y^+=\max (y,0)\) for \(y\in {\mathbb {R}}\).
Also we require that \(\zeta \) has a full support on [0, 1], that is, \({\mathbb {P}}(\zeta \in (a,b))>0\) for all a, b such that \(0\le a<b\le 1\).
2 The Case \(p<1\)
Throughout this Section we assume that \(0<p<1\) and that \(\mathop {\mathrm {supp}}\zeta =[0,1]\). Because of the scaling invariance, our results may be trivially extended to the case when \(\mathop {\mathrm {supp}}\zeta =[0,A]\), \(A\in (0,\infty )\); some of them are even true when \(A=\infty \); however, to simplify the presentation from now on we will deal only with the case \(A=1\).
First, we present some general statements; more precise results will follow in case where \(\zeta \sim U[0,1]\).
Proposition 1
We have
- (a)
\(\liminf _{t\rightarrow \infty }x_{(N-1)}(t)=0\);
- (b)
\({\mathbb {P}}\left( \exists \lim _{t\rightarrow \infty } {{\mathcal {X}}}'(t)\in (0,1]\right) =0\);
- (c)
if \(p<\frac{1}{2} + \frac{1}{2(N-1)}\) then \({\mathbb {P}}\left( \lim _{t\rightarrow \infty } {{\mathcal {X}}}'(t)=0\right) =1\);
- (d)
if \(p< \frac{1}{2} + \frac{1}{N-2}\) then \(\{x_{(1)}(t)\rightarrow 0\}\underset{\text {a.s.}}{\Longrightarrow }\{\lim _{t\rightarrow \infty }{{\mathcal {X}}}'(t)= 0\}\).
Proof
(a) Since \(\zeta \) has full support on [0, 1] it follows that (see [4, Proposition 1]) there exists a function \(f:{\mathbb {R}}^+\rightarrow {\mathbb {R}}^+\) such that
Also, to simplify notations, we write \(\mu =\mu _N({{\mathcal {X}}}(t))\) throughout the proof.
Fix a small positive \(\varepsilon \) such that \(p+2\varepsilon <1\). Suppose that for some t we have \(x_{(N-1)}(t)\le b\le 1\). We will show that \(x_{(N-1)}(t+N)\le b(1-\varepsilon )\) with a strictly positive probability which only depends on \(p,b,\epsilon \) and N. Assume that we have \(\zeta _{t+1},\dots ,\zeta _{t+N-1}\in \left( pb,(p+\varepsilon )b \right) \subset (pb,b)\); this happens with probability no less than \(\left[ f(p\varepsilon b)\right] ^{N-1}\). We claim that by the time \(t+N\) we have \(x_{(N-1)}(t+N-1)<(p+\varepsilon )b\). Indeed, \(p\mu \le pb\) always lies to the left of the newly sampled points, therefore either there are no more points to the right of \((p+\varepsilon )b\) at some time \(s\in [t,t+N-1]\) (which implies that there will be no points there at time \(t+N\) due to the sampling range of the new points), or one of the older points, i.e. present at time t, gets removed (it can be the one to the left of pb). Since we eventually have to replace all the \(N-1\) old points, then \(x_{(N-1)}(t+N)\le b(1-\varepsilon )\).
Fix a \(\delta >0\) and find M so large that \((1-\varepsilon )^M<\delta \). Let the event \(C(s)=\{x_{(N-1)}(s) <\delta \}\). By iterating the above argument, we get that \({\mathbb {P}}(C(t+NM) | {{\mathcal {F}}}_t) \ge \prod _{i=1}^M \left[ f(p\varepsilon (1-\varepsilon )^{i-1})\right] ^{N-1}\), since at time t we can set \(b=1\). Therefore, \(\sum _m {\mathbb {P}}(C(NM(m+1)) | {{\mathcal {F}}}_{NMm})=\infty \) and by Lévy’s extension of the Borel–Cantelli lemma (see e.g. [8]) infinitely many C(s) occur. Since \(\delta >0\) is arbitrary, we get \(\liminf _{t\rightarrow \infty } x_{(N-1)}(t)=0\).
(b) Let \(r=\frac{1+p^{-1}}{2}>1\). Suppose that the core converges to some point \(x\in (0,1]\); then there exist a rational \(q\in (0,1]\) and a \(T>0\) such that \({{\mathcal {X}}}'(t)\in (q,rq)\) for all \(t\ge T\), formally
where \(A_{q,t}=\{{{\mathcal {X}}}'(t)\in (q,rq)\}\). We will show that
for some \(\nu _q>0\). This will imply, in turn, that
and hence the RHS (and thus the LHS as well) of (2.2) has the probability 0.
Suppose \(A_{q,t}\) has occurred and the newly sampled point \(\zeta \in (pq,q)\). Then
Consequently, \(x_{(N-1)}\) lies further from the p-center of mass, and hence it should be removed. The new configuration will, however, contain the point \(\zeta \notin (q,rq)\) and hence \(A_{q,t+1}\) does not occur. Thus
as required.
(c) First, we will show that it is the right-most point of the configuration which should be always removed; note that it suffices to check this only when \(x_{(N)}>0\). Indeed, by the assumption on p we have
implying that
Therefore, \(x_{(N)}\) is the farthest point from the p-centre of mass. This implies that \(x_{(N-1)}(t)\) is non-increasing and therefore result now easily follows from part (a) since \(x_{(N-1)}(t)\) is an upper bound for all the core points.
(d) Apply Corollary 3 with \(k=1\); it is possible because of Remark 3. \(\square \)
We are ready to present the main result of this Section.
Theorem 1
Suppose that \(\zeta \sim U[0,1]\). Then \({{\mathcal {X}}}'(t)\rightarrow 0\) a.s.
Proof
Proposition 1(c) implies that we now only need to consider the case \(p\ge \frac{N}{2(N-1)}\), which we will assume from now on.
Let us introduce a modification of this process on \([0,+\infty )\) which we will call the borderless p-contest; it is essentially the same process as the one in [3, Sect. 3.4]. In order to do this, we need the following statement. \(\square \)
Lemma 1
Suppose that \(x_1,\dots ,x_{N-1}> 0\). Then there exists an \(R=R(x_{(N-1)})\ge 0\) such that x is the farthest point from \(p\mu =\frac{p}{N}(x_1+\dots +x_{N-1}+x)\) whenever \(x>R\).
Proof of Lemma 1
Set \(R=6 x_{(N-1)}\). Then \(x>x_{(1)}\) is farther from the centre of mass than \(x_{(1)}\) if and only if
This is true, due to the fact that \(x>R\) and
since \(p<1\) and \(N\ge 3\). \(\square \)
The borderless process is constructed as follows. Our core configuration starts as before in [0, 1], and we use the same rejection/acceptance criteria for new points. However, we will now allow points to be generated to the right of 1 as well. Let \(R_t=R(x_{(N-1)}(t))\) where R is taken from Lemma 1. Then a new point is sampled uniformly and independently of the past on the interval \([0,R_t]\); formally, it is given by \(R_t U_t\) where \(U_t\) are i.i.d. uniform [0, 1] random variables independent of everything. Observe that if we consider the embedded process only at the times when the core configuration changes, then the exact form of the function \(R(\cdot )\) is irrelevant, due to the fact that the uniform distribution conditioned on a subinterval is also uniform on that subinterval.
Next, for \(y=\{y_1,\dots ,y_{N-1}\}\) define the function
where
We continue with the following
Lemma 2
For the borderless p-contest the sequence of random variables \(h\left( {{\mathcal {X}}}'(t)\right) \ge 0\), \(t=1,2,\dots \), is a supermartingale.
Remark 1
Note that the function \(F(\cdot )\) defined above is a Lyapunov function for the process in [3]; this is no longer the case as long as \(p\ne 1\); that is why we have to use a carefully chosen “correction” factor which involves the barycentre of the configuration.
Proof of Lemma 2
Assume that \(x_{(N-1)}(t)>0\) (otherwise the process has already stopped, and the result is trivial). The inequality, which we want to obtain is
for all \(y=(y_1,\dots ,y_{N-1})\) with \(y_i \in [0,1]\). Note that the function h(y) is homogeneous of degree 2 in y, therefore w.l.o.g. we can assume that \(\max y\equiv 1\).
For simplicity let \(M=N-1\ge 2\), and let
Note also that
Define
Thus we need to establish
First of all, observe that if \({\tilde{y}}=(y\setminus \{y_i\})\cup \{z\}\), that is, \({\tilde{y}}\) is obtained from y by replacing \(y_i\) with \(y_0\), then
In particular, if we replace point a by the new point z, then
and if we replace point 1, then
Note that both \(\Delta _a\) and \(\Delta _1\) depend only on four variables \((a,z,\mu ,M)\) but not the whole configuration. Let us also define
the p-centre of mass of the old core and the newly sampled point.
There are three different cases that can occur: either (a) the point a is removed, (b) 1, the rightmost point of the previous core, is removed, or (c) the newly sampled point z is removed. In the third case the core remains unchanged, and the change in the value of the function h is trivially zero. The point a can only be removed if \(z>a\); the point 1 can only be removed if \(z<1\); the point z can be possibly removed only if \(z\in (0,a)\) or \(z\in (1,\infty )\). Let us compute the critical values for z, for which there is a tie between the farthest points. \(\square \)
2.1 Which Point to Remove?
(i) Suppose \(\boxed {z<a}\). Then there is a tie between z and 1 if and only if \(m(z)=\frac{z+1}{2}\), that is if
Thus, we have:
when \(p<p_1\), point 1 is removed;
when \(p_1<p<p_2\), if \(z<t_{z1}\) then z is removed; if \(z>t_{z1}\) point 1 is removed;
when \(p>p_2\), point z is removed.
(ii) Suppose \(\boxed {a<z<1}\). There is a tie between a and 1 if and only if \(m(z)=\frac{a+1}{2}\), that is if
Thus, we have:
when \(p<p_3\), point 1 is removed;
when \(p_3<p<p_2\), if \(z<t_{a1}\) then 1 is removed; if \(z>t_{a1}\) then point a is removed;
when \(p>p_2\), point a is removed.
(iii) Suppose \(\boxed {z>1}\). There is a tie between z and a if and only if \(m(z)=\frac{z+a}{2}\), that is if
Thus, we have:
when \(p<p_3\), point z is removed;
when \(p>p_3\), if \(z<t_{za}\) then a is removed; if \(z>t_{za}\) then point z is removed.
We always have \(p_1<p_2\), \(p_3<p_2\) since
while
The final observation is that \(t_{za}<6\), so there is indeed no need to sample the new point outside of the range (0, 6); this holds since \(M\ge 2\) and
2.2 The Five Cases for the Removal
\(p<\min \{p_1,p_3\}\):
when \(z<1\), point 1 is removed
when \(z>1\), point z is removed
\(p>p_2\):
when \(z<a\) or \(z>t_{za}\in (1,\infty )\) point z is removed
when \(a<z<t_{za}\), point a is removed
\(\max \{p_1,p_3\}<p<p_2\)
when \(z<t_{z1}\in (0,a)\) or \(t>t_{za}\in (1,+\infty )\), point z is removed
when \(t_{z1}<z<t_{a1}\in (a,1)\), point 1 is removed
when \(t_{a1}<z<t_{za}\), point a is removed
\(p_1<p<p_3\ (<p_2)\):
when \(z<t_{z1}\in (0,a)\) or \(z>1\), point z is removed
when \(t_{z1}<z<1\), point 1 is removed
\(p_3<p<p_1\ (<p_2)\):
when \(z<t_{a1}\in (a,1)\), point 1 is removed
when \(t_{a1}<z<t_{za}\in (1,+\infty )\), point a is removed
when \(z>t_{za}\), point z is removed
Let
Define
Since \(\max y=1\), because of the comment on the restriction of the uniform distribution on a subinterval, we have \(\tilde{\mathbf{I}}_j=c_j \mathbf{I}_j\), \(j=1,2,3,4,5\), where \(c_j\)’s are some positive constants and
Thus to establish (2.5), it suffices to show that \(\mathbf{I}_j\le 0\) for each \(j=1,2,3,4,5\). This is done by very extensive and tedious calculations, which can be found in the Appendix. \(\square \)
We now return to our original p-contest process \({{\mathcal {X}}}(t)\). For \(L\ge 2\) define
note that \(\tau _L\) is a.s. finite for every L by Proposition 1. Let \(W(s)=\{w_1(s),\dots ,w_N(s)\}\) be a borderless p-contest with \(W(0)={{\mathcal {X}}}(\tau _L)\); let \(W'(s)\) be its core. By Lemma 2 the quantity \(\xi _t=h(W'(t\wedge \eta _L))\) is a supermartingale, that converges to some \(\xi _\infty \). Since \(\xi _t\) is bounded,
since on \(\{\eta _L<\infty \}\) we have \(\xi _\infty =W'( \eta _L)\) and the largest coordinate of \(W'( \eta _L)\) is larger than 1/2, implying that \(\mu (W'(\eta _L))\ge \frac{1}{2(N-1)}\) and thus \(h(W'(\eta _L))=F(W'(\eta _L))+k\mu (W'(\eta _L))^2 \ge \frac{k}{(2(N-1))^2}\). We also have
since \({{\mathcal {X}}}'(\tau _L)\subset [0,1/L]\) and so \(\mu ({{\mathcal {X}}}'(\tau _L))\in [0,1/L]\).
Combining the above inequalities, we conclude that \({\mathbb {P}}(\eta _L<\infty )\rightarrow 0\) as \(L\rightarrow \infty \). However, on \(\eta _L=\infty \) the core of the regular p-contest process can be trivially coupled with the core of the borderless process \(W'(s)\) which converges to zero, so \({{\mathcal {X}}}'(t)\rightarrow 0\) as well. Since \({\mathbb {P}}(\eta _L=\infty )\) can be made arbitrarily close to 1 by choosing a large L, we conclude that \({{\mathcal {X}}}'(t)\rightarrow 0\) a.s. \(\square \)
3 The Case \(p>1\)
Throughout this section we suppose that \(\zeta \) has a full support on [0, 1], and, unless explicitly stated otherwise, that \(p>1\).
Theorem 2
-
(a)
\( {\mathbb {P}}\left( \{{{\mathcal {X}}}'(t)\rightarrow 0\}\cup \{{{\mathcal {X}}}'(t)\rightarrow 1\} \right) =1; \)
-
(b)
if \(x_{(1)}(0)\ge 1/p\) then \({\mathbb {P}}({{\mathcal {X}}}'(t)\rightarrow 1)=1\);
-
(c)
if \(x_{(k)}(0)>0\), where k satisfies
$$\begin{aligned} \left\{ 2p(N-k)>N-2p \right\} \Longleftrightarrow \left\{ k<N-\frac{N}{2p}+1\right\} , \end{aligned}$$(3.1)then \( {\mathbb {P}}\left( {{\mathcal {X}}}'(t)\rightarrow 1 \right) >0.\)
Remark 2
In general, both convergences can have a positive probability. Let \(N=3\), \(p\in (1,3/2)\), and
where \(U\in U[0,1]\) (so \(\zeta \) has full support). Suppose we sample the points of \({{\mathcal {X}}}(0)\) from \(\zeta \). If they all start off in 0, then \(p\mu \le p/3 < 1/2\), so they cannot escape from 0. On the other hand, there is a positive probability they all start in (1/p, 1], and then Theorem 2(b) says that they converge to 1.
The key idea behind the proof of Theorem 2 is that one can actually find the “ruling” order statistic of the core; namely, there exists some non-random \(k=k(N,p)\in \{1,2,\dots , N-1\}\) such that \(x_{(k)}(t)\rightarrow 0\) implies \({{\mathcal {X}}}'(t)\overset{\text {a.s.}}{\longrightarrow }0\), while \(x_{(k)}(t)\not \rightarrow 0 \) implies that \({{\mathcal {X}}}'(t)\overset{\text {a.s.}}{\longrightarrow }1\).
We start with the following two results, which tells us that there is an absorbing area\([\frac{1}{p},1]\) for the process, such that, once the core enters this area, it will never leave it, and moreover the core will keep moving to the right.
Claim 1
Suppose that \(x_1\le x_2\le x_3\le \dots \le x_{N}\le 1\) and \(x_2\ge p^{-1}\). Then \(\{x_1,\cdots ,x_N\}'=\{x_2,\cdots ,x_N\}\)
Proof
Let \(\mu =\frac{x_1+\cdots x_N}{N}\). If \(p\mu \ge x_N\) then the claim follows immediately; assume instead that \(p\mu < x_N\). We need to check if \(p\mu -x_1>x_N-p\mu \), that is, if
However, since \(x_i\ge x_2\) for \(i=3,\dots , N-1\) we have
while \((N-2p)(x_1+x_N)\le 2(N-2p)<2(N-2).\) Hence (3.2) follows. \(\square \)
Lemma 3
If \(x_{(1)}(t_0)\ge 1/p\) for some \(t_0\), then \({{\mathcal {X}}}'(t)\rightarrow 1\) a.s.
Proof
If \(x_{(1)}(t_0) \ge 1/p\), then any point that lands in [0, 1/p) is extreme, so \(x_{(2)}(t)\ge 1/p\) for all \(t\ge t_0\). Choose any positive \(\varepsilon <1-\frac{1}{p}\), and let \(A_t=\left\{ \zeta _{t+1},\dots ,\zeta _{t+N-1}\in \left( 1-\varepsilon ,1\right] \right\} \). Then if \(A_t\) happens for \(s > t_0\), any point in \(\left[ 0, 1 - \varepsilon \right] \) is removed in preference to any of the new points coming in, so \(x_{(2)}(s+N-1) > 1 -\varepsilon \). As a result, by Claim 1 we get that \({{\mathcal {X}}}'(t)\in \left[ 0, 1 - \varepsilon \right] \) for all \(t\ge s\).
On the other hand, \({\mathbb {P}}(A_t)\ge [f(\varepsilon )]^{N-1}>0\) (see (2.1)) for any t, and the events \(A_t,A_{t+N},A_{t+2N},\dots \) are independent. Hence, eventually with probability 1, one of the \(A_t\)’s must happen for some \(t>t_0\), so a.s. \({{\mathcal {X}}}'(t)\in \left[ 0, 1-\varepsilon \right] \) for all large t. Since \(\varepsilon \) can be chosen arbitrary small, we get the result. \(\square \)
The next two results show that if the is some \(\varepsilon >0\) such that infinitely often the core does not have any points in \([0,\varepsilon )\), then it must, in fact, converge to 1.
Lemma 4
If \(x_{(1)}(t_0)\ge \varepsilon \) for some \(t_0\) and \(\varepsilon >0\), then \({\mathbb {P}}(x_{(1)}(t_0+\ell )\ge p^{-1} | {{\mathcal {F}}}_t)\ge \delta \) for some \(\ell =\ell (\varepsilon )\) and \(\delta =\delta (\varepsilon )>0\).
Proof
Suppose that for some t we have \(x_{(1)}(t)\ge \varepsilon \). We claim that it is possible to move \(x_{(1)}\) to the right of \(\frac{1+p}{2}\varepsilon \) in at most \(N-1\) steps with positive probability, depending only on p and \(\varepsilon \). Indeed, if \(x_{(1)}(t)>\frac{1+p}{2}\varepsilon \) then we are already done. Otherwise, if the new point \(\zeta _{t+1}\) is sampled in \(\left( \frac{1+p}{2} \varepsilon ,p\varepsilon \right] \subset [0,1]\) it cannot be rejected. If at this stage \(x_{(1)}(t+1)>\frac{1+p}{2}\varepsilon \), then we are done. If not, we proceed again by sampling \(\zeta _{t+2} \in \left( \frac{1+p}{2}\varepsilon ,p\varepsilon \right] \), etc. After at most \(N-1\) steps of sampling new points in \(\left( \frac{1+p}{2}\varepsilon ,p\varepsilon \right] \), the leftmost point \(x_{(1)}\) will have moved to the right of \(\frac{1+p}{2}\varepsilon \).
Thus, in no more than \(N-1\) steps, with probability no less than \(\left[ f\left( \frac{p-1}{2}\varepsilon \right) \right] ^{N-1}>0\), \(x_{(1)}\) is to the right of \(\frac{1+p}{2}\varepsilon \). By iterating this argument at most m times, where \(m\in {\mathbb {N}}\) is chosen such that \(\left[ \frac{1+p}{2}\right] ^m \, \varepsilon >1/p\), we achieve that \(x_{(1)}\) is to the right of 1/p (for definiteness, one can chose \(\ell =(N-1)m\) and \(\delta = \left[ f\left( \frac{p-1}{2}\varepsilon \right) \right] ^{(N-1)m}\).) \(\square \)
Lemma 5
Let \(\varepsilon \in (0,1)\), and define \(B(\varepsilon ):= \{x_{(1)}(t)\ge \varepsilon \text { i.o.}\}\) Then \( B(\varepsilon ) \underset{\text {a.s.}}{\Longrightarrow }\{ {{\mathcal {X}}}'(t)\rightarrow 1\}. \)
Corollary 1
We have \(\left\{ \liminf _{t\rightarrow \infty } x_{(1)}(t)>0\right\} \underset{\text {a.s.}}{\Longrightarrow }\{ {{\mathcal {X}}}'(t)\rightarrow 1\}.\)
Proof of Lemma 5
Assume that \(\varepsilon <\frac{1}{p}\) (otherwise the result immediately follows from Lemma 3). Also suppose that \({\mathbb {P}}\left( B(\varepsilon )\right) >0\), since otherwise the result is trivial. Let \(\ell \) and \(\delta \) be the quantities from Lemma 4.
Define
with the convention that if \(\tau _k=\infty \) then \(\tau _m=\infty \) for all \(m>k\). Notice that \(B(\varepsilon )=\bigcap _{k=0}^\infty \{\tau _k<\infty \}\). On \(B(\varepsilon )\) we can also define \(D_{\tau _k}=\{x_{(1)}(\tau _k+\ell )\ge 1/p\}\). Since \(\tau _k-\tau _{k-1}>\ell \) whenever both are finite, we have from Lemma 4 we have \({\mathbb {P}}(D_{\tau _{k+1}} | {{\mathcal {F}}}_{\tau _k}) \ge \delta \). Therefore,
hence by Lévy’s extension of the Borel–Cantelli lemma it follows that a.s. on \(B(\varepsilon ))\) infinitely many (and hence at least one) of \(D_{\tau _k}\) occur, that is, \(x_{(1)}(\tau _k+\ell )\ge 1/p\). Now the result follows from Lemma 3. \(\square \)
Assume for now that \(p<\frac{N}{2}\); in this case \(N-\frac{N}{2p}+1<N\) (see (3.1)). The case \(p\ge \frac{N}{2}\) will be dealt with separately.
The following statement shows that if all the points to the right of \(x_{(k)}\) lie very near each other, while the left-most one lies near zero, then it is to be removed.
Claim 2
Let \(a\in (0,1]\) and suppose that \(k\in \{2,\dots ,N-1\}\) satisfies (3.1). Then there exist small \(\delta ,\Delta >0\), depending on N, k, p, a such that if
then \(\{x_1,\dots ,x_{N}\}'=\{x_2,\dots ,x_N\}\).
Proof
The condition to remove the leftmost point is \(p\mu -\frac{x_1+x_N}{2}> 0\) where \(\mu =(x_1+\dots +x_N)/N\). However,
The RHS is linear in \(\delta \) and \(\Delta \), and when \(\delta =\Delta =0\) it is strictly positive by the assumption on k; hence it can also be made positive, by allowing \(\delta >0\) and \(\Delta >0\) to be sufficiently small. \(\square \)
Corollary 2
Suppose that \({{\mathcal {X}}}(t)=\{x_1,\dots ,x_N\}\) satisfies the conditions of Claim 2 for some a and k. Let \(\delta \) be the quantity from this claim. Then
Proof
The probability to sample a new point \(\zeta \in (a(1-\Delta ),a]\) is bounded below by \(f(a\Delta )\) where f is the same function as in (2.1). On the other hand, if the new point is sampled in \((a(1-\Delta ),a]\) then \({{\mathcal {X}}}(t+1)\) continues to satisfy the conditions of Claim 2 as long as the leftmost point is in \([0,\delta ]\). By repeating this argument at most k times and using the induction, we get the result with \(c=\left[ f(a\Delta )\right] ^k>0\). \(\square \)
Lemma 6
Let \(k\in {\mathbb {N}}\) satisfy (3.1). Then
Proof
Note that by Lemma 5, it suffices to show that \(\left\{ x_{(k)}(t)\not \rightarrow 0 \right\} \underset{\text {a.s.}}{\Longrightarrow }\left\{ x_{(1)}(t)\not \rightarrow 0\right\} \).
If \(x_{(k)}(t)\not \rightarrow 0\), there exists an \(a>0\) such that \(x_{(k)}(t)\ge a\) for infinitely many t’s. Let s be such a time. Now suppose that \(\zeta _{s+i}\in I:=(a(1-\Delta ),a]\) for \(i=0,1,\dots , N-1\) where \(\Delta \) is defined in Claim 2; the probability of this event is strictly positive and depends only on a and \(\delta \) (see (2.1)). As long as there are points of \({{\mathcal {X}}}(s+i)\) on both sides of the interval I, none of the points inside I can be removed; hence, for some \(u\in \{s,s+1,\dots ,s+N-1\}\) we have that either \(\min {{\mathcal {X}}}(u)> a(1-\Delta )\) or \(\max {{\mathcal {X}}}(u)\le a\). In the first case, \(x_{(1)}(u)>a(1-\Delta )\).
In the latter case, both \(x_{(N)}(u)\in I\) and \(x_{(k)}(u)\in I\), since every time we replaced a point, the number of points to the left of I did not increase (and there were initially at most \(k-1\) of them). As a result
Together with Corollary 2, this yields
which proves Lemma 6. \(\square \)
Claim 3
Let \(A_i:=\left\{ x_{(i)}(t)\rightarrow 0\right\} \) and suppose that for some \(1\le k\le N-2\) we have
Then \( A_k\subseteq \{\exists \lim _{t\rightarrow \infty } x_{(k+1)}(t)\}. \)
Proof
Fix any \(a>0\). Let \(\delta >0\) be so small that
In the event \(A_k\) there exists a finite \(\tau =\tau _\delta (\omega )\) such that
From now on assume that \(t\ge \tau \). We will show below that \(x_{(k+1)}(t+1)\le \max \{x_{(k+1)}(t),a\}\).
To begin, let us prove that \(x_{(k+1)}(t+1)\le x_{(k+1)}(t)\) as long as \(x_{(k+1)}(t)>\delta \). Indeed, if the new point \(\zeta \) is sampled to the left of \(x_{(k+1)}(t)\), then regardless of which point is to be removed, \(x_{(k+1)}(t+1)\le x_{(k+1)}(t)\). If the new point \(\zeta \) is sampled to the right, then the farthest point from the \(p-\)centre of mass must be the rightmost one (and hence \(x_{(k+1)}(t+1)=x_{(k+1)}(t)\)) since there are exactly k points in \([0,\delta ]\) and none of these can be removed by the definition of \(\tau \).
On the other hand, if \(x_{(k+1)}(t)\le \delta \) then either \(x_{(k+2)}(t)\le a\) or \(x_{(k+2)}(t)> a\). In the first case, \(x_{(k+1)}(t+1)\le x_{(k+2)}(t)\le a\) even if \(x_{(1)}\) is removed. In the other case, when \(x_{(k+2)}(t)> a\), we have \(x_{(N-1)}>a\) as well, and
by (3.4), so \(x_{(N)}=x_{(N)}(t)\) must be removed and thus \(x_{(k+1)}(t+1)\le x_{(k+1)}(t)\).
Consequently, we obtained
since \(a>0\) is arbitrary, we get
\(\square \)
Lemma 7
Suppose that (3.3) holds for some \(1\le k\le N-2\). Then \(A_k \underset{\text {a.s.}}{\Longrightarrow }A_{k+1}\).
Proof
Let \({{\tilde{A}}}_{k+1}^{\ge a}:=\left\{ \lim _{t\rightarrow \infty } x_{(k+1)}(t)\ge a\right\} \) (the existence of this limit on \(A_k\) follows from Claim 3). It suffices to show that \({\mathbb {P}}\left( A_k \cap {\tilde{A}}_{k+1}^{\ge a}\right) =0\) for all \(a>0\); then from the continuity of probability we get that \({\mathbb {P}}\left( A_k \cap \{\lim _{t\rightarrow \infty } x_{(k+1)}(t)>0\}\right) =0\) and hence \(A_k\underset{\text {a.s.}}{\Longrightarrow }A_{k+1}\).
Fix an \(a>0\). Let
then
If the probability of the LHS is positive, then, using the continuity of probability and the fact that \({\bar{C}}_T\) is an increasing sequence of events, we obtain that \(\lim _{T\rightarrow \infty } {\mathbb {P}}({\bar{C}}_T)>0\). Consequently, there exists a non-random\(T_0\) such that \({\mathbb {P}}({\bar{C}}_{T_0})>0\).
This is, however, impossible, as at each time point t, with probability at least f(a/3) (see (2.1)) the new point \(\zeta _t\) is sampled in \(B:=\left( \frac{a}{3},\frac{2a}{3}\right) \) and then either \(x_{(k)}(t+1)\in B\) or \(x_{(k+1)}(t+1)\in B\). Formally, this means that
By induction, for all \(k\ge 1\),
Since k is arbitrary, and \(f(a/3)>0\), by taking the expectation, we conclude that \({\mathbb {P}}({\bar{C}}_{T_0})=0\) yielding a contradiction.
Hence the probability of the event \(A_k\cap {\tilde{A}}_{k+1}^{\ge a}\) is zero. \(\square \)
Corollary 3
Suppose that (3.3) holds for some \(1\le k\le N-2\). Then
Proof
Observe that if k satisfies (3.3) then \(k+1\) satisfies (3.3) as well. Thus by iterating Lemma 7 we obtain that \(A_k\underset{\text {a.s.}}{\Longrightarrow }A_{k+1}\underset{\text {a.s.}}{\Longrightarrow }A_{k+2}\underset{\text {a.s.}}{\Longrightarrow }\dots \underset{\text {a.s.}}{\Longrightarrow }A_{N-1}\), i.e. \(x_{(N-1)}(t)\rightarrow 0\), which is equivalent to the statement of Corollary. \(\square \)
Remark 3
Note that the condition (3.3) does not assume \(p>1\); hence the conclusion of Corollary 3 holds for the case \(0<p\le 1\) as well.
For the case \(p\ge \frac{N}{2}\) we have
Lemma 8
If \(p\ge \frac{N}{2}\) then \({{\mathcal {X}}}'(t)\rightarrow 1\) a.s.
Proof
The case \(p>\frac{N}{2}\) is easy: with a positive probability the newly sampled point \(\zeta >0\) and then
hence it is the left-most point which is always removed, implying \(\liminf _{t\rightarrow \infty } x_{(1)}(t)>0\). Hence by Corollary 1, \({{\mathcal {X}}}'(t)\rightarrow 1\) a.s.
For the case \(p=\frac{N}{2}\) we notice that at each moment of time we either have a tie (between the left-most and right-most point) or remove the left-most point. However, we can only have a tie if \(x_{(1)}(t)=\ldots =x_{(N-1)}(t)=0\); in this case, eventually the right-most point will be kept and the left-most removed. After this moment of time, there will be more ties, and the left-most point will always be removed, leading to the same conclusion as in the case \(p>N/2\). \(\square \)
Proof of Theorem 2
Part (b) follows from Lemma 3.
To prove part (c), note that unless \(x_{(1)}(0)>0\) already, by repeating the arguments from the beginning of the proof of Lemma 6, with a positive probability we can “drag” the whole configuration in at most \(N-1\) steps to the right of zero, that is, there is \(0\le t_0\le N-1\) such that \({\mathbb {P}}(\min {{\mathcal {X}}}'(t_0)>0)>0\). Now we can apply Lemma 4 and then Lemma 3.
Let us now prove part (a). First, assume \(p<\frac{N}{2}\). It is always possible to find an integer k which satisfies both (3.1) and (3.3), so let k be such that
(if \(N/(2p)\in {\mathbb {N}}\) this k will be unique). Now the statement of the theorem follows from Corollary 3 and Lemma 6.
Finally, in case \(p\ge \frac{N}{2}\) the theorem follows from Lemma 8. \(\square \)
4 Non-convergence to Zero for \(p>1\) and \(N=3\)
In this section we prove the following
Theorem 3
Suppose that \(N=3\), \(p>1\) and \(\zeta \), restricted to some neighbourhood of zero, is a continuous random variable with a non-decreasing density (e.g. uniformly distributed). Then \({{\mathcal {X}}}'(t)\rightarrow 1\) as \(t\rightarrow \infty \) a.s.
Remark 4
-
In case \(p\ge 3/2\) we already know that \({{\mathcal {X}}}'(t)\rightarrow 1\) for any initial configuration and any distribution (see Lemma 8), so we have to prove the theorem only for \(p\in (1,3/2)\).
-
Simulations suggest that the statement of Theorem 3 holds, in fact, for a much more general class of distributions \(\zeta \).
Let \(\varepsilon \in (0,1/2)\) be such that \(\zeta \) conditionedFootnote 2 on \(\{\zeta \le 2\varepsilon \}\) has a non-decreasing density; according to the statement of the Theorem 3 such an \(\varepsilon \) must exist. Let us fix this \(\varepsilon \) from now on.
The idea of the proof will be based on finding a non-negative function \(h:(0,1]^2\rightarrow {\mathbb {R}}_+\) which has the following three properties:
- (i)
\(h(\cdot ,\cdot )\) is non-increasing in each of its arguments;
- (ii)
\(h\left( x_{(1)}(t),x_{(2)}(t)\right) \) is a supermartingale as long as \(x_{(2)}(t)\le \varepsilon \);
- (iii)
\(h(\cdot ,\cdot )\) goes to infinity when the first coordinate goes to zero.
From the supermartingale convergence theorem it would then follow that
Let us formally prepare for the proof of Theorem 3. As before, denote by \(x_1,\dots ,x_N\)N distinct points on [0, 1], and let \(x_{(1)},\dots ,x_{(N)}\) be this unordered N-tuple sorted in the increasing order. Let
be the unordered N-tuple \(\{x_1,\dots ,x_N\}\) with the farthest point from p-centre of mass removed; w.l.o.g. assume that \(y_i\) are already in the increasing order.
Lemma 9
The operation \(\{\dots \}'_p\) is monotone in p, that is, if \({\hat{p}}\ge {\tilde{p}}\) and
then \({\hat{y}}_i\ge {\tilde{y}}_i\), \(i=1,\dots ,N-1\).
Proof
Assume w.l.o.g. \(x_1\le \ldots \le x_N\), and let \(\mu = \mu \left( \{x_1,\dots ,x_N\}\right) \). Notice that, regardless of the value of p, the only points which can possibly be removed are \(x_1\) or \(x_N\) (since they are the two extreme points). Therefore, it suffices to show that \(\{x_1,\dots ,x_N\}'_{\tilde{p}}=\{x_2,\dots ,x_N\}\) implies \(\{x_1,\dots ,x_N\}'_{{\hat{p}}}=\{x_2,\dots ,x_N\}\). Note also that \(|x_1-p\mu |=p\mu -x_1\) for all \(p\ge 1\).
If \({\tilde{p}}\mu -x_1>|{\tilde{p}}\mu -x_N|\) and \({\tilde{p}}\mu -x_N>0\), that is, the p-centre of mass lies to the right of \(x_N\), then \({\hat{p}}\mu>{\tilde{p}}\mu >x_N\) as well, and hence \(x_1\) is discarded.
On the other hand, if \({\tilde{p}}\mu -x_1>|{\tilde{p}}\mu -x_N|\) and \({\tilde{p}}\mu <x_N\) then either \({\hat{p}}\mu <x_N\), or \({\hat{p}}\mu \ge x_N\). In the first case,
so \(x_1\) is discarded. In the second case, p-centre of mass lies to the right of \(x_N\) and so \(x_1\) is also discarded. \(\square \)
Lemma 10
Let h be a real-valued function on the sets of N real numbers. Suppose that h is non-increasing in each of its arguments, namely
whenever \(x_i'\ge x_i\). Let \({{\mathcal {E}}}_t\) be some \({{\mathcal {F}}}_t\)-measurable event, and suppose that
for \(p=1\). Then (4.1) holds for \(p>1\) as well.
Proof
Let
be the new core after the new point \(\zeta _{t+1}\) is sampled and the farthest point from the p-centre of mass is removed; note that \({{\mathcal {X}}}'(t+1)=G_p({{\mathcal {X}}}'(t),\zeta _{t+1})\). Then on \({{\mathcal {E}}}_t\)
since the operation \(\{\dots \}'_p\) is monotone in p by Lemma 9 and h is decreasing in each argument. \(\square \)
From now on assume \(N=3\) and \(p=1\). Denote \(x_{(1)}(t)=a\), \(x_{(2)}(t)=b\) and consider the events
(we assume that b is smaller than 1/2, yielding \(2b-a<1\).) If \(x_{(2)}(t)\le \varepsilon \) then \({{\mathcal {X}}}'(t+1)\not ={{\mathcal {X}}}'(t)\) implies that one of the events \(L_b\), \(B_b\), \(B_a\) or \(R_a\) occurs (i.e. all points sampled outside of \(\left( (2a-b)^+,2b-a \right) \) are rejected at time \(t+1\)). Let us study the core \({{\mathcal {X}}}'(t+1)=\{\zeta _{t+1},a,b\}'\) on these events: on \(L_b\) and \(B_b\) we have \({{\mathcal {X}}}'(t+1)=\{x,a\}\), while on \(B_a\) and \(R_a\) we have \({{\mathcal {X}}}'(t+1)=\{x,b\}\).
We have, assuming \(x_{(1)}(t)=a\) and \(x_{(2)}(t)=b\),
When \(0\le a\le b\le \varepsilon \) we have \(2b-a\le 2\varepsilon \). Define
which is positive in the first two cases, and negative in the next two. Let \(\varphi (x)\) be the density of \(\zeta \) conditioned on \(\{\zeta \in [0,2\varepsilon ]\}\). By the monotonicity of \(\varphi \) and the positivity (negativity resp.) of g on the first (second resp.) interval,
where
So if we can establish that \(\Lambda \le 0\) for a suitable function h, then indeed \(\Delta (a,b)\le 0\), and the supermartingale property follows.
Remark 5
Notice that the method of proof, presented here, could possibly work for \(N>3\) as well; that is, if one can find a function \(h(x_1,\dots ,x_{N-1})\), which is positive and decreasing in each of its arguments, and \(h({{\mathcal {X}}}'(t))\) is a supermartingale provided \(\max {{\mathcal {X}}}'(t)<\varepsilon \) for some \(\varepsilon >0\). Unfortunately, however, we were not able to find such a function.
Set
it is easy to check h is indeed monotone in each of its arguments as long as \(x,y\in (0,1]\). Let us now compute the integrals in the expression for \(\Lambda \). We have
It turns out that \(h({{\mathcal {X}}}'(t))\) indeed has a non-positive drift, provided \(0<a\le b\le \varepsilon \), as is shown by the following
Lemma 11
\(\Lambda \le 0\) for \(a,b\in (0,1/2]\).
Proof
Substitute \(a=b\nu \) in the expression for \(\Lambda \). Then for \(\nu \le 1/3\) we easily obtain \( \Lambda =-b\left[ 3\nu (1-\ln 2) + \ln 8 -2\right] \le 0. \)
For \(1/3<\nu \le 1/2\) we have \(2\Lambda =-b C_1(\nu )\le 0\) where
since \(\frac{\partial ^2 C_1(\nu )}{\partial ^2 \nu }=-\frac{1}{\nu ^2(1+\nu )}<0\) and hence \(\min _{1/3\le \nu \le 1/2}C_1(\nu )\) is achieved at one of the endpoints \(\nu =1/3\) or \(\nu =1/2\); the values there are \(C_1(1/3)=\ln (4)-1>0\) and \(C(1/2)=\frac{1}{2} \ln \left( \frac{512}{27}\right) -1>0\) respectively.
For \(1/2<\nu \le 2/3\) we have \(\Lambda =-b C_2(\nu )\le 0\) where
since \(\frac{\partial ^2 C_2(\nu )}{\partial ^2 \nu }=\frac{2\nu ^2+6\nu +3}{\nu ^2(1+\nu )}>0\) and \(\left. \frac{\partial C_2(\nu )}{\partial \nu }\right| _{\nu =2/3}=\ln \left( \frac{256}{45}\right) -\frac{5}{2}<0\) implies that \(\frac{\partial C_2(\nu )}{\partial \nu }<0\) for all \(\nu \in [1/2,2/3]\) and hence \(\min _{1/2\le \nu \le 2/3} C_2(\nu )=C_2(2/3)=\frac{1}{3}\ln \left( \frac{104976}{3125}\right) -1>0.\)
Finally, for \(2/3<\nu \le 1\) we have \(\Lambda =-bC_3(\nu )\le 0\), where
since
changes its sign from − to \(+\) at \(1/\sqrt{2}\in (2/3,1)\) and therefore \(\frac{\partial C_3(\nu )}{\partial \nu }\) achieves its maximum at the endpoints of the interval; thus
Therefore, \(C_3(\nu )\) is decreasing and hence \(\min _{2/3\le \nu \le 1} C_3(\nu )=C_3(1)=0\). \(\square \)
Proof of Theorem 3
We will show that \({\mathbb {P}}({{\mathcal {X}}}'(t)\rightarrow 0)=0\), which will imply by Theorem 2(a) that \({\mathbb {P}}({{\mathcal {X}}}'(t)\rightarrow 1)=1\); we shall do it by showing that
Indeed, fix some \(\varepsilon \in (0,1/2)\) and let \(\tau _0=0\). For \(\ell =1,2,\dots ,\) define the sequence of stopping times
so that \(\tau _0<\eta _1<\tau _1<\eta _2<\tau _2<\dots \) with the conventions that if one of the stopping times is infinite, so is the rest of them. Define also \(\ell ^*=\inf \{\ell \ge 1:\ \tau _\ell =+\infty \}\).
If \(\ell ^*=\infty \), that is, \(\tau _\ell <\infty \) for all \(\ell \), we immediately have \(\limsup _{t\rightarrow \infty } x_{(2)}(t)\ge \varepsilon \) and we are done; so assume that for some \(\ell ^*\ge 1\) we have \(\tau _{\ell ^*-1}<\infty =\tau _{\ell ^*}\). If \(\eta _{\ell ^*}=\infty \), then \(x_{(2)}(t)>\varepsilon \) for all \(t\ge \tau _{\ell ^*-1}\) and thus again \(\limsup _{t\rightarrow \infty } x_{(2)}(t)\ge \varepsilon \); hence \(\ell ^*<\infty \) and \(\eta _{\ell ^*}<\infty \) on the event \(\left\{ \limsup _{t\rightarrow \infty } x_{(2)}(t)<\varepsilon \right\} \).
On the other hand, as long as \(\eta _\ell <\infty \), we can define
where h is given by (4.2).
hence \(\tilde{\xi }^{(\ell )}_s\), the process \(\xi ^{(\ell )}\) stopped at the time when \(x_{(2)}\) exceeds \(\varepsilon \), is a non-negative supermartingale, hence it must converge to a finite value. In case \(\tau _\ell =+\infty \) this means that \(\liminf _{t\rightarrow \infty } x_{(1)}(t)>0\) since the function h(a, b) goes to infinity when \(a\downarrow 0\). Thus we have established (4.3). \(\square \)
Notes
Rather than of \({{\mathcal {X}}}(t)\)
Note that the full support assumption ensures that the probability of this event is positive.
References
De Giorgi, E., Reimann, S.: The \(\alpha \)-beauty contest: choosing numbers, thinking intervals. Games Econ. Behav. 64, 470–486 (2008)
Grinfeld, M., Knight, P.A., Wade, A.R.: Rank-driven Markov processes. J. Stat. Phys. 146, 378–407 (2012)
Grinfeld, M., Volkov, S., Wade, A.R.: Convergence in a multidimensional randomized Keynesian beauty contest. Adv. Appl. Probab. 47, 57–82 (2015)
Kennerberg, P., Volkov, S.: Jante’s law process. Adv. Appl. Probab. 50, 414–439 (2018)
Keynes, J.M.: The General Theory of Employment. Interest and Money. Macmillan, London (1936)
Moulin, H.: Game Theory for the Social Sciences, 2nd edn. New York University Press, New York (1986)
Seel, C., Tsakas, E.: Rationalizability and Nash equilibria in guessing games. Games Econ. Behav. 106, 75–88 (2017)
Williams, D.: Probability with Martingales. Cambridge University Press, Cambridge (1991)
Acknowledgements
Open access funding provided by Lund University. We would like to thank the anonymous referees for very careful reading our manuscript and for giving many useful suggestions and corrections.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Eric A. Carlen.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The research is partially supported by the Swedish Research Council Grant VR 2014-5157 and by Crafoord foundation Grant 20190667.
Appendix: The Calculations for the Proof of Lemma 2.
Appendix: The Calculations for the Proof of Lemma 2.
Observe that all expressions for \(\mathbf{A}_j\) are fractions of the polynomials in (a, f, p, M); moreover, their denominators
are always positive. Throughout the rest of the proof let \(\mathsf{n}(w)\) denote the numerator of such a fraction w.
1.1 Case 1: \(\mathbf{I}_1\le 0\)
Observe that
and the term in the square brackets is positive as \(M\mu \ge 1\), so the maximum of \(\mathsf{n}(\mathbf{A}_1)\) is achieved at the highest possible value of p. However, in this case we have \(p\le p_1\), hence
where
Hence \(s_1\ge 0\) for \(M=2,3,\dots \) and thus \(\mathbf{I}_1\le 0\).
1.1.1 Case 2: \(\mathbf{I}_2\le 0\)
Here
where
and we need to show that \(s_2\ge 0\).
Assume first \(M=2\). Then (using the fact that \(\mu =(1+a)/2\))
which is impossible; so from now on \(M\ge 3\).
To establish \(\mathbf{I}_2\le 0\), it will suffice to demonstrate that
as \(\mathbf{I}_2\) has a factor \(1_{X_2\ge 0}\), and \(s_2 1_{X_2\ge 0} \ge \frac{s_3}{2M} 1_{X_2\ge 0}\). Substituting
where \(w\in [0,1)\) corresponding to the condition (2.4), we get
The expression in the square brakets is non-negative for \(M\ge 3\), so the minimum of \(s_3\) is achieved when \(a=1\); i.e.
But
so
1.2 Case 3: \(\mathbf{I}_3\le 0\)
Here
and it suffices to show that \(s_5\ge 0\). If \(M=2\), then \(\mu =(a+1)/2\) and \(p\ge 3/4\), so
For \(M\ge 3\), let \(M=3+\delta \), \(\delta =0,1,\dots \). Then \( s_5=\sum _{i=0}^5 e_{i+1} \delta ^i \) where we will show that all \(e_i\ge 0\). Indeed, we have
The fact that \(e_6\ge 0\) is trivial; we will prove separately that \(e_1,\dots ,e_5\ge 0\) below. In what follows, we substitute \(p=\frac{1+\nu }{2}\), where \(\nu \in (0,1)\).
1.3 Proof that \(e_1\ge 0\)
We have
hence \(e_1\) achieves its minimum at
which solves \(\frac{\partial e_1}{\partial a}=0\). Note that it is possible that \(a_{cr}\not \in [0,1]\). However, in any case,
so it will suffice to show that
is positive. We have
Similarly,
where
and
So, \(c_{1b1},c_{1b2}>0\)\(\Longrightarrow \)\(c_{1b}>0\) and since \(c_{1a}>0\) we have \(c_1\ge 0\) and thus \(e_1\ge 0\).
1.4 Proof that \(e_2\ge 0\)
We have
so, similarly to the previous case,
where
solves \(\frac{\partial e_2}{\partial a}=0\) and
Now,
hence the minimum of \(c_2\) w.r.t. \(\mu \in [0,1]\) can be achieved either at \(\mu =0\) or at \(\mu =1\). At the same time
so \(c_2\ge 0\) and hence \(e_2\ge 0\).
1.5 Proof that \(e_3\ge 0\)
We have
so, similarly to the previous case,
where
solves \(\frac{\partial e_3}{\partial a}=0\) and
Hence \(e_3\ge 0\).
1.6 Proof that \(e_4\ge 0\)
We have
so, similarly to the previous case,
where
solves \(\frac{\partial e_4}{\partial a}=0\) and
Hence \(e_4\ge 0\).
1.7 Proof that \(e_5\ge 0\)
We have
so, similarly to the previous case,
where
solves \(\frac{\partial e_5}{\partial a}=0\) and
Hence \(e_5\ge 0\). As a result, \(s_5\ge 0\) and thus \(\mathbf{I}_3\le 0\).
1.8 Case 4: \(\mathbf{I}_4\le 0\)
Here
Then, substituting \(M=2+\delta \),
and as a result for \(\delta \ge 0\) we have
1.9 Case 5: \(\mathbf{I}_5\le 0\)
Here
We need to show that \(s_7\ge 0\) when \(X_1\le 0\) and \(X_3\ge 0\).
Since \(X_1\le 0\), we have \(2Mp\mu \le M+1\). Together with \(X_3\ge 0\) this implies
whence
Let us show that for this a we have \(s_7\ge 0\); substitute \(a=b\cdot \frac{2p}{M+1}\), where \(b\in [0,1]\).
First, let \(\boxed {M=2}\) , then \(\mu =\frac{1+a}{2}\), \(p\in [3/4,1)\), and \(s_7=\frac{3-2p}{27} s_8\) where
Note that we can write \(s_8=e_1+e_2 (1-p) +e_3 (1-p)^2\), where
with \(p=\frac{3+\nu }{4}\), \(\nu \in [0,1]\). Consequently, since \((1-p)^2<1\) and \(e_1>0\),
and thus \(s_7\ge 0\) as required.
For \(\boxed {M\ge 3}\) , set \(M=3+\delta \), \(\delta \ge 0\). Then
where
and the expressions for \(e_9\) and \(e_{10}\) are given a little bit further.
First, we will show that \(e_i\ge 0\), \(i=1,\dots ,8\).
1.10 Proof that \(e_1,\dots , e_8> 0\)
It turns out that it is easiest is to use a computer-assisted proof in this case; to this end we developed the method which we call a Box method; it may have been described by other authors, but since we do not have the reference to the right source, we give its description below.
First of all, we substitute
Let \(\mathrm{m}=\min _{a_i\le x_i\le b_i,i=1,2,3} f(x_1,x_2,x_3)\) where
and \(f_+\) and \(f_-\) are polynomials with non-negative coefficients. We want to show that \(\mathrm{m}>0\).
Let
Since
as \(M\rightarrow \infty \), we conclude that \(\mathrm{m}>0\) if and only if \(G_{f,M}\ge 0\) for some \(M\ge 1\). Checking that \(G_{f,M}\ge 0\) can be quite tedious and time-consuming for large M, however, this could be easily accomplished with the help of a computer; please note, that the results are still completely rigorous, unlike e.g. simulations.
The results of application of this method to \(e_1,\dots ,e_8\) are presented in the following table:
Consequently, \(e_j>0\) for all \(j=1,\dots ,8\).
1.11 Proof that \(e_9\ge 0\) and \(e_{10}\ge 0\)
The Box method of the previous section would not work for \(e_9\) and \(e_{10}\), since these functions do touch zero in the required area, and hence the minimum is, in fact, 0. Therefore, we have to handle these two cases analytically.
We have
hence, the minimum is achieved either at \(b=0\) or \(b=1\).
For \(\mu <1/(2p)\) we have \(e_9\ge e_{9a}\), where
where \(s=2p\mu \in [0,1]\).
In case \(\mu \ge 1/(2p)\) we have \(e_9\ge e_{9b}\), where
where \(\mu =\frac{1}{2p}+s\left( 1-\frac{1}{2p}\right) \), \(s\in [0,1]\). Now,
so the minimum of \(e_{9b}\) w.r.t. s is achieved where \(\frac{\partial }{\partial s}e_{9b}=0\), i.e.
and equals
for \(p\ge 1/2\).
Finally, trivially, we have \(e_{10}=3p(2\mu p-1)^2\ge 0.\) Consequently, \(s_7\ge 0\) and \(\mathbf{I}_5\le 0\).
Combining this with the previously established inequalities \(\mathbf{I}_j\le 0\), \(j=1,2,3,4\), we complete the proof Lemma 2.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kennerberg, P., Volkov, S. Convergence in the p-Contest. J Stat Phys 178, 1096–1125 (2020). https://doi.org/10.1007/s10955-020-02491-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10955-020-02491-6