Abstract
Using recent results concerning non-uniqueness of the center of the mix for completely mixable probability distributions, we obtain the following result: For each\(d\in {\mathbb {N}}\) and each non-empty bounded Borel set \(B\subset {\mathbb {R}}^d\), there exists a d-dimensional probability distribution \(\varvec{\mu }\) satisfying the following: For each \(n\ge 3\) and each probability distribution \(\varvec{\nu }\) on B, there exist d-dimensional random vectors \({\textbf{X}}_{\varvec{\nu },1},{\textbf{X}}_{\varvec{\nu },2},\dots ,{\textbf{X}}_{\varvec{\nu },n}\) such that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \varvec{\nu }\) and \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) for \(i=1,2,\dots ,n\). We also show that the assumption regarding the boundedness of the set B cannot be completely omitted, but it can be substantially weakened.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Let \(\mu \) be a probability distribution on \({\mathbb {R}}\) and \(n\in {\mathbb {N}}\). We say that \(\mu \) is n-completely mixable if there exists a random vector \(X=(X_1,X_2,\dots ,X_n)\) such that for each \(i=1,2,\dots ,n\) the random variable \(X_i\sim \mu \) (i.e., \(X_i\) has distribution \(\mu \)) and \(\sum _{i=1}^n X_i\) is a.s. constant. If such a vector X exists, then it is called a complete mix (or a mix for short). The number \(C=\frac{1}{n}\sum _{i=1}^n X_i\) is called the center of the mix.
The problem of n-complete mixability is mostly theoretical in nature but it became popular because of its connections with applications, especially with risk theory. For more details, see, e.g., [1, 2, 4, 6].
Obviously, only one-point probability distributions are 1-completely mixable and each symmetric distribution is 2-completely mixable (the symmetry is the necessary and sufficient condition for 2-complete mixability). However, for \(n\ge 3\) the problem of characterizing n-complete mixability is only partially solved and is still an object of thorough research (see [9] for an exhaustive list of references). In particular, the following sufficient conditions for n-complete mixability have been established:
If \(\mu \) has a symmetric and unimodal density, then \(\mu \) is n-completely mixable for all \(n\ge 2\) ( [7]).
If \(\mu \) has a density, which is monotone on its support \([a,b]\subset {\mathbb {R}}\), then \(\mu \) is n-completely mixable if and only if the expected value \(E(\mu )\) is in \([a+\frac{b-a}{n},b-\frac{b-a}{n}]\) ([8]).
If \(\mu \) has a density, which is concave on its support, then \(\mu \) is n-completely mixable for all \(n\ge 3\) ( [5]).
If \(\mu \) has a density f on a finite interval [a, b] and \(f(x)\ge \frac{3}{n(b-a)}\), then \(\mu \) is n-completely mixable ([6]).
For a long time, it was not known whether the center of the mix is always unique for an n-completely mixable probability distribution (for \(n\ge 3\)). The problem was recently solved in [3]. The center of the mix is not necessarily unique. In our paper, we use that fact to prove that for each \(d\in {\mathbb {N}}\) and each non-empty bounded Borel set \(B\subset {\mathbb {R}}^d\) there exists a d-dimensional probability distribution \(\varvec{\mu }\) satisfying what follows: For each \(n\ge 3\) and each probability distribution \(\varvec{\nu }\) on B, there exist d-dimensional random vectors \({\textbf{X}}_{\varvec{\nu },1},{\textbf{X}}_{\varvec{\nu },2},\dots ,{\textbf{X}}_{\varvec{\nu },n}\) such that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \varvec{\nu }\) and \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) for \(i=1,2,\dots ,n\) (see Theorem 1). The assumption about the boundedness of the support of measure \(\varvec{\nu }\) can be replaced with the weaker assumption about the concentration of measure \(\varvec{\nu }\) (see Theorem 2 and Corollary 1).
Our results have connections with both risk theory and statistics. Namely, if we have a number of observations from some unknown probability distribution \(\mu \), and if we do not know anything about the dependence structure of these observations, then the sample mean is not a good statistic for inference about \(\mu \). For the connections with risk theory, see [1, 2, 4, 6].
In Sect. 2, we prove Theorems 1 and 2, and Corollary 1, but first we present some lemmas. The results presented in Sect. 3 show that certain generalizations of Theorem 1 are not possible.
2 The Main Results
Before proving our main results (Theorems 1 and 2), we need to present some auxiliary results.
Lemma 1
(See [3, Example 2.6].) Let N and M be independent random variables such that \(P(M=0)=P(M=1)=\frac{1}{2}\) and \(P(N=2^n)=\frac{1}{2^{n+1}}\) for \(n=0,1,\dots \). We define random variables \(U_0=V_0=N\), \(W_0=W_1=-2N\), \(U_1=2(1-M)N+M\) and \(V_1=2MN+1-M\). Then, \(U_0\sim U_1\), \(V_0\sim V_1\), and \(W_0\sim W_1\) and
Moreover, if X is any of random variables \(U_0\), \(U_1\), \(V_0\), \(V_1\), \(W_0\), and \(W_1\), then
In particular, \(E|X|=\infty \) and \(E|X|^\alpha <\infty \) for each \(\alpha \in (0,1)\).
Now, we use Lemma 1 and binary representations of numbers from [0, 1] to obtain the following proposition.
Proposition 1
There exist random variables \(U_t\), \(V_t\), \(W_t\) with \(t\in [0,1]\), such that
and the distributions of \(U_t\), \(V_t\), and \(W_t\) each do not depend on t.
Proof
Let \((U_{0,j}, V_{0,j}, W_{0,j}, U_{1,j}, V_{1,j}, W_{1,j})\) for \(j=1,2,\dots \) be a sequence of independent copies of the random vector defined in Lemma 1. We define
where \(t=\sum _{j=1}^\infty \frac{t_j}{2^j}\) is a binary representation of t, and \(t_1,t_2,\dots \in \{0,1\}\).
All the above series are almost surely convergent. Indeed, by Lemma 1, we have
and by the Borel–Cantelli lemma, we obtain that \(\left| \frac{U_{t_j,j}}{2^j}\right| \le \frac{1}{2^{j/2}}\) for all but finitely many \(j=1,2,\dots \). It follows that \(\sum _{j=1}^\infty \frac{U_{t_j,j}}{2^j}\) is almost surely absolutely convergent. The same holds for \(\sum _{j=1}^\infty \frac{V_{t_j,j}}{2^j}\) and \(\sum _{j=1}^\infty \frac{W_{t_j,j}}{2^j}\).
The distribution of \(U_t\) does not depend on t. Indeed, if \(s=\sum _{j=1}^\infty \frac{s_j}{2^j}\) and \(t=\sum _{j=1}^\infty \frac{t_j}{2^j}\), then \((U_{s_j,j}:j\in {\mathbb {N}})\sim (U_{t_j,j}:j\in {\mathbb {N}})\) since both sequences are i.i.d. with the same marginal distribution. Hence, \(U_s\sim U_t\). A similar argument shows that the distributions of \(V_t\) and \(W_t\) do not depend on t. For later use, we denote these distributions as follows: \(U_t\sim \mu _U\), \(V_t\sim \mu _V\), and \(W_t\sim \mu _W\).
Finally, for each \(t\in [0,1]\) we have
\(\square \)
Theorem 1
Let \(d\in {\mathbb {N}}\) and \(B\subset {\mathbb {R}}^d\) be a non-empty bounded Borel subset of the d-dimensional Euclidean space. There exists a d-dimensional probability distribution \(\varvec{\mu }\) satisfying what follows: For each \(n\ge 3\) and each probability distribution \(\varvec{\nu }\) on B, there exist d-dimensional random vectors \({\textbf{X}}_{\varvec{\nu },1},{\textbf{X}}_{\varvec{\nu },2},\dots ,{\textbf{X}}_{\varvec{\nu },n}\) such that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \varvec{\nu }\) and \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) for \(i=1,2,\dots ,n\).
Proof
Without loss of generality, we may assume that \(B\subset [0,1]^d\). (If \(B\not \subset [0,1]^d\), then we may apply an affine transformation to each of the d coordinates of \({\mathbb {R}}^d\) to put B into \([0,1]^d\).)
Let \(\varvec{\mu }\) be the probability distribution on \({\mathbb {R}}^d\) such that all d marginals of \(\varvec{\mu }\) are independent and equal to the convolution of the distributions \(\mu _U\), \(\mu _V\), and \(\mu _W\) defined in the proof of Proposition 1 (i.e., \(\varvec{\mu }=\mu \otimes \dots \otimes \mu \), where \(\mu =\mu _U*\mu _V*\mu _W\)).
We fix an arbitrary \(n\ge 3\). Let \(\varvec{\nu }\) be any probability distribution on B and let \({\textbf{T}}=(T^{(1)},T^{(2)},\dots ,T^{(d)})\) be a random vector satisfying \({\textbf{T}}\sim \varvec{\nu }\). Using the binary representation of \(T^{(1)},T^{(2)},\dots ,T^{(d)}\), we define random variables \(T_j^{(m)}\) with \(m=1,2,\dots ,d\), and \(j=1,2,\dots \), such that \(T^{(m)}=\sum _{j=1}^\infty \frac{T_j^{(m)}}{2^j}\) for \(m=1,2,\dots ,d\), and random variables \(T_j^{(m)}\) take values in \(\{0,1\}\). Let \((U_{0,j}^{(m),k}, V_{0,j}^{(m),k}, W_{0,j}^{(m),k}, U_{1,j}^{(m),k}, V_{1,j}^{(m),k}, W_{1,j}^{(m),k})\) (with \(m=1,2,\dots ,d\), \(k=0,1,\dots ,n-1\) and \(j = 1,2,\dots \)) be a system of independent copies of the random vector defined in Lemma 1. We assume that this system is independent of \({\textbf{T}}\).
For \(k=0,1,\dots ,n-1\), let \({\textbf{U}}_{{\textbf{T}}}^k=(U_{{\textbf{T}}}^{(1),k},U_{{\textbf{T}}}^{(2),k},\dots ,U_{{\textbf{T}}}^{(d),k})\), \({\textbf{V}}_{{\textbf{T}}}^k=(V_{{\textbf{T}}}^{(1),k},V_{{\textbf{T}}}^{(2),k},\dots ,V_{{\textbf{T}}}^{(d),k})\) and \({\textbf{W}}_{{\textbf{T}}}^k=(W_{{\textbf{T}}}^{(1),k},W_{{\textbf{T}}}^{(2),k},\dots ,W_{{\textbf{T}}}^{(d),k})\) be given by
for \(m=1,2,\dots ,d\)
We define the requested random vectors \({\textbf{X}}_{\varvec{\nu },i}=(X_{\varvec{\nu },i}^{(1)},X_{\varvec{\nu },i}^{(2)},\dots ,X_{\varvec{\nu },i}^{(d)})\) as follows:
Now, we use the property that for each \(t\in \{0,1\}\) and each m, j and k, we have \(U_{t,j}^{(m),k}+V_{t,j}^{(m),k}+W_{t,j}^{(m),k}=t\). As a consequence, we obtain that for \(m=1,2,\dots ,d\), we have
It follows that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})={\textbf{T}}\sim \varvec{\nu }\).
It remains to show that for each \(i=1,2,\dots ,n\), we have \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\). First, we observe that the conditional distribution of \(X_{\varvec{\nu },i}^{(m)}= U_{{\textbf{T}}}^{(m),i\text { mod }n} +V_{{\textbf{T}}}^{(m),(i+1)\text { mod }n} +W_{{\textbf{T}}}^{(m),(i+2)\text { mod }n}\), given T, is \(\mu =\mu _U*\mu _V*\mu _W\). Moreover, the coordinates \(X_{\varvec{\nu },i}^{(1)}\), \(X_{\varvec{\nu },i}^{(2)}\), ...,\(X_{\varvec{\nu },i}^{(d)}\) are conditionally independent (given T). It follows that \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }=\mu \otimes \dots \otimes \mu \), conditionally. Since this distribution does not depend on the condition, we obtain that \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) unconditionally. \(\square \)
In the following results, we strengthen Theorem 1 by replacing the assumption about the boundedness of the support of measure \(\varvec{\nu }\) with the weaker assumption about the concentration (small tails) of measure \(\varvec{\nu }\).
Theorem 2
Let \(d\in {\mathbb {N}}\), \(B_1\subset B_2\subset \dots \subset {\mathbb {R}}^d\) be non-empty bounded Borel sets and \(0\le p_1\le p_2\le \dots \) be a sequence of real numbers satisfying \(\lim _{k\rightarrow \infty }p_k=1\). There exists a d-dimensional probability distribution \(\varvec{\mu }\) satisfying what follows: For each \(n\ge 3\) and each probability distribution \(\varvec{\nu }\) on \({\mathbb {R}}^d\) satisfying \(\varvec{\nu }(B_k)\ge p_k\) for \(k=1,2,\dots \), there exist d-dimensional random vectors \({\textbf{X}}_{\varvec{\nu },1},{\textbf{X}}_{\varvec{\nu },2},\dots ,{\textbf{X}}_{\varvec{\nu },n}\) such that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \varvec{\nu }\) and \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) for \(i=1,2,\dots ,n\).
Proof
If \(p_k=1\) for some k, then the result follows by Theorem 1 applied to the Borel set \(B=B_k\). In the sequel, we assume that \(p_k<1\) for each k. We may also assume that the sequence \((p_k)\) is strictly increasing. Indeed, if \(p_k=p_{k+1}\), then we can exclude both \(p_{k+1}\) and \(B_{k+1}\) from the respective sequences.
We put \(p_0=0\) and \(B_0=\emptyset \). For \(k=1,2,\dots \) let \(\varvec{\mu }_k\) be the probability measure given by Theorem 1 for the Borel set \(B=B_k\) and let \(\varvec{\mu }=\sum _{k=1}^\infty (p_k-p_{k-1})\varvec{\mu }_k\).
Given the measure \(\varvec{\nu }\), we will define a sequence \((\varvec{\nu }_k)\) of measures on \({\mathbb {R}}^d\) by formula \(\varvec{\nu }_k=\sum _{l=1}^kA_{k,l}\cdot \varvec{\nu }|_{B_l{\setminus } B_{l-1}}\), where \(A_{k,l}=\prod _{j=l}^{k-1}(\varvec{\nu }(B_j)-p_j)/\prod _{j=l}^k(\varvec{\nu }(B_j)-p_{j-1})\) for \(1\le l\le k\), and the restriction \(\varvec{\nu }|_B\) is defined by \(\varvec{\nu }|_B(A)=\varvec{\nu }(A\cap B)\) for each Borel set \(A\subset {\mathbb {R}}^d\). It is easy to verify that for each \(1\le l\le k\) we have \(A_{k,l}\ge 0\), \(\sum _{l=1}^kA_{k,l}\cdot (\varvec{\nu }(B_l)-\varvec{\nu }(B_{l-1}))=1\), and \(\sum _{k=l}^\infty A_{k,l}\cdot (p_k-p_{k-1})=1\). Consequently, \(\varvec{\nu }_k\) is a probability measure on \(B_k\), and \(\sum _{k=1}^\infty (p_k-p_{k-1})\varvec{\nu }_k=\sum _{(l,k):1\le l\le k}(p_k-p_{k-1})A_{k,l}\varvec{\nu }|_{B_l{\setminus } B_{l-1}}=\sum _{l=1}^\infty \varvec{\nu }|_{B_l{\setminus } B_{l-1}}=\varvec{\nu }\).
Now, for each \(k=1,2,\dots \) let \({\textbf{X}}_{\varvec{\nu }_k,1},{\textbf{X}}_{\varvec{\nu }_k,2},\dots ,{\textbf{X}}_{\varvec{\nu }_k,n}\) be d-dimensional random vectors given by Theorem 1. Moreover, let L be a random variable, independent of these vectors, such that \(P(L=k)=p_k-p_{k-1}\) for \(k=1,2,\dots \). We define \({\textbf{X}}_{\varvec{\nu },i}={\textbf{X}}_{\varvec{\nu }_L,i}\) for \(i=1,2,\dots ,n\). The conditional distribution of \({\textbf{X}}_{\varvec{\nu },i}\) (under the condition \(L=k\)) is \(\varvec{\mu }_k\). Consequently, \({\textbf{X}}_{\varvec{\nu },i}\sim \sum _{k=1}^\infty (p_k-p_{k-1})\varvec{\mu }_k=\varvec{\mu }\). Similarly, \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \sum _{k=1}^\infty (p_k-p_{k-1})\varvec{\nu }_k=\varvec{\nu }\). \(\square \)
We recall that for two probability distributions \(\mu \) and \(\nu \) on \({\mathbb {R}}\), we say that \(\mu \) is stochastically dominated by \(\nu \) (denoted by \(\mu \le _{st}\nu \)) when their cumulative distribution functions satisfy the inequality \(F_\mu (u)\ge F_\nu (u)\) for each \(u\in {\mathbb {R}}\). Equivalently, \(\mu \le _{st}\nu \) if and only if there exist random variables \(X\sim \mu \) and \(Y\sim \nu \) satisfying \(X\le Y\).
Corollary 1
Let \(d\in {\mathbb {N}}\), \(f:{\mathbb {R}}^d\rightarrow {\mathbb {R}}\) be a Borel function such that \(\lim _{{\textbf{x}}\rightarrow \infty }f({\textbf{x}})=+\infty \), and \(\nu _0\) be a fixed probability distribution on \({\mathbb {R}}\). We consider a family \(\varvec{{\mathcal {V}}}\) consisting of probability distributions of all d-dimensional random vectors \({\textbf{Z}}\) such that the distribution of \(f({\textbf{Z}})\) is stochastically dominated by \(\nu _0\).
There exists a d-dimensional probability distribution \(\varvec{\mu }\) satisfying what follows: For each \(n\ge 3\) and each probability distribution \(\varvec{\nu }\in \varvec{{\mathcal {V}}}\), there exist d-dimensional random vectors \({\textbf{X}}_{\varvec{\nu },1},{\textbf{X}}_{\varvec{\nu },2},\dots ,{\textbf{X}}_{\varvec{\nu },n}\) such that \(\frac{1}{n}({\textbf{X}}_{\varvec{\nu },1}+{\textbf{X}}_{\varvec{\nu },2}+\dots +{\textbf{X}}_{\varvec{\nu },n})\sim \varvec{\nu }\) and \({\textbf{X}}_{\varvec{\nu },i}\sim \varvec{\mu }\) for \(i=1,2,\dots ,n\).
The case when f is the Euclidean norm in \({\mathbb {R}}^d\) and \(\nu _0\) is a probability distribution on \((0,\infty )\) seems to be the most useful.
Proof
For \(k=1,2,\dots \) let \(B_k=\{{\textbf{x}}\in {\mathbb {R}}^d:f({\textbf{x}})\le k+a\}\), where \(a\in {\mathbb {R}}\) is chosen in such a way that the sets \(B_k\) are non-empty and let \(p_k=\nu _0((-\infty ,k+a])\). By the assumption about the limit of f at infinity, the sets \(B_k\) are bounded. Applying Theorem 2 for sequences \((B_k)\) and \((p_k)\) completes the proof of the corollary. \(\square \)
3 Some Necessary Conditions for the Center of the Mix
In this section, we present some necessary conditions that should be satisfied by each complete mix. In particular, we show that we can neither replace the [0, 1] interval in Proposition 1 nor skip the boundedness assumption about set B in Theorem 1. (However, we can weaken the assumptions about the distribution \(\varvec{\nu }\) as in Theorem 2 and Corollary 1.) We also generalize the following necessary condition for a probability distribution to be n-completely mixable, formulated by Wang and Wang in 2011:
Proposition 2
( [8], see also [5, 6, 9]) Suppose the probability distribution \(\mu \) is n-completely mixable, centered at C. Let \(a=\inf \{x:\mu ((-\infty ,x])>0\}\) and \(b=\sup \{x:\mu ((-\infty ,x])<1\}\). If one of a and b is finite, then the other one is finite, and \(a+\frac{b-a}{n}\le C\le b-\frac{b-a}{n}\).
The following proposition and Corollary 2 significantly strengthen the above result (cf. Remark 1).
Proposition 3
Let \(\nu \) be a probability distribution on \({\mathbb {R}}\). If \(X_1,X_2,\dots ,X_n\) are random variables satisfying \(X_1+X_2+\dots +X_n\sim \nu \), then for each \(a_1,a_2,\dots ,a_n\in {\mathbb {R}}\) and \(i=1,2,\dots ,n\) we have
and
Proof
We show the first inequality. Using the implication \((\forall _j\ X_j\le a_j)\Rightarrow \sum _{j=1}^nX_j\le \sum _{j=1}^na_j\), we obtain
Similarly, by \((\sum _{j=1}^nX_j\le \sum _{j=1}^na_j,\ \forall _{j\ne i}\ -X_j\le -a_j)\Rightarrow X_i\le a_i\) we obtain
The proof of the second inequality is analogous. \(\square \)
If we put \(\nu =\delta _t\) (the one-point probability distribution concentrated at \(t\in {\mathbb {R}}\)) and \(a_i=t-\sum _{\begin{array}{c} j=1\\ j\ne i \end{array}}^na_j\) into Proposition 3, then we obtain the following corollary.
Corollary 2
If \(X_1,X_2,\dots ,X_n\) are random variables satisfying \(X_1+X_2+\dots +X_n=t\in {\mathbb {R}}\), then for each \(a_1,a_2,\dots ,a_n\in {\mathbb {R}}\) and \(i=1,2,\dots ,n\) we have
and
If \(\sum _{j=1}^n a_j<t\), then \(\sum _{j=1}^nP(X_j>a_j)\ge 1\). If \(\sum _{j=1}^n a_j>t\), then \(\sum _{j=1}^nP(X_j<a_j)\ge 1\).
Remark 1
We will show that Corollary 2 is a generalization of Proposition 2.
Let \(X_1,X_2,\dots ,X_n\) be random variables satisfying \(X_1,X_2,\dots ,X_n\sim \mu \) and \(\frac{1}{n}(X_1+X_2+\dots +X_n)=C\). We recall that \(a=\inf \{x:\mu ((-\infty ,x])>0\}\) and \(b=\sup \{x:\mu ((-\infty ,x])<1\}\).
Assume that a is finite. Then, for each \(j=1,2,\dots ,n\) we have \(P(X_j<a)=\mu ((-\infty ,a))=0\). By (2) applied to \(a_1=a_2=\dots =a_n=a\) and \(t=nC\) we obtain \(P(X_i>nC-(n-1)a)=0\). Consequently, b is finite and \(b\le nC-(n-1)a\), which is equivalent to \(a+\frac{b-a}{n}\le C\).
Now, assume that b is finite. Then, for each \(i=1,2,\dots ,n\) we have \(P(X_i>b)=\mu ((b,\infty ))=0\). By (3) applied to \(a_1=a_2=\dots =a_n=b\) and \(t=nC\) we obtain \(P(X_i<nC-(n-1)b)=0\). Consequently, a is finite and \(a\ge nC-(n-1)b\), which is equivalent to \(C\le b-\frac{b-a}{n}\).
As a consequence of Corollary 2, we immediately obtain the following corollary.
Corollary 3
Let \(\mu _1,\mu _2,\dots ,\mu _n\) be probability distributions. If there exist random variables \(X_1\sim \mu _1\), \(X_2\sim \mu _2,\dots \), \(X_n\sim \mu _n\) and \(t\in {\mathbb {R}}\) such that \(X_1+X_2+\dots +X_n=t\), then
Corollary 3 shows that we cannot replace the [0, 1] interval in Proposition 1 by any unbounded set. Additionally, it shows that we cannot skip the assumption that the set B is bounded in Theorem 1. (Indeed, Corollary 3 implies that the projection of B onto each coordinate has to be bounded.)
Availability of Data and Materials
Data sharing is not applicable to this article as no datasets were generated or analyzed during the current study.
References
Dhaene, J., Denuit, M.: The safest dependence structure among risks. Insurance Math. Econ. 25(1), 11–21 (1999)
Embrechts, P., Puccetti, G.: Bounds for the sum of dependent risks having overlapping marginals. J. Multivariate Anal. 101(1), 177–190 (2009)
Puccetti, G., Rigo, P., Wang, B., Wang, R.: Centers of probability measures without the mean. J. Theor. Prob. 32, 1482–1501 (2019)
Puccetti, G., Rüschendorf, L.: Bounds for joint portfolios of dependent risks. Stat. Risk Model. 29(2), 107–131 (2012)
Puccetti, G., Wang, B., Wang, R.: Advances in complete mixability. J. Appl. Prob. 49(2), 430–440 (2012)
Puccetti, G., Wang, B., Wang, R.: Complete mixability and asymptotic equivalence of worst-possible VaR and ES estimates. Insurance Math. Econ. 53(3), 821–828 (2013)
Rüschendorf, L., Uckelmann, L.: Variance minimization and random variables with constant sum. In: Cuadras, C.M., Fortiana, J., Rodriguez-Lallena, J.A. (eds.) Distributions with Given Marginals and Statistical Modelling, pp. 211–222. Kluwer Academic Publishers, Dordrecht (2002)
Wang, B., Wang, R.: The complete mixability and convex minimization problems with monotone marginal densities. J. Multivariate Anal. 102(10), 1344–1360 (2011)
Wang, R.: Current open questions in complete mixability. Prob. Surv. 12, 13–32 (2015)
Funding
No funding was received to assist with the preparation of this manuscript.
Author information
Authors and Affiliations
Contributions
Andrzej Komisarski and Jacques Labuschagne wrote the main manuscript text. All authors reviewed the manuscript. The article is a part of Jacques Labuschagne’s doctoral research.
Corresponding author
Ethics declarations
Conflict of interest
The authors have no competing interests to declare that are relevant to the content of this article.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Komisarski, A., Labuschagne, J. The Distributions of the Mean of Random Vectors with Fixed Marginal Distribution. J Theor Probab 37, 2121–2129 (2024). https://doi.org/10.1007/s10959-023-01277-2
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10959-023-01277-2
Keywords
- Sums of random vectors
- Distributions of sums of random variables
- Multivariate dependence
- Complete mixability