Abstract
We give necessary and sufficient conditions for an orthogonal series to converge in the mean-squares to a nonnegative function. We present many examples and applications, in analysis and probability. In particular, we give necessary and sufficient conditions for a Lancaster-type of expansion \( \sum _{n\ge 0}c_{n}\alpha _{n}(x)\beta _{n}(y)\) with two sets of orthogonal polynomials \(\left\{ \alpha _{n}\right\} \) and \(\left\{ \beta _{n}\right\} \) to converge in means-squares to a nonnegative bivariate function. In particular, we study the properties of the set \(C(\alpha ,\beta )\) of the sequences \(\left\{ c_{n}\right\} ,\) for which the above-mentioned series converge to a nonnegative function and give conditions for the membership to it. Further, we show that the class of bivariate distributions for which a Lancaster type expansion can be found, is the same as the class of distributions having all conditional moments in the form of polynomials in the conditioning random variable.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 Notation, terminology and basic settings
First, let us fix notation that mostly comes from the measure theory. All signed measures considered in the paper will be \(\sigma -\)finite, consequently, the Radon-Nikodym theorem can be applied. If \(\chi \) is a signed measure and \(\chi =\chi ^{+}-\chi ^{-}\) is its Hahn-Jordan decomposition then \(|\chi |=\chi ^{+}+\chi ^{-}\) is a measure. Obviously, a signed measure \(\chi \) is a measure if \(\chi ^{-}=0\). We will use the notation \(\int f(x)d\mu (x)\) interchangeably with \(\int f(x)\mu (dx)\) or even \(\int fd\mu \) if the set of integration is evident, to denote integral with respect to the (possibly signed) measure \(\mu \). Sometimes \(d\mu (.)\) will denote measure \(\mu \) itself.
Let \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\) denote the set of all functions \(f: \mathbb {R}^{m}\longrightarrow \mathbb {R}\) that are square-integrable with respect to the measure \(\left| \mu \right| \). Let us also agree that since all functions from the set \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\) are defined only on \(\mathrm{supp\,}( \mu )\) (\(\mu -a.s.\) in fact), hence we will use notation \(L_{2}(\mu )\) instead \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\).
In the sequel, we will be interested only in signed measures that have one-dimensional marginal measures that are identified by their moments (for the definition and basic properties see the Appendix below). Following [5] or [24] this is assured for those one-dimensional measures \( \mu \) that they satisfy the so-called Cramer’s conditionFootnote 1, that is that there exists \(\delta >0\) such that
In fact condition (1.1) can have a weaker form (i.e. the so-called Hardy’s condition if the measure \(\mu \) has support contained in \(\left\{ x:x\ge 0\right\} \). But we will not go into these details.
Let us denote by Cra the set of all signed measures \(\chi \) on \(\mathbb {R}\) such that satisfy the condition (1.1) for some positive number \(\delta \) . Notice that Cra contains all measures with bounded supports.
Further, let us introduce the following set of signed measures \(AC2(\mu ),\) generated by a measure \(\mu \):
In other words, the set \(AC2(\mu )\) contains all signed measures \(\nu \) that are absolutely continuous with respect to \(\mu \) with their Radon-Nikodym derivative \(\frac{d\nu }{d\mu }\) (i.e. function f) being square integrable with respect to the measure \(\mu .\) Note, that in the definition of the set \( AC2(\mu )\), \(\mu \) can be a multidimensional \(\sigma -\)finite measure.
We have the following simple observation:
Proposition 1
If a one-dimensional measure \(\chi \) belongs to the set Cra, then \( AC2\left( \chi \right) \subset Cra\).
Proof
Let \(f\in L^{2}(\chi )\), then by by the Cauchy-Schwarz inequality we have:
Consequently, \(fd\chi \) satisfies condition (1.1), that is, it belongs to the set Cra. \(\square \)
In other words, if a signed measure \(\mu \) is identifiable by moments, then every element of \(AC2\left( \mu \right) \) is identifiable by moments.
Let \(\mu \) be a measure from the set Cra, by \(AC2^{+}(\mu )\) let us denote the subset of \(AC2(\mu )\) that contains only measures, i.e. \( \int _{A}f(x)d\mu (x)\ge 0\) for any \(\mu -\) measurable set \(A\subset \mathbb { R}\). Notice that \(AC2^{+}(\mu )\) is in fact a closed cone in \(AC2(\mu )\).
Now if \(\mu \) is a measure, \(\left\{ p_{n}\right\} _{n\ge 0}\) denotes the set of polynomials that are orthogonal with respect to the measure \(\mu .\) Let us also define numbers \(\hat{p}_{n}\) by the following orthogonality relationship
with \(\delta _{nm}\) denoting traditionally, the Kronecker’s delta. Let us agree that in the sequel the “hat” over the symbol of a polynomial will denote the positive number defined by (1.2). Further, let \(\left\{ c_{n}\right\} _{n\ge 0}\) be the sequence of reals. The infinite series of the form
is called an orthogonal series. It is known (see, e.g. [2]), that if the following condition
is satisfied, then the series (1.3) converges in \(L_{2}(\mu )\).
Remark 1
Let us recall, that basically, all series considered will converge in the mean-square sense for some specified measure. However, let us recall that due to the Rademacher-Men’shov theorem (see, e.g. [2]), assuming sometimes only little stronger conditions, one can obtain the convergence almost everywhere with respect to the specified measure. More precisely, the Rademacher-Meshov theorem states that if the following condition is satisfied
then the series (1.3) converges in the means-squares and almost everywhere \(\mathrm{mod }\mu .\)
1.2 The problem
The main idea of the paper is to present necessary and sufficient conditions for the positivity of the sum of the orthogonal series (1.3) for almost all x belonging to the closed subset \(\mathcal {M}\) of the support of the positive measure \(\mu \) with respect to which polynomials \(\left\{ p_{n}\right\} \) are orthogonal. The sufficient part of the theorem has been presented in 2011 in [17]. Later over the years, slight generalizations of the original formulation and many examples were presented in [15, 16].
However, only recently I have realized, that the sufficient conditions for the coefficients \(c_{n}\) to assure positivity of (1.3), are also necessary.
The paper is organized as follows. In the next Sect. 2, we present our main result together with its simple proof. We also quote papers where many examples illustrating the assertions of the theorem are presented. The last Sect. 3, presents applications of our result to probability theory in particular to the so-called Lancaster expansions. There is also an appendix in which we recall basic facts about the moments and the moments’ problem and moment sequences.
2 General results
Our main result is the following:
Theorem 1
Let a measure \(\mu \in Cra\), \(\left\{ p_{n}\right\} \) be the sequence of polynomials orthogonal with respect to the measure \(\mu \). Let us consider the orthogonal series (1.3) and by f let us denote the mean-square sum of it. Let \(\mathcal {M}\) be some closed subset of the \( \mathrm{supp\,}( \mu ).\)
The following two conditions are equivalent:
a) \(f\left( x\right) \ge 0\) \(\mu \) - a.s. on \(\mathcal {M}\),
b) There exists \(\nu \in AC2^{+}(\mu )\), with \(\mathrm{supp\,}( \nu )=\mathcal {M}\) such that \(f(x)=\frac{d\nu }{d\mu }(x)\).
If one of the conditions a) and b) is satisfied, then the coefficients \( c_{n} \) are given by the following formula:
Remark 2
Notice, that, if additionally, condition (1.5) is satisfied by the coefficients \(\left\{ c_{n}\right\} \), then the convergence to f is not only in mean-square but also almost surely for almost all \(x\in \mathcal {M}\) \(\mathrm{mod }\mu \).
Remark 3
Notice, that, if \(\left\{ r_{n}\right\} \) denotes the sequence of polynomials orthogonal with respect to the measure \(\nu \), then the number \( \int _M p_{n}(x)d\nu (x)\) is equal to the free coefficient in the connection coefficient expansion of \(p_{n}(x)\) in terms of \(\left\{ r_{n}\right\} .\) More precisely considering connection coefficient expansions
we have
Consequently, the assertion of the Theorem 1 can be rephrased in the following way.
An orthogonal series (1.3) with coefficients satisfying condition (), is nonnegative for almost all (mod \(\mu )\) \(x\in M \) if and only if another sequence \(\left\{ r_{n}\right\} \) of orthogonal polynomials can be found such that considering connection coefficient expansion of \(p_{n}(x)\) in terms of \(\left\{ r_{n}\right\} \) given by () we have:
Proof of Theorem 1
b) \(\Rightarrow \) a). First, let us assume that the coefficients \(c_{n}\) are given by (2.1) and let us denote by f(x) the sum of (1.3). By assumptions, we know that it exists and it is square-integrable with respect to \(\mu .\) We have
Knowing numbers \(c_{n}\hat{p}_{n}\), \(n=0,1,2\ldots \) and the form of polynomials \(\left\{ p_{n}(x)\right\} \), we can find numbers \(\left\{ \int _{\mathcal {M}}x^{n}d\nu (x)\right\} _{n\ge 0}\) and \(\left\{ \int _{\mathcal {M}}x^{n}f(x)d\mu (x)\right\} \). We see that they are identical and the two measures are, by assumption, identifiable by moments so the two measures must be identical i.e.
But \(\nu \) was chosen to be nonnegative. So \(f(x)\ge 0\) on the \(\mathcal {M}\) mod \(\mu \). Besides, we see that
a) \(\Rightarrow \) b). Now, let us assume, that we want to find an expansion of the Radon-Nikodym derivative of two nonnegative measures \(\nu<<\mu \) that is additionally square-integrable (mod \(\mu )\) in an infinite orthogonal series. That is, we are looking for the coefficients of the expansion of the form of (1.3). Then, following our assumptions, we have
where \(\left\{ r_{n}(x)\right\} \) are polynomials orthogonal with respect to \(\nu \). These polynomials exist since for every positive measure satisfying condition (1.1) one can define such polynomials. Naturally, having two sets of orthogonal polynomials one has a set of connection coefficients between them. Since f is square-integrable with respect to \(d\mu (x)\) we know that the coefficients \(c_{n}\) are defined uniquely. Besides we have
\(\square \)
There are numerous examples of expansions of the type (1.3). They appeared over the years in [17] (Sect. 5 concerning mostly polynomials from the so-called Askey-Wilson scheme) or recently in [20], as well as in [15, 18, 19] (concerning. among others, Charlier (3.7) or Jacobi (3.6) polynomials).
Remark 4
Notice also that coefficient \(\gamma _{n,0}\) is equal to
where the coefficients \(\left\{ \pi _{n.j}\right\} \) are defined by the expansion
while the numbers \(\left\{ m_{j}\right\} \) form a moment sequence of some distribution absolutely continuous with respect to the measure \(\mu .\) This observation can be derived directly from (2.1) or from the formula given by Lemma 1 of [16]).
This observation leads also to the following method of checking if a given sequence \(\left\{ c_{n}\right\} \) applied in the series (1.3) can result in the series’ positive-sum. Namely, considering formulae (2.4 ) and (2.3) we can find a sequence \(\left\{ m_{n}\right\} \) by recursively solving a sequence of equations:
for \(n\ge 0\). Since \(\left\{ m_{n}\right\} \) has to be a moment sequence, we can apply one of the known criteria some of which are presented in the Appendix.
Remark 5
Continuing the previous remark, the assertion of the theorem (in case when \( \mathcal {M=}\mathrm{supp\,}\mu \) can be expressed in the following way.
There exists a linear map: K : \(L^{2}(\mu )\longrightarrow L^{2}(\mu )\) that can be symbolically expressed by the following formula:
that maps every function \(f\in L^{2}(\mu )\) on itself, since, as it is easily seen, we have:
3 Probabilistic aspects
In this section, to avoid confusion, we will assume that all considered measures will be probabilistic that is they will integrate up to 1. Further, we will consider bivariate distributions dF(x, y) with marginal distributions \(d\mu (x)\) and \(d\nu (y)\) (i.e. \(d\mu (x)=\int F(x,dy)\) and similarly for \(d\nu \)). Naturally, we will assume, that both marginal measures belong to the set Cra in order to be identified by their moments. Moreover, we will consider only such bivariate distributions F satisfying the following condition:
where \(\frac{\partial ^{2}F}{\partial \mu \partial \nu }(x,y)\), denotes Radon-Nikodym derivative of the measure F with respect to the product measure \(\mu \times v.\) That is, in other words, that \(dF\in AC2(d\mu \times d\nu ),\) where \(d\mu \times d\nu \) denotes the product measure of \(d\mu \) and \(d\nu .\)
Let us denote by \(\left\{ \alpha _{n}(x)\right\} \) and \(\left\{ \beta _{n}(y)\right\} \) two sets of polynomials orthogonal with respect to the measures respectively \(d\mu (x)\) and \(d\nu (y).\) Now, for all distributions satisfying (3.1) the following expansion is valid:
with \(\sum _{i,j\ge 0}\lambda _{i,j}^{2}\hat{\alpha }_{i}\hat{\beta } _{j}<\infty .\) Conditional distributions \(\zeta (dx|y)\) and \(\xi (dy|x)\) are defined respectively, for almost all y (\(\mathrm{mod }v\)) and almost all x (\(\mathrm{mod }\mu )\) by the following relationships:
One shows that both these distributions do exist and are defined uniquely respectively \(\mathrm{mod }v\) and \(\mathrm{mod }\mu \).
Notice that making use of the definition of marginal distribution and the orthogonal polynomials and changing, if necessary, the order of integration that we have:
for all \(n\ge 1\) and likewise for polynomials \(\left\{ \alpha _{n}\right\} \) .
Now applying the above-mentioned definitions and properties to the expansion (3.2) we deduce that
and also, that:
We can now rephrase the above-mentioned Theorem 1 in the form that is important for the probabilists.
Theorem 2
Let \(\mu \in Cra\), and let \(\left\{ \alpha _{n}\right\} \) be a set of polynomials orthogonal with respect to \(\mu \). Then, the orthogonal series:
where, as above, \(\hat{\alpha }_{n}\) is defined by (1.2) and such that
for all y belonging to some closed set \(\mathrm{supp\,}\nu \), converges in mean square (mod\(\mu )\) to a nonnegative function iff there exists a family of probability measures \(\zeta (.|y)\) indexed by y, such that for all \( y\in \mathrm{supp\,}\nu ,\) \(\zeta (.|y)<<\mu \) and \(\forall n\ge 0:\)
Moreover
If additionally there exists a probability measure \(\nu \) such that for \( \forall n\ge 0\):
then one can define a bivariate measure \(F<<\mu \times \nu \) by the formula
and for which
for all Borel subsets A of \(\mathrm{supp\,}( \mu )\) almost everywhere.
Proof
Suppose, that F satisfies (3.1). Let \(\mu \) and \(\nu \) denote its marginal measures and let the sets \(\left\{ \alpha _{n}\right\} \) and \( \left\{ \beta _{n}\right\} \) denote sets of polynomials orthogonal with respect to measures \(\mu \) and \(\nu \) respectively. Let the conditional distributions be defined by (3.5). Notice that by (3.4) and ( 3.5) we have \(\int _{\mathrm{supp\,}\mu }\zeta (dx,y)=1\), a.e. \((\text {mod}\,\nu \) and similarly for the \(\xi (dy,x))\). Now, changing the order of summation and denoting by
we have
Further, utilizing (3.3) and (3.4) we have:
with \(\delta _{n,m}\) denoting traditionally Kronecker’s delta. By assumptions concerning polynomials \(\left\{ \alpha _{n}\right\} \) and by Theorem 1 we see that for all \(y\in \mathrm{supp\,}\nu \), we have:
a.e. \(\mathrm{mod }v\).
Now let us assume the converse statement, i.e. that we have the converging to a nonnegative function in mean-square, series (3.6) with polynomials \( \left\{ \alpha _{n}\right\} \) and the measure \(\mu ,\) as described in the assumptions, together with the condition (3.7) satisfied for almost every y belonging to some closed set that we will denote by \(\mathrm{supp\,} \nu \). By Theorem 1, we deduce that if the series (3.6) converges to a nonnegative function, then there exists a family \(d\zeta (x|y) \) of positive measures absolutely continuous with respect to \(\mu \) such that \(\forall n\ge 0\):
Moreover, we have
Now, if there exists a probability measure \(\nu \) such that \(\forall n\ge 0:\int _{\mathrm{supp\,}\nu }h_{n}(y)v(dy)=\delta _{n,0}\), then we
a.e. mod \(\mu \), hence
as claimed. \(\square \)
The rest of this section will be dedicated to the so-called Lancaster expansions. In particular, we will be able to give now necessary and sufficient conditions for these types of expansions to exist. Let us recall that Lancaster, in the series of papers [8,9,10,11], considered and developed the following question: given a bivariate distribution say dF(x, y), its two marginal distributions say \(d\mu (x)\) and \(d\nu (y)\) and the two sets of polynomials, when is it possible to find the set of numbers \(\left\{ c_{n}\right\} \) such that
almost everywhere in \(\mathrm{supp\,}( \mu )\times \mathrm{supp\,}( \nu )\) with respect to the product measure. In fact, Lancaster in his papers and also his followers in their papers confined the problem to such bivariate distributions dF satisfying condition (3.1).
Definition 1
A class of bivariate distributions with margins identifiable by moments, satisfying (3.1) and having expansion (3.9) will be called Lancaster class (of bivariate distributions), briefly (LC distributions).
Remark 6
Notice, that if F is of LC distribution, then we have:
In other words, in terms used in probability, we can easily deduce that \( \forall n\ge 1\) the conditional moments, i.e.:
respectively \(\mathrm{mod }\nu \) and \(\mathrm{mod }\mu \), where \(p_{n}\) and \( q_{n}\), are some polynomials of the full orderFootnote 2n.
Definition 2
Class of bivariate distributions with margins identifiable by moments, having the property that all its conditional moments of the order, say, n are polynomials of the full order n will be called polynomial class (of distributions) briefly PC distributions.
As a corollary we have the following characterization of the Lancaster class of distributions.
Theorem 3
Let us consider a bivariate distribution F satisfying (3.1) with margins identifiable by moments. Then F is an LC distribution iff it is a PC distribution.
Proof
The fact that every distribution of the Lancaster class belongs also to the PC class was noted in Remark 6. So now, let us assume that F belongs to the PC class. By Theorem 2 we know that it can be expanded in the series (3.6). Now we see that \(\forall n\ge 1:h_{n}(y)=\int \alpha _{n}(x)\zeta (dx,y)=E((\alpha _{n}(X)|Y=y).\) But, by our assumption, \(h_{n}(y)\) has to be a polynomial of the full order n i.e.
where \(\left\{ \beta _{n}\right\} \) are the polynomials orthogonal with respect to the marginal measure \(\nu .\) Hence we musta have \(\gamma _{n,j}=0\) for \(j>n\). Now changing the order of summation in (3.6) we get:
But by our assumption \(E((\beta _{j}(Y)|X=x)\) is a polynomial of the full order j. So we have:
Now, by the uniqueness of expansion, we deduce that \(\forall n>j:\gamma _{n,j}=0.\) \(\square \)
Szabłowski in the series of papers [21,22,23] considered Markov, stochastic processes having two-dimensional finite distributions belonging to PC class. Hence, now, in light of the above-mentioned theorem, there is a possibility of expanding in the Lancaster-like series, the transition functions of such Markov processes.
Let us apply Theorem 2 to the analysis of the LC distributions or more precisely to the analysis when the series
converges to a nonnegative function of \(\left( x,y\right) ,\) where polynomials \(\left\{ \alpha _{n}\right\} \) and \(\left\{ \beta _{n}\right\} \) are defined as in the introduction to Sect. 3. To simplify the formulation of the theorem and the applications following it, let us assume additionally that both families of polynomials \(\left\{ \alpha _{n}\right\} \) and \(\left\{ \beta _{n}\right\} \) are orthonormal with respect to the measures \(\mu \) and v respectively.
Let us also denote by \(C(\alpha ,\beta )\) set of all sequences \(\left\{ c_{n}\right\} \) for which the sum (3.10) exists and is positive. Koudu in [12] showed that this set is convex (which is trivial, see, e.g. Appendix) and moreover, compact with respect to the weak topology. Hence the Choquet’s theorem about extreme points can be applied. As a corollary of the Theorem 2 we have the following result:
Theorem 4
Let the numbers \(\left\{ a_{n,j}\right\} \) and \(\left\{ b_{n,j}\right\} \) be defined by the polynomials \(\alpha _{n}\) and \(\beta _{n} \) in the following way:
The series (3.10) converges in the mean-squares to a positive function if and only if the following system of recurrent equations
for \(n\ge 0\), is satisfied for almost all \(y\in \mathrm{supp\,}v\) and \(x\in \mathrm{supp\,}\mu \), by the two polynomial moment sequences \(\left\{ m_{n}^{(a)}(y)\right\} \) and \(\left\{ m_{n}^{(b)}(x)\right\} \), that are defined by some measures, that are absolutely continuous with respect to the measures \(\mu \) and v respectively.
Proof
Firstly, under our assumptions we have \(\hat{\alpha }_{n}=\hat{\beta }_{n}=1,\) hence following the previous theorem we deduce that the series (3.10) converges in mean squares to some positive function iff
where the expectation is taken with respect to some absolutely continuous measure that is additionally parametrized by the parameter y, that belongs to the \(\mathrm{supp\,}v.\) From this remark follows directly the first of the above-mentioned equation. By a similar argument we deduce that the second equation holds. \(\square \)
Remark 7
Notice that if for \(c_{n}=\rho ^{n},\) for \(\left| \rho \right| <1,\) the series (3.10) is convergent to a positive bivariate probability density, then so is the series
for any probability distribution \(\gamma \) such that \(\gamma (\left\{ -1\right\} \cup \left\{ 1\right\} )=0.\)
Remark 8
In fact, the equations (3.12) and (3.13) should be written in the following, less legible but more precise, recursive way:
for \(n\ge 0\) with \(m_{0}(x)=m_{0}(y)=1\).
Corollary 1
Let coefficients \(\left\{ a_{n,j}\right\} \) and \(\left\{ b_{n,j}\right\} \) be defined by (3.11). If the series (3.10) converges to a positive sum, then
a) \(\sum _{n\ge 0}c_{n}^{2}<\infty ,\)
further additionally:
b) if \(\left\{ 0\right\} \in \mathrm{supp\,}\mu \) and supp v then, we have: \(\infty >\sum _{n\ge 0}c_{n}a_{n,0}b_{n,0}\ge 0,\)
c) if \(\mathrm{supp\,}\mu \) is unbounded, then the sequence \(\left\{ c_{n} \frac{a_{n,n}}{b_{n,n}}\right\} \), is a moment sequence, if additionally \( \mathrm{supp\,}v\) is also unbounded, then \(\left\{ c_{n}^{2}\right\} \) must be a moment sequence.
d) if measures \(\mu \) and v are the same and have unbounded supports, then \(\left\{ c_{n}\right\} \) must be a moment sequence.
Proof
Part a) follows the fact that the series (1.3) converges in mean-squares and that \(\hat{p}_{n}=1\) since we consider only orthonormal polynomials. b) is obvious. c) Firstly, notice that in all cases from the system of equations (3.14) and (3.15) it follows that the leading coefficient in \(m_{n}^{(a)}(y)\) and \( m_{n}^{(b)}(y)\) must be respectively \(c_{n}b_{n,n}/a_{n,n}\) and \( c_{n}a_{n,n}/b_{n,n}.\) Now, if, say \(\mathrm{supp\,}\mu \) is unbounded, then from the fact that \(\left\{ m_{n}^{(b)}(y)\right\} \) must be a moment sequence, then so must be the sequence \(\left\{ y^{n}c_{n}a_{n,n}/b_{n,n}\right\} .\) From this fact the first assertion follows immediately. Now, if both sequences \(\left\{ c_{n}\frac{a_{n,n}}{ b_{n,n}}\right\} ,\) \(\left\{ c_{n}\frac{b_{n,n}}{a_{n,n}}\right\} \) are the moment ones, then is their their product (see the Appendix, below). Part d) follows directly from c). \(\square \)
Remark 9
The assertion d) of the above-mentioned corollary repeats in fact the result of Tyan et al. presented in [25].
Remark 10
Theorem 4, at least theoretically, closes the problem of finding conditions for the convergence to a positive bivariate function of the infinite series (3.10). Namely, having two sequences of moments (what is important given by the recursive formula) one can find their two Laplace transforms of the distributions identified by these sequences and invert them obtaining two conditional measures \(\chi (.|x)\) and \(\zeta (.|y)\) that are also defined by the conditions
The procedure to get these inverses is very difficult and long. On the way, the procedure utilizes Nevalinna’s theory as described say in [1]. Now, the question of summing the series (3.10) is solved by the formula (3.8).
Example 1
We will now present an example in which we show how having a given family of orthogonal polynomials, and a moment sequence \(\left\{ c_{n}\right\} ,\) one finds a sequence of moments \(\left\{ m_{n}(y)\right\} \). Then, having this sequence one finds a sequence of orthogonal polynomials parametrized by y. Then, by different means, including the analysis of the three-term recurrence of this sequence, one finds the properties of the measure having moment sequence \(\left\{ m_{n}(y)\right\} \) and thus conclude, basing on Theorem 4, that the series (3.10) with \(\alpha _{n}(x)=\beta _{n}(x)\) converges to a positive sum. The way is long and it seems that each case would be enough for an article. To shorten the conclusions and the description, the example will concern Hermite polynomials that are well known and the sequence \(c_{n}=\rho ^{n}\) for some \(\left| \rho \right| <1,\) just to illustrate the process of calculation. Note, that such sequence \(\left\{ c_{n}\right\} \) is a moment sequence. On the way, we will make use of the well-known properties of these polynomials. Besides, in this case it is easy just to guess the conditioning measure \(d\chi (.|y).\)
Let us recall that the so-called (probabilistic) Hermite polynomials are defined by the following three-term recurrence
with \(H_{0}(x)=1,\) and \(H_{-1}(x)=0.\) It is known, that we have
Moreover, for all complex x, y and a the following expansions are true:
Notice that
and also that the orthonormal version of Hermite polynomials is equal to \( H_{n}(x)/\sqrt{n!}\), so for even n we have
We used here the well-known approximation \(\left( {\begin{array}{c}2k\\ k\end{array}}\right) \frac{1}{2^{2k}} \cong \frac{1}{\sqrt{\pi k}}\). Thus, applying Corollary 1, we see that every applicable sequence \(\left\{ c_{n}\right\} \) must satisfy the following conditions:
\(\left\{ c_{n}\right\} \) is a moment sequence.
Now, let us recall Remark 7 and examine the case \( c_{n}=\rho ^{n}\) for some \(\left| \rho \right| <1.\) Notice, that this sequence satisfies the above-mentioned conditions, since \(\sum _{n\ge 0}\rho ^{n}/\sqrt{n+1}=Li(\frac{1}{2};\rho )/\rho \ge 0,\) \(\rho \in (-1,1)\) , where \( Li(s;\rho )\) is the so-called polylogarithm function of order s.
We guess, that the measure v(dx|y) has the density
From this, we immediately deduce, that
for \(n\ge 1.\) Let us denote:
Now we change variables to \(z=\frac{x-\rho y}{\sqrt{ 1-\rho ^{2}}}\) getting \(x=\rho y+\sqrt{1-\rho ^{2}} z. \) Now, applying (3.17) we get:
since for \(k=0,1,\ldots \) :
Moreover, we get
after changing variables, as above, and applying (3.17). Thus, there is no need to check that indeed the sequence (3.18) is the solution of the system of equations (3.14) with \(\alpha _{n}(x)=\beta _{n}(x)=H_{n}(x)\) for all \(n\ge 0 \) and \(c_{n}=\rho ^{n}.\)
By the way, we have also:
Returning to Remark 7 we see that the series
converges to a positive density iff \(c_{n}=\int _{-1}^{1}\rho ^{n}d\gamma (\rho ),\) \(n=0,1,\ldots ,\) where \(\gamma \) is some probability distribution on \([-1,1],\) such that \(\gamma (\left\{ -1\right\} \cup \left\{ 1\right\} )=0\). This fact was already noticed by Sarmanov and Bratoeva in 1967 in [14]. Later this result was generalized by Griffith in [7] and Koudu in [13] with Hermite polynomials replaced by polynomials orthogonalizing gamma distribution (Griffith) and Poisson and negative binomial (Koudu). The Koudu’s results were later applied to parameter testing of the chosen Lancaster bivariate distributions by Chen in [6].
Notes
The names of this condition as well as Hardy’s condition, were recalled by Prof. Jordan Stoyanov in a private letter.
\(p_{n}(x)\) is of full order n iff coefficient by \(x^{n}\) is nonzero.
References
Akhiezer, N.I.: The classical moment problem and some related questions in analysis. Translated by N. Kemmer. Hafner Publishing Co., New York 1965 x+253 pp. MR0184042 (32 #1518)
Alexits, G.: Convergence problems of orthogonal series. Translated from the German by I. Földer. International Series of Monographs in Pure and Applied Mathematics, Vol. 20 Pergamon Press, New York-Oxford-Paris 1961 \(\text{rm}\) ix+350 pp. MR0218827
Boas, R.P.: The Stieltjes moment problem for functions of bounded variation. Bull. Am. Math. Soc. 45(6), 399–404 (1939)
Bennett, G.: Hausdorff means and moment sequences. Positivity 15(1), 17–48 (2011)
Chihara, T.S.: An introduction to orthogonal polynomials. Mathematics and its Applications, Vol. 13. Gordon and Breach Science Publishers, New York-London-Paris, 1978. xii+249 pp. ISBN: 0-677-04150-0 MR0481884 (58 #1979)
Chen, X.: A strong law of large numbers for simultaneously testing parameters of Lancaster bivariate distributions. Statist. Probab. Lett. 167, 8 (2020)
Griffiths, R.C.: The canonical correlation coefficients of bivariate gamma distributions. Ann. Math.l Stat. 40, 1401–1408 (1969)
Lancaster, H.O.: The structure of bivariate distributions. Ann. Math. Stat. 29(3), 719–736 (1958)
Lancaster, H.O.: Correlation and complete dependence of random variables. Ann. Math. Stat. 34(4), 1315–1321 (1963)
Lancaster, H.O.: Correlations and canonical forms of bivariate distributions. Ann. Math. Stat. 34(2), 532–538 (1963)
Lancaster, H.O.: Joint probability distributions in the Meixner classes. J. Roy. Stat. Soc. Ser. B 37(3), 434–443 (1975)
Koudu, A.E.: Probabilités de Lancaster (French) [[Lancaster]] probabilities. Expos. Math. 14(3), 247–275 (1996)
Koudu, A.E.: Lancaster bivariate probability distributions with Poisson, negative binomial and gamma margins. Test 7(1), 95–110 (1998)
Sarmanov, O.V., Bratoeva, Z.N.: Probabilistic properties of bilinear expansions in Hermite polynomials. (Russian) Teor. Verojatnost. i Primenen. 12, 520–531 (1967)
Szabłowski, P.J.: On summable, positive Poisson-Mehler kernels built of Al-Salam-Chihara and related polynomials. Infin. Dimens. Anal. Quantum Probab. Relat. Top. 15(3), 1250014 (2012)
Szabłowski, P.J.: A few remarks on orthogonal polynomials. Appl. Math. Comput. 252, 215–228 (2015). arXiv:1207.1172
Szabłowski, P.J.: Expansions of one density via polynomials orthogonal with respect to the other. J. Math. Anal. Appl. 383(1), 35–54 (2011)
Szabłowski, P.J.: On affinity relating two positive measures and the connection coefficients between polynomials orthogonalized by these measures. Appl. Math. Comput. 219(12), 6768–6776 (2013)
Szabłowski, P.J.: Befriending Askey-Wilson polynomials, Infin. Dimens. Anal. Quant. Probab. Relat. Top. 17(3), 1450015 (2014)
Szabłowski, P. J.: On the families of polynomials forming a part of the so-called Askey–Wilson scheme and their probabilistic applications. submitted, arXiv:2007.03267, in print in Infin. Dimens. Anal. Quant. Probab. Relat. Top
Szabłowski, P.J.: On Markov processes with polynomials conditional moments. Trans. Amer. Math. Soc. 367, 8487–8519 (2015). arXiv:1210.6055
Szabłowski, P.J.: On stationary Markov processes with polynomial conditional moments. Stoch. Anal. Appl. 35(5), 852–872 (2017). arXiv:1312.4887
Szabłowski, P.J.: Markov processes, polynomial martingales and orthogonal polynomials. Stochastics 90(1), 61–77 (2018)
Simon, B.: The classical moment problem as a self-adjoint finite difference operator. Adv. Math. 137(1), 82–203 (1998)
Tyan, S.G., Derin, H., Thomas, J.B.: Two necessary conditions on the representation of bivariate distributions by polynomials. Ann. Statist. 4(1), 216–222 (1976)
Author information
Authors and Affiliations
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
The author is very grateful to the unknown referee for suggestions improving the readability of the paper and also for pointing out numerous misprints.
Appendix: A few facts of the moment problem and the orthogonal polynomials
Appendix: A few facts of the moment problem and the orthogonal polynomials
Let be \(\alpha \) be a signed, finite measure on the real line. Then the sequence of reals \(\left\{ m_{n}\right\} _{n}\) defined by: \(m_{n}=\int x^{n}d\alpha (x)\), is called a sequence of moments or the moment sequence (sm) of the measure \(\alpha \text {.}\ \)Below we have the surprising general result by Boas [3].
Theorem 5
Any sequence \(\left\{ m_{n}\right\} _{n\ge 0}\) of real numbers can be represented in the form: \(m_{n}=\int x^{n}d\alpha (x)\ \), where \(\int |d\alpha (x)|<\infty \).
In other words any sequence of numbers is a moment sequence of some signed measure.
We will be interested in sequences of moments of positive measures.
It turns out (see, e.g. [24]), that the sequence \(\left\{ m_{n}\right\} \) is a moment sequence of some nonnegative measure \(\alpha \), iff it satisfies the following condition for \(\forall n\ge 0\!:\)
The sequence \(\left\{ d_{n}\right\} \) related to the sequence \(\left\{ m_{n}\right\} \) and defined by (3.19), is called the Hankel transform of the sequence \(\left\{ m_{n}\right\} .\) It is also known (see, e.g. [24]) that if the sharp inequality in (3.19) holds for all \( n\ge 1\), then the support of \(\alpha \) is of infinite cardinality. If additionally \( \forall n\ge 0\):
then \(\mathrm{supp\,}\alpha \subset [0,\infty ).\)
Sequences \(\left\{ m_{n}\right\} \) that are the moment sequences of some nonnegative measures will me called p(ositive)m(oment) sequences i.e. pm sequences.
Let us mention the two simple necessary conditions for a sequence of reals to be a pm sequence.
Condition 1
(Necessary) Let \(\left\{ d_{n}\right\} _{n\ge 0}\) a pm sequence. Then a) \(d_{2k}\ge 0,\) \(k=0,1,\ldots \) . b) \(\left| d_{k}\right| ^{2}\le d_{2k}d_{0},\) c) sequence \(\left\{ d_{2k}^{1/(2k)}\right\} \) is non-decreasing.
Proof
Assertion a) is obvious, b) follows directly Cauchy-Schwarz inequality while c) follows Jensen’s inequality. \(\square \)
In the sequel we will assume that the measure \(\alpha \) is a probability measure i.e. \(\int d\alpha (x)=1\) which results in the fact that all pm sequences will have \(m_{0}=1.\)
The generating function \(\varphi \) of the pm sequence is defined by the following formula:
Hence, if the moment problem is determinate (i.e. the measure \(\alpha \) is identified by its sequence of moments), this Laplace transform exists for even small neighborhood of zero.
For the aims of this paper, this criterion of determinacy is more important. However, for the sake of completeness, let us remark that there exists, however, another criterion given by Carleman (see, e.g. [1] or [5]) where the determinacy follows the properties of the moment sequence itself. Namely, Carleman’s criterion reads: If only
then the sequence of moments \(\left\{ m_{n}\right\} \) defines uniquely the measure that created this sequence.
It is known (see, e.g. [1] or [5]), that the sequence of polynomials orthogonal with respect to the measure that produced a given moment sequence \(\left\{ m_{n}\right\} \) is given by the following sequence:
\(n=1,2,\ldots \) .
Further, it is also known (see, e.g. [1] or [5]) that for every orthogonal polynomial sequence \(\left\{ p_{n}\right\} \) one can define three sequences of numbers \(\left\{ A_{n}\right\} ,\) \(\left\{ B_{n}\right\} ,\) \(\left\{ C_{n}\right\} \), such that for every \(n\ge 0{:}\)
and for \(n\ge 1:C_{n}A_{n}A_{n-1}>0\), provided the support of the measure making these polynomials orthogonal is of infinite cardinality. These real sequences are defined by numbers \(\left\{ d_{n}\right\} \) given by (3.19). For details see again [1] or [5].
We have also the following simple observations concerning the properties of pm sequences.
Proposition 2
Let \(\left\{ a_{n}\right\} _{n\ge 0}\) and \(\left\{ b_{n}\right\} _{n\ge 0}\) be two pm sequences. Then, so are the following sequences:
1. \(\left\{ pa_{n}+b_{n}(1-p)\right\} _{n\ge 0}\text {, }\)for \(p\in [0,1]\), \(\left\{ \sum _{i=0}^{n}(\pm 1)^{i}\left( {\begin{array}{c}n\\ i\end{array}}\right) \alpha ^{i}\beta ^{n-i}a_{i}b_{n-i}\right\} _{n\ge 0}\), for \(\alpha ,\beta \in \mathbb {R}\) , \(\left\{ a_{n}b_{n}\right\} _{n\ge 0},\)
2. \(\left\{ a_{kn}\right\} _{n\ge 0}\text {, }k\in \mathbb {N}\), \( c_{n}=\left\{ \begin{array}{ccc} a_{2k} &{} if &{} n=2k \\ 0 &{} if &{} n=2k+1 \end{array} \right. \ \text {,}\ \) \(k=0,1,\ldots \) . If a pm sequence \(\left\{ a_{n}\right\} _{n\ge 0}\) is nonnegative then also pm is the following sequence: \(b_{n}^{\ ^{\prime }}=\left\{ \begin{array}{ccc} 0 &{} if &{} n=2k+1 \\ a_{k} &{} if &{} n=2k \end{array} \right. \text {.}\)
3. The following sequences : \(\forall a\in \mathbb {R}:\)
\(k=1,2,\ldots \ \text {,}\ \) Catalan numbers i.e. \(\left\{ \left( {\begin{array}{c}2n\\ n\end{array}}\right) /(n+1)\right\} _{n\ge 0}\text {,}\) \(\left\{ n!\right\} _{n\ge 0}\text {,}\forall k>-1:\{1/(n+1)^{k+1}\}_{n\ge 0},\) \(\left\{ F_{n+1}\right\} _{n\ge 0}\text {,}\) \(\left\{ F_{n+1}/(n+1)\right\} _{n\ge 0}\text {,}\) \(\left\{ F_{2n+2}/(n+1)\right\} _{n\ge 0}\text {,}\) \(\big \{ (F_{2n+1}-1)/(n+1)\big \} _{n\ge 0}\) where \(F_{n}\) denotes \( n- \)th Fibonacci number, are pm sequences.
Proof
1. The arguments are probabilistic. Let X and Y be two independent random variables having respectively moments \(\left\{ a_{n}\right\} \) and \( \left\{ b_{n}\right\} \text {.}\ \) Then \(\left\{ a_{n}b_{n}\right\} \) is the moment sequence of \(XY\ \text {,}\ \) \(\left\{ \sum _{i=0}^{n}(\pm 1)^{i}\left( {\begin{array}{c} n\\ i\end{array}}\right) \alpha ^{i}\beta ^{n-i}a_{i}b_{n-i}\right\} _{n\ge 0}\) is the moment sequence of \((\beta Y\pm \alpha X).\) Let Z have the so-called mixture distribution of the distributions of random variables X and Y having moment sequences, respectively \(\left\{ a_{n}\right\} \) and \(\left\{ b_{n}\right\} \) . Then \(\left\{ pa_{n}+b_{n}(1-p)\right\} _{n\ge 0}\) is the moment sequence of \(Z\text {.} \)
2. \(\left\{ a_{kn}\right\} _{n\ge 0}\) is the moment sequence of \(X^{k}\). To get remaining statements of this we consider special mixtures of the independent copies of X and \(-X\) to get first statement and \(\sqrt{X}\) and \(-\sqrt{X}\) to get the second.
2. \(\left\{ a^{n}\right\} _{n\ge 0}\) is the moment sequence of the one-point distribution concentrated at \(a\ \),
\(k=1,2,\ldots \) is the moment sequence of Normal N(0, 1) distribution, Catalan numbers are moments of distribution with the density \(\frac{1}{2\pi } \sqrt{\frac{4-x}{x}}\ \text {,}\ \) \(x\in (0,4)\ \text {.}\ \) \(\left\{ n!\right\} _{n\ge 0}\) are moments of distribution with the density \(\exp (-x),x\ge 0,\) \(\{1/(n+1)^{k+1}\}_{n\ge 0}\) are the moment sequence of distributions with the densities \((-\log (x))^{k}/\Gamma (k+1),\) \(x\in (0,1)\) , \(k>-1\). For sequences composed of Fibonacci numbers, see [4]. \(\square \)
Hence in particular the following families of polynomials are the pm sequences for every \(x\in \mathbb {R}\):
\(\left\{ \sum _{k=0}^{n}\left( {\begin{array}{c}n\\ k\end{array}}\right) a_{k}(\pm 1)^{n-k}x^{n-k}\right\} _{n\ge 0},\) where \(\left\{ a_{n}\right\} _{n\ge 0}\) is the pm sequence hence some of the Appell polynomial sequences are pm.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Szabłowski, P.J. On positivity of orthogonal series and its applications in probability. Positivity 26, 19 (2022). https://doi.org/10.1007/s11117-022-00883-4
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s11117-022-00883-4
Keywords
- Orthogonal series
- Orthogonal polynomials
- Lancaster bivariate distributions
- Moment sequences
- Absolute continuity of measures