1 Introduction

1.1 Notation, terminology and basic settings

First, let us fix notation that mostly comes from the measure theory. All signed measures considered in the paper will be \(\sigma -\)finite, consequently, the Radon-Nikodym theorem can be applied. If \(\chi \) is a signed measure and \(\chi =\chi ^{+}-\chi ^{-}\) is its Hahn-Jordan decomposition then \(|\chi |=\chi ^{+}+\chi ^{-}\) is a measure. Obviously, a signed measure \(\chi \) is a measure if \(\chi ^{-}=0\). We will use the notation \(\int f(x)d\mu (x)\) interchangeably with \(\int f(x)\mu (dx)\) or even \(\int fd\mu \) if the set of integration is evident, to denote integral with respect to the (possibly signed) measure \(\mu \). Sometimes \(d\mu (.)\) will denote measure \(\mu \) itself.

Let \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\) denote the set of all functions \(f: \mathbb {R}^{m}\longrightarrow \mathbb {R}\) that are square-integrable with respect to the measure \(\left| \mu \right| \). Let us also agree that since all functions from the set \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\) are defined only on \(\mathrm{supp\,}( \mu )\) (\(\mu -a.s.\) in fact), hence we will use notation \(L_{2}(\mu )\) instead \(L_{2}(\mathrm{supp\,}( \mu ),\mu )\).

In the sequel, we will be interested only in signed measures that have one-dimensional marginal measures that are identified by their moments (for the definition and basic properties see the Appendix below). Following [5] or [24] this is assured for those one-dimensional measures \( \mu \) that they satisfy the so-called Cramer’s conditionFootnote 1, that is that there exists \(\delta >0\) such that

$$\begin{aligned} \int \exp (\delta \left| x\right| )d\left| \mu \right| (x)<\infty . \end{aligned}$$
(1.1)

In fact condition (1.1) can have a weaker form (i.e. the so-called Hardy’s condition if the measure \(\mu \) has support contained in \(\left\{ x:x\ge 0\right\} \). But we will not go into these details.

Let us denote by Cra the set of all signed measures \(\chi \) on \(\mathbb {R}\) such that satisfy the condition (1.1) for some positive number \(\delta \) . Notice that Cra contains all measures with bounded supports.

Further, let us introduce the following set of signed measures \(AC2(\mu ),\) generated by a measure \(\mu \):

$$\begin{aligned} AC2(\mu )=\left\{ fd\mu :f\in L_{2}(\mu )\right\} . \end{aligned}$$

In other words, the set \(AC2(\mu )\) contains all signed measures \(\nu \) that are absolutely continuous with respect to \(\mu \) with their Radon-Nikodym derivative \(\frac{d\nu }{d\mu }\) (i.e. function f) being square integrable with respect to the measure \(\mu .\) Note, that in the definition of the set \( AC2(\mu )\), \(\mu \) can be a multidimensional \(\sigma -\)finite measure.

We have the following simple observation:

Proposition 1

If a one-dimensional measure \(\chi \) belongs to the set Cra,  then \( AC2\left( \chi \right) \subset Cra\).

Proof

Let \(f\in L^{2}(\chi )\), then by by the Cauchy-Schwarz inequality we have:

$$\begin{aligned}&\left( \int _{\mathrm{supp\,}\mu }\left| f(x)\right| \exp (\delta |x|/2)d\left| \chi \right| (x)\right) ^{2}\\&\quad \le \int _{\mathrm{supp\,}\mu }\left| f\right| ^{2}d\left| \chi \right| \int _{\mathrm{supp\,}\mu }\exp (\delta \left| x\right| )d\left| \chi \right| (x)<\infty . \end{aligned}$$

Consequently, \(fd\chi \) satisfies condition (1.1), that is, it belongs to the set Cra. \(\square \)

In other words, if a signed measure \(\mu \) is identifiable by moments, then every element of \(AC2\left( \mu \right) \) is identifiable by moments.

Let \(\mu \) be a measure from the set Cra,  by \(AC2^{+}(\mu )\) let us denote the subset of \(AC2(\mu )\) that contains only measures, i.e. \( \int _{A}f(x)d\mu (x)\ge 0\) for any \(\mu -\) measurable set \(A\subset \mathbb { R}\). Notice that \(AC2^{+}(\mu )\) is in fact a closed cone in \(AC2(\mu )\).

Now if \(\mu \) is a measure, \(\left\{ p_{n}\right\} _{n\ge 0}\) denotes the set of polynomials that are orthogonal with respect to the measure \(\mu .\) Let us also define numbers \(\hat{p}_{n}\) by the following orthogonality relationship

$$\begin{aligned} \hat{p}_{n}\delta _{mn}=\int _{\mathrm{supp\,}\mu }p_{n}(x)p_{m}(x)d\mu (x), \end{aligned}$$
(1.2)

with \(\delta _{nm}\) denoting traditionally, the Kronecker’s delta. Let us agree that in the sequel the “hat” over the symbol of a polynomial will denote the positive number defined by (1.2). Further, let \(\left\{ c_{n}\right\} _{n\ge 0}\) be the sequence of reals. The infinite series of the form

$$\begin{aligned} \sum _{n\ge 0}c_{n}p_{n}(x), \end{aligned}$$
(1.3)

is called an orthogonal series. It is known (see, e.g. [2]), that if the following condition

$$\begin{aligned} \sum _{n\ge 0}c_{n}^{2}\hat{p}_{n}<\infty \end{aligned}$$
(1.4)

is satisfied, then the series (1.3) converges in \(L_{2}(\mu )\).

Remark 1

Let us recall, that basically, all series considered will converge in the mean-square sense for some specified measure. However, let us recall that due to the Rademacher-Men’shov theorem (see, e.g. [2]), assuming sometimes only little stronger conditions, one can obtain the convergence almost everywhere with respect to the specified measure. More precisely, the Rademacher-Meshov theorem states that if the following condition is satisfied

$$\begin{aligned} \sum _{n\ge 0}c_{n}^{2}\hat{p}_{n}\log ^{2}(n+1)<\infty , \end{aligned}$$
(1.5)

then the series (1.3) converges in the means-squares and almost everywhere \(\mathrm{mod }\mu .\)

1.2 The problem

The main idea of the paper is to present necessary and sufficient conditions for the positivity of the sum of the orthogonal series (1.3) for almost all x belonging to the closed subset \(\mathcal {M}\) of the support of the positive measure \(\mu \) with respect to which polynomials \(\left\{ p_{n}\right\} \) are orthogonal. The sufficient part of the theorem has been presented in 2011 in [17]. Later over the years, slight generalizations of the original formulation and many examples were presented in [15, 16].

However, only recently I have realized, that the sufficient conditions for the coefficients \(c_{n}\) to assure positivity of (1.3), are also necessary.

The paper is organized as follows. In the next Sect. 2, we present our main result together with its simple proof. We also quote papers where many examples illustrating the assertions of the theorem are presented. The last Sect. 3, presents applications of our result to probability theory in particular to the so-called Lancaster expansions. There is also an appendix in which we recall basic facts about the moments and the moments’ problem and moment sequences.

2 General results

Our main result is the following:

Theorem 1

Let a measure \(\mu \in Cra\), \(\left\{ p_{n}\right\} \) be the sequence of polynomials orthogonal with respect to the measure \(\mu \). Let us consider the orthogonal series (1.3) and by f let us denote the mean-square sum of it. Let \(\mathcal {M}\) be some closed subset of the \( \mathrm{supp\,}( \mu ).\)

The following two conditions are equivalent:

a) \(f\left( x\right) \ge 0\) \(\mu \) - a.s. on \(\mathcal {M}\),

b) There exists \(\nu \in AC2^{+}(\mu )\), with \(\mathrm{supp\,}( \nu )=\mathcal {M}\) such that \(f(x)=\frac{d\nu }{d\mu }(x)\).

If one of the conditions a) and b) is satisfied, then the coefficients \( c_{n} \) are given by the following formula:

$$\begin{aligned} c_{n}=\left( \int _{\mathcal {M}}p_{n}(x)d\nu (x)\right) /\hat{p} _{n}. \end{aligned}$$
(2.1)

Remark 2

Notice, that, if additionally, condition (1.5) is satisfied by the coefficients \(\left\{ c_{n}\right\} \), then the convergence to f is not only in mean-square but also almost surely for almost all \(x\in \mathcal {M}\) \(\mathrm{mod }\mu \).

Remark 3

Notice, that, if \(\left\{ r_{n}\right\} \) denotes the sequence of polynomials orthogonal with respect to the measure \(\nu \), then the number \( \int _M p_{n}(x)d\nu (x)\) is equal to the free coefficient in the connection coefficient expansion of \(p_{n}(x)\) in terms of \(\left\{ r_{n}\right\} .\) More precisely considering connection coefficient expansions

$$\begin{aligned} p_{n}(x)=\sum _{j=0}^{n}\gamma _{n,j}r_{j}(x), \end{aligned}$$
(2.2)

we have

$$\begin{aligned} \int _{M}p_{n}(x)d\nu (x)=\gamma _{n,0}. \end{aligned}$$

Consequently, the assertion of the Theorem 1 can be rephrased in the following way.

An orthogonal series (1.3) with coefficients satisfying condition (), is nonnegative for almost all (mod \(\mu )\) \(x\in M \) if and only if another sequence \(\left\{ r_{n}\right\} \) of orthogonal polynomials can be found such that considering connection coefficient expansion of \(p_{n}(x)\) in terms of \(\left\{ r_{n}\right\} \) given by () we have:

$$\begin{aligned} c_{n}=\gamma _{n,0}/\hat{p}_{n}. \end{aligned}$$
(2.3)

Proof of Theorem 1

b) \(\Rightarrow \) a). First, let us assume that the coefficients \(c_{n}\) are given by (2.1) and let us denote by f(x) the sum of (1.3). By assumptions, we know that it exists and it is square-integrable with respect to \(\mu .\) We have

$$\begin{aligned} \int _{\mathcal {M}}p_{n}(x)f\left( x\right) d\mu (x)=c_{n}\hat{p}_{n}=\int _{\mathcal {M}}p_{n}(x)d\nu (x).\end{aligned}$$

Knowing numbers \(c_{n}\hat{p}_{n}\), \(n=0,1,2\ldots \) and the form of polynomials \(\left\{ p_{n}(x)\right\} \), we can find numbers \(\left\{ \int _{\mathcal {M}}x^{n}d\nu (x)\right\} _{n\ge 0}\) and \(\left\{ \int _{\mathcal {M}}x^{n}f(x)d\mu (x)\right\} \). We see that they are identical and the two measures are, by assumption, identifiable by moments so the two measures must be identical i.e.

$$\begin{aligned} f\left( x\right) d\mu (x)=d\nu (x). \end{aligned}$$

But \(\nu \) was chosen to be nonnegative. So \(f(x)\ge 0\) on the \(\mathcal {M}\) mod \(\mu \). Besides, we see that

$$\begin{aligned} f(x)=\frac{d\nu }{d\mu }(x). \end{aligned}$$

a) \(\Rightarrow \) b). Now, let us assume, that we want to find an expansion of the Radon-Nikodym derivative of two nonnegative measures \(\nu<<\mu \) that is additionally square-integrable (mod \(\mu )\) in an infinite orthogonal series. That is, we are looking for the coefficients of the expansion of the form of (1.3). Then, following our assumptions, we have

$$\begin{aligned} d\nu (x)= & {} f(x)d\mu (x), \\ p_{n}(x)= & {} \sum _{j=0}^{n}\gamma _{n,j}r_{j}(x), \end{aligned}$$

where \(\left\{ r_{n}(x)\right\} \) are polynomials orthogonal with respect to \(\nu \). These polynomials exist since for every positive measure satisfying condition (1.1) one can define such polynomials. Naturally, having two sets of orthogonal polynomials one has a set of connection coefficients between them. Since f is square-integrable with respect to \(d\mu (x)\) we know that the coefficients \(c_{n}\) are defined uniquely. Besides we have

$$\begin{aligned} \int _{\mathcal {M}}f(x)p_{n}(x)d\mu \left( x\right) =c_{n}\hat{p}_{n}=\int _{ \mathcal {M}}p_{n}(x)d\nu (x)=\gamma _{n,0}. \end{aligned}$$

\(\square \)

There are numerous examples of expansions of the type (1.3). They appeared over the years in [17] (Sect. 5 concerning mostly polynomials from the so-called Askey-Wilson scheme) or recently in [20], as well as in [15, 18, 19] (concerning. among others, Charlier (3.7) or Jacobi (3.6) polynomials).

Remark 4

Notice also that coefficient \(\gamma _{n,0}\) is equal to

$$\begin{aligned} \gamma _{n,0}=\sum _{j=0}^{n}\pi _{n,j}m_{j}, \end{aligned}$$
(2.4)

where the coefficients \(\left\{ \pi _{n.j}\right\} \) are defined by the expansion

$$\begin{aligned} p_{n}(x)=\sum _{j=0}^{n}\pi _{n,j}x^{j}, \end{aligned}$$

while the numbers \(\left\{ m_{j}\right\} \) form a moment sequence of some distribution absolutely continuous with respect to the measure \(\mu .\) This observation can be derived directly from (2.1) or from the formula given by Lemma 1 of [16]).

This observation leads also to the following method of checking if a given sequence \(\left\{ c_{n}\right\} \) applied in the series (1.3) can result in the series’ positive-sum. Namely, considering formulae (2.4 ) and (2.3) we can find a sequence \(\left\{ m_{n}\right\} \) by recursively solving a sequence of equations:

$$\begin{aligned} m_{n}=\frac{1}{\pi _{n,n}}\left( \hat{p}_{n}c_{n}-\sum _{j=0}^{n-1}\pi _{n,j}m_{j}\right) , \end{aligned}$$

for \(n\ge 0\). Since \(\left\{ m_{n}\right\} \) has to be a moment sequence, we can apply one of the known criteria some of which are presented in the Appendix.

Remark 5

Continuing the previous remark, the assertion of the theorem (in case when \( \mathcal {M=}\mathrm{supp\,}\mu \) can be expressed in the following way.

There exists a linear map: K :  \(L^{2}(\mu )\longrightarrow L^{2}(\mu )\) that can be symbolically expressed by the following formula:

$$\begin{aligned} K(f)(x)=\int _{\mathrm{supp\,}\mu }\left( \sum _{i\ge 0}p_{i}(x)p_{i}(y)/ \hat{p}_{n}\right) f(y)d\mu (y), \end{aligned}$$

that maps every function \(f\in L^{2}(\mu )\) on itself, since, as it is easily seen, we have:

$$\begin{aligned} \int _{\mathrm{supp\,}\mu }\left( \sum _{i\ge 0}p_{i}(x)p_{i}(y)/\hat{p} _{n}\right) \left( \sum _{n\ge 0}c_{n}p_{n}(y)\right) d\mu (y)=\sum _{n\ge 0}c_{n}p_{n}(x). \end{aligned}$$

3 Probabilistic aspects

In this section, to avoid confusion, we will assume that all considered measures will be probabilistic that is they will integrate up to 1. Further, we will consider bivariate distributions dF(xy) with marginal distributions \(d\mu (x)\) and \(d\nu (y)\) (i.e. \(d\mu (x)=\int F(x,dy)\) and similarly for \(d\nu \)). Naturally, we will assume, that both marginal measures belong to the set Cra in order to be identified by their moments. Moreover, we will consider only such bivariate distributions F satisfying the following condition:

$$\begin{aligned} \int _{\mathrm{supp\,}\mu }\int _{\mathrm{supp\,}\nu }\left( \frac{\partial ^{2}F}{ \partial \mu \partial \nu }(x,y)\right) ^{2}d\mu (x)d\nu (y)<\infty , \end{aligned}$$
(3.1)

where \(\frac{\partial ^{2}F}{\partial \mu \partial \nu }(x,y)\), denotes Radon-Nikodym derivative of the measure F with respect to the product measure \(\mu \times v.\) That is, in other words, that \(dF\in AC2(d\mu \times d\nu ),\) where \(d\mu \times d\nu \) denotes the product measure of \(d\mu \) and \(d\nu .\)

Let us denote by \(\left\{ \alpha _{n}(x)\right\} \) and \(\left\{ \beta _{n}(y)\right\} \) two sets of polynomials orthogonal with respect to the measures respectively \(d\mu (x)\) and \(d\nu (y).\) Now, for all distributions satisfying (3.1) the following expansion is valid:

$$\begin{aligned} dF(x,y)=d\mu (x)d\nu (y)\sum _{i,j=0}\lambda _{i,j}\alpha _{i}(x)\beta _{j}(y), \end{aligned}$$
(3.2)

with \(\sum _{i,j\ge 0}\lambda _{i,j}^{2}\hat{\alpha }_{i}\hat{\beta } _{j}<\infty .\) Conditional distributions \(\zeta (dx|y)\) and \(\xi (dy|x)\) are defined respectively, for almost all y (\(\mathrm{mod }v\)) and almost all x (\(\mathrm{mod }\mu )\) by the following relationships:

$$\begin{aligned} F(dx,dy)=\zeta (dx|y)v(dy)=\xi (dy|x)\mu (dx). \end{aligned}$$
(3.3)

One shows that both these distributions do exist and are defined uniquely respectively \(\mathrm{mod }v\) and \(\mathrm{mod }\mu \).

Notice that making use of the definition of marginal distribution and the orthogonal polynomials and changing, if necessary, the order of integration that we have:

$$\begin{aligned} \int _{\mathrm{supp\,}\mu }\int _{\mathrm{supp\,}v}\beta _{n}(y)F(dx,dy)=\int _{ \mathrm{supp\,}v}\beta _{n}(x)v(dx)=0, \end{aligned}$$

for all \(n\ge 1\) and likewise for polynomials \(\left\{ \alpha _{n}\right\} \) .

Now applying the above-mentioned definitions and properties to the expansion (3.2) we deduce that

$$\begin{aligned} \int \int _{\mathrm{supp\,}\mu \times \mathrm{supp\,}v}dF(x,y)=\lambda _{0,0}=1,~~\forall n\ge 1:\lambda _{0,n}=\lambda _{n,0}=0, \end{aligned}$$
(3.4)

and also, that:

$$\begin{aligned} \zeta (dx|y) = \left( \sum \limits _{{i,j = 0}} {\lambda _{{ij}} } \alpha _{i} (x)\beta _{j} (y)\right) \mu (dx),~\xi \left( dy|x \right) = \left( \sum \limits _{{i,j = 0}} {\lambda _{{ij}} } \alpha _{i} (x)\beta _{j} (y)\right) v\left( dy \right) . \end{aligned}$$
(3.5)

We can now rephrase the above-mentioned Theorem 1 in the form that is important for the probabilists.

Theorem 2

Let \(\mu \in Cra\), and let \(\left\{ \alpha _{n}\right\} \) be a set of polynomials orthogonal with respect to \(\mu \). Then, the orthogonal series:

$$\begin{aligned} g(x,y)=\sum _{n\ge 0}h_{n}(y)\alpha _{n}(x)/\hat{\alpha }_{n}, \end{aligned}$$
(3.6)

where, as above, \(\hat{\alpha }_{n}\) is defined by (1.2) and such that

$$\begin{aligned} \sum _{m\ge 0}\left| h_{n}(y)\right| ^{2}/\hat{\alpha }_{n}<\infty , \end{aligned}$$
(3.7)

for all y belonging to some closed set \(\mathrm{supp\,}\nu \), converges in mean square (mod\(\mu )\) to a nonnegative function iff there exists a family of probability measures \(\zeta (.|y)\) indexed by y,  such that for all \( y\in \mathrm{supp\,}\nu ,\) \(\zeta (.|y)<<\mu \) and \(\forall n\ge 0:\)

$$\begin{aligned} h_{n}(y)=\int _{\mathrm{supp\,}\mu }\alpha _{n}(x)\zeta (dx,y). \end{aligned}$$

Moreover

$$\begin{aligned} d\zeta (x|y)=g(x,y)d\mu (x). \end{aligned}$$

If additionally there exists a probability measure \(\nu \) such that for \( \forall n\ge 0\):

$$\begin{aligned} \int _{\mathrm{supp\,}\nu }h_{n}(y)d\nu (y)=\delta _{n,0}, \end{aligned}$$

then one can define a bivariate measure \(F<<\mu \times \nu \) by the formula

$$\begin{aligned} F(dx,dy)=d\zeta (x|y)d\nu (y)=g(x,y)d\mu (x)d\nu (y) \end{aligned}$$
(3.8)

and for which

$$\begin{aligned} \mu (A)=\int _{\mathrm{supp\,}\nu }\zeta (A|y)d\nu (y), \end{aligned}$$

for all Borel subsets A of \(\mathrm{supp\,}( \mu )\) almost everywhere.

Proof

Suppose, that F satisfies (3.1). Let \(\mu \) and \(\nu \) denote its marginal measures and let the sets \(\left\{ \alpha _{n}\right\} \) and \( \left\{ \beta _{n}\right\} \) denote sets of polynomials orthogonal with respect to measures \(\mu \) and \(\nu \) respectively. Let the conditional distributions be defined by (3.5). Notice that by (3.4) and ( 3.5) we have \(\int _{\mathrm{supp\,}\mu }\zeta (dx,y)=1\), a.e. \((\text {mod}\,\nu \) and similarly for the \(\xi (dy,x))\). Now, changing the order of summation and denoting by

$$\begin{aligned} h_{n}(y)=\hat{\alpha }_{n}\sum _{j=0}^{\infty }\lambda _{nj}\beta _{j}(y), \end{aligned}$$

we have

$$\begin{aligned} dF(x,y)=d\mu (x)d\nu (y)\sum _{n\ge 0}h_{n}(y)\alpha _{n}(x)/\hat{\alpha } _{n}. \end{aligned}$$

Further, utilizing (3.3) and (3.4) we have:

$$\begin{aligned}&\zeta (dx|y)=\left( \sum _{n\ge 0}h_{n}(y)\alpha _{n}(x)/\hat{\alpha }_{n}\right) d\mu (x),\\&\int _{\mathrm{supp\,}v}h_{n}(y)v(dy)=\delta _{n,0}. \end{aligned}$$

with \(\delta _{n,m}\) denoting traditionally Kronecker’s delta. By assumptions concerning polynomials \(\left\{ \alpha _{n}\right\} \) and by Theorem 1 we see that for all \(y\in \mathrm{supp\,}\nu \), we have:

$$\begin{aligned}&h_{n}(y)=\int _{\mathrm{supp\,}\mu }\alpha _{n}(x)d\zeta (x|y),\sum _{j\ge 0}\int _{\mathrm{supp\,}v}\left| h_{j}(y)\right| ^{2}v(dy)/\hat{\alpha }_{j}<\infty , \\&\text {consequently }\sum _{j\ge 0}\left| h_{j}(y)\right| ^{2}/\hat{ \alpha }_{j}<\infty , \end{aligned}$$

a.e. \(\mathrm{mod }v\).

Now let us assume the converse statement, i.e. that we have the converging to a nonnegative function in mean-square, series (3.6) with polynomials \( \left\{ \alpha _{n}\right\} \) and the measure \(\mu ,\) as described in the assumptions, together with the condition (3.7) satisfied for almost every y belonging to some closed set that we will denote by \(\mathrm{supp\,} \nu \). By Theorem 1, we deduce that if the series (3.6) converges to a nonnegative function, then there exists a family \(d\zeta (x|y) \) of positive measures absolutely continuous with respect to \(\mu \) such that \(\forall n\ge 0\):

$$\begin{aligned} h_{n}(y)=\int _{\mathrm{supp\,}\mu }\alpha _{n}(x)\zeta (dx|y). \end{aligned}$$

Moreover, we have

$$\begin{aligned} \zeta (dx|y)=g(x,y)d\mu (x). \end{aligned}$$

Now, if there exists a probability measure \(\nu \) such that \(\forall n\ge 0:\int _{\mathrm{supp\,}\nu }h_{n}(y)v(dy)=\delta _{n,0}\), then we

$$\begin{aligned} \int _{\mathrm{supp\,}\nu }g(x,y)d\nu (y)=1, \end{aligned}$$

a.e. mod \(\mu \), hence

$$\begin{aligned} \int _{\mathrm{supp\,}\nu }\zeta (dx|y)d\nu (y)=d\mu (x), \end{aligned}$$

as claimed. \(\square \)

The rest of this section will be dedicated to the so-called Lancaster expansions. In particular, we will be able to give now necessary and sufficient conditions for these types of expansions to exist. Let us recall that Lancaster, in the series of papers [8,9,10,11], considered and developed the following question: given a bivariate distribution say dF(xy), its two marginal distributions say \(d\mu (x)\) and \(d\nu (y)\) and the two sets of polynomials, when is it possible to find the set of numbers \(\left\{ c_{n}\right\} \) such that

$$\begin{aligned} dF(x,y)=d\mu (x)d\nu (y)\sum _{n\ge 0}c_{n}\alpha _{n}(x)\beta _{n}(y) \end{aligned}$$
(3.9)

almost everywhere in \(\mathrm{supp\,}( \mu )\times \mathrm{supp\,}( \nu )\) with respect to the product measure. In fact, Lancaster in his papers and also his followers in their papers confined the problem to such bivariate distributions dF satisfying condition (3.1).

Definition 1

A class of bivariate distributions with margins identifiable by moments, satisfying (3.1) and having expansion (3.9) will be called Lancaster class (of bivariate distributions), briefly (LC distributions).

Remark 6

Notice, that if F is of LC distribution, then we have:

$$\begin{aligned} \int _{\mathrm{supp\,}\mu }\alpha _{n}(x)dF(dx,y)= & {} c_{n}\beta _{n}(y)d\nu (y), \\ \int _{\mathrm{supp\,}\nu }\beta _{n}(x)dF(x,dy)= & {} c_{n}\alpha _{n}(x)d\mu (x). \end{aligned}$$

In other words, in terms used in probability, we can easily deduce that \( \forall n\ge 1\) the conditional moments, i.e.:

$$\begin{aligned} E(X^{n}|Y)= & {} p_{n}(Y), \\ E(Y^{n}|X)= & {} q_{n}(X), \end{aligned}$$

respectively \(\mathrm{mod }\nu \) and \(\mathrm{mod }\mu \), where \(p_{n}\) and \( q_{n}\), are some polynomials of the full orderFootnote 2n.

Definition 2

Class of bivariate distributions with margins identifiable by moments, having the property that all its conditional moments of the order, say, n are polynomials of the full order n will be called polynomial class (of distributions) briefly PC distributions.

As a corollary we have the following characterization of the Lancaster class of distributions.

Theorem 3

Let us consider a bivariate distribution F satisfying (3.1) with margins identifiable by moments. Then F is an LC distribution iff it is a PC distribution.

Proof

The fact that every distribution of the Lancaster class belongs also to the PC class was noted in Remark 6. So now, let us assume that F belongs to the PC class. By Theorem 2 we know that it can be expanded in the series (3.6). Now we see that \(\forall n\ge 1:h_{n}(y)=\int \alpha _{n}(x)\zeta (dx,y)=E((\alpha _{n}(X)|Y=y).\) But, by our assumption, \(h_{n}(y)\) has to be a polynomial of the full order n i.e.

$$\begin{aligned} h_{n}(y)=\sum _{j=0}^{n}\gamma _{n,j}\beta _{j}(y), \end{aligned}$$

where \(\left\{ \beta _{n}\right\} \) are the polynomials orthogonal with respect to the marginal measure \(\nu .\) Hence we musta have \(\gamma _{n,j}=0\) for \(j>n\). Now changing the order of summation in (3.6) we get:

$$\begin{aligned} F(dx,dy)=d\mu (x)d\nu (y)\sum _{j=0}^{\infty }\beta _{j}(y)\sum _{n\ge j}\gamma _{n,j}\alpha _{n}(x)/\hat{\alpha }_{n}. \end{aligned}$$

But by our assumption \(E((\beta _{j}(Y)|X=x)\) is a polynomial of the full order j. So we have:

$$\begin{aligned} \int _{\mathrm{supp\,}\nu }\beta _{j}(y)F(dx,dy)=d\mu (x)\sum _{n\ge j}\gamma _{n,j}\alpha _{n}(x)/\hat{\alpha }_{n}. \end{aligned}$$

Now, by the uniqueness of expansion, we deduce that \(\forall n>j:\gamma _{n,j}=0.\) \(\square \)

Szabłowski in the series of papers [21,22,23] considered Markov, stochastic processes having two-dimensional finite distributions belonging to PC class. Hence, now, in light of the above-mentioned theorem, there is a possibility of expanding in the Lancaster-like series, the transition functions of such Markov processes.

Let us apply Theorem 2 to the analysis of the LC distributions or more precisely to the analysis when the series

$$\begin{aligned} \sum _{n\ge 0}c_{n}\alpha _{n}(x)\beta _{n}(y) \end{aligned}$$
(3.10)

converges to a nonnegative function of \(\left( x,y\right) ,\) where polynomials \(\left\{ \alpha _{n}\right\} \) and \(\left\{ \beta _{n}\right\} \) are defined as in the introduction to Sect. 3. To simplify the formulation of the theorem and the applications following it, let us assume additionally that both families of polynomials \(\left\{ \alpha _{n}\right\} \) and \(\left\{ \beta _{n}\right\} \) are orthonormal with respect to the measures \(\mu \) and v respectively.

Let us also denote by \(C(\alpha ,\beta )\) set of all sequences \(\left\{ c_{n}\right\} \) for which the sum (3.10) exists and is positive. Koudu in [12] showed that this set is convex (which is trivial, see, e.g. Appendix) and moreover, compact with respect to the weak topology. Hence the Choquet’s theorem about extreme points can be applied. As a corollary of the Theorem 2 we have the following result:

Theorem 4

Let the numbers \(\left\{ a_{n,j}\right\} \) and \(\left\{ b_{n,j}\right\} \) be defined by the polynomials \(\alpha _{n}\) and \(\beta _{n} \) in the following way:

$$\begin{aligned} \alpha _{n}(x)=\sum _{j=0}^{n}a_{n,j}x^{j},~~\beta _{n}(y)=\sum _{j=0}^{n}b_{n,j}y^{j}. \end{aligned}$$
(3.11)

The series (3.10) converges in the mean-squares to a positive function if and only if the following system of recurrent equations

$$\begin{aligned} \sum _{j=0}^{n}a_{n,j}m_{j}^{(a)}(y)= & {} c_{n}\beta _{n}(y), \end{aligned}$$
(3.12)
$$\begin{aligned} \sum _{j=0}^{n}b_{n,j}m_{j}^{(b)}(x)= & {} c_{n}\alpha _{n}(x), \end{aligned}$$
(3.13)

for \(n\ge 0\), is satisfied for almost all \(y\in \mathrm{supp\,}v\) and \(x\in \mathrm{supp\,}\mu \), by the two polynomial moment sequences \(\left\{ m_{n}^{(a)}(y)\right\} \) and \(\left\{ m_{n}^{(b)}(x)\right\} \), that are defined by some measures, that are absolutely continuous with respect to the measures \(\mu \) and v respectively.

Proof

Firstly, under our assumptions we have \(\hat{\alpha }_{n}=\hat{\beta }_{n}=1,\) hence following the previous theorem we deduce that the series (3.10) converges in mean squares to some positive function iff

$$\begin{aligned} E\alpha _{n}(x)=c_{n}\beta _{n}(y), \end{aligned}$$

where the expectation is taken with respect to some absolutely continuous measure that is additionally parametrized by the parameter y,  that belongs to the \(\mathrm{supp\,}v.\) From this remark follows directly the first of the above-mentioned equation. By a similar argument we deduce that the second equation holds. \(\square \)

Remark 7

Notice that if for \(c_{n}=\rho ^{n},\) for \(\left| \rho \right| <1,\) the series (3.10) is convergent to a positive bivariate probability density, then so is the series

$$\begin{aligned} \sum _{j\ge 0} \left( {\int _{{[ - 1,1]}} {\rho ^{j} } d\gamma (\rho )} \right) \alpha _{j} (x)\beta _{j} (y) , \end{aligned}$$

for any probability distribution \(\gamma \) such that \(\gamma (\left\{ -1\right\} \cup \left\{ 1\right\} )=0.\)

Remark 8

In fact, the equations (3.12) and (3.13) should be written in the following, less legible but more precise, recursive way:

$$\begin{aligned} m_{n}^{(a)}(y)= & {} c_{n}\frac{b_{n,n}}{a_{n,n}}y^{n}+ \sum _{j=0}^{n-1}(c_{n}b_{n,j}y^{j}-a_{n,j}m_{j}^{(a)}(y))/a_{n,n}, \end{aligned}$$
(3.14)
$$\begin{aligned} m_{n}^{(b)}(y)= & {} c_{n}\frac{a_{n,n}}{b_{n,n}}y^{n}+ \sum _{j=0}^{n-1}(c_{n}a_{n,j}y^{j}-b_{n,j}m_{j}^{(b)}(y))/b_{n,n}. \end{aligned}$$
(3.15)

for \(n\ge 0\) with \(m_{0}(x)=m_{0}(y)=1\).

Corollary 1

Let coefficients \(\left\{ a_{n,j}\right\} \) and \(\left\{ b_{n,j}\right\} \) be defined by (3.11). If the series (3.10) converges to a positive sum, then

a) \(\sum _{n\ge 0}c_{n}^{2}<\infty ,\)

further additionally:

b) if \(\left\{ 0\right\} \in \mathrm{supp\,}\mu \) and supp v then, we have: \(\infty >\sum _{n\ge 0}c_{n}a_{n,0}b_{n,0}\ge 0,\)

c) if \(\mathrm{supp\,}\mu \) is unbounded, then the sequence \(\left\{ c_{n} \frac{a_{n,n}}{b_{n,n}}\right\} \), is a moment sequence, if additionally \( \mathrm{supp\,}v\) is also unbounded, then \(\left\{ c_{n}^{2}\right\} \) must be a moment sequence.

d) if measures \(\mu \) and v are the same and have unbounded supports, then \(\left\{ c_{n}\right\} \) must be a moment sequence.

Proof

Part a) follows the fact that the series (1.3) converges in mean-squares and that \(\hat{p}_{n}=1\) since we consider only orthonormal polynomials. b) is obvious. c) Firstly, notice that in all cases from the system of equations (3.14) and (3.15) it follows that the leading coefficient in \(m_{n}^{(a)}(y)\) and \( m_{n}^{(b)}(y)\) must be respectively \(c_{n}b_{n,n}/a_{n,n}\) and \( c_{n}a_{n,n}/b_{n,n}.\) Now, if, say \(\mathrm{supp\,}\mu \) is unbounded, then from the fact that \(\left\{ m_{n}^{(b)}(y)\right\} \) must be a moment sequence, then so must be the sequence \(\left\{ y^{n}c_{n}a_{n,n}/b_{n,n}\right\} .\) From this fact the first assertion follows immediately. Now, if both sequences \(\left\{ c_{n}\frac{a_{n,n}}{ b_{n,n}}\right\} ,\) \(\left\{ c_{n}\frac{b_{n,n}}{a_{n,n}}\right\} \) are the moment ones, then is their their product (see the Appendix, below). Part d) follows directly from c). \(\square \)

Remark 9

The assertion d) of the above-mentioned corollary repeats in fact the result of Tyan et al. presented in [25].

Remark 10

Theorem 4, at least theoretically, closes the problem of finding conditions for the convergence to a positive bivariate function of the infinite series (3.10). Namely, having two sequences of moments (what is important given by the recursive formula) one can find their two Laplace transforms of the distributions identified by these sequences and invert them obtaining two conditional measures \(\chi (.|x)\) and \(\zeta (.|y)\) that are also defined by the conditions

$$\begin{aligned} \int \beta _{n}(y)d\chi (y|x)= & {} c_{n}\alpha _{n}(x), \\ \int \alpha _{n}(x)d\zeta (x|y)= & {} c_{n}\beta _{n}(y). \end{aligned}$$

The procedure to get these inverses is very difficult and long. On the way, the procedure utilizes Nevalinna’s theory as described say in [1]. Now, the question of summing the series (3.10) is solved by the formula (3.8).

Example 1

We will now present an example in which we show how having a given family of orthogonal polynomials, and a moment sequence \(\left\{ c_{n}\right\} ,\) one finds a sequence of moments \(\left\{ m_{n}(y)\right\} \). Then, having this sequence one finds a sequence of orthogonal polynomials parametrized by y. Then, by different means, including the analysis of the three-term recurrence of this sequence, one finds the properties of the measure having moment sequence \(\left\{ m_{n}(y)\right\} \) and thus conclude, basing on Theorem 4, that the series (3.10) with \(\alpha _{n}(x)=\beta _{n}(x)\) converges to a positive sum. The way is long and it seems that each case would be enough for an article. To shorten the conclusions and the description, the example will concern Hermite polynomials that are well known and the sequence \(c_{n}=\rho ^{n}\) for some \(\left| \rho \right| <1,\) just to illustrate the process of calculation. Note, that such sequence \(\left\{ c_{n}\right\} \) is a moment sequence. On the way, we will make use of the well-known properties of these polynomials. Besides, in this case it is easy just to guess the conditioning measure \(d\chi (.|y).\)

Let us recall that the so-called (probabilistic) Hermite polynomials are defined by the following three-term recurrence

$$\begin{aligned} H_{n+1}(x)=xH_{n}(x)-nH_{n-1}(x), \end{aligned}$$

with \(H_{0}(x)=1,\) and \(H_{-1}(x)=0.\) It is known, that we have

$$\begin{aligned} \frac{1}{2\pi }\int _{-\infty }^{\infty }H_{n}(x)H_{m}(x)\exp (-x^{2}/2)dx=\delta _{nm}n!. \end{aligned}$$

Moreover, for all complex xy and a the following expansions are true:

$$\begin{aligned} H_{n}(x)= & {} n!\sum _{m=0}^{\left\lfloor n/2\right\rfloor }\frac{ (-1)^{m}x^{n-2m}}{2^{m}m!(n-2m)!}, \end{aligned}$$
(3.16)
$$\begin{aligned} H_{n}(ax+\sqrt{1-a^{2}}y)= & {} \sum _{m=0}^{n}\left( {\begin{array}{c}n\\ m\end{array}}\right) a^{m}(1-a^{2})^{(n-m)/2}H_{m}(x)H_{n-m}(y). \end{aligned}$$
(3.17)

Notice that

$$\begin{aligned} H_{n}(0)=\left\{ \begin{array}{ccc} 0 &{} if &{} n\text { is odd} \\ \frac{(-1)^{k}(2k)!}{2^{k}k!} &{} if &{} n=2k. \end{array} \right. \end{aligned}$$

and also that the orthonormal version of Hermite polynomials is equal to \( H_{n}(x)/\sqrt{n!}\), so for even n we have

$$\begin{aligned} \left| H_{2k}(0)\right| ^{2}/(2k)!=\frac{ (2k)!}{2^{2k}(k!)^{2}}=\frac{1}{2^{2k}}\left( {\begin{array}{c}2k\\ k\end{array}}\right) =\frac{1}{ \sqrt{\pi k}}. \end{aligned}$$

We used here the well-known approximation \(\left( {\begin{array}{c}2k\\ k\end{array}}\right) \frac{1}{2^{2k}} \cong \frac{1}{\sqrt{\pi k}}\). Thus, applying Corollary 1, we see that every applicable sequence \(\left\{ c_{n}\right\} \) must satisfy the following conditions:

$$\begin{aligned} \sum _{n\ge 0}c_{n}^{2}<\infty ,~\infty >\sum _{n\ge 0}c_{n}/\sqrt{n+1}\ge 0, \end{aligned}$$

\(\left\{ c_{n}\right\} \) is a moment sequence.

Now, let us recall Remark 7 and examine the case \( c_{n}=\rho ^{n}\) for some \(\left| \rho \right| <1.\) Notice, that this sequence satisfies the above-mentioned conditions, since \(\sum _{n\ge 0}\rho ^{n}/\sqrt{n+1}=Li(\frac{1}{2};\rho )/\rho \ge 0,\) \(\rho \in (-1,1)\) , where \( Li(s;\rho )\) is the so-called polylogarithm function of order s.

We guess, that the measure v(dx|y) has the density

$$\begin{aligned} g(x|y,\rho )=\frac{1}{\sqrt{2\pi (1-\rho ^{2})}}\exp \left( -\frac{(x-\rho y)^{2}}{ 2(1-\rho ^{2})}\right) . \end{aligned}$$

From this, we immediately deduce, that

$$\begin{aligned} \int _{-\infty }^{\infty }H_{n}\left( \frac{x-\rho y}{\sqrt{1-\rho ^{2}}} \right) g(x|y,\rho )dx=0, \end{aligned}$$

for \(n\ge 1.\) Let us denote:

$$\begin{aligned} m_{n}(y)=\int _{-\infty }^{\infty }x^{n}g(x|y,\rho )dx. \end{aligned}$$

Now we change variables to \(z=\frac{x-\rho y}{\sqrt{ 1-\rho ^{2}}}\) getting \(x=\rho y+\sqrt{1-\rho ^{2}} z. \) Now, applying (3.17) we get:

$$\begin{aligned} m_{n}(y)=\sum _{j=0}^{\left\lfloor n/2\right\rfloor }\left( {\begin{array}{c}n\\ 2j\end{array}}\right) (1-\rho ^{2})^{j}(2j-1)!!\rho ^{n-2j}y^{n-2j}, \end{aligned}$$
(3.18)

since for \(k=0,1,\ldots \) :

$$\begin{aligned} \int _{-\infty }^{\infty }z^{n}\exp \left( -\frac{z^{2}}{2}\right) dz=\left\{ \begin{array}{ccc} 0 &{} if &{} n=2k+1 \\ (2k-1)!! &{} if &{} n=2k \end{array} .\right. \end{aligned}$$

Moreover, we get

$$\begin{aligned} \int _{-\infty }^{\infty }H_{n}(x)g(x|y,\rho )dx=\int _{-\infty }^{\infty }H_{n}(\rho y+z\sqrt{1-\rho ^{2}})\exp \left( -\frac{z^{2} }{2}\right) dz=\rho ^{n}H_{n}(y), \end{aligned}$$

after changing variables, as above, and applying (3.17). Thus, there is no need to check that indeed the sequence (3.18) is the solution of the system of equations (3.14) with \(\alpha _{n}(x)=\beta _{n}(x)=H_{n}(x)\) for all \(n\ge 0 \) and \(c_{n}=\rho ^{n}.\)

By the way, we have also:

$$\begin{aligned} \sum _{n\ge 0}m_{n}(y)\frac{t^{n}}{n!}=\exp \left( -t^{2}(1-\rho ^{2})/2+yt\right) . \end{aligned}$$

Returning to Remark 7 we see that the series

$$\begin{aligned} \sum _{n\ge 0}\frac{c_{n}}{n!}H_{n}(x)H_{n}(y), \end{aligned}$$

converges to a positive density iff \(c_{n}=\int _{-1}^{1}\rho ^{n}d\gamma (\rho ),\) \(n=0,1,\ldots ,\) where \(\gamma \) is some probability distribution on \([-1,1],\) such that \(\gamma (\left\{ -1\right\} \cup \left\{ 1\right\} )=0\). This fact was already noticed by Sarmanov and Bratoeva in 1967 in [14]. Later this result was generalized by Griffith in [7] and Koudu in [13] with Hermite polynomials replaced by polynomials orthogonalizing gamma distribution (Griffith) and Poisson and negative binomial (Koudu). The Koudu’s results were later applied to parameter testing of the chosen Lancaster bivariate distributions by Chen in [6].