1 Introduction

Let \({\mathcal {A}}=(a_{i,j})_{i,j=0}^{\infty }\) be an infinite matrix of real numbers. For \(m\in \mathbb {Z}\), we say that the entry \(a_{i,j}\) lies in the m-diagonal if \(j=i+m\). Obviously, the 0-diagonal is the usual main diagonal of \({\mathcal {A}}\). The matrix \({\mathcal {A}}\) is an m-diagonal matrix if all of its nonzero elements lie in its m-diagonal and lower (upper) triangular matrix if \(a_{i,j}=0\) whenever \(j>i\) (\(j<i\)). The symbols \({\mathcal {A}}^{\textsf{T}}\) and \([{\mathcal {A}}]_n\) denote the transposed matrix and the squared matrix of the first n rows and columns of \({\mathcal {A}}\), respectively. \({\mathcal {I}}\) is called the unit matrix; its (ij)th entry is \(\delta _{i,j}\) where \(\delta _{i,j}=0\) if \(i\ne j\) and \(\delta _{i,i}=1\). \({\mathcal {A}}\) is called positive definite of infinite order if \(\det ([{\mathcal {A}}]_n)>0\) for all \(n\ge 1\), where \(\det ([{\mathcal {A}}]_n)\) is the determinant of \([{\mathcal {A}}]_n\). If \(\det ([{\mathcal {A}}]_n)>0\) for all \(1 \le n\le k\) and \(\det ([{\mathcal {A}}]_n)=0\) for all \( n> k\), we say that \({\mathcal {A}}\) is a positive definite matrix of order k.

According to the definitions given in [5, Ch. II], if \({\mathcal {A}}\) and \({\mathcal {B}}\) are two infinite matrices such that \({\mathcal {A}}\cdot {\mathcal {B}}={\mathcal {I}}\), then \({\mathcal {B}}\) is called a right-hand inverse of \({\mathcal {A}}\), denoted by \({\mathcal {A}}^{-1}\); and \({\mathcal {A}}\) is called a left-hand inverse of \({\mathcal {B}}\), denoted by \(^{-1}{\mathcal {B}}\). The transposes of \({\mathcal {A}}^{-1}\) (\(^{-1}\!{\mathcal {A}}\) ) and \({\mathcal {A}}^{m}\)  (the mth power of the matrix \({\mathcal {A}}\), with \(m \in \mathbb {Z}_{+}\)) are denoted by \({\mathcal {A}}^{-\textsf{T}}\) (\(^{-\textsf{T}}\!{\mathcal {A}}\)) and \({\mathcal {A}}^{m\textsf{T}}\), respectively. Moreover, \({\mathcal {A}}^{-m}= \left( {\mathcal {A}}^{-1}\right) ^m\), where \(m \in \mathbb {Z}_{+}\).

A difficulty of dealing with infinite matrices is that matrix products can be ill-defined (c.f. [4, (1)]). Nevertheless, in this paper we will only consider the product of infinite matrices \({\mathcal {A}}=(a_{i,j})_{i,j=0}^{\infty }\) and \({\mathcal {B}}=(b_{i,j})_{i,j=0}^{\infty }\) when \({\mathcal {A}}{\mathcal {B}}=(\sum _{k} a_{i,k}b_{k,j})_{i,j=0}^{\infty }\) is such that each sum (ij-dependent) involves only a finite number of non-null summands (c.f. [4, Def. 1]).

We will denote by \({\mathcal {U}}\) the infinite matrix whose (ij)th entry is \(\delta _{i+1,j}\) for \(i,j\in \mathbb {Z}_{+}\); i.e. the upper (or backward) shift infinite matrix given by the expression

$$\begin{aligned} {\mathcal {U}}=\left( \begin{array}{cccc} 0 &{} 1 &{} 0 &{} \cdots \\ 0 &{} 0 &{} 1 &{} \cdots \\ 0 &{}0 &{} 0 &{} \cdots \\ \vdots &{} \vdots &{} \vdots &{} \ddots \end{array} \right) . \end{aligned}$$

The matrix \({\mathcal {U}}^{\textsf{T}}\) is called the lower (or forward) shift infinite matrix and it is easy to check that \(\displaystyle {\mathcal {U}}\cdot {\mathcal {U}}^{\textsf{T}}= {\mathcal {I}};\) i.e. \({\mathcal {U}}^{\textsf{T}}={\mathcal {U}}^{-1}\), the right-hand inverse of \({\mathcal {U}}\) (c.f. [9, Sec.   0.9]).

An infinite Hankel matrix is an infinite matrix in which each ascending skew-diagonal from left to right is constant. In other words, \({\mathcal {H}}=(h_{i,j})_{i,j=0}^{\infty }\) is a Hankel matrix if \(h_{i,j+1}=h_{i+1,j}\) for all \(i,j \in \mathbb {Z}_{+}\) or equivalently if

$$\begin{aligned} {\mathcal {U}} {\mathcal {H}}-{\mathcal {H}} {\mathcal {U}}^{-1}={\mathcal {O}}, \end{aligned}$$
(1)

where \({\mathcal {O}}\) denote the infinite null matrix. If \(\{ r_i\}_{i=0}^{\infty }\) is a sequence of real numbers, we denote \({\mathcal {D}}\!\left( r_i\right) \) the infinite diagonal matrix whose ith main diagonal entry is \(r_i\), and by \({\mathcal {H}}\left[ r_i \right] \) the associated Hankel matrix, defined as

$$\begin{aligned} {\mathcal {H}}\left[ r_i \right] =\left( \begin{array}{cccc} r_0 &{} r_1 &{} r_2 &{} \cdots \\ r_1 &{} r_2 &{} r_3 &{} \cdots \\ r_2 &{} r_3 &{} r_4 &{} \cdots \\ \vdots &{} \vdots &{} \vdots &{} \ddots \end{array}\right) . \end{aligned}$$

We say that a matrix \({\mathcal {M}}=(m_{i,j})_{i,j=0}^{\infty }\) is a Hankel-Sobolev matrix if there exists a sequence of Hankel matrices \(\left\{ {\mathcal {H}}_k\right\} _{k=0}^{\infty }\) such that

$$\begin{aligned} {\mathcal {M}}= \sum _{k=0}^{\infty } \left( {\mathcal {U}}^{-k}\;{\mathcal {D}}_k \; {\mathcal {H}}_{k} \; {\mathcal {D}}_k\;{\mathcal {U}}^{k}\right) , \end{aligned}$$
(2)

where \({\mathcal {D}}_0={\mathcal {I}} \) and \({\mathcal {D}}_k={\mathcal {D}}\!\left( \frac{(k+i)!}{i!}\right) \) for each \(k>0\), e.g.

$$\begin{aligned} {\mathcal {D}}_1=\begin{pmatrix}1 &{} 0 &{} 0 &{} \cdots \\ 0 &{} 2 &{} 0 &{} \cdots \\ 0 &{} 0 &{} 3 &{}\cdots \\ \vdots &{}\vdots &{} \vdots &{} \ddots \end{pmatrix}, \; {\mathcal {D}}_2=\begin{pmatrix}2 &{} 0 &{} 0 &{}\cdots \\ 0 &{} 6 &{} 0 &{} \cdots \\ 0 &{} 0 &{} 12 &{} \cdots \\ \vdots &{}\vdots &{} \vdots &{}\ddots \end{pmatrix} \text { and } {\mathcal {D}}_3=\begin{pmatrix}6 &{} 0 &{} 0 &{}\cdots \\ 0 &{} 24 &{} 0 &{} \cdots \\ 0 &{} 0 &{} 60 &{} \cdots \\ \vdots &{}\vdots &{} \vdots &{} \ddots \end{pmatrix}. \end{aligned}$$

We say that a Hankel-Sobolev matrix \({\mathcal {M}}\) is of index \(d \in \mathbb {Z}_{+}\) if \({\mathcal {H}}_d\ne {\mathcal {O}}\) and \({\mathcal {H}}_k={\mathcal {O}}\) for all \(k>d\). Otherwise, we will say that \({\mathcal {M}}\) is of infinite index.

Let \({\mathcal {M}}\) be a Hankel-Sobolev matrix of index \(d \in \overline{\mathbb {Z}}_{+}=\mathbb {Z}_{+}\cup \{\infty \}\). If \({\mathcal {H}}_k\ne {\mathcal {O}}\) for all \(k<d\), we say that \({\mathcal {M}}\) is non-lacunary and lacunary in any other case.

Hankel-Sobolev matrices appeared for the first time in [3, 16] in close connection with the moment problem for a Sobolev inner product. Some of the properties of this class of infinite matrices have also been studied in [6, 13, 14, 17, 19, 22].

Let \(\mathbb {M}\) be the linear space of all infinite matrices of real numbers. For each \(\eta \in \mathbb {Z}_{+}\) fixed, we denote by \(\Phi ( \cdot ,\eta )\) the operator from \(\mathbb {M}\) to itself given by the expression

$$\begin{aligned} \Phi ( {\mathcal {A}},\eta ):=&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell }, \end{aligned}$$
(3)

where \({\mathcal {A}} \in \mathbb {M}\) and \(\left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) \) denote binomial coefficients. Obviously, \(\Phi ( \cdot ,\eta )\) is a linear operator.

One of the main results of this work is the following intrinsic characterization of the Hankel-Sobolev matrices using the operator \(\Phi ( \cdot ,\eta )\), which we will be prove in section 2.

Theorem 1

An infinite matrix \({\mathcal {M}}\) is a Hankel-Sobolev matrix of index \(d \in {\mathbb {Z}_{+}}\), if and only if \({\mathcal {M}}\) is a symmetric matrix and

$$\begin{aligned} \Phi ( {\mathcal {M}},2d+1)={\mathcal {O}} \quad \text { and } \quad \Phi ( {\mathcal {M}},2d)\ne {\mathcal {O}}. \end{aligned}$$
(4)

Moreover, for \(k=0,\,1,\dots ,\;d\); the Hankel matrix \({\mathcal {H}}_{d-k}\) in (2) is given by

$$\begin{aligned} {\mathcal {H}}_{d-k}= \frac{(-1)^{d-k}}{(2d-2k)!}\; \Phi ( {\mathcal {M}}_{d-k},2 d-2k), \end{aligned}$$
(5)

where \(\displaystyle {\mathcal {M}}_d={\mathcal {M}}\) and \(\displaystyle {\mathcal {M}}_{d-k}={\mathcal {M}}_{d-k+1}- {\mathcal {U}}^{-d-1+k} {\mathcal {D}}_{d+1-k} {\mathcal {H}}_{d+1-k} {\mathcal {D}}_{d+1-k} {\mathcal {U}}^{d+1-k}\) for \(k=1,\;2,\dots , \; d\).

An infinite matrix \(\displaystyle {\mathcal {G}}=(g_{i,j})_{i,j=0}^{\infty }\) is a lower Hessenberg infinite matrix if \(g_{i,j}=0\) whenever \(j-i>1\) and at least one entry of the 1-diagonal is different from zero, i.e.

$$\begin{aligned} {\mathcal {G}}= \left( \begin{array}{cccccc} g_{0,0} &{} g_{0, 1} &{} 0 &{}\cdots &{} 0 &{}\cdots \\ g_{1,0} &{} g_{1,1} &{} g_{1, 2} &{}\cdots &{} 0 &{}\cdots \\ \vdots &{} \vdots &{} \vdots &{} \ddots &{} \vdots &{} \cdots \\ g_{n,0} &{} g_{n,1} &{} g_{n,2} &{}\cdots &{} g_{n, n+1} &{} \cdots \\ \vdots &{} \vdots &{} \vdots &{} \vdots &{} \vdots &{} \ddots \end{array} \right) . \end{aligned}$$
(6)

Let us denote by \(\mathbb {H}\) the set of all lower Hessenberg infinite matrices. Additionally, if \({\mathcal {G}} \in \mathbb {H}\) and all the entries in the 1-diagonal are equal to 1, we say that \({\mathcal {G}}\) is monic. If all the entries of the 1-diagonal of \({\mathcal {G}}\) are nonzero, we say that \({\mathcal {G}}\) is a non-degenerate lower Hessenberg infinite matrix (for brevity hereafter referred as non-degenerate Hessenberg matrix). An upper Hessenberg matrix is a matrix whose transpose is a lower Hessenberg matrix.

For each \(\eta \in \mathbb {Z}_{+}\) fixed, we denote by \(\Psi ( \cdot ,\eta )\) the operator from \(\mathbb {H}\) to \(\mathbb {M}\) given by the expression

$$\begin{aligned} \Psi ({\mathcal {B}}, \eta )=&\sum _{k=0}^{\eta }(-1)^{k}\, \left( {\begin{array}{c}\eta \\ k\end{array}}\right) \;{\mathcal {B}}^{k}\; {\mathcal {B}}^{(\eta -k)\; \textsf{T}},\; {\mathcal {B}} \in \mathbb {H}. \end{aligned}$$
(7)

Theorem 9 establishes the relation between the operators (3) and (7).

Given a non-degenerate matrix \({\mathcal {G}}\in \mathbb {H}\), we can generate a sequence of polynomials \(\displaystyle \{Q_n\}_{n=0}^{\infty }\) as follows. Assume that \(Q_{0}(z) = t_{0,0}>0\), then

$$\begin{aligned} g_{0, 1} Q_{1}(x) =&x Q_0(x)-g_{0,0} Q_0(x), \nonumber \\ g_{1, 2} Q_2(x) =&xQ_1(x)-g_{1,1} Q_1(x)- g_{1,0} Q_0(x), \nonumber \\ \vdots \qquad \vdots&\qquad \vdots \nonumber \\ g_{n, n+1} Q_{n+1}(x) =&x Q_{n}(x) - \sum _{k=0}^{n}g_{n,k} \, Q_{k}(x) ,\\ \vdots \qquad \vdots&\qquad \vdots \nonumber \end{aligned}$$
(8)

Hereafter, we will say that \(\displaystyle \{Q_n\}\) is the sequence of polynomials generated by \(\displaystyle {\mathcal {G}}\). As \({\mathcal {G}}\) is non-degenerate, \(Q_n\) is a polynomial of degree n.

Let \({\mathcal {T}}\) be the lower triangular infinite matrix whose entries are the coefficients of the sequence of polynomials \(\{Q_n\}\), i.e.

$$\begin{aligned} {\mathcal {Q}}(x)= {\mathcal {T}}\;{\mathcal {P}}(x) \quad \text { where } \quad {\mathcal {T}}= \left( \begin{array}{ccccc} t_{0,0} &{} 0 &{} \cdots &{} 0 &{}\cdots \\ t_{1,0} &{} t_{1, 1} &{} \cdots &{} 0 &{}\cdots \\ \vdots &{} \vdots &{} \ddots &{} \vdots &{}\vdots \\ t_{n,0} &{} t_{n,1} &{} \cdots &{} t_{n, n} &{}\cdots \\ \vdots &{} \vdots &{} \cdots &{} \vdots &{}\ddots \end{array} \right) , \end{aligned}$$
(9)

\({\mathcal {Q}}(x)=(Q_0(x),Q_1(x),\cdots ,Q_n(x), \cdots )^{\textsf{T}}\) and \({\mathcal {P}}(x)=(1,x,\cdots ,x^n, \cdots )^{\textsf{T}}\).

As \({\mathcal {G}}\) is non-degenerate, \(t_{i,i}=t_{0,0}\,\left( \prod _{k=0}^{i-1} g_{k,k+1}\right) ^{-1}\ne 0\) for all \(i \ge 1\). Therefore, there exists a unique lower triangular infinite matrix \({\mathcal {T}}^{-1}\) such that \({\mathcal {T}}\cdot {\mathcal {T}}^{-1}={\mathcal {I}}\) (c.f. [5, (2.1.I)]), i.e. \({\mathcal {T}}\) has a unique right-hand inverse. Furthermore, in this case \({\mathcal {T}}^{-1}\) is also a left-hand inverse of \({\mathcal {T}}\) and it is its only two-sided inverse (c.f. [5, Remark (a) pp. 22]),

$$\begin{aligned} {\mathcal {T}}^{-1}= \left( \begin{array}{ccccc} \tau _{0,0} &{} 0 &{} \cdots &{} 0 &{}\cdots \\ \tau _{1,0} &{} \tau _{1, 1} &{} \cdots &{} 0 &{}\cdots \\ \vdots &{} \vdots &{} \ddots &{} \vdots &{}\vdots \\ \tau _{n,0} &{} \tau _{n,1} &{} \cdots &{} \tau _{n, n} &{}\cdots \\ \vdots &{} \vdots &{} \cdots &{} \vdots &{}\ddots \end{array} \right) , \; \text {where}\quad \tau _{i,i}=t_{i,i}^{-1}.\end{aligned}$$
(10)

We will denote by \({\mathcal {M}}_{{\mathcal {G}}}\) the matrix of formal moments associated with \({\mathcal {G}}\) (a non-degenerate Hessenberg matrix) defined by

$$\begin{aligned} {\mathcal {M}}_{{\mathcal {G}}}= \left( m_{i,j}\right) _{i,j=0}^{\infty }={\mathcal {T}}^{-1}\; {\mathcal {T}}^{-\textsf{T}}. \end{aligned}$$
(11)

We say that a non-degenerate Hessenberg matrix \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in {\overline{\mathbb {Z}}_{+}}\) if its associated matrix of formal moments is a Hankel-Sobolev matrix of index d. In the following theorem, we give a characterization of these matrices.

Theorem 2

A non-degenerate Hessenberg matrix \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in {\mathbb {Z}_{+}}\), if and only if

$$\begin{aligned} \Psi ( {\mathcal {G}},2d+1)={\mathcal {O}} \quad \text { and } \quad \Psi ( {\mathcal {G}},2d)\ne {\mathcal {O}}.\end{aligned}$$

The proof of this theorem is an immediate consequence of Theorem 1 and Theorem 9 (stated in section 3).

Let \(\mathbb {P}\) be the linear space of polynomials with real coefficients, \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix and \({\mathcal {M}}_{{\mathcal {G}}}\) its associated matrix of formal moments. If \(p(x)=\sum _{i=0} ^{n_1} a_i x^i\) and \(q(x)=\sum _{j=0} ^{n_2} b_j x^j\) are two polynomials in \(\mathbb {P}\) of degree \(n_1\) and \(n_2\), respectively. Then, the bilinear form

$$\begin{aligned} \langle p,q \rangle _{{\mathcal {G}}}= (a_0,\cdots ,a_{n_1},0,\cdots ) {\mathcal {M}}_{{\mathcal {G}}} (b_0,\cdots ,b_{n_2},0,\cdots )^{\textsf{T}} = \sum _{i=0} ^{n_1} \sum _{j=0} ^{n_2} a_i \,m_{i,j}\,{b_j};\nonumber \\ \end{aligned}$$
(12)

defines an inner product on \(\mathbb {P}\) and \(\Vert \cdot \Vert _{{\mathcal {G}}}=\sqrt{\langle \cdot ,\cdot \rangle _{{\mathcal {G}}}}\) is the norm induced by (12) on \(\mathbb {P}\) (c.f. Theorem 7).

Let \(d\in \mathbb {Z}_{+}\) and \(\vec {\varvec{\mu }}_d= (\mu _0,\mu _1,\dots ,\mu _d)\) be a vector of \(d+1\) measures, we write \(\vec {\varvec{\mu }}_d \in {\mathfrak {M}}_{d}(\mathbb {R})\) if for each k (\(0\le k \le d\)) the measure \(\mu _k\) is a non-negative finite Borel measure with support \(\Delta _k \subset \mathbb {R}\), \(\mathbb {P}\subset {\textbf{L}}_1\left( \mu _k\right) \), \(\mu _0\) is positive and \(\Delta _0\) contains infinite points. If \(d=\infty \) and \(\vec {\varvec{\mu }}_{\infty }\) is a sequence of measures that satisfy the above conditions, we write \(\vec {\varvec{\mu }}_{\infty }\in {\mathfrak {M}}_{\infty }(\mathbb {R})\).

For \(d\in \overline{\mathbb {Z}}_{+}\) and \(\vec {\varvec{\mu }}_d \in {\mathfrak {M}}_{d}(\mathbb {R})\), we define on \(\mathbb {P}\) the Sobolev inner product

$$\begin{aligned} \langle f, g \rangle _{\vec {\varvec{\mu }}_d} = \sum _{k=0}^{d} \int f^{(k)}(x) g^{(k)}(x) d\mu _{k}(x) = \sum _{k=0}^{d} \langle f^{(k)}, g^{(k)}\rangle _{\mu _k}\;, \end{aligned}$$
(13)

where \(f,g \in \mathbb {P}\) and \(f^{(k)}\) denote the kth derivative of f. The symbol \(\Vert \cdot \Vert _{\vec {\varvec{\mu }}_d}=\sqrt{\langle \cdot , \cdot \rangle _{\vec {\varvec{\mu }}_d}}\) denotes the Sobolev norm associated with (13). Note that although d is usually considered a non-negative integer (c.f. [12]), the case \(d=\infty \) has sense on \(\mathbb {P}\). If all the measures \(\mu _k\) involved in (13) are positive, we say that the Sobolev inner product is non-lacunary and lacunary in any other case.

Taking into account the nature of the support of the measures involved in (13), we have the following three cases:

Continuous case.:

The measures \(\mu _0, \cdots , \mu _d\) are supported on infinite subsets.

Discrete case.:

The support of the measure \(\mu _0\) is an infinite subset and the measures \(\mu _1, \cdots , \mu _d\) are supported on finite subsets.

Discrete-continuous case.:

The support of the measure \(\mu _d\) is an infinite subset and the measures \(\mu _0, \cdots , \mu _{d-1}\) are supported on finite subsets.

The notion of Sobolev moment and several related topics were firstly introduced in [3]. The (nk)-moment associated with the inner product (13) is defined as \(s_{n,k}=\langle x^n,x^k\rangle _{\vec {\varvec{\mu }}_d}\) (\(n,k\ge 0\)), provided the integral exists. In [3], it was proved that the infinite matrix of moments \({\mathcal {S}}\) with entries \(s_{n,k}\), (\(n,k\ge 0\)) is a Hankel-Sobolev matrix (c.f. [3] and Sect. 2.1 of this paper). Furthermore, if \({Q_n}\) is the sequence of orthonormal polynomials with respect to (13) with leading coefficient \(c_{n} >0\), then the infinite matrix \({\mathcal {G}}_{\vec {\varvec{\mu }}_d}\) with entries \(g_{i,j}=\langle xQ_i, Q_j \rangle _{\vec {\varvec{\mu }}_d} \) is a non-degenerate Hessenberg matrix. In this case, the sequence of orthonormal polynomials \({Q_n}\) is the sequence of polynomials generated by \({\mathcal {G}}_{\vec {\varvec{\mu }}_d}\).

The following theorem gives a characterization of the non-degenerate Hessenberg matrices whose sequence of generated polynomials is orthogonal with respect to a Sobolev inner product as (13).

Theorem 3

(Favard type theorem for continuous case) Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix. Then, there exists \(d \in \mathbb {Z}_{+}\) and \(\vec {\varvec{\mu }}_d \in {\mathfrak {M}}_{d}(\mathbb {R})\) such that \( \langle p, q \rangle _{\vec {\varvec{\mu }}_d} =\langle p, q \rangle _{{\mathcal {G}}} \) if and only if  

  1. 1.

    \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in \mathbb {Z}_{+}\).

  2. 2.

    For each \(k=0,\,1,\dots ,\;d\); the Hankel matrix \({\mathcal {H}}_{d-k}\) defined by (5), is a positive definite matrix of infinite order.

The Favard type theorems for the cases discrete and the discrete-continuous are Theorems 11 and 12, respectively. Some basic aspects about the classical moment problem and the Sobolev moment problem are revisited in Sect. 2.1.

In Sect. 2, we proceed with the study of the properties of the matrix operator \(\Phi (\cdot ,\eta )\), the Hankel-Sobolev matrices and the proof of Theorem 1. We revisit the Sobolev moment problem in Sect. 2.1. The third section is devoted to study the properties of the bilinear form (12) and the nexus between the operators \(\Phi (\cdot ,\eta )\) and \(\Psi ( \cdot ,\eta )\). In the last section, we prove the extension of the Favard Theorem for Sobolev orthogonality stated in Theorem 3.

2 Hankel-Sobolev Matrices

First of all, we need to prove that the notion of a Hankel-Sobolev matrix introduced in (2) is well defined.

Proposition 2.1

Let \({\mathcal {M}}\) be a Hankel-Sobolev matrix, then the decomposition of \({\mathcal {M}}\) established in (2) is unique.

Proof

We first recall that for each \(k \in \mathbb {Z}_{+}\), \({\mathcal {D}}_k\) is a diagonal matrix with positive entries in the main diagonal. Furthermore, if \({\mathcal {A}}\) is an infinite matrix and \(k\in \mathbb {Z}_{+}\) is fixed, the matrix \(\left( {\mathcal {U}}^{-k}\; {\mathcal {A}} \;{\mathcal {U}}^{k}\right) \) is obtained adding to \({\mathcal {A}}\) the first k rows and columns of zeros.

Suppose there are two sequences of Hankel matrices, \(\left\{ {\mathcal {H}}_k\right\} _{k=0}^{\infty }\) and \(\left\{ \widehat{{\mathcal {H}}}_k\right\} _{k=0}^{\infty }\), such that

$$\begin{aligned} {\mathcal {M}}= \sum _{k=0}^{\infty } \left( {\mathcal {U}}^{-k}\;{\mathcal {D}}_k \; {\mathcal {H}}_{k} \; {\mathcal {D}}_k\;{\mathcal {U}}^{k}\right) \quad \text { and }\quad {\mathcal {M}}= \sum _{k=0}^{\infty } \left( {\mathcal {U}}^{-k}\;{\mathcal {D}}_k \; \widehat{{\mathcal {H}}}_{k} \; {\mathcal {D}}_k\;{\mathcal {U}}^{k}\right) . \end{aligned}$$

Therefore,

$$\begin{aligned} \sum _{k=0}^{\infty } \left( {\mathcal {U}}^{-k}\;{\mathcal {D}}_k \; \left( {\mathcal {H}}_{k} -\widehat{{\mathcal {H}}}_{k} \right) \; {\mathcal {D}}_k\;{\mathcal {U}}^{k}\right) ={\mathcal {O}}. \end{aligned}$$

Hence, for each \(k\in \mathbb {Z}_{+}\) fixed, the matrix \(\left( {\mathcal {H}}_{k} -\widehat{{\mathcal {H}}}_{k} \right) \) is a Hankel matrix whose first row has all its entries equal to zero, i.e. \( {\mathcal {H}}_{k} =\widehat{{\mathcal {H}}}_{k}\), which completes the proof. \(\square \)

Obviously, the matrix operator \(\Phi ( \cdot ,\eta )\) defined in (3) is linear. Before proving Theorem 1, we need to study some other properties of this operator and some auxiliary results.

Proposition 2.2

(Recurrence) Let \(\eta \in \mathbb {Z}_{+}\) fixed and \({\mathcal {A}} \in \mathbb {M}\), then

$$\begin{aligned} \Phi ( {\mathcal {A}},\eta +1)={\mathcal {U}}\,\Phi ( {\mathcal {A}},\eta )-\Phi ( {\mathcal {A}},\eta ) \,{\mathcal {U}}^{-1} . \end{aligned}$$
(14)

Proof

$$\begin{aligned} \Phi ( {\mathcal {A}},\eta +1)&= {\mathcal {U}}^{\eta +1} {\mathcal {A}} + \left( \sum _{\ell =1}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta +1\\ \ell \end{array}}\right) {\mathcal {U}}^{\eta +1-\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell } \right) + (-1)^{\eta +1} {\mathcal {A}} {\mathcal {U}}^{-\eta -1}\\&= {\mathcal {U}}^{\eta +1} {\mathcal {A}} + \left( \sum _{\ell =1}^{\eta } (-1)^\ell \left( \left( {\begin{array}{c}\eta \\ \ell -1\end{array}}\right) + \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) \right) {\mathcal {U}}^{\eta +1-\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell } \right) \\&\quad + (-1)^{\eta +1} {\mathcal {A}} {\mathcal {U}}^{-\eta -1}\\&= \sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta +1-\ell }{\mathcal {A}} {\mathcal {U}}^{-\ell } +\sum _{\ell =1}^{\eta +1} (-1)^{\ell } \left( {\begin{array}{c}\eta \\ \ell -1\end{array}}\right) {\mathcal {U}}^{\eta +1-\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell }\\&= {\mathcal {U}} \left( \sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell } \right) -\left( \sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell }\right) {\mathcal {U}}^{-1}\\&= {\mathcal {U}} \, \Phi ( {\mathcal {A}},\eta )-\Phi ( {\mathcal {A}},\eta ) {\mathcal {U}}^{-1} \end{aligned}$$

\(\square \)

The following proposition is an immediate consequence of Proposition 2.2 and (1).

Proposition 2.3

If for a matrix \({\mathcal {A}} \in \mathbb {M}\), there exists \(\eta \in \mathbb {Z}_{+}\) such that \(\Phi ( {\mathcal {A}},\eta )={\mathcal {O}}\), then

  1. (a)

    \(\Phi ( {\mathcal {A}},\eta _1)={\mathcal {O}}\) for all \(\eta _1 \ge \eta .\)

  2. (b)

    For all \(c\in \mathbb {R}\) and \(\eta \ge 1\), the matrix \(c\,\Phi ( {\mathcal {A}},\eta -1)\) is a Hankel matrix.

Proposition 2.4

Assume that \({\mathcal {A}}\in \mathbb {M}\) is a symmetric matrix, then \(\Phi ( {\mathcal {A}},\eta )\) is a symmetric (antisymmetric) matrix if and only if \(\eta \) is an even (odd) integer number.

Proof

$$\begin{aligned} \left( \Phi ( {\mathcal {A}},\eta )\right) ^{\textsf{T}}=&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) \left( {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell } \right) ^{\textsf{T}}= \sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}\eta \\ \eta -\ell \end{array}}\right) {\mathcal {U}}^{\ell } {\mathcal {A}} {\mathcal {U}}^{\ell -\eta } \\ =&\sum _{\ell =0}^{\eta } (-1)^{\eta -\ell } \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell }= (-1)^{\eta } \sum _{\ell =0}^{\eta } (-1)^{\ell } \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {A}} {\mathcal {U}}^{-\ell } \\ =&(-1)^{\eta } \Phi ( {\mathcal {A}},\eta ).\\ \end{aligned}$$

\(\square \)

Theorem 4

Let \(d\in \overline{\mathbb {Z}}_{+}\) and \({\mathcal {M}}\) be a Hankel-Sobolev matrix of index d, as in (2). Denote

$$\begin{aligned} {\mathcal {M}}_\eta = \sum _{k=0}^\eta {\mathcal {U}}^{-k} {\mathcal {D}}_k {\mathcal {H}}_{k} {\mathcal {D}}_k {\mathcal {U}}^{k}, \quad 0\le \eta \le d. \end{aligned}$$

Then,

  1. (a)

    \(\Phi ( {\mathcal {M}}_\eta ,2\eta +1)={\mathcal {O}}\).

  2. (b)

    \(\displaystyle {\mathcal {H}}_\eta = \frac{(-1)^\eta }{(2\eta )!}\Phi ( {\mathcal {M}}_\eta ,2\eta )\).

Before proving the previous theorem, we need the next two lemmas. The first one is a version of the famous Euler’s finite difference theorem (c.f. [18, Sec. 6.1]).

Lemma 2.1

Let \(\displaystyle f(z)\) be a complex polynomial of degree n and leading coefficient \(a_n \in \mathbb {C}\). Then, for all \(\nu \in \mathbb {Z}_{+}\)

$$\begin{aligned} \sum _{\ell =0}^{\nu } (-1)^{\ell } \left( {\begin{array}{c}\nu \\ \ell \end{array}}\right) f(\ell )=\left\{ \begin{array}{rl} 0, &{}\text {if } 0 \le n < \nu , \\ (-1)^{\nu } \;\nu ! \; a_ \nu , &{}\text {if } n=\nu . \end{array}\right. \end{aligned}$$

Lemma 2.2

Let \({\mathcal {A}}=[a_{i,j}]\) be an infinite symmetric matrix, whose (ij) entry is \(a_{i,j}\). Then, for all \(\eta ,\nu \in \mathbb {Z}_{+}\), the matrix \(\left( {\mathcal {U}}^{\eta } \,{\mathcal {A}}\,{\mathcal {U}}^{-\nu }\right) \) is symmetric.

Proof

Given a sequence of double indexes \(\{a_{i,j}\}\) and two functions \(f,\,g\) on \(\mathbb {Z}_{+}\), we denote by \({\mathcal {A}}=[a_{f(i),g(j)}]\) the corresponding infinite matrix, whose (ij) entry is \(a_{f(i),g(j)}\). Therefore, as \(a_{i,j}=a_{j,i}\) we get

$$\begin{aligned} {\mathcal {U}}^{\eta } \, {\mathcal {A}}\,{\mathcal {U}}^{-\nu }{} & {} = \left[ a_{\eta +i,\nu +j}\right] =\left[ a_{\nu +j,\eta +i}\right] ={\mathcal {U}}^{\nu } \, {\mathcal {A}}\,{\mathcal {U}}^{-\eta }\\{} & {} =\left( {\mathcal {U}}^{\eta } \, {\mathcal {A}}\,{\mathcal {U}}^{-\nu }\right) ^{\textsf{T}}; \; i,j=1,2,\dots . \end{aligned}$$

\(\square \)

Proof of Theorem 4

Let \(0\le k \le d\) fixed and \(\displaystyle {\mathcal {R}}_k={\mathcal {U}}^{-k}{\mathcal {D}}_k {\mathcal {H}}_{k} {\mathcal {D}}_k{\mathcal {U}}^{k} \). Then, \(\displaystyle {\mathcal {M}}_\eta = \sum _{k=0}^\eta {\mathcal {R}}_k \) and from linearity \(\displaystyle \Phi ( {\mathcal {M}}_\eta ,2\eta +1)= \sum _{k=0}^\eta \Phi \left( {\mathcal {R}}_k,2\eta +1 \right) \).

$$\begin{aligned} \Phi ({\mathcal {R}}_k,2\eta +1)=&\sum _{\ell =0}^{2\eta +1} (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } \nonumber \\ =&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell }\nonumber \\&+ \sum _{\ell =\eta +1}^{2\eta +1} (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } \nonumber \\ =&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } \nonumber \\&+ \sum _{\ell =\eta +1}^{2\eta +1} (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ 2\eta +1-\ell \end{array}}\right) {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } \nonumber \\ =&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) \; {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } \nonumber \\ {}&\quad - \sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) \; {\mathcal {U}}^{\ell }{\mathcal {R}}_k{\mathcal {U}}^{\ell -2\eta -1} \nonumber \\ =&\sum _{\ell =0}^{\eta } (-1)^\ell \left( {\begin{array}{c}2\eta +1\\ \ell \end{array}}\right) \; \left( {\mathcal {U}}^{2\eta +1-\ell }{\mathcal {R}}_k{\mathcal {U}}^{-\ell } - {\mathcal {U}}^{\ell } {\mathcal {R}}_k{\mathcal {U}}^{\ell -2\eta -1} \right) . \end{aligned}$$
(15)

Clearly, \(\displaystyle {\mathcal {R}}_k\) is a symmetric matrix. Therefore, from Lemma 2.2, \( \displaystyle {\mathcal {U}}^{2\eta +1-\ell } {\mathcal {R}}_k{\mathcal {U}}^{-\ell } \) is a symmetric matrix too and

$$\begin{aligned} {\mathcal {U}}^{2\eta +1-\ell } {\mathcal {R}}_k{\mathcal {U}}^{-\ell } = \left( {\mathcal {U}}^{2\eta +1-\ell } {\mathcal {R}}_k{\mathcal {U}}^{-\ell } \right) ^{\textsf{T}}= {\mathcal {U}}^{\ell } {\mathcal {R}}_k{\mathcal {U}}^{\ell -2\eta -1}. \end{aligned}$$
(16)

Combining (15)–(16), we get (a) in Theorem 4.

From (a) and Proposition 2.2, it follows that

$$\begin{aligned} \Phi ( {\mathcal {M}}_\eta ,2\eta )=&\Phi ( {\mathcal {R}}_\eta ,2\eta ) + \sum _{k=0}^{\eta -1} \Phi ( {\mathcal {R}}_k,2\eta ) \\ =&\Phi ( {\mathcal {R}}_\eta ,2\eta ) + {\mathcal {U}} \sum _{k=0}^{\eta -1} \Phi ( {\mathcal {R}}_k,2\eta -1)- \sum _{k=0}^{\eta -1} \Phi ( {\mathcal {R}}_k,2\eta -1) {\mathcal {U}}^{-1}\\ =&\Phi ( {\mathcal {R}}_\eta ,2\eta ) + {\mathcal {U}} \, \Phi ( {\mathcal {M}}_{\eta -1},2\eta -1)- \Phi ( {\mathcal {M}}_{\eta -1},2\eta -1) {\mathcal {U}}^{-1}\\&= \Phi ( {\mathcal {R}}_\eta ,2\eta ) .\\ \Phi ( {\mathcal {R}}_\eta ,2\eta ) =&\sum _{\ell =0}^{2\eta } (-1)^\ell \left( {\begin{array}{c}2\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{2\eta -\ell } {\mathcal {R}}_\eta {\mathcal {U}}^{-\ell }= \sum _{\ell =0}^{2\eta } (-1)^\ell \left( {\begin{array}{c}2\eta \\ \ell \end{array}}\right) {\mathcal {U}}^{\eta -\ell } {\mathcal {D}}_\eta {\mathcal {H}}_{\eta } {\mathcal {D}}_\eta {\mathcal {U}}^{\eta -\ell }. \end{aligned}$$

Let \(\{m_{k,i}\}\) be the sequence of real numbers such that \(m_{k,i+j-2}\) is the (ij)th entry of the infinite Hankel matrix \( {\mathcal {H}}_{k}\), where \(i,j=1,2,\dots \) In that case, we write \( \displaystyle {\mathcal {H}}_{k}=\left[ m_{k,i+j-2}\right] \). Thus,

$$\begin{aligned} {\mathcal {D}}_\eta {\mathcal {H}}_{\eta } {\mathcal {D}}_\eta =&\left[ \frac{(\eta +i-1)!}{(i-1)!}\frac{(\eta +j-1)!}{(j-1)!}m_{\eta ,i+j-2}\right] \\ =&\left( \eta ! \right) ^2\, \left[ \left( {\begin{array}{c}\eta +i-1\\ \eta \end{array}}\right) \left( {\begin{array}{c}\eta +j-1\\ \eta \end{array}}\right) m_{\eta ,i+j-2}\right] , \end{aligned}$$

and therefore

$$\begin{aligned} \Phi ( {\mathcal {R}}_\eta ,2\eta ) =&\left( \eta ! \right) ^2 \left[ \left( \sum _{\ell =0}^{2\eta } (-1)^{\ell }\left( {\begin{array}{c}2\eta \\ \ell \end{array}}\right) \left( {\begin{array}{c}\eta +i-1+(\eta -\ell )\\ \eta \end{array}}\right) \right. \right. \nonumber \\&\quad \left. \left. \left( {\begin{array}{c}\eta +j-1-(\eta -\ell )\\ \eta \end{array}}\right) \right) \;m_{\eta ,i+j-2}\right] \nonumber \\ =&\left( \eta ! \right) ^2\,\left[ \left( \sum _{\ell =0}^{2\eta } (-1)^{\ell }\left( {\begin{array}{c}2\eta \\ \ell \end{array}}\right) \left( {\begin{array}{c}2\eta +i-1-\ell \\ \eta \end{array}}\right) \right. \right. \nonumber \\&\quad \left. \left. \left( {\begin{array}{c}j-1+\ell \\ \eta \end{array}}\right) \right) \;m_{\eta ,i+j-2}\right] . \end{aligned}$$
(17)

Clearly \(\displaystyle f(\ell )= \left( {\begin{array}{c}2\eta +i-1-\ell \\ \eta \end{array}}\right) \left( {\begin{array}{c}j-1+\ell \\ \eta \end{array}}\right) \) is a polynomial of degree \(2\eta \) in \(\ell \) and leading coefficient \(\displaystyle \frac{(-1)^\eta }{(\eta !)^2} \). By Lemma 2.1, we deduce that

$$\begin{aligned} \sum _{\ell =0}^{2\eta } (-1)^{\ell }\left( {\begin{array}{c}2\eta \\ \ell \end{array}}\right) \left( {\begin{array}{c}2\eta +i-1-\ell \\ \eta \end{array}}\right) \left( {\begin{array}{c}j-1+\ell \\ \eta \end{array}}\right) =(-1)^{\eta }\; \left( {\begin{array}{c}2\eta \\ \eta \end{array}}\right) . \end{aligned}$$
(18)

Hence, from (17)–(18) we get \(\displaystyle \Phi ( {\mathcal {R}}_\eta ,2\eta ) =(-1)^\eta (2\eta )![m_{\eta ,i+j-2}]=(-1)^\eta \;(2\eta )!\,{\mathcal {H}}_\eta \) and (b). \(\square \)

We will assume that \({\mathcal {A}}\) is an infinite symmetric matrix because this is obviously a necessary condition for (2) to take place since the Hankel matrices \({\mathcal {H}}_k\) are symmetric.

Theorem 5

Let \({\mathcal {A}}\) be an infinite symmetric matrix, \(\eta \in \mathbb {Z}_{+}\) (fixed) such that \(\Phi ( {\mathcal {A}},2\eta +1)={\mathcal {O}}\). Then,

$$\begin{aligned} \Phi ( {\mathcal {A}}_\eta ,2\eta -1)={\mathcal {O}}, \end{aligned}$$
(19)

where \( {\mathcal {A}}_\eta ={\mathcal {A}}-{\mathcal {R}}_\eta \) and \(\displaystyle {\mathcal {R}}_\eta = \frac{(-1)^\eta }{(2\eta )!}{\mathcal {U}}^{-\eta } {\mathcal {D}}_\eta \Phi ( {\mathcal {A}},2\eta ) {\mathcal {D}}_\eta {\mathcal {U}}^{\eta }\).

Proof

If \({\mathcal {A}}_\eta ={\mathcal {O}}\), the theorem is obvious. Assume that \({\mathcal {A}}_\eta \ne {\mathcal {O}}\), from Theorem 4, we get \(\Phi ( {\mathcal {R}}_\eta ,2\eta )= \frac{(-1)^\eta }{(2\eta )!}\Phi ( {\mathcal {A}},2\eta )\), i.e. \(\Phi ( {\mathcal {A}}_\eta ,2\eta )= {\mathcal {O}}\). According to the recurrence formula (14), we have \( {\mathcal {U}}\,\Phi ( {\mathcal {A}}_\eta ,2\eta -1) = \Phi ( {\mathcal {A}}_\eta ,2\eta -1){\mathcal {U}}^{-1}\) which is equivalent to stating that \(\Phi ( {\mathcal {A}}_\eta ,2\eta -1)\) is a Hankel matrix and therefore it is a symmetric matrix.

On the other hand, \({\mathcal {A}}_\eta \) is a symmetric matrix since it is the difference of two symmetric matrices. Hence, from Proposition 2.4 we get that \(\Phi ( {\mathcal {A}}_\eta ,2\eta -1)\) is antisymmetric which establishes (19). \(\square \)

Proof of Theorem 1

From Theorem 4, a Hankel-Sobolev matrix of index \(d\in \mathbb {Z}_{+}\) satisfies the conditions (4) and each Hankel matrix \({\mathcal {H}}_k\) holds (5), which establishes the first implication of the theorem.

For the converse, assume that \({\mathcal {M}}\) is a symmetric infinite matrix and there exits \(d\in \mathbb {Z}_{+}\) such that the conditions (4) are satisfied. From Proposition 2.3, \(H_d=\frac{(-1)^d}{(2d)!}\Phi ( {\mathcal {M}},2d)\ne {\mathcal {O}}\) is a Hankel matrix.

Denote \({\mathcal {M}}_{d-1}={\mathcal {M}}_{d}-{\mathcal {R}}_d\), where \({\mathcal {M}}_{d}={\mathcal {M}}\) and \(\displaystyle {\mathcal {R}}_d= {\mathcal {U}}^{-d} {\mathcal {D}}_d {\mathcal {H}}_d {\mathcal {D}}_d {\mathcal {U}}^{d}\). From (19) and Proposition 2.3, \(H_{d-1}=\frac{(-1)^{d-1}}{(2d-2)!}\Phi ( {\mathcal {M}}_{d-1},2d-2)\) is a Hankel matrix.

Let \({\mathcal {M}}_{d-k}={\mathcal {M}}_{d+1-k}-{\mathcal {R}}_{d+1-k}\) and \(\displaystyle {\mathcal {R}}_{d+1-k}= {\mathcal {U}}^{-d-1+k} {\mathcal {D}}_{d+1-k} {\mathcal {H}}_{d+1-k} {\mathcal {D}}_{d+1-k} {\mathcal {U}}^{d+1-k}\). Repeating the previous argument, we get that \(H_{d-k}=\frac{(-1)^{d-k}}{(2d-2k)!}\Phi ( {\mathcal {M}}_{d-k},2d-2k)\) is a Hankel matrix for \(k=2,\dots , d\).

By construction, it is clear that

$$\begin{aligned} {\mathcal {M}}=\sum _{k=0}^{d}{\mathcal {R}}_{d-k}= \sum _{k=0}^{d} {\mathcal {U}}^{k-d} {\mathcal {D}}_{d-k} {\mathcal {H}}_{d-k} {\mathcal {D}}_{d-k} {\mathcal {U}}^{d-k}, \end{aligned}$$

i.e. \({\mathcal {M}}\) is a Hankel-Sobolev matrix and the proof is complete. \(\square \)

2.1 The Sobolev Moment Problem

Let \(\mu \) be a finite positive Borel measure supported on the real line and \({\textbf{L}}_2(\mu )\) be the usual Hilbert space of square integrable functions with respect to \(\mu \) with the inner product

$$\begin{aligned} \langle f,g \rangle _{\mu }= \int _{\mathbb {R}} f(x) g(x) d\mu (x), \quad \text{ for } \text{ all } f,g \in {\textbf{L}}_2(\mu ). \end{aligned}$$
(20)

The nth moment associated with the inner product (20) (or the measure \(\mu \)) is defined as \(m_n=\langle x^n,1\rangle _{\mu }\) (\(n\ge 0\)), provided the integral exists. The Hankel matrix \({\mathcal {H}}\left[ m_n \right] \) is called the matrix of moments associated with \(\mu \).

The classical moment problem consists in solving the following question: given an arbitrary sequence of real numbers \(\{m_n\}_{n\ge 0}\) (or equivalently the associated Hankel matrix \({\mathcal {H}}\left[ m_n \right] \))   and a closed subset \(\Delta \subset \mathbb {R}\), find a positive Borel measure \(\mu \) supported on \(\Delta \), whose nth moment is \(m_n\), i.e.

$$\begin{aligned} m_n= \int _{\Delta } x^n d\mu (x), \quad \text { for all }n\ge 0. \end{aligned}$$

It is said that the moment problem \(({\mathcal {H}};\Delta )\) is definite, if it has at least one solution and determinate if the solution is unique. There are three named classical moment problems: the Hamburger moment problem when the support of \(\mu \) is on the whole real line, the Stieltjes moment problem if \(\Delta =[0,\infty )\), and the Hausdorff moment problem for a bounded interval \(\Delta \) (without loss of generality, \(\Delta =[0, 1]\)).

As H. J. Landau write in the introduction of [11, p.1]: “The moment problem is a classical question in analysis, remarkable not only for its own elegance, but also for the extraordinary range of subjects, theoretical and applied, which it has illuminated”. For more details on the classical moment problem, the reader is referred to [2, 8, 11, 20, 21] and for historical aspects to [10] or [8, Sec. 2.4].

Without restriction of generality, we now turn our attention to the Hamburger moment problem referring to the following lemma as a necessary and sufficient condition for the problem of moments to be defined and determined.

Lemma 2.3

([21, Th. 1.2]) Let \(\{ m_n\}_{n=0}^{\infty }\) be a sequence of real numbers and denote by \({\mathcal {H}}={\mathcal {H}}\left[ m_n \right] \) the associated Hankel matrix. Then,

  1. 1.

    The Hamburger moment problem \(({\mathcal {H}};\mathbb {R})\) has a solution, whose support is not reducible to a finite set of points, if and only if \({\mathcal {H}}\) is a matrix positive definite of infinite order (i.e. \(\det ([{\mathcal {H}}]_n)>0\) for all \(n\ge 1\)).

  2. 2.

    The Hamburger moment problem \(({\mathcal {H}};\mathbb {R})\) has a solution, whose support consists of precisely k distinct points, if and only if \({\mathcal {H}}\) is a matrix positive definite of order k (i.e. \(\det ([{\mathcal {H}}]_n)>0\) for all \(1 \le n\le k\), and \(\det ([{\mathcal {H}}]_n)=0\) for all \( n> k\)). The moment problem is determined in this case.

The analogous results for the moment problem of Stieltjes \(({\mathcal {H}};\mathbb {R}_{+})\) or the moment problem of Hausdorff \(({\mathcal {H}};[0,1])\) are [21, Th. 1.3] and [21, Th. 1.5], respectively. Other equivalent formulations of these results can be seen in [15, Ch. 1, Sec. 7] or [21, Sec. 3.2].

The (nk)-moment associated with the inner product (13) is defined as \(m_{n,k}=\langle x^n,x^k\rangle _{\vec {\varvec{\mu }}}\) (\(n,k\ge 0\)), provided the integral exists. In the sequel, the values \(\langle x^n,x^k\rangle _{\vec {\varvec{\mu }}}\) are called S-moments. Here, instead of a sequence of moments, we have the infinite matrix of moments \({\mathcal {M}}\) with entries \(m_{n,k}=\langle x^n,x^k\rangle _{\vec {\varvec{\mu }}}\), (\(n,k\ge 0\)).

Now, the Sobolev moment problem (or \(S\!\)-moment problem) consists of solving the following question: given an infinite matrix \( {\mathcal {M}}=(m_{i,j})_{i,j=0}^{\infty }\) and \(d+1\) subsets \(\Delta _{k} \subset \mathbb {R}\) (\(0\le k \le d\)), find a set of \(d+1\) measures \(\{\mu _0, \mu _1, \dots , \mu _d \}\), where \(\mu _d \ne 0 \) and \(\text {supp} \left( \mu _k \right) \subset \Delta _k\), such that \(m_{i,j}=\langle x^i,x^j\rangle _{\vec {\varvec{\mu }}}\) for \(i,j=0,1, \dots \) As in the standard case, the problem is considered definite if it has at least one solution, and determinate if this solution is unique. There are three conventional cases of \(S\!\)-moment problems: the Hamburger \(S\!\)-moment problem when \(\Delta _0=\cdots =\Delta _d=\mathbb {R}\); the Stieltjes \(S\!\)-moment problem if \(\Delta _0=\cdots =\Delta _d=[0,\infty )\), and the Hausdorff \(S\!\)-moment problem for \(\Delta _0=\cdots =\Delta _d=[0,1]\). Nonetheless, other combinations of the sets \(\Delta _{k} \subset \mathbb {R}\) (\(0\le k \le d\)) are possible too. An equivalent formulation of the Sobolev moment problem is made for the special case \(d=\infty \).

The following result was proved in [3, Th. 1]. In virtue of Theorem 1, we can now reformulate it in the following equivalent form.

Theorem 6

Given an infinite symmetric matrix \( {\mathcal {M}}=(m_{i,j})_{i,j=0}^{\infty }\) and \(d+1\) subset \(\Delta _{k} \subset \mathbb {R}\) (\(0\le k \le d \in \mathbb {Z}_{+}\)), the \(S\!\)-moment problem is definite (or determinate) if and only if   \(\Phi ( {\mathcal {M}},2d+1)={\mathcal {O}}\), \(\Phi ( {\mathcal {M}},2d)\ne {\mathcal {O}}\) and for each \(k=0,1, \dots , d\) the Hankel matrix \( {\mathcal {H}}_{k}\) (defined in (5)) is such that the classical moment problem \(( {\mathcal {H}}_{k};\Delta _k)\) is definite (or determinate).

Although [3] is devoted to the study of the case in which d is finite and the measures involved are supported on subset of the real line, there are no difficulties in extending these results when \( d =\infty \) or if the measures are supported on the unit circle, as confirmed by the authors of [13, 14]. The \(S\!\)-moments problem for discrete Sobolev-type inner products was studied in [22].

3 Hessenberg–Sobolev Matrices

From the definition of the matrix of formal moments \({\mathcal {M}}\) in (11), we have two immediate consequences.

Proposition 3.1

Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix and \({\mathcal {M}}_{{\mathcal {G}}}\) be its associated matrix of formal moments. Then, \({\mathcal {M}}_{{\mathcal {G}}}\) is a symmetric and positive definite infinite matrix.

Proof

Obviously, \(\displaystyle {\mathcal {M}}_{{\mathcal {G}}}^{\textsf{T}} =\left( {\mathcal {T}}^{-1}\; {\mathcal {T}}^{-\textsf{T}}\right) ^{\textsf{T}}={\mathcal {T}}^{-1}\; {\mathcal {T}}^{-\textsf{T}}={\mathcal {M}}_{{\mathcal {G}}}.\) Moreover, for all \(n \ge 1\),

$$\begin{aligned} \det \left( [{\mathcal {M}}_{{\mathcal {G}}}]_k\right) =\det \left( [{\mathcal {T}}]_k^{-1}\right) \det \left( [{\mathcal {T}}]_k^{-\textsf{T}}\right) =\tau _{0, 0}^2 \cdot \tau _{1, 1}^2 \cdot \ldots \cdot \tau _{k-1, k-1}^2 > 0, \end{aligned}$$

where \(\tau _{i, i}\) is the (ii)-entry of \({\mathcal {T}}^{-1}\). \(\square \)

The following theorem clarifies the relation between the sequence of polynomials generated by \({\mathcal {G}}\) and the matrix of formal moments.

Theorem 7

Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix and \({\mathcal {M}}_{{\mathcal {G}}}= \left( m_{i,j}\right) \) be its associated matrix of formal moments.

  1. 1.

    If \(p(x)=\sum _{i=0} ^{n_1} a_i x^i\) and \(q(x)=\sum _{j=0} ^{n_2} b_j x^j\) are two polynomials in \(\mathbb {P}\) of degree \(n_1\) and \(n_2\), respectively. Then, the bilinear form (12) defines an inner product on \(\mathbb {P}\) and \(\Vert \cdot \Vert _{{\mathcal {G}}}=\sqrt{\langle \cdot ,\cdot \rangle _{{\mathcal {G}}}}\) is the norm induced by (12) on \(\mathbb {P}\).

  2. 2.

    Let \(m_{i,j}\) be the (ij)th entry of \({\mathcal {M}}_{{\mathcal {G}}}\), as in (11), then \(m_{i,j}=\langle x^{i},x^{j} \rangle _{{\mathcal {G}}}\).

  3. 3.

    \(\{Q_n\}\), the sequence of polynomials generated by \({\mathcal {G}}\), is the sequence of orthonormal polynomials with respect to the inner product (12).

  4. 4.

    \(g_{j,k}=\langle x Q_{j}, Q_k\rangle _{{\mathcal {G}}}\), where \(g_{j,k}\) is the (jk)-entry of the matrix \({\mathcal {G}}\) given in (6), with \(j\ge 0\) and \(0\le k \le j+1\).

Proof

From Proposition 3.1, the statement 1) is straightforward. The assertion 2) follows from (12).

Let \({\mathcal {E}}_j\) be the infinite column-vector whose i-entry is \(\delta _{i,j}\), where \(i,j\in \mathbb {Z}_{+}\). Denote by \(Q_n(x)=\sum _{i=0} ^{n} t_{n,i} \,x^i\) the nth polynomial generated by \({\mathcal {G}}\), as in (9). Then, for \(j=0, \ldots , n\)

$$\begin{aligned} \langle Q_n, x^j \rangle _{{\mathcal {G}}} =&( t_{n,0},\cdots , t_{n,n},0,\cdots ) \; {\mathcal {M}}_{{\mathcal {G}}} \; {\mathcal {E}}_j ={\mathcal {E}}_n^{\textsf{T}}{\mathcal {T}}{\mathcal {M}}_{{\mathcal {G}}} {\mathcal {E}}_j={\mathcal {E}}_n^{\textsf{T}}{\mathcal {T}} \, {\mathcal {T}}^{-1}\, {\mathcal {T}}^{-\textsf{T}}\, {\mathcal {E}}_j\\ =&{\mathcal {E}}_n^{\textsf{T}}\, {\mathcal {T}}^{-\textsf{T}} \, {\mathcal {E}}_j= \tau _{n,n} \, {\mathcal {E}}_n^{\textsf{T}}\, {\mathcal {E}}_j= \tau _{n,n} \, \delta _{n,j}; \end{aligned}$$

where \(\tau _{n,n}\ne 0\). Furthermore,

$$\begin{aligned} \langle Q_n, Q_n \rangle _{{\mathcal {G}}} =&( t_{n,0},\cdots , t_{n,n},0,\cdots ) \; {\mathcal {M}}_{{\mathcal {G}}} \; ( t_{n,0},\cdots , t_{n,n},0,\cdots )^{\textsf{T}}\\ =&{\mathcal {E}}_n^{\textsf{T}}\, {\mathcal {T}}\,{\mathcal {M}}_{{\mathcal {G}}} \,{\mathcal {T}}^{\textsf{T}} \, {\mathcal {E}}_n = {\mathcal {E}}_n^{\textsf{T}}\,{\mathcal {T}}\,{\mathcal {T}}^{-1}\, {\mathcal {T}}^{-\textsf{T}}\, {\mathcal {T}}^{\textsf{T}} \, {\mathcal {E}}_n =1. \end{aligned}$$

Hence, \(Q_n\) is the nth orthonormal polynomial with respect to (12). The fourth assertion is straightforward from (8) and the orthogonality. \(\square \)

Remark 3.1

From (8)–(9), we have that the leading coefficient of \(Q_n\) is

$$\begin{aligned} t_{n,n}=t_{0,0}\left( \prod _{k=0}^{n-1} g_{k,k+1}\right) ^{-1}\ne 0; \quad \text { for all \;} n \ge 1. \end{aligned}$$

Therefore, the corresponding nth-monic polynomial orthogonal with respect to \(\langle \cdot , \cdot \rangle _{{\mathcal {G}}}\) is \(q_n=\tau _{n,n} \,Q_n\) and \(\Vert q_n\Vert _{{\mathcal {G}}}=\tau _{n,n}=t_{n,n}^{-1}\) as in (10).

Theorem 8

The matrices \({\mathcal {G}}\) and \({\mathcal {M}}_{{\mathcal {G}}}\) are closely related by the expression

$$\begin{aligned} {\mathcal {G}}={\mathcal {T}} {\mathcal {U}}{\mathcal {M}}_{{\mathcal {G}}} {\mathcal {T}}^{\textsf{T}}. \end{aligned}$$
(21)

Proof

From (9) \(\displaystyle Q_{n}(x)=\sum _{i=0}^n t_{n,i} x^{i}\), therefore

$$\begin{aligned} g_{k,\ell }=&\langle x Q_{k}, Q_\ell \rangle _{{\mathcal {G}}}= \sum _{i=0}^k\sum _{j=0}^\ell t_{k,i} \, t_{\ell ,j} \,\langle x^{i+1}, x^{j}\rangle _{{\mathcal {G}}}=\sum _{i=0}^k\sum _{j=0}^\ell t_{k,i} \, t_{\ell ,j} \, m_{i+1,j} \end{aligned}$$

which is the \((k,\ell )\) entry of matrix \({\mathcal {T}} {\mathcal {U}}{\mathcal {M}}_{{\mathcal {G}}}{\mathcal {T}}^{\textsf{T}}\) and (21) is proved. \(\square \)

Theorem 9

Let \({\mathcal {G}} \in \mathbb {M}\) be a non-degenerate Hessenberg matrix, \({\mathcal {M}}_{{\mathcal {G}}} \in \mathbb {M}\) be its matrix of formal moments associated and \(\eta \in \mathbb {Z}_{+}\) fixed. Then 

$$\begin{aligned} \Psi ( {\mathcal {G}},\eta )= (-1)^{n}\,{\mathcal {T}} \, \Phi ( {\mathcal {M}}_{{\mathcal {G}}},\eta ) \, {\mathcal {T}}^{\textsf{T}}. \end{aligned}$$

Proof

From (11) and (21), we get that \({\mathcal {G}}={\mathcal {T}}{\mathcal {U}} {\mathcal {T}}^{-1}\). Therefore, for each \(k \in \mathbb {Z}_{+}\) we obtain

$$\begin{aligned} {\mathcal {G}}^k={\mathcal {T}} {\mathcal {U}}^{k} {\mathcal {T}}^{-1} \quad \text { and } \quad {\mathcal {G}}^{k\textsf{T}}={\mathcal {T}}^{-\textsf{T}} {\mathcal {U}}^{-k} {\mathcal {T}}^{\textsf{T}} . \end{aligned}$$
(22)

Now, from (3), (7), (11) and (22) it follows that

$$\begin{aligned} \Psi ( {\mathcal {G}},\eta ) =&\sum _{k=0}^{\eta }(-1)^{k}\, \left( {\begin{array}{c}\eta \\ k\end{array}}\right) \;{\mathcal {G}}^{k}\; {\mathcal {G}}^{(\eta -k)\; \textsf{T}}=\sum _{k=0}^{\eta }(-1)^{k}\, \left( {\begin{array}{c}\eta \\ k\end{array}}\right) \;{\mathcal {T}} \,{\mathcal {U}}^{k} \, {\mathcal {M}}_{{\mathcal {G}}}\, {\mathcal {U}}^{k-\eta }\, {\mathcal {T}}^{\textsf{T}}\\ =&{\mathcal {T}} \, \left( (-1)^{n}\, \sum _{\ell =0}^{\eta }(-1)^{\ell }\, \left( {\begin{array}{c}\eta \\ \ell \end{array}}\right) \,{\mathcal {U}}^{\eta -\ell } \, {\mathcal {M}}_{{\mathcal {G}}}\, {\mathcal {U}}^{-\ell } \right) \, {\mathcal {T}}^{\textsf{T}}. \end{aligned}$$

\(\square \)

4 Favard Type Theorems

One of the main problems in the general theory of orthogonal polynomials is to characterize the non-degenerate Hessenberg matrices, for which there exists a non-discrete positive measure \(\mu \) supported on the real line such that the inner product (12) can be represented as

$$\begin{aligned} \langle p,q \rangle _{{\mathcal {G}}}=\langle p,q \rangle _{\mu }:= \int p\,q\,d\mu . \end{aligned}$$
(23)

The aforementioned characterization is the well-known Favard Theorem (c.f. [7] or [1] for an overview of this theorem and its extensions), that we revisit according to the view-point of this paper.

Theorem 10

(Favard theorem) Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix and \(\langle \cdot , \cdot \rangle _{{\mathcal {G}}}\) be the inner product on \(\mathbb {P}\) defined by (12). Then, there exists a non-discrete positive measure \(\mu \) such that \(\langle p,q \rangle _{{\mathcal {G}}}= \langle p, q \rangle _{\mu }\) for all \(p,q \in \mathbb {P}\) if and only if   \(\Psi ( {\mathcal {G}},1)={\mathcal {O}}\).

Proof

Assume that there exists a non-discrete positive measure \(\mu \) such that \(\langle p,q \rangle _{{\mathcal {G}}}= \langle p, q \rangle _{\mu }\) for all \(p,q \in \mathbb {P}\), where \({\mathcal {G}}\) is a non-degenerate Hessenberg matrix. From the orthogonality of the generated polynomials \(Q_n\) (Theorem 7) and the fact that the operator of multiplication by the variable is symmetric with respect to \(\langle \cdot , \cdot \rangle _{\mu }\) (\(\langle x p, q \rangle _{\mu }=\langle p, x q \rangle _{\mu }\)), it is clear that \({\mathcal {G}}\) is a symmetric tridiagonal matrix, which is equivalent to \(\Psi ( {\mathcal {G}},1)={\mathcal {O}}\).

On the other hand, if \({\mathcal {G}}\) is a non-degenerate Hessenberg matrix such that \(\Psi ( {\mathcal {G}},1)={\mathcal {O}}\), we get that \({\mathcal {G}}\) a symmetric Hessenberg matrix or equivalently a non-degenerate tridiagonal matrix. From Theorem 9,

$$\begin{aligned} {\mathcal {O}}=\Phi ( {\mathcal {M}}_{{\mathcal {G}}},1)={\mathcal {U}} {\mathcal {M}}_{{\mathcal {G}}}-{\mathcal {M}}_{{\mathcal {G}}} {\mathcal {U}}^{-1}, \end{aligned}$$

i.e. \({\mathcal {M}}_{{\mathcal {G}}} \) is a Hankel matrix, which from Proposition 3.1 is positive definite. From Lemma 2.3, the proof is complete. \(\square \)

Obviously, under the assumptions of Theorems 7 and 10, the sequence \(\{Q_n\}\) of polynomials generated by \({\mathcal {G}}\) is the sequence of orthogonal polynomials with respect to the measure \(\mu \) (i.e. with respect to the inner product (23)).

Example 4.1

The sequence of polynomials \(\{1, x, x^2, \ldots , x^n, \ldots \}\) is generated by the non-degenerated Hessenberg matrix

$$\begin{aligned} {\mathcal {G}} = \left( \begin{array}{cccc} 0 &{} 1 &{} 0 &{} \ldots \\ 0 &{} 0 &{} 1 &{} \ldots \\ 0 &{} 0 &{} 0 &{} \ldots \\ \vdots &{} \vdots &{} \vdots &{} \ddots \end{array} \right) , \end{aligned}$$

hence from Theorem 7, the sequence is orthonormal with respect to the inner product (12). As \({\mathcal {G}}\) is a non-symmetric matrix, \(\Psi ( {\mathcal {G}},1)\ne {\mathcal {O}}\). Then, from Theorem 10 there does not exist a non-discrete positive measure \(\mu \), such that \(\langle p,q \rangle _{{\mathcal {G}}}= \langle p, q \rangle _{\mu }\) for all \(p,q \in \mathbb {P}\).

Proof of Theorem 3

Let \(p(x)=\sum _{i=0} ^{n_1} a_i x^i\) and \(q(x)=\sum _{j=0} ^{n_2} b_j x^j\) be polynomials in \(\mathbb {P}\) of degree \(n_1\) and \(n_2\), respectively. Then, from the Sobolev inner product (13) we have the representation

$$\begin{aligned} \langle p, q \rangle _{\vec {\varvec{\mu }}_d}= (a_0,\cdots ,a_{n_1},0,\cdots ) {\mathcal {S}} (b_0,\cdots ,b_{n_2},0,\cdots )^{\textsf{T}} = \sum _{i=0} ^{n_1} \sum _{j=0} ^{n_2} a_i \,s_{i,j}\,{b_j} \;, \end{aligned}$$
(24)

where \( {\mathcal {S}}=(s_{i,j})_{i,j=0}^{\infty }\) with \(s_{i,j}=\langle x^i,x^j\rangle _{\vec {\varvec{\mu }}_d}\), is a Hankel-Sobolev matrix of index d. If \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in {\mathbb {Z}_{+}}\), from (12) and (24), to prove that \( \langle p, q \rangle _{\vec {\varvec{\mu }}_d} =\langle p, q \rangle _{{\mathcal {G}}} \) is equivalent to prove that \({\mathcal {S}} ={\mathcal {M}}_{{\mathcal {G}}}\), where \({\mathcal {M}}_{{\mathcal {G}}}\) is the matrix of formal moments associated with \({\mathcal {G}}\).

Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix. Assume that there exists \(d \in {\mathbb {Z}_{+}}\) and \(\vec {\varvec{\mu }}_d= (\mu _0,\mu _1,\dots ,\mu _d) \in {\mathfrak {M}}_{d}(\mathbb {R})\) (continuous case), such that \({\mathcal {S}} ={\mathcal {M}}_{{\mathcal {G}}}\). From Theorem 6, \({\mathcal {S}}\) is a Hankel-Sobolev matrix. Therefore, combining Theorem 1 and Theorem 9, we get that \(\Psi ( {\mathcal {G}},2d+1)={\mathcal {O}}\)   and   \(\Psi ( {\mathcal {G}},2d)\ne {\mathcal {O}}\). Furthermore, each matrix \({\mathcal {H}}_k\) defined by (5), is the moment matrix of the measure \(\mu _k\), which is a non-negative finite Borel measure whose support is an infinite subset. Hence, from Lemma 2.3 we have that \({\mathcal {H}}_k\) s a positive definite matrix of infinite order.

Reciprocally, let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix satisfying 1 and 2. From Theorems 2 and 9, we conclude that \({\mathcal {M}}_{{\mathcal {G}}}\), the matrix of formal moments associated with \({\mathcal {G}}\), is a Hankel-Sobolev matrix of index d, i.e. there exist Hankel matrices \({\mathcal {H}}_{k}\), \(k=0,\,1,\dots ,\,d\), such that

$$\begin{aligned} {\mathcal {M}}_{{\mathcal {G}}}=\sum _{k=0}^{d} \left( {\mathcal {U}}^{-k}\;{\mathcal {D}}_k \; {\mathcal {H}}_{k} \; {\mathcal {D}}_k\;{\mathcal {U}}^{k}\right) . \end{aligned}$$

From Theorem 6 and Lemma 2.3, the \(S\!\)-moment problem for \({\mathcal {M}}_{{\mathcal {G}}}\) is defined. Let \(\mu _{d-k}\) be a solution of the problem of moments with respect to \({\mathcal {H}}_{d-k}\) for each \(k=0,\,1,\dots ,\,d\). If \(\langle p, q \rangle _{\vec {\varvec{\mu }}_d}\) is as in (13), from Proposition 2.1 we get \( {\mathcal {S}} ={\mathcal {M}}_{{\mathcal {G}}}\). \(\square \)

The following result may be proved in much the same way as Theorem 3, using the appropriate assertions of Lemma 2.3 for the case of measures supported on finite subsets.

Theorem 11

(Favard type theorem for discrete case) Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix. Then, there exists \(d \in \mathbb {Z}_{+}\) and \(\vec {\varvec{\mu }}_d \in {\mathfrak {M}}_{d}(\mathbb {R})\) such that \( \langle p, q \rangle _{\vec {\varvec{\mu }}_d} =\langle p, q \rangle _{{\mathcal {G}}} \) if and only if  

  1. 1.

    \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in \mathbb {Z}_{+}\).

  2. 2.

    The Hankel matrix \({\mathcal {H}}_{0}\) defined by (5) is a positive definite matrix of infinite order and for each \(k=0,\,1,\dots ,\;d-1\); the matrix \({\mathcal {H}}_{d-k}\) is a positive definite matrix of order \(m_k \in \mathbb {Z}_{+}\).

The previous theorem is a refinement of [7, Lemma 3].

Theorem 12

(Favard type theorem for discrete-continuous case) Let \({\mathcal {G}}\) be a non-degenerate Hessenberg matrix. Then, there exists \(d \in \mathbb {Z}_{+}\) and \(\vec {\varvec{\mu }}_d \in {\mathfrak {M}}_{d}(\mathbb {R})\) such that \( \langle p, q \rangle _{\vec {\varvec{\mu }}_d} =\langle p, q \rangle _{{\mathcal {G}}} \) if and only if  

  1. 1.

    \({\mathcal {G}}\) is a Hessenberg–Sobolev matrix of index \(d \in \mathbb {Z}_{+}\).

  2. 2.

    The Hankel matrix \({\mathcal {H}}_{d}\) defined by (5), is a positive definite matrix of infinite order and for each \(k= 1,\,2\dots ,\;d\); the matrix \({\mathcal {H}}_{d-k}\) is a positive definite matrix of order \(m_k \in \mathbb {Z}_{+}\).