Abstract
The famous Jones–Ruscheweyh theorem states that n distinct points on the unit circle can be mapped to n arbitrary points on the unit circle by a Blaschke product of degree at most \(n-1\). In this paper, we provide a new proof using real algebraic techniques. First, the interpolation conditions are rewritten into complex equations. These complex equations are transformed into a system of polynomial equations with real coefficients. This step leads to a “geometric representation” of Blaschke products. Then another set of transformations is applied to reveal some structure of the equations. Finally, the following two fundamental tools are used: a Positivstellensatz by Prestel and Delzell describing positive polynomials on compact semialgebraic sets using Archimedean module of length N. The other tool is a representation of positive polynomials in a specific form due to Berr and Wörmann. This, combined with a careful calculation of leading terms of occurring polynomials finishes the proof.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Interpolating with Blaschke products has a large literature. In this paper, we focus exclusively on boundary interpolation, that is when we interpolate with Blaschke products and the nodes are on the unit circle. Necessarily, the values are on the unit circle too. The possibility of such interpolation was established in Cantor and Phelps [3]. Jones and Ruscheweyh [15] sharpened that result and showed that any m pairs of data can be interpolated with a Blaschke product of degree at most \(m-1\).
We denote the set of Blaschke products of degree m by
Here, \(\mathbf {B}_0\) consists of constants (with modulus 1). We also write
Theorem 1
(Jones–Ruscheweyh, 1987) Let \(0\le \varphi _1<\varphi _2<\ldots<\varphi _{m}<2\pi \) and \(\psi _1,\psi _2,\ldots ,\psi _m\in [0,2\pi )\). Then there exists a Blaschke product \(B\in \mathbf {B}_{\le m-1}\) such that \(B(\exp (i \phi _j))=\exp (i \psi _j)\), \(j=1,2,\ldots ,m\).
Actually, there are several proofs of this result and there is a very nice overview in the excellent paper [17] by Semmler and Wegert from 2006. Some earlier references are (not exhaustive list): [1, 6,7,8,9,10,11,12,13, 18, 19] and it is also worth mentioning the books [4, 5].
The outline of the paper is the following. After this introduction, we present a new proof in detail. The proof consists of two main steps: first, we parametrize Blaschke products and transform the boundary Blaschke interpolation problem to interpolation of real values with special real rational functions. After this transformation, we have two sets: one coming from the Blaschke product representation, and another one coming from interpolation data. These sets are subsets of some high dimensional real euclidean space and the assertion of the Jones–Ruscheweyh theorem is equivalent to that their intersection is not empty. We show this by employing tools from real algebraic geometry. In particular, the main tool in this step is a Positivstellensatz of Prestel and Delzell. Finally, we present some technical lemmas and their proofs in Sect. 5. It would be interesting to compare our approach with that of Semmler and Wegert, and also investigate the structure of solutions as a future paper.
2 First Part of the Proof
In this part, we rephrase our problem completely. First, we apply the Cayley transform on finite Blaschke products. Let us mention that Semmler and Wegert used this approach, and they established a natural description (see [17, Lem. 3]). As we need the exact dependence of the coefficients on the zeros of the Blaschke product, we detail this step. In this way, we transform Blaschke products (of degree at most m, denoted by \(\mathbf {B}_{\le m}\)) to a special subset of real rational functions (denoted by \(\mathbf {H}_m\)). Here the coordinates represent the coefficients, and we use a natural choice to exclude the ambiguity (caused by multiplying the numerator and the denominator by the same constant). We also transform the interpolation data and this yields a homogeneous set of data (\(\mathbf {S}\), solution set). If the number of interpolation pairs (n) and the degree of Blaschke products (\(\le m\)) differ by one only (i.e. \(n=m+1\)), then there is a nice description of the solution set \(\mathbf {S}\). Formulating this description (linear parametrization of \(\mathbf {S}\)) finishes the rephrasing of the problem to real algebraic equations and it is the last step of the first part.
2.1 Parametrizing the Blaschke Products
We use the unit disk \(\mathbb {D}:=\{z\in \mathbb {C}:\ |z|<1\}\) and the unit circle \(\mathbb {T}:=\{z\in \mathbb {C}:\ |z|=1\}\). We introduce a parametrization as follows. Let
where the closure of \(\mathbf {E}_m\) is \(\mathbf {F}_m\),
Now we investigate how \(\mathbf {E}_m\) and \(\mathbf {F}_m\) can be used to parametrize Blaschke products. Consider the parametrization mapping
where we set
for convenience (see (10), (11) and Lemma 1 below). Note that if \(a\rightarrow a_0\), \(|a|<1\), \(|a_0|=1\), then \((z-a)/(1-{\overline{a}}z)\rightarrow -a_0\) locally uniformly in \(\mathbb {D}\), and also on \(\mathbb {D}\cup \mathbb {T}\). Hence, \(\mathcal {P}_m\) is a continuous mapping (from the euclidean topology to the locally uniform convergence topology), see also Lemma 3. Roughly speaking, \(\mathcal {P}_m\) maps \(\mathbf {E}_m\), and \(\mathbf {F}_m\) to Blaschke products of degree m and Blaschke products of degree at most m. More precisely,
2.2 The Cayley Transform
In this section, we give a description of how the Cayley transform behaves on Blaschke products of degree at most m. We also need the inverse Cayley transform too. We denote the Cayley transform by \(z=T(u)\):
The set of real rational functions of degree at most m is:
It is important that if \(B\in \mathbf {B}_{\le m}\) is a Blaschke product, then
for some polynomials P, Q with real coefficients and \(\deg (P), \deg (Q)\le m\). This follows immediately from [17, Lem. 3] (Representation Lemma) and we reprove and investigate it in detail in the next subsection.
2.3 Structure of the Coefficients After the Cayley Transform
We also need the structure of the coefficients of P and Q when B is a Blaschke product (\(|\gamma |=1\)) later, so we detail this calculation.
So let \(B\in \mathbf {B}_m\) and we write
To simplify this, we introduce
so we can write
Denote the coefficients of C and D by \(c_j\) and \(d_j\) respectively:
Note that
are (holomorphic) polynomials in \(a_1,\ldots ,a_m\).
We also have \(\overline{D(u)}=(-1)^m C(u)\) (when \(u\in \mathbb {R}\)), therefore
for \(j=0,1,\ldots ,m\). We can express the leading coefficients, \(c_m=\prod _{j=1}^m (1+\overline{a_j})\) and \(d_m=\prod _{j=1}^m (-1-a_j)\).
We are going to express P and Q using C, D and (7). First, the leading coefficient of \(i C(u)-i \gamma D(u)\) is
and the leading coefficient of \(C(u)+\gamma D(u)\) is
We use Lemma 1 (with \(W=(-1)^m d_m\)) since \(|\gamma |=1\) and Lemma 2; Lemmas 1 and 2 can be found in Sect. 5. This way we get that
are polynomials with real coefficients and
Let us remark here that P(u) and Q(u) cannot have common zeros. Otherwise, if \(u\in \mathbb {R}\) is such that \(P(u)=0\) and \(Q(u)=0\), then \(C(u)=0\) and \(D(u)=0\) too. Considering the definitions of C(u) and D(u), there should be \(a_j, a_k\) with \(|a_j|< 1\), \(|a_k|< 1\) and \(i(\overline{a_j}-1)/(\overline{a_j}+1)=i(a_k-1)/(-a_k-1)\), a contradiction. This is also in accord with [17, Rem. 1].
Up to now, any branch of the square root can be used (\(\pm \sqrt{\gamma }\)). Now using (4), hence \(\sqrt{\gamma }=\delta \), we use \(\delta \) instead below.
Continuing the calculation for the coefficients of \(P(u)= \sum _{j=0}^m p_j u^j\) (and using \(|\delta |=1\) again), we have
The coefficients of \(Q(u)=\sum _{j=0}^m q_j u^j\) are
2.4 Parametrization of the Real Rational Functions
We investigate the parametrization of rational functions. In the previous section, \(p_j\)’s and \(q_j\)’s are polynomials which are the coefficients of rational functions depending on the zeros of the Blaschke product. In this paragraph, we use \(p_j\)’s and \(q_j\)’s as variables which are the coefficients of rational functions for convenience (and slightly abusing the notation). Afterward, we use them again as polynomials. Let
It is standard that
is a surjective but not bijective mapping, and if \(\mathbf {p} \in \mathbf {A}_1\subset \mathbb {R}^{2m+2}\), then \(c \mathbf {p}\) determines the same rational function (\(c\in \mathbb {R}{\setminus }\{0\}\)). In other words, the coefficients of the numerator and denominator of a rational function from \(\mathbf {H}_m\) are not uniquely determined (unless some type of normalization is imposed on the denominator and numerator).
Therefore we directly define the coefficients
where we used (12), (13) and (6). We know that
where \(J^c=\{1,2,\ldots ,m\}{\setminus } J\); in particular, if \(j=0\), then \(J=\emptyset \) and the sum consists of only one term, hence
Obviously, \(\mathcal {L}_{m,\mathbb {C}}(\mathbf {F}_m) \subset \mathbf {A}_1\).
Finally, we switch to real algebraic language, in particular, we use new variables as follows
With these substitutions, we introduce
and
and as above, \(\mathcal {L}_{m}(\mathbf {F}_{m,\mathbb {R}}) \subset \mathbf {A}_1\).
Let \(U_\ell \) and \(V_\ell \) be the following polynomials
where \(U_\ell ,V_\ell \in \mathbb {R}[\delta _1,\delta _2,x_1,y_1,\ldots ,x_m,y_m]\), and actually
for \(\ell =0,1,\ldots ,m\). Note that for \(\ell =0,1,\ldots ,m\)
and
where \(d_\ell \) depends on \(a_1,\ldots ,a_m\) while \(U_\ell \) and \(V_\ell \) depend on \(x_1,y_1,\ldots ,x_m,y_m\) but they are connected by (15).
2.5 Applying Cayley Transform on the Interpolation Data
Here we consider the interpolation data and transform it with the Cayley transform.
Suppose that pairwise distinct \(z_1,\ldots ,z_n\in \mathbb {C}\) with \(|z_1|=\ldots =|z_n|=1\) are given and \(w_1,\ldots ,w_n\in \mathbb {C}\) with \(|w_1|=\ldots =|w_n|=1\) are also given. We transform these and consider
Note that if \(w_j=-1\) for some j, then \(v_j=\infty \); also if \(z_j=-1\) for some j, then \(u_j=\infty \). By appropriate rotations, this can be avoided. To be precise, let \(\omega ,\chi \in \mathbb {C}\), \(|\omega |=|\chi |=1\) such that none of \(\omega w_1, \ldots , \omega w_n\) is equal to \(-1\) and none of \(z_1 \chi ,\ldots , z_n \chi \) is equal to \(-1\). Then we find a Blaschke product \(B(\cdot )\) so that \(B(z_j \chi )=\omega w_j\) for \(j=1,\ldots ,n\) where
Then just take
and this will interpolate \(w_j\) at \(z_j\): \({\widetilde{B}}(z_j)=w_j\), \(j=1,\ldots ,m\).
Therefore, we may assume that all \(u_j\) and \(v_j\) are finite, i.e.
2.6 Introducing Two Real Sets Coming from Blaschke Products and Interpolation Data
Here we would like to find real polynomials
\(Q(u)\not \equiv 0\) such that
This equation is equivalent to \(P(u_j)/Q(u_j)=v_j\), provided that
We will return to this condition later.
An equivalent form of (22) is
Note that this is a homogeneous linear equation in \(\alpha _0,\beta _0,\ldots ,\alpha _m,\beta _m\), so it always has a solution. Let \(\mathbf {U}\) be the following Vandermonde matrix
and put \(\mathbf {D}:=\mathrm {diag}(-v_1,\ldots ,-v_n)\) and \(\mathbf {v}:=(\alpha _0,\alpha _1,\ldots ,\alpha _m,\beta _0,\beta _1,\ldots ,\beta _m)^{\top }\) (where \(.^{\top }\) means transpose hence \(\mathbf {v}\) is a column vector) and introduce
for short. Hence, (24) can be written as
We assume that
So \(\mathbf {U}\) is a square matrix, \(\mathbf {U}\in \mathbb {R}^{n\times n}\), and since \(u_1,\ldots ,u_n\) are pairwise different (see (21)), \(\mathbf {U}\) is a non-singular matrix. Note that \(\mathbf {M}\) can be thought of as a mapping from \(\mathbf {A}=\mathbb {R}^{2m+2}\). Therefore,
Using the rank-nullity theorem, this implies that
Note that \(\mathbf {S}=\mathbf {S}(u_1,\ldots ,u_n,v_1,\ldots ,v_n)\subset \mathbf {A}\) depends on the interpolation data (i.e. on \(z_1,\ldots ,z_n\) and \(w_1,\ldots ,w_n\)) and contains all the solutions of (24) while the ”geometrical representation” of general Blaschke products,
is independent of the interpolation data.
We return to the condition (23). It holds because
hence the corresponding \(P(u)=\sum _{j=0}^m \alpha _j u^j\), \(Q(u)=\sum _{j=0}^m \beta _j u^j\) satisfy that \(P(u)/Q(u)=T^{-1}\circ B\circ T(u)\) holds for some Blaschke product \(B\in \mathbf {B}_{\le m}\). Therefore P and Q cannot have common zeros.
Note that \(\mathbf {G}_{\le m}\subset \mathbf {A}_1\), moreover
Also, observe that \(\mathbf {S}{\setminus }\{0\}\subset \mathbf {A}_1\). Indeed, if \(\mathbf {v}\in \mathbf {S}{\setminus }\{0\}\) and \(\mathbf {v}\in \mathbf {A}{\setminus }\mathbf {A}_1=\mathbf {A}_1^c\), then \(\mathbf {v}\) has the form \((\alpha _1,\ldots ,\alpha _n,0,\ldots ,0)\) where \((\alpha _1,\ldots ,\alpha _n)\ne 0\). Substituting this into (24), we get that \(\mathbf {U} (\alpha _1,\ldots ,\alpha _n)^\top =0\) which contradicts the fact that \(\mathbf {U}\) is invertible.
It is straightforward to see that the Jones-Ruscheweyh theorem is equivalent to
2.7 Parametrizing the Set Coming from Interpolation Data
We now parametrize \(\mathbf {S}\). In the previous section, \(\alpha _j\)’s and \(\beta _j\)’s were unknowns, coming from a subspace: \((\alpha _0,\alpha _1,\ldots ,\alpha _m,\beta _0,\beta _1,\ldots ,\beta _m)\in \mathbf {S}\). As we parametrize \(\mathbf {S}\), we use the same symbols: \(\alpha _j=\alpha _j(t_1,\ldots ,t_n)\), \(\beta _j=\beta _j(t_1,\ldots ,t_n)\) to keep the notation simple. Using \(t=(t_1,\ldots ,t_n)\in \mathbb {R}^n\), we consider
where \(\alpha _0,\alpha _1,\ldots ,\alpha _m\) and \(\beta _0,\beta _1,\ldots ,\beta _m\) are linear polynomials without constant terms:
Note that (29) is equivalent to that the system
has a solution under the conditions
where \(\delta _1,\delta _2,x_1,y_1,\ldots ,x_m,y_m,t_1,\ldots ,t_n\in \mathbb {R}\).
This reformulation is expressed in terms of real algebraic geometry only.
To exploit the dimension condition (28), we introduce the matrix A by collecting the coefficients of \(\alpha _0,\alpha _1,\ldots ,\alpha _m,\beta _0,\beta _1,\ldots ,\beta _m\). Let \(A\in \mathbb {R}^{2n \times n}\) be the matrix for which
where \(t^\top \) is a column vector. Now we use the dimension condition (28). Hence, it is standard (see e.g. [14, p. 13, 0.4.6 (f)]) that there is an invertible matrix \(B_1\in \mathbb {R}^{n\times n}\) and a set \(I_1\subset \{1,2,\ldots ,2n\}\) where \(|I_1|=n\) such that the rows of \(AB_1\) with indices from \(I_1\) is the identity matrix of size n.
For ease of notation, we do not introduce new variables for \(B_1^{-1} t^\top \), that is, we assume that the rows of A with indices from \(I_1\) give the identity matrix. For simplicity, we label the set of \(I_1\) with j(.): \(j(1),\ldots ,j(n)\) are distinct, \(I_1=\{j(1),\ldots ,j(n)\}\). Put \(I_2:=\{1,2,\ldots ,2n\}{\setminus } I_1\) for the remaining row indices.
For convenience, we introduce
for \(\ell =0,1,\ldots ,m\). Therefore
3 A Second Transformation Applied on the Two Sets
Our ultimate goal is to show that the two sets (\(\mathbf {S}\) and \(\mathbf {G}_{\le m}\)) coming from different ”sides” of the problem have non-empty intersection, i.e (29) holds.
In this section, we apply a second transformation which is an adapted form of the rational parametrization of the unit circle. This transformation changes the occurring polynomials and reveals a crucial property of the polynomials (see (44), (45), and (46)). We exploit this property with a Positivstellensatz of Delzell and Prestel which is a special description of positive polynomials on compact, semialgebraic sets. Instead of sums of squares of polynomials, it features higher powers of polynomials.
3.1 Describing the Second Transformation
We will transform our system of equations by substituting variables and replacing equations so that we can apply a Positivstellensatz.
It is known that
is a rational parametrization of the unit circle, more precisely, it is a bijective mapping from \(\mathbb {R}\) to \(\{(x,y)\in \mathbb {R}^2: x^2 + y^2 =1\}{\setminus }\{(-1,0)\}\). Similarly,
is a rational mapping from \((\sigma ,r)\in [0,\infty )\times [-1,1]\) to \(\{(x,y)\in \mathbb {R}^2: x^2 + y^2 \le 1\}{\setminus }\{(-1,0)\}\) and bijective from \((0,\infty )\times [-1,1]\) to \(\{(x,y)\in \mathbb {R}^2: x^2 + y^2 \le 1\} {\setminus }\{(-1,0),(1,0)\}\) and \(\{0\}\times [-1,1]\) is mapped to the point (1, 0). We also use \(\sigma =s/(1-s)\) and \(\tau =u/(1-u^2)\), so if s runs over the interval (0, 1), then \(\sigma \) runs over the interval \((0,\infty )\) and if u runs over the interval \((-1,1)\), then \(\tau \) runs over the real numbers. Therefore we consider the composite mapping:
which is a bijective mapping from the open interval \((-1,1)\) to \(\{(x,y)\in \mathbb {R}^2: x^2 + y^2 =1\}{\setminus }\{(-1,0)\}\) and maps \(-1\) and \(+1\) to the point \((-1,0)\). Similarly, we consider
which has the following mapping properties. It maps \([0,1]\times (-1,1)\) bijectively to the open unit disk \(\{(x,y)\in \mathbb {R}^2:\ x^2+y^2<1\}\), it is also bijective mapping from \((0,1)\times [-1,1]\) to \(\{(x,y)\in \mathbb {R}^2:\ x^2+y^2\le 1\}{\setminus }\{(1,0),(-1,0)\}\), maps \(\{1\}\times [-1,1]\) to the point \((-1,0)\) and maps \(\{0\}\times [-1,1]\) to the point (1, 0).
Based on these, we introduce the following substitutions:
We will apply these substitutions and to obtain polynomials, we will multiply them by the denominators. To see it precisely, we focus on the structure of \(d_\ell \) and use formula (14) with the substitutions (15), (16) and the substitutions above so we can write for \(\ell =0,1,\ldots ,m\)
where
Slightly rewriting it, we have
and, in particular,
We also use
Observe that \(S_u\ge 3/4\) (for \(u\in [-1,1]\)) and \(1-2s_j+2s_j^2\ge 1/2\), \(j=1,\ldots ,m\) hence \(S\ge 3/2^{m+2}\).
We rewrite the polynomials \(U_\ell \) and \(V_\ell \)’s using substitutions (38) and (39). So we introduce
for \(\ell =0,1,\ldots ,m-1,m\). Note that \(R_j\)’s are polynomials from \(\mathbb {R}[u,s_1,r_1,\ldots ,s_m,r_m]\). For simplicity, we define
It is very important that \(R_j\), \(j=1,2,\ldots ,2m\) do not have constant terms (because of the first factor on the right of (41))
In other words, using the substitution
we can also write that
The last two \(R_j\)’s, namely \(R_{2m+1}\) and \(R_{2m+2}\) behave differently. The expressions (19) and (20) for \(U_m\) and \(V_m\) show that
-
if m is odd, then
$$\begin{aligned} U_m&= \frac{(-2)^{m+1}}{S_s} \left( \delta _1(1+\ldots )+ \delta _2\cdot \ldots \right) , \quad \\ V_m&= \frac{2 (-2)^{m}}{S_s} \left( \delta _1\cdot \ldots + \delta _2(1+\ldots ) \right) , \end{aligned}$$ -
if m is even, then
$$\begin{aligned} U_m&= \frac{2(-2)^{m}}{S_s} \left( \delta _1\cdot \ldots + \delta _2 (1+\ldots ) \right) ,\\ V_m&= \frac{2 (-2)^{m}}{S_s} \left( \delta _1(1+\ldots )+ \delta _2\cdot \ldots \right) \end{aligned}$$
where the \(\ldots \) stand for terms that are multiplied with an \(s_j\).
Taking into account these four lines above and the substitutions (38) and (39), we can write
Hence \(R_{j_0}\) has zero constant term and \(R_{j_1}\) has non-zero constant term. These observations will be crucial for the argument later.
The system (31) with conditions (32) is equivalent to the system
with the condition that \((u,s_1,\ldots ,s_m,r_1,\ldots ,r_m)\) are from the set
We remark that W is compact and \(L_j\in \mathbb {R}[t_1,\ldots ,t_n]\) and \(R_j\in \mathbb {R}[u,s_1,r_1,\ldots ,s_m,r_m]\).
For simplicity, we introduce \(X_1=u\), \(X_{1+j}=s_j\), \(j=1,2,\ldots ,m\), \(X_{m+1+j}=r_j\), \(j=1,2,\ldots ,m\), and
3.2 Eliminating the Parametrizing Auxiliary Variables
Since there is an identity submatrix within the matrix A (coming from the coefficients of \(L_j\)’s; see the definition of A, (33), and that of \(L_k\)’s, (34) and (35)) there exists an invertible matrix \(B_2\in \mathbb {R}^{2n\times 2n}\) such that with
we have
This is a simple elimination on the left-hand sides (using row operations on A instead of column operations which we used in Sect. 2.7). We transform the right-hand sides accordingly, hence we introduce \({\widetilde{R}}_1,{\widetilde{R}}_2,\ldots ,{\widetilde{R}}_{2n}\) as
The system (47) with condition (48) is equivalent to
with conditions (48), that is, when \(\mathbf {x}=(u,s_1,\ldots ,s_m,r_1,\ldots ,r_m) \in W\).
Introduce
Using (52), we can express \(t_k\)’s with \(u,s_1,\ldots ,s_m,r_1,\ldots ,r_m\) as follows
Therefore if \(\mathbf {x}\in W_2\), then there is a unique \(t\in \mathbb {R}^n\) such that \((\mathbf {x},t)\in {\widetilde{W}}\), because of the following. Note that \({\widetilde{R}}_j(\mathbf {x})\) are polynomials independent of \(t_1,\ldots ,t_n\). Furthermore, \(1-u^2+u^4 \ge 3/4\) when \(u\in [-1,1]\) and \(S\ge 3/2^{m+2}\) holds also. Therefore, for all possible \(\mathbf {x}\in W\), \({\widetilde{R}}_j(\mathbf {x})/S(\mathbf {x})\) is continuous and bounded.
This implies that \({\widetilde{W}}\) is compact.
Hence, the t’s coming from a solution are bounded, i.e. there is \(M_0>0\) such that for all \(t\in \mathbb {R}^n\) such that \((\mathbf {x},t)\in {\widetilde{W}}\), we have \(|t_j| \le M_0\), \(j=1,2,\ldots ,n\).
4 Application of the Positivstellensatz
We are going to use a form of Positivstellensatz which can be found in the book of Prestel and Delzell [16]. Briefly, it is for sums of even powers and for compact (semialgebraic) sets.
First, we introduce our new set of notations. Then we apply the Positivstellensatz to find a solution. This indirect argument features a step-by-step simplification of the representation (61) provided by the Positivstellensatz. As the first step of simplification, we apply a substitution (63) which turns the representation into a univariate identity (68). Then a careful comparison of the leading terms and degrees leads to an even more simplified identity (69). Finally, exploiting the special structure of the equation (comparing (72) and (71)) leads to a contradiction. We remark that in this section we do not use the polynomials \(c_0,c_1,\ldots ,c_m\) from Sect. 2.3 (and we use \(c_0,c_1\) as new symbols).
As the next step, we set
Recall that \(\mathbf {x}=(u,s_1,\ldots ,s_m,r_1,\ldots ,r_m)\) and for unifying the notation, we introduce the following:
Put \(N_1:=2m+2\).
Note that \(h_1(\mathbf {x})\ge 0\) if and only if \(u\in [-1,1]\) and \(h_{1+j}(\mathbf {x})\ge 0\) if and only if \(s_j\in [0,1]\) and \(h_{1+m+j}(\mathbf {x})\ge 0\) if and only if \(r_j\in [-1,1]\). Hence,
i.e. \({\widehat{W}}\) is a compact, semialgebraic set.
Introduce
Obviously, \(f(\mathbf {x},t)=0\) at some \((\mathbf {x},t)\in {\widehat{W}}\) if and only if \((\mathbf {x},t)\) is a solution of (47) and also \(f\ge 0\).
As the next step, we apply a form of Positivstellensatz, more precisely [16, Thm. 7.3.11, p. 174]. We verify the conditions now. The highest homogeneous parts of \(h_1,\ldots ,h_{N_1}\) are \(-u^8, -s_1^8,\ldots ,-s_m^8, -r_1^8,\ldots ,-r_m^8, -(t_1^8+ \cdots +t_n^8)\) respectively, and it is easy to see that at least one of them is negative at every \((\mathbf {x},t)\in \mathbb {R}^{3n-1}{\setminus }\{(0,\ldots ,0)\}\), i.e. condition (7.3.11.1) is satisfied. Of course, they have the same degree, N, and (59) is compact. The theorem states that if \(f>0\) on \({\widehat{W}}\), then f is in the (Archimedean) module generated by \(h_1,h_2,\ldots ,h_{N_1}\) of level N, i.e.
where \(\sigma _0,\sigma _1,\ldots ,\sigma _{N_1}\) are sums of N-th powers, i.e. they are from
Indirectly, assume that \(f>0\) on \({\widehat{W}}\) which implies that (61) holds.
As the next step, we apply a substitution for (61) to simplify it. The expression on the right of (61) after substitution \(\mathbf {Y}\), where
has the following structure. Obviously,
Also, if \(\sigma \in \sum ^N[\mathbf {x},t]\), then \(\sigma |_\mathbf {Y}\in \sum ^N[u]\). Therefore the right-hand side will have this form:
where slightly abusing the notation, we write \(\sigma |_\mathbf {Y}=\sigma \).
The substitution simplifies greatly the left-hand side of (61)
where we used (45), (46) and also (44).
As the next step, we rewrite \(u^2+1\) with a result of Berr and Wörmann. Obviously, \(u^2+1\) is strictly positive on \([-1,1]\), so it can be written as
where \(\tau _0,\tau _1,\ldots ,\tau _7\in \sum \nolimits ^N[u]\) (recall \(N=8\)). This representation follows from [2, Ex. 4.5, p. 834]. It would be interesting to establish this expansion directly.
Using (66), we rewrite the right-hand side of (64):
where
As the next step, we collect the results of substitution and simplification. Using (61), (66) and (67), the right-hand side is relatively simple, while regarding the left-hand side, we use (65) and we write
To compare the two sides, we need the powers of \(1-u^6\):
We investigate the degrees and the leading terms in (68). On the left-hand side, \(\deg f_1=N=8\). The right-hand side is more involved. We remark that \(\deg {\widetilde{\sigma }}_j=N k_j\) for some \(k_j\in \mathbb {N}\) and the leading coefficient \(\mathrm {lc}({\widetilde{\sigma }}_j)\) of \({\widetilde{\sigma }}_j\) is positive. We investigate the degrees of \({\widetilde{\sigma }}_j (1-u^6)^j\), \(j=0,1,\ldots ,7\) modulo N. Consider the following groups of powers: \(\{0,4\}\) and \(\{1,5\}\) and \(\{2,6\}\) and \(\{3,7\}\). In each group they have the same degrees mod N and the same signs of leading coefficients, e.g. if we take \(\{1,5\}\), then \(\deg {\widetilde{\sigma }}_1 (1-u^6)^1= N k_1 + 6\), \(\deg {\widetilde{\sigma }}_5 (1-u^6)^5= N k_5 + 30= N (k_5 + 3) + 6\) and the signs of the leading coefficients are the same \(\mathrm {sign}\,\mathrm {lc}\big ( {\widetilde{\sigma }}_1 (1-u^6)^1\big ) =\mathrm {sign}\,\mathrm {lc}\big ( {\widetilde{\sigma }}_5 (1-u^6)^5)=-1\). Since the signs are the same, the leading terms in the same group cannot cancel. Also, since the degrees of \({\widetilde{\sigma }}_j (1-u^6)^j\) are different modulo \(N=8\), the leading terms of different groups cannot cancel either. Hence
Taking into account that \(\deg (f_1)=8\), this can happen only when
i.e. \({\widetilde{\sigma }}_1\) is a constant.
Finally, we obtain that
where \(c_1\ge 0\).
So we have
where \(N_2\) is a positive integer and \(A_j(u)\) are polynomials. Again, degree considerations show that the \(A_j\)’s must be constant or linear polynomials, so we can write
where \(\lambda _j>0\), \(\zeta _j\in \mathbb {R}\) are pairwise different and \(c_0\in \mathbb {R}\) and \(N_3\le N_2\).
We finally reach a contradiction by showing that (71) cannot hold. Note that
hence \(f_1(\cdot )\) is even, \(f_1(-u)=f_1(u)\). Exploiting this (and that N is even), we write
Observe that all coefficients of \(\left( (u-\zeta _j)^N +(u+\zeta _j)^N\right) /2\) are non-negative (e.g. the coefficient of \(u^2\) is \(28\zeta _j^6\)). Therefore the coefficient of \(u^2\) is \(4^{m+1} (-2)<0\) (according to the left-hand side of (71)) and it is \(\sum _{j=1}^{N_3} \lambda _j \cdot 28\zeta _j^6 \ge 0\) (according to the right-hand side of (71)). This gives a contradiction.
Therefore, we have \(f\not >0\). Obviously \(f\ge 0\). These two imply that f must have a zero on \({\widehat{W}}\), \((\mathbf {x},t)\in {\widehat{W}}\) with \(f(\mathbf {x},t)=0\), so the system (52) and (53) with \(\mathbf {x}\in W\) has a solution. In turn, this implies that \({\widetilde{W}} \ne 0\) which leads to the fact that (47) has a solution from W (see (48)), which implies that (31) has a solution under the conditions (32), that is, (29) holds which is equivalent to the assertion of Theorem 1. Therefore the proof is complete.
5 Some Technical Lemmas
Lemma 1
Let \(W\in \mathbb {C}\), \(W\ne 0\), \(\gamma \in \mathbb {C}\), \(|\gamma |=1\). Then there exists \(f\in \mathbb {C}\), \(|f|=1\) such that
and the numerator and denominator on the right are real, i.e.
In particular, \(f=\sqrt{\gamma }\) will do. Furthermore, a similar identity also holds: there exists \(f\in \mathbb {C}\), \(|f|=1\) such that
and again, the numerator and the denominator on the right are real. In this case, \(f=i/\sqrt{\gamma }\) is a good choice.
Proof
We use a separate set of notation in this proof.
To see the first assertion, write W, \(\gamma \) and f in polar form in this proof: \(W=re^{i\omega }\), where \(r>0\), \(\gamma =e^{i\alpha }\) and \(f=e^{i\varphi }\). We also have
and similarly for the denominator,
So we have to find \(\varphi \) for the given \(\alpha \) and \(\omega \) such that
So \(\varphi =-\alpha /2\) will do.
For the second assertion, a similar argument yields that \(\varphi =(\pi -\alpha )/2\) will do.
\(\square \)
Lemma 2
Let P, Q be complex polynomials without common zeros. Assume that \(H(u)=P(u)/Q(u)\) is a real rational function, i.e. if \(u\in \mathbb {R}\) and H(u) is finite, then \(H(u)\in \mathbb {R}\). Also assume that the leading coefficients of P and Q are real. Then all the coefficients of P and Q are real.
Proof
We prove it by induction as follows. Write \(P(u)= a u^n + P_1(u)\) and \(Q(u)= b u^m + Q_1(u)\) where \(\deg (P_1)<\deg (P)\) and \(\deg (Q_1)<\deg (Q)\). By the assumptions, \(a,b\in \mathbb {R}\), \(a\ne 0\), \(b\ne 0\).
If \(n\ge m>0\), we can write
which implies that with \(P_2(u):=P(u)- \frac{a}{b} u^{n-m} Q(u)\), \(P_2(u)/Q(u)\) is a real rational function. Note that \(n_1:=\deg (P_2)<\deg (P)=n\). Denote the leading coefficient of \(P_2\) by c, \(c\in \mathbb {C}\), \(c\ne 0\). It is standard to see that
where the left-hand side is real and b on the right is also real. Hence the leading coefficient c of \(P_2\) is real.
If \(m> n>0\), then consider \(1/H(u)=Q(u)/P(u)\) which is again a real rational function.
If \(m=0\), then \(H(u)=P(u)/Q(u)\) is actually a polynomial. Also, \(Q(u)=Q(0)\in \mathbb {R}\), and hence H(u) is a real polynomial, that is, if \(u\in \mathbb {R}\), then \(H(u)\in \mathbb {R}\). It is then standard that the polynomial H(u) must have real coefficients.
Finally, if \(n=0\), then consider 1/H(u) and this way we reduce this case to the case discussed in the previous paragraph.
\(\square \)
We also need the following lemma ( [17, Lem. 2])
Lemma 3
For any sequence \((B_n)\) of Blaschke products of degree m there exist a Blaschke product B of degree \(k\le m\) and a subsequence of \((B_n)\) which converges to B locally uniformly on a set which contains all points of \( \mathbb {D}\cup \mathbb {T}\) with the possible exception of at most \(m-k\) boundary points.
References
Agler, J., Lykova, Z.A., Young, N.J.: Finite Blaschke products and the construction of rational \(\Gamma \)-inner functions. J. Math. Anal. Appl. 447(2), 1163–1196 (2017). https://doi.org/10.1016/j.jmaa.2016.10.035
Berr, R., Wörmann, T.: Positive polynomials and tame preorderings. Math. Z. 236(4), 813–840 (2001). https://doi.org/10.1007/PL00004853
Cantor, D.G., Phelps, R.R.: An elementary interpolation theorem. Proc. Am. Math. Soc. 16, 523–525 (1965). https://doi.org/10.2307/2034689
Daepp, U., Gorkin, P., Shaffer, A., Voss, K.: Finding ellipses. In: Carus Mathematical Monographs, vol. 34. MAA Press, Providence (2018). What Blaschke products, Poncelet’s theorem, and the numerical range know about each other. https://doi.org/10.1090/car/034
Garcia, S.R., Mashreghi, J., Ross, W.T.: Finite Blaschke products and their connections. Springer, Cham (2018). https://doi.org/10.1007/978-3-319-78247-8
Glader, C.: Rational unimodular interpolation on the unit circle. Comput. Methods Funct. Theory 6(2), 481–492 (2006). https://doi.org/10.1007/BF03321625
Glader, C.: Minimal degree rational unimodular interpolation on the unit circle. Electron. Trans. Numer. Anal. 30, 88–106 (2008)
Glader, C., Lindström, M.: Finite Blaschke product interpolation on the closed unit disc. J. Math. Anal. Appl. 273(2), 417–427 (2002). https://doi.org/10.1016/S0022-247X(02)00249-4
Gorkin, P., Mortini, R.: Radial limits of interpolating Blaschke products. Math. Ann. 331(2), 417–444 (2005). https://doi.org/10.1007/s00208-004-0588-0
Gorkin, P., Rhoades, R.C.: Boundary interpolation by finite Blaschke products. Constr. Approx. 27(1), 75–98 (2008). https://doi.org/10.1007/s00365-006-0646-3
Hjelle, G.A.: Approximations by interpolating Blaschke products. Ph.D. thesis, Norwegian University of Science and Technology, Trondheim (2006). http://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.527.9301&rep=rep1&type=pdf
Hjelle, G.A.: Unimodular functions and interpolating Blaschke products. Proc. Am. Math. Soc. 134(1), 207–214 (2006). https://doi.org/10.1090/S0002-9939-05-07968-2
Hjelle, G.A.: Constructing interpolating Blaschke products with given preimages. Comput. Methods Funct. Theory 7(1), 43–54 (2007). https://doi.org/10.1007/BF03321630
Horn, R.A., Johnson, C.R.: Matrix Analysis, 2nd edn. Cambridge University Press, Cambridge (2013)
Jones, W.B., Ruscheweyh, S.: Blaschke product interpolation and its application to the design of digital filters. Constr. Approx. 3(4), 405–409 (1987). https://doi.org/10.1007/BF01890578
Prestel, A., Delzell, C.N.: Positive polynomials. In: Springer Monographs in Mathematics. Springer, Berlin (2001). From Hilbert’s 17th problem to real algebra.https://doi.org/10.1007/978-3-662-04648-7
Semmler, G., Wegert, E.: Boundary interpolation with Blaschke products of minimal degree. Comput. Methods Funct. Theory 6(2), 493–511 (2006). https://doi.org/10.1007/BF03321626
Semmler, G., Wegert, E.: Finite Blaschke products with prescribed critical points, Stieltjes polynomials, and moment problems. Anal. Math. Phys. 9(1), 221–249 (2019). https://doi.org/10.1007/s13324-017-0193-5
Younis, R.: Interpolation by a finite Blaschke product. Proc. Am. Math. Soc. 78(3), 451–452 (1980).https://doi.org/10.2307/2042343
Acknowledgements
The research of Sergei Kalmykov was supported by Moscow Center for Fundamental and Applied Mathematics, Agreement with the Ministry of Science and Higher Education of the Russian Federation, No. 075-15-2019-1623 and the Russian Science Foundation grant 19-71-30004 (Section 2). Béla Nagy was supported by DAAD-TKA Research Project ”Harmonic Analysis and Extremal Problems” #308015. The authors thank the referee for careful reading and helpful suggestions.
Funding
Open access funding provided by University of Szeged.
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by Elias Wegert.
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Kalmykov, S., Nagy, B. Positive Polynomials and Boundary Interpolation with Finite Blaschke Products. Comput. Methods Funct. Theory 23, 49–72 (2023). https://doi.org/10.1007/s40315-021-00430-4
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s40315-021-00430-4
Keywords
- Blaschke product
- Boundary interpolation
- Rational interpolation
- Real algebraic geometry
- Positivstellensatz