1 Introduction

Finding exact constants is one of the most desirable goals in constructive approximation. Many classical results of this sort are collected in the addendum to the Akhiezer book [3]. We are happy to mention in this connection [7]. This paper deals with a certain clarification of the famous exact Remez inequality, that is, we provide the exact constant in the Andrievskii problem defined below.

Based on several results by Erdélyi, Saff and himself [5, 6, 15, 16], Andrievskii posed the following problem.

Problem 1.1

Let \({\mathcal {P}}_n\) be the collection of polynomials of degree at most n. Let E be a closed subset of \([-1,1]\), and |E| denote its Lebesgue measure. For \(x_0\in [-1,1]\), define

$$\begin{aligned} M_n(x_0,E)=\sup \{|P_n(x_0)|: P_n\in {\mathcal {P}}_n, |P_n(x)|\le 1 \ \text {for}\ x\in E\}. \end{aligned}$$
(1.1)

For \(\delta \in (0,1)\), find

$$\begin{aligned} L_{n,\delta }(x_0)=\sup \{M_n(x_0,E): E\ \text {such that}\ |E|\ge 2-2\delta \}. \end{aligned}$$
(1.2)

Let us comment the setting with the following three evident remarks:

  1. (1)

    \(L_{n,\delta }(x_0)\) is even, thus we will consider only \(x_0\in [-1,0]\).

  2. (2)

    \({\mathcal {L}}_{n,\delta }:=\sup _{x_0\in [-1,0]}L_{n,\delta }(x_0)\) is the famous Remez constant [24]. It is attained at the endpoint \(-1\) by the Chebyshev polynomial \( R_{n,\delta }(x)\) for the interval \([-1+2\delta ,1]\), i.e.,

    $$\begin{aligned} {\mathcal {L}}_{n,\delta }=\mathfrak {T}_n\left( \frac{1+\delta }{1-\delta }\right) ,\quad R_{n,\delta }(x)=\mathfrak {T}_n\left( \frac{\delta - x}{1-\delta }\right) , \end{aligned}$$

    where \(\mathfrak {T}_n\) denotes the standard Chebyshev polynomial of degree n associated with \([-1,1]\),

    $$\begin{aligned} \mathfrak {T}_n(x)=\frac{1}{2}\left( (x+\sqrt{x^2-1})^n+(x-\sqrt{x^2-1})^n\right) . \end{aligned}$$

    We will henceforth call \(R_{n,\delta }\) the Remez polynomial.

  3. (3)

    Clearly, for Problem 1.1, \( R_{n,\delta }(x)\) cannot be extremal for all \(x_0\in [-1,0]\) as soon as \(\delta \) is sufficiently small. Let \(\delta =1/2\). Then, \(R_{n,1/2}(0)=1\), while for the so-called Akhiezer polynomial \(A_{2m,\delta }(x)\) we get \(A_{2m,1/2}(0)=\mathfrak {T}_m\left( 5/3\right) >1\). Recall [1] or [4, Chapter 10]

    $$\begin{aligned} A_{2m,\delta }(x)=\mathfrak {T}_m\left( \frac{1+\delta ^2-2x^2}{1-\delta ^2}\right) \end{aligned}$$

    is the even Chebyshev polynomial on the set \([-1,1]\setminus (-\delta ,\delta )\).

Andrievskii raised his question on several international conferences, including Jaen Conference on Approximation Theory, 2018. The third remark highlights that the problem is non-trivial. We found it highly interesting and in this paper we provide its complete asymptotic solution.

In [1, 2], Akhiezer studied various extremal problems for polynomials bounded on two disjoint intervals \(E(\alpha ,\delta )=[-1,1]\setminus (\alpha -\delta ,\alpha +\delta )\), \(\delta -1<\alpha \le 0\), see also [3, Appendix Section 36 and Section 38 in German translation].

Definition 1.2

We say that a polynomial \(A_{n,\alpha ,\delta }(x)\) is the Akhiezer polynomial for \(E(\alpha ,\delta )\) with respect to an internal gap if it solves the following extremal problem

$$\begin{aligned} A_{n,\alpha ,\delta }(x_0)=\sup \{|P_n(x_0)|: \ P_n\in {\mathcal {P}}_n,\ |P_n(x)|\le 1\ \text {for}\ x\in E(\alpha ,\delta )\}, \end{aligned}$$
(1.3)

where \(x_0\in (\alpha -\delta ,\alpha +\delta )\).

We point out that \(A_{2m,\delta }\) as introduced above corresponds to the symmetric case \(\alpha =0\) and even \(n=2m\). We describe properties of Akhiezer polynomials in Sect. 2.1. Note, in particular, that the extremal property of \(A_{n,\alpha ,\delta }(x)\) does not depend on which point \(x_0\in (\alpha -\delta ,\alpha +\delta )\) was fixed in (1.3).

In Sect. 3, we prove the following theorem.

Theorem 1.3

The extremal value \(L_{n,\delta }(x_0)\) is assumed either on the Remez polynomial \(R_{n,\delta }(x)\) or on an Akhiezer polynomial \(A_{n,\alpha ,\delta }(x)\) with a suitable \(\alpha \).

Our main result, the asymptotic behavior of \(L_{n,\delta }(x_0)\), is presented in Sect. 4. Nowadays, the language of potential theory is so common and widely accepted, see, e.g., [6, 11, 13, 19, 27], that we will formulate our asymptotic result using this terminology.

Let \(G_{\alpha ,\delta }(z,z_0)\) be the Green function in the domain \(\Omega :=\overline{{\mathbb {C}}}\setminus E(\alpha ,\delta )\) with respect to \(z_0\in \Omega \). In particular, we set \(G_{\alpha ,\delta }(z)=G_{\alpha ,\delta }(z,\infty )\). Respectively, \(G_{\delta }(z)\) is the Green function of the domain \(\overline{{\mathbb {C}}}\setminus E(\delta )\) with respect to infinity, where \(E(\delta )=[-1+2\delta ,1]\). It is well known, that

$$\begin{aligned} R_{n,\delta }(x)=\frac{e^{n G_{\delta }(x)}+e^{-n G_{\delta }(x)}}{2},\quad x\in [-1, -1+2\delta ], \end{aligned}$$
(1.4)

and

$$\begin{aligned} \lim _{\alpha \rightarrow -1+\delta }G_{\alpha ,\delta }(x)=G_{\delta }(x),\quad x\in (-1, -1+2\delta ). \end{aligned}$$

Lemma 1.4

Let \(\Phi _\delta (x)\) be the upper envelope

$$\begin{aligned} \Phi _\delta (x)=\sup _{\alpha \in (\delta -1,0]}G_{\alpha ,\delta }(x). \end{aligned}$$
(1.5)

If for \(x_0\) the supremum is attained for some internal point \(\alpha \in (\delta -1,0]\), then \(x_0=x_0(\alpha )\) is a unique solution of the equation

$$\begin{aligned} {\partial }_\alpha G_{\alpha ,\delta }(x)=0. \end{aligned}$$
(1.6)

Otherwise, it is attained as \(\alpha \rightarrow \delta -1\) and \(\Phi _\delta (x_0)=G_{\delta }(x_0)\).

Note that the non-trivial claim in Lemma 1.4 is the uniqueness of \(x_0(\alpha )\). We will provide a proof of this fact in Sect. 4.2. Thus, we show that the upper envelope (1.5) is in fact the upper envelope of just two curves. One of them is the graph of \(G_{\delta }(x)\) given in terms of elementary functions, see (4.8). The other one is obtained in the following way: first we find the unique solution \(x_0(\alpha )\) of (1.6) and then we form the curve \((x_0(\alpha ), G_{\alpha ,\delta }(x_0(\alpha )))\). Although each individual Green function \(G_{\alpha ,\delta }(x)\) can be expressed in terms of elliptic functions [4, \(\S \) 55 equation (4)], due to the implicit definition of \(x_0(\alpha )\) by means of (1.6), we don’t believe that a parametric description of the upper envelope in terms of elliptic functions exists. However, for any fixed \(\delta >0\) it can be solved numerically, as we demonstrate in Example 4.5 including a graph of the curve \((x_0(\alpha ), G_{\alpha ,\delta }(x_0(\alpha )))\), cf. Fig. 3. Note that the endpoints of this curve are also given explicitly.

Our main theorem below shows that the asymptotics of \(L_{n,\delta }(x_0)\) are described by the upper envelope \(\Phi _\delta (x_0)\).

Theorem 1.5

The following limit exists

$$\begin{aligned} \lim _{n\rightarrow \infty }\frac{1}{n}\log 2L_{n,\delta }(x_0)=\Phi _\delta (x_0). \end{aligned}$$

To be more precise, if \(\Phi _\delta (x_0)=G_{\delta }(x_0)\), then

$$\begin{aligned} \log 2L_{n,\delta }(x_0)=n \Phi _\delta (x_0)+o(1). \end{aligned}$$
(1.7)

If \(\Phi _\delta (x_0)=G_{\delta ,\alpha }(x_0)\) for some \(\alpha \in (\delta -1,0]\), then the Totik–Widom bound

$$\begin{aligned} \log 2L_{n,\delta }(x_0)=n \Phi _\delta (x_0)+O(1) \end{aligned}$$
(1.8)

holds.

For the classical Chebyshev problem related to a fixed set E, the asymptotics are described in terms of the Green function associated with this set. In contrast to this, \(\Phi _\delta (x_0)\), respectively, \(E(\alpha _0,\delta )\) for some extremal \(\alpha _0\) depend on \(x_0\) in a quite sophisticated way. We present an asymptotic diagram which allows to describe completely the asymptotic behavior of \(L_{n,\delta }(x_0)\) for all \(x_0\in [-1,1]\) and a given \(\delta \).

The organization of the paper is as follows. In Sect. 2, we prove part 1 of Theorem 1.3 and present technical tools that will be used throughout the paper. The structure of Chebyshev polynomials can be revealed by their representation in terms of conformal mappings on so-called comb domains. However, the relation between the parameters describing a given set E and the related comb domain are highly non-trivial. We provide a comprehensive description of Akhiezer polynomials and their dependence on the position \(\alpha \). In Sect. 2.3, we finish the proof of Theorem 1.3. The proof of Theorem 1.5 and a description of the upper envelope (1.5) are given in Sect. 4.

2 Preliminaries

In [26], the sharp constant in the Remez inequality for trigonometric polynomials on the unit circle was given. The proof was based on the following two steps:

  1. (i)

    The structure of possible extremal polynomials was revealed with the help of their comb representations,

  2. (ii)

    The principle of harmonic measure (a monotonic dependence on a domain extension) allows to get an extremal configuration for the comb parameters in the given problem.

We also start with recalling comb representations for extremal polynomials, see Sect. 2.1. We refer the reader to [17, 25] and the references therein for more information about the use of comb mappings in approximation theory. However, we doubt that the step (ii) is applicable to the Andrievskii problem. That is, that comparably simple arguments from potential theory such as the principle of harmonic measure, would also allow us to bring a certain fixed configuration (comb) to an extremal one. Instead, we develop here an infinitesimal approach, closely related to the ideas of Loewner chains [12, 22].

Using this method, we will prove in Sect. 3 that the extremal solution for \(L_{n,\delta }(x_0)\) is either a Chebyshev or an Akhiezer polynomial. For this reason, we continue in Sect.  2.3 with the complete discussion of Akhiezer polynomials \(A_{n,\alpha ,\delta }\), see (1.3). Recall that \(A_{n,\alpha ,\delta }\) is extremal on two given intervals with respect to points \(x_0 \in (\alpha -\delta ,\alpha +\delta )\).Footnote 1 Generically, \(A^{-1}_{n,\alpha ,\delta }([-1, 1])\) is a union of three intervals—the set contains also an additional interval outside of \([-1, 1]\). We demonstrate our infinitesimal approach showing dependence of this additional interval on \(\alpha \) for fixed n and \(\delta \). In the language of comb domains, we will observe a rather involved dependence of the comb parameters in a simple monotonic motion of \(\alpha \). The domain can undergo all possible variations: to narrow, expand, or a combination of both, see Theorem 2.14. Essentially, this is the base for our belief that simple arguments, in the spirit of (ii), in the Andrievskii problem are hardly possible.

If for fixed \(x_0\) the extremal polynomial is a Remez polynomial, there is no additional interval outside of \([-1,1]\). Intuitively, the same considerations as were used in [26] should work. However, a technical difference prevents direct applications of the principle of harmonic measure. Namely, the Lebesgue measure on the circle corresponds to the harmonic measure evaluated at 0, while in the sense of potential theory the Lebesgue measure on \({\mathbb {R}}\) corresponds to the Martin or Phragmén–Lindelöf measure, [20]. Although we are convinced that a limiting process would allow to overcome this technical issue, we provide an alternative proof below; cf. Lemma 2.5.

2.1 Comb Representation for Extremal Polynomials

By a regular comb, we mean a half strip with a system of vertical slits

$$\begin{aligned} \Pi =\{z: \mathfrak {R}z\in (\pi n_-,\pi n_+), \ \mathfrak {I}z>0 \}\setminus \{z=\pi k+iy: y\in (0,h_k), k\in (n_-,n_+)\} \end{aligned}$$

where \(n_+-n_-=n\in {\mathbb {N}}\) and the k’s are integers. We call \(h_k\), \(h_k\ge 0\), the height of the k-th slit and point out that the degeneration \(h_k=0\) is possible. Let \(\theta :{\mathbb {C}}_+\rightarrow \Pi \) be a conformal mapping of the upper half-plane onto a regular comb such that \(\theta (\infty )=\infty \). Then,

$$\begin{aligned} T_n(z)=\cos \theta (z) \end{aligned}$$

defines a polynomial of degree n. Let \(E\subset [-1,1]\) be compact and \(x_0\in [-1,1]\setminus E\). Moreover, let (ab) denote the maximal open interval in \({\mathbb {R}}\setminus E\) that contains \(x_0\). Then, using the technique of Markov correction terms we obtain the following representation of the extremizer of (1.1)

Theorem 2.1

([25, 7.5. Basic theorem], [14, Theorem 3.2.]) Under the normalization \(T_n(x_0)>0\), there exists a unique extremal polynomial for (1.1) and it only depends on the gap (ab) and not on the particular point \(x_0\in (a,b)\). Moreover, let \(\theta :{\mathbb {C}}_+\rightarrow \Pi \) be a comb mapping onto a regular comb \(\Pi \) and E be such that:

  1. (i)

    \(h_0>0\) and \(a=\theta ^{-1}(-0), b=\theta ^{-1}(+0)\),

  2. (ii)

    E contains at least one of the points \(\theta ^{-1}(k\pi \pm 0)\), for all \(k\in (n_-,n_+)\),

  3. (iii)

    E contains at least one of the points \(\theta ^{-1}(\pi n_-), \theta ^{-1}(\pi n_+)\),

Then,

$$\begin{aligned} T_n(z)=\cos \theta (z) \end{aligned}$$
(2.1)

is an extremal polynomial for E and the interval (ab). Vice versa, if \(T_n\) is an extremal polynomial for a set E and an interval (ab), then there exists a regular comb with these properties such that (2.1) holds.

Let us now specify to the case of Akhiezer polynomials, where \(E=E(\alpha ,\delta )=[-1,1]\setminus (\alpha -\delta ,\alpha +\delta )\). By the above theorem, the so-called n-extension \(E_n=E_n(\alpha ,\delta ):=A_{n,\alpha ,\delta }^{-1}([-1,1])\) can be of the following types:

  1. (i)

    there is an additional interval to the right of 1,

  2. (ii)

    E is extended at 1,

  3. (iii)

    E is extended at \(-1\),

  4. (iv)

    there is an additional interval to the left of \(-1\).

The corresponding comb-mapping realization is collected in the following Corollary:

Corollary 2.2

For a fixed set \(E=E(\alpha ,\delta )\) and degree n the extremal polynomial \(A_{n,\alpha ,\delta }(z)\) of (1.1) is of the form (2.1). The unique comb \(\Pi _{n,\alpha ,\delta }\) has one of four possible shapes shown in Figs. 1 and 2. Moreover, the following normalizations distinguish the cases

  1. (i)

    \(\theta ^{-1}(\pi n_-)=-1\), \(\theta ^{-1}(\pi (n_+-1)-0)=1\), see Fig. 1, left,

  2. (ii)

    \(\theta ^{-1}(\pi n_--0)=-1\), \(\theta (1)\in [\pi (n_+-1), \pi n_+]\) see Fig. 1, right,

  3. (iii)

    \(\theta (-1)\in [\pi (n_-), \pi (n_-+1)]\), \(\theta ^{-1}(\pi n_--0)=-1\), see Fig. 2, left,

  4. (iv)

    \(\theta ^{-1}(\pi (n_-+1)+0)=-1\), \(\theta ^{-1}(\pi n_+)=1\), see Fig. 2, right.

Fig. 1
figure 1

Comb domains for the cases (i) and (ii)

Fig. 2
figure 2

Comb domains for the cases (iii) and (iv)

Remark 2.3

Let us denote the additional interval for the cases (i) and (iv) by \(I_n\). These cases include the limit cases \(h_{n_+-1}=\infty \) and \(h_{n_-+1}=\infty \), respectively. That is, the extremal polynomial is of the degree \(n-1\). Note that \(A_{n,\alpha ,\delta }\) has a zero on \(I_n\). The aforementioned degeneration corresponds to “the zero being at \(\infty \).” On the other hand, also the degenerations \(h_{n_+-1}=0\) and \(h_{n_-+1}=0\) are possible, which allow for a smooth transition to the cases (ii) and (iii), respectively.

2.2 Reduction to Remez Polynomials (Proof of Theorem 1.3, Part 1)

As we have mentioned in the beginning of this section, we cannot use the ideas of the principle of harmonic measure directly. We overcome this technical problem by using transition functions from the theory of Loewner chains [12, 22]. Let us consider an arbitrary regular comb \(\Pi =\Pi (0)\) as in Theorem 2.1 and let us fix a slit with height \(h_k>0\). Let \(h_k({\epsilon })\) be strictly monotonically decreasing such that \(h_k(0)=h_k\) and \(\Pi ({\epsilon })\) be the comb which coincides with \(\Pi \), but \(h_k\) is reduced to \(h_k({\epsilon })\). Let \(\theta \) and \(\theta _{\epsilon }\) be the corresponding comb mappings. Then, clearly \(\Pi \subset \Pi _{{\epsilon }}\) and thus the transition function

$$\begin{aligned} w(z,{\epsilon })=w_{\epsilon }(z):=\theta ^{-1}_{\epsilon }(\theta (z)) \end{aligned}$$

is well defined and is an analytic map from \({\mathbb {C}}_+\) into \({\mathbb {C}}_+\). If we define \(I({\epsilon }):=\theta ^{-1}([\pi k+ih_k({\epsilon }), \pi k+ih_k])\) and \(c_k({\epsilon })=\theta _{\epsilon }^{-1}(\pi k+ih_k({\epsilon }))\), then \(w_{\epsilon }:{\mathbb {R}}\setminus I({\epsilon })\rightarrow {\mathbb {R}}\setminus \{c_k({\epsilon })\}\) is one-to-one and onto. The arc \(J({\epsilon }):=\theta _{\epsilon }^{-1}([\pi k+ih_k({\epsilon }), \pi k+ih_k])\) lies, except its endpoint in \({\mathbb {C}}_+\) and \(w_{\epsilon }:{\mathbb {C}}_+\rightarrow {\mathbb {C}}_+\setminus J({\epsilon })\) is conformal.

Lemma 2.4

The Nevanlinna function \(w_{\epsilon }\) admits the representation

$$\begin{aligned} w(z,{\epsilon })=z+\int _{I({\epsilon })}\frac{1-z^2}{1-\xi ^2}\frac{\mathrm {d}\sigma _{\epsilon }(\xi )}{\xi -z}. \end{aligned}$$
(2.2)

Proof

It follows from the above, that \(w_{\epsilon }\) is a Nevanlinna function, i.e., is an analytic function with positive imaginary part on \({\mathbb {C}}_+\), and that the measure in its integral representation is supported on \(I({\epsilon })\). That is, we can write

$$\begin{aligned} w_{\epsilon }(z)=A_{\epsilon }z+B_{\epsilon }+\int _{I({\epsilon })}\frac{\mathrm {d}\sigma _{\epsilon }(\xi )}{\xi -z},\quad A_{\epsilon }\ge 0,B_{\epsilon }\in {\mathbb {R}}. \end{aligned}$$

Moreover, it satisfies the normalization conditions

$$\begin{aligned} w_{\epsilon }(-1)=-1,\quad w_{\epsilon }(1)=1,\quad w_{\epsilon }(\infty )=\infty , \end{aligned}$$

which yields

$$\begin{aligned} B_{\epsilon }=1-A_{\epsilon }-\int _{I({\epsilon })}\frac{\mathrm {d}\sigma _{\epsilon }(\xi )}{\xi -1},\quad A_{\epsilon }=1+\int _{I({\epsilon })}\frac{\mathrm {d}\sigma _{\epsilon }(\xi )}{1-\xi ^2}. \end{aligned}$$
(2.3)

From this, we obtain (2.2). \(\square \)

With the aid of this lemma, one could already recover Remez’ result, that for points on the boundary the extremal configuration is an interval.

Lemma 2.5

Let

$$\begin{aligned} E=[b_0,1]\setminus \bigcup _{j=1}^g(a_j,b_j),\quad x_0\in [-1,b_0),\quad |E|=2-2\delta \end{aligned}$$

and \(T_n(x,E)\) be the extremal polynomial for \(x_0\) as described in Theorem 2.1. Then,

$$\begin{aligned} R_{n,\delta }(x_0)> T_n(x_0,E). \end{aligned}$$

Proof

For an elementary proof even in the multi-dimensional case, we refer to [8, 9, Lemma 2]. \(\square \)

Remark 2.6

Continuing the heuristics provided at the beginning of this section, we give an interpretation of this result in terms of the harmonic measure. Assume that \(\theta _{\epsilon }\) is a conformal map for which one of the slit heights was reduced, \(E_{\epsilon }=\theta _{\epsilon }^{-1}([\pi n_-,\pi n_+])\) and \(w_{\epsilon }\) the corresponding transition function. Lemma 2.5 follows easily, provided that \(|E_{\epsilon }|>|E|\) and \(w_{\epsilon }(x_0)>x_0\). Let us assume we were interested in the harmonic measure of E instead of its Lebesgue measure. Then, the above properties have a probabilistic interpretation: namely, the first one corresponds to the probability that a particle which starts at a point which is close to infinity in the domain \(\Pi _{\epsilon }\) and \(\Pi \) and terminates on the base of the comb. From this perspective, it is clear that this increases if one of the slit heights is decreased. Similarly, writing the second one as \(w_{\epsilon }(x_0)+1>x_0+1\), it corresponds to the probability that a particle terminates in \([\theta (-1),\theta (x_0)]\) in \(\Pi \) and \(\Pi _{\epsilon }\), respectively. Again, this explains, why the value should be increased. This can be made rigorous by using Lemma 2.4.

The above interpretation uses the fact that there is no additional portion of E outside of \([-1,1]\). For Akhiezer polynomials, this is not true. This makes the problem essentially more delicate and we extend our tools by using the infinitesimal variation \({\dot{w}}\) as defined in (2.4); cf. also Remark 2.15.

2.3 Transformation of the Akhiezer Polynomials as \(\alpha \) is moving

We believe that the observations presented in this section are of independent interest. Our goal is to describe the transformation of the comb as the interval starting from the center moves continuously to the left. This should correspond to a continuous transformation of the comb, which seems at the first sight impossible, since the base of the slits are positioned only at integers. We show in the following how the cases (i), (ii), (iii), (iv) can be connected by a continuous transformation: let us start with case (i) in the degenerated configuration \(h_{n_+-1}=\infty \). Then, we start decreasing \(h_{n_+-1}\) until we reach \(h_{n_+-1}=0\). Now we are in the situation that \(\theta (1)=\pi (n_+-1)\) and we continue with case (ii). That is, \(\theta (1)\) increases until it reaches \(\theta (1)=\pi n_+\). All the time \(\theta (-1)=\pi n_-\). Now we continue with case (iii) and increase \(\theta (-1)\) until the point \(\theta (-1)=\pi (n_-+1)\). We continue with case (iv) and increase \(h_{n_-+1}\) from \(h_{n_-+1}=0\) to \(h_{n_-+1}=\infty \). We have arrived at our initial configuration but the base of the comb was shifted from position \((\pi n_-,\pi (n_+-1))\) to position \((\pi (n_-+1),\pi n_+)\).

We believe it is helpful to understand these transformations also on the level of \(E_n\). Our initial configuration corresponds to \(E=E_n\). We view this as the additional interval is hidden at \(\infty \). Starting case (i) corresponds to the motion that the position of the additional interval decreases from \(+\infty \) until it joins E. This corresponds to the change from case (i) to (ii). When it is fully absorbed, we arrive at case (iii), i.e., E starts to be extended to the left until the extension separates and starts to decrease to \(-\infty \). And the circle starts from the beginning.

Theorem 2.7

Moving (ab) to the left corresponds to a succession of continuous transformations of the comb as described above. If n is odd, we start with case (i) and \(h_{n_+-1}=\infty \). If n is even, we start with case (iii) and \(\theta (-1)=\pi n_-\).

Remark 2.8

When \(\alpha =0\) by symmetry, the extremal polynomial is even and all critical points of the extremal polynomial are equally distributed on \([-1,-\delta ]\) and \([\delta ,1]\). The above procedure describes how all critical points are moved from the interval \([-1,\alpha -\delta ]\) to \([\alpha +\delta ,1]\) as \(\alpha \) is decreasing. In the limit when \(\alpha \) approaches \(-1+\delta \), all critical points will be on \([-1+2\delta ,1]\) and the extremal polynomial will transform into the Remez polynomial \(R_{n,\delta }\).

As we have already indicated at the end of Section 2.2 the proof of Theorem 2.7 requires in addition to the transition function \(w_{\epsilon }\) its infinitesimal transform \({\dot{w}}\). Let us introduce the notation

$$\begin{aligned} w'(z,{\epsilon })=\partial _zw(z,{\epsilon }),\quad {\dot{w}}(z,{\epsilon })=\partial _{\epsilon }w(z,{\epsilon })\quad \text {and}\quad {\dot{w}}(z)={\dot{w}}(z,0). \end{aligned}$$
(2.4)

Lemma 2.9

Under an appropriate choice of \({\epsilon }\), \(w(z,{\epsilon })\) is differentiable with respect to \({\epsilon }\) and there exists \(\sigma _k>0\) such that

$$\begin{aligned} {\dot{w}}(z):=\lim _{{\epsilon }\rightarrow 0}\frac{w(z,{\epsilon })-w(z,0)}{{\epsilon }}= \frac{1-z^2}{1-c_k^2}\frac{\sigma _k}{c_k-z}, \end{aligned}$$
(2.5)

where \(c_k=c_k(0)=\theta ^{-1}(\pi k+ih_k)\).

Proof

We note that if \(k=n_-+1\) or \(k=n_+-1\) and we are in the situation of Theorem 2.1 (i) or (iv), then due to (2.3) \(A_{\epsilon }<1\) and \(A_{\epsilon }>1\) otherwise. In any case, we obtain from the properties of \(w_{\epsilon }\) that \(A_{\epsilon }\) is strictly monotonic. Therefore, if we fix \({\epsilon }_0\) and define \( \beta _0=\left| \int _{I({\epsilon }_0)}\frac{\mathrm {d}\sigma _{{\epsilon }_0}(\xi )}{1-\xi ^2}\right| , \) then \( {\epsilon }\mapsto \left| \int _{I({\epsilon })}\frac{\mathrm {d}\sigma _{\epsilon }(\xi )}{1-\xi ^2}\right| \) maps \([0,{\epsilon }_0]\) continuously and bijectively on \([0,\beta _0]\). Through a reparametrization we can achieve that \( \left| \int _{I({\epsilon }')}\frac{\mathrm {d}\sigma _{{\epsilon }'}(\xi )}{1-\xi ^2}\right| =\frac{{\epsilon }'\beta _0}{{\epsilon }_0}. \) Thus, the measures \(\frac{\mathrm {d}\sigma _{{\epsilon }'}(\xi )}{(1-\xi ^2){\epsilon }'}\) are normalized and we get by passing to a subsequence

$$\begin{aligned} \lim _{k\rightarrow \infty }\frac{w(z,{\epsilon }'_k)-z}{{\epsilon }'_k}=\lim \limits _{k\rightarrow \infty }\int _{I({\epsilon }'_k)}\frac{1-z^2}{1-\xi ^2}\frac{1}{\xi -z}\frac{\mathrm {d}\sigma _{{\epsilon }'_k}(\xi )}{{\epsilon }_k'}=\int \frac{1-z^2}{1-\xi ^2}\frac{\mathrm {d}\sigma _\infty (\xi )}{\xi -z}. \end{aligned}$$

Since the support of \(\sigma _{{\epsilon }'_k}\) shrinks to \(\{c_k\}\), we get that \(\sigma _\infty =\sigma _k\delta _{c_k}\), for \(\sigma _k/|1-c_k^2|=\beta _0/{\epsilon }_0\). This concludes the proof. \(\square \)

We are also interested in the case of growing slit heights. This corresponds to

$$\begin{aligned} u_{\epsilon }(x)=w_{\epsilon }^{-1}(x), \end{aligned}$$

for some transition function as defined above. Note that since \(c_k({\epsilon })\rightarrow c_k\) as \({\epsilon }\rightarrow 0\), this inverse is well defined on \({\mathbb {R}}\) away from a vicinity of \(c_k\) and we conclude since \(w'(z,0)=1\) that

$$\begin{aligned} \dot{u}(x)=-{\dot{w}}(x). \end{aligned}$$

The following lemma discusses the case (i). We show that decreasing \(h_{n_+-1}\) and simultaneously increasing \(h_0\) appropriately allows us to move the gap \((\alpha -\delta ,\alpha +\delta )\) to the left without changing its size. Let us set \((a,b)=(\alpha -\delta ,\alpha +\delta )\) let c be the critical point in (ab) and \(c_+\) be the critical point outside \([-1,1]\), i.e., \(c=\theta ^{-1}(ih_0)\), \(c_+=\theta ^{-1}(\pi (n_+-1)+ih_{n_+-1})\).

Lemma 2.10

Let \(E(\alpha ,\delta )\) with \(\alpha \le 0\) be such that the corresponding extremal polynomial \(A_{n,\alpha ,\delta }(z)\) corresponds to the case (i). Then, the infinitesimal variation \({\dot{w}}(x)\) generated by decreasing the height \(h_{n_+-1}\) under the constraint of a constant gap length (\(\delta =\mathrm{const}\)) leads to an increase in \(h_0\). Moreover, in this variation the gap is moving to the left, that is, \(\alpha ({\epsilon })\) is decreasing.

Proof

Let \(w_{n_+-1}\) and \(w_0\) be transforms corresponding to a variation of the slit heights \(h_{n_+-1}\) and \(h_0\). In (2.5), we chose \(\sigma _{n_+-1}>0\) and determine the sign of \(\sigma _0\) by the condition of constant gap length. Thus, the total transform corresponds to \(w(z,{\epsilon })=w_{n_+-1}(w_0(z,{\epsilon }),{\epsilon })\) and hence, by the previous computations, we find that

$$\begin{aligned} \dot{w}(x)=\frac{1-x^2}{1-c^2}\frac{\sigma _0}{c-x}+\frac{1-x^2}{1-c_+^2}\frac{\sigma _+}{c_+-x}, \end{aligned}$$

with \(\sigma _+>0\). The value \(\sigma _0\) is determined due to the constraint \(w(b,{\epsilon })-w(a,{\epsilon })=b-a\), i.e., \({\dot{w}}(b)={\dot{w}}(a)\). Thus, we obtain

$$\begin{aligned} 0&=\left( \frac{1-b^2}{c-b}-\frac{1-a^2}{c-a} \right) \frac{\sigma _0}{1-c^2}+\left( \frac{1-b^2}{c_+-b}-\frac{1-a^2}{c_+-a} \right) \frac{\sigma _+}{1-c_+^2} \nonumber \\&=(b-a)\left( \frac{1+ab-c(a+b)}{(c-a)(c-b)} \frac{\sigma _0}{1-c^2}+\frac{1+ab-c_+(a+b)}{(c_+-a)(c_+-b)}\frac{\sigma _+}{1-c_+^2}\right) . \end{aligned}$$
(2.6)

Let

$$\begin{aligned} \ell (x)=1+ab-(a+b)x. \end{aligned}$$
(2.7)

Since \(a+b=2\alpha \le 0\), \(\ell \) is non-decreasing. Moreover, since \(a,b>-1\), \(\ell (-1)=(1+a)(1+b)>0\) and hence, \(\ell (c)>0\) and \(\ell (c_+)>0\). Using that the numerator is negative for both summands, we find that \(\sigma _+>0\) implies \(\sigma _0<0\). In other words, the compression of the height \(h_{n_+-1}\) leads to a growth of \(h_0\). Since both summands are negative,

$$\begin{aligned} {\dot{w}}(a)=\frac{1-a^2}{1-c^2}\frac{\sigma _0}{c-a}+\frac{1-b^2}{1-c_+^2}\frac{\sigma _+}{c_+-a}<0. \end{aligned}$$

Consequently, \({\dot{w}}(b)={\dot{w}}(a)<0\), and we find that the ends of the interval \((a_{\epsilon },b_{\epsilon })\) are moving to the left. This concludes the proof. \(\square \)

Case (iv) is similar to case (i). However, we will encounter certain specific phenomena. First of all, we will increase the value of \(h_{n_-+1}\) to move the given gap to the left. But in order to fulfill the constraint of fixed gap length both an increasing or a decreasing of \(h_0\) is possible.

Lemma 2.11

Let \(E(\alpha ,\delta )\) with \(\alpha \le 0\) be such that the corresponding extremal polynomial \(A_{n,\alpha ,\delta }(z)\) corresponds to the case (iv). Let \(\ell \) be defined as in (2.7) and

$$\begin{aligned} \eta :=\frac{1}{2}\left( \alpha +\frac{1-\delta ^2}{\alpha }\right) , \quad (\eta <-1), \end{aligned}$$
(2.8)

be its zero. If \(c_-<\eta \), the infinitesimal variation \({\dot{w}}(x)\) generated by increasing the height \(h_{n_-+1}\) under the constraint of a constant gap length (\(\delta =\mathrm{const}\)) leads to an increase in \(h_0\) and it leads to a decrease in \(h_0\) if \(c_-\in (\eta ,-1)\). In both cases, the gap is moving to the left, that is, \(\alpha ({\epsilon })\) is decreasing.

Proof

As before, we find that the infinitesimal variation is of the form

$$\begin{aligned} {\dot{w}}(x)=\frac{1-x^2}{1-c^2}\frac{\sigma _0}{c-x}+\frac{1-x^2}{1-c_-^2}\frac{\sigma _-}{c_--x},\quad \sigma _-<0. \end{aligned}$$

Respectively, the constraint \({\dot{w}}(b)={\dot{w}}(a)\) corresponds to

$$\begin{aligned} \frac{\ell (c)}{(c-a)(c-b)}\frac{\sigma _0}{1-c^2}+\frac{\ell (c_-)}{(c_--a)(c_--b)}\frac{\sigma _-}{1-c_-^2}=0. \end{aligned}$$

We have that \(\ell (c_-)<0\) for \(c_-<\eta \). Since \(c_-<-1<a<c<b\), this implies that \(\sigma _0<0\). As before, we conclude that \( \dot{w}(a)<0, \) and the interval is moving to the left. If \(c_-\in (\eta ,-1)\), then \(\ell (c_-)>0\) and therefore \(\sigma _0>0\). In this case, \( {\dot{w}}(b)<0 \) and again the interval is moving to the left. Finally, if \(c_-=\eta \), we have \(\ell (c_-)=0\) and therefore (2.6) implies that \(\sigma _0=0\). \(\square \)

2.3.1 The cases (ii) and (iii)

We will discuss case (ii) and case (iii) simultaneously. Recall that case (ii) corresponds to an extension of E to the right and case (iii) to an extension to the left, i.e., \(\theta (1)\in (\pi (n_+-1),\pi n_+)\) or \(\theta (-1)\in (\pi n_-,\pi ( n_-+1))\). Let \(\Pi \equiv \Pi ({\epsilon })\) but let us increase the normalization point \(\theta _{\epsilon }(\pm 1)\). Let \(w^\pm _{\epsilon }(z)=\theta _{\epsilon }^{-1}(\theta (z))\) be the corresponding transition functions.

Lemma 2.12

Let \(w^\pm _{\epsilon }\) be defined as above. Then, there exists \(\rho ^\pm _{\epsilon }>0\), such that

$$\begin{aligned} w^+_{\epsilon }(z)=-1+\rho ^+_{\epsilon }(z+1),\quad w^-_{\epsilon }(z)=1+\rho ^-_{\epsilon }(z-1),\quad \rho ^+_{\epsilon }<1,\rho _{\epsilon }^->1 \end{aligned}$$
(2.9)

and

$$\begin{aligned} {\dot{w}}^\pm (z)=\tau ^\pm (1\pm z),\quad \tau ^\pm <0. \end{aligned}$$
(2.10)

Proof

We only prove the claim for \(w^+_{\epsilon }\). Since \(\Pi ({\epsilon })\equiv \Pi (0)\), \(w_{\epsilon }\) is just an affine transformation and using \(w_{\epsilon }(-1)=-1\) and \(w_{\epsilon }(\infty )=\infty \) we find (2.9). Since \(\theta _{\epsilon }(1)>\theta (1)\), we obtain that \(w_{\epsilon }(1)=\theta _{\epsilon }^{-1}(\theta (1))<\theta _{\epsilon }^{-1}(\theta _{\epsilon }(1))=1\) and thus \(\rho _{\epsilon }<1\). Therefore,

$$\begin{aligned} {\dot{w}}(z)=\lim \limits _{{\epsilon }\rightarrow \infty }\frac{w_{\epsilon }(z)-z}{{\epsilon }}=(z+1)\lim \limits _{{\epsilon }\rightarrow \infty }\frac{\rho _{\epsilon }-1}{{\epsilon }}=\tau <0. \end{aligned}$$

\(\square \)

Lemma 2.13

Let \(E(\alpha ,\delta )\) with \(\alpha \le 0\) be such that the corresponding extremal polynomial \(A_{n,\alpha ,\delta }(z)\) corresponds to the case (ii) (case (iii)). Then, the infinitesimal variation \({\dot{w}}(x)\) generated by increasing \(\theta (1)\) (increasing \(\theta (-1)\)) under the constraint of a constant gap length (\(\delta =\mathrm{const}\)) leads to an increase (decrease) of \(h_0\). Moreover, in this variation the gap is moving to the left, that is, \(\alpha ({\epsilon })\) is decreasing.

Proof

We only prove the case (ii). We have

$$\begin{aligned} {\dot{w}}(x)=\tau (x+1)+\frac{1-x^2}{1-c^2}\frac{\sigma _0}{c-x}, \quad \tau <0. \end{aligned}$$
(2.11)

Applying the constraint \({\dot{w}}(a)={\dot{w}}(b)\), we obtain

$$\begin{aligned} 0=(b-a)\left( \tau +\frac{1+ab-(a+b)c}{(c-a)(c-b)}\frac{\sigma _0}{1-c^2} \right) . \end{aligned}$$

Since \(\ell (c)>0\), this implies \(\sigma _0<0\) and thus \(h_0\) is increasing. Moreover, \({\dot{w}}(a)<0\), which concludes the proof. \(\square \)

We summarize our results in the following theorem:

Theorem 2.14

Let \(\eta \) be defined by (2.8). Then, we have:

  1. (i)

    \(h_0\) is increasing,

  2. (ii)

    \(h_0\) is increasing,

  3. (iii)

    \(h_0\) is decreasing,

  4. (iv)

    if \(c_-<\eta \), then \(h_0\) is decreasing, if \(c_->\eta \), then \(h_0\) is increasing.

In all cases \(\alpha \) is decreasing.

Remark 2.15

We have seen in the proof of Lemma 2.5 that \(w_{\epsilon }(x_0)\) increased monotonically, if some other slit height was decreased. Theorem 2.14, in particular case (iii) and case (iv) show that such a monotonicity is lacking for Akhiezer polynomials, which makes the situation essentially different to the Remez extremal problem.

3 Reduction to Akhiezer Polynomials (Proof of Theorem 1.3, Part 2)

The goal of this section is to finish the proof of Theorem 1.3. That is, if \(x_0\) is in an internal gap, then the extremal polynomial is an Akhiezer polynomial. Recall that in contrast to Section 2.2 now it is possible that there is an extension outside of \( [-1,1]\), moreover, this is a generic position. All possible types of extremizer were described in Corollary 2.2 and the discussion above the corollary. Thus, it remains to show that on \([-1,1]\) the extremal configuration in fact only has one gap, i.e., the extremal set is of the form \(E(\alpha ,\delta )\) for some \(\alpha \in (-1+\delta ,1-\delta )\).

First, we will show that E has at most two gaps on \([-1,1]\). Let \(T_n(z,E)\) denote the extremizer of (1.1) for the set E.

Lemma 3.1

Let \(E\subset [-1,1]\) and let \(x_0\) be in an internal gap of E. Then, there exists a two-gap set \({\tilde{E}}\), such that

$$\begin{aligned} T_n(x_0,E)=T_n(x_0,{{\tilde{E}}}),\quad |E|=|{{\tilde{E}}}|. \end{aligned}$$
(3.1)

Proof

We will only prove the case that there are no boundary gaps. Moreover, let us assume that E is already maximal, i.e., \(T_n^{-1}([-1,1],E)\cap [-1,1]=E\). Let us write

$$\begin{aligned} E=[-1,1]\setminus \bigcup _{j=1}^g(a_j,b_j), \end{aligned}$$

and let us denote the gap which contains \(x_0\) by (ab). Let \(\Pi \) be the comb related to E and let us assume that the slit corresponding to (ab) is placed at 0 and let c denote the critical point of \(T_n\) in (ab). Assume that there are two additional gaps \((a_1,b_1)\) and \((a_2,b_2)\) with slit heights \(h_k\) and critical points \(c_k\), \(k=1,2\). We will vary the slit heights \(h_k\) and h such that \(w_{\epsilon }(x_0)=x_0\) and \(|E_{\epsilon }|=|E|\). Therefore, we get

$$\begin{aligned} {\dot{w}}(z)=\frac{z^ 2-1}{c_1^ 2-1}\frac{\sigma _{1}}{c_{1}-z}+\frac{z^ 2-1}{c_2^ 2-1}\frac{\sigma _{2}}{c_{1}-z}+\frac{z^ 2-1}{c^ 2-1}\frac{\sigma }{c-z}. \end{aligned}$$
(3.2)

The constraints

$$\begin{aligned} \sum _{j=1}^{g}({\dot{w}}(b_j)-{\dot{w}}(a_j))=0, \quad {\dot{w}}(x_0)=0 \end{aligned}$$
(3.3)

will guarantee that (3.1) is satisfied. Our goal is to find \(\sigma _1>0\) and \(\sigma ,\sigma _2\), such that (3.3) is satisfied. Let us define

$$\begin{aligned} H(z)=\frac{(z^2-1)(z-x_0)}{(z-c_1)(z-c_2)(z-c)}. \end{aligned}$$

Due to the second constraint in (3.3), we have

$$\begin{aligned} {\dot{w}}(z)=f(z)H(z), \end{aligned}$$
(3.4)

where f(z) is linear. Thus \(f(z)=K(z-\xi )\) or \(f(z)=K\). We need to check that we can always find \(\xi \) such that the first constraint in (3.3) is satisfied. If \( \sum _{j=1}^{g}(H(b_j)-H(a_j))=0, \) we set \(f(z)=K\). If \( \sum _{j=1}^{g}(H(b_j)-H(a_j))\ne 0, \) we define

$$\begin{aligned} \xi =\frac{\sum _{j=1}^{g}(b_jH(b_j)-a_jH(a_j))}{\sum _{j=1}^{g}(H(b_j)-H(a_j))}. \end{aligned}$$

In any case, we define \({\dot{w}}\) by (3.4) and \(\sigma _{1}\), \(\sigma _2\), \(\sigma \) as the residues of this function at \(c_1, c_2, c\), respectively. Now we have to distinguish two cases. If \(\xi \ne c_1\), we can choose K so that \(\sigma _1>0\) and decrease \(h_1\). If at some point \(\xi =c_1\), we can choose K so that \(\sigma _{2}\) will be decreased. Note that in this case \(h_1\) remains unchanged. In particular, it won’t increase again. Hence, this procedure allows us to “erase” all but one additional gap. The case of boundary gaps works essentially in the same way, only using instead of the variations used above, variations as described in Lemma 2.12. \(\square \)

Ending Proof of Theorem 1.3

First, assume that the extremizer is in a generic position, that is, there is an extra interval \(I_n\subset {\mathbb {R}}\setminus [-1,1]\). We have \(T_n(z,E)\) with

$$\begin{aligned} E=[-1,1]\setminus \left( (a,b)\cup (a_1,b_1)\right) \end{aligned}$$
(3.5)

and \(|T_n(z,E)|\le 1\) for \(z\in E\cup I_n\). The corresponding comb has three slits, which heights we denote by \(h_{out}, h, h_1\). Our goal is to show that we can reduce the value \(h_1\). Varying all three values, we get that the corresponding infinitesimal variation as described by expression (3.2), with the parameters \(\sigma _{out},\sigma , \sigma _1\). Therefore, we still can satisfy the two constraints in (3.3), and choose one of the parameters positive. Since the direction of the variation of the heights \(h_{out}\) and h is not essential for us, we choose \(\sigma _1>0\), and therefore reduce the size of \(h_1\).

In a degenerated case we can use variations, which were described in Section 2.3. This is possible as long as either \(\theta (1)\) or \(\theta (-1)\) can be moved in both directions. Thus, it remains to consider the case that E is of the form (3.5), but the corresponding comb has only two non-trivial teeth of the heights h and \(h_1\). WLOG we assume that \(b<a_1\). We have

$$\begin{aligned} T_n(z,E)=\cos \theta (z), \quad \theta (-1)=\pi n_-,\quad \theta (1)=\pi n_+. \end{aligned}$$

We will apply the third variation, see Fig. 2, left. That is, we will reduce the value \(h_1\) and move the preimage of \(-1\) in the positive direction. According to (2.10), see also (2.11), we obtain

$$\begin{aligned} {\dot{w}}(z)=-\tau (z-1)+\frac{1-z^2}{1-c_1^2}\frac{\sigma _1}{c_1-z}, \quad \tau <0,\quad \sigma _1>0. \end{aligned}$$

Let us point out that with an arbitrary choice of the parameter \(\sigma _1>0\) and \(\tau <0\) we get an increasing function. Therefore,

$$\begin{aligned} {\dot{w}}(1)-{\dot{w}}(b_1)>0,\ {\dot{w}}(a_1)-{\dot{w}}(b)>0,\ {\dot{w}}(a)-{\dot{w}}(-1)>0. \end{aligned}$$

Thus, with a small variation \({\epsilon }\) of this kind we get

$$\begin{aligned} E_{\epsilon }=[w_{\epsilon }(-1),w_{\epsilon }(1)]\setminus \left( (w_{\epsilon }(a),w_{\epsilon }(b))\cup (w_{\epsilon }(a_1),w_{\epsilon }(b_1)\right) \end{aligned}$$

with

$$\begin{aligned} |E_{\epsilon }|=&(w_{\epsilon }(1)-w_{\epsilon }(b_1))+( w_{\epsilon }(a_1)-w_{\epsilon }(b))+ (w_{\epsilon }(a)-w_{\epsilon }(-1))\\ >&(1-b_1)+ (a_1-b)+(a-(-1))=|E|. \end{aligned}$$

On the other hand \({\dot{w}}(z)\) has a trivial zero \(z=1\) and a second one, which we denote by \(y_*\). Since

$$\begin{aligned} \tau +\frac{1+y_*}{1-c_1^2}\frac{\sigma _1}{c_1-y_*}=0, \end{aligned}$$

we get

$$\begin{aligned} y_*=y_*(\rho _1,\rho _2)=\rho _1(-1)+\rho _2 c_1, \end{aligned}$$

where

$$\begin{aligned} \rho _2 =\frac{-\tau }{-\tau +\frac{\sigma _1}{1-c^2_1}},\quad \rho _1=1-\rho _2,\quad \rho _{1,2}>0. \end{aligned}$$

Thus, with different values of \(\tau <0\) and \(\sigma _1>0\), we can get an arbitrary value \(y_*\in (-1,c_1)\).

Assume that \(x_0<c\). We choose \(\rho _1,\rho _2\) such that \(y_*>c\) (recall our assumption \(b<a_1\), therefore this is possible). Then, \({\dot{w}}(x_0)<0\). For a small \({\epsilon }\), we get \(w_{\epsilon }(x_0)<x_0\). By definition, \(T_n(w_{\epsilon }(x_0),E_{\epsilon })=T_n(x_0,E)\). Since in this range \(T_n(w_{\epsilon }(x_0),E_{\epsilon })\) is increasing, we get \(T_n(x_0,E_{\epsilon })>T_n(x_0,E)\), that is \(T_n(z,E)\) was not an extremal polynomial for the Andrievskii problem.

If \(x_0\in (c,b)\), we choose \(y_*<c\). Then, \({\dot{w}}(x_0)>0\). We repeat the same arguments, having in mind that in this range \(T_n(z,E)\) is decreasing. Thus, we arrive to the same contradiction \(T_n(x_0,E_{\epsilon })>T_n(x_0,E)\). \(\square \)

4 Asymptotics (Proof of Theorem 1.5)

The goal of this section is to prove Theorem 1.5 and to give a description of the upper envelope (1.5).

4.1 Totik–Widom bounds

We need an asymptotic result for Akhiezer polynomials. Recall that \(E(\alpha ,\delta )=[-1,1]\setminus (\alpha -\delta ,\alpha +\delta )\) and that \(A_{n,\alpha ,\delta }\) denotes the associated Akhiezer polynomial. Moreover, we denote \({\hat{E}}_n(\alpha ,\delta )=A_{n,\alpha ,\delta }^{-1}([-1,1])=E(\alpha ,\delta )\cup I_n\) and \(\hat{\Omega }_n={\overline{{\mathbb {C}}}}\setminus {\hat{E}}_n\). We have described the shape of the additional interval \(I_n\) in Theorem 2.1 and the discussion below. The following Lemma is known in a much more general context [11, Proposition 4.4.].

Lemma 4.1

Let \(\alpha ,\delta \) be fixed and \(A_{n,\alpha ,\delta }\) be the associated Akhiezer polynomial. Let \(x_n\in I_n\) denote the single zero of \(A_{n,\alpha ,\delta }\) outside of \([-1,1]\). For any \(y\in (\alpha -\delta ,\alpha +\delta )\)

$$\begin{aligned} \log 2|A_{n,\alpha ,\delta }(y)|=nG(y,\infty ,\hat{\Omega }_n)+o(1). \end{aligned}$$
(4.1)

If we pass to a subsequence such that \(\lim _{k\rightarrow \infty }x_{n_k}=x_\infty \in ({\mathbb {R}}\cup \{\infty \})\setminus (-1,1)\), then

$$\begin{aligned} \lim \limits _{k\rightarrow \infty }n_k(G(y,\infty ;\Omega (\alpha ,\delta ))-G(y,\infty ;\hat{\Omega }_{n_k}))=G(y,x_\infty ,\Omega (\alpha ,\delta )), \end{aligned}$$
(4.2)

where \(\Omega (\alpha ,\delta )={\overline{{\mathbb {C}}}}\setminus E(\alpha ,\delta )\).

Proof of Theorem 1.5

We start with the case that the sup in (1.5) is attained at some internal point \(\alpha _0\in (-1+\delta ,0]\). Let \(T_{n,\delta ,x_0}\) be the extremizer of (1.2) and set \({\hat{E}}_n=T_{n,\delta ,x_0}^{-1}([-1,1])\) and \(\hat{\Omega }_n={\overline{{\mathbb {C}}}}\setminus {\hat{E}}_n\). Due to Theorem 1.3, \({\hat{E}}_n\) is either \([-1+2\delta ,1]\) or \({\hat{E}}_n=E(\alpha _n,\delta )\cup I_n\) for some \(\alpha _n\) and some additional interval outside of \([-1,1]\). In the following, we will denote \(G(x,\infty ,\Omega )=G(x,\Omega )\) and we note that by definition \(G(x,\Omega (\alpha ,\delta ))=G_{\alpha ,\delta }(x)\). Put \(E_n={\hat{E}}_n\cap [-1,1]\) and \(\Omega _n={\overline{{\mathbb {C}}}}\setminus E_n\). Due to extremality of \(\alpha _0\), we have

$$\begin{aligned} G(x_0,\Omega (\alpha _0,\delta ))\ge G(x_0,\Omega _n). \end{aligned}$$

Since \({\hat{\Omega }}_n\subset \Omega _n\), we get

$$\begin{aligned} G(x_0,\Omega _n)\ge G(x_0,\hat{\Omega }_n). \end{aligned}$$

Using (4.1) or (1.4) and the extremality property of \(T_{n,\delta ,x_0}(x_0)\), we get

$$\begin{aligned} nG(x_0,\hat{\Omega }_n)&=\log 2|T_{n,\delta ,x_0}(x_0)|+o(1)\\&\ge \log 2|T_{n}(x_0,E(\alpha _0,\delta ))|+o(1). \end{aligned}$$

On the other hand, equation (4.2) yields

$$\begin{aligned} \log 2|T_n(x_0,E(\alpha _0,\delta ))|\ge nG(x_0,\Omega (\alpha _0,\delta ))-C+o(1), \end{aligned}$$

where

$$\begin{aligned} C=\sup _{x\in ({\mathbb {R}}\cup \{\infty \})} G(x,x_0,\Omega (\alpha _0,\delta )). \end{aligned}$$
(4.3)

By definition, \(|T_{n,\delta ,x_0}(x_0)|=L_n(x_0)\) and therefore combining all inequalities finishes the proof of (1.8). The proof for the boundary case is essentially the same. Only in the last step, due to representation (1.4), there is no extra constant C (due to the fact that there is no extension of the domain) and we get (1.7). \(\square \)

4.2 The Asymptotic Diagram

In this section, we introduce an asymptotic diagram, which provides a description of the upper envelope \(\Phi _\delta (x)\). First of all, we prove Lemma 1.4.

Proof of Lemma 1.4

Recall the explicit representation of Green functions for two intervals as elliptic integrals. For \(a=\alpha -\delta , b=\alpha +\delta \), we have

$$\begin{aligned} G_{\alpha ,\delta }(x)=\int _a^x\frac{(c-\xi )\mathrm{d}\xi }{\sqrt{(\xi ^2-1)(\xi -a)(\xi -b)}}, \end{aligned}$$
(4.4)

where

$$\begin{aligned} c=c(\alpha )=\frac{\int _a^b\frac{\xi \mathrm{d}\xi }{\sqrt{(\xi ^2-1)(\xi -a)(\xi -b)}}}{\int _a^b\frac{\mathrm{d}\xi }{\sqrt{(\xi ^2-1)(\xi -a)(\xi -b)}}}. \end{aligned}$$
(4.5)

If for fixed \(x_0\) the sup is attained at an internal point, then clearly

$$\begin{aligned} {\partial }_\alpha G_{\alpha ,\delta }(x)=0 \end{aligned}$$
(4.6)

holds. Thus, it remains to show that (4.6) has a unique solution \(x_0(\alpha )\). Due to (4.4), we have

$$\begin{aligned} {\partial }_\alpha \log {\partial }_xG_{\alpha ,\delta }(x)=\frac{\dot{c}}{c-x}-\frac{1}{2}\frac{\dot{a}}{a-x}-\frac{1}{2}\frac{\dot{b}}{b-x}. \end{aligned}$$

Since \(\dot{a}=\dot{b}=1\), we get

$$\begin{aligned} {\partial }_x {\partial }_\alpha G_{\alpha ,\delta }(x)=\left\{ \dot{c}+\frac{(c-x)(x-\alpha )}{(x-a)(x-b)}\right\} \frac{1}{\sqrt{(x^2-1)(x-a)(x-b)}}. \end{aligned}$$
(4.7)

Note that

$$\begin{aligned} \inf _{x\in (a,b)}\frac{(c-x)(x-\alpha )}{(x-a)(x-b)}>-1. \end{aligned}$$

Since the distance \(|c(\alpha )-b(\alpha )|\) monotonically increases with \(|\alpha |\), we have \(\dot{c}(\alpha )>1\). Therefore, we get \({\partial }_x{\partial }_\alpha G_{\alpha ,\delta }(x)>0\) in (4.7). That is, \({\partial }_\alpha G_{\alpha ,\delta }(x)\) is increasing for \(x\in (a,b)\). Moreover, \(\lim _{x\rightarrow a} {\partial }_\alpha G_{\alpha ,\delta }(x)=-\infty \) and \(\lim _{x\rightarrow b} {\partial }_\alpha G_{\alpha ,\delta }(x)=+\infty \) and we obtain that a zero \(x_0(\alpha )\) of the function \({\partial }_\alpha G_{\alpha ,\delta }(x)\) in (ab) exists and is unique. \(\square \)

Thus, the limiting behavior of \(L_{n,\delta }(x_0)\), \(n\rightarrow \infty \), can be distinguished by a diagram with two curves, which we describe in the following proposition, see also Example 4.5.

Proposition 4.2

In the range \(x\in [-1,0]\), \(\Phi _\delta (x)\) represents the upper envelope of the following two curves. The first one is given explicitly

$$\begin{aligned} y=G_\delta (x)=\log \left( \frac{\delta -x}{1-\delta }+\sqrt{\left( \frac{\delta -x}{1-\delta }\right) ^2-1}\right) ,\quad x\in [-1,-1+2\delta ], \end{aligned}$$
(4.8)

and the second one is given in parametric form

$$\begin{aligned} x=x_0(\alpha ),\quad y=G_{\alpha ,\delta }(x_0(\alpha )), \quad \alpha \in (-1+\delta ,0]. \end{aligned}$$
(4.9)

Moreover, the end points of the last curve are given explicitly by

$$\begin{aligned} \left( 0,\frac{1}{2}\log \frac{1+\delta }{1-\delta }\right) \ \text {for}\ \alpha =0,\quad (-1+2\delta ,0)\ \text {for}\ \alpha \rightarrow -1+\delta . \end{aligned}$$
(4.10)

Proof

According to Lemma 1.4, if \(\Phi _\delta (x)\) is assumed at the end point \(\alpha \rightarrow -1+\delta \), then it is the Green function in the complement to the interval \([-1+2\delta ,1]\), which has a well-known representation (4.8). Alternatively, \(x=x_0(\alpha )\) and \(\Phi _\delta (x)=G_{\alpha ,\delta }(x_0(\alpha ))\) for a certain \(\alpha \) in the range, what is (4.9).

Further, for \(\alpha =0\), \(G_{0,\delta }(x)\) is the Green function related to two symmetric intervals. Due to the symmetry \(x_0(0)=0\) and \(G_{0,\delta }(x)\) can be reduced to the Green function of a single interval \([\delta ^2,1]\). We get

$$\begin{aligned} G_{0,\delta }(0)=\frac{1}{2}\log \frac{1+\delta }{1-\delta }. \end{aligned}$$

Thus, it remains to prove the last statement of the proposition. Our proof is based on expressing \({\partial }_\alpha G_{\alpha ,\delta }\) in terms of elliptic integrals and manipulating those. It will be convenient to make a standard substitution in (4.4). Let \(\xi (\psi ,\alpha )=\alpha -\delta \cos \psi \), then

$$\begin{aligned} G_{\alpha ,\delta }(x)=\int _{\phi (x,\alpha )}^\pi \frac{\xi (\psi ,\alpha )-c(\alpha )}{\sqrt{1-\xi (\psi ,\alpha )^2}}\mathrm{d}\psi ,\quad \phi (x,\alpha )= \arccos \frac{\alpha -x}{\delta }. \end{aligned}$$
(4.11)

Differentiating (4.11), we get

$$\begin{aligned} {\partial }_\alpha G_{\alpha ,\delta }(x)=I_1(x,\alpha )+I_2(x,\alpha )-\dot{c}(\alpha ) I_3(x,\alpha ) \end{aligned}$$

where

$$\begin{aligned} I_1(x,\alpha )=&\int _{\phi (x,\alpha )}^\pi \frac{1- c(\alpha )\xi (\psi ,\alpha )}{(1-\xi (\psi ,\alpha )^2)\sqrt{1-\xi (\psi ,\alpha )^2}}\mathrm{d}\psi \\ I_2(x,\alpha )=&\frac{x-c(\alpha )}{\sqrt{(1-x^2)(\delta ^2-(x-\alpha )^2)}}, \quad I_3(x,\alpha )=\int _{\phi (x,\alpha )}^\pi \frac{\mathrm{d}\psi }{\sqrt{1-\xi (\psi ,\alpha )^2}}. \end{aligned}$$

Let \({\epsilon }={\epsilon }(\alpha )\) be such that \({\epsilon }\rightarrow 0\) as \(\alpha \rightarrow -1+\delta \), to be chosen later on, and let

$$\begin{aligned} x(\alpha )=\alpha -\delta \cos (\pi -{\epsilon }). \end{aligned}$$
(4.12)

Direct estimations show that

$$\begin{aligned} I_1(x(\alpha ),\alpha )=o(1),\quad I_2(x(\alpha ),\alpha )=\frac{a_1(\alpha )}{{\epsilon }}+o(1), \end{aligned}$$

where \(\lim _{\alpha \rightarrow -1+\delta }a_1(\alpha )>0\). The integral \(I_3(x(\alpha ),\alpha )\) also tends to zero, but we will need a more precise decomposition

$$\begin{aligned} I_3(x(\alpha ),\alpha )=b_1(\alpha ){\epsilon }+\frac{b_2(\alpha )}{3}{\epsilon }^3+O({\epsilon }^5) \end{aligned}$$
(4.13)

with \(\lim _{\alpha \rightarrow -1+\delta }b_1(\alpha )>0\). Indeed, we have

$$\begin{aligned} I_3= \int _{\phi (x(a),\alpha )}^\pi&\frac{\mathrm{d}\psi }{\sqrt{1-\xi (\psi ,\alpha )^2}}=\int _{0}^{\epsilon }\frac{\mathrm{d}\psi }{\sqrt{1-(\alpha +\delta \cos \psi )^2}}. \end{aligned}$$

Therefore, for \(\psi \) sufficiently small we can use the expansion

$$\begin{aligned} \frac{1}{\sqrt{1-(\alpha +\delta \cos \psi )^2}}=b_1(\alpha )+b_2(\alpha )\psi ^2+O(\psi ^4), \end{aligned}$$

where \(O(\psi ^4)\) is uniform in \(\alpha \). We get (4.13). In the same time,

$$\begin{aligned} \lim _{\alpha \rightarrow -1+\delta }b_1(\alpha )=\frac{1}{\sqrt{1-(-1+2\delta )^2}}>0. \end{aligned}$$

Collecting all three terms, we obtain

$$\begin{aligned} {\partial }_\alpha G_{\alpha ,\delta }(x(\alpha ))&=\frac{a_1(\alpha )}{{\epsilon }}-\dot{c}(\alpha )\left( b_1(\alpha ){\epsilon }+\frac{b_2(\alpha )}{3}{\epsilon }^3\right) +o(1) \nonumber \\&=\varkappa _1(\alpha )\left( \frac{\varkappa (\alpha )}{{\epsilon }}-\dot{c}(\alpha )\left( \frac{{\epsilon }}{\varkappa (\alpha )}+\frac{\varkappa _2(\alpha )}{3}{\epsilon }^3\right) \right) +o(1), \end{aligned}$$
(4.14)

where \(\varkappa _1(\alpha )=\sqrt{a_1(\alpha )b_1(\alpha )}\), \(\varkappa (\alpha )=\sqrt{a_1(\alpha )/b_1(\alpha )}\) and \(\varkappa _2(\alpha )\) is chosen appropriately. Recall that \(\varkappa _1(\alpha )\) and \(\varkappa (\alpha )\) have positive and finite limits as \(\alpha \rightarrow -1+\delta \).

Similar manipulations with elliptic integrals show that \(\dot{c}(\alpha )\rightarrow +\infty \) as \(\alpha \rightarrow -1+\delta \). In fact, the rate of this divergence is (see Appendix)

$$\begin{aligned} \dot{c}(\alpha )\simeq (\varepsilon \log \varepsilon )^{-2}, \quad \alpha =-1+\delta (1+\frac{1}{2} \varepsilon ^2),\ \varepsilon \rightarrow 0, \end{aligned}$$

but such accuracy is not needed for our purpose.

Having these estimations, we can find a suitable interval \([x_-(\alpha ),x_+(\alpha )]\), with the limit

$$\begin{aligned} \lim _{\alpha \rightarrow -1+\delta }x_\pm (\alpha )= -1+2\delta , \end{aligned}$$
(4.15)

such that \({\partial }_\alpha G_{\alpha ,\delta }(x)\) changes its sign in it and therefore this interval contains the unique solution of the equation \({\partial }_\alpha G_{\alpha ,\delta }(x)=0\).

Define

$$\begin{aligned} {\epsilon }_\pm (\alpha )=\frac{\varkappa (\alpha )}{\sqrt{\dot{c}(\alpha )}\pm 1}. \end{aligned}$$

Since \(\dot{c}(\alpha ){\epsilon }_\pm (\alpha )^3=o(1)\), (4.14) gets the form

$$\begin{aligned} \varkappa _1(\alpha )^{-1}{\partial }_{\alpha }G(x_\pm (\alpha ),\alpha )&=\sqrt{\dot{c}(\alpha )}\pm 1-\frac{\dot{c}(\alpha )}{\sqrt{\dot{c}(\alpha )}\pm 1}+o(1)\nonumber \\&=\frac{\pm 2\sqrt{\dot{c}(\alpha )}+1}{\sqrt{\dot{c}(\alpha )}\pm 1}+o(1), \end{aligned}$$
(4.16)

where \(x_\pm (\alpha )\) is defined by (4.12) for \({\epsilon }={\epsilon }_\pm (\alpha )\). For \(\dot{c}(\alpha )\) sufficiently large, we obtain in (4.16) both positive and negative values, and simultaneously we have (4.15). Consequently, \(x_0(\alpha )\rightarrow -1+2\delta \). \(\square \)

Corollary 4.3

Let \(\delta _*\) be a unique solution of the equation

$$\begin{aligned} \delta _*^2=\frac{1-\delta _*}{1+\delta _*},\quad \delta _*\in (0,1), \end{aligned}$$

numerically \(\delta _*=0.543689...\). Then, for \(\delta <\delta _*\) \(\Phi _{\delta }(x)\) does not coincide identically with \(G_\delta (x)\) (in its range \(x\in (-1,0]\)).

On the other hand, for an arbitrary \(\delta >0\) there exists \(x_*(\delta )>-1\) such that \(\Phi _{\delta }(x)\) and \(G_\delta (x)\) coincide in \([-1,x_*(\delta ))\).

Proof

The first claim follows by a direct comparison of \(G_\delta (0)\) and \(G_{0,\delta }(0)\), see (4.8) and (4.10).

For a fixed \(\delta \), we define

$$\begin{aligned} x_*(\delta )=\inf _{\alpha \in (-1+\delta ,0]} x_0(\alpha ) \end{aligned}$$
(4.17)

By (4.10) and continuity, \(x_*(\delta )>-1\). Thus, the curve (4.9) does not intersect the range \([-1,x_*(\delta ))\). \(\square \)

Remark 4.4

We do not claim here that \([-1,x_*(\delta ))\) with \(x_*(\delta )\) given by (4.17) is the biggest possible interval on which \(\Phi _\delta (x)=G_\delta (x)\), see Example 4.5 for details.

Fig. 3
figure 3

The asymptotic diagram for \(\delta =0.4\)

Example 4.5

A numerical example of the asymptotic diagram for \(\delta =0.4\) is given in Fig. 3 (diagrams for other values of \(\delta <0.5\) look pretty the same). Let \(x_s(\delta )\) be the switching point between two (Remez and Akhiezer) extremal configurations, i.e.,

$$\begin{aligned} \Phi _\delta (x_s)=G_\delta (x_s(\delta ))=G_{\alpha _s,\delta }(x_0(\alpha _s)),\quad x_0(\alpha _s)=x_s(\delta ). \end{aligned}$$

Recall that \(x_*(\delta )\) was defined in (4.17). On the diagram, we can observe the following four regions: \((-1, x_*(\delta )), (x_*(\delta ),x_s(\delta )), (x_s(\delta ),-1+2\delta )\) and \((-1+2\delta ,0)\). Note that we discuss the case \(-1+2\delta <0\), i.e., \(\delta<0.5<\delta _*\).

  1. a)

    \(x\in (-1+2\delta ,0)\). In this case \(x\in (\alpha -\delta ,\alpha +\delta )\) implies that such an interval is a subset of \((-1,1)\) even in the leftmost position \(\alpha =-\delta +x\). Therefore, the function \(G_{\alpha ,\delta }(x)\) for a fixed x and \(\alpha \in (x-\delta ,x+\delta )\) attains its maximum at some internal point and we get the case \({\partial }_\alpha G_{\alpha ,\delta }(x)=0\).

  2. b)

    \(x\in (x_s(\delta ),-1+2\delta )\). As soon as \(x<-1+2\delta \) the left boundary for a possible value of \(\alpha \) is given by \(\alpha -\delta =-1\). Respectively, the supremum of \(G_{\alpha ,\delta }(x)\) for a fixed x can be attained either at an internal point \(\alpha \in (-1+\delta ,x+\delta )\) or as the limit at the left end point. In this range, it is still attained at an internal point. Note that besides the local maximum the function gets a local minimum (the second branch of the curve (4.9) with the same coordinate \(x_0(\alpha )=x\)).

  3. c)

    \(x\in (x_*(\delta ),x_s(\delta ))\). For such x, the function \(G_{\alpha ,\delta }(x)\) has still its local maximum and minimum, but the biggest value is attained at the boundary point \(\alpha =-1+\delta \), i.e., \(\Phi _\delta (x)=G_\delta (x)\).

  4. d)

    \(x\in (-1,x_*(\delta ))\). At \(x=x_*(\delta )\) the points of local maximum and minimum of the function \(G_{\alpha ,\delta }(x)\) collide, that is, in fact, they produce an inflection point. The function \(G_{\alpha ,\delta }(x)\) become monotonic in this region. Its supremum is the limit at the boundary point \(\alpha =-1+\delta \), see the second claim in Corollary 4.3.