Convex hulls of monomial curves, and a sparse positivstellensatz

Consider the closed convex hull $K$ of a monomial curve given parametrically as $(t^{m_1},\ldots,t^{m_n})$, with the parameter $t$ varying in an interval $I$. We show, using constructive arguments, that $K$ admits a lifted semidefinite description by $\mathcal{O}(d)$ linear matrix inequalities (LMIs), each of size $\left\lfloor \frac{n}{2} \right\rfloor +1$, where $d= \max \{m_1,\ldots,m_n\}$ is the degree of the curve. On the dual side, we show that if a univariate polynomial $p(t)$ of degree $d$ with at most $2k+1$ monomials is non-negative on $\mathbb{R}_+$, then $p$ admits a representation $p = t^0 \sigma_0 + \cdots + t^{d-k} \sigma_{d-k}$, where the polynomials $\sigma_0,\ldots,\sigma_{d-k}$ are sums of squares and $\operatorname{deg} (\sigma_i) \le 2k$. The latter is a univariate positivstellensatz for sparse polynomials, with non-negativity of $p$ being certified by sos polynomials whose degree only depends on the sparsity of $p$. Our results fit into the general attempt of formulating polynomial optimization problems as semidefinite problems with LMIs of small size. Such small-size descriptions are much more tractable from a computational viewpoint.


Introduction
Polynomial optimization studies the task of minimizing an n-variate polynomial p ∈ R[x 1 , . . ., x n ] over a semialgebraic set S ⊆ R n .The convexification approach to this problem consists of picking finite sets A ⊇ supp(p) of Z n + , considering the monomials x α = x α1 1 • • • x αn n with α ∈ A and finding exact or approximate descriptions of the closed convex hulls K = conv {(x α ) α∈A : x ∈ S} within some optimization paradigm.The basic idea is that optimizing a polynomial p = c 0 + α∈A c α x α ∈ R[x 1 , . . ., x n ] means to optimize the affine-linear function λ : (y α ) α∈A → c 0 + α∈A c α y α over K. Usually, descriptions of the sets K arise from positivstellensätze from real algebra, since non-negativity of λ on K corresponds to non-negativity of p on S. Most positivstellensätze in real algebra employ sum-of-squares (sos) certificates.This fact establishes a connection from real algebra to semidefinite optimization, since the cone Σ n,2d of n-variate sos polynomials of degree ≤ 2d is a linear image of the cone of symmetric positive semidefinite matrices of size n+d n .As suggested by Lasserre in his seminal work [12] and following this line of ideas, Putinar's positivstellensatz [16] leads to a hierarchy of outer approximations of K in terms of semidefinite constraints.Although Lasserre's approach provides a universal template for converting polynomial problems to semidefinite ones, further adjustment is usually needed to make it computationally tractable.Despite the fact that semidefinite programs are solvable in polynomial time under Date: March 8, 2023.mild assumptions within a given error tolerance [3,15], solving a semidefinite program can quickly become extremely challenging in practice since the size of the semidefinite constraints is a critical parameter.We view a semidefinite problem as the problem of optimization of a linear function subject to finitely many linear matrix inequalities (LMIs).An LMI of size d is the condition M (y) 0 that imposes semidefiniteness of a symmetric d × d matrix M (y) whose entries are affine-linear functions in the variables y = (y 1 , . . ., y N ).To get a first impression of how the workload is increased by large size of the LMIs, consider the approximation of a semidefinite problem inf{L(y) : M 1 (y) 0, . . ., M ℓ (y) 0} by the convex problem L(y) − ǫ ℓ i=1 log det M i (y) that uses the logarithmic barrier and a small parameter ǫ > 0. Solving the convex problem with the gradient descent method would involve computating the gradients of the barriers, which requires to invert the matrices M i (y) [6, 4.3.1].But inverting M i (y) of a large size is known to be expensive.So, while in theoretical considerations it is customary to model ℓ LMIs as a single LMI of size ℓd, using a block-diagonal matrix, this reduction conceals the aspect of efficiency, since a general LMI of size ℓd has a much higher computational cost than ℓ LMIs of size d.
To address this issue, the following terminology was introduced in By studying sxdeg(C) one keeps track of the size of the LMIs needed to optimize linear functions over C, disregarding their number.For computations, both size and number of the LMIs play a role, but size is more critical as a parameter.Clearly sxdeg(C) = 1 if and only if C is a polyhedron, and it is easy to see that sxdeg(C) ≤ 2 if and only if C is second-order cone representable [9].In [4] it was shown that the cone Σ n,2d of n-variate sos polynomials of degree ≤ 2d has sxdeg = n+d n , matching the size of the LMIs in Lasserre's approach.The rapid growth of n+d n in n and d explains why the computational time needed to solve Lasserre's relaxations is extremely sensitive to the choice of n and d.This issue with the size of the LMIs in was detected by a number of researchers.Some suggestions for how to cope with it in practice were made in [1,5,21,22].The common general idea in [1,5,21,22] is to deliberately choose a cone C, which has small sxdeg(C) by its construction, to serve as a tractable outer approximation of K. Or, from the dual viewpoint, to choose a cone P of non-negative polynomials of more specific structure than in Putinar's positivstellensatz, such that P has small sxdeg(P ).While these approaches seem to help in practice, they are purely heuristic, since positivstellensätze are still missing that would guarantee that positivitiy can indeed always be certified in the intended, computationally less expensive, manner.
In this paper, we make a first step in filling this gap by studying the size of semidefinite representations for sparse polynomial optimization problems in one variable.Consider a set A = {m 1 , . . ., m n } of positive integers m 1 > • • • > m n > 0 and a non-degenerate interval I ⊆ R. For optimizing the univariate polynomial p = a∈A c a t a ∈ R[t] on the interval I, we are interested in finding a lifted semidefinite representation of K = conv{(t m1 , . . ., t mn ) : t ∈ I} ⊆ R n .Since the curve (t m1 , . . ., t mn ) is a projection of the rational normal curve (t i ) 1≤i≤m1 of degree m 1 , Lasserre's approach gives a description of size m1 2 + 1.When the number n = |A| of monomials is small compared to the degree m 1 , it is desirable to find an alternative description of smaller size.Our main result (Theorem 3.19) shows that the semidefinite extension degree of such K is at most n 2 + 1, which is the best possible bound.Consequently, the size of the descripttion depends only on the number of monomials, and not on the degree.The description is completely explicit and follows from a sparse positivstellensatz (Theorem 3.9).The latter characterizes non-negativity of a degree d polynomial p with at most 2k + 1 terms on an interval I ⊆ R + , by using the cone , which is of semidefinite extension degree k + 1.The key technical ingredient in our proof is Jacobi's bialternant formula for Schur polynomials from the theory of symmetric polynomials.
Our result about non-negativity of univariate sparse polynomials should help to understand the impact of sparsity-based approaches to optimization problems of arbitrary dimension.Univariate problems may, on the one hand, demonstrate phenomena that occur in every dimension.On the other, whenever some limitation can be identified in the one-dimensional case, it will also be present in a certain form in an n-dimensional setting as well.Furthermore, our main result can serve as a starting point for further studies in several directions.It would obviously be interesting to obtain similar results in the multivariate case.While our K is given as the closed convex hull of a monomial curve, one may ask for bounds on the semidefinite extension degree of closed convex hulls K of arbitrary semialgebraic curves S ⊆ R n .For n = 2, it was proved in [17] that every closed convex semialgebraic set K in the plane is second-order cone representable, i.e. has sxdeg(K) ≤ 2. It can be shown that the bound sxdeg(K) ≤ n 2 + 1 holds for the closed convex hull of an arbitrary semialgebraic curve S ⊆ R n , but the proof is much more involved [18].
The paper is organized as follows.After a few preliminaries in Section 2, the main result is obtained in Section 3. In Section 4 we introduce the cones of sums of copositive fewnomials, which form a sparse counterpart of the cones of nonnegative polynomials.We show that in the one-dimensional case these cones admit semidefinite descriptions with LMIs of small size, and explain that such descriptions lead to variations of Lasserre's relaxations that are based on LMIs of small size.

Preliminaries
For univariate polynomials we mostly use t to denote the variable, and we write By S m we denote the space of real symmetric m × m matrices, and S m + is the cone of positive semidefinite matrices in S m .The linear span of a subset M of a vector space is denoted lin(M ).

Let
x, y ≥ 0}, where x, y denotes the standard inner product on R n .The bi-dual cone in an arbitrary real vector space V , the dual cone is C * := {y ∈ V ∨ : ∀x ∈ C y(x) ≥ 0}, where V ∨ is the dual vector space of V .
A face of C is a convex cone F with F ⊆ C such that x, y ∈ C and x + y ∈ F imply x, y ∈ F .For any x ∈ C there is a unique inclusion-minimal face F of C with x ∈ F , called the supporting face of x in C. One-dimensional faces of convex cones are called extreme rays.It is well-known that a finite-dimensional closed and pointed convex cone is the Minkowski sum of its extreme rays.

2.3.
We briefly explain the conic duality behind the approaches in polynomial optimization by providing a generic version of the discussion in [13] Ch. 10.We are given a subset S of R n and polynomials p, q in a finite-dimensional vector subspace V of R[x 1 , . . ., x n ] such that q > 0 on S. A common choice for q is the constant q = 1.The problem of minimizing the quotient p/q over S can be relaxed (i.e., lower bounded) by making use of a closed convex cone C ⊆ V that satisfies g ≥ 0 on S for every g ∈ C. One has where the supremum is a conic optimization problem dual to the problem In the following proposition, the mentioned duality is phrased without any reference to polynomial by identifying V and its dual space V ∨ with R N , where N := dim(V ).
Proposition 2.4.Let N be a positive integer and C ⊆ R N be a closed and pointed convex cone.Then for every p ∈ C and q ∈ C {0} one has This is a special case of the duality of conic optimization problems, cf. the discussion in [7] 2.1.4.We omit the proof which is quite straightforward.

Convex hulls of monomial curves
Let I ⊆ R be a non-degenerate closed interval, let V ⊆ R[t] be a linear subspace of finite dimension, and let P = {f ∈ V : f ≥ 0 on I}, a closed and pointed convex cone in V .We start by showing that every face of the cone P is described by suitable vanishing conditions at points of I, or at infinity when For extreme rays of P , this implies: Corollary 3.2.Assume that I is compact, let f ∈ V span an extreme ray of P .Then f has at least dim(V ) − 1 roots in I, counting with multiplicity.
Proof.By assumption we have F f = R + f , so W f = Rf by Proposition 3.1.Since W f consists of the elements in V with at least the same roots in I as f , there have to be at least codim(W f ) = dim(V ) − 1 many roots of f in I, counting with multiplicities.
In the non-compact case, the result reads as follows: and similarly in the case of a left half-line.By passing to a finite subcovering of I we find a constant c as required.
By the preceding remark we find a linear basis g 1 , . . ., g r of U f with the property that |g i | ≤ f on I for i = 1, . . ., r.It follows that f − g i ∈ F f (i = 1, . . ., r), and so We continue to assume that V ⊆ R[t] is a linear subspace of finite dimension dim(V ) = n + 1. Assume we are given n vanishing conditions for elements of V at points on the real line, where we allow conditions of higher order vanishing.If these conditions are independent on V , the following lemma gives an explicit formula for the essentially unique solution of these conditions in V : Lemma 3.4.Let p 0 , . . ., p n be a linear basis of V .Let ξ = (ξ 1 , . . ., ξ r ) be a tuple of pairwise different real numbers, and let b 1 , . . ., b r ≥ 1 be integers with r i=1 b i = n.The following conditions are equivalent: (i) The matrix of size (n + 1) × (n + 1) and with entries in R[t] has non-zero determinant; (ii) the subspace {f ∈ V : ord ξi (f ) ≥ b i for i = 1, . . ., r} of V has dimension one.When (i) and (ii) hold, the polynomial det A(t; ξ) is the unique (up to scaling) element of V that vanishes in ξ i of order at least b i , for i = 1, . . ., r.
Proof.Note that the dimension in (ii) is always ≥ 1.The matrix B formed by the lower n rows of (3) is the matrix of the linear map φ : V → R n , p → p(ξ 1 ), . . ., p (b1−1) (ξ 1 ), . . ., p(ξ r ), . . ., p (br −1) (ξ r ) with respect to the basis p 0 , . . ., p n of V .Subspace (ii) is just the kernel of φ.The determinant of ( 3) is an element of V , and is non-zero if and only if B has a non-vanishing n × n-minor.This is equivalent to φ being surjective, and hence also to (ii).The last assertion is clear since det A(t; ξ) is an element of V that has a zero of multiplicity ≥ b i at ξ i , for i = 1, . . ., r.

3.5.
Next we recall some background on Schur polynomials.Let n ∈ N be a fixed integer and let x = (x 0 , . . ., x n ) be a tuple of indeterminates.We consider partitions µ = (m 0 , . . ., m n ) into n+1 pieces, with m 0 ≥ • • • ≥ m n ≥ 0. A particular partition is δ = (n, n − 1, . . ., 1, 0).Given a partition µ as above, the determinant is identically zero unless the m i are pairwise distinct.In this latter case, The co-factor is the Schur polynomial s λ (x) of the partition λ.In other words, the bialternant formula holds.Depending on how Schur polynomials are introduced, identity (4) is either the definition of s λ (x) or a theorem, see [19] Theorem 7.15.1.The Schur polynomial s λ (x) is symmetric as a polynomial in the variables x i , homogeneous of degree |λ|, and of degree λ 0 with respect to each variable x i .Schur polynomials have the remarkable property that all their coefficients are nonnegative integers.In fact there exists a combinatorial description of the coefficients, see Section 7.10 in [19].In our context, the integrality of the coefficients plays no role, but the non-negativity is a crucial property.
3.6.We use Schur polynomials to deduce a product formula for the determinant in Lemma 3.4, in the case where the p i are monomials.Let µ = (m 0 , . . ., m n ) be a partition into different parts, i.e. with m 0 > • • • > m n ≥ 0. Write p i (t) = t mi for i = 0, . . ., n, and denote the j-th derivative of p i (t) by p (j) i (t) := d j dt j p(t) (j ≥ 0).Moreover we introduce the tuples p(t) = p 0 (t), . . ., p n (t) and p (j) (t) = p and let again λ = µ − δ.
And so on.After b 0 many steps, the rows have become and at that point we have thrown in 0 2 many minus signs.We can now repeat this procedure.The next step consists in doing b 1 many steps on the variables x b0 , . . ., x b0+b1−1 .And so on.Finally, relabel the variables that have survived as y 0 , . . ., y r .
3.8.After these preparations we come to the main result of our paper.Let µ = (m 0 , . . ., m n ) with be the ring of polynomials in the variable t.We consider the subspace V = V µ of R[t] that is spanned by the monomials t m0 , . . ., t mn .Let I ⊆ R be an interval, and let Note that P is a closed convex cone in V .The key result is the following nichtnegativstellensatz: Then every f ∈ P can be written as a finite sum where g i , p i , h j , q j are polynomials with deg(g i ), deg(h j ) ≤ ⌊ n 2 ⌋ = k − 1, such that the degree of every summand is ≤ m 0 , and such that the coefficients of p i , q j are all non-negative.
Conversely, when f has a representation as in the theorem, it is obvious that f ≥ 0 on I.

3.10.
To start the proof of Theorem 3.9 (in either case (a) or (b)), let Q be the set of all f ∈ V that have a representation (6).Then Q is a convex cone, and Q ⊆ P .Let E ⊆ P be the set of all polynomials f that generate an extreme ray of P .Every element of P is a sum of finitely many elements of E, since the cone P is closed and pointed.To prove Q = P , it therefore suffices to show f ∈ Q for any given f ∈ E.

3.11.
Let us make two more reduction steps.Assume that f ∈ Q has been shown (for all monomial subspaces) whenever f ∈ E satisfies f (0) > 0. Let f ∈ E with f (0) = 0. Writing f = t w f where w = ord 0 (f ), there is an index 0 ≤ s ≤ n with m s = w.The subspace Ṽ = lin(t m0−w , . . ., t ms−1−w , 1) of R[t] has dimension s + 1 ≤ n + 1 and contains f .Moreover f ≥ 0 on I and f (0) > 0. To fix ideas, assume we are in case (a), so I = [0, 1].By assumption, the theorem holds for f and the subspace Ṽ .This means that there is an identity where deg(g i ), deg(h j ) ≤ ⌊ s 2 ⌋ and the p i , q j have non-negative coefficients, such that every summand in (7) has degree ≤ m 0 − w.Multiplying the identity with t w , we get an identity for f as claimed in Theorem 3.9(b).In case (b), the argument is exactly the same.
3.12.So we need to establish an identity (6) for every f ∈ E with f (0) > 0. For this, we clearly may discard all monomials of degree greater than d = deg(f ).In other words, we may assume that deg(f ) = m 0 .This reduction step plays a role only in case (b).(In fact, deg(f ) = m 0 is automatic if I = [0, 1], as will be seen from the proof below.)3.13.From Descartes' rule of signs (e.g.[11] Cor. 1.10.3) it follows that every nonzero f ∈ V has at most n strictly positive roots, counting with multiplicity.On the other hand, Corollaries 3.2 and 3.3 show that every f ∈ E with deg(f ) = m 0 has at least n roots in I.If in addition f (0) > 0, it follows that f has precisely n strictly positive roots, and they all lie in I.

3.14.
Let f ∈ E with f (0) > 0 and deg(f ) = m 0 .Note that m n = 0 now since f (0) = 0.By 3.13, f has precisely n positive roots, and they all lie in I.It suffices to show that every such f satisfies an identity (6).
Let ξ 1 , . . ., ξ r be the different positive roots of f , and let b i = ord ξi (f ) (i = 1, . . ., r).Each b i is an even integer, except possibly in case (a) when ξ i = 1.We have r i=1 b i = n since f has n roots in I. Consider the determinant (5) in 3.6, with b 0 := 1 and p i = t mi (i = 0, . . ., n).After substituting y 0 = t and y i = ξ i (i = 1, . . ., r), Proposition 3.7 shows that the determinant has a factorization , . . ., ξ r , . . ., ξ r br with γ = 0 a constant.The last factor s λ (t, ξ 1 , . . ., ξ r ) is a polynomial in t.Since all coefficients of the Schur polynomial s λ are ≥ 0, and since ξ i > 0 for all i, this polynomial is not identically zero (and has degree λ 0 = m 0 − n).Hence the determinant is non-zero, and Lemma 3.4 implies that it agrees with f up to scaling.This means that f has a factorization with a constant γ ′ = 0.The first factor has the form where deg(g) = k − 1 = ⌊ n 2 ⌋ and u(t) = 1 or t − 1.The case u(t) = t − 1 occurs only in case (a) (I = [0, 1]) when n is odd.The last factor s λ (t, ξ 1 , . . . ) in ( 8) is a polynomial in t with non-negative coefficients.So, up to a non-zero constant factor, the right hand side has the form claimed in Theorem 3.9.Since f | I ≥ 0 we see that γ ′ > 0 in case (a) with n even, or in case (b).When n is odd in case (a), we have ξ i = 1 for some i, and b i is odd, so γ ′ < 0 in this case.The theorem is proved.
1 + g 2 2 with polynomials g 1 , g 2 of degree ≤ k.Theorem 3.9 can be stated in the following alternative form: where p i , q i are polynomials of degree ≤ ⌊ m−1 2 ⌋.This point of view will be expanded in more detail in the next section.
given by For (b), consider the linear map φ : In either case we have P = V ∩ im(φ), according to Corollary 3.15.This is an explicit block semidefinite representation of P of block size k + 1.
Remark 3.18.For sparse polynomials f in R[t] that are non-negative on the whole real axis, there does not in general exist a sparse decomposition similar to the one described in Remark 3.16.For example, f = 3t 4 − 4t 3 + 1 = (t − 1) 2 (3t 2 + 2t + 1) is non-negative on R but cannot be written f = g 1 (t) + t 2 g 2 (t) with g 1 , g 2 sums of squares of linear polynomials.
Still, we may easily produce a block semidefinite representation of block size k+1 for P = {f ∈ V : f ≥ 0 on R}, from the case I = R + in 3.17.It suffices to remark that f (t) ≥ 0 on R is equivalent to f (t) and f (−t) both being ≥ 0 on R + .
By cone duality we get the desired theorem for convex hulls of monomial space curves: Theorem 3.19.Let m 1 , . . ., m n ≥ 1 be integers, and let I ⊆ R be a semialgebraic set.The closed convex hull K of the set S = (t m1 , . . ., t mn ) : t ∈ I in R n has semidefinite extension degree at most ⌊ n 2 ⌋ + 1. Proof.The set K is identified with an affine-linear slice of the dual cone P * of P .Since sxdeg(P * ) = sxdeg(P ) ([17] Prop.1.7), it suffices to prove sxdeg(P ) ≤ ⌊ n 2 ⌋+1.We may assume that the m i are pairwise distinct.When I = [0, 1] or I = R + , the claim sxdeg(P ) ≤ ⌊ n 2 ⌋ 1 was shown in Remark 3.17.We sketch how the case of other sets I can essentially be reduced to these two cases, without going into full details.Clearly I can be assumed to be closed.If I = I 1 ∪ I 2 with I 1 , I 2 semialgebraic, it is enough to prove the claim for both I 1 and I 2 , in view of [17] Prop.1.6.In this way we reduce to considering The cases a = 0 are already done (assuming b = 1 in the compact case was nowhere essential).Assume I = [a, b] with 0 < a < b < ∞.Then every f ∈ P that generates an extreme ray of P has exactly n roots in I, counting with multiplicities.So the reduction step 3.11 in the proof of Theorem 3.9 is not needed.Otherwise we may just follow the proof of this theorem.In this way we arrive at a representation of every element of P in a form similar to 3.9(a), but with weights 1, t − a, b − t and (t − a)(b − t) instead of only 1 and 1 − t.When I = [a, ∞[ with a > 0, we may proceed in a similar way.Remark 3.20.For I = [0, 1] or I = R + , an explicit block semidefinite representation of K of block size k = ⌊ n 2 ⌋ + 1 can be obtained from Remarks 3.17, 3.18 by dualizing.
Remark 3.21.It is natural to ask whether the bound sxdeg(K) ≤ ⌊ n 2 ⌋+ 1 extends to cases more general than convex hulls of monomial curves.In fact, the same bound is true in general whenever K ⊆ R n is the closed convex hull of a one-dimensional semialgebraic set in R n [18].However the proof gets much more difficult than in the monomial case.Even when the curve is parametrized by polynomials instead of monomials, one has to argue locally on sufficiently small intervals on the curve, and there does not seem to be an explicit form for a block semidefinite representation.When the curve is non-rational, the proof becomes technically even much more complicated.

Sparse non-negativity certificates in polynomial optimization
4.1.Sums of copositive fewnomials and sparse semidefinite relaxations.We are going to take a second look at Corollary 3.15, concentrating on part (b).A univariate polynomial f ∈ R[t] will be called a k-nomial if f is a linear combination of at most k monomials t i .We say that f is copositive if f ≥ 0 on R + .Given a finite set J ⊆ Z + := {0, 1, 2, . . .}, let CP(J) := {p ∈ R[t] : p ≥ 0 on R + and supp(p) ⊆ J} .This is the cone P considered in Sect.
For d = 2k, note that SOCF 2k+1,2k is just the cone of all copositive univariate polynomials of degree at most 2k, i.e., it is the cone where the latter description of copositivity in the univariate case is well known.
Remark 4.4.For the case k = 0, relaxations in Proposition 4.3 are trivial with the optimal value f (0) or −∞, depending on whether or not all coefficients of f − f (0) are non-negative.But already the simplest non-trivial case k = 1 exhibits connections to active ongoing research on sparse relaxations in polynomial optimization, namely to the relaxations that rely on the so-called sage and sonc polynomials; see [8] and [14,10], respectively.Both kinds of polynomials emerge from the same idea applied to non-negativity on R n + and R n , respectively.A sage polynomial is a sum of polynomials that are non-negative on R n + and whose supports are contained in a simplicial curcuit.Here, a simplicial circuit is an inclusion-minimal affinely dependent set whose convex hull is a simplex.Since simplicial circuits within R are merely three-element sets, we see that the cone SOCF 3,d is the cone of all univariate sage polynomials of degree at most d.Thus, the case k = 1 provides a primal and a dual sage relaxation of min R+ f .See, for example, [10] for the discussion of the duality for the sonc and sage cones and the applications in optimization.We also note that the special case k = 1 of Corollary 4.1 amounts to a description of non-negative sage polynomials in terms of the so-called reduced circuits; see [10].That is, in the univariate situation, an analog of a reduced circuit in the setting of copositive (2k + 1)-nomials is a set of 2k + 1 consecutive integer values, since the support of polynomials in t i Σ 2k is a subset of {i, . . ., i + 2k}.
We stress that the sparsity threshold of sage and sonc polynomials is bound to the dimension, since a circuit in dimension n cannot have more than n+ 2 elements.In contrast to this, in our setting the sparsity threshold 2k + 1 may vary arbitrarily, bridging the sage relaxations with the dense standard relaxations of Lasserre.While being aware that our results are limited to the univariate case, we hope that such results might serve as an inspiration for similar studies in an arbitrary dimension n.That is, it would be interesting to understand properties of sos and moment relaxations that have a variable sparsity threshold.
Remark 4.5.The above discussion of sparse polynomial optimization over R + can also be used to handle sparse polynomial optimization on an arbitrary interval I ⊆ R + .For example, when 2k < d − 1 and I = [0, 1], in view of Corollaries 3.15(a) and 4.1, the cone of sums of (2k + 1)-nomials of degree at most d that are nonnegative on [0, 1] can be described as SOCF

4.2.
Sparse semidefinite relaxations with a chordal-graph sparsity.We consider graphs G = (V, E) where |V | < ∞ and edges e ∈ E are two-element subsets of V .A cycle of length k in G is a connected subgraph of G with k nodes and each node having degree two.A chord of a cycle C in G is an edge of G that connects two nodes of C but is not an edge of C. A graph G is said to be chordal if every cycle of G of length at least four has a chord in G.A clique W ⊆ V of G = (V, E) is a set of nodes with any two distinct nodes in W connected by an edge.The following result provides a convenient representation of the cone of psd matrices with the chordal sparsity pattern: Theorem 4.6 (see [2] and [20] Sect.9.2).Let G = (V, E) be a graph with n = |V | < ∞.With G we associate the cone +,E := (a ij ) i,j∈V ∈ S n + : a ij = 0 when i = j and {i, j} ∈ E .If G is chordal, then this cone admits the decomposition Remark 4.7.S V +,E is the so-called cone of psd matrices with the sparsity pattern of +,E is known to have favorable theoretical and computational properties, which allow one to solve conic optimization problems with respect to the cones S V +,E more efficiently than the problems with respect to S n + .For example, computations with barrier functions can be done more efficiently on S V +,E since there are more efficient ways to carry out Cholesky factorization for matrices in S V +,E when the graph G = (V, E) is chordal.We refer to [23] Appendix A for further details.
Remark 4.8.As a consequence of Theorem 4.6 we see that, for a chordal graph G = (V, E), the semidefinite extension degree of S V +,E is the clique number of G, i.e., the maximum clique size of G.
For positive integers n and k with k < n, the cone +,E for the graph G = (V, E) with vertex set V = {0, . . ., n − 1} and edge set E consisting of {i, j} that satisfy 0 < |i − j| ≤ k.In [20] Sect.8.2, S n +,k is called the cone of psd matrices with the band sparsity pattern of the band width 2k + 1.It is clear that the graph G defining S n +,k is chordal and that inclusionmaximal cliques of G are sets of the form {i, . . ., i + k} with 0 ≤ i < n− k.We show that, similarly to how S n + is used for representing sos cones and truncated quadratic modules, the cone S n +,k can be used for representing the cone SOCF 2k+1,n .That is, optimization problem (1) from Proposition 4.3 can be formulated as a conic problem with respect to the cones S n +,k .Such formulations may have computational advantages if used in solvers that can exploit the chordal sparsity.respectively.This gives the desired assertion.
Remark 4.10.Wang, Magron and Lasserre [21,22] suggest to use the cones S V +,E from chordal graphs G = (V, E) to approximate sparse sos polynomials.For a given finite set A ⊆ Z n + , they consider the cone Σ(A) := {f ∈ R[x 1 , . . ., x n ] : f sos and supp(f ) ⊆ A} They suggest two iterative algorithms (one in [22] and another one in [22]) that take A as an input and produce a chordal graph G = (V, E) with V ⊆ Z n + in order to use the image of the repsective cone S V +,E under the map ϕ : (a α,β ) α,β∈V → α,β∈V a α,β x α+β as an approximation of Σ(A).The algorithm in [21] produces a G = (V, E) being a disjoint union of cliques.The graph G = (V, E) in [21] is guaranteed to satisfy the equality Σ(A) = f ∈ φ(S V +,E ) : supp(f ) ⊆ A (10) (see Thm. 3.3 in [21]), but there is no guarantee that the clique number of G is small, or rather, the dependence of the clique number of G on the properties of A remains unexplored.Since the graph G = (V, E) generated by an algorithm from [21] may have large cliques, in [22] another approach is suggested that generates a graph G = (V, E) with a smaller number of edges.This other approach is heuristic in the sense that there are no theoretical guarantees for the equality Σ(A) = φ(S V +,E ) (see Example 3.5.in [22]).It would be interesting to study the semidefinite extension degree of the cones Σ(A) and try to relate these cones to the cones S V +,E in a non-heuristic way.

[4]: Definition 1 . 1 .
Let C ⊆ R n be a set.A description of C in the formC = {x ∈ R n : M 1 (x, y) 0, . . ., M ℓ (x, y) 0 for some y ∈ R m }(1)where the M i are LMIs in (x, y), is called a lifted (or extended ) semidefinite representation of C. If every M i has size ≤ d, we say that (1) is a lifted representation of size ≤ d.The minimal d for which C admits a lifted representation of size ≤ d is the semidefinite extension degree of C, denoted sxdeg(C).If no such d exists one puts sxdeg(C) = ∞.

Corollary 3 . 3 .
Let I be unbounded, let f ∈ P span an extreme ray of P .Write d = deg(f ) and V d := {g ∈ V : deg(g) ≤ d}.Then f has at least dim(V d ) − 1 roots in I, counting with multiplicity.Proof.Same argument as for Corollary 3.2, since Rf consists of the elements in V d with at least the same roots in I as f .Proof of Proposition 3.1.The supporting face of f is

Proposition 4 . 9 .
Let d, k be positive integers with 2k < d.Then the cone SOCF 2k+1,d admits a representation as a linear image of S e+1 +,k × S e +,k , if d = 2e is even, and as a linear image of S e+1 +,k × S e+1 +,k , if d = 2e + 1 is odd.Proof.We only consider the case of an even d with d = 2e, because the case of an odd d is completely analogous.We use the linear map ϕ : S e+1 → R[t] d with ϕ(A) := e i,j=0 a ij t i+j for A = (a ij ) i,j=0,...,e ∈ S e+1 from Remark 3.17.As mentioned in Remark 3.17, ϕ(S W + ) = Σ 2k for W = {0, . . ., k}, where the notation S W + is borrowed from Theorem 4.6.The latter implies ϕ(S Vi + ) = t 2i Σ 2k for V i := {i, . . ., i + k} with 0 ≤ i ≤ d − 2k.Consequently, splitting the sum representing SOCF 2k+1,d in (9) into two according to the parity of the exponents in t 0 , . . ., t d−2k , we obtain SOCF 2k+1,d = 2k+1,d is a linear image of K × L under the linear map (A, B) → ϕ(A) + t ϕ(B).In view of Theorem 4.6, K and L are copies of S e+1 +,k and S e +,k 16mark 3.16.Part (b) of Corollary 3.15 can be considered to be a sparse (or fewnomial ) positivstellensatz (or rather, nichtnegativstellensatz ) for univariate poly- nomials: Every polynomial f ∈ R[t] with m monomials that is non-negative on R + can be written as a (finite) sum The acronym SOCF stands for sums of copositive fewnomials.So f ∈ R[t] d lies in SOCF k,d if, and only if, f can be written as a finite sum of copositive k-nomials of degree at most d.It is obvious how to generalize this setup from univariate polynomials to a multivariate setting.This could be an interesting setup to explore in the future.As before, let Σ 2k ⊆ R[t] denote the set of sums of squares polynomials of degree ≤ 2k.The main result of Sect. 3 implies the following positivstellensatz for the cones SOCF: Corollary 4.1.For all integers k, d ≥ 1 with d > 2k, we have