1 Introduction

Let \((X, \Vert \,{ \cdot }\, \Vert )\) be a normed space. A set \(S\subseteq X\) is called c-equilateral if \(\Vert x-y\Vert =c\) for all distinct \(x,y\in S\). S is called equilateral if it is c-equilateral for some \(c>0\). The equilateral number e(X) of X is the cardinality of the largest equilateral set of X. Petty [7] made the following conjecture regarding lower bounds on e(X).

Conjecture 1.1

For all normed spaces X of dimension n, \(e(X)\ge n+1\).

Petty [7] proved Conjecture 1.1 for \(n=3\), and Makeev [6] for \(n=4\). For \(n\ge 5\) the conjecture is still open, except for some special classes of norms. The best general lower bound is \(e(X)\ge \exp \Omega (\sqrt{\log n})\), proved by Swanepoel and Villa [10]. Regarding upper bounds on the equilateral number, a classical result of Petty [7] and Soltan [8] shows that \(e(X)\le 2^n\) for any X of dimension n, with equality if and only if the unit ball of X is an affine image of the n-dimensional cube. For more background on the equilateral number see Sect. 3 of the survey [9].

The norm \(\Vert \,{\cdot }\,\Vert _{\infty }\) of \(x\in \mathbb {R}^n\) is defined as \(\Vert x\Vert _{\infty }=\max _{1\le i\le n}|x_i|\), and \(\ell _{\infty }^n\) denotes the normed space \((\mathbb {R}^n, \Vert \,{\cdot }\,\Vert _{\infty })\). In [5] Kobos studied subspaces of \(\ell _{\infty }^n\) of codimension 1, and proved the lower bound \(e(X)\ge 2^{\lfloor {n}/{2}\rfloor }\), which in particular implies Conjecture 1.1 for these spaces for \(n\ge 6\). In the same paper he proposed as a problem to prove Petty’s conjecture for subspaces of \(\ell _{\infty }^n\) of codimension 2. In Theorem 1.2 we prove exponential lower bounds on the equilateral number of subspaces of \(\ell _{\infty }^n\) of codimension k. This, in particular, solves Kobos’ problem if \(n\ge 9\).

Theorem 1.2

Let X be an \((n-k)\)-dimensional subspace of \(\ell _\infty ^n\). Then

$$\begin{aligned}&e(X) \ge \frac{2^{n-k}}{(n-k)^k}, \end{aligned}$$
(1)
$$\begin{aligned}&e(X)\ge 1+\frac{1}{2^{k-1}} \displaystyle \sum _{r=1}^\ell \left( {\begin{array}{c}n-k\ell \\ r\end{array}}\right) \quad for \ every \ 1\le \ell \le \frac{n}{k+1},\quad {{and}}\end{aligned}$$
(2)
$$\begin{aligned}&e(X) \ge 1+ \sum _{r=1}^{\ell }\left( {\begin{array}{c}n-2k\ell \\ r\end{array}}\right) \quad for \ every \ 1\le \ell \le \frac{n}{2k+1}. \end{aligned}$$
(3)

Note that none of the three bounds follows from the other two in Theorem 1.2, hence none of them is redundant. Comparing (1) and (3), for fixed k we have

$$\begin{aligned} \max _{\ell } \sum _{1\le r \le \ell }\left( {\begin{array}{c}n-2k\ell \\ r\end{array}}\right) =O( 2^{c_kn}) \end{aligned}$$

for some \(0<c_k <1\), while \({2^{n-k}}/{(n-k)^k}=2^{n-k-k\log (n-k)}=2^{n-o(n)}\). On the other hand, when we let k vary, it can be as large as \(\Omega (n)\) in (3) to still give a non-trivial estimate, while k can only be chosen up to \(O(n/\log n)\) for (1) to be non-trivial. Finally, (2) is beaten by (1) or (3) in most cases, however, for \(k=2, 3\) and for small values of n, (2) gives the best bound. Table 1 shows the lower bounds given by Theorem 1.2 for small values of k and n, whenever it is at least 3.

Table 1 Lower bounds provided by Theorem 1.2 for small values of k and n

For two n-dimensional normed spaces XY by \(d_{\text {BM}}(X,Y)=\inf _T\{\Vert T\Vert \cdot \Vert T^{-1}\Vert \}\) we denote their Banach–Mazur distance, where the infimum is over all linear isomorphisms \(T:X\rightarrow Y\). The metric space of isometry classes of normed spaces endowed with the logarithm of the Banach–Mazur distance is the Banach–Mazur compactum. It is not hard to see that e(X) is upper semi-continuous on the Banach–Mazur compactum. This, together with the fact that any convex polytope can be obtained as a section of a cube of sufficiently large dimension (see for example p. 72 of Grünbaum’s book [4]) implies that it would be sufficient to prove Conjecture 1.1 for k-codimensional subspaces of \(\ell _{\infty }^n\) for all \(1\le k \le n-4\) and \(n\ge 5\). (This was also pointed out in [5].) Unfortunately, our bounds are only non-trivial if n is sufficiently large compared to k. However, we deduce an interesting corollary.

Corollary 1.3

Let P be an origin-symmetric convex polytope in \(\mathbb {R}^d\) with at most

$$\begin{aligned} \frac{4d}{3}-\frac{1+\sqrt{8d+9}}{6}=\frac{4d}{3}-o(d) \end{aligned}$$

opposite pairs of facets. If X is a d-dimensional normed space with P as a unit ball, then \(e(X)\ge d+1\).

There have been some extensions of lower bounds obtained on the equilateral number of certain normed spaces to other norms that are close to them according to the Banach–Mazur distance. These results are based on using Brouwer’s fixed point theorem, first applied in this context by Brass [2] and Dekster [3]. Swanepoel and Villa [10] proved that if \(d_{\text {BM}}(Y,\ell ^n_{\infty })\le {3}/{2}\), then \(e(Y)\ge n+1\). Kobos [5] proved that if X is a subspace of \(\ell _{\infty }^n\) of dimension \(n-1\) and Y is a normed space of dimension \(n-1\) such that \(d_{\text {BM}}(X,Y)\le 2\), then \(e(Y)\ge \lceil {n}/{2}\rceil \). We prove a similar result for subspaces of \(\ell _{\infty }^n\) of codimension at least 1.

Theorem 1.4

For \(k\ge 1\) let X be an \((n-k)\)-dimensional subspace of \(\ell _{\infty }^n\), and Y be an \((n-k)\)-dimensional normed space such that

$$\begin{aligned} d_ \mathrm{BM}(X,Y)\le 1+\frac{\ell }{2(n-2k-\ell k-1)} \end{aligned}$$

for some integer \(1\le \ell \le ({n-2k})/{k}\). Then \(e(Y)\ge n-k(2+\ell )\).

2 Norms with Polytopal Unit Ball and Small Codimension

We recall the following well-known fact to prove Corollary 1.3. (For a proof, see for example [1].)

Lemma 2.1

Any centrally symmetric convex d-polytope with \(f\ge d\) opposite pairs of facets is a d-dimensional section of the f-dimensional cube.

Proof of Corollary 1.3

By Lemma 2.1, P can be obtained as an d-dimensional section of the \( ({4d}/{3}-({1+\sqrt{8d+9}})/{6})\)-dimensional cube. Choose

$$\begin{aligned} n=\frac{4d}{3}-\frac{1+\sqrt{8d+9}}{6},\qquad \ell =2, \qquad k=\frac{d}{3}-\frac{d+\sqrt{8d+9}}{6}, \end{aligned}$$

and apply inequality (3) from Theorem 1.2. This yields \(e(X)\ge d+1\). \(\square \)

To confirm Petty’s conjecture for subspaces of \(\ell _{\infty }^n\) of codimension 2 and 3 when \(n\ge 9\) and respectively \(n\ge 15\), apply inequality (2) from Theorem 1.2 with \(\ell =2\).

3 Large Equilateral Sets

3.1 Notation

We denote vectors by bold lowercase letters, and the i-th coordinate of a vector \(\mathbf {a}\in \mathbb {R}^n\) by \(a^i\). We treat vectors by default as column vectors. By subspace we mean linear subspace. We write \({{\,\mathrm{span}\,}}{(\mathbf {a}_1,\dots ,\mathbf {a}_k)}\) for the subspace spanned by \(\mathbf {a}_1,\dots ,\mathbf {a}_k\in \mathbb {R}^n\). For a subspace \(X\subseteq \mathbb {R}^n\) we denote by \(X^{\bot }\) the orthogonal complement of X. We denote by [n] the set \(\{1,\dots , n\}\) and by \(2^{[n]}\) the set of all subsets of [n]. \(\mathbf {0}\) denotes the vector \((0,\dots ,0)\in \mathbb {R}^n\). For two vectors \(\mathbf {a}\) and \(\mathbf {b}\), let \(\mathbf {a}\cdot \mathbf {b}=\sum _{i=1}^n a^ib^i\) be their scalar product.

3.2 Idea of the Constructions

For two vectors \(\mathbf {x},\mathbf {y}\in X\) we have \(\Vert \mathbf {x}-\mathbf {y}\Vert _{\infty }=c\) if and only if the following hold:

$$\begin{aligned}&\text {there is an}~ 1\le i \le n~ \text {such that}~ |x^i-y^i|=c,~ \text {and}\end{aligned}$$
(4)
$$\begin{aligned}&|x^j-y^j|\le c ~\text {for all} ~1\le j \le n. \end{aligned}$$
(5)

In our constructions of c-equilateral sets \(S\subseteq X\), we split the index set [n] of the coordinates into two parts, \([n]=N_1\cup N_2\). In the first part \(N_1\), we choose all the coordinates from the set \(\{0,1,-1\}\), so that for each pair from S there will be an index in \(N_1\) for which (4) holds, and (5) is not violated by any index in \(N_1\). We use \(N_2\) to ensure that all of the points we choose are indeed in the subspace X. For each vector, this will lead to a system of linear equations. The main difficulty will be to choose the values of the coordinates in \(N_1\) so that the coordinates in \(N_2\), obtained as a solution to those systems of linear equations, do not violate (5).

3.3 Proof of Theorem 1.2

For vectors \(\mathbf {v}_1,\dots ,\mathbf {v}_k\in \mathbb {R}^k\) let \(B(\mathbf {v}_1,\dots ,\mathbf {v}_k)\in \mathbb {R}^{k\times k}\) be the matrix whose i-th column is \(\mathbf {v}_i\). For a matrix \(B\in \mathbb {R}^{k\times k}\), a vector \(\mathbf {v}\in \mathbb {R}^k\), and an index \(i\in [k]\), we denote by \(B(i,\mathbf {v})\) the matrix obtained from B by replacing its i-th column by \(\mathbf {v}\).

Let \(\{\mathbf {a}_i:1\le i \le k\}\) be a set of k linearly independent vectors in \(\mathbb {R}^n\) spanning \(X^{\bot }\). That is, \(\mathbf {x}\in X\) if and only if \(\mathbf {a}_i\cdot \mathbf {x}=0\) for all \(1\le i \le k\). Further, let \(A\in \mathbb {R}^{k\times n}\) be the matrix whose i-th row is \(\mathbf {a}_i^T\), and let \(\mathbf {b}_j=(a_1^j,\dots , a_k^j)\) be the j-th column of A. For \(I\subseteq [n]\) and for \(\varvec{\sigma }\in \{ \pm 1\}^n\) let \(\mathbf {b}_{I}=\sum _{i\in I}\mathbf {b}_i\) and \(\mathbf {b}_{I,\varvec{\sigma }}=\sum _{i\in I}\sigma ^i\mathbf {b}_i\). For a subset \(N\subseteq [n]\) we denote by \(A|_N\) the submatrix of A formed by those columns of A whose index is in N. Similarly, for a vector \(\mathbf {v}\in \mathbb {R}^n\) we denote by \(\mathbf {v}|_N\) the vector that is formed by those coordinates of \(\mathbf {v}\) whose index is in N (without changing the order of the indices).

Proof of (1)

We construct a 2-equilateral set of size \({2^{n-k}}/{(n-k)^k}\). Let B be a \(k\times k\) submatrix of A for which \(|{\det B}|\) is maximal among all \(k\times k\) submatrices. Further, let \(N_2\) be the set of indices of the columns of B and \(N_1=[n]\setminus N_2\). Note that \(\det B \ne 0\), since the vectors \(\{\mathbf {a}_i:i\in [k]\}\) are linearly independent. First, we find many vectors in X whose coordinates in \(N_1\) are from \(\{1,-1\}\), and then we select a subset of these that form an equilateral set. For every \(J\subseteq N_1\) we define the vector \(\mathbf {w}(J)\in \mathbb {R}^n\) as

$$\begin{aligned} w(J)^i= {\left\{ \begin{array}{ll} 1 &{}\quad \text {if}~ i\in J,\\ -1 &{}\quad \text {if}~ i\in N_1\setminus J,\\ \displaystyle \frac{\det B(j,\mathbf {b}_{N_1\setminus J}-\mathbf {b}_{J})}{\det B} &{}\quad \text {if}~ i\in N_2 ~\text {is the}~ j\text {-th element of}~N_2. \end{array}\right. } \end{aligned}$$

To see that \(\mathbf {w}(J)\) is in X, we need to check that \(A\mathbf {w}(J)=\mathbf {0}\). This is indeed the case, since by Cramer’s rule \(\mathbf {w}(J)|_{N_2}\) is the solution of

$$\begin{aligned} B\mathbf {x}=\mathbf {b}_{N_1\setminus J}-\mathbf {b}_J. \end{aligned}$$

Further, by the multilinearity of the determinant for every J and j we have

$$\begin{aligned} \det B(j,\mathbf {b}_{N_1\setminus J}-\mathbf {b}_{J})=\sum _{r\in N_1\setminus J}\!\!\det B(j,\mathbf {b}_r)-\sum _{r\in J}\det B(j,\mathbf {b}_r). \end{aligned}$$

By the maximality of \(|{\det B}|\) and by the triangle inequality we have

$$\begin{aligned} \bigl |{\det B(j,\mathbf {b}_{N_1\setminus J}-\mathbf {b}_{J})}\bigr | \le (n-k)|{\det B}|. \end{aligned}$$

This implies that for each J and \(i\in N_2\) we have \(-(n-k)\le w(J)^i \le n-k\). Consider the set \(W=\{\mathbf {w}(J):J\in 2^{N_1}\}\). W is not necessarily a 2-equilateral set, because for \(J_1,J_2\in 2^{N_1}\) and for \(i\in N_2\) we only have that \(|w(J_1)^i-w(J_2)^i|\le 2(n-k)\). However, we can find a 2-equilateral subset of W that has large cardinality, as follows.

For each vector \(\mathbf {s}\in \{-(n-k),(-n-k)+2,\dots ,n-k-2\}^k=T^k\) let \(W(\mathbf {s})\) be the set of those vectors \(\mathbf {w}(J)\) for which

$$\begin{aligned} w(J)^{j}\in [s^i,s^i+2]\quad \text { for every }~\ i\in k, \end{aligned}$$

where j is the i-th element of \(N_2 \). Since \(W\subseteq \bigcup _{\mathbf {s}\in T^k} W(\mathbf {s})\), there is an \(\mathbf {s}\) for which \(|W(\mathbf {s})|\ge {2^{n-k}}/{(n-k)^k}\). It is not hard to check that \(W(\mathbf {s})\) is 2-equilateral. Indeed, for every \(J_1,J_2\in W(\mathbf {s})\), we have \(|w^i(J_1)-w^i(J_2)|\le 2\) for \(i\in N_2\) by the definition of \(W(\mathbf {s})\), and for \(i\in N_1\) by the definition of \(\mathbf {w}(J)\). Further, by the definition of \(\mathbf {w}(J)\) there is an index \(j\in N_1\) for which \(\{w(J_1)^j, w(J_2)^j\}=\{1,-1\}\) (assuming \(J_1\ne J_2\)). \(\square \)

Proof of (2)

Fix some \(1\le \ell \le n/(k+1)\). We will construct a 1-equilateral set of cardinality

$$\begin{aligned} \frac{1}{2{^{k-1}}}\sum _{1\le r \le \ell }\left( {\begin{array}{c}n-k\ell \\ r\end{array}}\right) +1. \end{aligned}$$

Let \(I_1,\dots , I_k\subseteq [n]\) be sets of cardinality at most \(\ell \), and \(\varvec{\sigma }\in \{\pm 1\}^n\) be a sign vector with the following properties:

(\(*\)):

The determinant of \(B=B(\mathbf {b}_{I_1, \varvec{\sigma }},\dots , \mathbf {b}_{I_k, \varvec{\sigma }})\) is maximal among all possible choices of k disjoint sets \(I_1,\dots , I_k\subseteq [n]\) of cardinality at most \(\ell \) and \(\sigma \in \{\pm 1\}^n\).

(\(**\)):

\(\det {B(k,\sigma ^j\mathbf {b}_j)}\ge 0\) for every \(j\in [n]\setminus (I_1\cup \cdots \cup I_k)\).

Note that \(\det B>0\), since the vectors \(\mathbf {a}_1,\dots ,\mathbf {a}_k\) are linearly independent. Let \(N_2=\bigcup _{i\in [k]} I_i\) and \(N_1=[n]\setminus N_2\). We find a set of vectors whose coordinates in \(N_1\) are from \(\{0,1,-1\}\), and then we find a large subset of them that forms an equilateral set.

For every subset \(J\subseteq N_1\) of cardinality at most \(\ell \) we define the vector \(\mathbf {w}(J)\in \mathbb {R}^n\) as

$$\begin{aligned} {w}(J)^i = {\left\{ \begin{array}{ll} -\sigma ^i &{}\quad \text {if}~ i\in J , \\ 0 &{}\quad \text {if}~ i\in N_1\setminus J, \\ \displaystyle \frac{\sigma ^i\det B(j,\mathbf {b}_{J,\sigma })}{\det B} &{}\quad \text {if}~ i\in I_j~ \text {for some}~ j\in [k]. \end{array}\right. } \end{aligned}$$

To see that \(\mathbf {w}(J)\) is in X, we have to check that \(A\mathbf {w}(J)=\mathbf {0}\). This follows by showing that \(\mathbf {w}(J)|_{N_2}\) is a solution of

$$\begin{aligned} A|_{N_2}\mathbf {x}=\mathbf {b}_{J,\sigma }. \end{aligned}$$
(6)

By Cramer’s rule \(\mathbf {y}\in \mathbb {R}^k\) defined as

$$\begin{aligned} y^j=\frac{\det B(j,\mathbf {b}_{J,\sigma })}{\det B} \end{aligned}$$

is a solution of \(B\mathbf {y}=\mathbf {b}_{J,\sigma }.\) Thus, \(\mathbf {w}(J)|_{N_2}\) is indeed a solution of (6) of a special form where \(x^i=\sigma ^iy^j\) if \(i\in I_j\). Note that \(B(j,\mathbf {b}_{J,\sigma })=B(\mathbf {b}_{J_1,\sigma },\dots ,\mathbf {b}_{J_k,\sigma })\) for some disjoint sets \(J_1,\dots , J_k\), hence by property (\(*\)) we have

$$\begin{aligned} |w(J)^i|\le 1 \quad \text { for each }\, 1\le i \le n. \end{aligned}$$
(7)

Further, the multilinearity of the determinant together with property (\(**\)) implies that for \(i\in I_k\) we have

$$\begin{aligned} \begin{aligned} \sigma ^i w(J)^i&=\frac{\det {B(k,\mathbf {b}_{J,\sigma })}}{\det B}=\frac{\det {B\bigl (k,\sum _{j\in J}\sigma ^j\mathbf {b}_j\bigr )}}{\det B}\\&=\frac{\sum _{j\in J}\det {B(k,\sigma ^j\mathbf {b}_j)}}{\det B}\ge 0. \end{aligned} \end{aligned}$$
(8)

Consider the set \(W=\{\mathbf {w}(J):J\subseteq N_1,\, |J|\le \ell \}\). W is not necessarily a 1-equilateral set, because for \(J_1,J_2\subseteq N_1\) with \(|J_1|,|J_2|\le \ell \) and for some \(i_1\in I_1\cup \cdots \cup I_{k-1}\) we only have that \(w(J_1)^i, w(J_2)^i\in [-1,1]\), and thus \(|w(J_1)^i-w(J_2)^i|\le 2\). However, we can find a 1-equilateral subset of W of large cardinality.

For each vector \(\mathbf {s}\in \{\pm 1\}^{k-1}\) let \(W(\mathbf {s})\subseteq W\) be the set of those vectors \(\mathbf {w}(J)\in W\) for which

$$\begin{aligned} s^jw(J)^i\sigma ^i\ge 0\quad \text { for each }\, i\in I_1\cup \cdots \cup I_{k-1}, \end{aligned}$$

where \(j\in [k-1]\) is such that \(i\in I_j\). Then \(\bigcup _{\mathbf {s}\in \{\pm 1\}^{k-1}} W(\mathbf {s})\) is a partition of W, hence there is an \(\mathbf {s}\) for which

$$|W(\mathbf {s})|\ge \frac{1}{2^{k-1}}|W|\ge \frac{1}{2{^{k-1}}}\sum _{1\le r \le \ell }\left( {\begin{array}{c}n-k\ell \\ r\end{array}}\right) .$$

\(W(\mathbf {s})\) is a 1-equilateral set, because for any two vectors \(\mathbf {w}_1,\mathbf {w}_2\in W(\mathbf {s})\), there is an index \(i\in N_1\) for which either \(\{w_1^i, w_2^i\}=\{0,-1\}\) or \(\{w_1^i, w_2^i\}=\{0,1\}\), and for all \(i\in [n]\) we have \(|w_1^i-w_2^i|\le 1\) by (7), by the definition of \(W(\mathbf {s})\), and by (8). Finally, it is not hard to see that we can add \(\mathbf {0}\) to \(W(\mathbf {s})\). Thus \(W(\mathbf {s})\cup \{\mathbf {0}\}\) is a 1-equilateral set of the promised cardinality. \(\square \)

In the proof of (3) we will need the following simple lemma.

Lemma 3.1

Let A be a real matrix of size \(k\times n\). Then for every \(\varepsilon >0\) there exists a real matrix \(A'\) of size \(k\times n\) such that \(|a_j^i-{a_j^i}'|\le \varepsilon \) for all ij, and every \(k\times k\) minor of \(A'\) is non-zero.

Proof

Associating matrices with points of \(\mathbb {R}^{n\times k}\), the set of those matrices in which a given \(k\times k\) minor is 0 is the zero set of a non-zero polynomial, which is nowhere dense. Thus, the set of those matrices for which there is a \(k\times k\) minor which is zero is nowhere dense, which implies the statement. \(\square \)

Proof of (3)

Fix some \(1\le \ell \le n/(2k+1)\). We will construct a 1-equilateral set of cardinality

$$\sum _{1\le r \le \ell }\left( {\begin{array}{c}n-2k\ell \\ r\end{array}}\right) +1.$$

Pick \(2\ell \) disjoint submatrices \(B_1,\dots ,B_{2\ell }\) of A of size \(k\times k\), such that for every \(1\le m \le k\) and \(1\le i \le 2\ell \) and for any column \(\mathbf {b}_r\) that is not a column of any \(B_j\) we have

$$\begin{aligned} |{\det B_i}| \ge |{\det B_i(m,\mathbf {b}_r)}|. \end{aligned}$$
(9)

This we may do by choosing the \(B_i\)’s after each other, always choosing the submatrix with the largest determinant disjoint from the previous submatrices. Using Lemma 3.1 and the upper semi-continuity of the equilateral number, we may further assume that \(|{\det B_{i}}|>0\) for all \(1\le i \le 2 \ell \).

Let \(U_i\) be the set of indices of the columns of \(B_i\), let \(N_2=U_1\cup \dots \cup U_{2\ell }\) and \(N_1=[n]\setminus N_2\). We will find vectors in X (denoted by \(\mathbf {y}(J)\)) whose coordinates in \(N_1\) are from the set \(\{0,-1\}\), and whose coordinates in \(N_2\) have absolute value at most 1/2. We do not construct them directly, but as the sum of some other vectors \(\mathbf {w}(J,i), \mathbf {z}(J,i)\in X\), whose coordinates in \(N_1\) are from \(\{0,-{1}/{2}\}\).

For every set \(J\!=\!\{q_1,\dots ,q_{|J|} \}\!\subseteq \![N_1]\) of cardinality at most \(\ell \) with \(q_1<\dots <q_{|J|}\), and for every \(1\le i \le |J|\) we define \(\mathbf {w}(J,i)\in \mathbb {R}^n\) and \(\mathbf {z}(J,i)\in \mathbb {R}^n\) as

$$\begin{aligned} w(J,i)^j&= {\left\{ \begin{array}{ll} \displaystyle -\frac{1}{2} &{}\quad \text {if}~ j=q_i,\\ 0 &{}\quad \text {if}~ j\in [n]\setminus (\{q_i\}\cup U_{2i}), \\ \displaystyle \frac{\det B_{2i} (m,\mathbf {b}_{q_i}/2)}{\det B_{2i}} &{}\quad \text {if}~ j ~\text {is the}~ m\text {-th element of}~ U_{2i}, \end{array}\right. }\\ z(J,i)^j&={\left\{ \begin{array}{ll} \displaystyle -\frac{1}{2} &{}\quad \text {if}~ j=q_i,\\ 0 &{}\quad \text {if}~ j\in [n]\setminus (\{q_i\}\cup U_{2i-1}),\\ \displaystyle \frac{\det B_{2i-1}(m,\mathbf {b}_{q_i}/2)}{\det B_{2i-1}} &{}\quad \text {if}~ j ~\text {is the}~ m\text {-th element of}~ U_{2i-1}. \end{array}\right. } \end{aligned}$$

To see that \(\mathbf {w}(J,i)\) and \(\mathbf {z}(J,i)\) are in X, we need to check that \(A\mathbf {w}(J,i)=0\) and \(A\mathbf {z}(J,i)=\mathbf {0}\). This is indeed the case, since by Cramer’s rule both \(\mathbf {w}(J,i)|_{U_{2i}}\) and \(\mathbf {z}(J,i)|_{U_{2i-1}}\) are solutions of

$$\begin{aligned} A|_J\mathbf {x}=\frac{\mathbf {b}_{q_i}}{2}. \end{aligned}$$

Therefore \(\mathbf {y}(J)=\sum _{1\le i \le |J|} (\mathbf {w}(J,i)+\mathbf {z}(J,i))\) is also in X. Note that by assumption (9) and by the multilinearity of the determinant we have \(|w(J,i)^j|,|z(J,i)^j|\le {1}/{2}\) for all \(1\le j \le n\). For the coordinates of \(\mathbf {y}(J)\) we have

$$\begin{aligned} y(J)^j&= -1\quad \text {if}~ j\in J,\\ y(J)^j&= 0\qquad \text {if}~ j\in [N_1]\setminus J,\\ |y(J)^j|&\le \frac{1}{2}\quad \text {if}~ j\in N_2. \end{aligned}$$

Thus, for any two distinct sets \(J_1,J_2\subseteq [N_1]\) of cardinality at most \(\ell \) there is a coordinate \(j\in [N_1]\) for which \(\{y(J_1)^j,y(J_2)^j\}=\{0,-1\}\), and for all \(1\le j \le n\) we have \(|y(J_1)^j-y(J_2)^j|\le 1\). This means \(\Vert \mathbf {y}(J_1)-\mathbf {y}(J_2)\Vert _{\infty }=1\), and \(\{\mathbf {y}(J):J \subseteq [N_1], \,|J|\le \ell \}\cup \{\mathbf {0}\}\) is a 1-equilateral set of cardinality

$$\begin{aligned} \sum _{1\le r \le \ell }\left( {\begin{array}{c}N_1\\ r\end{array}}\right) +1=\sum _{1\le r \le \ell }\left( {\begin{array}{c}n-2k\ell \\ r\end{array}}\right) +1. \end{aligned}$$

\(\square \)

4 Equilateral Sets in Normed Spaces Close to Subspaces of \(\ell _{\infty }^n\)

In this section we prove Theorem 1.4. The construction we use is similar to the one from [10]. Fix some \(1\le \ell \le ({n-2k})/{k}\), and let \(N=n-k(2+\ell )\) and \(c={\ell }/({2(N-1)})>0\). We assume that the linear structure of Y is identified with the linear structure of X and the norm \(\Vert \,{\cdot }\, \Vert _Y\) of Y satisfies

$$\begin{aligned} \Vert x\Vert _Y\le \Vert x\Vert _{\infty } \le (1+c)\Vert x\Vert _Y \end{aligned}$$

for each \(x\in X\). Further, let \(M=\{(i,j):1\le i < j \le N\}\). For every \(\varvec{\varepsilon }= (\varepsilon _j^i)_{(i,j)\in M}\in [0,c]^{M}\) and \(j\in [N]\) we will define a vector \(\mathbf {p}_j(\varvec{\varepsilon })\in Y\) such that

$$\begin{aligned} p_j(\varepsilon )^i&= -1 \qquad \text {if} \,\, i=j,\end{aligned}$$
(10)
$$\begin{aligned} p_j(\varepsilon )^i&= \varepsilon _j^i\qquad ~\text {if} \,\, i<j,\end{aligned}$$
(11)
$$\begin{aligned} p_j(\varepsilon )^i&= 0 \qquad ~~~\text {if} \,\, i\in [N]\setminus [j], \end{aligned}$$
(12)
$$\begin{aligned} |p_j(\varepsilon )^i|&\le \frac{1}{2} \qquad ~~\text {if} \,\, i\in [n]\setminus [N]. \end{aligned}$$
(13)

Conditions (10)–(13) imply that \(\Vert \mathbf {p}_s(\varvec{\varepsilon })-\mathbf {p}_t(\varvec{\varepsilon }) \Vert _{\infty }=1+\varepsilon _t^s\) for every \(1\le s <t \le N\). Define \(\varphi :[0,c]^{M}\rightarrow \mathbb {R}^{M}\) by

$$\begin{aligned} \varphi _j^i(\varvec{\varepsilon })=1+\varepsilon _j^i-\Vert \mathbf {p}_i(\varvec{\varepsilon }) -\mathbf {p}_j(\varvec{\varepsilon })\Vert _Y, \end{aligned}$$

for every \(1\le i <j \le N\). From

$$\begin{aligned} 0=1+\varepsilon _j^i-\Vert \mathbf {p}_i(\varvec{\varepsilon })-\mathbf {p}_j(\varvec{\varepsilon }) \Vert _{\infty }&\le \varphi _j^i(\varvec{\varepsilon })=1+\varepsilon _j^i-\Vert \mathbf {p}_i (\varvec{\varepsilon })-\mathbf {p}_j(\varvec{\varepsilon })\Vert _Y \\&\le 1+\varepsilon _j^i- (1+c)^{-1}\Vert \mathbf {p}_i(\varvec{\varepsilon })-\mathbf {p}_j (\varvec{\varepsilon })\Vert _{\infty }\le c, \end{aligned}$$

it follows that the image of \(\varphi \) is contained in \([0,c]^M\). Since \(\varphi \) is continuous, by Brouwer’s fixed point theorem, \(\varphi \) has a fixed point \(\varvec{\varepsilon }_0\in [0,c]^M\). Then \(\{\mathbf {p}_j(\varvec{\varepsilon }_0):j\in [N]\}\) is a 1-equilateral set in Y of cardinality \(N=n-k(2+\ell )+1\).

To finish the proof, we only have to find vectors \(\mathbf {p}_j(\varvec{\varepsilon })\) that satisfy conditions (10)–(13). We construct them in a similar way as the equilateral sets in the proof of Theorem 1.2. Select \(2+\ell \) disjoint submatrices \(B_1,\dots ,B_{2\ell }\) of A of size \(k\times k\), such that for every \(1\le m \le k\) and \(1\le i \le \ell +2\), and for any column \(\mathbf {b}_r\) that is not a column of any \(B_j\), we have

$$\begin{aligned} |{\det B_i}| \ge |{\det B_i(m,\mathbf {b}_r)}|. \end{aligned}$$
(14)

This we may do by choosing the \(B_i\)’s after each other, always choosing the submatrix with the largest determinant disjoint from the previous submatrices. Using Lemma 3.1 and the upper semi-continuity of the equilateral number, we may further assume that \(|{\det B_{i}}|>0\) for all \(1\le i \le \ell +2\). Let \(U_i\) be the set of indices of the columns of \(B_i\), and let \(N_2=U_1\cup \cdots \cup U_{\ell +2}\). By permuting the coordinates, we may assume that \(N_2\cap [N]=\emptyset \). Indeed, permuting the coordinates gives a subspace that is linearly isometric to the initial one, and the equilateral number is the same for isometric normed spaces.

We construct \(\mathbf {p}_j(\varvec{\varepsilon })\) as a sum of \(2+\ell \) other vectors \(\mathbf {p}_j(\varvec{\varepsilon },\!1),\!\dots \!, \mathbf {p}_j(\varvec{\varepsilon },2+\ell )\). For \(1\le j \le N\) we define \(\mathbf {p}_j(\varvec{\varepsilon },q)\) as follows. For \(q\in \{1,2\}\) let

$$\begin{aligned} p_j(\varvec{\varepsilon },q)^i= {\left\{ \begin{array}{ll} \displaystyle -\frac{1}{2}&{}\quad \text {if}~ i=j,\\ 0 &{}\quad \text {if}~ i\in [n]\setminus (\{j\}\cup U_q),\\ \displaystyle \frac{\det B_q(m,{\mathbf{b}_{j}}/{2})}{\det B_q} &{}\quad \text {if}~ i ~\text {is the}~ m\text {-th element of}~ U_q. \end{array}\right. } \end{aligned}$$

Let \(\mathbf {s}(\varvec{\varepsilon },j)=\sum _{r<j}{\varepsilon _j^r}\mathbf {b}_r/{\ell }\) and for \(q\in \{3,\dots , 2+\ell \}\) let

$$\begin{aligned} p_j(\varvec{\varepsilon },q)^i= {\left\{ \begin{array}{ll} \displaystyle \frac{\varepsilon _j^i}{\ell } &{}\quad \text {if}~ i<j,\\ 0 &{}\quad \text {if}~ i\in [n]\setminus ([j-1]\cup U_q),\\ \displaystyle \frac{\det B_q (m,-\mathbf {s}(\varvec{\varepsilon },j))}{\det B_q} &{}\quad \text {if}~ i~ \text {is the}~ m\text {-th element of}~ U_q. \end{array}\right. } \end{aligned}$$

As before, by Cramer’s rule we obtain that \(\mathbf {p}_j({\varvec{\varepsilon }},q)\) is contained in Y for every q and j, thus \(\mathbf {p}_j(\varvec{\varepsilon })=\sum _{q\in [2+\ell ]}\mathbf {p}_j(\varvec{\varepsilon },q)\) is also contained in Y. It follows immediately that \(\mathbf {p}_j(\varvec{\varepsilon })\) satisfies conditions (10)–(12). It only remains to check condition (13).

By the multilinearity of the determinant, (14), and the triangle inequality, for every j, for \(q\in \{1,2\}\) and \(i\in U_q\) we have

$$\begin{aligned} |p_j(\varvec{\varepsilon },q)^i| = \biggl | \frac{\det B_q(m, {\mathbf{b}_{j}}/{2})}{\det B_q}\biggr |\le \frac{1}{2}. \end{aligned}$$

For every j, for \(q\in \{3,\dots ,2+\ell \}\) and \(i\in U_q\) we have

$$\begin{aligned} |p_j(\varvec{\varepsilon },q)^i|&= \biggl |\frac{\det B_q (m,-\mathbf {s} (\varvec{\varepsilon },j))}{\det B_m} \biggr | \le \Biggl |\sum _{r<j} \frac{\varepsilon _j^r}{\ell }\cdot \frac{\det B_q (m, -\mathbf {b}_r )}{\det B_q}\Biggr |\\&\le \sum _{r<j} \frac{\varepsilon _j^r}{\ell }\cdot \biggl | \frac{\det B_q(m, -\mathbf {b}_r)}{\det B_q}\biggr |\le \sum _{r<j} \frac{\varepsilon _j^r}{\ell }\le (N-1)\frac{c}{\ell }=\frac{1}{2}. \end{aligned}$$

The above bounds on \(|p_j(\varvec{\varepsilon },q)^i|\) imply that condition (13) holds for \(\mathbf {p}_j(\varvec{\varepsilon })\), finishing the proof.

5 Concluding Remarks

In Theorem 1.4 we did not make an attempt to find the most optimal bound on the equilateral number in terms of the Banach–Mazur distance that can be achieved with this approach. Our goal with this theorem was to illustrate how to find non-trivial lower bounds on the equilateral number depending on the codimension and the Banach–Mazur distance, while keeping the proof short.

One possible immediate improvement is to increase the lower bound by 1 without changing the bound on the Banach–Mazur distance. This can be done by adding a vector to the construction of the form \(\mathbf {p}_{N+1}(\varepsilon )=(\varepsilon _{N+1}^1,\dots ,\varepsilon _{N+1}^{N}, 0,\dots ,0,\le {1}/{2},\dots ,\le {1}/{2})\). As this improvement is negligible for large k, we decided not to include it in the proof to keep it simpler.

With this plus one improvement, the lower bound on e(Y) for \(k=0\) would match the bound from [10]. However, the bound on the Banach–Mazur distance in the assumption would be weaker. To explain this, note that our proof consists of two parts. In the first part we find possible candidates for points of a large equilateral set. This part is very similar to the proof in [10]. In the second part we have to fit these candidates in a subspace, and with doing this we lose some flexibility with the Banach–Mazur distance. For \(k=0\), however, we would not have a second part, thus the proof would be exactly the same as in [10].

Note also that for \(k=1\), \(\ell ={n}/{2}+1\) our lower bound in Theorem 1.4 is the same as the bound in [5], but again with a worse bound on the Banach–Mazur distance. The reason for this is that our intention was to present a non-trivial general bound that can be proven in a short way in various settings of the parameters.

Finally, it would be interesting to find some applications of Theorem 1.4 (or of a more optimized form of it). One corollary which was pointed out by an anonymous referee is that if Y is a subspace of \(\ell _{p}^n\) of codimension k for \(p\ge 2\), then \(e(Y)\ge \Omega _{k,p}(n^{1-1/p})\), since \(d_{\text {BM}}(\ell _{\infty }^n,\ell _{p}^n)=n^{1/p}\).