1 Introduction

Assume that the function f belongs to a given k-parameter explicit family of functions. Can we recover f by looking at k (or, at least, not many more than k) of its Fourier coefficients? This situation is often called “a signal with a finite rate of innovation” in the engineering literature [1]. We are particularly interested in families where the dependence on the parameters is non-linear.

The recovery of a function by not too many of its Fourier coefficients may be viewed as a problem in the general field of sparse representation (see, e.g., [2]). The specific rules that apply to this paper are the following:

  • We consider functions or measures on the torus \({{\mathbb {T}}}={{\mathbb {R}}}/{{\mathbb {Z}}}\) or \({{\mathbb {T}}}^d\), \(d\in {{\mathbb {N}}}\). The classes of functions we examine have a finite number of degrees of freedom whose number is constrained by the parameter N. For instance we might consider sums of point masses on \({{\mathbb {T}}}\) with the number of points being at most N.

  • We seek an a priori known finite set \(\Omega = \Omega _N\) of Fourier coefficients (a subset of \({{\mathbb {Z}}}\) or \({{\mathbb {Z}}}^d\)) which are assumed to be known for our class of functions. The set \(\Omega \) is allowed to depend on N and on nothing else.

  • The aim is to show that the mapping \(f \rightarrow \widehat{f}\restriction _\Omega \) is one to one. Though our proofs can often be turned into algorithms for the recovery of f we do not concern ourselves with matters of numerical stability or efficiency.

We deal with two problems in this paper.

  1. (1)

    Intervals in \({{\mathbb {T}}}\). Our function f is the indicator function of the union of at most N open intervals on \({{\mathbb {T}}}\). Courtney [3] has shown that such a function is determined by its Fourier coefficients at the locations \(0, 1, \ldots , N\). We give a new proof of this fact which is completely elementary (Courtney’s proof uses Blaschke products and conformal mapping). We further discuss the problem of whether functions of this class are determined by different sets of Fourier coefficients. We need to emphasize here that, apart from Courtney [3] this problem has not been considered elsewhere. There are papers [1] where one recovers a function on \({{\mathbb {T}}}\) which is piecewise constant (or even piecewise a polynomial) with the correct number of samples (roughly equal to the number of degrees of freedom) but all these methods fail to take into account the fact that the function only takes two values (0 or 1) and will accordingly use a number of samples that is larger than the minimum by at least \(\mathrm{const.} N\) samples. The “extreme” nonlinearity of this problem (not only in allowing variable nodes in the decomposition of \({{\mathbb {T}}}\), but also in the values of the function) does not seem to make it amenable to the usual methods such as Prony’s method, if one wants to use the minimal number of samples (or close to the minimal). This we do in §2.

  2. (2)

    Point masses in \({{\mathbb {T}}}^d\). We examine the class of measures which are sums of at most N complex point masses on \({{\mathbb {T}}}^d\). Using the corresponding question in dimension 1 (solvable with the so-called Prony’s method) we show an explicit set of locations \(\Omega \) such that the Fourier coefficients on \(\Omega \) determine the measure. This set \(\Omega \) is of size \(O(N \log ^{d-1} N)\). We believe it is the first such set given of size \(o(N^2)\), though several other methods have been described for this problem under additional assumptions on the locations of the point masses [4,5,6,7,8,9,10,11,12]. We emphasize that the set \(\Omega \) depends only on N and is not determined on the fly by looking at the Fourier coefficients of the measure. This we show in §3 where we also give a set \(\Omega \) of size \(O(k^{d-1} N)\) when we assume that the set of point masses has at most k points with the same x-coordinate. In §3 we also describe a general connection of this problem with the problem of interpolation.

2 At most N intervals on \({{\mathbb {T}}}\)

2.1 Determination from the Fourier coefficients at \(0, 1, \ldots , N\)

We consider sets \(E \subset {{\mathbb {T}}}\) of the form

$$\begin{aligned} E = \bigcup _{j=1}^k (a_j, b_j) \end{aligned}$$

where \(k \le N\) and the open intervals \((a_j, b_j)\) are disjoint. We show that \(f = \textbf{1}_E\) is determined by the complex data

$$\begin{aligned} \widehat{f}(0), \widehat{f}(1),\ldots ,\widehat{f}(N). \end{aligned}$$

The family has \(\le 2N\) real degrees of freedom and the data has \(2N+1\), since \(\widehat{f}(0)\) is always real.

Several similar problems with functions supported on intervals are treated in [4].

Theorem 1

Suppose that the sets \(E,E' \subseteq {{\mathbb {T}}}\) are both unions of at most N open arcs and that \(\widehat{\chi _E}(\nu ) = \widehat{\chi _{E'}}(\nu )\) for \(\nu =0,1,\ldots ,N\). Then \(E = E'\).

Proof

For \(x=(x_1,\ldots ,x_n)\) let \(\sigma _k(x)\) denote the k-th elementary symmetric function of the variables \(x_i\) and let \(s_k(x) = \sum _{j=1}^n x_j^k\) denote the k-th power sum of the \(x_i\).

We use the Newton-Girard formulas

$$\begin{aligned} k \cdot \sigma _k(x) = \sum _{i=1}^k (-1)^{i-1} \sigma _{k-i}(x) s_i(x),\ \ (k \ge 1). \end{aligned}$$
(1)

Note that \(\sigma _0(x)=1\). What is important about these formulas is that if we know \(s_1,\ldots ,s_\nu \) then we know also the numbers \(\sigma _1,\ldots ,\sigma _\nu \), for all \(\nu \ge 1\). The precise dependence is irrelevant for our purposes.

Suppose N is given and that the sets \(E = \bigcup _{j=1}^n I_j\) and \(E' = \bigcup _{j=1}^{n'} I_j'\) (with \(n, n' \le N\) and the \(I_j, I_j'\) being arcs) have the same Fourier coefficients of order up to N:

$$\begin{aligned} \widehat{\chi _E}(\nu ) = \widehat{\chi _{E'}}(\nu ),\ \ \nu =0,1,2,\ldots ,N. \end{aligned}$$

If \(I_j=(a_j, b_j)\) and \(I_j'=(a_j', b_j')\) then, differentiating the functions \(\chi _E, \chi _{E'}\), we obtain that the measures

$$\begin{aligned} \mu = \sum _{j=1}^n \delta _{a_j} - \delta _{b_j},\ \ \mu ' = \sum _{j=1}^{n'} \delta _{a_j'} - \delta _{b_j'}, \end{aligned}$$

have the same Fourier coefficients of order up to N. Writing \(z_j = e^{-2\pi i a_j}, w_j = e^{-2\pi i b_j}, z_j' = e^{-2\pi i a_j'}, w_j' = e^{-2\pi i b_j'}\) we obtain the relations

$$\begin{aligned} \sum _{j=1}^n z_j^\nu - w_j^\nu = \sum _{j=1}^{n'} z_j'^{\nu }-w_j'^{\nu },\ \ \ \nu =1,2,\ldots ,N. \end{aligned}$$

From this we get \(s_\nu (z,w') = s_\nu (z',w)\), for \(\nu =1,2,\ldots ,N\), where

$$\begin{aligned} s_\nu (z,w') := \sum _{j=1}^n z_j^\nu + \sum _{j=1}^{n'} w_j'^\nu , \end{aligned}$$

and

$$\begin{aligned} s_\nu (z',w) = \sum _{j=1}^{n'}z_j'^{\nu } + \sum _{j=1}^n w_j^\nu . \end{aligned}$$

By the Newton-Girard formulas the numbers \(s_1,\ldots ,s_N\) determine the numbers \(\sigma _1,\ldots ,\sigma _N\), hence we have

(2)

Write \(M=n+n'\le 2N\) and observe that

$$\begin{aligned} \sigma _M(z,w') = \sigma _M(z',w), \end{aligned}$$

i.e., \(\prod z_j w_j^{-1} = \prod z_j' w_j'^{-1}\). This comes from the fact that the total length of E and \(E'\) is the same, as testified by \(\widehat{\chi _E}(0) = \widehat{\chi _{E'}}(0)\).

We now use the fact that \({\left| {z_j}\right| }={\left| {w_j}\right| }={\left| {z_j'}\right| }={\left| {w_j'}\right| }=1\):

$$\begin{aligned} \overline{\sigma _k(z,w')} = \sigma _k\left( \frac{1}{z},\frac{1}{w'}\right) = \frac{\sigma _{M-k}(z,w')}{\sigma _M(z,w')},\ \ \ k=0,1,\ldots ,M, \end{aligned}$$
(3)

and similarly for the elementary symmetric functions of the vector \((z', w)\). For \(k=1,2,\ldots ,M-N-1\le N\) we obtain from (3) the missing values of \(\sigma _\nu \) for \(\nu =N+1,\ldots ,M-1\).

We have proved that

$$\begin{aligned} \sigma _\nu (z,w') = \sigma _\nu (z',w),\ \ \ \nu =0,1,2,\ldots ,M, \end{aligned}$$

hence the multisets \({\left\{ {z_j, w_j'}\right\} }\) and \({\left\{ {z_j', w_j}\right\} }\) are equal, since the elementary symmetric functions determine the polynomials p and q with roots at \({\left\{ {z,w'}\right\} }\) and \({\left\{ {z',w}\right\} }\) respectively and they are equal. But \({\left\{ {z_j}\right\} } \cap {\left\{ {w_j}\right\} } = {\left\{ {z_j'}\right\} } \cap {\left\{ {w_j'}\right\} } = \emptyset \) so the only possibility is that \({\left\{ {z_j}\right\} } = {\left\{ {z_j'}\right\} }\) and \({\left\{ {w_j}\right\} } = {\left\{ {w_j'}\right\} }\), as we had to show. \(\square \)

Remark 1

Of course it is also possible to solve this problem using Prony’s method, which we will introduce in Sect. 3.1. However, Prony’s method will require more samples, as it cannot exploit the fact that the coefficients of \(\hat{\chi _E}\) alternate between plus and minus one.

Remark 2

The determination of the elementary symmetric functions \(\sigma _\nu \) from the power sums \(s_\nu \) (essentially the equal coefficient case for Prony’s problem) has also recently been pointed out in the PhD thesis of M. Wageringel [13, §1.4], in the form of determining the numbers \(\xi _1, \ldots , \xi _r\) from the power sums

$$\begin{aligned} m_k = \xi _1^k + \cdots + \xi _r^k,\ \ k=0, 1, \ldots , r, \end{aligned}$$

thus also using about half as many power sums than what Prony’s method alone would require.

2.2 Determination from other sets of Fourier coefficients

The problem is sensitive to the choice of which Fourier coefficients to use in order to determine the set, even in the case of one interval \(E=(a,b)\). In this case, we have

$$\begin{aligned} 2\pi i\nu \widehat{\chi _E}(\nu )=e^{-2\pi i\nu a}-e^{-2\pi i\nu b}:=z^{\nu }-w^{\nu },\; \nu \ne 0. \end{aligned}$$
(4)

(Here, again, \(z = e^{-2\pi i a}\), \(w = e^{-2\pi i b}\).)

A single Fourier coefficient is not enough to determine the interval uniquely: for \(\nu =0\) this is obvious as \(\widehat{\chi _E}(0)=b-a\). For \(\nu =1\), consider \(E'=(a',b')\) such that \(z'=-w\), \(w'=-z\) (but this is the only other option). Last, for \(\nu \ge 2\), we may take \(E=(0, \frac{1}{\nu })\) and \(E'=(a', a'+\frac{1}{\nu })\), for some \(a'\ne 0\).

For two coefficients, apart from the case when \(\widehat{\chi _E}(0), \widehat{\chi _E}(1)\) are known, thus defining E uniquely, it is also easy to see that knowing \(\widehat{\chi _E}(1), \widehat{\chi _E}(2)\) also determines E. No other combination of two Fourier coefficients \(\widehat{\chi _E}(m), \widehat{\chi _E}(n)\) determines the set (we skip the details).

For \(N\ge 2\), let us point out that even equality of all Fourier coefficients of two sets

$$\begin{aligned} E=\bigcup _{j=1}^N (a_j, b_j),\ \ \ E'=\bigcup _{j=1}^N(a_j',b_j') \end{aligned}$$

at \(1, 2, \ldots 2N-1\), is not enough to conclude \(E=E'\), while, by Theorem 1, the Fourier coefficients from 0 to N suffice.

To prove the claim, let us begin with a simple observation. Let \(x=(x_1, \ldots , x_{2N})\). Then,

$$\begin{aligned} s_1(x)=\ldots s_{2N-1}(x)=0 \iff \sigma _{1}(x)=\ldots \sigma _{2N-1}(x)=0. \end{aligned}$$
(5)

(We keep the notation for power sums and elementary symmetric functions that was used in the proof of Theorem 1.) The fact that vanishing power sums give vanishing elementary symmetric functions follows directly from the Newton-Girard formulas (1). For the converse, observe that always \(s_1(x)= \sigma _{1}(x)\) and then apply (1) consecutively for \(k=2, \dots , 2N-1\) (or observe that the power sums are themselves symmetric functions, hence they can be expressed via the elementary symmetric functions).

Fix some \(\theta \in (0, \pi /N)\) and consider two regular 2N-gons \({\mathcal {P}}\) and \({\mathcal {Q}}\) on the unit circle, with vertices arranged counterclockwise, defined by

$$\begin{aligned} V_{{\mathcal {P}}} = {\left\{ {z_1, \ldots , z_N, w_1',\ldots , w_N'}\right\} } \end{aligned}$$

and

$$\begin{aligned} V_{{\mathcal {Q}}} = {\left\{ {w_1,\ldots , w_N, z_2', \ldots , z_N', z_1'}\right\} } \end{aligned}$$

(see Fig. 1, where \(V_{{\mathcal {P}}}\) is blue and \(V_{{\mathcal {Q}}}\) is red) where

$$\begin{aligned} z_1 = 1,\ w_1 = e^{i\theta }. \end{aligned}$$
Fig. 1
figure 1

One possible selection of arcs \(\bigcup \nolimits _{j=1}^N(z_j, w_j)\) and \(\bigcup \nolimits _{j=1}^N(z_j', w_j')\) with the same Fourier coefficients of order \(1, 2, \ldots , 2N-1\) (shown for \(N=5\))

It follows that the numbers in \(V_{{\mathcal {P}}}\) are the roots of the polynomial \(z^{2N}-1\) and that the numbers in \(V_{{\mathcal {Q}}}\) are the roots of the polynomial \(z^{2N}-e^{i2N\theta }\). Since the elementary symmetric functions of the roots of a polynomial are the coefficients of the polynomial, it follows that the numbers \(\sigma _{\nu }(z,w')\), \(\sigma _{\nu }(z',w)\) vanish for \(\nu =1, \dots , 2N-1\). By (5), we also have \(s_{\nu }(z,w')=s_{\nu }(z',w) = 0\) for \(\nu =1, \dots , 2N-1\). Then, we have \(\widehat{\chi _E}(\nu ) = \widehat{\chi _{E'}}(\nu )\) for all \(\nu =1, \dots , 2N-1\), however the sets E, \(E'\) do not coincide, as implied by the arrangement of \(V_{{\mathcal {P}}}\), \(V_{{\mathcal {Q}}}\).

It is interesting to see that many more examples are possible with the points \(z_1, \ldots , z_N, w_1', \ldots , w_N'\) located at the vertices of a regular 2N-gon and the points \(z_1', \ldots , z_N', w_1, \ldots , w_N\) located at the vertices of a rotated regular 2N-gon, but not necessarily in the order shown in Fig. 1. As explained above these locations guarantee that the two sets \(\bigcup _{j=1}^N(z_j, w_j)\) and \(\bigcup _{j=1}^N(z_j', w_j')\) (with an obvious and excusable abuse of notation) have the same Fourier coefficients of order \(1, 2, \ldots , 2N-1\) and they are of course not equal.

One needs to find the arrangements of the points \(z_j, w_j, z_j', w_j'\) on the vertices of these two polygons so that the following rules are satisfied:

  1. (1)

    The points \(z_1, \ldots , z_N, w_1, \ldots , w_N\) appear on the circle in the counterclockwise order. Same for the points \(z_1', \ldots , z_N', w_1', \ldots , w_N'\). This rule ensures that the arcs \((z_i, w_i)\) are non-overlapping and the same for the arcs \((z_i', w_i')\).

  2. (2)

    The \(z_j\) and \(w_j'\) are blue (polygon \({{\mathcal {P}}}\)) and the \(z_j'\) and \(w_j\) are red (polygon \({{\mathcal {Q}}}\)).

We can enumerate these arrangements by viewing this problem as a variant of the so-called Terquem’s problem (see, e.g., [14, Problem 30 on p. 120, and solution on p. 170]). Terquem’s problem asks in how many ways we can choose a sequence \(a_1< a_2< \ldots < a_k\) from the set \({\left\{ {1, 2, \ldots , n}\right\} }\) whose elements alternate between odd and even. By viewing blue as odd and red as even on our polygons we see that we can restate our problem as follows:

In how many ways can we select an alternating sequence \(a_1< a_2< \ldots < a_{2N}\) from the set \({\left\{ {1, 2, \ldots , 4N}\right\} }\) such that its complement is also alternating.

The numbers \(a_1, \ldots , a_{2N}\) correspond to the choices for the labels \(z_1, w_1, \ldots , z_N, w_N\) and the complementary set corresponds to the labels \(z_1', w_1', \ldots , z_N', w_N'\).

Following [14, p. 170] we can encode the sequence \(a_i\) via the sequence \(b_i\) defined by

$$\begin{aligned} b_i = a_i - i + 1. \end{aligned}$$

This sequence is increasing

$$\begin{aligned} b_1 \le b_2 \le \ldots \le b_{2N} \end{aligned}$$

and gives back the sequence \(a_i\) as \(a_i = b_i + i -1\) (which is strictly increasing). It also satisfies the bounds

$$\begin{aligned} 1 \le b_i \le 2N+1. \end{aligned}$$

The alternating property of the sequence \(a_i\) translates exactly to the \(b_i\) being all odd. If we did not care about the complement of the sequence \(a_i\) being also alternating then, as explained in [14], all we would have to do is select with replacement the 2N numbers \(b_i\) among the odd numbers of the set \({\left\{ {1, 2, \ldots , 2N+1}\right\} }\), that is from \(N+1\) numbers. To ensure that the complement is also an alternating sequence it is necessary and sufficient to ensure that the intervals (consecutive values) defined by the sequence \(a_i\) are all of even length. But an interval of the \(a_i\) translates into an interval of constancy for the corresponding \(b_i\). Summarizing, the \(b_i\) must be odd and be selected an even number of times each. To achieve this we select with replacement N numbers from the odd numbers of the set \({\left\{ {1, 2, \ldots , 2N+1}\right\} }\) and then double the number of times each selection appears. This enumerates the \(b_i\) and therefore also the \(a_i\). We omit the details.

3 Point masses on \({{\mathbb {T}}}^d\)

3.1 Point masses on \({{\mathbb {T}}}\)

The one-dimensional problem has a very long history, going back to Gaspard de Prony’s work [15] from 1795. Since then, many solutions have been proposed, like Pisarenko’s method [16], MUSIC [17] or ESPRIT [18]. Still, there is ongoing research on further improvements, see [19] for a recent approach.

We show here another approach from the Electrical Engineering literature (see e.g. [1]) with so-called annihilation filters. We will make use of Theorem 2 repeatedly when solving the same problem on \({{\mathbb {T}}}^d\).

Theorem 2

Suppose \(\mu \) is a measure on \({{\mathbb {T}}}\) which is a sum of at most N complex point masses. Then \(\mu \) is determined by the data

$$\begin{aligned} \widehat{\mu }(j),\ \ \ j=-N+1, -N+2, \ldots , N. \end{aligned}$$
(6)

Proof

Suppose that \(\mu =\sum _{j=1}^K c_j \delta _{\theta _j}\), with \(c_j \in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} }\), the \(\theta _j\) all different and \(K \le N\). It follows that

$$\begin{aligned} \widehat{\mu }(n) = \sum _{j=1}^K c_j \rho _j^{-n},\ \ \ (\rho _j=e^{2\pi i \theta _j},\ n \in {{\mathbb {Z}}}). \end{aligned}$$
(7)

Define the polynomial

$$\begin{aligned} a(z) = \prod _{j=1}^K (z-\rho _j) = \widehat{a}(0) + \widehat{a}(1)z + \ldots + \widehat{a}(K-1)z^{K-1}+z^K, \end{aligned}$$

and note that \(\widehat{a}(n)\) also denotes its Fourier coefficients when viewed as a function on \({{\mathbb {T}}}\). Since \(a\mu = 0\) it follows that

$$\begin{aligned} \widehat{a} * \widehat{\mu }(n) = 0,\ \ \ (n \in {{\mathbb {Z}}}). \end{aligned}$$
(8)

We now show that the polynomial a(z) is determined up to constant multiples by the conditions

$$\begin{aligned} \deg {a}\le & {} K \end{aligned}$$
(9)
$$\begin{aligned} \widehat{a} * \widehat{\mu }(n)= & {} 0,\ \ \ (n \in [1, K]). \end{aligned}$$
(10)

It is enough to show that (9) and (10) together imply that a(z) vanishes on the \(\rho _j\) (this is the same as (8)), whose number is K, the same as the degree of a.

Substituting (7) in (10) we get for \(l=1,2,\ldots ,K\)

$$\begin{aligned} 0= & {} \sum _{k=0}^K \widehat{a}(k) \widehat{\mu }(l-k),\\= & {} \sum _{k=0}^K \widehat{a}(k) \sum _{j=1}^K c_j \rho _j^{k-l},\\= & {} \sum _{j=1}^K c_j \rho _j^{-l} \sum _{k=0}^K \widehat{a}(k) \rho _j^k,\\= & {} \sum _{j=1}^K c_j \rho _j^{-l} a(\rho _j). \end{aligned}$$

Observe that the \(K\times K\) Vandermonde matrix \(\rho _j^{-\nu }\), \(j,\nu =1,\ldots ,K\), is nonsingular and, therefore, all \(c_j a(\rho _j)\) are 0, \(j=1,2,\ldots ,K\). Since all \(c_j\) are nonzero this implies that

$$\begin{aligned} a(\rho _j)=0,\ \ \ j=1,2,\ldots ,K. \end{aligned}$$

Suppose \(\mu _1, \mu _2\) are two measures with the same Fourier data (6):

$$\begin{aligned} \mu _1 = \sum _{j=1}^{K_1} c_{1,j} \delta _{\theta _{1,j}},\ \ \mu _2 = \sum _{j=1}^{K_2} c_{2,j} \delta _{\theta _{2,j}}, \ K_1 \le K_2 \le N, \end{aligned}$$

and

$$\begin{aligned} \widehat{\mu _1}(n) = \widehat{\mu _2}(n),\ \ (n=-N+1,\ldots ,N-1,N). \end{aligned}$$
(11)

We are assuming that all \(\theta _{1,j}\) are distinct and so are all \(\theta _{2,j}\), and that \(c_{1,j}, c_{2,j} \in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} }\). Write also \(\rho _{1,j}=e^{2\pi i \theta _{1,j}}\) and \(\rho _{2,j}=e^{2\pi i \theta _{2,j}}\).

Write

$$\begin{aligned} a_1(z) = \prod _{j=1}^{K_1} (z-\rho _{1,j}),\ \ a_2(z) = \prod _{j=1}^{K_2} (z-\rho _{2,j}). \end{aligned}$$

We have \(\widehat{a_1}*\widehat{\mu _1}(n) = \widehat{a_2}*\widehat{\mu _2}(n) = 0\) for all n but we also have \(\widehat{a_1}*\widehat{\mu _2}(n) = 0\) for \(n=1,\ldots ,N\), because of (11). Applying the fact that a(z) is determined by (9) and (10) with \(K=K_2\) we obtain that \(a_1(z)=a_2(z)\) hence \(K_1=K_2\) and \({\left\{ {\rho _{1,j}}\right\} } = {\left\{ {\rho _{2,j}}\right\} }\).

It remains to show that the linear map

$$\begin{aligned} (c_1,\ldots ,c_K) \rightarrow (\widehat{\mu }(1),\ldots ,\widehat{\mu }(K)) \end{aligned}$$

is injective. This map is given by (7) and it is easily seen to be nonsingular as its determinant is a multiple of the Vandermonde determinant. \(\square \)

Remark 3

The number of samples is sharp, as we recover the 2N parameters using 2N samples. However, note that we use complex samples, to recover N complex parameters and N parameters in \({{\mathbb {T}}}\). One can easily check that the proof extends to the case of point measures on \({{\mathbb {T}}}+ i{{\mathbb {R}}}\), where (7) becomes

$$\begin{aligned} \widehat{\mu }(n) = \sum _{j=1}^K c_j \rho _j^{-n},\ \ \ (\rho _j=e^{2\pi i (\theta _j+i\xi _j)},\ n \in {{\mathbb {Z}}}). \end{aligned}$$

In case of real coefficients and point measures on \({{\mathbb {T}}}\) one can utilize \(\overline{{\hat{\mu }}(-k)} = {\hat{\mu }}(k)\) to use only \({\hat{\mu }}(k),~k=0,\ldots , N\). This observation is an important part of the unitary ESPRIT algorithm [20], which uses only real-valued computation to solve the problem.

3.2 Connections to interpolation

Suppose \(\mu = \sum _{j=1}^N c_j \delta _{u_j}\), where \(u_j =(u_{j1},\ldots ,u_{jd}) \in {{\mathbb {T}}}^d\) are distinct points, all \(c_j\) are non-zero and \(d>1\). Can we recover \(\mu \) from a number of Fourier coefficients that is close to the number of degrees of freedom (in this case \((d+1)N\) or \((d+2)N\) depending on whether \(c_j \in {{\mathbb {R}}}\) or \(c_j \in {{\mathbb {C}}}\))?

Here the existing results do not seem to be final. In the special case where all \(u_{j1}\) are different (or equivalently all \(u_{jk}\) for a fixed \(k=1,\ldots ,d\)) the problem is solved using the one-dimensional theory with O(dN) Fourier coefficients, namely by using the Fourier coefficients at the locations \((m, \epsilon _2,\dots ,\epsilon _d)\) with \(m=0, 1, \ldots , N\) and \(\epsilon _j\in {\left\{ {0,1}\right\} }\), assuming all \(u_{j1}\) are different (we generalize this in Theorem 8). In the general case and without imposing any restrictions on the locations \(u_j\), it has only been known until this work how to recover \(\mu \) using \(O(N^2)\) Fourier coefficients [5, Section iii.C] in the two dimensional case and using \(O(N^2\log ^{2d-2}N)\) coefficients in the general case [12].

If one allows for the collection of Fourier coefficients used to depend on the data then one can recover the parameters with O(N) Fourier coefficients, see [4] for the case \(d=2\) and [9] for arbitrary d. It was conjectured in [4] that recovery of the parameters in this problem with \(d=2\) is always possible with O(N) Fourier coefficients which do not depend on the data and are on four predetermined lines. This conjecture was disproved in [7] but the possibility remains that some more general set of O(N) Fourier coefficients suffices. (In Theorem 9 we show that \(O(N \log N)\) Fourier coefficients suffice.) For the general d dimensional case it was shown in [21] that taking a total of \(O(N^2)\) samples on scattered line allows for a reconstruction.

Definition 3.1

Suppose \(\Omega \subseteq {{\mathbb {Z}}}^d\) and \(k=1,2,3,\ldots \). We call \(\Omega \) \(\underline{k-\hbox {interpolating}}\) if whenever \(u_1,\ldots ,u_{\ell } \in {{\mathbb {T}}}^d\), \(\ell \le k\), are distinct and \(d_1, \ldots , d_{\ell } \in {{\mathbb {C}}}\) we can find coefficients \(c_\omega \), \(\omega \in \Omega \), such that

$$\begin{aligned} d_j = \sum _{\omega \in \Omega } c_\omega e^{2\pi i \omega \cdot u_j},\ \ j=1,2,\ldots ,\ell . \end{aligned}$$

We call \(\Omega \) \(\underline{k-\hbox {sufficient}}\) if we can recover any measure \(\mu = \sum _{j=1}^{\ell } c_j \delta _{u_j}\), \(\ell \le k\), (with unknown \(c_j\in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} }\), unknown and distinct \(u_j\in {{\mathbb {T}}}^d\)) from its Fourier coefficients at \(\Omega \)

$$\begin{aligned} \widehat{\mu }(\omega ) = \sum _{j=1}^\ell c_j e^{-2\pi i \omega \cdot u_j},\ \ \omega \in \Omega . \end{aligned}$$

The connection between the concepts of k-interpolation and k-sufficiency is the following.

Theorem 3

$$\begin{aligned} \Omega \ \textrm{is}\ (2N)-\textrm{interpolating}\ \Longrightarrow \Omega \ \mathrm{is \ N-sufficient}\ \Longrightarrow \Omega \ \mathrm{is\ N-interpolating}.\nonumber \\ \end{aligned}$$
(12)

Proof

To prove Theorem 3 let us first make the following remark, which says that if we can solve the problem of sufficiency with the locations fixed then we can also solve the problem with unknown (but fewer) locations.

Lemma 4

Suppose \(\Omega \subseteq {{\mathbb {Z}}}^d\) is such that the mapping

$$\begin{aligned} \mu \rightarrow (\widehat{\mu }(\omega ),\ \omega \in \Omega ) \end{aligned}$$
(13)

is injective on the set

$$\begin{aligned} {\left\{ {\mu = \sum _{j=1}^{k} c_j \delta _{u_j}:\ c_j \in {{\mathbb {C}}}}\right\} }, \end{aligned}$$
(14)

for any choice of \(k \le 2N\) and distinct points \(u_1,\ldots ,u_k \in {{\mathbb {T}}}^d\). Then the mapping (13) is injective also on the set

$$\begin{aligned} {\left\{ {\mu = \sum _{j=1}^{\ell } c_j \delta _{u_j}:\ \ell \le N, c_j \in {{\mathbb {C}}}, u_j \in {{\mathbb {T}}}^d, c_j\ne 0\ \mathrm{and\ the}\ u_j \ \mathrm{are\ distinct}}\right\} }.\nonumber \\ \end{aligned}$$
(15)

Suppose \(\mu =\sum _{j=1}^{\ell _1} c_j \delta _{u_j}\) and \(\nu =\sum _{j=1}^{\ell _2} d_j \delta _{v_j}\), \(\ell _1, \ell _2 \le N\), are two different measures with the same image under (13). Then \(\mu -\nu \) is a non-zero measure supported on \(\le 2N\) points and is mapped to 0 under (13). This conflicts with the injectivity of (13) on the set (14).

\(\square \)

For the mapping (13) on the set (14) to be injective it is necessary and sufficient that the matrix

$$\begin{aligned} \left( e^{2\pi i u_j\cdot \omega } \right) _{j=1,\ldots ,k,\ \omega \in \Omega } \end{aligned}$$
(16)

has rank k. Write \(e_\omega (x) = e^{2\pi i \omega \cdot x}\).

Lemma 5

The mapping (13) on the set (14) is injective if and only if the functions \(e_\omega \) form an interpolating set for the set \({\left\{ {u_1,u_2,\ldots ,u_k}\right\} }\), i.e. for any values \(d_j \in {{\mathbb {C}}}\) there is a \({{\mathbb {C}}}\)-linear combination of the \(e_\omega \) which takes the value \(d_j\) at \(u_j\).

Proof

The \(e_\omega \) are interpolating at the \(u_j\) if and only if the rank of the matrix in (16) is k. By the preceding remark this is equivalent to the mapping (13) on the set (14) being injective. \(\square \)

Let us complete the proof of Theorem 3. If \(\Omega \) is (2N)-interpolating and \(u_1,\ldots ,u_{2N}\) are distinct points in \({{\mathbb {T}}}^d\) it follows from Lemma 5 that the mapping \(\mu \rightarrow \widehat{\mu }|_\Omega \) is injective on the set (14) (with \(k=2N\)). From Lemma 4 it follows that it is also injective on the set (15), hence \(\Omega \) is N-sufficient.

If \(\Omega \) is N-sufficient and \(u_1,\ldots ,u_k \in {{\mathbb {T}}}^d\), \(k\le N\), are distinct points then the matrix (16) has rank k. Therefore the mapping (13) is injective (with the \(u_j\) fixed) and from Lemma 5 we get that the functions \(e_\omega (x)\) are interpolating, which is what it means for \(\Omega \) to be N-interpolating. \(\square \)

Now it becomes clear that the situation in dimension 2 is significantly harder than in dimension 1. The reason is that interpolation is harder. Indeed, in dimension 1 one can easily find a set of N functions the linear combinations of which can interpolate any data on any N points. One such example of functions are the monomials \(1, x, x^2, \ldots , x^{N-1}\) and another example are the functions \(1, e^{2\pi ix}, e^{2\pi i 2x}, \ldots , e^{2\pi i (N-1)x}\) (when all \(u_j \in {{\mathbb {T}}}\)).

Such a set of functions is called a Chebyshev or Haar system and it is well known and easy to prove that continuous Chebyshev systems do not exist except in dimension 1 [22]. Indeed, suppose that \(S \subseteq {{\mathbb {R}}}^2\) is an open set and the continuous functions \(f_j:S\rightarrow {{\mathbb {R}}}\), \(j=1,2,\ldots ,N\), are such that for any set of N distinct points \(u_j \in S\) we can find a linear combination of the \(f_j\) which interpolates any given \(\underline{\hbox {real}}\) data at the \(u_j\). This means that for any choice of the distinct points \(u_j\) the determinant of the matrix \(f_i(u_j)\), \(i,j=1,2,\ldots ,N\), is non-zero. Choose then the \(u_j\) to belong to an open disk in S and carry out a continuous movement of the points \(u_1\) and \(u_2\) so that they do not collide between themselves and with any of the other points and such that, at the end of the motion, the two points have exchanged their positions. The determinant of the matrix has changed sign and, since it has varied continuously during the motion, it follows that the determinant has vanished at some point during the exchange, a contradiction. It is proved in [22] that the existence of a (continuous) Chebyshev system on a set \(S \subseteq {{\mathbb {R}}}^d\) is only possible when S is homeomorphic to a closed subset of a circle.

This argument is strictly for the real case of course but it has been extended [23] to the case of complex functions: there is a complex Chebyshev system for domains in \({{\mathbb {C}}}\) but not for domains in \({{\mathbb {C}}}^2\) or in higher dimension. More specifically, it is proved in [23, 24] that a \(\underline{\hbox {complex}}\) continuous Chebyshev system exists on a locally connected set S if and only if S is homeomorphic to a closed subset of \({{\mathbb {R}}}^2\). This result allows us to prove that the situation in \({{\mathbb {T}}}^2\) is strictly worse than in \({{\mathbb {T}}}\).

Theorem 6

Let \(d>1\). Suppose \(\Omega \subseteq {{\mathbb {Z}}}^d\) is N-sufficient. Then \({\left| {\Omega }\right| } > N\).

Proof

If such an \(\Omega \) had size N then, according to Theorem 3, the corresponding set of characters \(e_\omega (x) = e^{2\pi i \omega \cdot x}\), \(\omega \in \Omega \), would be a continuous Chebyshev system on \({{\mathbb {T}}}^d\), According to [23, 24] this would make \({{\mathbb {T}}}^d\) embeddable into the plane, which it is not.

\(\square \)

Again using the connection to interpolation let us now give a new proof, different from the one given in [5, Section iii.C] for the case \(d=2\) and [6, 11] for general d, of the following fact.

Theorem 7

There is \(\Omega \subseteq {{\mathbb {Z}}}^d\) of size \(O(N^d)\) which is N-sufficient.

Proof

It is enough to produce a set \(\Omega \subseteq {{\mathbb {Z}}}^d\) of size \(O(N^d)\) such that the set of corresponding exponentials \(e^{2\pi i \omega \cdot x}\) is 2N-interpolating, i.e. its linear combinations can interpolate any values at any 2N distinct points in \({{\mathbb {T}}}^d\). We use the fact that for any set of 2N distinct points in \({{\mathbb {C}}}^d\) and any complex data on them there is a complex polynomial in d variables of degree at most \(2N-1\) (the degree of each monomial is the sum of the exponents of the variables) which interpolates the data. To see this observe that for any set of 2N distinct points \(x_1, \ldots , x_{2N} \in {{\mathbb {C}}}^d\) we can find a vector \(u \in {{\mathbb {C}}}^d\) such that the complex numbers \(t_i = v\cdot x_i\) are all different. Let now p be a one-variable polynomial of degree \(\le 2N-1\) which interpolates the given data on the points \(t_i\). Then \(q(x) = p(v\cdot x)\) is a two-variable polynomial that interpolates the given data on the points \(x_i\). The degree of q is no larger than \(2N-1\).

Take \(\Omega ={\left\{ {m\in {{\mathbb {N}}}_0^d: \sum _{j=1}^d m_j\le 2N-1}\right\} }\). Suppose \( u_1,\ldots ,u_{2N}\in {{\mathbb {T}}}^d \) are distinct and \(d_1,\ldots ,d_{2N} \in {{\mathbb {C}}}\). Let \(p(z)=\sum _{m}p_{m}z^m\) have degree at most \(2N-1\) and interpolate the data \(d_k\) at the points \((e^{2\pi i u_{j1}}, \ldots , e^{2\pi i u_{jd}}) \in {{\mathbb {C}}}^d\), \(j=1,2,\ldots ,2N\) (note that these are distinct points as the \(u_d\) are in \({{\mathbb {T}}}^d\) not in \({{\mathbb {R}}}^d\)). We have

$$\begin{aligned} d_j = p(e^{2\pi i u_{j1}}, \ldots , e^{2\pi i u_{jd}}) = \sum _{m\in \Omega } p_{m} e^{2\pi i u_j\cdot m}, \end{aligned}$$

which means that the functions \(e^{2\pi i \omega \cdot x}\), \(\omega \in \Omega \), are interpolating the arbitrary data \(d_j\) at the 2N arbitrary points \(u_j\), as we had to prove. \(\square \)

3.3 Small sufficient sets for \({{\mathbb {T}}}^d\)

Next we provide a case where the sufficient number of coefficients for N points of \({\mathbb {T}}^d\) is \(O(k^{d-1}N)\), for some \(1\le k\le N\).

Theorem 8

Let \(\mu = \sum _{j=1}^N c_j \delta _{u_j}\), where \(u_j =(x_j,y_j) \in {{\mathbb {T}}}\times {{\mathbb {T}}}^{d-1}\) are distinct points and all \(c_j \in {\mathbb {C}}\backslash \{0\}\). Assume that the number of points \(u_j\) that share the same x coordinate is at most k, for some \(1\le k\le N\). Then \(\mu \) can be recovered by \(O(k^{d-1}N)\) Fourier coefficients.

Proof

Write \(U = {\left\{ {u_j:\;j=1,\ldots ,N}\right\} }\) and \(X = {\left\{ {x_j:\; j=1,\ldots ,N}\right\} }\) for the set of distinct x that appear as first coordinates for the points in U. Notice that X may have fewer than N points.

Recall that

$$\begin{aligned} \widehat{\mu }(m, n)&= \sum \limits _{j=1}^{N}c_je^{-2\pi i (mx_j+n\cdot y_j)}\\&= \sum _{x \in X} \left( \sum _{y: \; (x, y) \in U} c_{(x, y)} e^{-2\pi i n \cdot y} \right) e^{-2\pi i m x}, \end{aligned}$$

for \(m\in {\mathbb {Z}}, \, n\in {{\mathbb {Z}}}^{d-1}\). For fixed n the numbers \(\widehat{\mu }(m, n)\) are the Fourier coefficients of a collection of point masses at the points of X (some of these point masses may be 0).

Consider the data

$$\begin{aligned} \widehat{\mu }(-N, \ell ),\; \ldots ,\; \widehat{\mu }(0, \ell ),\; \widehat{\mu }(1, \ell ),\; \ldots ,\; \widehat{\mu }(N, \ell ), \end{aligned}$$

where \(\ell \in {{\mathbb {Z}}}^{d-1}\) is fixed. By Theorem 2, we can recover the sums

$$\begin{aligned} S(x, \ell ) = \sum \limits _{y: \; (x, y) \in U} c_{(x, y)} e^{-2\pi i\ell \cdot y}, \quad \ell \in {{\mathbb {Z}}}^{d-1}. \end{aligned}$$
(17)

and the corresponding \(x \in X\). Notice that we only “see” the x for which \(S(x, \ell ) \ne 0\).

For fixed \(x \in X\), define the measure on \({{\mathbb {T}}}^{d-1}\)

$$\begin{aligned} \lambda _x = \sum \limits _{y: \; (x, y) \in U} c_{(x, y)} \delta _{y}, \end{aligned}$$

which is supported on at most k locations on \({{\mathbb {T}}}^{d-1}\). By Theorem 7, there is \(\Omega \subset {{\mathbb {Z}}}^{d-1}\) of size at most \(O(k^{d-1})\) that is k-sufficient. Thus, knowing the Fourier coefficients of \(\lambda _x\) at \(\ell \in \Omega \) is sufficient to recover the measure. Knowing these Fourier coefficients means precisely knowing the sums in (17), so we recover the points y sitting over each \(x \in X\) and the corresponding coefficients. The proof of the Theorem is complete.

\(\square \)

Finally we come to the main result of this section. One can view Theorem 9 as a more sophisticated version of Theorem 8, where the gain comes from being able to distinguish which \(x \in {{\mathbb {T}}}\) have many points projected onto them. This set of x cannot be large.

This theorem was first proved in [12], using techniques from computational algebra. We give an elementary proof.

Theorem 9

There is \(\Omega \subseteq {{\mathbb {Z}}}^d\) of size \({\left| {\Omega }\right| } \le C_d N \log ^{d-1} (N)\) which is N-interpolating. One such set is the positive octant of the hyperbolic cross

$$\begin{aligned} \Gamma _N^d = \left\{ n\in {{\mathbb {N}}}_0^d ~:~\prod _{j=1}^d\left( n_j + 1 \right) \le N\right\} . \end{aligned}$$

Proof

Let \(U={\left\{ {u_j:\; j=1,\ldots , N}\right\} }\). As noted in (16), it suffices to show that the vectors

$$\begin{aligned} v_\Gamma (u_j) = (e^{2\pi i u_j\cdot w})_{\omega \in \Gamma _N^d},\quad j=1,\ldots ,N \end{aligned}$$

are linearly independent. We use induction in d. For the base case \(d=1\), observe that the matrix \((e^{2\pi i u_j\omega })_{j=1, \ldots , N, \; \omega =0, \ldots , N-1}\) is Vandermonde. For \(d\ge 2\), assume on the contrary that there are \(c_u\in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} },\;u\in U\) (we can exclude nodes with zero coefficients) satisfying

$$\begin{aligned} \sum _{u\in U} c_{u} v_\Gamma (u) = 0. \end{aligned}$$

Let \(X={\left\{ {x_j\in {{\mathbb {T}}}^{d-1}:\; j=1,\ldots ,N}\right\} }\) be the set of distinct x that appear as the first \(d-1\) entries of points in U (again, X may have fewer than N points). As in the proof of Theorem 8, we note that the condition of linear dependence rewrites as

$$\begin{aligned} \sum _{j=1}^N c_{u_j} e^{-2\pi i (m\cdot x_j+ky_j)}&= \sum _{x \in X} \left( \sum _{y: \; (x, y) \in U} c_{(x, y)} e^ {-2\pi i k y} \right) e^{-2\pi i m \cdot x}\nonumber \\&=0, \end{aligned}$$
(18)

for \(m\in {{\mathbb {Z}}}^{d-1}, k\in {{\mathbb {Z}}}, \;(m,k)\in \Gamma _N^{d}\).

Observe that we have

$$\begin{aligned}&(m, 0) \in \Gamma _N^d \; \text { for } m\in \Gamma _{N}^{d-1} \text { and }\nonumber \\&(m, k-1) \in \Gamma _N^d \; \text { for } m\in \Gamma _{\lfloor N/k\rfloor }^{d-1}, \; k\in {{\mathbb {N}}}. \end{aligned}$$
(19)

Then, as (18) holds for \((m, 0) \in \Gamma _N^d\) for all \(m\in \Gamma _N^{d-1}\), we can use the induction hypothesis to conclude that

$$\begin{aligned} \sum _{y: \; (x, y) \in U} c_{(x, y)} e^{-2\pi i yk } = 0,\quad \text {for } k=0, \; \text {for all } x\in X. \end{aligned}$$
(20)

We partition X according to how many points of U project to each point:

$$\begin{aligned} X = X_1 \sqcup \ldots \sqcup X_r\ \ \ (r \le N), \end{aligned}$$

where \(X_t = {\left\{ {x \in X:\ {\left| {{\left\{ {y: (x, y) \in U}\right\} }}\right| } = t}\right\} }\). By (20), we see that

$$\begin{aligned} c_{(x,y)} = 0 \quad \text {for all } x\in X_1, \; (x,y)\in U, \end{aligned}$$

which contradicts \(c_u\in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} }\) and thus \(X_1=\emptyset \). Now we use the crucial observation that

$$\begin{aligned} |U| = \sum _{j=1}^r j|X_j| \end{aligned}$$
(21)

which implies with \(X_1=\emptyset \) that \(|X|\le \lfloor N/2\rfloor \).

Taking data at (19) for \(k=0, 1\), by the induction hypothesis we see that (20) holds true for \(k=0,1\). If \(X_2\) is not empty, then, for any \(x\in X_2\), the summands in (20) are exactly 2, for each \(k\in \{0,1\}\). Thus, we have a homogeneous \(2\times 2\) system, with a Vandermonde matrix of coefficients, so

$$\begin{aligned} c_{(x,y)} = 0 \quad \text {for all } x\in X_2, \; (x,y)\in U, \end{aligned}$$

a contradiction. That allows us to deduce \(X_2=\emptyset \), giving us \(|X|\le \lfloor N/3\rfloor \). Repeating the argument r times results in the contradiction \(X=\emptyset \). \(\square \)

We use Theorem 3 to see that \(O(N\log ^{d-1} N)\) samples are sufficient for unique determination of a point measure \(\mu \) of at most N peaks.

Corollary 10

There is \(\Omega \subseteq {{\mathbb {Z}}}^d\) of size \({\left| {\Omega }\right| } = O( N \log ^{d-1} (N))\) which is N-sufficient. One such set is \(\Gamma _{2N}^d\).

Note, however, that the proof cannot be converted in an algorithm recovering the measure from its Fourier samples. Next, we show that such an algorithm exists for a slightly larger sampling set.

Theorem 11

Any measure \(\mu = \sum _{j=1}^K c_j\delta _{u_j},~K\le N\), where \(u_j=(x_j,y_j)\in {{\mathbb {T}}}^{d-1}\times {{\mathbb {T}}}\) are distinct points and \(c_j\in {{\mathbb {C}}}\setminus {\left\{ {0}\right\} }\), is determined by its Fourier coefficient on the set

$$\begin{aligned} {\tilde{\Gamma }}_N^d = \left\{ n\in {{\mathbb {N}}}_0^d ~:~\prod _{j=1}^d \left\lceil \frac{n_j+1}{2} \right\rceil \le N\right\} . \end{aligned}$$
(22)

The measure can be recovered with an algorithm using a finite number of steps.

Proof

The proof works by using one-dimensional methods to give a large set of candidates. These candidates are all point measures with at most N summands. By Corollary 10, only one can fit all the available data, as \(\Gamma _{2N}^d\subset {\tilde{\Gamma }}_N^d\) . We use the same notation as in the proof of Theorem 9.

Again, we use induction in d, where the case \(d=1\) is a consequence of Theorem 2. Further, note that for \(r\in {{\mathbb {N}}}\setminus {\left\{ {0}\right\} }\) we have

$$\begin{aligned} (m, k)\in {\tilde{\Gamma }}_N^d \text { for } m\in {\tilde{\Gamma }}_{\lfloor N/r\rfloor }^{d-1} \text{ and } k=2r-2, \; 2r-1. \end{aligned}$$
(23)

We again use the decomposition (18) and introduce the notation

$$\begin{aligned} c_x(k) = \sum _{y: \; (x, y) \in U} c_{(x, y)} e^{-2\pi i yk }, \end{aligned}$$

to write

$$\begin{aligned} \widehat{\mu }(m, k)=\sum _{x \in X} c_x(k)e^{-2\pi i m \cdot x}, \quad m\in {{\mathbb {Z}}}^{d-1}, \; k\in {{\mathbb {Z}}}. \end{aligned}$$

Applying our algorithm for \(d-1\) to the samples \(\widehat{\mu }(m, 0)\), where \(m\in {\tilde{\Gamma }}_N^{d-1}\), we determine the quantities

$$\begin{aligned} c_x(0) = \sum _{y:\; (x, y) \in U} c_{(x, y)} \end{aligned}$$

as well as all \(x \in X\) for which \(c_x(0) \ne 0\). This includes all \(x \in X_1\). Again from our algorithm for \(d-1\), the numbers \(\widehat{\mu }(m, 1)\), \(m\in {\tilde{\Gamma }}_N^{d-1}\), determine the quantities

$$\begin{aligned} c_x(1) = \sum _{y:\; (x, y) \in U} c_{(x, y)} e^{-2\pi i y}, \end{aligned}$$

and all \(x \in X\) for which \(c_x(1) \ne 0\). So, by the the data (22) corresponding to \(n=(m, k)\), with \(m\in {\tilde{\Gamma }}_N^{d-1}\) and \(k=0, 1\), we can determine, for each point x, the Fourier coefficients

$$\begin{aligned} \widehat{\lambda _x}(0),\ \ \widehat{\lambda _x}(1) \end{aligned}$$

of the one-dimensional measure sitting over x:

$$\begin{aligned} \lambda _x = \sum _{y:\; (x, y) \in U} c_{(x, y)} \delta _y. \end{aligned}$$

We also determine those \(x \in X\) for which at least one of the numbers

$$\begin{aligned} c_x(0),\; c_x(1) \end{aligned}$$

is non-zero. We collect them in the set \(Z_1\). As we said above this includes all \(x \in X_1\):

$$\begin{aligned} X_1\subset Z_1\subset X. \end{aligned}$$

Now assume for the moment that we could identify the subset \(X_1\) in \(Z_1\). Then we could determine the part of \(\mu \) sitting over \(X_1\):

$$\begin{aligned} \mu _1 = \sum _{x \in X_1} \sum _{y:\;(x, y) \in U} c_{(x, y)} \delta _{(x, y)}. \end{aligned}$$

Subtracting \(\widehat{\mu _1}(m, k)\) from \(\widehat{\mu }(m, k)\) we see that we know the Fourier coefficients of the measure \(\mu -\mu _1\) for all indices in (22).

The next stage is to determine \(\mu _2\), the part of \(\mu \) sitting over \(X_2\), the points on the x-axis with two point masses over them. We will do this using the data

$$\begin{aligned}&\widehat{\mu -\mu _1}(m, k)\text { for }(m, k) \in {\tilde{\Gamma }}_{N}^{d-1}\times \{0,1\} \text { and } (m,k)\in {\tilde{\Gamma }}^{d-1}_{{\left\lfloor {\frac{N}{2}}\right\rfloor }}\times \{ 2,3\}.\qquad \end{aligned}$$
(24)

From (21) the measure \(\mu -\mu _1\) contains at most N/2 point masses, therefore, by the induction hypothesis, the data (24) are now enough to determine the quantities

$$\begin{aligned} c_x(k) = \sum _{y:\;(x, y) \in U} c_{(x, y)} e^{-2\pi i k y}, \quad x \in X_2 \sqcup X_3 \sqcup \cdots \sqcup X_r, \end{aligned}$$

for \(k = 0, \ldots , 3\), and those \(x \in X\) for which at least one of the numbers \(c_x(k)\), \(k=0, \ldots , 3\), is non-zero, which we collect in the set \(Z_2\). This includes all \(x \in X_2\) by using the induction hypothesis. Again, assume we were somehow able to identify \(X_2\) from the larger set \(Z_2\).

If \(x \in X_2\) this information suffices, because of Theorem 2, to determine \(\lambda _x\), that is, the part of measure \(\mu \) sitting over x. So the data \((m, k)\in {{\mathbb {N}}}_0^{d-1}\times {{\mathbb {N}}}_0\) in (22) with \(k \le 3\) determine \(\mu _2\), the part of measure \(\mu \) sitting over \(X_2\).

This process continues. The next step is to find, using the data

$$\begin{aligned} (\mu -\mu _1-\mu _2)^\wedge (m, k), \end{aligned}$$

where

$$\begin{aligned} (m, k)\in {{\mathbb {N}}}_0^{d-1}\times {{\mathbb {N}}}_0 \text { as in } (22) \text { with } k \le 5 \end{aligned}$$

the measure \(\mu _3\), the part of \(\mu \) sitting over \(X_3\). This is again possible since \(\mu -\mu _1-\mu _2\) contains at most N/3 point masses.

Continuing like this we determine the measure \(\mu \) completely.

However, we do not know which subset of \(Z_1\) is \(X_1\). Instead, we run the whole procedure for every possible choice, not only for \(X_1\) but also \(X_2\subset Z_2\) and so on. As all sets are finite and only one solution exists, this procedure will recover \(\mu \) in a finite number of operations. \(\square \)

Theorem 11 is a new result, showing that in principle \(O(N\log ^{d-1} N)\) samples enable not only to conclude uniqueness (which allows to search the continuous parameter space to recover \(\mu \)), but to recover the measure using a finite number of computational steps. It was sketched in the PhD thesis [25, Thm. 3.24] of the first-named author. However, the algorithm does not have polynomial runtime. It would be interesting to examine whether this is a conceptual barrier or whether more efficient methods exist. The result should be contrasted with the result of Sauer [12], which uses \(O(N^2\log ^{2d-2} N)\) samples but has polynomial runtime.