# The Fixed Angle Scattering Problem with a First-Order Perturbation

## Abstract

We study the inverse scattering problem of determining a magnetic field and electric potential from scattering measurements corresponding to finitely many plane waves. The main result shows that the coefficients are uniquely determined by 2n measurements up to a natural gauge. We also show that one can recover the full first-order term for a related equation having no gauge invariance, and that it is possible to reduce the number of measurements if the coefficients have certain symmetries. This work extends the fixed angle scattering results of Rakesh and Salo (SIAM J Math Anal 52(6):5467–5499, 2020) and (Inverse Probl 36(3):035005, 2020) to Hamiltonians with first-order perturbations, and it is based on wave equation methods and Carleman estimates.

## Introduction and Main Theorems

In this work, we study the inverse scattering problem of recovering a first-order perturbation from fixed angle scattering measurements. Let $$\lambda > 0$$, $$n\ge 2$$, and let $$\omega \in S^{n-1}$$ be a fixed unit vector. Suppose that for $$m\in \mathbb {N}$$, $$\mathcal V(x,D)$$ is a first-order differential operator with $$C^m(\mathbb {R}^n)$$ coefficients having compact support in $$B = \{x\in \mathbb {R}^n:|x|<1\}$$, the open ball of radius 1. We consider a Hamiltonian $$H_{\mathcal V} = -\Delta + {\mathcal V}(x,D)$$ in $$\mathbb {R}^n$$ and the problem

\begin{aligned} {\left\{ \begin{array}{ll} (H_{\mathcal V} -\lambda ^2) \psi _{\mathcal V} = 0 \\ \psi _{\mathcal V}(x,\lambda ,\omega ) = e^{i\lambda \omega \cdot x} + \psi ^s_{\mathcal V}(x,\lambda ,\omega ), \end{array}\right. } \end{aligned}
(1.1)

where $$\psi ^s_{\mathcal V}(x,\lambda ,\omega )$$ is known as a the scattering solution. It is well known that in order to have uniqueness for this problem, one needs to put further restrictions on the function $$\psi _{\mathcal V}^s$$. See, e.g.  [27] for the following facts. The outgoing Sommerfeld radiation condition (SRC for short)

\begin{aligned} \partial _r \psi _{\mathcal V}^s -i \lambda \psi _{\mathcal V}^s = o(r^{-(n-1)/2}) \quad \text {as } r \rightarrow \infty , \end{aligned}

where $$r = |x|$$, selects the solutions that heuristically behave as Fourier transforms in time of spherical waves that propagate toward infinity. A function $$\psi ^s_{\mathcal V}$$ satisfying (1.1) and the SRC is called an outgoing scattering solution. This solution is given by the so-called outgoing resolvent operator

\begin{aligned} R_{\mathcal V}(\lambda ) : = (H_{ \mathcal V} - (\lambda + i0)^2)^{-1} , \end{aligned}
(1.2)

so that, formally,

\begin{aligned} \psi _{\mathcal V}^s = R_{\mathcal V}(\lambda )(- \mathcal V(x,D)e^{i\lambda \omega \cdot x}). \end{aligned}

Notice that, assuming that such a solution $$\psi ^s_{\mathcal V}$$ exists—which would happen if the resolvent is well defined and bounded in appropriate spaces—it must satisfy the Helmholtz equation

\begin{aligned}(-\Delta -\lambda ^2) \psi ^s_{\mathcal V} = 0 \quad \text {in} \; \; \mathbb {R}^n \setminus \overline{B}, \end{aligned}

since the coefficients of $${\mathcal V}(x,D)$$ are compactly supported in B. It is well known that a solution of Helmholtz equation satisfying the SRC has always the asymptotic expansion

\begin{aligned} \psi ^s _{{\mathcal V}}(x,\lambda ,\omega ) = e^{i\lambda |x|} |x|^{-\frac{n-1}{2}} a_{\mathcal V}(\lambda ,\theta ,\omega ) + o(|x|^{-\frac{n-1}{2}}), \quad \text {as} \;\; |x|\rightarrow \infty , \end{aligned}
(1.3)

where $$\theta = \frac{x}{|x|}$$ and $$a_{\mathcal V}(\lambda ,\theta ,\omega )$$ is called the scattering amplitude or far field pattern.

In this setting, the main objective of an inverse scattering problem consists in reconstructing the coefficients of $$\mathcal V(x,D)$$ from partial or full knowledge of $$a_{\mathcal V}(\lambda ,\theta ,\omega )$$. Depending on the data that is assumed to be known, we can distinguish several types of inverse scattering problems:

1. 1.

Full data Recover the coefficients of $$\mathcal V(x,D)$$ from the knowledge of $$a_{\mathcal V}(\lambda ,\theta ,\omega )$$ for all $$(\lambda , \theta , \omega )\in (0,\infty ) \times S^{n-1}\times S^{n-1}$$.

2. 2.

Fixed frequency (or fixed energy) Recover the coefficients of $$\mathcal V(x,D)$$ from the knowledge of $$a_{\mathcal V}(\lambda _0,\theta ,\omega )$$ for a fixed $$\lambda _0>0$$ and all $$(\theta , \omega )\in S^{n-1}\times S^{n-1}$$.

3. 3.

Backscattering Recover the coefficients of $$\mathcal V(x,D)$$ from the knowledge of $$a_{\mathcal V}(\lambda ,\omega ,-\omega )$$ for all $$(\lambda , \omega )\in (0,\infty ) \times S^{n-1}$$.

4. 4.

Fixed angle (single measurement) Recover the coefficients of $$\mathcal V(x,D)$$ from the knowledge of $$a_{\mathcal V}(\lambda ,\theta ,\omega _0)$$ for a fixed $$\omega _0\in S^{n-1}$$ and all $$(\lambda , \theta )\in (0,\infty ) \times S^{n-1}$$.

In the case of fixed angle scattering, it is also interesting to consider analogous inverse problems in which $$a_{\mathcal V}(\, \cdot \,, \, \cdot \,, \omega )$$ is assumed to be known for each $$\omega$$ in a fixed subset (usually finite) of $$S^{n-1}$$.

Let $$D = -i\nabla$$. We consider the Hamiltonian

\begin{aligned} H_{\mathcal V} = H_{\mathbf {A},q} =(D+ \mathbf {A})^2 +q= -\Delta + 2\mathbf {A}\cdot D+ D\cdot \mathbf {A}+ \mathbf {A}^2+ q ,\end{aligned}

where both the magnetic potential $$\mathbf {A}$$ and the electrostatic potential q are real. Then, $$H_{\mathcal V}$$ is self-adjoint, and if $$\mathbf {A}$$ and q are compactly supported, the resolvent (1.2) is bounded in appropriate spaces under very general assumptions on the regularity of $$\mathbf {A}$$ and q. This implies that the problem (1.1) has a unique solution $$\psi ^s_{\mathcal V}$$ and hence that the scattering amplitude $$a_{\mathcal V} =a_{\mathbf {A},q}$$ is well defined, so that the fixed angle scattering problem can be appropriately stated.

In [19, 21], it has been proved that for $$H_\mathcal V = -\Delta +q$$, knowledge of the fixed angle scattering data $$a_{\mathcal V}(\, \cdot \, , \, \cdot \, , \omega )$$ in two opposite directions $$\omega = \pm \omega _0$$ for a fixed $$\omega _0 \in S^{n-1}$$ determines uniquely the potential q. The present article extends the results of [19, 21] to the case of non-vanishing first-order coefficients and proves that from 2n measurements, or just $$n+1$$ measurements under symmetry conditions, one can determine both the first- and zeroth-order coefficients up to natural gauges. To prove these results, we follow the approach used in [19]; that is, we show the equivalence of the fixed angle scattering problem with an appropriate inverse problem for the wave equation. This inverse scattering problem in time domain consists in recovering information on $$\mathbf {A}$$ and q from boundary measurements of the solution $$U_{\mathbf {A},q}$$ of the initial value problem

\begin{aligned} (\partial _t^2 + H_{\mathbf {A},q})U_{\mathbf {A},q} =0 \; \; \text {in}\; \mathbb {R}^{n+1}, \qquad U_{\mathbf {A},q}|_{\left\{ t<-1\right\} } = \delta (t- x \cdot \omega ). \end{aligned}
(1.4)

If the support of $$\mathbf {A}$$ and q is contained in B, the boundary measurements of $$U_{\mathbf {A},q}$$ are made in the set $$\partial B \times (-T,T) \cap \{(x,t) : t \ge x\cdot \omega \}$$, where $$\partial B$$ denotes the boundary of the ball.

We now describe some previous results on the inverse scattering problem of recovering a potential q(x) from fixed angle measurements. As discussed above, this problem can be considered in the frequency domain, as the problem of determining q from the scattering amplitude $$a_q(\, \cdot \,, \, \cdot \,, \omega )$$ for the Schrödinger operator $$-\Delta + q$$ with a fixed direction $$\omega \in S^{n-1}$$, or alternatively in the time domain as the problem of recovering q from boundary or scattering measurements of the solution $$U_q$$ of the wave equation. The equivalence of these problems is discussed in [19] (see [16, 17, 26] for the odd-dimensional case).

The one-dimensional case is quite classical, see [6, 15]. In dimensions $$n \ge 2$$, uniqueness has been proved for small or generic potentials [1, 23], recovery of singularities results are given in [18, 20], and uniqueness of the zero potential is considered in [2]. Recently, in [19, 21] it was proved that measurements for two opposite fixed angles uniquely determine a potential $$q \in C^{\infty }_c(\mathbb {R}^n)$$. The problem with one measurement remains open, but [19, 21] prove uniqueness for symmetric or horizontally controlled potentials (similar to angularly controlled potentials in backscattering [25]), and Lipschitz stability estimates are given for the wave equation version of the problem. We also mention the recent work [14] which studies the fixed angle problem when the Euclidean metric is replaced by a Riemannian metric, or sound speed, satisfying certain conditions, and the upcoming work [13] which studies fixed angle scattering for time-dependent coefficients also in the case of first-order perturbations.

We now introduce the main results in this work. Since the metric is Euclidean, the vector potential $$\mathbf {A}$$ can equivalently be seen as a 1-form $$\mathbf {A}= A^j dx_j$$. We denote by $$d\mathbf {A}$$ the exterior derivative of $$\mathbf {A}$$. Let $$\{e_1,\dots ,e_n\}$$ be any orthonormal basis in $$\mathbb {R}^n$$. Our first result shows that the magnetic field $$d\mathbf {A}$$ and the electrostatic potential q are uniquely determined by the knowledge of the fixed angle scattering amplitude $$a_{\mathbf {A},q}(\, \cdot \,, \, \cdot \,, \omega )$$ for the n orthogonal directions $$\omega = e_j$$, $$1\le j \le n$$ and the n opposite ones, $$\omega = -e_j$$. From now on, in this paper we will fix $$m$$ to be the integer

\begin{aligned} m= \frac{3}{2} n +10 \; \; \; \text {if } n \text { is even}, \quad m=\frac{3}{2}(n+1) +10 \; \;\;\text {if } n \text { is odd}. \end{aligned}
(1.5)

In general, we consider $$\mathbf {A}\in C^{m+2}(\mathbb {R}^n; \mathbb {R}^n)$$ and $$q \in C^m(\mathbb {R}^n; \mathbb {R})$$. This is required in order to guarantee that the solutions of (1.4) satisfy certain regularity properties.

### Theorem 1.1

Let $$n\ge 2$$ and $$\lambda _0>0$$, and let $$\{e_1,\dots ,e_n\}$$ be any orthonormal basis in $$\mathbb {R}^n$$. Assume that the pairs of potentials $$\mathbf {A}_1,\mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {R}^n)$$ and $$q_1,q_2 \in C^{m}_c(\mathbb {R}^n; \mathbb {R})$$ are compactly supported in B. Assume also that the following condition holds:

\begin{aligned}&\int _{-\infty }^\infty e_n \cdot \mathbf {A}_k(x_1,\dots ,x_{n-1},s) \, ds = 0\nonumber \\&\quad \text {for} \; k=1,2 \; \, \text {and for every} \;(x_1,\dots ,x_{n-1}) \in \mathbb {R}^{n-1}. \end{aligned}
(1.6)

If for all $$\theta \in S^{n-1}$$ and $$\lambda \ge \lambda _0$$ we have

\begin{aligned}a_{ \mathbf {A}_1,q_1} (\lambda ,\theta , \pm e_j) = a_{ \mathbf {A}_2,q_2}(\lambda ,\theta ,\pm e_j) \quad \text {for all } j =1,\dots ,n ,\end{aligned}

then $$d \mathbf {A}_1 = d \mathbf {A}_2$$ and $$q_1 = q_2$$.

The condition (1.6) is a technical restriction necessary to decouple the information on q from the information on $$\mathbf {A}$$ at some point in the proof of this uniqueness result.

In Theorem 1.1, one cannot recover completely the magnetic potential $$\mathbf {A}$$ due to the phenomenon of gauge invariance. This consists simply in the observation that if $$H_{\mathbf {A},q} u = v$$ for some functions u and v, then $$H_{\mathbf {A}+ \nabla f,q} \, (e^{ -if} u) = e^{-if}v$$, for any $$f\in C^2(\mathbb {R}^n)$$. Therefore, if f is compactly supported, the scattering amplitude is not going to be affected by f, so that $$a_{\mathbf {A},q} = a_{\mathbf {A}+ \nabla f,q}$$. On the other hand, if we consider Hamiltonians

\begin{aligned} \tilde{H}_{\mathbf {A},V} := -\Delta -2i \mathbf {A}\cdot D + V,\end{aligned}

where V is a fixed function, then the gauge invariance is broken and knowledge of the associated scattering amplitude $$\tilde{a}_{\mathbf {A},V}(\, \cdot \, , \,\cdot \,,\omega )$$ for the 2n directions $$\omega = \pm e_j$$ determines completely $$\mathbf {A}$$.

### Theorem 1.2

Let $$n\ge 2$$ and $$\lambda _0>0$$, and let $$\{e_1,\dots ,e_n\}$$ be any orthonormal basis in $$\mathbb {R}^n$$. Assume that $$\mathbf {A}_1,\mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {R}^n)$$ and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ have compact support in B, and that the Hamiltonians $$\tilde{H}_{\mathbf {A}_1,V}$$ and $$\tilde{H}_{\mathbf {A}_2,V}$$ are both self-adjoint operators.

If for all $$\theta \in S^{n-1}$$ and $$\lambda \ge \lambda _0$$ we have

\begin{aligned} \tilde{a}_{ \mathbf {A}_1,V} (\lambda ,\theta , \pm e_j) = \tilde{a}_{ \mathbf {A}_2,V}(\lambda ,\theta ,\pm e_j) \quad \text {for all } j =1,\dots ,n, \end{aligned}

then $$\mathbf {A}_1 = \mathbf {A}_2$$.

Notice that in this statement (1.6) is not assumed. This is related to the fact that V is fixed, so it is not necessary to decouple V from $$\mathbf {A}$$ in the proof. Similarly, if $$\mathbf {A}$$ is a fixed vector potential, it would be possible to determine q from the knowledge of $$a_{\mathbf {A},q}(\,\cdot \,,\,\cdot \,,\pm \omega )$$ for a fixed $$\omega \in S^{n-1}$$.

In both the previous theorems, we have considered an orthonormal basis $$\{e_1,\dots ,e_n\}$$ of $$\mathbb {R}^n$$ in order to simplify the notation and computations in some parts of the arguments, but we remark that our proofs can be easily adapted to allow non-orthonormal directions of measurements. Also, in both theorems the measured quantities are the scattering amplitudes generated by 2n waves incoming each one from one of the 2n different directions $$\pm e_j$$. Given $$e_j$$, one can avoid the need of sending a wave also from the opposite direction $$-e_j$$ provided the potentials satisfy certain symmetry properties. As an example of this phenomenon, we state the following result.

### Theorem 1.3

Let $$n\ge 2$$ and $$\lambda _0>0$$, and let $$\{e_1,\dots ,e_n\}$$ be any orthonormal basis in $$\mathbb {R}^n$$. Assume that the pairs of potentials $$\mathbf {A}_1,\mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {R}^n)$$ and $$q_1,q_2 \in C^{m}_c(\mathbb {R}^n; \mathbb {R})$$ are compactly supported in B and satisfy (1.6). Assume also that

\begin{aligned} \mathbf {A}_k(-x) = -\mathbf {A}_k(x) , \quad \text {for} \quad k=1,2. \end{aligned}
(1.7)

If for all $$\theta \in S^{n-1}$$ and $$\lambda \ge \lambda _0$$ we have

\begin{aligned} \begin{aligned} a_{ \mathbf {A}_1,q_1} (\lambda ,\theta , e_j)&= a_{ \mathbf {A}_2,q_2}(\lambda ,\theta , e_j) \quad \text {for all } j =1,\dots ,n-1, \text { and} \\ a_{ \mathbf {A}_1,q_1} (\lambda ,\theta , \pm e_n)&= a_{ \mathbf {A}_2,q_2}(\lambda ,\theta ,\pm e_n), \end{aligned} \end{aligned}

then $$d \mathbf {A}_1 = d \mathbf {A}_2$$ and $$q_1 = q_2$$.

We assume that $$a_{ \mathbf {A}_1, q_1} (\lambda ,\theta , \omega ) = a_{ \mathbf {A}_2,q_2}(\lambda ,\theta ,\omega )$$ for $$\omega = \pm e_n$$ instead of just $$\omega = e_n$$ since we have not considered any symmetry on the potential q. (A result assuming symmetries on q to reduce further the data could also be proved modifying slightly the arguments used to prove this theorem.) We also prove in time domain a more technical result analogous to Theorem 1.2 that requires just n measurements instead of 2n, and that is compatible with less restrictive symmetry conditions than (1.7) (see Theorem 5.1).

As already mentioned, the previous theorems follow from corresponding results for the time domain inverse problem (Theorems 2.1, 2.2, and 5.2, respectively). We now state the precise result that establishes the equivalence between the inverse scattering problem in frequency domain and the inverse scattering problem in time domain, extending the results in [19] to first-order perturbations.

### Theorem 1.4

Let $$n\ge 2$$, $$\omega \in S^{n-1}$$, and $$\lambda _0>0$$. Assume that $$\mathbf {A}_1,\mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {R}^n)$$ and $$q_1,q_2 \in C^{m}_c(\mathbb {R}^n; \mathbb {R})$$ are supported in B. For $$k=1,2$$, let $$U_{\mathbf {A}_k, q_k}(x,t;\omega )$$ be the unique distributional solution of the initial value problem

\begin{aligned} ( \partial _t^2 + (D +\mathbf {A}_k )^2 + q_k )U_{\mathbf {A}_k, q_k} = 0 \; \, \text {in}\; \; \mathbb {R}^{n+1}, \quad U_{\mathbf {A}_k, q_k} |_{\left\{ t<-1\right\} } = \delta (t- x \cdot \omega ). \end{aligned}

Then, one has that

\begin{aligned} a_{\mathbf {A}_1, q_1}(\lambda , \theta , \omega )= a_{\mathbf {A}_2, q_2}(\lambda , \theta , \omega )\; \; \text {for}\; \;\lambda \ge \lambda _0 \; \;\text {and}\; \;\theta \in S^{n-1}, \end{aligned}

if and only if

\begin{aligned} U_{\mathbf {A}_1, q_1}(x,t;\omega )= U_{\mathbf {A}_2, q_2}(x,t ;\omega ) \; \; \text {for all}\;\; (x,t)\in (\partial B \times \mathbb {R}) \cap \left\{ t \ge x\cdot \omega \right\} . \end{aligned}

We remark that the restriction of the distribution $$U_{\mathbf {A}_k,q_k}$$ to the surface $$\partial B \times \mathbb {R}$$ is always well defined and vanishes in the open set $$(\partial B \times \mathbb {R}) \cap \left\{ t <x\cdot \omega \right\}$$. This can be seen from the explicit formula for $$U_{\mathbf {A}_k,q_k}$$ that we will compute in Sect. 2.

The proof of the main results in time domain is based on a Carleman estimate method introduced in [19, 21], which in turn adapts the method introduced in [4]. See [3, 11, 12] for more information and references on the Bukhgeim–Klibanov method, and [5, 10] for its use in inverse boundary problems for the magnetic Schrödinger operator.

Essentially, the Carleman estimate is applied to the difference of two solutions of (1.4). The general idea is to choose an appropriate Carleman weight function for the wave operator that is large on the surface $$\{ t = x \cdot \omega \}$$ and allows one to control a source term on the right-hand side of the equation. Then, one needs an additional energy estimate to absorb the error coming from the source term. This will allow one to control the difference of the potentials $$\mathbf {A}_1-\mathbf {A}_2$$ or $$q_1-q_2$$. This step is the key to get the uniqueness result. Unfortunately, after doing all this, there is a remaining boundary term in the Carleman estimate that cannot be appropriately controlled. However, this term can be canceled using an equivalent Carleman estimate for solutions of the wave equation coming from the opposite direction. This is why we require 2n measurements to recover n independent functions instead of just n measurements. Assuming symmetry properties on the coefficients like in Theorem 1.3 is essentially an alternative way to get around this difficulty.

An interesting point in the proof of the time domain results is how one decouples the information concerning $$\mathbf {A}$$ from the information on q. The method used here consists in considering the solutions of the initial value problem

\begin{aligned} (\partial _t^2 + H_{\mathbf {A},q}) U_{\mathbf {A},q} =0 \; \; \text {in}\; \mathbb {R}^{n+1}, \qquad U_{\mathbf {A},q}|_{\left\{ t<-1\right\} } = H(t- x \cdot \omega ), \end{aligned}
(1.8)

where H stands for the Heaviside function. Since (1.4) is essentially the time derivative of the previous IVP, it turns out that it is equivalent to formulate the inverse scattering problem in time domain using any of these initial value problems. The advantage is that the solutions of (1.8) contain information only about $$\mathbf {A}$$ at the surface $$\{t=x \cdot \omega \}$$. By using these ideas, we are able to estimate both $$\mathbf {A}_1-\mathbf {A}_2$$ in terms of $$q_1-q_2$$ and $$q_1-q_2$$ in terms of $$\mathbf {A}_1-\mathbf {A}_2$$. Using these two estimates in tandem allows us to recover both the magnetic field and electric potential under the assumption (1.6).

This paper is structured as follows. In Sect. 2, we state the time domain results, Theorems 2.1 and 2.2, from which Theorems 1.1 and 1.2 follow by Theorem 1.4. We also analyze the structure of the solutions of the initial value problems (1.4) and (1.8), and we state several of their properties that will play an essential role later on. In Sect. 3, we introduce the Carleman estimate, and in Sect. 4, we combine the results of the previous two sections to prove Theorems 2.1 and 2.2. In the last section of the paper, we state and prove Theorems 5.1 and 5.2 in order to illustrate how the number of measurements can be reduced in time domain by imposing symmetry assumptions on the potentials. (Theorem 1.3 follows from the second result.) The proof of Theorem 1.4 given in “Appendices A and B” is devoted to adapting several known results for the wave operator to our purposes.

## The Inverse Problem in Time Domain

### Main Results in Time Domain

Let $$\omega \in S^{n-1}$$ be fixed. In the time domain setting, we consider the initial value problem

\begin{aligned} ( \partial _t^2 + H_{\mathcal V})U_{\mathcal V}=0 \; \; \text {in}\; \mathbb {R}^{n+1}, \qquad U_{\mathcal V}|_{\left\{ t<-1\right\} } = \delta (t- x \cdot \omega ), \end{aligned}
(2.1)

where $$\delta$$ represents the 1-dimensional delta distribution and $$H_{\mathcal V} = -\Delta + {\mathcal V}(x,D)$$ . Formally, the problem (1.1) is the Fourier transform in the time variable of (2.1). As we will show later in this section, there is a unique distributional solution of $$U_{\mathcal V}$$ if the first-order coefficients of $$\mathcal V$$ are in $$C^{m+2}_{c}(\mathbb {R}^n)$$ and the zero-order coefficient is $$C^{m}_{c}(\mathbb {R}^n)$$, for $$m$$ as in (1.5).

The inverse problem in the time domain consists in determining the coefficients of $$\mathcal {V}$$ from certain measurements of $$U_{\mathcal V}$$ at the boundary $$\partial B \times (-T, T)\subset \mathbb {R}^{n+1}$$ for some fixed $$T>0$$. To simplify the notation, we define

\begin{aligned} \Sigma : = \partial B \times (-T,T) .\end{aligned}

From now, depending on the context, it will be useful to write the Hamiltonian $$H_\mathcal V$$ both in the forms

\begin{aligned} H_{\mathbf {A},q} = (D + \mathbf {A})^2 + q , \end{aligned}
(2.2)

and

\begin{aligned} L_{\mathbf {W},V} = -\Delta + 2\mathbf {W}\cdot \nabla + V , \end{aligned}
(2.3)

where $$\mathbf {A},\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$q,V\in C^{m}_c(\mathbb {R}^n ;\mathbb {C})$$. Since the coefficients have high regularity and are complex valued, both forms are completely equivalent, but the first notation is specially convenient in the cases where there is gauge invariance. In fact, this inverse problem has an invariance equivalent to the gauge invariance present in the frequency domain problem. A straightforward computation shows that if U is a solution of

\begin{aligned} (\partial _t^2 + H_{\mathbf {A},q})U=0 \; \; \text {in}\; \mathbb {R}^{n+1}, \qquad U|_{\left\{ t<-1\right\} } = \delta (t- x \cdot \omega ), \end{aligned}

then $$\widetilde{U}= e^{-f} U$$ is a solution of

\begin{aligned} (\partial _t^2 + H_{\mathbf {A}+\nabla f, q})\widetilde{U}=0 \; \; \text {in}\; \mathbb {R}^{n+1}, \qquad \widetilde{U}|_{\left\{ t<-1\right\} } = \delta (t- x \cdot \omega ), \end{aligned}

where f is any $$C^2(\mathbb {R}^n)$$ function with compact support in B. The initial condition satisfied by $$\widetilde{U}$$ is not affected by the exponential factor $$e^{-f}$$ since for $$t<-1$$ the distribution $$\delta (t-x\cdot \omega )$$ is supported in $$\{ x \cdot \omega < -1\}$$, a region where f vanishes. On the other hand, we also have that $$\widetilde{U}|_{\Sigma }= U|_{\Sigma }$$ since the support of f is contained in B. Hence, at best one can recover the magnetic field $$d\mathbf {A}$$ from the boundary data $$U|_{\Sigma }$$. We now state two uniqueness results for the inverse problem that we have just introduced.

### Theorem 2.1

Let $$\mathbf {A}_1,\mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$q_1,q_2 \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ with compact support in B and such that

\begin{aligned} \int _{-\infty }^\infty e_n \cdot \mathbf {A}_k(x_1,\dots ,x_{n-1},s) \, ds = 0 \quad \text {for} \; k=1,2, \; \text {and all} \;(x_1,\dots ,x_{n-1}) \in \mathbb {R}^{n-1}. \end{aligned}

Also, let $$1 \le j \le n$$ and consider the 2n solutions $$U_{k, \pm j}(x,t)$$ of

\begin{aligned} ( \partial _t^2 + (D +\mathbf {A}_k )^2 + q_k )U_{k, \pm j}=0 \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, \pm j}|_{\left\{ t<-1\right\} } = \delta (t- \pm x_j).\nonumber \\ \end{aligned}
(2.4)

If for each $$1 \le j \le n$$ one has $$U_{1,\pm j} = U_{2, \pm j}$$ on the surface $$\Sigma \cap \{ t \ge \pm x_j \}$$, then $$d \mathbf {A}_1 = d \mathbf {A}_2$$ and $$q_1 = q_2$$.

As in the introduction, we highlight that the restriction of the distribution $$U_{1,\pm j}$$ to the surface $$\Sigma$$ is well defined and vanishes in the open set $$\Sigma \cap \{ t < \pm x_j \}$$, see the comments after Proposition 2.4 for more details. Theorem 1.1 follows directly from this result and Theorem 1.4. On the other hand, if we fix the zero-order term to be always the same, then the gauge invariance disappears and one can recover completely the first-order term of the perturbation. To state this result, we use the Hamiltonian in the form (2.3).

### Theorem 2.2

Let $$\mathbf {W}_1,\mathbf {W}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ with compact support in B. Let $$1 \le j \le n$$ and $$k=1,2$$, and consider the corresponding 2n solutions $$U_{k, \pm j}$$ satisfying

\begin{aligned} ( \partial _t^2 - \Delta + 2\mathbf {W}_k \cdot \nabla + V) U_{k, \pm j} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, \pm j}|_{\left\{ t<-1\right\} }= \delta (t- \pm x_j).\nonumber \\ \end{aligned}
(2.5)

If for each $$1 \le j \le n$$ one has $$U_{1,\pm j} = U_{2, \pm j}$$ on the surface $$\Sigma \cap \{ t \ge \pm x_j \}$$, then $$\mathbf {W}_1 = \mathbf {W}_2$$.

As in the previous case, Theorem 1.2 follows from Theorem 2.2 with $$\mathbf {W}_k = -i\mathbf {A}_k$$ and Theorem 1.4. To see this, notice that for $$k=1,2$$ the Hamiltonians $$\tilde{H}_{\mathbf {A}_k,V}$$ in Theorem 1.2 satisfy $$\tilde{H}_{\mathbf {A}_k,V} = (D +\mathbf {A}_k )^2 + q_k$$ for $$q_k := V-\mathbf {A}_k^2 -D \cdot \mathbf {A}_k$$. And since by assumption $$\mathbf {A}_k$$ is real and $$\tilde{H}_{\mathbf {A}_k,V}$$ is self-adjoint, $$q_k$$ must be a real function. This means that the conditions required to apply Theorem 1.4 are satisfied.

In Theorem 2.2, one needs 2n measurements to obtain the unique determination of the first-order coefficient $$\mathbf {W}$$. The number of measurements in Theorem 2.1 stays the same even if one now also proves the unique determination of q. This is due to the gauge invariance phenomenon, since now one uniquely determines just the magnetic field $$d\mathbf {A}$$ and not the first-order term $$\mathbf {A}$$. In fact, in the uniqueness proof, gauge invariance reduces one degree of freedom in the first-order term by making it possible to choose a gauge in which the nth component of $$\mathbf {A}_1 -\mathbf {A}_2$$ vanishes. (As we shall see later on, the fact that the solutions $$U_{k, \pm n}$$ coincide at $$\Sigma \cap \{ t = \pm x_n \}$$ guarantees that there are no obstructions for this gauge transformation.)

### The Direct Problem

In order to prove the previous theorems, we need to study the direct problem (2.1) in more detail. Let $$\omega \in S^{n-1}$$. Assume $$\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V\in C^{m}_c(\mathbb {R};\mathbb {C})$$, and consider the initial value problems for the wave operator

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) U_\delta =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad U_\delta |_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega ), \end{aligned}
(2.6)

and

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) U_H =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad U_H|_{\left\{ t<-1\right\} }= H (t-x\cdot \omega ), \end{aligned}
(2.7)

where $$L_{\mathbf {W},V}$$ was defined in (2.3). As mentioned in the introduction, the reason we also consider the second equation is that the $$\delta$$-wave $$U_\delta$$ and H-wave $$U_H$$ contain equivalent information about $$\mathbf {W}$$ and V (see Proposition 2.5), but H-waves decouple the information on $$\mathbf {W}$$ from the information on V.

To study (2.6) and (2.7), it is convenient to use certain coordinates in $$\mathbb {R}^n$$ associated with the fixed vector $$\omega$$. Specifically, we take any orthonormal coordinate system such that, for $$x\in \mathbb {R}^n$$, we have $$x = (y,z)$$ where $$y \in \{ \omega \}^{\perp }$$ (identified with $$\mathbb {R}^{n-1}$$) and $$z = x \cdot \omega$$.

For a fixed $$T> 7$$, it will be helpful to introduce the following subsets of $$\mathbb {R}^{n+1}$$ (see Fig. 1):

\begin{aligned} \begin{aligned} Q&:= B \times (-T, T) ,&\qquad \Sigma&:= \partial B \times (-T, T), \\ Q_{\pm }&:= Q\cap \left\{ \pm (t-z)>0 \right\} ,&\qquad \Sigma _{\pm }&:= \Sigma \cap \left\{ \pm (t-z)>0\right\} ,\\ \Gamma&:= \overline{Q}\cap \left\{ t=z \right\} ,&\qquad \Gamma _{\pm T}&:= \overline{Q}\cap \left\{ t=\pm T\right\} . \end{aligned} \end{aligned}
(2.8)

We now give a heuristic motivation of the existence of solutions $$U_\delta$$ and $$U_H$$ of (2.6) and (2.7). For the interested reader, we give a proof in Sect. B.1 of the properties that we now state, by means of the progressing wave expansion method. We start by making the ansatz of looking for possible solutions of (2.6) and (2.7) in the family of functions satisfying

\begin{aligned} U(y, z, t)= f(y,z,t) H(t-z) + g(y,z,t) \delta (t-z), \end{aligned}

where f(yzt) and g(yzt) are $$C^2$$ functions in $$\{t \ge z\}$$. A straightforward computation shows that

\begin{aligned} \begin{aligned} (\partial _t^2+ L_{\mathbf {W},V})U&= \left[ (\partial _t^2+ L_{\mathbf {W},V})f\right] H(t-z)\\&\quad +\, \left[ (\partial _t^2+ L_{\mathbf {W},V})g +2(\partial _t+\partial _z - \omega \cdot \mathbf {W})f \right] \delta (t-z) \\&\quad +\, 2\left[ (\partial _t+\partial _z- \omega \cdot \mathbf {W})g\right] \partial _t \delta (t-z). \end{aligned} \end{aligned}
(2.9)

In the case of equation (2.7) to satisfy the initial condition, we need to have $$f = u$$ where u is a function satisfying $$u(x,t) = 1$$ for all $$t<-1$$, and $$g(x,t)=0$$ for all $$(x,t) \in \mathbb {R}^{n+1}$$. Then, (2.9) implies that u must satisfy

\begin{aligned} \begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) u&=0 \qquad \text {in}\; \left\{ t> z\right\} , \\ \partial _t u + \partial _z u - \omega \cdot \mathbf {W}u&= 0 \qquad \text {in } \left\{ t= z\right\} . \end{aligned} \end{aligned}
(2.10)

The unique solution of the last ODE is

\begin{aligned} u(y,z,z) = e^{\int _{-\infty }^{z} \omega \cdot \mathbf {W}(y,s) \,ds}, \end{aligned}

since it has to satisfy the initial condition $$u(y,z,z)=1$$ for $$z<-1$$. We now state this and further results about the solution of (2.7).

### Proposition 2.3

Let $$\omega \in S^{n-1}$$ be fixed. Let $$\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$. Define

\begin{aligned} \psi (x) := \int _{-\infty }^{0} \omega \cdot \mathbf {W}(x + s\omega ) \,ds. \end{aligned}
(2.11)

There is a unique distributional solution $$U_H(x,t;\omega )$$ of (2.7), and it is supported in the region $$\left\{ t\ge x \cdot \omega \right\}$$. In particular,

\begin{aligned} U_H(x,t;\omega ) = u(x,t)H(t-x \cdot \omega ), \end{aligned}

where u is a $${C^2}$$ function in $$\left\{ t\ge x \cdot \omega \right\}$$ satisfying the IVP

\begin{aligned} \begin{aligned}&(\partial _t^2 + L_{\mathbf {W},V}) u =0&\qquad&\text {in} \; \; \left\{ t> x \cdot \omega \right\} , \\&u(x, x \cdot \omega ) = e^{\psi (x)}&\qquad&\\&u(x,t)=1&\qquad&\text {in} \; \; \left\{ x \cdot \omega \le t<-1 \right\} . \end{aligned} \end{aligned}
(2.12)

Notice that the boundary value of u at $$\{t=x\cdot \omega \}$$ depends only on $$\mathbf {W}$$ and not on the zero-order term V.

We now study the solutions of (2.6). In this case, we have to consider (2.9) with $$g =1$$ for $$t<-1$$, and $$f= v$$ where v is a function satisfying $$v(x,t) = 0$$ for $$t< -1$$. Then, the following conditions must be satisfied:

\begin{aligned} \begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) v&=0&\quad&\text {in} \; \; \{t> z\}, \\ \partial _t v + \partial _z v - \omega \cdot \mathbf {W}v&= -1/2( \partial _t^2 + L_{\mathbf {W},V})g&\quad&\text {in} \; \; \left\{ t= z\right\} , \\ \partial _t g + \partial _z g - \omega \cdot \mathbf {W}g&= 0&\quad&\text {in} \; \; \left\{ t= z\right\} ,\\ \partial _t (\partial _t g) + \partial _z (\partial _t g) - \omega \cdot \mathbf {W}(\partial _t g)&= 0&\quad&\text {in} \; \; \left\{ t= z\right\} . \end{aligned} \end{aligned}
(2.13)

The last two conditions are required so that $$\left[ (\partial _t+\partial _z- \omega \cdot \mathbf {W})g\right] \partial _t \delta (t-z)$$ vanishes completely in all $$\mathbb {R}^{n+1}$$ as a distribution. As in the case of (2.10), the third ODE and the initial condition imply that

\begin{aligned} g(y,z,z) = e^{\psi (y,z)} = e^{\int _{-\infty }^{z} \omega \cdot \mathbf {W}(y,s) \,ds}, \end{aligned}

and in fact we are going to choose $$g(y,z,t) = e^{\psi (y,z)}$$ for all $$(y,z,t) \in \mathbb {R}^{n+1}$$. We can freely do this: $$g_1(y,z,t) \delta (t-z) = g_2(y,z,t) \delta (t-z)$$ iff $$g_1(y,z,z) = g_2(y,z,z)$$. Also, the previous choice implies that $$\partial _t g = 0$$, so that the last condition is satisfied too. Computing explicitly $$L_{\mathbf {W},V}(g)$$, the second equation in (2.13) becomes the ODE

\begin{aligned} \partial _t v + \partial _z v - \omega \cdot \mathbf {W}v = - \frac{1}{2} e^\psi (-\Delta \psi -| \nabla \psi |^2 + 2\mathbf {W}\cdot \nabla \psi +V)\quad \text {in } \left\{ t= z \right\} , \end{aligned}

with initial condition $$v(y,z,z) = 0$$ if $$z<-1$$. The unique solution is then

\begin{aligned} v(y,z,z)= - \frac{1}{2} e^{ \psi (y,z)} \int _{-\infty }^{z} \left[ -\Delta \psi - |\nabla \psi |^2 + 2\mathbf {W}\cdot \nabla \psi + V \right] (y,s) \, \mathrm{d}s. \end{aligned}

We state this rigorously in the following proposition.

### Proposition 2.4

Let $$\omega \in S^{n-1}$$ be fixed. Let $$\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$. Consider $$\psi$$ as in (2.11). There is a unique distributional solution $$U_\delta (x,t;\omega )$$ of (2.6), and it is supported in the region $$\left\{ t\ge x \cdot \omega \right\}$$. Moreover,

\begin{aligned} U_\delta (x,t;\omega ) = v(x,t) H(t-x \cdot \omega ) + e^{\psi (x)}\delta (t-x \cdot \omega ), \end{aligned}
(2.14)

where v is a $$C^2$$ function in $$\left\{ t\ge x \cdot \omega \right\}$$ satisfying

\begin{aligned} \begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) v&=0&\quad&\text {in}\; \; \left\{ t> x \cdot \omega \right\} , \\ v(x, x \cdot \omega )&= F(x),&\quad&\\ v(x,t)&=0&\quad&\text {in} \; \; \left\{ x \cdot \omega \le t<-1 \right\} , \end{aligned} \end{aligned}
(2.15)

with

\begin{aligned} F(x)= -\frac{1}{2} e^{ \psi (x)} \int _{-\infty }^{0} \left[ -\Delta \psi - |\nabla \psi |^2 + 2\mathbf {W}\cdot \nabla \psi + V \right] (x +s\omega ) \, ds. \end{aligned}

One of the consequences of the formula (2.14) is that the restriction of $$U_\delta (x,t;\omega )$$ to the surface $$\Sigma$$ is well defined. This essentially follows from the fact that the wave front set of the distribution $$\delta (t-x\cdot \omega )$$ is disjoint from the normal bundle of $$\Sigma$$. We emphasize that $$U_\delta$$ satisfies the initial condition $$U_\delta |_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega )$$, even if it does not look that way at a first glance. This is due to the fact that when $$t<-1$$ the distribution $$\delta (t-x\cdot \omega )$$ is supported in $$\{ x \cdot \omega < -1\}$$, a region in which $$\psi$$ vanishes. Therefore,

\begin{aligned} e^{\psi (x)}\delta (t-x\cdot \omega ) = \delta (t-x\cdot \omega ) \quad \text {for} \; \, t<-1. \end{aligned}

As mentioned previously, for more details about the proofs of Propositions 2.3 and 2.4 see Sect. B.1. We remark that the condition (1.5) on the regularity of the coefficients appears in the proofs of these propositions in order to have $$C^2$$ solutions u and v of (2.12) and (2.15).

An important fact later on is that the solutions of (2.6) and (2.7) satisfy that $$\partial _t U_H= U_\delta$$. This is consequence of the independence of V and $$\mathbf {W}$$ from t together with the uniqueness of solutions for both equations. Of particular relevance for the scattering problem will be that this equivalence also holds for the boundary data: knowledge of $$U_H|_{\Sigma _+}$$ gives $$U_\delta |_{\Sigma _+}$$ and vice versa.

### Proposition 2.5

Let $$\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$, and let $$\omega \in S^{n-1}$$ be fixed. If $$U_\delta$$ and $$U_H$$ are, respectively, the unique distributional solutions of (2.6) and (2.7), then one has that $$\partial _t U_H = U_\delta$$ in the sense of distributions. In fact, it also holds that $$v(x,t) = \partial _t u(x,t)$$, so that (2.14) can be written as

\begin{aligned} U_\delta (x,t;\omega ) = (\partial _t u(x,t)) H(t-x \cdot \omega ) + u(x,t) \delta (t-x \cdot \omega ). \end{aligned}
(2.16)

Specifically, if $$\psi$$ is given by (2.11) we have that

\begin{aligned} u (x,t) = \int _{x \cdot \omega }^{t} v(x,\tau ) \, d\tau + e^{\psi (x)}, \quad \text {for} \;\, t \ge x \cdot \omega . \end{aligned}
(2.17)

Notice that the previous identity holds in particular for every $$x \in \partial B$$, so we can write that

\begin{aligned} u|_{\Sigma _+}(x,t) = \int _{x \cdot \omega }^{t} v|_{\Sigma _+}(x,\tau ) \, d\tau + e^{\psi (x)}, \quad \text {for} \;\, (x,t) \in \Sigma _+. \end{aligned}
(2.18)

### Proof of Proposition 2.5

Since $$\mathbf {W}$$ and V are independent of t, we can take a time derivative of both sides of (2.7). This implies that $$\partial _t U_H$$ satisfies

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) \partial _t U_H =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad \partial _t U_H|_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega ). \end{aligned}

Computing explicitly $$\partial _t U_H$$, we get

\begin{aligned} \partial _t U_H (x,t;\omega ) = (\partial _t u(x,t)) H(t-x \cdot \omega ) + u(x,t) \delta (t-x \cdot \omega ). \end{aligned}

Proposition 2.4 implies there is a unique distributional solution of (2.7), and hence, $$\partial _t U_H = U_\delta$$. Then, $$\partial _t u = v$$. We also get that $$u(x,x\cdot \omega ) = e^{\psi }$$, but we already knew this from Proposition 2.3. Therefore, (2.16) holds true. Identity (2.17) follows directly by the fundamental theorem of calculus. $$\square$$

As an immediate consequence of identity (2.18), we get the following lemma.

### Lemma 2.6

Let $$\mathbf {W}_1,\mathbf {W}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$V_1,V_2\in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$. For $$k=1,2$$ and $$\omega \in S^{n-1}$$, consider the solutions $$U_{H,k}$$ and $$U_{\delta ,k}$$ of

\begin{aligned} ( \square + 2\mathbf {W}_k \cdot \nabla + V_k) U_{H,k} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{H,k}|_{\left\{ t<-1\right\} }= H (t- x \cdot \omega ), \end{aligned}

and

\begin{aligned} ( \square + 2\mathbf {W}_k \cdot \nabla + V_k) U_{\delta ,k} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{\delta ,k}|_{\left\{ t<-1\right\} }= \delta (t- x \cdot \omega ). \end{aligned}

Then, $$U_{H,1} = U_{H,2}$$ in $$\Sigma \cap \{t \ge x\cdot \omega \}$$ if and only if $$U_{\delta ,1} = U_{\delta ,2}$$ in $$\Sigma \cap \{ t \ge x\cdot \omega \}$$.

### Energy Estimates

To finish this section, we state three different estimates related to the wave operator that will be useful later on. They are analogues of the estimates given in [21, Lemmas 3.3-3.5], modified in order to account for the presence of a first-order perturbation not considered in the mentioned paper. For completeness, we have included the proofs in Sect. B.2. The first two lemmas will be used to control certain boundary terms appearing in the Carleman estimate. We denote by $$\nabla _\Gamma \alpha$$ the component of $$\nabla \alpha$$ tangential to $$\Gamma$$.

### Lemma 2.7

Let $$T>1$$. Let $$\mathbf {W}\in L^\infty (B,\mathbb {C}^n)$$ and $$V\in L^\infty (B,\mathbb {C})$$. Then, the estimate

\begin{aligned}&\Vert \alpha \Vert _{L^2(\Gamma _T)} + \Vert \nabla _{x,t}\alpha \Vert _{L^2(\Gamma _T)} \\&\quad \lesssim \Vert \alpha \Vert _{H^1(\Gamma )} + \Vert (\square + 2\mathbf {W}\cdot \nabla + V)\alpha \Vert _{L^2(Q_+)} +\Vert \alpha \Vert _{H^1(\Sigma _+)} + \Vert \partial _\nu \alpha \Vert _{L^2(\Sigma _+)} , \end{aligned}

holds trueFootnote 1 for every $$\alpha \in C^{\infty }(\overline{Q}_+)$$. The implicit constant depends on $$\Vert \mathbf {W} \Vert _{L^\infty (B)}$$, $$\Vert V \Vert _{L^\infty (B)}$$ and T.

### Lemma 2.8

Let $$T>1$$. Let $$\mathbf {W}\in L^\infty (B,\mathbb {C}^n)$$, $$V\in L^\infty (B,\mathbb {C})$$ and $$\phi \in C^2(\overline{Q}_+)$$. Then, there is a constant $$\sigma _0>0$$ such that the following estimate

\begin{aligned} \begin{aligned}&\sigma ^2 \Vert e^{\sigma \phi }\alpha \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi }\nabla _{\Gamma }\alpha \Vert _{L^2(\Gamma )}^2\lesssim \sigma ^3 \Vert e^{\sigma \phi }\alpha \Vert _{L^2(Q_+)}^2 + \sigma \Vert e^{\sigma \phi }\nabla _{x,t}\alpha \Vert _{L^2(Q_+)}^2 \\&\quad + \Vert e^{\sigma \phi }(\square + 2\mathbf {W}\cdot \nabla + V)\alpha \Vert _{L^2(Q_+)}^2 + \sigma ^2 \Vert e^{\sigma \phi }\alpha \Vert _{L^2(\Sigma _+)}^2\\&\quad + \Vert e^{\sigma \phi } \nabla _{x,t}\alpha \Vert _{L^2(\Sigma _+)}^2 , \end{aligned} \end{aligned}
(2.19)

holds for every $$\alpha \in C^\infty (\overline{Q}_+)$$ and for every $$\sigma \ge \sigma _0$$. The implicit constant depends on $$\left\| \phi \right\| _{C^2(\overline{Q}_+)}$$, $$\Vert \mathbf {W} \Vert _{L^\infty (B)}$$, $$\Vert V \Vert _{L^\infty (B)}$$ and T.

The following lemma will be used to show that the normal derivative at $$\Sigma _{+}$$ of a solution of the free wave equation outside the unit ball vanishes at $$\Sigma _{+}$$ provided that certain extra conditions are met. Here, $$\nu$$ denotes the unit vector field normal to $$\Sigma _+$$.

### Lemma 2.9

Let $$T>1$$. Let $$\alpha (y,z,t)$$ be a $$C^2$$ function on $$\left\{ t\ge z\right\}$$ satisfying

\begin{aligned} \begin{aligned} \square \, \alpha&=0&\qquad&\text {in}\;\; \left\{ (y,z,t): |(y,z)|>1, \, t>z \right\} , \\ (\partial _t +\partial _z) \alpha (y, z,z)&= 0&\qquad&\text {in} \;\; \left\{ |(y,z)|>1, \, t=z\right\} ,\\ \alpha (y, z, t)&=0&\qquad&\text {in} \;\; \left\{ z< t<-1 \right\} . \end{aligned} \end{aligned}
(2.20)

Assume that on the region $$|(y,z) |\ge 1$$, we have

\begin{aligned} \alpha (y,z,z)= {\left\{ \begin{array}{ll} 0, \text {if} \quad |y |\ge 1, \\ 0, \text {if} \quad |y |\le 1, \; \text {and}\; \; z \le - \sqrt{1-|y|^2},\\ \beta (y), \text {if} \quad |y |\le 1, \; \text {and}\; \; z \ge \sqrt{1-|y|^2}, \end{array}\right. } \end{aligned}
(2.21)

for some $$\beta \in C^2_c(\mathbb {R}^{n-1})$$ compactly supported on $$|y |\le 1-\varepsilon$$, where $$\varepsilon \in (0,1)$$. Then,

\begin{aligned} \Vert \partial _\nu (\chi \alpha ) \Vert _{L^2(\Sigma _+)} \lesssim \varepsilon ^{-1/2} \Vert \chi \Vert _{C^1} \left( \Vert \alpha \Vert _{H^1(\Sigma _+)} + \Vert \alpha \Vert _{H^1(\Sigma _+\cap \Gamma )} \right) , \end{aligned}

for any function $$\chi \in C^1(\overline{Q}_+)$$. The implicit constant depends on T.

## The Carleman Estimate and Its Consequences

In this section, we are going to apply a suitable Carleman estimate in order to be able to control the difference of the potentials with the boundary data. For this purpose, we adapt a Carleman estimate for general second-order operators stated in [19, Theorem A.7]. The trick is to choose an appropriate weight function to obtain a meaningful estimate for the wave operator.

First, take any $$\vartheta \in \mathbb {R}^n$$ such that $$|\vartheta |=2$$ and consider the following smooth function,

\begin{aligned} \eta (y,z,t) := |x-\vartheta |^2 - \frac{1}{4}(t-z)^2. \end{aligned}
(3.1)

The weight in the Carleman estimate is going to be the function $$\phi = e^{\lambda \eta }$$ for some $$\lambda >0$$ large enough. This choice is made in order to have several properties. On the one hand, we want $$\phi$$ to be sufficiently pseudoconvex so that the Carleman estimate holds for the wave operator. On the other hand, we want $$\phi$$ to decay very fast when $$t >z$$ in order to deal with certain terms appearing in the Carleman estimate (see Lemma 3.2 and its application in the proof of Lemma 3.3). And finally, since the z coordinate is going to be determined by the direction $$\omega$$ of the traveling wave, we require the restriction of $$\phi$$ to the surface $$\{t=z\}$$ to be independent of $$\omega$$ (or in other words, dependent on x but independent of the choice of coordinates $$x=(y,z)$$). This is of great help since we recover $$\mathbf {A}$$ combining measurements made from waves traveling in different directions.

In order to state the Carleman estimate, we fix an open set $$D \subset \{ (x,t) \in \mathbb {R}^{n+1} : |x|<3/2\}$$ such that $$Q \subset D$$. (Recall the notation introduced in (2.8).)

### Proposition 3.1

Let $$\phi = e^{\lambda \eta }$$ for $$\lambda >0$$ large enough. Let $$\Omega \subset D$$ be any open set with Lipschitz boundary. Then, there exists some $$\sigma _0>0$$ such that

\begin{aligned}&\sigma \int _{\Omega } e^{2\sigma \phi } (|\nabla u |^2 + \sigma ^2 |u|^2) \, dx + \sigma \int _{\partial \Omega } \sum _{j=0}^n \nu _j E^j(x,\sigma {{\,\mathrm{\mathrm {Re}}\,}}(u), \nabla {{\,\mathrm{\mathrm {Re}}\,}}(u)) \, dS\\&\qquad + \sigma \int _{\partial \Omega } \sum _{j=0}^n \nu _j E^j(x, \sigma {{\,\mathrm{\mathrm {Im}}\,}}(u), \nabla {{\,\mathrm{\mathrm {Im}}\,}}(u)) \, dS \lesssim \int _{\Omega }e^{2\sigma \phi }|\square u |^2 dx \end{aligned}

holds for all $$u\in C^{2}(\overline{\Omega })$$ and all $$\sigma \ge \sigma _0$$. The implicit constant depends on n, $$\lambda$$, and $$\Omega$$. Here, $$\nu$$ is the outward pointing unit vector normal to $$\partial \Omega$$, and for real v and $$0\le j \le n$$, $$E^j$$ is given by

\begin{aligned} E^j \big (x, \sigma e^{-\sigma \phi } v,\nabla (e^{-\sigma \phi } v)\big )= & {} -\partial _j \phi (|\nabla _x v|^2-\partial _t v^2) \nonumber \\&-\, \sigma ^2 \partial _j \phi (| \nabla _x \phi |^2 - \partial _j \phi ^2)v^2\nonumber \\&-2 \partial _j v( \nabla _x v \cdot \nabla _x \phi - \partial _t v \partial _t\phi ) -\,g(x,t)\partial _j v v,\nonumber \\ \end{aligned}
(3.2)

where g is some real-valued and bounded function independent of $$\sigma$$ and v. (Here, the index 0 corresponds to t, so that $$\partial _0 \phi = \partial _t \phi$$.)

### Proof

Since $$\square h$$ is real if h is a real function, the statement follows from applying Theorem A.7 of [19] to the real and imaginary parts of u, and then adding the resulting estimates. However, to apply the mentioned result, one needs to verify that $$\phi$$ is strongly pseudoconvex in the domain D with respect to the wave operator $$\square$$. The reader can find the precise definition of this condition in [19, Appendix]), though it is not necessary for the discussion that follows.

Denote by $$(\xi ,\tau ) \in \mathbb {R}^{n+1}$$ the Fourier variables corresponding to (xt), where $$\xi \in \mathbb {R}^n$$ and $$\tau \in \mathbb {R}$$.

It can be proved that a function $$\phi = e^{\lambda \eta }$$ will be strongly pseudoconvex for $$\lambda >0$$ large enough provided $$\eta$$ satisfies certain technical conditions. By Propositions A.3 and A.5 in [19], these conditions are the following: One needs to verify that the level surfaces of $$\eta$$ are pseudoconvex with respect to the wave operator $$\square$$, and that $$|(\nabla _{x,t} \eta )(x,t) |>0$$ for all $$(x,t)\in \overline{D}$$.

In our case, the second property is immediate since $$|\vartheta |=2$$ implies that $$|x-\vartheta |^2$$ has non-vanishing gradient in $$\overline{D}$$. The reader can find in [19, Definition A.1] a precise definition of the first property. For the purpose of this work, it is enough to use that the level surfaces of a function f are pseudoconvex w.r.t. $$\square$$ in a domain D if for all $$(x,t) \in \overline{D}$$ and $$(\xi ,\tau ) \in \mathbb {R}^{n+1}$$

\begin{aligned}&\tau ^2 \partial _t^2 f -2\tau \xi \cdot \nabla _x \partial _t f + \sum _{j,k=1}^{n} \xi _j\xi _k \, \partial ^2_{jk} f > 0 \; \text { when } \tau ^2\nonumber \\&\quad = |\xi |^2 \text { and } \tau \partial _t f - \xi \cdot \nabla _x f = 0. \end{aligned}
(3.3)

. (Notice that $$\tau ^2 - |\xi |^2$$ is the symbol of the wave operator.)

We now consider f given by

\begin{aligned} f(y,z,t)= \frac{1}{2}(b|x-\vartheta |^2 - (t-z)^2), \end{aligned}

where $$|\vartheta |=2$$ and $$b>0$$. Here, we are using the orthonormal coordinates $$(x,t) = (y,z,t)$$, $$y\in \mathbb {R}^{n-1}$$ introduced previously. Let $$(\zeta , \rho ,\tau ) \in \mathbb {R}^{n-1} \times \mathbb {R}\times \mathbb {R}$$ be the Fourier variables counterpart to (yzt). A straightforward computation with $$j, k =1, \ldots , n-1$$ and $$\partial _j = \partial _{y_j}$$, shows that the only non-vanishing second-order derivatives of f are

\begin{aligned} \partial ^2_{t}f= -1, \quad \partial _{jk}^2 f= b\delta _{j,k} , \quad \partial _z^2 f= b-1, \quad \partial ^2_{tz}f=1, \end{aligned}

Then, in particular, (3.3) is verified if we show that

\begin{aligned} -\tau ^2+ b|\zeta |^2+ (b-1)\rho ^2- 2\tau \rho >0,\quad \text {whenever } \; \tau ^2= |\zeta |^2 + \rho ^2. \end{aligned}
(3.4)

Note that both conditions are homogeneous in the variables $$(\zeta ,\rho ,\tau )$$, so it is enough to study the case $$\tau =1$$. Thus, (3.4) becomes

\begin{aligned} -1 + b + (b-2)\rho ^2 - 2\rho >0 \qquad \text { for} \; -1 \le \rho \le 1. \end{aligned}

It is not difficult to verify that for $$b>3$$, this inequality always holds. This proves that for $$b=4$$, the level surfaces of f are strongly pseudoconvex with respect to $$\square$$, and therefore, the same holds for $$\eta = \frac{1}{2}f$$.

To finish, we mention that precise formula (3.2) of the quadratic forms $$E_j$$ is computed in detail in [19, Section A.2]. $$\square$$

In the following lemma, we prove a couple of properties of the weight $$\phi$$ that will be important later in order to show that some terms appearing in the Carleman estimate are suitably small in the parameter $$\sigma$$.

### Lemma 3.2

Let $$\vartheta \in \mathbb {R}^n$$ with $$|\vartheta |=2$$, and $$\eta$$ as in (3.1). Then, the following properties are satisfied for any $$T>7$$:

1. (i)

The smallest value of $$\phi$$ on $$\Gamma$$ is strictly larger than the largest value of $$\phi$$ on $$\Gamma _{-T} \cup \Gamma _{T}$$.

2. (ii)

The function

\begin{aligned} \kappa (\sigma ) = \sup _{(y,z) \in \overline{B}} \int _{-T}^T e^{2\sigma (\phi (y,z,t)-\phi (y,z,z))} \, dt \end{aligned}

satisfies that $$\underset{\sigma \rightarrow \infty }{\lim } \kappa (\sigma ) =0$$.

### Proof

Let $$\eta _0(y, z):= \vert x-\vartheta \vert ^2$$. For the first assertion, it is enough to prove that

\begin{aligned} \max _{\Gamma _{-T}\cup \Gamma _T} e^{\lambda (\eta _0 - \frac{1}{4}(t-z)^2)} <\min _{\Gamma } e^{\lambda \eta _0}, \end{aligned}

for T large enough, since $$\eta (y,z,z) = \eta _0(y, z)$$. Observing that $$|(y,z)| \le 1$$ in $$\Gamma$$ and $$\Gamma _{\pm T}$$, the previous inequality will hold if

\begin{aligned} \max _B \eta _0- \min _B \eta _0< \frac{1}{4}(T-1)^2, \end{aligned}

and since $$\max _B \eta _0- \min _B \eta _0 = 8$$, this is true for any $$T > 7$$. This yields the first assertion. To prove the second assertion, we are going to use the following inequality

\begin{aligned} 1- e^{-s} \ge \min \left\{ 1/2, s/2\right\} , \quad s\ge 0. \end{aligned}

Since $$e^{\lambda \eta _0}>1$$ always, we have

\begin{aligned} \phi (y,z,z)-\phi (y,z,t)= & {} e^{\lambda \eta _0} \left( 1- e^{-\frac{\lambda }{4}(t-z)^2} \right) \\\ge & {} 1- e^{-\frac{\lambda }{4}(t-z)^2}\ge \frac{1}{2} \min \left\{ 1, \frac{\lambda }{4} (t-z)^2 \right\} , \end{aligned}

and hence, since $$|z| \le 1$$,

\begin{aligned} 0&\le \int ^{T}_{-T} e^{2\sigma (\phi (y,z,t)-\phi (y,z,z))}\mathrm{d}t \le \int _{-T}^T e^{-\sigma \min \left\{ 1, \frac{\lambda }{4} (t-z)^2\right\} }dt\\&\le \int _{-T-z}^{T-z}e^{-\sigma \min \left\{ 1, \frac{\lambda }{4} t^2\right\} }\mathrm{d}t\le \int _{-T-1}^{T+1}e^{-\sigma \min \left\{ 1, \frac{\lambda }{4} t^2\right\} }dt. \end{aligned}

By the dominated convergence theorem, the last integral goes to zero when $$\sigma \rightarrow \infty$$, and therefore, $$\displaystyle {\lim _{\sigma \rightarrow \infty } \kappa (\sigma ) =0}$$. This completes the proof. $$\square$$

We now adapt the Carleman estimate of Proposition 3.1 to our purposes. First, define

\begin{aligned} \phi _0(x,\vartheta ) := e^{\lambda \vert x-\vartheta \vert ^2}, \end{aligned}

for $$x\in \mathbb {R}^n$$ and $$|\vartheta | =2$$, so that $$\phi (y,z,z) = \phi _0(y,z)$$. From now on, it is convenient to use the notation

\begin{aligned} Z : =\partial _t + \partial _z, \qquad N : =\partial _t - \partial _z. \end{aligned}

We are interested in applying Proposition 3.1 for $$\Omega = Q_\pm$$. The boundary of $$Q_\pm$$ is composed of the following regions: $$\partial Q_\pm = \Gamma \cup \Sigma _\pm \cup \Gamma _{\pm T}$$. We are going to use the precise formula (3.2) only in $$\Gamma$$, where in fact an explicit computation yields that

\begin{aligned}&\sum _{j=0}^n \nu _j E^j \big (x, \sigma e^{-\sigma \phi } v, \nabla (e^{-\sigma \phi } v)\big ) \nonumber \\&\quad = 4(N\phi )((Zv)^2 + \sigma ^2 (Z \phi )^2 v^2) +\, 2(Z\phi )(|\nabla _y v|^2 - \sigma ^2 |\nabla _y \phi |^2v^2) \nonumber \\&\qquad - 4(Zv)(\nabla _y v\cdot \nabla _y \phi ) -2(Zv)gv \end{aligned}
(3.5)

for any real function $$v \in C^2(\overline{Q}_\pm )$$ (for a detailed derivation of this formula, see [19, Section A.2]). The important thing about this identity is that it does not depend on $$\nabla v$$ but just on Zv and $$\nabla _y v$$, which are derivatives along directions tangent to $$\Gamma$$.

The following lemma is the consequence of the Carleman estimate in Proposition 3.1 that is relevant for our fixed angle scattering problem. It is an analogue of [19, Proposition 3.2] adapted to the case of magnetic potentials.

### Lemma 3.3

Let $$T>7$$ and $$\omega \in S^{n-1}$$, and let $$\phi _0(x,\vartheta )$$ be as above. Assume that

\begin{aligned} \left| (\square + 2\mathbf {E}_\pm \cdot \nabla + f_\pm ) w_\pm (x,t) \right| \lesssim |\mathbf {A}_\pm (x)| + |q_\pm (x)| \quad \text {for all} \quad (x,t) \in Q_\pm ,\nonumber \\ \end{aligned}
(3.6)

and that

\begin{aligned} \left| (\partial _t + \partial _z - \omega \cdot \mathbf {E}_\pm ) w_\pm \right| \gtrsim |h_\pm (x)| \quad \text {for all} \quad (x,t) \in \Gamma , \end{aligned}
(3.7)

hold for some vector fields $$\mathbf {E}_\pm (x), \mathbf {A}_\pm (x)$$ in $$C(\mathbb {R}^n;\mathbb {C}^n)$$, some functions $$f_\pm (x) , h_\pm (x), q_\pm (x)$$ in $$C(\mathbb {R}^n;\mathbb {C})$$, and $$w_\pm (x,t)$$ in $$C^2(\overline{Q}_\pm )$$. Then, there is a constant $$c>0$$ such that, for $$\sigma >0$$ large enough,

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} h_\pm \Vert _{L^2(B)}^2\lesssim & {} \sigma ^3 e^{c \sigma } \Vert w_+ - w_- \Vert _{H^1(\Gamma )}( \Vert w_+ \Vert _{H^1(\Gamma )} + \Vert w_- \Vert _{H^1(\Gamma )} )\nonumber \\&\quad +\, \gamma (\sigma ) \sum _{\pm } \left( \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B )}^2 \right) \nonumber \\&\quad +\, \sigma ^3 e^{c \sigma } \sum _{\pm } \left( \Vert w_\pm \Vert _{H^1(\Sigma _\pm )}^2 + \Vert \partial _\nu w_\pm \Vert _{L^2(\Sigma _\pm )}^2 \right) , \end{aligned}
(3.8)

where $$\gamma$$ is a positive function satisfying $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$. The implicit constant is independent of $$\sigma$$, $$\vartheta$$ and $$w_\pm$$.

We remark that since the functions $$\phi _0(x,\vartheta )$$ and $$h_\pm (x)$$ are independent of t, the norms $$\Vert e^{\sigma \phi _0} h_\pm \Vert _{L^2(B)}$$ and $$\Vert e^{\sigma \phi _0} h_\pm \Vert _{L^2(\Gamma )}$$ are equivalent. The same holds for $$\mathbf {A}_\pm (x)$$ and $$q_\pm (x)$$.

In the proof of the lemma, it will be useful to introduce the notation

\begin{aligned} F^j(x,\sigma u, \nabla u):= & {} e^{-2\sigma \phi } E^j \big (x,\sigma {{\,\mathrm{\mathrm {Re}}\,}}(u), \nabla {{\,\mathrm{\mathrm {Re}}\,}}(u)\big ) \nonumber \\&+ e^{-2\sigma \phi }E^j \big (x, \sigma {{\,\mathrm{\mathrm {Im}}\,}}(u), \nabla {{\,\mathrm{\mathrm {Im}}\,}}(u) \big ) . \end{aligned}
(3.9)

We remark that $$F^j$$ depends on the function g in Proposition 3.1 which could in principle be different in the domains $$Q_+$$ and $$Q_-$$, but as discussed in [19, footnote 1 in the proof of Proposition 3.2] we can choose g to be the same both in $$Q_+$$ and $$Q_-$$.

### Proof

We first apply Proposition 3.1 with $$\Omega = Q_+ \subset D$$. To simplify notation, we use w, $$\mathbf {A}$$, $$\mathbf {E}$$, q, and h instead of $$w_+$$, $$\mathbf {A}_+$$, $$\mathbf {E}_+$$, $$q_+$$, and $$h_+$$. For $$\sigma >0$$ large enough, Proposition 3.1 and (3.9) yield the estimate

\begin{aligned}&\sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(Q_+)}^2 +\sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(Q_+)}^2 \\&\quad +\, \sigma \int _{\partial Q_+ } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S \lesssim \Vert e^{\sigma \phi } \square w \Vert ^2_{L^2(Q_+)}, \end{aligned}

where to shorten notation, we are temporarily adopting the convention that $$F^j\nu _j$$ stands for $$\sum _{j=0}^n F^j\nu _j$$. Since $$\mathbf {E}$$ and f are bounded, a direct perturbation argument (one can absorb the extra terms in the left-hand side for $$\sigma$$ large enough) allows to obtain from the previous estimate that

\begin{aligned}&\sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(Q_+)}^2 +\sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(Q_+)}^2 + \sigma \int _{\partial Q_+ } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S \\&\quad \lesssim \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f )w \Vert ^2_{L^2(Q_+)}, \end{aligned}

for $$\sigma$$ large enough. Then, since $$\partial Q_+ = \Gamma \cup \Sigma _+ \cup \Gamma _T$$, we have

\begin{aligned}&\sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(Q_+)}^2 +\sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(Q_+)}^2 + \sigma \int _{\Gamma } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S \nonumber \\&\quad \lesssim \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f )w \Vert ^2_{L^2(Q_+)} \nonumber \\&\qquad + \sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(\Sigma _+ \cup \Gamma _T)}^2 + \sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(\Sigma _+ \cup \Gamma _T)}^2. \end{aligned}
(3.10)

The energy estimate in Lemma 2.8 yields

\begin{aligned}&\sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi } \nabla _\Gamma w \Vert _{L^2(\Gamma )}^2\\&\quad \lesssim \sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(Q_+)}^2 \\&\qquad +\,\sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(Q_+)}^2 + \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f )w \Vert ^2_{L^2(Q_+)} + \sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Sigma _+)}^2 \\&\qquad +\, \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(\Sigma _+)}^2. \end{aligned}

Combining this estimate with (3.10) gives

\begin{aligned}&\sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi } \nabla _\Gamma w \Vert _{L^2(\Gamma )}^2 + \sigma \int _{ \Gamma } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S \nonumber \\&\quad \lesssim \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f)w \Vert ^2_{L^2(Q_+)} + \sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(\Sigma _+ \cup \Gamma _T)}^2 \nonumber \\&\qquad + \sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(\Sigma _+ \cup \Gamma _T)}^2. \end{aligned}
(3.11)

For the terms over $$\Gamma _T$$, using the energy estimate in Lemma 2.7 one has

\begin{aligned} \Vert w \Vert _{L^2(\Gamma _T)}^2 + \Vert \nabla w \Vert _{L^2(\Gamma _T)}^2&\lesssim \Vert w \Vert _{H^1(\Gamma )}^2 + \Vert (\square + 2\mathbf {E}\cdot \nabla + f )w \Vert ^2_{L^2(Q_+)} \nonumber \\&\quad +\, \Vert w \Vert _{L^2(\Sigma _+ )}^2 + \Vert \nabla w \Vert _{L^2(\Sigma _+ )}^2 \nonumber \\&\lesssim \Vert w \Vert _{H^1(\Gamma )}^2 + \Vert \mathbf {A} \Vert _{L^2(\Gamma )}^2 + \Vert q \Vert _{L^2(\Gamma )}^2 + \Vert w \Vert _{L^2(\Sigma _+ )}^2 \nonumber \\&\quad +\, \Vert \partial _\nu w \Vert _{L^2(\Sigma _+ )}^2, \end{aligned}
(3.12)

where in the last line we have used (3.6), which implies that

\begin{aligned}\Vert (\square + 2\mathbf {E}\cdot \nabla +f)w \Vert ^2_{L^2(Q_+)} \lesssim \Vert \mathbf {A} \Vert _{L^2(B )}^2 + \Vert q \Vert _{L^2(B)}^2 \lesssim \Vert \mathbf {A} \Vert _{L^2(\Gamma )}^2 + \Vert q \Vert _{L^2(\Gamma )}^2, \end{aligned}

since $$\mathbf {A}$$ and q do not depend on t. We now multiply (3.12) by $$e^{\sigma \sup _{\Gamma _T} \phi }$$. Then, by Lemma 3.2 one can use in the right-hand side that $$\sup _{\Gamma _T} \phi \le \inf _\Gamma \phi -\delta$$ for some $$\delta >0$$. This gives

\begin{aligned}&\sigma ^3 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma _T)}^2 + \sigma \Vert e^{\sigma \phi } \nabla w \Vert _{L^2(\Gamma _T)}^2\\&\quad \lesssim \sigma ^3 e^{-2\delta \sigma } ( \Vert e^{\sigma \phi }w \Vert _{H^1(\Gamma )}^2 + \Vert e^{\sigma \phi }\mathbf {A} \Vert _{L^2(\Gamma )}^2 +\, \Vert e^{\sigma \phi }q \Vert _{L^2(\Gamma )}^2) \\&\qquad +\, \sigma ^3 e^{2 \sigma \sup _{\Gamma _T }\phi }( \Vert e^{\sigma \phi }w \Vert _{L^2(\Sigma _+)}^2 + \Vert e^{\sigma \phi } \partial _\nu w \Vert _{L^2(\Sigma _+)}^2). \end{aligned}

Inserting this estimate in (3.11), and taking $$\sigma$$ large enough to absorb the term $$\sigma ^3 e^{-2\delta \sigma } \Vert e^{\sigma \phi }w \Vert _{H^1(\Gamma )}^2$$ on the left, yields

\begin{aligned}&\sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi } \nabla _\Gamma w \Vert _{L^2(\Gamma )}^2 + \sigma \int _{ \Gamma } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S\nonumber \\&\quad \lesssim \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f)w \Vert ^2_{L^2(Q_+)} + \sigma ^3 e^{-2\delta \sigma } \left( \Vert e^{\sigma \phi }\mathbf {A} \Vert _{L^2(\Gamma )}^2 +\Vert e^{\sigma \phi }q \Vert _{L^2(\Gamma )}^2 \right) \nonumber \\&\qquad + \sigma ^3 e^{c \sigma } \left( \Vert w \Vert _{L^2(\Sigma _+)}^2 + \Vert \nabla w \Vert _{L^2(\Sigma _+)}^2 \right) . \end{aligned}
(3.13)

By (3.7), we can relate the left-hand side of the previous inequality with h using that

\begin{aligned} \Vert e^{\sigma \phi } h \Vert _{L^2(\Gamma )}^2 \lesssim \Vert e^{\sigma \phi } (Z - \omega \cdot \mathbf {E})w \Vert _{L^2(\Gamma )}^2 , \end{aligned}

and that for $$\sigma$$ large enough

\begin{aligned} \Vert e^{\sigma \phi } (Z - \omega \cdot \mathbf {E})w \Vert _{L^2(\Gamma )}^2\lesssim & {} \sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi } Z w \Vert _{L^2(\Gamma )}^2 \\\le & {} \sigma ^2 \Vert e^{\sigma \phi }w \Vert _{L^2(\Gamma )}^2 + \Vert e^{\sigma \phi } \nabla _\Gamma w \Vert _{L^2(\Gamma )}^2. \end{aligned}

Inserting these two estimates in (3.13) gives

\begin{aligned}&\Vert e^{\sigma \phi } h \Vert _{L^2(\Gamma )}^2 + \sigma \int _{ \Gamma } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S\nonumber \\&\quad \lesssim \Vert e^{\sigma \phi } (\square + 2\mathbf {E}\cdot \nabla +f)w \Vert ^2_{L^2(Q_+)} +\, \sigma ^3 e^{-2\delta \sigma } \left( \Vert e^{\sigma \phi }\mathbf {A} \Vert _{L^2(\Gamma )}^2 +\Vert e^{\sigma \phi }q \Vert _{L^2(\Gamma )}^2 \right) \nonumber \\&\qquad + \sigma ^3 e^{c \sigma } \left( \Vert w \Vert _{L^2(\Sigma _+)}^2 + \Vert \nabla w \Vert _{L^2(\Sigma _+)}^2 \right) . \end{aligned}
(3.14)

By (3.6) and Lemma 3.2, we have

\begin{aligned} \Vert e^{\sigma \phi } (\square +2 \mathbf {E}\cdot \nabla +f)w \Vert ^2_{L^2(Q_+)}\lesssim & {} \Vert e^{\sigma \phi } \mathbf {A} \Vert ^2_{L^2(Q_+)} +\Vert e^{\sigma \phi } q \Vert ^2_{L^2(Q_+)} \\\lesssim & {} \kappa (\sigma ) \left( \Vert e^{\sigma \phi } \mathbf {A} \Vert ^2_{L^2(\Gamma )} + \Vert e^{\sigma \phi } q \Vert ^2_{L^2(\Gamma )} \right) , \end{aligned}

where $$\kappa (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$. Using this in (3.14) yields

\begin{aligned}&\Vert e^{\sigma \phi } h \Vert _{L^2(\Gamma )}^2 + \sigma \int _{ \Gamma } e^{2\sigma \phi } F^j (x,\sigma w, \nabla w)\nu _j \, \mathrm{d}S \\&\quad \lesssim \gamma (\sigma ) \left( \Vert e^{\sigma \phi } \mathbf {A} \Vert ^2_{L^2(\Gamma )} + \Vert e^{\sigma \phi } q \Vert ^2_{L^2(\Gamma )} \right) + \sigma ^3 e^{c \sigma } \left( \Vert w \Vert _{L^2(\Sigma _+)}^2 + \Vert \nabla w \Vert _{L^2(\Sigma _+)}^2 \right) , \end{aligned}

where $$\gamma (\sigma ) := \kappa (\sigma ) + \sigma ^3 e^{-2\delta \sigma }$$ also satisfies $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$. We now use that $$\phi (y,z,z) = e^{\lambda |x-v|^2} = \phi _0(x)$$, which means that we can write the previous estimate changing the $$L^2(\Gamma )$$ norms to $$L^2(B)$$ norms. (This is possible since the integrands do not depend on t anymore.) Also, (3.5) and (3.9) imply that

\begin{aligned}F^j(x,\sigma w, \nabla w)\nu _j |_{\Gamma } = F^j(x,\sigma w, \nabla _\Gamma w)\nu _j |_{\Gamma } ;\end{aligned}

that is, $$F^j\nu _j$$ on $$\Gamma$$ only depends on the part of the gradient of w tangential to $$\Gamma$$. Applying these observations and rewriting the previous estimate with $$w = w_+$$ and $$\mathbf {A}= \mathbf {A}_+$$ yield

\begin{aligned}&\Vert e^{\sigma \phi _0} h_+ \Vert _{L^2(B)}^2 + \sigma \int _{ \Gamma } e^{2\sigma \phi _0} F^j (x,\sigma w_+, \nabla _\Gamma w_+)\nu _j \, \mathrm{d}S \nonumber \\&\quad \lesssim \gamma (\sigma ) \left( \Vert e^{\sigma \phi _0}\mathbf {A}_+ \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0}q_+ \Vert _{L^2(B)}^2 \right) \nonumber \\&\qquad + \sigma ^3 e^{c \sigma } \left( \Vert w_+ \Vert _{L^2(\Sigma _+)}^2 + \Vert \nabla w_+ \Vert _{L^2(\Sigma _+)}^2 \right) . \end{aligned}
(3.15)

Fix $$\nu$$ to be the downward pointing unit normal to $$\Gamma$$, so $$\nu$$ is an exterior normal for $$Q_+$$. An analogous argument in $$Q_-$$ yields the estimate

\begin{aligned}&\Vert e^{\sigma \phi _0} h_- \Vert _{L^2(B)}^2 - \sigma \int _{ \Gamma } e^{2\sigma \phi _0} F^j (x,\sigma w_-, \nabla _\Gamma w_-)\nu _j \, \mathrm{d}S\nonumber \\&\quad \lesssim \gamma (\sigma ) \left( \Vert e^{\sigma \phi _0}\mathbf {A}_- \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0}q_- \Vert _{L^2(B)}^2 \right) \nonumber \\&\qquad + \sigma ^3 e^{c \sigma } \left( \Vert w_- \Vert _{L^2(\Sigma _-)}^2 + \Vert \nabla w_- \Vert _{L^2(\Sigma _-)}^2 \right) , \end{aligned}
(3.16)

where the minus sign in the boundary term comes from the fact that the outward pointing normal at $$\Gamma$$ seen as part of the boundary of $$Q_-$$ is the opposite to that of the case of $$Q_+$$. Now, since $$F^j(x,\sigma w, \nabla _\Gamma w) \nu _j$$ is quadratic in w and $$\nabla _\Gamma w$$ (and the coefficients are bounded functions), we have that

\begin{aligned}&\int _{ \Gamma } e^{2\sigma \phi _0} |F^j (x,\sigma w_+, \nabla _\Gamma w_+) \nu _j- F^j (x,\sigma w_-, \nabla _\Gamma w_-) \nu _j | \, \mathrm{d}S \\&\quad \lesssim \sigma ^2 e^{c \sigma } \Vert w_+ - w_- \Vert _{H^1(\Gamma )}( \Vert w_+ \Vert _{H^1(\Gamma )} + \Vert w_- \Vert _{H^1(\Gamma )} ) . \end{aligned}

Therefore, adding (3.15) and (3.16) and applying the previous estimate give the desired result. $$\square$$

Lemma 3.3 is going to be used for two different purposes, and as a consequence, it will be convenient to restate the estimate in a more specific way. We do this in the following couple of lemmas.

### Lemma 3.4

Let $$T>7$$. Let $$\mathbf {E}_\pm$$, $$\mathbf {A}_\pm$$, $$f_\pm$$, $$q_\pm$$ and $$\gamma (\sigma )$$ be as in Lemma 3.3, and suppose that $$h_\pm = q_\pm$$. Assume that for a fixed $$\omega \in S^{n-1}$$, there exist $$w_\pm \in H^2(Q_\pm )$$ such that (3.6) and (3.7) hold with $$w_+ = w_-$$ on $$\Gamma$$. Assume also that $$w_\pm |_{\Sigma _\pm } = 0$$ and $$\partial _\nu w_\pm |_{\Sigma _\pm } = 0$$. Then, one has that

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2, \end{aligned}
(3.17)

for $$\sigma >0$$ large enough.

The proof is immediate from Lemma 3.3. This lemma will be applied in the following section to appropriate $$w_\pm$$ functions that we will construct using the $$\delta$$-wave solutions given by Proposition 2.4. The following lemma will be used in a similar setting, in this case for $$w_\pm$$ functions generated by H-wave solutions like the ones given by Proposition 2.4.

### Lemma 3.5

Let $$T>7$$. Also, let $$\mathbf {A}_\pm$$, $$q_\pm$$ and $$\gamma (\sigma )$$ be as in Lemma 3.3.

Suppose that for a fixed $$\omega \in S^{n-1}$$ there exist $$w_\pm$$, $$\mathbf {E}_\pm$$ and $$f_\pm$$ satisfying the assumptions of Lemma 3.3, and such that (3.6) and (3.7) hold with $$w_+ = w_-$$ on $$\Gamma$$ and $$h_\pm = \omega \cdot \mathbf {A}_\pm$$. Assume also that $$w_\pm |_{\Sigma _\pm } = 0$$ and $$\partial _\nu w_\pm |_{\Sigma _\pm } = 0$$. Then, one has that

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} \omega \cdot \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \left( \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B)}^2 +\Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 \right) .\nonumber \\ \end{aligned}
(3.18)

Moreover, let $$\{e_1 , \dots , e_n\}$$ be an orthonormal basis of $$\mathbb {R}^n$$, and let $$J \subseteq \{1,\dots ,n\}$$ be the set of natural numbers satisfying that at least one of the components $$e_j \cdot \mathbf {A}_+$$ and $$e_j \cdot \mathbf {A}_-$$ does not vanish completely in $$\mathbb {R}^n$$. Suppose that for each $$j\in J$$, estimate (3.18) holds with $$\omega = e_j$$. Then, one has that

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B)}^2 , \end{aligned}
(3.19)

for $$\sigma >0$$ large enough.

### Proof

Estimate (3.18) follows immediately from Lemma 3.3 under the assumptions in the statement. Let us prove (3.19). By assumption for each $$j \in J$$, we have

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} e_j \cdot \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \left( \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B )}^2 \right) .\nonumber \\ \end{aligned}
(3.20)

But notice that, by definition, $$e_j \cdot \mathbf {A}_\pm = 0$$ in $$\mathbb {R}^n$$ if $$j \notin J$$, so that $$\mathbf {A}_\pm = \sum _{j\in J} (e_j \cdot \mathbf {A}_\pm )e_j$$. This means that adding (3.20) for all $$j\in J$$ gives

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \left( \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0} q_\pm \Vert _{L^2(B )}^2 \right) . \end{aligned}

Then, using that $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$, to finish the proof is enough to take $$\sigma$$ large enough in order to absorb in the left-hand side the first term on the right. $$\square$$

We have already seen in Sect. 2 that the restriction to $$\Gamma$$ of the solutions of (2.7) and (2.6) is in a certain sense related to the coefficients of the perturbation. In the next section, we use this fact to construct appropriate functions $$w_+$$ and $$w_-$$ so that estimates (3.17) and (3.19) hold simultaneously with $$\mathbf {A}_\pm$$ and $$q_\pm$$ being quantities related to the differences of potentials $$\mathbf {A}_1 - \mathbf {A}_2$$ and $$q_1 - q_2$$. This will yield the proof of Theorem 2.1 (and similarly of 2.2), since the function $$\gamma (\sigma )$$ in the previous lemmas goes to zero when $$\sigma \rightarrow \infty$$.

## Proof of the Uniqueness Theorems with 2n measurements

With Lemmas 3.4 and 3.5, we can finally prove the uniqueness results, Theorem 2.1 and 2.2. Both theorems are stated in terms of pairs of measurements, one for each direction $$\pm e_j$$. Due to this fact, it is convenient to give an explicit expression in the same coordinate system for solutions of (2.7) and (2.6) that correspond to the opposite directions $$\omega = \pm \omega _0$$, where $$\omega _0 \in S^{n-1}$$ is fixed.

Let $$x \in \mathbb {R}^n$$, and choose any orthonormal coordinate system in $$\mathbb {R}^n$$ such that $$x = (y,z)$$, where $$y \in \{ \omega _0 \}^{\perp }$$ (identified with $$\mathbb {R}^{n-1}$$) and $$z= x\cdot \omega _0$$. Consider the solutions $$U_{H,\pm }(y,z,t) := U_{H}(y,z,t; \pm \omega _0)$$ given by Proposition 2.3 when $$\omega = \omega _0$$ and when $$\omega = -\omega _0$$. By Proposition 2.3, in this coordinate system it holds that $$U_{H,\pm }(y,z,t) = u_\pm (y,z,t)H(t- (\pm z))$$ with

\begin{aligned} {\left\{ \begin{array}{ll} (\partial _t^2 + L_{\mathbf {W},V}) u_\pm =0 \qquad \text {in }\; \; \left\{ t > \pm z \right\} , \\ u_\pm (y,z, \pm z) = e^{ \psi _{\pm }(y, z)} \end{array}\right. } \end{aligned}
(4.1)

where

\begin{aligned} \psi _{\pm }(y, z) := \pm \int _{-\infty }^{\pm z} \omega _0 \cdot \mathbf {W}(y, \pm s) \,\mathrm{d}s. \end{aligned}

In the case of Proposition 2.4, it is convenient to state the results for a Hamiltonian written in the form (2.2). This is easily obtained making the change $$\mathbf {W}= -i\mathbf {A}$$ and $$V= \mathbf {A}^2 + D\cdot \mathbf {A}+ q$$ in the previous results. Therefore, using the same coordinates as before, the solutions $$U_{\delta ,\pm }(y,z,t):= U_{\delta }(y,z,t;\pm \omega _0)$$ of (2.6) given by Proposition 2.4 satisfy

\begin{aligned}U_{\delta ,\pm } (y,z,t) = v_\pm (y,z,t)H(t- \pm z) + e^{\psi _\pm (y,z)} \delta (t- \pm z) \quad \text {in} \; \mathbb {R}^n,\end{aligned}

where $$\psi _\pm (y,z) := \pm (-i) \int _{-\infty }^{\pm z} \omega _0 \cdot \mathbf {A}(y, \pm s) \,ds$$, and $$v_\pm$$ satisfies

\begin{aligned} {\left\{ \begin{array}{ll} (\partial _t^2 + L_{\mathbf {W},V}) v_\pm =0 \qquad \text {in}\; \left\{ t> \pm z \right\} , \\ v_\pm (y,z, \pm z) = \frac{1}{2} e^{ \psi _\pm (y,z)} \int _{-\infty }^{\pm z} \left[ \nabla \cdot \left( i\mathbf {A}+ \nabla \psi _\pm \right) + \left( i\mathbf {A}+ \nabla \psi _\pm \right) ^2 -q \right] (y, \pm s) \, \mathrm{d}s. \end{array}\right. }\nonumber \\ \end{aligned}
(4.2)

Having these explicit coordinate expressions, we now prove Theorems 2.1 and 2.2. We start with the second one which is the simplest, since the zeroth-order term V is fixed. The proof consists in the construction of two appropriate functions $$w_+$$ and $$w_-$$ using the solutions $$U_{k, \pm j}$$ of (2.7) in order to apply the results introduced in the previous section. For a fixed direction $$\omega = e_j$$ and $$k=1,2$$, we use $$U_{k, + j}$$ to construct $$w_+$$ and, with a certain gauge change, we use $$U_{k, - j}$$ to construct $$w_-$$. The gauge change is a technical requirement necessary to have $$w_+ = w_-$$ in $$\Gamma$$, as assumed in Lemmas 3.4 and 3.5. Notice that the reason for this assumption comes from the fact that one wants to get rid of the first term on the right-hand side of the estimate (3.8) which is large when the parameter $$\sigma$$ is large.

### Proof of Theorem 2.2

By Lemma 2.6, we know that it is completely equivalent to consider that the $$U_{k, \pm j}$$ are solutions of the initial value problem

\begin{aligned} ( \partial _t^2 - \Delta + 2\mathbf {W}_k \cdot \nabla + V) U_{k, \pm j} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, \pm j}|_{\left\{ t<-1\right\} }= H (t- \pm x_j), \end{aligned}

instead of the IVP (2.5). Here, it is convenient to work with H-waves instead of $$\delta$$-waves since, as mentioned previously, the former have boundary values on $$\Gamma$$ that are independent of V (see Proposition 2.3).

Fix $$1 \le j \le n$$, and take any orthonormal coordinate system in $$\mathbb {R}^n$$ such that $$x= (y,z)$$ with $$z= x_j$$ and $$y \in \mathbb {R}^{n-1}$$. Let $$k=1,2$$. By the previous discussion, we know that the $$U_{k,\pm j}$$ functions satisfy that

\begin{aligned} U_{k,\pm j}(y,z,t) = u_{k,\pm j}(y,z,t)H(t-\pm z) , \end{aligned}
(4.3)

where $$u_{k,\pm j}$$ is given by (4.1) with $$\mathbf {W}=\mathbf {W}_k$$, $$\omega = \pm e_j$$, and $$\omega _0 = e_j$$. Writing this in detail, we obtain that

\begin{aligned} {\left\{ \begin{array}{ll} (\square + 2\mathbf {W}_k \cdot \nabla + V)u_{k,\pm j} =0 \quad \text {in} \quad \{t> \pm z\}, \\ u_{k,\pm j}(y,z,\pm z) = e^{\int _{-\infty }^{\pm z} (\pm e_j) \cdot \mathbf {W}_k(y,\pm s) \, \mathrm{d}s}. \end{array}\right. } \end{aligned}
(4.4)

By assumption, we have that $$U_{1,\pm j} = U_{2,\pm j}$$ on $$\Sigma _\pm = \Sigma \cap \{ t \ge \pm z\}$$, so that

\begin{aligned} u_{1,\pm j} = u_{2, \pm j} \quad \text {on} \quad \Sigma _\pm , \end{aligned}
(4.5)

and hence, in particular $$u_{1,j} = u_{2, j}$$ in $$\Sigma _+ \cap \Gamma = \Sigma \cap \{ t = z\}$$. From this and (4.4), we get that there is a function $$\mu _j : \mathbb {R}^{n-1} \rightarrow \mathbb {C}$$ such that

\begin{aligned} \mu _j(y): = \int _{-\infty }^{\infty } e_j \cdot \mathbf {W}_1(y, s) \, \mathrm{d}s = \int _{-\infty }^{\infty } e_j \cdot \mathbf {W}_2(y, s) \, \mathrm{d}s . \end{aligned}
(4.6)

We now define $$w_+ := u_{1,j} - u_{2,j}$$ in $$\overline{Q}_+$$. With this choice, Proposition 2.3 yields that $$w_+ \in C^2(\overline{Q}_+)$$, and (4.4) that

\begin{aligned} w_+(y,z, z) = e^{\int _{-\infty }^z e_j \cdot \mathbf {W}_1(y,s) \, \mathrm{d}s} - e^{\int _{-\infty }^z e_j \cdot \mathbf {W}_2(y,s) \, \mathrm{d}s}. \end{aligned}

To apply Lemma 3.3 and Lemma 3.5, we need also to define an appropriate function $$w_-$$ in $$\overline{Q}_-$$. To obtain a useful choice, we now consider the solutions $$u_{k,-j}$$ of (4.4) and we take

\begin{aligned}w_-(y,z,t) = e^{\mu _{j}(y)}(u_{1,-j}(y,z,-t) -u_{2,-j}(y,z,-t)).\end{aligned}

Then, $$w_- \in C^2(\overline{Q}_-)$$ as desired. Also, if $$t=z$$

\begin{aligned} \begin{aligned} w_-(y,z,z)&= e^{\mu _{j}(y)} u_{1,-j}(y,z,-z) - e^{\mu _{j}(y)} u_{2,-j}(y,z,-z)\\&= e^{ \mu _{j}(y) + \int _{-\infty }^{-z} (-e_j) \cdot \mathbf {W}_1(y,-s) \, ds} - e^{ \mu _{j}(y) + \int _{-\infty }^{-z} (-e_j) \cdot \mathbf {W}_2(y,-s) \, \mathrm{d}s} \\&= e^{ \mu _{j}(y) - \int _{z}^{\infty } e_j \cdot \mathbf {W}_1(y,s) \, \mathrm{d}s} - e^{ \mu _{j}(y) - \int _{z}^{\infty } e_j \cdot \mathbf {W}_2(y,s) \, \mathrm{d}s} = w_+(y,z,z). \end{aligned} \end{aligned}
(4.7)

Therefore, $$w_+$$ satisfies

\begin{aligned} \begin{aligned} (\square + 2\mathbf {W}_1 \cdot \nabla + V)w_+&= 2(\mathbf {W}_2 -\mathbf {W}_1) \cdot \nabla u_{2,j}&\quad&\text {in} \quad Q_+, \\ (\partial _t + \partial _z -e_j \cdot \mathbf {W}_1)w_+&= e_j \cdot (\mathbf {W}_1-\mathbf {W}_2) u_{2,j}&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}
(4.8)

and $$w_-$$ satisfies

\begin{aligned} \begin{aligned}&(\square + 2 (\mathbf {W}_1 + \nabla \mu _{j}) \cdot \nabla + f_-) w_- = 2e^{\mu _{j}} (\mathbf {W}_2 -\mathbf {W}_1) \cdot \nabla u_{2,-j}(x,-t)&\quad&\text {in} \quad Q_-, \\&(\partial _t + \partial _z -e_j \cdot \mathbf {W}_1) w_- = e^{\mu _{j}} \, e_j \cdot (\mathbf {W}_1-\mathbf {W}_2) u_{2,-j}(y,z,-z)&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}

where $$f_- = V-|\nabla \mu _{j}|^2 + \Delta \mu _{j} -2\mathbf {W}_1 \cdot \nabla \mu _{j}$$.

Observe that $$e^{\mu _{j}(y)}$$, $$u_{2,\pm j}(x,\pm t)$$ and $$|\nabla u_{2,\pm j}(x,\pm t) |$$ are bounded functions in $$\overline{Q}_\pm$$. Also, we have that $$|u_{2,\pm j}(x, \pm t)| \gtrsim 1$$ in $$\Gamma$$ by (4.4). Therefore, it holds that

\begin{aligned} \begin{aligned} |(\square + 2 \mathbf {E}_\pm \cdot \nabla + f_\pm )w_\pm |&\lesssim |\mathbf {W}_2 - \mathbf {W}_1|&\quad&\text {in} \quad Q_\pm , \\ |(\partial _t + \partial _z -e_j \cdot \mathbf {E}_\pm ) w_\pm |&\gtrsim |e_j \cdot (\mathbf {W}_1-\mathbf {W}_2)|&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}

with $$f_+ = V$$, $$\mathbf {E}_+ = \mathbf {W}_1$$, $$\mathbf {E}_- = \mathbf {W}_1 +\nabla \mu _{j}$$, and $$f_-$$ as before (notice that $$e_j \cdot \mathbf {E}_\pm = e_j \cdot \mathbf {W}_1$$ since $$e_j \cdot \nabla \mu _j =0$$). Hence, (3.6) and (3.7) are satisfied with $$\mathbf {A}_\pm = \mathbf {W}_1-\mathbf {W}_2$$, $$q_\pm = 0$$, and $$h_\pm = e_j \cdot (\mathbf {W}_1-\mathbf {W}_2)$$.

On the other hand, (4.5) implies that $$w_+|_{\Sigma _+} = w_-|_{\Sigma _-} = 0$$. Then, applying Lemma 2.9, respectively, with $$\alpha = w_+$$ and $$\chi =1$$, or $$\alpha = e^{-\mu _j}w_-$$ and $$\chi = e^{\mu _j}$$ yields that $$\partial _\nu w_\pm |_{\Sigma _\pm } = 0$$.

Since $$w_+ = w_-$$ in $$\Gamma$$ by (4.7), the previous observations imply that all the conditions to apply Lemma 3.5 are satisfied, so that (3.18) holds for each $$j\in \{1,\dots ,n\}$$ with $$\mathbf {A}_\pm = \mathbf {W}_1-\mathbf {W}_2$$ and $$q_\pm = 0$$. Therefore, the same lemma implies that (3.19) must also hold, and this gives

\begin{aligned} \Vert e^{\sigma \phi _0} (\mathbf {W}_1 - \mathbf {W}_2) \Vert _{L^2(B)} \le 0. \end{aligned}

Hence, $$\mathbf {W}_1 =\mathbf {W}_2$$, since $$e^{\sigma \phi _0(x)}>0$$ always. This finishes the proof. $$\square$$

We now go to the remaining case, Theorem 2.1. In this result, the zero-order term is not fixed so that there is gauge invariance. (In fact, to simplify the proof it will be convenient to fix a specific gauge.) This is a harder proof than the previous one since we need to decouple information about q from the information about $$\mathbf {A}$$.

### Proof of Theorem 2.1

Let $$k=1,2$$. Since making a change of gauge $$\mathbf {A}_k -\nabla f_k$$ with $$f_k$$ compactly supported leaves invariant the measured values $$U_{k,\pm j}|_{\Sigma _{+ ,\pm j}}$$, we can freely choose a suitable $$f_k$$ in order to simplify the problem. In fact, to show that $$d\mathbf {A}_1= d\mathbf {A}_2$$ it is enough to prove that $$\mathbf {A}_1=\mathbf {A}_2$$ in a specific fixed gauge. Under the assumptions in the statement, one can always take

\begin{aligned} f_k(x) = \int _{-\infty }^{x_n} e_n \cdot \mathbf {A}_k (x_1,\dots ,x_{n-1},s) \, \mathrm{d}s, \end{aligned}

since (1.6) implies that $$f_k$$ must be compactly supported in B. With this choice, one has that

\begin{aligned} e_n \cdot (\mathbf {A}_1 -\nabla f_1) = e_n \cdot (\mathbf {A}_2 - \nabla f_2) = 0, \end{aligned}

in $$\mathbb {R}^n$$. Therefore, the previous arguments show that from now on, we can assume without loss of generality that we have fixed a gauge such that $$e_n \cdot \mathbf {A}_1 = e_n \cdot \mathbf {A}_2 = 0$$ in $$\mathbb {R}^n$$.

Fix $$1 \le j \le n-1$$. We again take any orthonormal coordinate system in $$\mathbb {R}^n$$ such that $$x= (y,z)$$, where $$y\in \mathbb {R}^{n-1}$$ and $$z= x_j$$. Let $$k=1,2$$. As in the proof of Theorem 2.2, by Lemma 2.6 we can assume that $$U_{k,\pm j}$$ satisfies the IVP

\begin{aligned} ( \partial _t^2 + (D +\mathbf {A}_k )^2 + q_k )U_{k, \pm j}=0 \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, \pm j}|_{\left\{ t<-1\right\} } = H (t- \pm x_j),\nonumber \\ \end{aligned}
(4.9)

instead of (2.4). By Proposition 2.3, we know that the $$U_{k,\pm j}$$ have the structure described in (4.3) for $$1\le j \le n-1$$ where $$u_{k,\pm j}$$ satisfies (4.1) with $$\mathbf {W}= -i\mathbf {A}_k$$, $$\omega = \pm e_j$$, and $$\omega _0 = e_j$$. Writing this in detail, we obtain that

\begin{aligned} {\left\{ \begin{array}{ll} (\partial _t^2 + (D +\mathbf {A}_k)^2 +q_k)u_{k,\pm j} =0 \quad \text {in} \quad \{t> \pm z\}, \\ u_{k,\pm j}(y,z,\pm z) = e^{-i\int _{-\infty }^{\pm z} (\pm e_j) \cdot \mathbf {A}_k(y,\pm s) \, \mathrm{d}s}. \end{array}\right. } \end{aligned}
(4.10)

By the assumption that $$U_{1,\pm j} = U_{2,\pm j}$$ on $$\Sigma _\pm$$ and (4.3), we have that (4.5) holds analogously in this case. Specifically, in $$\Sigma _+ \cap \Gamma$$ this implies that there is a function $$\mu _j$$ such that

\begin{aligned} \mu _j(y) = -i\int _{-\infty }^{\infty } e_j \cdot \mathbf {A}_1(y, s) \, \mathrm{d}s = -i\int _{-\infty }^{\infty } e_j \cdot \mathbf {A}_2 (y, s) \, \mathrm{d}s. \end{aligned}

We now define $$w_+ := u_{1,j} - u_{2,j}$$, so that $$w_+ \in C^2(\overline{Q}_+)$$ by Proposition 2.3, and

\begin{aligned} w_+(y,z, z) = e^{-i\int _{-\infty }^z e_j \cdot \mathbf {A}_1(y,s) \, \mathrm{d}s} - e^{-i \int _{-\infty }^z e_j \cdot \mathbf {A}_2(y,s) \, \mathrm{d}s}. \end{aligned}

To define $$w_-$$, we consider the solutions $$u_{k,-j}$$ of (4.10) and we take

\begin{aligned} w_-(y,z,t) = e^{\mu _j(y)}u_{1,-j}(y,z,-t) - e^{\mu _j(y)} u_{2,-j}(y,z,-t). \end{aligned}

Then, $$w_- \in C^2(\overline{Q}_-)$$ as desired. Also, for $$t=z$$

\begin{aligned} \begin{aligned} w_-(y,z,z)&= e^{ \mu _{j}(y) - i\int _{-\infty }^{-z} (-e_j) \cdot \mathbf {A}_1(y,-s) \, \mathrm{d}s} - e^{ \mu _{j}(y) - i\int _{-\infty }^{-z} (-e_j) \cdot \mathbf {A}_2(y,-s) \, \mathrm{d}s} \\&= e^{\mu _{j}(y) + i\int _{z}^{\infty } e_j \cdot \mathbf {A}_1(y,s) \, \mathrm{d}s} - e^{\mu _{j}(y) + i\int _{z}^{\infty } e_j \cdot \mathbf {A}_2(y,s) \, \mathrm{d}s} = w_+(y,z,z). \end{aligned} \end{aligned}
(4.11)

Hence, if we define

\begin{aligned} V_k := \mathbf {A}_k^2 + D \cdot \mathbf {A}_k + q_k, \end{aligned}
(4.12)

$$w_+$$ satisfies

\begin{aligned} \begin{aligned} (\square -i2 \mathbf {A}_1 \cdot \nabla + V_1) w_+&= 2(\mathbf {A}_2 -\mathbf {A}_1) \cdot D u_{2,j} + (V_2 -V_1) u_{2,j}&\quad&\text {in} \quad Q_+, \\ (\partial _t + \partial _z +ie_j \cdot \mathbf {A}_1)w_+&= - i e_j \cdot (\mathbf {A}_1-\mathbf {A}_2) u_{2,j}&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}

and $$w_-$$ satisfies

\begin{aligned} \begin{aligned}&e^{-\mu _{j}}(\square - 2(i\mathbf {A}_1 - \nabla \mu _j) \cdot \nabla + \widetilde{V}_1)w_- = \big ( 2 (\mathbf {A}_2 -\mathbf {A}_1) \cdot D + \widetilde{V}_2 -\widetilde{V}_1 \big ) u_{2,-j} (x,-t)&\;&\text {in} \;\; Q_-, \\&e^{- \mu _{j}} (\partial _t + \partial _z +i e_j \cdot \mathbf {A}_1) w_- = - i e_j \cdot (\mathbf {A}_1-\mathbf {A}_2) u_{2,-j}(y,z,-z)&\quad&\text {in} \;\; \Gamma , \end{aligned} \end{aligned}

where $$\widetilde{V}_k = V_k + |\nabla \mu _{j}|^2 + \Delta \mu _{j} + 2i \mathbf {A}_k \cdot \nabla \mu _{j}$$. Since $$|\nabla \mu _j|$$ is bounded, we have that

\begin{aligned} |\widetilde{V}_1 -\widetilde{V}_2| \lesssim |V_1 -V_2| + |\mathbf {A}_1 -\mathbf {A}_2| . \end{aligned}

On the other hand, $$u_{2,\pm j}(x, \pm t)$$ and $$|\nabla u_{2,\pm j}(x, \pm t) |$$ are also bounded in $$\overline{Q}_\pm$$, and on $$\Gamma$$ we have that $$|u_{2,\pm j} | \gtrsim 1$$ by (4.10). Therefore, if $$f_+ = V_1$$, $$f_- = \widetilde{V}_1$$, $$\mathbf {E}_+ = -i\mathbf {A}_1$$, and $$\mathbf {E}_- = -i\mathbf {A}_1 + \nabla \mu _j$$ (notice that $$e_j \cdot \nabla \mu _j = 0$$), one gets

\begin{aligned} \begin{aligned} |(\square + 2 \mathbf {E}_\pm \cdot \nabla + f_\pm )w_\pm |&\lesssim |\mathbf {A}_1 - \mathbf {A}_2| +|V_1-V_2|&\quad&\text {in} \quad Q_\pm , \\ |(\partial _t + \partial _z - e_j \cdot \mathbf {E}_\pm ) w_\pm |&\gtrsim |e_j \cdot (\mathbf {A}_1-\mathbf {A}_2)|&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}

so that (3.6) and (3.7) are satisfied with $$h_\pm = e_j \cdot (\mathbf {A}_1-\mathbf {A}_2)$$,

\begin{aligned} \mathbf {A}_\pm = \mathbf {A}_1-\mathbf {A}_2, \quad \text {and} \quad q_\pm = V_1-V_2. \end{aligned}
(4.13)

As mentioned previously, (4.5) holds analogously in this case, so that $$w_+|_{\Sigma _+} = w_-|_{\Sigma _-} = 0$$. Again, applying Lemma 2.9, respectively, with $$\alpha = w_+$$ and $$\chi =1$$, or $$\alpha = e^{-\mu _j}w_-$$ and $$\chi = e^{\mu _j}$$ yields that $$\partial _\nu w_\pm |_{\Sigma _\pm } = 0$$.

Also, (4.11) shows that $$w_+ = w_-$$ in $$\Gamma$$. These assertions hold for each $$j =1, \dots ,n-1$$, and the n-th component $$e_n \cdot (\mathbf {A}_1-\mathbf {A}_2)$$ vanishes in $$\mathbb {R}^n$$. This means that the assumptions of Lemma 3.5 are satisfied with $$J= \{ 1, \dots ,n-1 \}$$.

As a consequence, (3.18) holds for each $$j\in J$$ and hence (3.18) also holds with the choices established in (4.13). This gives

\begin{aligned} \Vert e^{\sigma \phi _0} (\mathbf {A}_1 - \mathbf {A}_2) \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \Vert e^{\sigma \phi _0} (V_1-V_2) \Vert _{L^2(B)}^2, \end{aligned}
(4.14)

where $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$.

We now use the information provided by $$U_{k,\pm n}$$. We use the same coordinates as before, in this case with $$z= x_n$$. Since $$e_n \cdot \mathbf {A}_1= e_n \cdot \mathbf {A}_2 =0$$ in all $$\mathbb {R}^n$$, we have that

\begin{aligned} \psi _\pm (y,z) = -(\pm i) \int _{-\infty }^{\pm z} e_n \cdot \mathbf {A}_k(y,\pm s) \, \mathrm{d}s = 0 . \end{aligned}
(4.15)

Then, Proposition 2.4 and (4.2) yield

\begin{aligned} U_{k,\pm n}(y,z,t) = v_{k,\pm n}(y,z,t)H(t-\pm z) + \delta (t - \pm z), \end{aligned}

where using the notation introduced in (4.12) we have

\begin{aligned} {\left\{ \begin{array}{ll} (\partial _t^2 + (D +\mathbf {A}_k)^2 +q_k)v_{k,\pm n} =0 \quad \text {in} \quad \{t> \pm z\}, \\ v_{k,\pm n}(y,z,\pm z) = -\frac{1}{2}\int _{-\infty }^{\pm z} V_k (y,\pm s) \, \mathrm{d}s. \end{array}\right. } \end{aligned}
(4.16)

The assumption $$U_{1,\pm n}|_{\Sigma _\pm } = U_{2,\pm n}|_{\Sigma _\pm }$$ in the statement implies that

\begin{aligned} v_{1,\pm n}|_{\Sigma _\pm } = v_{2,\pm n}|_{\Sigma _\pm }, \end{aligned}
(4.17)

and since $$\mathbf {A}_k$$ and $$q_k$$ are compactly supported, this means that

\begin{aligned} v_{1, n}|_{\Sigma _+ \cap \Gamma } = -\frac{1}{2}\int _{-\infty }^{\infty } V_1 (y,s) \, \mathrm{d}s = -\frac{1}{2}\int _{-\infty }^{\infty } V_2 (y,s) \, \mathrm{d}s = v_{2,n}|_{\Sigma _+ \cap \Gamma }.\nonumber \\ \end{aligned}
(4.18)

We define

\begin{aligned} w_+ (y,z,t)= & {} 2(v_{1, n} - v_{2, n})(y,z,t) \quad \text {and} \quad w_-(y,z,t)\\= & {} -2(v_{1, -n} - v_{2, -n})(y,z,-t), \end{aligned}

so that $$w_\pm \in C^2(\overline{Q}_\pm )$$ by Proposition 2.4. The combination of the identities (4.16) and (4.18) and a change of variables shows that

\begin{aligned} w_+ (y,z,z) - w_- (y,z,z)&= -\int _{-\infty }^{ z} (V_1-V_2)(y, s) \, \mathrm{d}s - \int _{-\infty }^{-z } (V_1-V_2)(y,- s) \, \mathrm{d}s \nonumber \\&= -\int _{-\infty }^{ \infty } (V_1-V_2)(y, s) \, \mathrm{d}s = 0. \end{aligned}
(4.19)

Also, by direct computation (4.16) yields

\begin{aligned} \begin{aligned} |(\square -i2 \mathbf {A}_1 \cdot \nabla + V_1)w_\pm |&\lesssim |\mathbf {A}_1 - \mathbf {A}_2| +|V_1-V_2 |&\quad&\text {in} \quad Q_\pm , \\ |(\partial _t + \partial _z + ie_n \cdot \mathbf {A}_1 ) w_\pm |&= |V_1-V_2|&\quad&\text {in} \quad \Gamma . \end{aligned} \end{aligned}

Hence, (3.6) and (3.7) are satisfied with $$h_\pm = V_1-V_2$$, $$\mathbf {E}_\pm =-i\mathbf {A}_1$$, $$f_\pm =V_1$$, and $$\mathbf {A}_\pm$$ and $$q_\pm$$ as in (4.13). Also, (4.17) implies that $$w_+|_{\Sigma _+} = w_-|_{\Sigma _-} = 0$$, and (4.19) that $$w_+|_\Gamma = w_-|_\Gamma$$. The condition $$\partial _\nu w_\pm |_{{\Sigma _\pm }} = 0$$ follows again from Lemma 2.9.

Therefore, all the assumptions to apply Lemma 3.4 hold with the previous choices for $$\mathbf {A}_\pm$$ and $$q_\pm$$, and this yields

\begin{aligned} \Vert e^{\sigma \phi _0} (V_1-V_2) \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \Vert e^{\sigma \phi _0} (\mathbf {A}_1-\mathbf {A}_2) \Vert _{L^2(B)}^2 . \end{aligned}
(4.20)

Since $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow \infty$$, combining the previous estimate with (4.14) immediately implies that $$\mathbf {A}_1-\mathbf {A}_2 =0$$ and $$V_1 -V_2 = 0$$ in this gauge. In a general gauge, then $$d(\mathbf {A}_1-\mathbf {A}_2) =0$$. And since $$V_k = \mathbf {A}_k^2 + D \cdot \mathbf {A}_k + q_k$$, one obtains that $$q_1 = q_2$$. This finishes the proof. $$\square$$

That the $$w_\pm$$ functions satisfy condition (4.19) is one of the key properties needed in order to prove Theorem 2.1, since otherwise Lemma 3.4 cannot be applied with this choice of $$w_\pm$$. In order to prove (4.19), we have used indirectly that (1.6) holds. (Notice that this last condition means that we can move to a gauge in which (4.15) is true.) In fact, (4.19) is no longer true in general if one removes (1.6), since there appear non-vanishing terms in the right-hand side of (4.19) related to $$\psi _+$$ and $$\psi _-$$.

## Reducing the Number of Measurements

In this section, we prove an analogous result to Theorem 2.2, in which the number of measurements is reduced to n. To compensate this, one needs assume that the potentials have certain symmetries. (In fact, each component of $$\mathbf {W}$$ must satisfy some kind of antisymmetry property.) The main change in the proof is in the definition of $$w_-$$ in $$Q_-$$, which now is constructed by symmetry from $$w_+$$, instead of using new information coming from the solution associated with the opposite direction. For each $$0\le j \le$$, the symmetry of $$e_j \cdot \mathbf {W}$$ plays an essential role since it is necessary to show that $$w_+ = w_-$$ in $$\Gamma$$, in order to extract meaningful results from the Carleman estimate.

### Theorem 5.1

Let $$\mathbf {W}_1,\mathbf {W}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$, and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ with compact support in B. Also, let $$1 \le j \le n$$ and $$k=1,2$$, and consider the n solutions $$U_{k, j}$$ of

\begin{aligned} ( \partial _t^2 - \Delta + 2\mathbf {W}_k \cdot \nabla + V) U_{k, j} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, j}|_{\left\{ t<-1\right\} }= \delta (t- x_j).\nonumber \\ \end{aligned}
(5.1)

Assume also that for each $$1 \le j \le n$$, there exists an orthogonal transformation $$\mathcal O_j$$ satisfying that $$\mathcal O_j(e_j) = -e_j$$ and such that

\begin{aligned} e_j \cdot \mathbf {W}_k(x)= - e_j \cdot \mathbf {W}_k(\mathcal O_j(x)) \quad \text {for all} \quad k=1,2. \end{aligned}
(5.2)

Then, if for all $$1 \le j \le n$$ one has $$U_{1,j} = U_{2,j}$$ on the surface $$\Sigma \cap \{ t \ge x_j \}$$, it holds that $$\mathbf {W}_1 = \mathbf {W}_2$$.

The simplest example of a vector field satisfying the previous conditions is the case of an antisymmetric vector fields $$\mathbf {W}_k$$, that is, such that $$\mathbf {W}_k (-x) = -\mathbf {W}_k(x)$$ (for example, the gradient of a radial function).

We remark that it is possible to show that the previous theorem also holds if one substitutes condition (5.2) by

\begin{aligned} e_j \cdot \mathbf {W}_k(x)= - e_j \cdot \overline{\mathbf {W}}_k({\mathcal {O}}_j(x)) \quad \text {for all} \quad k=1,2 \quad \text {and}\quad 1 \le j \le n, \end{aligned}

that is, a symmetry condition instead of an antisymmetry condition in the imaginary part of $$e_j \cdot \mathbf {W}_k$$.

### Proof of Theorem 5.1

By Lemma 2.6, we know that it is completely equivalent to assume that the $$U_{k, j}$$ solutions solve

\begin{aligned} ( \partial _t^2 - \Delta + 2\mathbf {W}_k \cdot \nabla + V) U_{k, j} = 0, \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, j}|_{\left\{ t<-1\right\} }= H (t- x_j), \end{aligned}

instead of (5.1). Take $$1 \le j \le n$$. In this proof, we fix $$x= (y,z)$$ in $$\mathbb {R}^n$$, where $$z= x_j$$ and $$y \in \mathbb {R}^{n-1}$$. Let $$k=1,2$$. By Proposition 2.3 and (4.1), we know that $$U_{k,j}(y,z,t) = u_{k, j}(y,z,t)H(t- z)$$ where $$u_{k, j}$$ is the same function as $$u_{k, +j}$$ in (4.4)

Under the assumption in the statement, we have that $$u_{1,j} = u_{2, j}$$ on the surface $$\Sigma _+$$ and hence also in $$\Sigma _+ \cap \Gamma = \Sigma \cap \{ t = z\}$$, which implies that there is a function $$\mu _j : \mathbb {R}^{n-1} \rightarrow \mathbb {C}$$ such that (4.6) holds.

We now define $$w_+ := u_{1,j} - u_{2,j}$$ in $$\overline{Q}_+$$. Then,

\begin{aligned} w_+(y,z, z) = e^{\int _{-\infty }^z e_j \cdot \mathbf {W}_1(y,s) \, \mathrm{d}s} - e^{\int _{-\infty }^z e_j \cdot \mathbf {W}_2(y,s) \, \mathrm{d}s}. \end{aligned}
(5.3)

In this coordinates, for each matrix $$\mathcal O_j$$ in the statement there is by definition a $$n-1 \times n-1$$ orthogonal matrix $$\mathcal T_j$$ such that

\begin{aligned} \mathcal O_j (y,z) = (\mathcal T_j(y),-z) .\end{aligned}

To apply Lemma 3.5, we need also to define an appropriate function $$w_-$$ in $$\overline{Q}_-$$. We take

\begin{aligned} w_-(y,z,t) = e^{\mu _{j}(y)} w_+( {\mathcal T}_j (y),-z,-t) . \end{aligned}

If $$t=z$$, using (5.3) and the symmetry condition (5.2), we get that

\begin{aligned} \begin{aligned} w_-(y,z,z)&= e^{ \mu _{j}(y) + \int _{-\infty }^{-z} e_j \cdot \mathbf {W}_1({\mathcal T}_j(y),s) \, \mathrm{d}s} - e^{ \mu _{j}(y) + \int _{-\infty }^{-z} e_j \cdot \mathbf {W}_2({\mathcal T}_j(y),s) \, \mathrm{d}s} \\&= e^{ \mu _{j}(y) - \int _{-\infty }^{-z} e_j \cdot \mathbf {W}_1(y,-s) \, \mathrm{d}s} - e^{ \mu _{j}(y) - \int _{-\infty }^{-z} e_j \cdot \mathbf {W}_2(y,-s) \, \mathrm{d}s} \\&= e^{ \mu _{j}(y) - \int _{z}^{\infty } e_j \cdot \mathbf {W}_1(y,s) \, \mathrm{d}s} - e^{ \mu _{j}(y) - \int _{z}^{\infty } e_j \cdot \mathbf {W}_2(y,s) \, \mathrm{d}s} = w_+(y,z,z). \end{aligned} \end{aligned}
(5.4)

Therefore, $$w_+$$ satisfies (4.8), and $$w_-$$ satisfies

\begin{aligned}&( \square + 2 \mathbf {E}_-\cdot \nabla + f_-) w_- (y,z,t) \\&\quad = 2 e^{\mu _{j}(y)} (\mathbf {W}_2 -\mathbf {W}_1)( {\mathcal T}_j(y),-z) \cdot \nabla \big ( u_{2,j}( {\mathcal T}_j(y),-z,-t) \big ) \; \; \text {in} \; \; Q_- , \end{aligned}

where

\begin{aligned} \begin{aligned} \mathbf {E}_-(y,z)&= \mathbf {W}_1( {\mathcal T}_j(y),-z) + \nabla \mu _{j}(y) \\ f_- (y,z)&{=} V ( {\mathcal T}_j(y),{-}z) {-} |\nabla \mu _{j}(y)|^2 {+} \Delta \mu _{j}(y) {-}2\mathbf {W}_1( {\mathcal T}_j(y),{-}z) \cdot \nabla \mu _{j}(y). \end{aligned} \end{aligned}
(5.5)

Also, since $$e_j \cdot \mathbf {W}_1( {\mathcal T}_j(y),-z) = - e_j \cdot \mathbf {W}_1( y,z)$$, we have that

\begin{aligned}&(\partial _t + \partial _z -e_j \cdot \mathbf {W}_1( y,z)) w_- ( y,z,z) \\&\quad = - e^{\mu _j(y)} \big (\partial _t + \partial _z - e_j \cdot \mathbf {W}_1( {\mathcal T}_j(y),-z) \big ) w_+({\mathcal T}_j(y),-z,-z) \\&\quad = - e^{\mu _{j}(y)} e_j \cdot (\mathbf {W}_1-\mathbf {W}_2)({\mathcal T}_j(y),z) u_{2,j}({\mathcal T}_j(y),-z,-z). \end{aligned}

Since $$|u_2(y,z,z)| \gtrsim 1$$ always, and $$|\nabla u_2|$$ is bounded above in $$\overline{Q}_+$$, the previous identities show that

\begin{aligned} \begin{aligned} |(\square + 2 \mathbf {E}_\pm \cdot \nabla + f_\pm )w_\pm |&\lesssim |\mathbf {A}_\pm |&\quad&\text {in} \quad Q_\pm , \\ |(\partial _t + \partial _z -e_j \cdot \mathbf {E}_\pm ) w_\pm |&\gtrsim |e_j \cdot \mathbf {A}_\pm |&\quad&\text {in} \quad \Gamma , \end{aligned} \end{aligned}

where $$\mathbf {E}_-$$ and $$f_-$$ were defined in (5.5), $$\mathbf {E}_+ = \mathbf {W}_1$$, $$f_+ = V$$, and

\begin{aligned} \mathbf {A}_+ = \mathbf {W}_1-\mathbf {W}_2 \quad \text {and} \quad \mathbf {A}_-(y,z) = (\mathbf {W}_1-\mathbf {W}_2)({\mathcal T}_j(y),-z). \end{aligned}
(5.6)

From (5.4), we get that $$w_+ = w_-$$ on $$\Gamma$$. Also, we have that $$w_+|_{\Sigma _+} = w_-|_{\Sigma _-} = 0$$, and $$\partial _\nu w_\pm |_{{\Sigma _\pm }} = 0$$ by Lemma 2.9. Hence, we can use (3.18) of Lemma 3.5 with $$q_\pm = 0$$ and $$\omega = e_j$$, which yields

\begin{aligned} \sum _{\pm } \Vert e^{\sigma \phi _0} e_j \cdot \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \Vert e^{\sigma \phi _0} \mathbf {A}_\pm \Vert _{L^2(B)}^2. \end{aligned}

Now, this holds for $$\phi _0(x) = \phi _0(x,\vartheta ) = e^{\lambda \vert x-\vartheta \vert ^2}$$, where $$\vartheta$$ is an arbitrary vector such that $$|\vartheta | = 2$$. Also, the implicit constant in the estimate is independent of $$\vartheta$$. Therefore, writing $$\vartheta = 2\theta$$ for $$\theta \in S^{n-1}$$, we can integrate both sides of the previous estimate in $$S^{n-1}$$ to get

\begin{aligned}&\int _{S^{n-1}} \sum _{\pm } \Vert e^{\sigma \phi _0(\cdot ,2\theta )} e_j \cdot \mathbf {A}_\pm \Vert _{L^2(B)}^2 \, d S(\theta )\\&\quad \lesssim \gamma (\sigma ) \sum _{\pm } \int _{S^{n-1}} \Vert e^{\sigma \phi _0(\cdot ,2\theta )} \mathbf {A}_\pm \Vert _{L^2(B)}^2 \, \mathrm{d} S(\theta ), \end{aligned}

where $$d S(\theta )$$ denotes integration against the surface measure of the unit sphere. Changing the order of integration with the $$L^2(B)$$ integrals gives

\begin{aligned} \sum _{\pm } \Vert r(x,\sigma ) e_j \cdot \mathbf {A}_\pm \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \sum _{\pm } \Vert r(x,\sigma ) \mathbf {A}_\pm \Vert _{L^2(B)}^2 , \end{aligned}
(5.7)

where it can be verified that $$r(x,\sigma ) : = \left( \int _{S^{n-1}} e^{2\sigma \phi _0(x,2\theta )} \, \mathrm{d} S(\theta ) \right) ^{1/2}$$ is a radial function.

Since r(x) is a radial, and $$\mathbf {A}_+$$ and $$\mathbf {A}_-$$ differ in an orthogonal transformation by (5.6), a direct change of variables shows that

\begin{aligned} \Vert r(x,\sigma ) e_j \cdot \mathbf {A}_- \Vert _{L^2(B)}^2= & {} \Vert r(x,\sigma ) e_j \cdot \mathbf {A}_+ \Vert _{L^2(B)}^2, \quad \text {and} \quad \\ \Vert r(x,\sigma ) \mathbf {A}_- \Vert _{L^2(B)}^2= & {} \Vert r(x,\sigma ) \mathbf {A}_+ \Vert _{L^2(B)}^2 , \end{aligned}

and therefore, taking into account that $$\mathbf {A}_+ = \mathbf {W}_1-\mathbf {W}_2$$ we get from (5.7) that

\begin{aligned} \Vert r(x,\sigma ) e_j \cdot (\mathbf {W}_1-\mathbf {W}_2) \Vert _{L^2(B)}^2 \lesssim \gamma (\sigma ) \Vert r(x,\sigma ) (\mathbf {W}_1-\mathbf {W}_2) \Vert _{L^2(B)}^2 . \end{aligned}

This estimate can be proved for any $$1 \le j \le n$$. Adding over all directions, and using that $$\gamma (\sigma ) \rightarrow 0$$ as $$\sigma \rightarrow 0$$ to absorb the resulting term on the right-hand side in the left, yields

\begin{aligned} \Vert r(x,\sigma ) (\mathbf {W}_1-\mathbf {W}_2) \Vert _{L^2(B)}^2 \le 0 , \end{aligned}

for $$\sigma >0$$ large enough. Since $$r(x,\sigma )>0$$ for all $$x\in \mathbb {R}^n$$, and $$\sigma >0$$, the previous estimate implies that $$\mathbf {W}_1 =\mathbf {W}_2$$. This finishes the proof. $$\square$$

Combining the techniques used in this proof with the techniques used in the proof of Theorem 2.1, the reader can obtain some similar results to the previous one, always interchanging some measurements for symmetry assumptions on $$\mathbf {A}$$ and q. A specially simple case is the following one.

### Theorem 5.2

Let $$\mathbf {A}_1, \mathbf {A}_2 \in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$ and $$q_1,q_2 \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ with compact support in B, satisfying (1.6), and such that

\begin{aligned} \mathbf {A}_k(-x) = -\mathbf {A}_k(x) , \quad \text {for} \quad k=1,2. \end{aligned}
(5.8)

Let $$1 \le j \le n-1$$ and consider the $$n-1$$ solutions $$U_{k, j}(x,t)$$ of

\begin{aligned} ( \partial _t^2 + (D +\mathbf {A}_k )^2 + q_k )U_{k, j}=0 \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, j}|_{\left\{ t<-1\right\} } = \delta (t- x_j), \end{aligned}

and the 2 solutions $$U_{k, \pm n}(x,t)$$ of

\begin{aligned} ( \partial _t^2 + (D +\mathbf {A}_k )^2 + q_k )U_{k, \pm n}=0 \; \, \text {in}\; \mathbb {R}^{n+1}, \quad U_{k, \pm n}|_{\left\{ t<-1\right\} } = \delta (t- (\pm x_n)). \end{aligned}

If for each $$1 \le j \le n$$ one has $$U_{1, j} = U_{2, j}$$ on the surface $$\Sigma \cap \{ t \ge x_j \}$$, then $$d \mathbf {A}_1 = d \mathbf {A}_2$$ and $$q_1 = q_2$$.

### Proof

We use the notation introduced in (4.12). Notice that $$U_{k,j}$$ is the same as $$U_{k,+j}$$ in the proof of Theorem 2.1. We only give a sketch of the main ideas in the proof. One can start as in the proof of Theorem 2.1 by making a change of gauge such that $$e_n \cdot \mathbf {A}_1 = e_n \cdot \mathbf {A}_2 =0$$. One can verify that this choice of gauge does not alter the antisymmetry property (5.8).

Let $$0\le j \le n-1$$. By Lemma 2.6, we can assume that the $$U_{k,j} =U_{k,+j}$$ satisfy (4.9) and (4.10) with $$\mathbf {W}_k = -i\mathbf {A}_k$$ and $$V= V_k$$. We define $$w_+ = u_{1,j}-u_{2,j}$$ and

\begin{aligned} w_-(y,z,t) = e^{-i\mu _{j}(y)} w_+( -y,-z,-t) . \end{aligned}

It follows that $$w_+ = w_-$$ in $$\Gamma$$, and this can be verified using (5.8), in complete analogy with the computations in (5.4). The remaining conditions necessary to apply Lemma 3.5 with $$\mathbf {A}_+ = \mathbf {A}_1 -\mathbf {A}_2$$, $$\mathbf {A}_-(x) = \mathbf {A}_+(-x)$$, $$q_+ = V_1-V_2$$ and $$q_-(x) =q_+(-x)$$ can be verified as in the proof of Theorem 2.1. Then, Lemma 3.5 and a change of variables to transform $$q_-$$ in $$q_+$$, and $$\mathbf {A}_-$$ in $$\mathbf {A}_+$$ yields the estimate

\begin{aligned}&\Vert e^{\sigma \phi _0(\cdot , \vartheta )} (\mathbf {A}_1 - \mathbf {A}_2) \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0(\cdot , -\vartheta )} (\mathbf {A}_1 - \mathbf {A}_2) \Vert _{L^2(B)}^2 \nonumber \\&\quad \lesssim \gamma (\sigma ) \Vert e^{\sigma \phi _0(\cdot , \vartheta )} (V_1-V_2) \Vert _{L^2(B)}^2 + \gamma (\sigma ) \Vert e^{\sigma \phi _0(\cdot , -\vartheta )} (V_1-V_2) \Vert _{L^2(B)}^2. \end{aligned}
(5.9)

We can now repeat exactly the same arguments used in the proof of Theorem 2.1 to prove (4.20). In fact, we have that (4.20) holds independently for both the weight functions $$\phi _0(\cdot , \vartheta )$$ and $$\phi _0(\cdot ,- \vartheta )$$. Adding these two possible estimates yields

\begin{aligned}&\Vert e^{\sigma \phi _0(\cdot , \vartheta )} (V_1-V_2) \Vert _{L^2(B)}^2 + \Vert e^{\sigma \phi _0(\cdot , -\vartheta )} (V_1-V_2) \Vert _{L^2(B)}^2 \\&\quad \lesssim \gamma (\sigma )\Vert e^{\sigma \phi _0(\cdot , \vartheta )} (\mathbf {A}_1 - \mathbf {A}_2) \Vert _{L^2(B)}^2 + \gamma (\sigma ) \Vert e^{\sigma \phi _0(\cdot , -\vartheta )} (\mathbf {A}_1 - \mathbf {A}_2) \Vert _{L^2(B)}^2 . \end{aligned}

The previous inequality and (5.9) imply that $$q_1 = q_2$$ and $$\mathbf {A}_1 = \mathbf {A}_2$$ in the gauge fixed at the beginning of the proof. $$\square$$

## Notes

1. 1.

Throughout the paper, we write $$a \lesssim b$$ or equivalently $$b \gtrsim a$$, when a and b are positive constants and there exists $$C > 0$$ so that $$a \le C b$$. We refer to C as the implicit constant in the estimate.

## References

1. 1.

Barceló, J.A., Castro, C., Luque, T., Meroño, C.J., Ruiz, A., Vilela, M.C.: Uniqueness for the inverse fixed angle scattering problem. J. Inverse Ill-Posed Probl. 28(4), 465–470 (2020)

2. 2.

Bayliss, A., Li, Y., Morawetz, C.: Scattering by a potential using hyperbolic methods. Math. Comput. 52(186), 321–338 (1989)

3. 3.

Bellassoued, M., Yamamoto, M.: Carleman Estimates and Applications to Inverse Problems for Hyperbolic Systems. Springer Monographs in Mathematics. Springer (2017)

4. 4.

Bukhgeim, A., Klibanov, M.: Global uniqueness of class of multidimensional inverse problems. Soviet Math. Dokl. 24, 244–247 (1981)

5. 5.

Cristofol, M., Soccorsi, E.: Stability estimate in an inverse problem for non-autonomous Schrödinger equations. Appl. Anal. 90(10), 1499–1520 (2011)

6. 6.

Deift, P., Trubowitz, E.: Inverse scattering on the line. Commun. Pure Appl. Math 32(2), 121–151 (1979)

7. 7.

Evans, L.: Partial differential equations. Graduate Student in Mathematics, vol. 19, American Mathematical Society (1997)

8. 8.

Hörmander, L.: Linear Partial Differential Operators, 4th Printing. Springer-Verlag, Berlin (1976)

9. 9.

Hörmander, L.: The Analysis of Linear Partial Differential Operators, vol. I–IV. Springer-Verlag, Berlin (1983–1985)

10. 10.

Huang, X., Kian, Y., Soccorsi, E., Yamamoto, M.: Carleman estimate for the Schrödinger equation and application to magnetic inverse problems. J. Math. Anal. Appl. 474, 116–142 (2019)

11. 11.

Imanuvilov, O., Yamamoto, M.: Global uniqueness and stability in determining coefficients of wave equations. Commun. PDE 26, 1409–1425 (2001)

12. 12.

Klibanov, M.: Carleman estimates for global uniqueness, stability and numerical methods for coefficient inverse problems. J. Inverse Ill-Posed Probl. 21(4), 477–560 (2013)

13. 13.

Krishnan, V., Rakesh, Senapati, S.: Stability for a formally determined inverse problem for a hyperbolic PDE with space and time dependent coefficients. Preprint (2020), arXiv:2010.11726

14. 14.

Ma, S., Salo, M.: Fixed angle inverse scattering in the presence of a Riemannian metric. Preprint (2020), arXiv:2008.07329

15. 15.

Marchenko, V.: Sturm–Liouville Operators and Applications. Operator Theory: Advances and Applications. Birkhäuser, Basel (2011)

16. 16.

Melrose, R.: Geometric Scattering Theory. Lectures at Stanford Geometric scattering theory. Lectures at Stanford Geometric scattering theory. Cambridge University Press (1995)

17. 17.

Melrose, R., Uhlmann, G.: An Introduction to Microlocal Analysis. Book in preparation, http://www-math.mit.edu/rbm/books/imaast.pdf

18. 18.

Meroño, C.J.: Fixed angle scattering: recovery of singularities and its limitations. SIAM J. Math. Anal. 50(5), 5616–5636 (2018)

19. 19.

Rakesh, Salo, M.: Fixed angle inverse scattering for almost symmetric or controlled perturbations. SIAM J. Math. Anal. 52(6), 5467–5499 (2020)

20. 20.

Ruiz, A.: Recovery of the singularities of a potential from fixed angle scattering data. Commun. Partial Differ. Equ. 26(9–10), 1721–1738 (2001)

21. 21.

Rakesh, Salo, M.: The fixed angle scattering problem and wave equation inverse problems with two measurements. Inverse Probl. 36(3), 035005 (2020)

22. 22.

Shiota, T.: An inverse problem for the wave equation with first order perturbation. Am. J. Math. 107(1), 241–251 (1985)

23. 23.

Stefanov, P.: Generic uniqueness for two inverse problems in potential scattering. Commun. Partial Differ. Equ. 17(1–2), 55–68 (1992)

24. 24.

Teschl, G.: Mathematical Methods in Quantum Mechanics. Graduate studies in Mathematics, vol. 99, AMS (2009)

25. 25.

Uhlmann, G.: Uniqueness for the inverse backscattering problem for angularly controlled potentials. Inverse Probl. 30, 065004 (2014)

26. 26.

Uhlmann, G.: A time-dependent approach to the inverse backscattering problem. Inverse Probl. 17(4), 703–716 (2001)

27. 27.

Yafaev, D.R.: Mathematical Scattering Theory: Analytic Theory. Mathematical Surveys and Monographs, vol. 99, AMS (2010)

## Acknowledgements

C.M. was supported by project MTM2017-85934-C3-3-P. L.P. and M.S. were supported by the Academy of Finland (Finnish Centre of Excellence in Inverse Modelling and Imaging, grant numbers 312121 and 309963), and M.S. was also supported by the European Research Council under Horizon 2020 (ERC CoG 770924). The authors would like to thank the anonymous referees for their comments and suggestions that have helped to improve this paper.

## Funding

Open access funding provided by University of Jyväskylä (JYU).

## Author information

Authors

### Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Communicated by Jan Derezinski.

## Appendices

### Appendix A. Stationary Scattering

In this section, we prove Theorem 1.4. We have adapted the proof of [19, Theorem 5.1] in order to allow for the presence of a first-order perturbation, but the main ideas and the exposition are similar to the work in that paper.

We define $$\mathbb {C}^+: = \{ \lambda \in \mathbb {C}: \, {{\,\mathrm{\mathrm {Im}}\,}}(\lambda )>0 \}$$, and we write $$R_{\mathcal V}(\lambda )$$ for the resolvent operator $$R_{\mathcal V}(\lambda ) =(H_{\mathcal V} -\lambda ^2)^{-1}$$; in case it is well defined. We also use the following non-standard convention for the Fourier transform and its inverse for Schwartz functions on the real line:

\begin{aligned} \tilde{f} (\lambda ) = \int _{-\infty }^\infty e^{i \lambda t} f(t) \, \mathrm{d}t \qquad {\breve{F}} (t) = \frac{1}{2\pi }\int _{-\infty }^\infty e^{-i \lambda t} F(\lambda ) \, \mathrm{d}\lambda , \end{aligned}

(and equally for the extension of the Fourier transform to tempered distributions).

In order to illustrate why it is reasonable to expect an equivalence between the stationary scattering data and the time domain data as stated in Theorem 1.4, we reproduce here the following heuristic argument given in [19]. Let $$U_{\mathcal V}(x,t;\omega )$$ be the solution of

\begin{aligned} (\partial _t^2 - \Delta + \mathcal V (x,D) ) U_{\mathcal V} =0 \; \; \text {in} \; \mathbb {R}^n \times \mathbb {R}, \qquad U_{\mathcal V}|_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega ). \end{aligned}

Suppose for the moment that the Fourier transform of $$U_{\mathcal V}$$ in the time variable is well defined. Then, for each $$\lambda \in \mathbb {R}$$ the function $$\widetilde{U}_{\mathcal V}(x,\lambda ;\omega )$$ should solve the equation

\begin{aligned} (- \Delta + \mathcal V(x,D) -\lambda ^2 ) \widetilde{U}_{\mathcal V}(x,\lambda ) = 0 \; \; \text {in} \; \mathbb {R}^n. \end{aligned}
(A.1)

If we define the time domain scattering solution to be $$u_{\mathcal V} = U_{\mathcal V } - \delta (t-x\cdot \omega )$$, one has that $$\widetilde{U}_{\mathcal V}(x,\lambda ) = e^{i\lambda x\cdot \omega } + \widetilde{u}_{\mathcal V}(x,\lambda )$$, where $$\widetilde{u}_{\mathcal V}(x,\lambda )$$ extends holomorphically to $$\{ {{\,\mathrm{\mathrm {Im}}\,}}(\lambda )>0 \}$$ since $$u_{\mathcal V}$$ vanishes for $$t<-1$$. Since these are the properties that characterize the outgoing eigenfunctions of (A.1), one might expect that

\begin{aligned} \widetilde{U}_{\mathcal V}(x,\lambda ; \omega ) = \psi _{\mathcal V}(x,\lambda ,\omega ), \end{aligned}

where $$\psi _{\mathcal V}$$ is the solution of (1.1). Now, the condition $$a_{\mathcal V_1}(\lambda ,\cdot ,\omega ) = a_{\mathcal V_2}(\lambda ,\cdot ,\omega )$$ implies by the Rellich uniqueness theorem that the outgoing eigenfunctions for $$H_{\mathcal V_1}$$ and $$H_{\mathcal V_2}$$ agree outside the support of the potentials:

\begin{aligned} \psi _{\mathcal V_1} (\lambda ,\cdot ,\omega )|_{\mathbb {R}^n \setminus \overline{B} } = \psi _{\mathcal V_2}(\lambda ,\cdot ,\omega ) |_{\mathbb {R}^n \setminus \overline{B} } . \end{aligned}
(A.2)

If the map $$\lambda \rightarrow \psi _{\mathcal V}(\lambda ,x,\omega )$$ was smooth near $$\lambda =0$$, then one could have (A.2) for all $$\lambda \in \mathbb {R}$$. Taking the inverse Fourier transform would imply that

\begin{aligned} U_{\mathcal V_1}(\cdot ,t;\omega )|_{\mathbb {R}^n \setminus \overline{B} } = U_{\mathcal V_2}(\cdot ,t;\omega )|_{\mathbb {R}^n \setminus \overline{B} }. \end{aligned}

The argument above is only formal since requires the regularity of the map $$\lambda \rightarrow \psi _{\mathcal V}(\lambda ,x,\omega )$$ on the real line. The regularity of this map is related to the poles of the meromorphic continuation of the resolvent $$R_{\mathcal {V}}(\lambda )$$, initially defined in the resolvent set of $$H_{\mathcal {V}}$$. Indeed, in some cases there is a pole located at $$\lambda =0$$ and thus the argument above does not work in general. To get around these difficulties, we start by recalling the following property of the Fourier transform.

### Lemma A.1

Suppose F(z) is analytic on $$\{{{\,\mathrm{\mathrm {Im}}\,}}(z) >r\}$$ for some $$r\in \mathbb {R}$$ and

\begin{aligned} |F(z)| \le C(1 + |z|)^N e^{R {{\,\mathrm{\mathrm {Im}}\,}}(z)}, \quad \text {for} \;\; {{\,\mathrm{\mathrm {Im}}\,}}(z)>r, \end{aligned}

for some positive RCN independent of z. There exist an $$f \in \mathcal D'(\mathbb {R})$$ with $${{\,\mathrm{\mathrm {supp}}\,}}(f)\subset [-R,\infty )$$ and $$e^{-(\mu -r)t} f \in \mathcal S'(\mathbb {R})$$ that also satisfies $$(e^{-(\mu -r)t} f)^{\sim }(\cdot ) =F(\cdot + i\mu )$$ for every $$\mu >r$$.

This is essentially a Paley–Wiener theorem that we have stated in the form given in [19, Lemma 5.3]. In the following proposition, we give the precise relation between the time domain and frequency measurements.

### Proposition A.2

Let $$\omega \in S^{n-1}$$, and let $${\mathcal V}(x,D) = \mathbf {W}\cdot \nabla + V$$ with $$\mathbf {W}\in C^{m+2}_c(\mathbb {R}^n; \mathbb {C}^n)$$, and $$V \in C^{m}_c(\mathbb {R}^n; \mathbb {C})$$ compactly supported in B. Let $$U_{\mathcal V}$$ be the solution of

\begin{aligned} (\partial _t^2 - \Delta + \mathcal V (x,D) ) U_{\mathcal V} =0 \; \; \text {in} \; \mathbb {R}^n \times \mathbb {R}, \qquad U_{\mathcal V}|_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega ),\nonumber \\ \end{aligned}
(A.3)

given by Proposition 2.4, and let $$u_{\mathcal V}(x,t;\omega ) = U_{\mathcal V} (x,t;\omega ) -\delta (t-x \cdot \omega )$$. Assume also that there exists some $$r \ge 0$$ such that for $${{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r$$

\begin{aligned} \Vert R_{\mathcal V}(\lambda ) \Vert _{L^2\rightarrow L^2} \le C_{r} , \end{aligned}
(A.4)

where $$C_r>0$$ is independent of $$\lambda$$. Then, if we define

\begin{aligned} \psi ^s_{\mathcal V}(\cdot , \lambda , \omega ):=- R_{\mathcal V}(\lambda )({\mathcal V}(x,D) e^{i\lambda x\cdot \omega } ), \quad \text {for } \; {{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) > r, \end{aligned}

the following identity holds

\begin{aligned} \langle u_{\mathcal V} (x,t; \omega ), \varphi (x)\chi (t) \rangle _{\mathbb {R}^n_x \times \mathbb {R}_t}= \langle \psi ^s_{\mathcal V}(x, \sigma + i\mu , \omega ), \varphi (x) (e^{\mu t}\chi )^\vee (\sigma ) \rangle _{\mathbb {R}^n_x \times \mathbb {R}_\sigma }, \end{aligned}

for all $$\mu > r$$, and all $$\varphi \in C^\infty _c(\mathbb {R}^n)$$ and $$\chi \in C^\infty _c(\mathbb {R})$$.

### Proof

If (A.4) holds for some $$\lambda$$, then $$z= \lambda ^2$$ is, by definition, in the resolvent set $$\rho (H_{\mathcal V})\subset \mathbb {C}$$ of the operator $$H_{\mathcal V}$$. It is well known that the resolvent map $$z \rightarrow (H_{\mathcal V} -z)^{-1}$$ forms a holomorphic family of bounded $$L^2$$ operators in the open set $$\rho (H_{\mathcal V})$$ (see, for example, [24, Theorem 2.15]). Since the map $$z= \lambda ^2$$ is also holomorphic and (A.4) holds for $${{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r$$, $$\lambda \rightarrow R_{\mathcal V}(\lambda ) = (H_{\mathcal V} -\lambda ^2)^{-1}$$ is also an holomorphic map for $${{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r$$.

On the other hand, using (A.4) we have

\begin{aligned} \left\| \psi _{\mathcal V}^s (\cdot , \lambda , \omega )\right\| _{L^2(\mathbb {R}^n)}\le C_{r, A, q}(1+ |\lambda |) e^{{{\,\mathrm{\mathrm {Im}}\,}}(\lambda )}, \quad {{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r. \end{aligned}
(A.5)

For any fixed $$\varphi \in C^\infty _x(\mathbb {R}^n)$$, define

\begin{aligned} F_\varphi (\lambda )= \int _{\mathbb {R}^n} \psi ^s_{\mathcal V}(x, \lambda , \omega )\varphi (x)\mathrm{d}x, \quad {{\,\mathrm{\mathrm {Im}}\,}}(\lambda )\ge r. \end{aligned}

From the previous observations, it follows that $$F_\varphi (\lambda )$$ must be an holomorphic function in the set $${{\,\mathrm{\mathrm {Im}}\,}}(\lambda )\ge r$$. By estimate (A.5), we get

\begin{aligned} |F_{\varphi }(\lambda )|\le C_{r, \mathbf {A}, q}(1+ |\lambda |) e^{{{\,\mathrm{\mathrm {Im}}\,}}(\lambda )}\left\| \varphi \right\| _{L^2}, \quad {{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r. \end{aligned}

Then, Lemma A.1 implies that there is a function $$f_\varphi \in \mathcal {D}^\prime (\mathbb {R})$$ supported on $$[-1, \infty )$$ such that for all $$\mu >r$$:

\begin{aligned} \langle e^{-(\mu -r)t}f_\varphi , \chi \rangle = \langle F_\varphi (\cdot + i\mu ), \breve{\chi } \rangle , \quad \chi \in C^\infty _c(\mathbb {R}). \end{aligned}

Now, given $$\mu >r$$, define the linear map $$\mathcal K: C^\infty _c(\mathbb {R}^n) \rightarrow \mathcal D'(\mathbb {R})$$ given by

\begin{aligned} \mathcal K \varphi = e^{-(\mu - r)t} f_{\varphi }. \end{aligned}

The map $$\mathcal {K}$$ is continuous. To see this, take a sequence $$\varphi _j \rightarrow 0$$ in $$C^\infty _c(\mathbb {R}^n)$$, then (A.5) implies that $$F_{\varphi _j} \rightarrow 0$$ when $${{\,\mathrm{\mathrm {Im}}\,}}(\lambda ) \ge r$$, and hence

\begin{aligned} \langle \mathcal K \varphi _j , \chi \rangle = \langle e^{-(\mu -r)t}f_{\varphi _j}, \chi \rangle = \langle F_{\varphi _j} (\cdot + i\mu ), \breve{\chi } \rangle \rightarrow 0 \quad {as} \;\; j\rightarrow \infty . \end{aligned}

Since $$\mathcal {K}$$ is continuous, the Schwartz kernel theorem ensures that there is a unique $$K \in \mathcal {D}'(\mathbb {R}^n\times \mathbb {R})$$ such that

\begin{aligned} \langle K, \varphi (x)\chi (t)\rangle&= \nonumber \langle \mathcal K \varphi , \chi \rangle = \langle e^{-(\mu -r)t} f_\varphi , \varphi (x), \chi \rangle = \langle F_\varphi (\cdot + i\mu ) ,\breve{\chi } \rangle \nonumber \\&= \langle \psi ^s_{\mathcal V}(x, \sigma +i\mu , \omega ) , \varphi (x) \breve{\chi }(\sigma )\rangle _{\mathbb {R}^n_x \times \mathbb {R}_\sigma }. \end{aligned}
(A.6)

Since $$f_\varphi$$ is supported in $$[-1,\infty )$$, it follows that K is supported in $$\{t\ge 1\}$$. We now define the distribution

\begin{aligned} v(x,t) := e^{\mu t} K(x,t) \in \mathcal {D}^\prime (\mathbb {R}^n\times \mathbb {R}). \end{aligned}

We claim that v is a solution in $$\mathbb {R}^{n+1}$$ of

\begin{aligned} (\square + {\mathcal V}(x,D))v = -{\mathcal V}(x,D) \delta (t-x\cdot \omega ) . \end{aligned}
(A.7)

Since, by (A.3), this is also the equation satisfied by $$u_{\mathcal V}$$, the uniqueness of distributional solutions of the wave equation supported in $$\{t\ge -1\}$$ (see [8, Theorem 9.3.2]) implies that $$u_{\mathcal V} = v$$, so

which finishes the proof of the Proposition.

To prove the previous claim, we use (A.6) in the following computations. First,

Also, if we denote by $$\mathcal V^*$$ the formal adjoin of $$\mathcal V$$ (with respect to the distribution pairing $$\langle \cdot ,\cdot \rangle$$), we have

and similarly, one gets

Then, putting this together we obtain that

Hence, v satisfies (A.7), which proves the claim. $$\square$$

The following proposition gives in the self-adjoint case the properties of the resolvent that we require to apply the previous proposition. Therefore, we assume that $${\mathcal V}(x,D)$$ can be written as $${\mathcal V}(x,D) = 2 \mathbf {A}\cdot D+ D\cdot \mathbf {A}+ q$$ for real $$\mathbf {A}$$ and q.

### Proposition A.3

Let $$\mathbf {A}\in C^1_c(\mathbb {R}^n, \mathbb {R}^n)$$, $$q\in C^1_c(\mathbb {R}^n, \mathbb {R})$$, and fix $$r_0 = (2\Vert \mathbf {A} \Vert _{L^\infty }^2 + \Vert q \Vert _{L^\infty })^{1/2}$$. For any $$\lambda \in \mathbb {C}_+\setminus i(0, r_0]$$, there is a bounded operator

\begin{aligned} R_{ \mathbf {A},q}(\lambda ): L^2(\mathbb {R}^n) \rightarrow L^2(\mathbb {R}^n) \end{aligned}

such that for any $$f\in L^2(\mathbb {R}^n)$$, the function $$u=R_{ \mathbf {A},q}(\lambda )f$$ is the unique solution in $$L^2(\mathbb {R}^n)$$ of

\begin{aligned} (H_{ \mathbf {A},q}-\lambda ^2)u=f. \end{aligned}

Moreover, if $$r>r_0$$, one has

\begin{aligned} \left\| R_{ \mathbf {A},q}(\lambda ) \right\| _{L^2\rightarrow L^2} \le C_{r, \mathbf {A},q}, \quad {{\,\mathrm{\mathrm {Im}}\,}}(\lambda )\ge r. \end{aligned}
(A.8)

For any $$\delta >1 /2$$ and $$\lambda$$ in the region $$\mathbb {C}_+\setminus i(0, r_0]$$, the family

\begin{aligned} \left( \langle x \rangle ^{-\delta } R_{ \mathbf {A},q}(\lambda ) \langle x \rangle ^{-\delta } \right) _{\lambda \in \mathbb {C}_+\setminus (0, r_0]} \end{aligned}

is a holomorphic family of bounded operators on $$L^2(\mathbb {R}^n)$$ that can be extended continuously in the weak operator topology to $$\overline{\mathbb {C}}_+\setminus i[0, r_0]$$.

Let $$\lambda \ge 0$$. As mentioned in the introduction, the direct problem (1.1) will have a unique scattering solution $$\psi ^s_{\mathcal V}$$ satisfying the SRC if the outgoing resolvent operator

\begin{aligned} (H_{\mathcal V} -(\lambda ^2 + i0) )^{-1} = \lim _{\varepsilon \rightarrow 0} (H_{\mathcal V} -(\lambda ^2 + i\varepsilon ) )^{-1} , \end{aligned}

is bounded in appropriate spaces. Then, one can take

\begin{aligned} \psi ^s_{\mathcal V} = (H_{\mathcal V} -(\lambda ^2 + i0)^2)^{-1}(-{\mathcal V}(x,D) e^{i\lambda \omega \cdot x}), \end{aligned}

as solution of (1.1). Under the assumptions of the previous proposition, for real $$\lambda >0$$ the operator $$R_{ \mathbf {A},q}(\lambda )$$ given by the continuous extension to $$\overline{\mathbb {C}}_+\setminus i[0, r_0]$$ of the resolvent is exactly the outgoing resolvent operator. On the other hand, if $$\lambda <0$$ the resolvent $$R_{ \mathbf {A},q}(\lambda )$$ is the operator known as the incoming resolvent operator. This can be seen in the following computation: For real $$\lambda \ne 0$$, taking the limits in an appropriate topology, one has

\begin{aligned} R_{ \mathbf {A},q}(\lambda )= & {} \lim _{\varepsilon \rightarrow 0} R_{ \mathbf {A},q}(\lambda +i\varepsilon ) = \lim _{\varepsilon \rightarrow 0} (H_{ \mathbf {A},q} -(\lambda +i\varepsilon )^2)^{-1} \\= & {} \lim _{\varepsilon \rightarrow 0} (H_{ \mathbf {A},q} -\lambda ^2 -i2\lambda \varepsilon -\varepsilon ^2 )^{-1} = (H_{ \mathbf {A},q} -(\lambda ^2 \pm i0))^{-1}, \end{aligned}

where the ± is given by the sign of $$\lambda$$.

### Proof of Proposition A.3

We have that

\begin{aligned} H_{ \mathbf {A},q} = (D+ \mathbf {A})^2 +q = -\Delta + {\mathcal V}(x,D),\end{aligned}

where $${\mathcal V}(x,D) = 2 \mathbf {A}\cdot D+ D\cdot \mathbf {A}+ q$$. The operator $$H_{ \mathbf {A},q}$$ is self-adjoint with domain $$H^2(\mathbb {R}^n)$$, and as such, it has real spectrum: The resolvent $$R_{ \mathbf {A},q}(\lambda )$$ is a bounded operator in $$L^2$$ if $$\lambda \in \mathbb {C}_+$$ satisfies $$\lambda ^2 \notin \mathbb {R}$$. Also, since $$\mathbf {A}$$ and q are compactly supported—$${\mathcal V}(x,D)$$ is a short range perturbation of $$-\Delta$$—it is well known that the continuous spectrum of $$H_{ \mathbf {A},q}$$ is $$(0,\infty )$$ without embedded eigenvalues (see, for example, [9, Chapter 14]). We now show that the point spectrum of $$H_{ \mathbf {A},q}$$ is contained in $$[-r_0^2,0]$$ where $$r_0^2= 2\Vert \mathbf {A} \Vert _{L^\infty }^2+ \Vert q \Vert _{L^\infty }$$. Indeed, assume that $$\lambda ^2 \in \mathbb {R}$$ and $$\psi \in L^2$$ are such that

\begin{aligned} H_{ \mathbf {A},q}\psi = \lambda ^2 \psi , \end{aligned}

is satisfied in the sense of distributions. Then, $$\psi \in H^2(\mathbb {R}^n)$$ by elliptic regularity, so taking the $$L^2$$ product with $${\psi }$$ and integrating by parts give us

\begin{aligned} \lambda ^2 \Vert \psi \Vert ^2= & {} \Vert \nabla \psi \Vert ^2 + ( \mathbf {A}\cdot D\psi , \psi )_{L^2} + ( \mathbf {A}\psi , D\psi )_{L^2} + ( ( \mathbf {A}^2+q)\psi , \psi )_{L^2}\\\ge & {} \Vert \nabla \psi \Vert ^2 - 2 \Vert \mathbf {A} \Vert _{L^\infty } \Vert \psi \Vert \Vert \nabla \psi \Vert - \Vert q \Vert _{L^\infty } \Vert \psi \Vert ^2\\\ge & {} \frac{1}{2}\Vert \nabla \psi \Vert ^2 - (2\Vert \mathbf {A} \Vert _{L^\infty }^2 + \Vert q \Vert _{L^\infty })\Vert \psi \Vert ^2, \end{aligned}

where we have used that $$\mathbf {A}^2 \ge 0$$ and Young’s inequality with $$\varepsilon$$. Hence,

\begin{aligned} (\lambda ^2 +2\Vert \mathbf {A} \Vert _{L^\infty }^2 + \Vert q \Vert _{L^\infty })\Vert \psi \Vert ^2 \ge \frac{1}{2}\Vert \nabla \psi \Vert ^2 \ge 0, \end{aligned}

and thus, we must necessarily have $$\lambda ^2 \ge -(2\Vert \mathbf {A} \Vert _{L^\infty }^2 + \Vert q \Vert _{L^\infty })$$. With this, we can conclude that the full spectrum of $$H_{ \mathbf {A},q}$$ is contained in $$[-r_0^2,\infty )$$, so that $$R_{ \mathbf {A},q}(\lambda )$$ is a bounded operator in $$L^2$$ for all $$\lambda \in \mathbb {C}_+ \setminus i(0, r_0]$$. Then, the theory of self-adjoint operators implies two important facts.

The first is that one has the estimate

\begin{aligned} \Vert R_{ \mathbf {A},q}(\lambda ) \Vert _{L^2 \rightarrow L^2} \le \frac{1}{{{\,\mathrm{dist}\,}}(\lambda ^2,[-r_0^2,\infty ) )} , \end{aligned}
(A.9)

(see, for example, [24, Theorem 2.15]). And the second is that $$R_{ \mathbf {A},q}(\lambda ): L^2 \rightarrow H^2$$ is an holomorphic family of operators for $$\lambda \in \mathbb {C}_+$$. This last statement follows from the fact that, outside the spectrum, for all $$\lambda ,\lambda _0 \in \mathbb {C}_+ \setminus i(0, r_0]$$ one has the resolvent formula

\begin{aligned} R_{ \mathbf {A},q}(\lambda ) = R_{ \mathbf {A},q}(\lambda _0) \left( \sum _{j=0}^{m} (\lambda ^2-\lambda ^2_0)^j R_{ \mathbf {A},q}^{j}(\lambda _0) + (\lambda ^2-\lambda ^2_0)^{m+1} R_{ \mathbf {A},q}^{m}(\lambda _0)R_{ \mathbf {A},q}(\lambda )\right) , \end{aligned}

see, for example, [24, p. 75]. One can take the limit $$m \rightarrow \infty$$ in the $$L^2 \rightarrow L^2$$ operator norm to obtain an analytic expansion of the resolvent around $$\lambda _0$$, since the remainder of the series goes to zero if $$\lambda$$ is close enough to $$\lambda _0$$. Then,

\begin{aligned} R_{ \mathbf {A},q}(\lambda ) = R_{ \mathbf {A},q}(\lambda _0) \sum _{j=0}^{\infty } (\lambda ^2-\lambda ^2_0)^j R_{ \mathbf {A},q}^{j}(\lambda _0), \end{aligned}
(A.10)

for $$\lambda$$ close enough to $$\lambda _0$$. Since $$R_{ \mathbf {A},q}(\lambda )$$ is also bounded from $$L^2$$ to $$H^2$$, (A.10) implies that $$R_{ \mathbf {A},q}(\lambda ) : L^2 \rightarrow H^2$$ is holomorphic in $$\mathbb {C}_+ \setminus i(0, r_0]$$, and therefore, so it is

\begin{aligned} \langle x \rangle ^{-\delta } R_{ \mathbf {A},q}(\lambda ) \langle x \rangle ^{-\delta } : L^2 \rightarrow H^2 \end{aligned}

since the weight $$\langle x \rangle ^{-\delta }$$ and all its derivatives are bounded in $$\mathbb {R}^n$$.

We now prove (A.8) using (A.9). To see this, write $$\lambda = \sigma + i \mu$$. Then, it is enough to use that $${{\,\mathrm{dist}\,}}(\lambda ^2,[-r_0^2,\infty ) ) \ge |{{\,\mathrm{\mathrm {Im}}\,}}(\lambda ^2)| \ge 2|\sigma \mu |$$ when $$\mu ^2 \ge \sigma ^2-r_0$$, and that $${{\,\mathrm{dist}\,}}(\lambda ^2,[-r_0^2,\infty ) ) \ge |{{\,\mathrm{\mathrm {Re}}\,}}(\lambda ^2)| = |\sigma ^2-\mu ^2- r_0|$$ otherwise.

The continuity of $$\langle x \rangle ^{-\delta } R_{ \mathbf {A},q}(\lambda ) \langle x \rangle ^{-\delta }$$ for $$\lambda \in \overline{\mathbb {C}}_+ \setminus i[0,r_0]$$ is the well-known limiting absorption principle. See, for example, [27, Proposition 1.7.1] for the free resolvent, and [9, Chapter 14] for the case of short range magnetic potentials (as in this case). A more specific statement of the limiting absorption principle (also including long range magnetic potentials) can be found in [9, Theorem 30.2.10], which implies that $$R_{ \mathbf {A},q}(\lambda )$$ is continuous as a function from $$\overline{\mathbb {C}}_+ \setminus i[0,r_0]$$ to the space of bounded operators between the Hörmander spaces B and $$B^*$$ considered with the weak operator topology. Since B is continuously embedded in $$\langle x \rangle ^{-\delta } L^2$$, this implies the (weak) continuity of $$\langle x \rangle ^{-\delta } R_{ \mathbf {A},q}(\lambda ) \langle x \rangle ^{-\delta }$$ in $$\overline{\mathbb {C}}_+ \setminus i[0,r_0]$$. $$\square$$

Putting together Propositions A.2 and A.3, we can now formalize the heuristic argument given at the beginning of this section in order to prove Theorem 1.4.

### Proof of Theorem 1.4

Let $$r_0 = \max _{k=1,2}(2\Vert \mathbf {A}_k \Vert _{L^\infty }^2 + \Vert q_k \Vert _{L^\infty })^{1/2}$$. By Proposition A.3, for all $$\lambda \in \overline{\mathbb {C}}_+ \setminus i[0,r_0]$$ we can define

\begin{aligned} \psi _{\mathbf {A}_k,q_k}^s(\cdot ,\lambda ,\omega ) =- R_{\mathbf {A}_k,q_k} ((\mathbf {A}_k^2 + D \cdot \mathbf {A}_k + 2\lambda \omega \cdot A + q_k )e^{i\lambda x \cdot \omega } ). \end{aligned}

Assume first that $$a_{\mathbf {A}_1,q_1}(\lambda ,\theta ,\omega ) = a_{\mathbf {A}_2,q_2}(\lambda ,\theta ,\omega )$$ for all $$\lambda \in \mathbb {R}$$ such that $$\lambda \ge \lambda _0$$ and all $$\theta \in S^{n-1}$$. Recall that $$a_{\mathbf {A}_k,q_k}(\lambda ,\theta ,\omega )$$ are defined by the asymptotic expansion

\begin{aligned} \psi ^s _{\mathbf {A}_k,q_k}(r\theta ,\lambda ,\omega ) = e^{i\lambda r} r^{-\frac{n-1}{2}} a_{\mathbf {A}_k,q_k}(\lambda ,\theta ,\omega )+o(r^{-\frac{n-1}{2}}), \qquad r\rightarrow \infty \; \; k=1,2 . \end{aligned}

Therefore, since $$\mathbf {A}_k$$ and $$q_k$$ are supported in B, for any fixed $$\lambda \ge \lambda _0$$ the function $$\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2}$$ satisfies

\begin{aligned} (-\Delta - \lambda )(\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2}) (\cdot ,\lambda ,\omega )&= 0 \quad \text {in} \; \; \mathbb {R}^n \setminus \overline{B} \\ (\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2}) (x,\lambda ,\omega )&= o(|x|^{-\frac{n-1}{2}}) \quad \text {as} \; \; |x|\rightarrow \infty . \end{aligned}

The Rellich uniqueness theorem implies that $$\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2}$$ vanishes outside $$\overline{B}$$. In particular, for any $$\varphi \in C^\infty _c(\mathbb {R}^n\setminus \overline{B})$$, the function

\begin{aligned} w_{\varphi }|_{[\lambda _0,\infty )}(\lambda ) = \langle (\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2})(\cdot ,\lambda ,\omega ) ,\varphi \rangle _{\mathbb {R}^n_x} \end{aligned}
(A.11)

satisfies

\begin{aligned} w_{\varphi }|_{[\lambda _0,\infty )} =0. \end{aligned}

By Proposition A.3, the map $$\lambda \rightarrow w_{\varphi }$$ is holomorphic in $${\mathbb {C}}_+ \setminus i(0,r_0]$$ and continuous in $$\overline{\mathbb {C}}_+\setminus i[0,r_0]$$. Since it vanishes on $$[\lambda _0,\infty )$$, we must have $$w_\varphi =0$$. In particular, for any $$\mu > r_0$$ and $$\sigma \in \mathbb {R}$$ one has

\begin{aligned} \langle (\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2})(x,\sigma + i\mu ,\omega ) ,\varphi (x)\rangle = 0. \end{aligned}

Then, Proposition A.2 implies that

\begin{aligned} \langle (u_{\mathbf {A}_1,q_1} -u_{\mathbf {A}_2,q_2})(x,t ;\omega ) ,\varphi (x)\chi (t)\rangle _{\mathbb {R}^n_x \times \mathbb {R}_t} = 0 \end{aligned}

for all $$\varphi \in C^\infty _c(\mathbb {R}^n\setminus \overline{B})$$ and $$\chi \in C^\infty _c(\mathbb {R})$$. Therefore,

\begin{aligned} (u_{\mathbf {A}_1,q_1} -u_{\mathbf {A}_2,q_2})(x,t ;\omega ) = 0 \quad (x,t) \in \mathbb {R}^n \times \overline{B}. \end{aligned}

Let us prove the converse statement. As in the previous sections, we take any orthonormal coordinate system in $$\mathbb {R}^n$$ such that $$x= (y,z)$$, where $$z = x\cdot \omega$$ and $$y \in \{ \omega \}^{\perp }$$ (identified with $$\mathbb {R}^{n-1}$$).

Assume that $$u_{\mathbf {A}_1,q_1}(x,t; \omega ) = u_{\mathbf {A}_2,q_2}(x,t; \omega )$$ for $$(x,t) \in (\partial B \times \mathbb {R}) \cap \{ t \ge z \}$$. By Proposition 2.4, the function $$\alpha = u_{\mathbf {A}_1,q_1} -u_{\mathbf {A}_2,q_2}$$ solves

\begin{aligned} \square \alpha =0 \qquad \text {in} \;\; \{(x,t): |x|>1 \; \text {and} \; t>z\}, \end{aligned}

and $$\alpha |_{(\partial B \times \mathbb {R}) \cap \{ t \ge z \}} = 0$$ and satisfies all the conditions required to apply Lemma 2.9. Thus, this lemma yields that one also has $$\partial _\nu \alpha |_{(\partial B \times \mathbb {R}) \cap \{ t \ge z \}} = 0$$. Now, the Cauchy data of $$\alpha$$ vanish on the lateral boundary of the set $$\{(x,t): |x| \ge 1 \text { and } t \ge z\}$$, and Holmgren’s uniqueness theorem applied in this set shows that $$\alpha$$ is identically zero in the relevant domain of dependence. However, by finite speed of propagation the support of $$\alpha$$ is contained in the same domain of dependence. Thus, $$\alpha$$ is identically zero in $$\{(x,t): |x| \ge 1 \text { and } t \ge z\}$$, which implies that

\begin{aligned}u_{\mathbf {A}_1,q_1}(x,t; \omega ) = u_{\mathbf {A}_2,q_2}(x,t; \omega ), \qquad (x,t) \in (\mathbb {R}^n \setminus \overline{B}) \times \mathbb {R}.\end{aligned}

The relation in Proposition A.2 gives that for any $$\mu > r_0$$ and for any $$\varphi \in C^\infty _c(\mathbb {R}^n \setminus \overline{B})$$

\begin{aligned} w_\varphi (\sigma +i\mu ) = \langle (\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2})(x,\sigma +i\mu ,\omega ) ,\varphi (x)\rangle _{\mathbb {R}^n_x} = 0 \qquad \sigma \in \mathbb {R}, \end{aligned}

using the notation in (A.11). A mentioned previously, $$w_\varphi$$ is holomorphic in $${\mathbb {C}}_+ \setminus i(0,r_0]$$ and has a continuous extension to $$\overline{\mathbb {C}}_+ \setminus i[0,r_0]$$ so, in particular, it follows that

\begin{aligned} w_\varphi (\lambda ) = \langle (\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2})(\cdot ,\lambda ,\omega ) ,\varphi \rangle _{\mathbb {R}^n_x} = 0 \end{aligned}

for all $$\lambda >0$$. Thus, $$(\psi _{\mathbf {A}_1,q_1} -\psi _{\mathbf {A}_2,q_2})(\cdot ,\lambda ,\omega )$$ vanishes outside $$\overline{B}$$ for any $$\lambda >0$$. By the asymptotics given in (1.3), this implies that $$a_{\mathbf {A}_1,q_1}(\lambda ,\theta ,\omega )=a_{\mathbf {A}_2,q_2}(\lambda ,\theta ,\omega )$$ for all $$\lambda >0$$ and $$\theta \in S^{n-1}$$. $$\square$$

### Appendix B. Some Results Concerning the Wave Operator

The first part of this section is devoted to the proof of Propositions 2.3 and 2.4. In the second part, we prove Lemmas 2.72.9.

### B.1. Existence and Uniqueness of Solutions

Here, we complete the proof of Propositions 2.3 and 2.4. Let $$\omega \in S^{n-1}$$. In what follows, we write $$\partial _z=\omega \cdot \nabla$$. We first prove the existence and uniqueness of distributional solutions to

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) U_\delta =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad U_\delta |_{\left\{ t<-1\right\} }= \delta (t-x\cdot \omega ) \end{aligned}
(B.1)

and

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V}) U_H =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad U_H|_{\left\{ t<-1\right\} }= H (t-x\cdot \omega ). \end{aligned}
(B.2)

We start by proving the uniqueness. In both cases, it is reduced to prove that zero is the unique distributional solution to the following homogenous equation

\begin{aligned} (\partial _t^2 + L_{\mathbf {W},V})\, \mathcal {U} =0 \; \text {in}\; \mathbb {R}^{n+1}, \quad \mathcal {U} \, |_{\left\{ t<-1\right\} }=0, \end{aligned}

which is true by [9, Theorem 23.2.7]. Let us now prove the existence of solutions. The method we shall use is the so-called progressing wave expansion method, see, for example, [22, Lemma 1] and [25, Theorem 1]. For any $$j\ge 0$$, define

\begin{aligned} s_+^j = \left\{ \begin{matrix} s^j , &{}s\ge 0, \\ 0 , &{}s<0. \end{matrix}\right. \end{aligned}
(B.3)

Note that $$s^0_+=H(s)$$ is the unidimensional Heaviside function at $$s\in \mathbb {R}$$. Let $$N\in \mathbb {N}$$ and suppose that the solutions to $$(\partial _t^2 + L_{\mathbf {W},V}) U =0$$ have the following expansion

\begin{aligned} U (x,t)= a_{-1}(x) \delta (t-x\cdot \omega ) + \sum _{j=0}^N a_j (x)(t- x\cdot \omega )_+^j + R_N(x,t). \end{aligned}
(B.4)

In the case of (B.1), the coefficients $$(a_j)_{j=-1}^N$$ and the remainder term $$R_N$$ must satisfy the following initial value conditions

\begin{aligned} R_N|_{t<-1}=0, \quad a_{-1}|_{x\cdot \omega<-1}=1, \quad a_{j}|_{x\cdot \omega <-1}=0, \quad j=0, 1, \dots , N \end{aligned}

and in the case of (B.2), the conditions

\begin{aligned} R_N|_{t<-1}=0, \quad a_{0}|_{x\cdot \omega<-1}=1, \quad a_{j}|_{x\cdot \omega <-1}=0, \quad j=-1, 1, \dots , N. \end{aligned}

A straightforward computation shows that the remainder term $$R_N$$ must satisfy

\begin{aligned} \begin{aligned} (\partial _t^2+ L_{\mathbf {W},V})R_N(x,t)&= -2\left( ( \partial _z- \omega \cdot \mathbf {W})a_{-1}\right) \partial _t \delta (t-x\cdot \omega )\\&\quad -\, \left( 2(\partial _z-\omega \cdot \mathbf {W})a_0+L_{\mathbf {W},V}a_{-1}\right) \delta (t-x\cdot \omega )\\&\quad -\, \sum _{j=0}^{N-1}\big (2(j+1)( \partial _z- \omega \cdot \mathbf {W})a_{j+1}\\&\quad + L_{\mathbf {W},V} \, a_j \big )(t-x\cdot \omega )^j_+ -\, \left( L_{\mathbf {W},V}\, a_N\right) (t-x\cdot \omega )^N_+. \end{aligned} \end{aligned}
(B.5)

The task now is to prove the existence of the coefficients $$(a_j)_{j=-1}^N$$ and $$R_N$$, satisfying the recursive identity (B.5) with the corresponding initial value conditions. One expects getting smoother remainder terms $$R_N$$ as N grows, or at least with better regularity than the Delta distribution and Heaviside function. This can be achieved by killing most the non-smooth terms on the right-hand side of (B.5). We now split the proof into two cases depending on the nature of the initial value conditions.

First case. Existence of solutions of (B.1). Above discussion motivates choosing the recursive formulae

\begin{aligned} (\partial _z- \omega \cdot \mathbf {W})a_{-1}&=0,&\quad&\text {in}\; \mathbb {R}^n,&\quad a_{-1}|_{x\cdot \omega <-1}&=1, \end{aligned}
(B.6)
\begin{aligned} (\partial _z- \omega \cdot \mathbf {W})a_{0}&=-\frac{1}{2} L_{\mathbf {W},V}a_{-1}&\quad&\text {in}\; \mathbb {R}^n,&\quad a_{0}|_{x\cdot \omega<-1}&=0, \nonumber \\ ( \partial _z- \omega \cdot \mathbf {W})a_{k+1}&=-\frac{1}{2(k+1)} L_{\mathbf {W},V}\, a_k,&\quad&\text {in}\; \mathbb {R}^n,&\quad a_{k+1}|_{x\cdot \omega <-1}&=0, \end{aligned}
(B.7)
\begin{aligned} (\partial _t^2+ L_{\mathbf {W},V})R_N(x,t)&= -\left( L_{\mathbf {W},V}\, a_N\right) (t-x\cdot \omega )^N_+,&\quad&\text {in}\; \mathbb {R}^{n+1},&\quad R_N|_{t<-1}&=0, \end{aligned}
(B.8)

where $$k=0, 1, \dots , N-1$$. By standard ODEs techniques, one can prove that if

\begin{aligned} \psi (x)=\int _{-\infty }^0 \omega \cdot \mathbf {W}(x+s\omega )\, \mathrm{d}s, \end{aligned}

then we have in $$\mathbb {R}^n$$ for $$k=0, 1, \dots , N-1$$:

\begin{aligned} \begin{aligned} a_{-1}(x)&= e^{\psi (x)}, \\ a_0(x)&=-\frac{1}{2}e^{\psi (x)}\int _{-\infty }^0 e^{-\psi (x+s\omega )} (L_{\mathbf {W},V}a_{-1})(x+s\omega )\,\mathrm{d}s\\&= -\frac{1}{2} e^{ \psi (x)} \int _{-\infty }^{0} \left[ -\Delta \psi - |\nabla \psi |^2 + 2 \mathbf {W}\cdot \nabla \psi + V \right] (x +s\omega ) \, \mathrm{d}s,\\ a_{k+1}(x)&= -\frac{1}{2(k+1)}e^{\psi (x)}\int _{-\infty }^0 e^{-\psi (x+s\omega )} (L_{\mathbf {W},V}\,a_k)(x+s\omega )\,\mathrm{d}s. \end{aligned} \end{aligned}
(B.9)

It remains to prove the existence of solutions to (B.8). To do that, assume that

\begin{aligned} \mathbf {W}\in C^{m+2}_c(\mathbb {R}^n) \quad \text {and} \quad V\in C^{m}_c(\mathbb {R}^n) \end{aligned}

for some $$m\in \mathbb {N}$$ large enough which will be fixed later. Note that

\begin{aligned} \Vert a_{-1} \Vert _{L^\infty }\lesssim \Vert \mathbf {W} \Vert _{C^{m+2}}, \end{aligned}

and for $$j=0, 1, \ldots , N$$

\begin{aligned} \Vert a_j \Vert _{L^\infty }\lesssim \Vert \mathbf {W} \Vert _{C^{m-2j}} + \Vert V \Vert _{C^{m-2j}}. \end{aligned}

Setting

\begin{aligned} \beta = \min \left\{ m-2N-2, N-1 \right\} , \end{aligned}

we deduce that $$\ss (t,x):=\left( L_{\mathbf {W},V}\, a_N\right) (t-x\cdot \omega )^N_+$$ belongs to $$C^{\beta }( \mathbb {R}^{n+1})$$. We actually claim that for every fixed $$t\ge -1$$, the function $$\ss (t, \cdot )$$ is supported in the following compact set of $$\mathbb {R}^n$$

\begin{aligned} \left\{ x\in \mathbb {R}^n \, : \, -1 \le x\cdot \omega \le t, \, |x-(x\cdot \omega ) \omega |\le 1 \right\} . \end{aligned}

Indeed, by (B.3) we immediately deduce that $$(t-x\cdot \omega )^N_+=0$$ when $$t<x\cdot \omega$$. Moreover, by (B.7), we already known that $$a_N(x)=0$$ when $$x\cdot \omega <-1$$. The explicit expression of $$a_N$$ given in (B.9), combined with the compact support of $$\mathbf {W}$$ and V in the unit ball, implies that $$a_N(x)=0$$ when $$|x-(x\cdot \omega )x|>1$$. This proves the claim. In particular, the function $$\ss$$ belongs to $$H^{\beta _1}_{loc}(\mathbb {R}; H^{\beta _2}_{loc}(\mathbb {R}^n))$$ with $$\beta _j\ge 0$$ (both will be fixed later) and $$\beta _1+\beta _2=\beta$$. Since $$\partial _t^2+ L_{\mathbf {W},V}$$ is a strictly hyperbolic operator, [8, Theorems 9.3.1 and 9.3.2] ensure that there exists a unique solution $$R_N\in H^{\beta _1+1}_{loc}(\mathbb {R}; H^{\beta _2}_{loc}(\mathbb {R}^n))$$ to (B.8) such that for any given $$T>-1$$ we have

\begin{aligned} \left\| R_N\right\| _{H^{\beta _1+1}((-1, T]; H^{\beta _2}(\mathbb {R}^n))}\lesssim \left\| \left( L_{\mathbf {W},V}\, a_N\right) (t-x\cdot \omega )^N_+\right\| _{H^{\beta _1}((-1, T]; H^{\beta _2}(\mathbb {R}^n))}. \end{aligned}

We claim that $$R_N\in C^2(\mathbb {R}^{n+1})$$ by suitably choosing the parameters $$m$$, N, $$\beta _1$$, and $$\beta _2$$. This regularity is needed to apply, for instance, the Carleman estimate with boundary terms in Proposition 3.1 and the estimate in Lemma 3.3. By [9, Theorem B.2.8/Vol III], this follows if, for instance,

\begin{aligned} \frac{n+3}{2}< \beta _1 + \beta _2=\beta \quad \text {and} \quad \frac{3}{2}<\beta _1, \end{aligned}

and furthermore

\begin{aligned} \begin{aligned} \Vert R_N \Vert _{C^2((-1,T]\times \mathbb {R}^n)}&\lesssim \left\| R_N\right\| _{H^{\beta _1+1}((-1, T]; H^{\beta _2}(\mathbb {R}^n))}\\&\lesssim \left\| \left( L_{\mathbf {W},V}\, a_N\right) (t-x\cdot \omega )^N_+\right\| _{H^{\beta _1}((-1, T]; H^{\beta _2}(\mathbb {R}^n))}. \end{aligned} \end{aligned}
(B.10)

Equating the two parameters involved in the definition of $$\beta$$, that is, $$m-2N-2=N-1$$, we choose $$m=3N+1$$, and hence, $$\beta =N-1$$. We distinguish two cases:

• When n is even we consider

\begin{aligned} N =\frac{n+6}{2}, \quad m=\frac{3}{2}n +10, \quad \beta _1=2, \quad \beta _2=\frac{n}{2}. \end{aligned}
• When n is odd we consider

\begin{aligned} N =\frac{n+7}{2}, \quad m=\frac{3}{2}(n+1) +10, \quad \beta _1=2, \quad \beta _2=\frac{n+1}{2}. \end{aligned}

The desired claim is proved by combining above choices with (B.10). On the other hand, by (B.4) and (B.9), we deduce that $$U_\delta$$ can be written as follows

\begin{aligned} U_\delta (x,t)= e^{\psi (x)} \delta (t-x\cdot \omega ) + v(x,t)H(t-x\cdot \omega ), \end{aligned}

where

\begin{aligned} v(x,t)= \sum _{j=0}^N a_j (x)(t- x\cdot \omega )_+^j + R_N(x,t) \end{aligned}

is of class $$C^2$$ in the region $$\left\{ t\ge x\cdot \omega \right\}$$. This shows that v satisfies all the properties stated in Proposition 2.4, and hence, the proof of existence and uniqueness of solutions to (B.1) is completed.

Second case. Existence of solutions of (B.2). This case is quite similar to the previous one, and hence, we only give a brief explanation of the proof. Here, we have $$a_{-1}=0$$ and according to identity (B.5), the function $$a_0$$ must satisfy (B.6) in place of $$a_{-1}$$. The remaining coefficients $$a_k$$ satisfy (B.7). The remainder term $$R_N$$ satisfies the same equation as in (B.8), and thus, its existence is guaranteed by [8, Theorems 9.3.1 and 9.3.2]. In this case, the solution $$U_H$$ to (B.2) has the form

\begin{aligned} U_H(x, t)=u(x,t) H(t- x\cdot \omega ), \quad u(x,t)= \sum _{j=0}^N a_j (x)(t- x\cdot \omega )_+^j + R_N(x,t), \end{aligned}

where u is of class $$C^2$$ in the region $$\left\{ t\ge x\cdot \omega \right\}$$. Moreover, it satisfies all the desired properties stated in Proposition 2.3.

### B.2. Proof of the Energy Lemmas

The proofs of the following lemmas are mainly based on standard multiplier techniques. Here, $$\text {div}_{\mathfrak {m}}$$ and $$\nabla _{\mathfrak {m}}$$ stand, respectively, for the divergence and gradient operators with respect to a set of variables $$\mathfrak {m}$$, while $$\nabla$$ is reserved for the gradient operator with respect to the full spatial variables, that is $$\nabla = \nabla _x$$. Let $$r=|x|$$. We define the radial and angular derivatives as

\begin{aligned} \partial _r = \frac{x}{|x|}\cdot \nabla , \quad \quad \Omega _{j k}=x_j \partial _k - x_k \partial _j, \quad j, k=1, 2, \ldots , n. \end{aligned}

Note that

\begin{aligned} |\nabla \alpha (x) |^2 = |\partial _r \alpha |^2 + \frac{1}{2r^2} \sum _{\underset{i, j=1, 2, \ldots , n}{i \ne j}} |\Omega _{i j}\alpha (x) |^2. \end{aligned}

We also adopt the notation: $$\alpha _z:= \partial _z \alpha$$, $$\alpha _t:=\partial _t \alpha$$ and $$\alpha _r:=\partial _r \alpha$$.

### Proof of Lemma 2.9

This result is quite similar to Lemma [21, Lemma 3.3]. Define

\begin{aligned} H_T= \left\{ (y,z,t): |(y,z) |\ge 1, \; -T\le t\le T,\; t\ge z \right\} . \end{aligned}

The following identities will be useful in our computations

\begin{aligned} 2\alpha _t \,\square \, \alpha&= {\text {div}}_{x,t} \left( -2 \alpha _t \nabla \alpha , \, \alpha _t^2+ |\nabla \alpha |^2\right) ,\\ 2(x\cdot \nabla \alpha )\, \square \,\alpha =&{\text {div}}_{x,t} \left( x(|\nabla \alpha |^2- \alpha _t^2)-2(x\cdot \nabla \alpha )\nabla \alpha , 2\alpha _t (x\cdot \nabla \alpha ) \right) \\&\quad +\, n\alpha _t^2- (n-2)|\nabla \alpha |^2. \end{aligned}

Thanks to the domain of dependence theorem for the wave equation $$\square \, \alpha =0$$ in $$H_T$$, see, for instance, [7, Section 2.4, Theorem 6], we deduce that $$\alpha$$ is compactly supported for each fixed t. In particular, for each fixed t, one integral involving $$\alpha$$ on $$|x |\ge 1$$ is actually on $$M\ge |x |\ge 1$$ for a large enough $$M>1$$. This fact will be used several times throughout this proof.

On the one hand, integrating the first identity over the region $$H_T\cap \left\{ t\le \tau \right\}$$ for any fixed $$\tau \in [-T, T]$$, and combining Stoke’s theorem with the third equation in (2.20), we obtain

\begin{aligned} \begin{aligned} 0&= \int _{\partial (H_T\cap \left\{ t\le \tau \right\} )} \nu (y,z,t)\cdot \left( -2 \alpha _t \nabla \alpha , \alpha _t^2+ |\nabla \alpha |^2\right) \mathrm{d}S\\&= \int _{H_T\cap \left\{ t=\tau \right\} } \left( \alpha _t^2+|\nabla \alpha |^2\right) \mathrm{d}x +2 \int _{\Sigma _+\cap \left\{ t\le \tau \right\} } \alpha _r \alpha _t \,\mathrm{d}S\\&\quad -\, \frac{1}{\sqrt{2}} \int _{H_T\cap \left\{ t=z \right\} } \left( \alpha _t^2+ |\nabla \alpha |^2+ 2\alpha _t\alpha _z\right) \mathrm{d}y \mathrm{d}z . \end{aligned} \end{aligned}
(B.11)

We have used that the unit normal vectors $$\nu (y,z,t)$$ are, respectively, equal to (0, 0, 1), $$\frac{1}{\sqrt{2}}(0,1,-1)$$ and $$-\frac{1}{|x |}(y,z,0)$$ on the regions $$H_T\cap \left\{ t=\tau \right\}$$, $$H_T\cap \left\{ t=z \right\}$$ and $$\Sigma _+\cap \left\{ t\le \tau \right\}$$. A domain of dependence argument shows that $$H_T\cap {{\,\mathrm{\mathrm {supp}}\,}}\alpha$$ is bounded and far away from the origin, and thus, $$1/|x |$$ is well defined on that region. In particular, $$\alpha _r$$ is well defined on $$H_T\cap {{\,\mathrm{\mathrm {supp}}\,}}\alpha$$.

On the other hand, note that $$\alpha _t^2+ |\nabla \alpha |^2+ 2\alpha _t\alpha _z= |\nabla _y \alpha |^2 + (Z\alpha )^2$$ and $$Z\alpha =0$$ on $$H_T\cap \left\{ t=z\right\}$$. Combining these facts with Young’s inequality applied with $$\epsilon >0$$, we get from identity (B.11) that

\begin{aligned} \begin{aligned} \int _{H_T\cap \left\{ t=\tau \right\} } \left( \alpha _t^2+|\nabla \alpha |^2\right) dx&\le \epsilon \int _{\Sigma _+} \alpha _r^2 \,\mathrm{d}S + \epsilon ^{-1}\int _{\Sigma _+} \alpha _t^2 \,\mathrm{d}S \\&\quad +\, \frac{1}{\sqrt{2}}\int _{H_T\cap \left\{ t=z \right\} } |\nabla _y\,\alpha (y,z,z ) |^2 \mathrm{d}y\mathrm{d}z. \end{aligned} \end{aligned}
(B.12)

In a similar fashion, now integrating the second useful identity over $$H_T$$, we obtain

\begin{aligned} \begin{aligned}&\int _{\Sigma _+} \left( 2 \alpha _r^2 +\alpha _t^2-|\nabla \alpha |^2\right) \mathrm{d}S= \int _{H_T}\left( (n-2)|\nabla \alpha |^2 - n\alpha _t^2\right) \mathrm{d}x\mathrm{d}t \\&\qquad -\, 2 \int _{\left\{ |x |\ge 1\right\} \cap \left\{ t=T\right\} }\alpha _t (x\cdot \nabla \alpha ) \, dx-\frac{1}{\sqrt{2}} \int _{H_T\cap \left\{ t=z\right\} }z (|\nabla \alpha |^2 - \alpha _t^2) \\&\quad = \int _{H_T}\left( (n-2)|\nabla \alpha |^2 - n\alpha _t^2\right) \mathrm{d}x\mathrm{d}t \\&\qquad -\, 2 \int _{\left\{ |x |\ge 1\right\} \cap \left\{ t=T\right\} }\alpha _t (x\cdot \nabla \alpha )\, \mathrm{d}x-\frac{1}{\sqrt{2}} \int _{H_T\cap \left\{ t=z\right\} }z |\nabla _y \alpha (y,z, z) |^2 \mathrm{d}y \mathrm{d}z, \end{aligned} \end{aligned}
(B.13)

where we used that $$|(y,z) |=1$$ on $$\Sigma _+$$ and $$\alpha _t^2=\alpha _z^2$$ on $$H_T\cap \left\{ t=z\right\}$$.

Integrating inequality (B.12) in time from $$-T$$ to T, and using the resulting estimate into (B.13), we deduce

\begin{aligned} \int _{\Sigma _+} (\alpha _r^2 + \alpha _t^2)\, \mathrm{d}S&\lesssim | \int _{\Sigma _+} \left( |\nabla \alpha (x) |^2-\alpha _r^2\right) \mathrm{d}S | + \int _{-T}^T\int _{H_T\cap \left\{ t=\tau \right\} } \left( \alpha _t^2 + |\nabla \alpha |^2 \right) \mathrm{d}x\mathrm{d}\tau \\&\quad +\, \int _{H_T \cap \left\{ t=T \right\} } \left( \alpha _t^2+ |\nabla \alpha |^2\right) dx+ \int _{H_T\cap \left\{ t=z\right\} } |\nabla _y \alpha (y, z, z) |^2 \mathrm{d}y \mathrm{d}z\\&\le \sum _{j<k} \int _{\Sigma _+}|\Omega _{jk}\alpha (x) |^2\, \mathrm{d}S + (1+ (2T+1)\epsilon ^{-1})\int _{\Sigma _+} \alpha _t^2\, \mathrm{d}S\\&\quad +\, (2T+1)\epsilon \int _{\Sigma _+} \alpha _r^2\, \mathrm{d}S+ (\sqrt{2}(T+1)+1) \\&\times \, \int _{H_T\cap \left\{ t=z\right\} } |\nabla _y \alpha (y,z,z) |^2 \mathrm{d}y \mathrm{d}z. \end{aligned}

Here, we have also used (B.12) with $$\tau =T$$ to bound the integral on $$H_T\cap \left\{ t=T \right\}$$. Choosing $$\epsilon$$ small enough, we obtain

\begin{aligned} \int _{\Sigma _+} \alpha _r^2 \, \mathrm{d}S \lesssim \Vert \alpha \Vert ^2_{H^1(\Sigma _+)}+ \int _{H_T\cap \left\{ t=z\right\} } |\nabla _y \alpha (y,z, z) |^2 \mathrm{d}y \mathrm{d}z. \end{aligned}
(B.14)

The last term on the right can be bounded using (2.21). Indeed,

\begin{aligned} \begin{aligned} \int _{H_T\cap \left\{ t=z\right\} } |\nabla _y \alpha (y,z, z) |^2 \mathrm{d}y \mathrm{d}z&\lesssim \int _{|y |\le 1}\int _{\sqrt{1-|y |^2}\le z \le T}|\nabla _y \alpha (y, z, z)) |^2 \mathrm{d}z\mathrm{d}y\\&= \int _{|y |\le 1}\int _{\sqrt{1-|y |^2}\le z \le T}|\nabla _y \beta (y) |^2 \mathrm{d}z\mathrm{d}y \\&\le T \varepsilon ^{-1} \int _{|y |\le 1} \frac{1}{\sqrt{1-|y |^2}}| {\sqrt{1-|y |^2}} \nabla _y \beta (y) |^2 \mathrm{d}y, \end{aligned} \end{aligned}

where we used that $$\beta$$ vanishes on $$|y |\ge 1-\varepsilon$$. Since $$\beta$$ is independent of z, and the tangential part of the gradient to the unit sphere is $$z\nabla _y- y\partial _z$$, we have

\begin{aligned} \int _{H_T\cap \left\{ t=z\right\} } |\nabla _y \alpha (y,z, z) |^2 \mathrm{d}y \mathrm{d}z\lesssim & {} \varepsilon ^{-1} \int _{|y |\le 1} | (z \nabla _y - y\nabla _z) \alpha (y,z,z) |^2 \mathrm{d}S \\\le & {} \varepsilon ^{-1} \Vert \alpha \Vert ^2_{H^1(\Sigma _+ \cap \left\{ t=z \right\} )}. \end{aligned}

This estimate and (B.14) imply

\begin{aligned} \int _{\Sigma _+} \alpha _r^2 \, \mathrm{d}S \lesssim \varepsilon ^{-1} \left( \Vert \alpha \Vert ^2_{H^1(\Sigma _+)}+ \Vert \alpha \Vert ^2_{H^1(\Sigma _+\cap \Gamma )}\right) . \end{aligned}

Let $$\chi \in C^{1}(\overline{Q})$$. The chain rule shows that it also holds for $$\chi \alpha$$ in place of $$\alpha$$, where the implicit constant is proportional to $$\Vert \chi \Vert _{C^1(\overline{Q})}$$. This finishes the proof. $$\square$$

### Proof of Lemma 2.7

Following the proof of [21, Lemma 3.4], we shall prove the estimate

\begin{aligned} \int _B (|\nabla _{x,t}\alpha |^2 +|\alpha |^2)(x, \tau )\mathrm{d}x\lesssim & {} \Vert \alpha \Vert ^2_{H^1(\Gamma )} + \Vert (\square + 2\mathbf {W}\cdot \nabla + V)\alpha \Vert ^2_{L^2(Q_+)} \\&\quad +\, \Vert \alpha \Vert ^2_{H^1(\Sigma _+)} + \Vert \partial _\nu \alpha \Vert ^2_{L^2(\Sigma _+)} \end{aligned}

for every $$\tau \in (-1,T]$$. Although the desired estimate is just the case $$\tau =T$$, our proof involves Gronwall’s inequality, so we need to consider the left-hand side information when $$\tau$$ is also far way from T. Due to this, we split the energy level into two cases depending on whether or not the intersection of $$t=\tau$$ with $$t=z$$ is inside $$Q_+$$. When $$\tau \in (-1, 1]$$, define

\begin{aligned} E(\tau )&:= \int _{B\cap \left\{ z\le \tau \right\} } (\alpha ^2+ \alpha _t^2+ |\nabla \alpha |^2)(y,z, \tau )\mathrm{d}z \mathrm{d}y,\\ J(\tau )&:= \int _{B\cap \left\{ z\le \tau \right\} } (\alpha ^2+ (Z\alpha )^2 + |\nabla _y \alpha |^2) (y,z,z)\mathrm{d}z\mathrm{d}y, \end{aligned}

and when $$\tau \in (1, T]$$, define

\begin{aligned} E(\tau )&:= \int _{B} (\alpha ^2+ \alpha _t^2+ |\nabla \alpha |^2)(y,z, \tau )\mathrm{d}z \mathrm{d}y,\\ J&:= J(1). \end{aligned}

A straightforward computation shows

\begin{aligned} 2\alpha _t (\square \, \alpha + \alpha )= \text {div}_{x,t}(-2\alpha _t \nabla \alpha , \alpha _t^2+|\nabla \alpha |^2+ \alpha ^2). \end{aligned}

For any $$\tau \in (-1, T]$$, integrating this identity over the region $$Q_+\cap \left\{ t\le \tau \right\}$$, and using Stoke’s theorem, we obtain

\begin{aligned} 2\int _{Q_+\cap \left\{ t\le \tau \right\} } \alpha _t (\square \, \alpha + \alpha )&= \int _{\partial (Q_+\cap \left\{ t\le \tau \right\} )} \nu (y,z,t)\cdot \left( -2 \alpha _t \nabla \alpha ,\alpha _t^2+|\nabla \alpha |^2+ \alpha ^2 \right) \mathrm{d}S\\&= -\int _{Q_+\cap \left\{ t=\tau \right\} } \left( \alpha _t^2+|\nabla \alpha |^2+\alpha ^2\right) \mathrm{d}x -2 \int _{\Sigma _+\cap \left\{ t\le \tau \right\} } \alpha _r \alpha _t \,\mathrm{d}S\\&\quad +\, \frac{1}{\sqrt{2}} \int _{Q_+\cap \left\{ t=z \right\} } \left( \alpha _t^2+ |\nabla \alpha |^2+ 2\alpha _t\alpha _z + \alpha ^2\right) \mathrm{d}y \mathrm{d}z. \end{aligned}

Now, using Young’s inequality with $$\epsilon =1$$, we get for all $$\tau \in (-1, T]$$

\begin{aligned} E(\tau )&\lesssim J + \int _{\Sigma _+} (\alpha _t^2 + \alpha _r ^2)\mathrm{d}S + \int _{Q_+\cap \left\{ t\le \tau \right\} } 2|\alpha _t \left( (\square +2\mathbf {W}\cdot \nabla +V ) \alpha \right. \\&\quad \left. +\, (1-V-2\mathbf {W}\cdot \nabla ) \alpha \right) |\\&\lesssim J + \int _{\Sigma _+}( \alpha _t^2 + \alpha _r ^2)\mathrm{d}S + \int _{Q_+}|(\square + 2\mathbf {W}\cdot \nabla + V)\alpha |^2 + \int _{Q_+} \alpha ^2+ \alpha _t^2 + |\nabla \alpha |^2\\&\le J + \int _{\Sigma _+}( \alpha ^2 + |\nabla \alpha |^2+\alpha _t^2+\alpha _r^2)\mathrm{d}S + \int _{Q_+}|(\square + 2\mathbf {W}\cdot \nabla + V)\alpha |^2 + \int _0^\tau E(t)\mathrm{d}t. \end{aligned}

Note that the implicit constant depends on $$\Vert V \Vert _{L^\infty }$$ and $$\Vert \mathbf {W} \Vert _{L^\infty }$$. It follows from the first line in above estimate. Applying Gronwall’s inequality, we have for all $$\tau \in (-1, T]$$

\begin{aligned} E(\tau ) \lesssim J + \int _{Q_+}|(\square + 2\mathbf {W}\cdot \nabla + V)\alpha |^2 + \int _{\Sigma _+}( \alpha ^2 + |\nabla \alpha |^2+\alpha _t^2+\alpha _r^2)\mathrm{d}S. \end{aligned}

The proof is now complete by taking $$\tau =T$$. $$\square$$

### Proof of Lemma 2.8

As in the proof of [21, Lemma 3.5], we first set $$v=e^{\sigma \psi }\alpha$$. For any $$\tau \in [1,T]$$, define the energy corresponding to $$t=\tau$$ and $$t=z$$ as follows

\begin{aligned} E(\tau )&:= \int _{B} (\sigma ^2 v^2+ v_t^2+ |\nabla v |^2)(y,z, \tau )\mathrm{d}z \mathrm{d}y,\\ J&:= \int _B (\sigma ^2 v^2+ (Zv)^2 + |\nabla _y v |^2) (y,z,z)\mathrm{d}z\mathrm{d}y. \end{aligned}

For any $$\tau \in [1, T]$$, integrating the identity

\begin{aligned} 2 v_t (\square \, v + \sigma ^2 v)= \text {div}_{x,t}(-2v_t \nabla v, v_t^2+|\nabla v |^2+ \sigma ^2 v^2) \end{aligned}

over the region $$Q_+\cap \left\{ t\le \tau \right\}$$, and using Stoke’s theorem, we obtain

\begin{aligned} 2\int _{Q_+\cap \left\{ t\le \tau \right\} } v_t (\square \, v + \sigma ^2 v)&= \int _{\partial (Q_+\cap \left\{ t\le \tau \right\} )} \nu (y,z,t)\cdot \left( -2v_t \nabla v, v_t^2+|\nabla v |^2+ \sigma ^2 v^2 \right) \mathrm{d}S\\&= - \int _{Q_+\cap \left\{ t=\tau \right\} } \left( v_t^2+|\nabla v |^2+\sigma ^2 v^2\right) dx -2 \int _{\Sigma _+\cap \left\{ t\le \tau \right\} } v_r v_t \,\mathrm{d}S\\&\quad +\, \frac{1}{\sqrt{2}} \int _{Q_+\cap \left\{ t=z \right\} } \left( v_t^2+ |\nabla v |^2+ 2v_tv_z + \sigma ^2 v^2\right) \mathrm{d}y \mathrm{d}z. \end{aligned}

Using Young’s inequality with $$2\sigma v_t(\sigma v)\le \sigma ( \sigma ^2 v^2 + v_t^2 )$$, we get

\begin{aligned} J&\lesssim E(\tau )+ 2\int _{\Sigma _+ \cap \left\{ t\le \tau \right\} } |v_t v_r | \mathrm{d}S + \int _{Q_+\cap \left\{ t\le \tau \right\} } 2|v_t \left( (\square +2\mathbf {W}\cdot \nabla +V ) v \right. \\&\quad \left. +\, (\sigma ^2-2\mathbf {W}\cdot \nabla -V) v\right) |\\&\lesssim E(\tau ) + 2\int _{\Sigma _+} |v_t v_r | \mathrm{d}S + \int _{Q_+}2|v_t(\square + 2\mathbf {W}\cdot \nabla + V)v | + \sigma \int _{Q_+} v_t^2\\&\quad + |\nabla v |^2+ \sigma ^2 v^2, \end{aligned}

whenever $$\sigma ^2\ge 1+ \Vert V \Vert _{L^\infty }+2\Vert \mathbf {W} \Vert _{L^\infty }$$. Since J is independent of $$\tau$$, integrating this estimate in $$\tau$$ from 1 to T, we have

\begin{aligned} J \lesssim \int _{\Sigma _+} |v_t v_r | \mathrm{d}S + \int _{Q_+}|v_t(\square + 2\mathbf {W}\cdot \nabla + V)v | + \sigma \int _{Q_+} v_t^2 + |\nabla v |^2+ \sigma ^2 v^2. \end{aligned}

The right-hand side can be bounded by integral terms involving weighted versions of $$\alpha$$, namely terms given by the right-hand side of (2.19). This can be seen by using the following inequalities:

\begin{aligned}&|v |\le e^{\sigma \psi } |\alpha |, \quad |v_t | + |\nabla v |\le e^{\sigma \psi } (|\alpha _t | + |\nabla \alpha |+ \sigma |\alpha |),\\&|(\square + 2\mathbf {W}\cdot \nabla + V) v |\lesssim e^{\sigma \psi } (|(\square + 2\mathbf {W}\cdot \nabla + V) \alpha | + \sigma (|\alpha _t |+ |\nabla \alpha |)+ \sigma ^2 |\alpha |),\\&2\sigma |\alpha _t | |\alpha | \le \alpha _t^2 + \sigma ^2 \alpha ^2, \quad 2\sigma |\nabla \alpha | |\alpha | \le |\nabla \alpha |^2 + \sigma ^2 \alpha ^2. \end{aligned}

Finally, we complete the proof by using the estimate below

\begin{aligned} \int _{B} e^{2\sigma \psi } (\sigma ^2 \alpha ^2+ (Z\alpha )^2 + |\nabla _y \alpha |^2) (y,z,z) \, \mathrm{d}z\mathrm{d}y\lesssim J. \end{aligned}

$$\square$$

## Rights and permissions

Reprints and Permissions