Abstract
We consider the approximation of an abstract evolution problem with inhomogeneous side constraint using A-stable Runge–Kutta methods. We derive a priori estimates in norms other than the underlying Banach space. Most notably, we derive estimates in the graph norm of the generator. These results are used to study convolution quadrature based discretizations of a wave scattering and a heat conduction problem.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
Many time dependent partial differential equations can be conveniently described in the language of strongly continuous semigroups. In this language, these initial boundary value problems resemble systems of ordinary differential equations, which suggests that they are amendable to the standard discretization schemes of multistep or Runge–Kutta type. Unlike the ODE case, one needs to pay special attention to the boundary conditions imposed by the generator of the semigroup. This, in most cases, leads to a reduction of order phenomenon, meaning that the convergence rates are (mainly) determined by the stage order of the Runge–Kutta method instead of the classical order. The a priori convergence of Runge–Kutta methods for semigroups has been extensively studied in the literature. Starting with the early works [8, 9], it has been established that conditions of the form \(u(t) \in {\text {dom}}(A^{\mu })\), where \(A\) is the generator of the semigroup and \({\text {dom}}\) denotes the domain of an operator, determine the convergence rates. In [25], this has been generalized to the case of non-integer \(\mu \ge 1\) using the theory of interpolation spaces. Finally, in [1], the case of \(\mu \in [0,1]\) was adressed, which is the case needed for PDEs with inhomogeneous boundary conditions. We point out that in the case of analytic semigroups, Lubich and Ostermann [19] had already established convergence also for inhomogeneous boundary conditions. All of these works focus on establishing convergence rates with respect to the norm of the underlying Banach space. In many applications one needs to establish convergence with respect to other norms, for example, in order to be able to bound boundary traces of the solution. Most notably, one might be interested in convergence of \(A_{\star }u\), where \(A_{\star }\) is an extension of the generator that disregards boundary conditions. If u is assumed to be in \({\text {dom}}(A)\), we get \(A_{\star }u=Au\) and the convergence result can be easily established by using the fact that the time evolution commutes with the generator of the underlying semigroup (both in the continuous and discrete settings). If the boundary conditions are inhomogeneous, such a strategy cannot be pursued. It is the goal of this paper to establish convergence results for \(A_{\star }u\) also for the case \(u(t) \in {\text {dom}}(A^{\mu })\) for \(\mu \in [0,1]\), again using the theory of interpolation spaces.
Similarly it is sometimes useful to compute discrete integrals of the time evolution by reusing the same Runge–Kutta method. Also in this case, we establish rigorous convergence rates.
Our interest in such estimates originally arose from the study of time domain boundary integral equations and their discretization using convolution quadrature (CQ). It has already been noticed in the early works (see e.g. [19]) that such discretizations have a strong relation to the Runge–Kutta approximation of the underlying semigroup. This approach of studying TDBIEs in a strictly time-domain way has recently garnered a lot of interest, see [3, 13, 15] and the monograph [31], as it potentially allows sharper bounds than the more standard Laplace domain based approach. Similar techniques have even been extended to the case of certain nonlinear problems in [4]. This paper can be seen as our latest addition to this effort. While the convergence rates provided by the Laplace-domain approach in [2] and the results in this current paper are essentially the same, the present new approach provides better insight into the dependence on the end-time of the computation (quadratic vs. general unknown polynomial behavior). This suggest that the present approach might be better suited for analyzing long term computations. It also fits more naturally with the time-domain analysis of the continuous problem and space discretization, as for example presented in [13].
The paper is structured as follows. Section 2 introduces the abstract setting and fixes notation, most notably for working with Runge–Kutta methods. Section 3 then contains the main estimates. Starting by summarizing known results from [1] in Sect. 3.1, we then formulate the main new results of this article in Sect. 3.2. After proving some preparatory lemmas related to Runge–Kutta methods in Sects. 4 and 5, we provide the proofs of the main estimates in Sect. 6. In Sect. 7, we show how our setting simplifies if we restrict our view to a subclass of admissible operators. In Sect. 8, to showcase how the theory developed in this paper is useful for this class of problems, we consider a simple exterior scattering problem in Sect. 8.3 and a heat transmission problem in Sect. 8.5. We note that Sect. 8.3 showcases the need for the bound on the discrete integral of the result, whereas Sect. 8.5 was chosen because, in order to bound the main quantity of interest on the boundary, we need to apply a trace theorem. This necessitates the use of the graph norm estimate.
2 Problem setting
We start by fixing the general setting used for the rest of the paper, first with respect to the equation to be solved and then with respect to its discretization.
2.1 Operator equation, functional calculus, and Sobolev towers
Assumption 2.I
We are given:
-
(a)
A closed linear operator \(A_{\star }:{\text {dom}}\left(A_{\star }\right)\subset {\mathcal {X}} \rightarrow {\mathcal {X}}\) in a Banach space \({\mathcal {X}}\).
-
(b)
A bounded linear operator \(B: {\text {dom}} \left(A_{\star }\right)\rightarrow {\mathcal {M}}\) with another Banach space \({\mathcal {M}}\).
We assume that \(A:=A_{\star }|_{\ker {B}}\) generates a \(C_0\)-semigroup and that B admits a bounded right inverse \({\mathscr {E}}\) such that \(\mathrm {range}\,{\mathscr {E}}\subset \ker (I-A_{\star })\), where \(I:{\mathcal {X}}\rightarrow {\mathcal {X}}\) is the identity operator.
We are given \(u_0\in {\text {dom}} (A)\) and data functions \(F \in {{\mathcal {C}}}^1([0,T], {\mathcal {X}})\), \(\varXi \in {\mathcal {C}}^1([0,T],{\mathcal {M}})\), and we consider the problem: find \(u \in {{\mathcal {C}}}^1([0,T],{\mathcal {X}})\) such that
For conditions on the well-posedness of this problem, see [13]. We start by recalling the following consequence of the Hille-Yosida theorem.
Proposition 2.1
([26, Corollary 3.8]) If A is the generator of a \(C_0\)-semigroup on a Banach space \({\mathcal {X}}\), then there exist constants \(\omega \ge 0\) and \(M\ge 1 \) such that the spectrum \(\sigma (A)\) of A satisfies \(\sigma (A) \subseteq \{z \in {\mathbb {C}}: \mathrm {Re}\,z \le \omega \}\) and the resolvent satisfies the estimates
When working with Runge–Kutta methods, it is useful to use a calculus that allows one to apply rational functions to (unbounded) operators, as long as the poles of the function are compatible with the spectrum of the operator.
Definition 2.2
(Rational functions of operators) Let q be a rational function that is bounded at infinity. Let \(\varLambda \) be the set of poles of q, which we can write in the form (note that we allow for some of the factors in the numerator to be constant)
If \(A:{\text {dom}}(A)\subset {\mathcal {X}}\rightarrow {\mathcal {X}}\) is a linear operator such that \(\sigma (A)\cap \varLambda =\emptyset \), we define
It is easy to see that different reorderings of the factors in the numerator and denominator of q produce the same result and that each factor in the definition of q(A) is a bounded linear operator in \({\mathcal {X}}\) since \(\lambda _i\not \in \sigma (A).\) The bounded linear operator \(q(A):{\mathcal {X}} \rightarrow {\mathcal {X}}\) satisfies
The error estimates of this paper use the theory of interpolation spaces. For Banach spaces \({\mathcal {X}}_1 \subset {\mathcal {X}}_0\) with continuous embedding and \(\mu \in (0,1)\), we define the space \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{\mu ,\infty }\) using real interpolation with the following norm:
We will not go into details of the definitions and instead refer to [35, 36] or [22, Appendix B]. For simplicity of notation we often drop the second parameter \(\infty \) and just write \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{\mu }\).
The most important property is the following: a bounded linear operator \(T: {\mathcal {X}}_0 \rightarrow {\mathcal {Y}}_0\) and \({\mathcal {X}}_1 \rightarrow {\mathcal {Y}}_1\) with \({\mathcal {X}}_1 \subseteq {\mathcal {X}}_0\) and \({\mathcal {Y}}_1 \subseteq {\mathcal {Y}}_0\) is also a bounded operator mapping \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{\mu } \rightarrow [{\mathcal {Y}}_0,{\mathcal {Y}}_1]_{\mu }\) with the following norm bound
We also note that for \(\mu _1 \le \mu _2\), the spaces are nested, i.e., \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{\mu _2} \subseteq [{\mathcal {X}}_0,{\mathcal {X}}_1]_{\mu _1}\) with continuous embedding. For notational convenience we write \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{0}:={\mathcal {X}}_0\) and \([{\mathcal {X}}_0,{\mathcal {X}}_1]_{1}:={\mathcal {X}}_1\). We will be interested in a collection of spaces defined by interpolating the domains of the powers of the operator A. The details of this construction can be found, for example in [11].
Definition 2.3
(Sobolev towers) Let \(A\) be a closed operator on a Banach space \({\mathcal {X}}\). For \(\mu \in {\mathbb {N}}_0\), we define the following spaces \({\mathcal {X}}_0:={\text {dom}} \left(A^0\right):={\mathcal {X}}\) and \({\mathcal {X}}_{\mu }:={\text {dom}}\left(A^{\mu }\right)\), equipped with the following norm
For \(\mu \in [0,\infty )\), we define \({\mathcal {X}}_{\mu }:=\left[ {\mathcal {X}}_{\lfloor \mu \rfloor }, {\mathcal {X}}_{\lfloor \mu \rfloor +1}\right] _{\mu -\lfloor \mu \rfloor }\) by interpolation.
We sometimes consider \({\text {dom}}(A)\) as a Banach space. It is to be understood carrying the graph norm, same as \({\mathcal {X}}_1\).
2.2 Runge–Kutta approximation and discrete stage derivative
An m-stage Runge–Kutta method is given by its Butcher tableau, characterized by \({\mathcal {Q}}\in {\mathbb {R}}^{m\times m}\) and \({\mathbf {b}}\), \({\mathbf {c}}\in {\mathbb {R}}^{m}\). The Runge–Kutta approximation of the problem (2.1a–2.1c) starts at \(u^k_0:=u_0\) and then computes for \(n\ge 0\) the stage vector \(U^k_n \in {\mathcal {X}}^m\) and the step approximation \(u^k_{n+1}\in {\mathcal {X}}\) by solving
We have used the following notation (the spaces \({\mathcal {Y}}\) and \({\mathcal {Z}}\) are generic):
-
(a)
For a function \(G:[0,T]\rightarrow {\mathcal {Y}}\) we write
$$\begin{aligned} G(t_n+k{\mathbf {c}}):= \left( \begin{array}{ccc} G(t_n+k c_1),&\ldots&, G(t_n+ k c_m)\end{array}\right) ^\top \in {\mathcal {Y}}^m. \end{aligned}$$ -
(b)
For a matrix \({\mathcal {S}}\in {\mathbb {R}}^{m\times m}\) and an operator \(C:{\mathcal {Y}}\rightarrow {\mathcal {Z}}\) we write
$$\begin{aligned} {\mathcal {S}}\otimes C:= \left[ \begin{array}{ccc} {\mathcal {S}}_{11} C &{} \cdots &{} {\mathcal {S}}_{1m} C \\ \vdots &{} &{} \vdots \\ {\mathcal {S}}_{m1} C &{} \cdots &{} {\mathcal {S}}_{mm} C \end{array}\right] : {\mathcal {Y}}^m \rightarrow {\mathcal {Z}}^m. \end{aligned}$$ -
(c)
For the vector \({\mathbf {b}}\) and an operator \(C:{\mathcal {Y}} \rightarrow {\mathcal {Z}}\) we write
$$\begin{aligned} {\mathbf {b}}^\top \otimes C := \left[ \begin{array}{ccc} b_1 C&\cdots&b_m C \end{array}\right] : {\mathcal {Y}}^m \rightarrow {\mathcal {Z}}. \end{aligned}$$ -
(d)
\({\mathcal {I}}\) is the \(m\times m\) identity matrix, and \({\mathbf {1}}=(1,\cdots ,1)^\top \).
-
(e)
We admit shortened expressions such as
$$\begin{aligned} {\mathcal {Q}}F(t_n+k{\mathbf {c}})&:=({\mathcal {Q}}\otimes I) F(t_n+k{\mathbf {c}}),\\ {\mathbf {1}}\, u&:= ({\mathbf {1}}\otimes I) u,\\ {\mathbf {b}}^\top F(t_n+k{\mathbf {c}})&:=({\mathbf {b}}^\top \otimes I) F(t_n+k{\mathbf {c}}). \end{aligned}$$
The following lemma involving inversion of matrices of operators associated to an operator can be proved by taking the Jordan canonical form of the matrix \({\mathcal {S}}\).
Lemma 2.4
If \(A:{\text {dom}}(A)\subset {\mathcal {X}}\rightarrow {\mathcal {X}}\) is a linear operator on a Banach space \({\mathcal {X}}\) and \({\mathcal {S}}\in {\mathbb {C}}^{m\times m}\) satisfies \(\sigma (A) \cap \sigma ({\mathcal {S}}) = \emptyset \), then
is invertible. Furthermore, there exists a constant \(C_{\mathcal {S}}\), depending only on \({\mathcal {S}}\), such that
Under Assumption 2.I, the internal stage computation in the RK method can be decomposed in the following form:
In (2.8b) we look for \(Z^k_n\in (\mathrm {dom}(A))^m\).
The stability function of the Runge–Kutta method is the rational function \(r(z):=1+z{\mathbf {b}}^\top (I-z{\mathcal {Q}})^{-1}{\mathbf {1}}\). We will not consider the full class of Runge–Kutta methods, but will restrict our considerations to those satisfying the following Assumptions:
Assumption 2.II
-
(i)
The matrix \({\mathcal {Q}}\) is invertible.
-
(ii)
The stability function r does not have poles in \(\{ z\,:\, \mathrm {Re}\,z<0\}\), and \(\left| r(it)\right| \le 1\) for all \(t \in {\mathbb {R}}\) (i.e., the method is A-stable). Equivalently, \(|r(z)|<1\) for all z with negative real part.
We note that Assumption 2.II (i) implies that the following limit exists
Assumption 2.II (ii) implies that
and that r is a rational function with poles only in \({\mathbb {C}}_+\) and bounded at infinity.
The computation of the internal stages in the numerical approximation (2.7a–2.7c) requires the inversion of
as can be seen from the equivalent form (2.8a–2.8c).
If A is the infinitesimal generator of a \(C_0\)-semigroup and \(\omega \) and M are given by Proposition 2.1 and if we choose (recall that \(\sigma ({\mathcal {Q}})\subset {\mathbb {C}}_+\))
then the RK method can be applied for any \(0 < k\le k_0\). By Proposition 2.1 and Lemma 2.4, it follows that
Using Definition 2.2, we can define \(r(k\, A)\) for an RK method satisfying Assumption 2.II and \(k\le k_0\) satisfying (2.9). We then define
This quantity is relevant for the study of the error propagation in the Runge–Kutta method.
Given an RK method, we consider the following matrix-valued rational function
(The verification that these two formulas correspond to the same matrix is simple by using the Sherman–Morrison–Woodbury formula.) This matrix is related to the discrete differentiation process associated to an RK method satisfying Assumption 2.II: on the one hand \(k^{-1}\delta (z)\) is the discrete symbol associated to the discrete operational calculus built with the RK method [19]; on the other hand, a direct interpretation of this symbol is possible using the Z-transformation (see [14, Sect. 6]). Given a sequence \(U:=\{ U_n\}\) (tagged from \(n\ge 0\)) on a space, its Z-transform \(\mathcal{Z} \{ U_n\}\) is the formal series
For a detailed treatment on formal power series, see [12].
Definition 2.5
Let \(U:=\{ U_n\}\) and \(V:=\{ V_n\}\) be two sequences in \({\mathcal {X}}^m\) and let \({\widehat{U}}\) and \({\widehat{V}}\) be their respective Z-transforms. If
we write
The above definition is consistent with the RK discrete operational calculus of Lubich and Ostermann, see Sect. 8.1 and [19]. We now show an explicit form of the computation of \(\partial ^k\) and its inverse.
Lemma 2.6
If \(U=\{U_n\}\) is a sequence in \({\mathcal {X}}^m\), then \(X:=(\partial ^k)^{-1}U\) can be computed with the recurrence
and \(V:=\partial ^k U\) can be computed with the inverse recurrence
Proof
The proof of (2.13) is a simple exercise in Z-transforms, while (2.14) follows from (2.13) by writing \(U_n\) in terms of \(X_n\) (and changing names to the sequences). \(\square \)
The first result of Lemma 2.6 expresses the fact that if we apply the RK method to the equation
and \(X:=\{ X_n\}\) is the sequence of vectors of internal stages, then \(X=(\partial ^k)^{-1} U\), where \(U_n:=u(t_n+k{\mathbf {c}})\).
Finally we note that we call a Runge–Kutta method stiffly accurate, if it satisfies \({\mathbf {b}}^\top {\mathcal {Q}}^{-1}=\mathbf{e}_m^\top :=(0,\dots ,0,1)\). Stiffly accurate methods satisfy [we use that \({\mathcal {Q}}{\mathbf {1}}={\mathbf {c}}\), see (5.1)]
and \(r(\infty )=0.\)
For stiffly accurate methods, taking the discrete derivative of a stage vector consisting of samples taken from a continuous function is particularly simple:
Lemma 2.7
Let \(t \mapsto F(t)\) be a continuous function with \(F(0) = 0\). For stiffly accurate RK methods the sequence \(G:=\partial ^k F\) with \(F_n=F(t_n+k{\mathbf {c}})\) satisfies
Proof
For stiffly accurate methods we have \(r(\infty )=0\) and therefore
However, since \(c_m=1\), we have \({\mathbf {e}}_m^\top F(t_{n-1}+ k{\mathbf {c}})=F(t_{n-1}+k c_m)=F(t_n),\) which proves the result. \(\square \)
We also make the following optional assumption, which allows us to increase the convergence order in some cases.
Assumption 2.III
For all \(t\in {\mathbb {R}}, t \ne 0\) the stability function satisfies \(\left| r(it)\right| < 1\) and \(r(\infty )<1\).
3 Error estimates
We are now in a position to formulate the main results of this article and put them into context with previous results, most notably from [1].
To simplify notation, we will write for \(v\in \mathcal C([0,T];{\mathcal {X}}_\mu )\) with \(\mu \ge 0,\)
For functions \(f:[0,T]\rightarrow {\mathcal {Y}}\), we will write \((\partial ^{-1} f)(t):=\int _0^t f(\tau )\mathrm d\tau \), where \({\mathcal {Y}}\) denotes a generic Banach space.
3.1 The estimates of Alonso-Mallo and Palencia
The following two propositions summarize the results of Alonso-Mallo and Palencia [1], rewritten with the notation of the present paper. The ‘proofs’ which we provide clarify how notation needs to be adapted and how the hypotheses of the main results of [1] are satisfied in our context.
Proposition 3.1
([1, Theorem 1]) Let Assumption 2.I hold and assume that the exact solution u satisfies \(u \in {\mathcal {C}}^{p+1}\left( \left[ 0,T\right] ,{\mathcal {X}}_{\mu }\right) \) for some \(\mu \ge 0\). Let \(\{ u^k_n\}\) denote the Runge–Kutta approximation from (2.7a–2.7c). Then there exist constants \(k_0 > 0\) and \(C>0\) such that for \(0<k\le k_0\) and \(0< n k\le T\) the following estimate holds:
The constant C depends on the Runge–Kutta method, \(\mu \), and the constants M and \(\omega \) from (2.2). The constant \(k_0\) depends only on \(\omega \) and the Runge–Kutta method.
Proof
We only remark on the differences in notation. A different definition of interpolation spaces is given in [1], but the proof only relies on estimates of the form (2.6). The choice of \(k_0\) follows from the fact that it is only needed to ensure that \((I - k\,{\mathcal {Q}}\otimes A)\) is invertible, see (2.10). The assumption \(\mu \le p-q\) in [1, Theorem 1] can be replaced by using the rate \(\min \{p,q+\mu \}\) in (3.1) as the spaces \({\mathcal {X}}_{\mu } \subseteq {\mathcal {X}}_{p-q}\) are nested for \(\mu \ge p-q\). We also lowered the regularity requirements on the highest derivative compared to their stated result. The fact that this holds true follows from inspection of the proof. See also Lemma 5.9 for the key ingredient. \(\square \)
For certain Runge–Kutta methods, these estimates can be improved:
Proposition 3.2
([1, Theorem 2]) Let the assumptions of Proposition 3.1 hold and assume that, in addition, the RK method satisfies Assumption 2.III. Then there exist constants \(k_0 > 0\), \(C>0\) such that for \(0<k\le k_0\) and \(0< n k\le T\) the following improved estimate holds:
The constant C depends on the Runge–Kutta method, \(\mu \), and the constants M and \(\omega \) from (2.2); \(k_0\) depends only on the constant \(\omega \) and the Runge–Kutta method.
Proof
Again, this is just a reformulation of [1, Theorem 2]. We first note that, due to our assumption on \(r(\infty )\), we are always in the case \(m=0\) of [1]. Since we assumed that on the imaginary axis \(\left| r(it)\right| <1\) for \(0\ne t \in {\mathbb {R}}\), we directly note that for sufficiently small \(k_0\), all the zeros of \(r(z)-1\) except \(z=0\) satisfy \(\mathrm {Re}\, z > k_0 \omega \). By the resolvent bound (2.2) we can therefore estimate for \(k\le k_0\)
i.e., we have a uniform resolvent bound in the set \(Z_{\alpha ,\delta }\) in [1]. We also note that we reformulated the convergence rate such that we do not have the restriction \(\mu \le p-q-1\), since the exceptional cases are already covered by Proposition 3.1. \(\square \)
Remark 3.3
The assumption \(\left| r(z)\right| <1\) for \({\text {Re}}(z) \le 0\) and \(r(\infty )\ne 1\) is satisfied by the Radau IIA family of Runge–Kutta methods, but is violated by the Gauss methods, which satisfy \(\left| r(z)\right| =1\) on the imaginary axis.
3.2 New results in this article
In this section we present some a priori estimates for the convergence of Runge–Kutta methods when applied to the abstract problem (2.1a–2.1c). These can be seen as a continuation of [1] to the case where the boundary conditions are not given exactly but stem from computing discrete integrals and differentials using the same Runge–Kutta method.
Theorem 3.4
(Integrated estimate) Let u solve (2.1a–2.1c) with \(u_0=0\) and assume that for some \(\mu \ge 0\) we have
Set \(x:=\partial ^{-1} u\). Let \(U^k=\{ U^k_n\}\) and let \(u^k=\{ u^k_n\}\) be the discrete approximation given by (2.7a–2.7c) for a method satisfying Assumption 2.II. If \(X^k:=(\partial ^k)^{-1} U^k\) and we define \(x^k=\{x^k_n\}\) with the recurrence
then there exists a constant \(k_0>0\) such that for all \(k<k_0\) and \(n \in {\mathbb {N}}\) with \(nk\le T\) the following estimate holds:
If Assumption 2.III holds and if we assume the stronger regularities
then
The constant \(k_0\) depends only on \(\omega \) from (2.2) and the Runge–Kutta method. If \(\omega =0\) then \(k_0\) can be chosen arbitrarily large. C depends on \(\omega \), M from (2.2), the Runge–Kutta method, and \(\mu \).
Theorem 3.5
(Differentiated estimate) Let u solve (2.1a–2.1c) with \(u_0=0\) and assume \({\dot{u}}(0)=0\). Assume that for some \(\mu \ge 0\) we have
and let \(v:=\dot{u}\). Let \(U^k=\{ U^k_n\}\) and \(u^k=\{ u^k_n\}\) be the discrete approximation given by (2.7a–2.7c) for a stiffly accurate method satisfying Assumption 2.II.
If \(V^k:=\partial ^kU^k\) and \(v^k_n={\mathbf {e}}_m^\top V^k_{n-1}\), then there exists a constant \(k_0>0\) such that for all \(k<k_0\) and \(n\ge 1\) such that \(nk\le T\) the following estimate holds:
If, in addition, the method satisfies Assumption 2.III and
then
The constant \(k_0\) depends only on \(\omega \) from (2.2) and the Runge–Kutta method. If \(\omega =0\), then \(k_0\) can be chosen arbitrarily large. C depends on \(\omega \), M from (2.2), the Runge–Kutta method, and \(\mu \).
Remark 3.6
Most of the effort in proving the above theorem is done in order to obtain a convergence rate higher than q, even though the constraint in the stages is approximated only with order q. This is possible by exploiting the additional structure of the discretization error of the side constraint.
Remark 3.7
We formulated all our results for homogeneous initial conditions, since it is sufficient for our purposes in time domain BEM and convolution quadrature. It should be possible to generalize these results to the case of \(u_0 \in {\text {dom}}(A^{s})\) for sufficiently large \(s \ge 1\) by considering the evolution of the semigroup with inhomogeneous side constraint but homogeneous initial condition and the semigroup of homogeneous constraint but inhomogeneous \(u_0\) separately.
Remark 3.8
The loss of order by 1 in Theorem 3.5 compared to Propositions 3.1 and 3.2 is to be expected. Indeed, if we look at the case \(u \in {\text {dom}}(A^{\mu })\) for \(\mu \ge 1\), this means \(A_{\star }u \in {\text {dom}}(A^{\mu -1})\). Applying Proposition 3.2 to this semigroup then also gives a reduced order of \(k^{\min (q+\mu ,p)}\).
4 Some computations related to the main theorems
We will collect the sampled data and the stage and step parts of the solutions in four formal series
If the data functions are polynomially bounded in time, the series in (4.1a) are convergent (in \({\mathcal {X}}^m\) and \(\mathcal M^m\) respectively) with at least unit radius of convergence. Because of the equivalent formulation of the numerical method in the form (2.8a–2.8c), and using (2.10), it follows that for \(k\le k_0\) [with \(k_0\) chosen using (2.9)], the numerical solution is at least bounded in the form \(\left\| U^k_n\right\| _{{\mathcal {X}}} \lesssim C^{n}\). Thus, the two series in (4.1b) also converge on a sufficiently small disk.
Proposition 4.1
The sequences \(\{ U^k_n\}\) and \(\{u^k_n\}\) satisfy equations (2.7a–2.7c) if and only if
Proof
Let us start by proving a simple result: the discrete equations (2.7a) and (2.7c) hold if and only if (2.7a) and
hold. To see this, note that (2.7a) is equivalent to
and therefore (2.7a) and (2.7c) imply
or equivalently (4.3). The reciprocal statement is proved similarly. The recurrence (4.3) is equivalent to (4.2c). At the same time, the recurrence (2.7a) is equivalent to
After inserting (4.2c) into (4.4), the formula (4.2a) follows. \(\square \)
Proposition 4.1 is a rephrasing of [28, Lemma 3.19], where the computation is also laid out in more detail. Note how equations (4.2a)–(4.2b) relate strongly to (2.1a–2.1c), with the discrete symbol \(k^{-1}\delta (z)\) playing the role of the time derivative and
playing the role of a discrete Dirac delta at time \(t=0\).
Lemma 4.2
([2, Lemma 2.6]) If the matrix \({\mathcal {Q}}\) of the RK method is invertible, then for \(\left| z\right| < 1\)
In particular, if the Runge–Kutta method is A-stable (Assumption 2.II), then \(\sigma (\delta (z))\subset {\mathbb {C}}_+\).
We need a corollary to the previous result:
Corollary 4.3
Let Assumption 2.II hold. Then, for all \(r_0<1\), there exists a constant \(d>0\) such that for all \(\left| z\right| <r_0\) there holds
Proof
In view of of Lemma 4.2, since \(\sigma ({\mathcal {Q}})\) is finite, independent of z, and contained in \({\mathbb {C}}_+\), we are mainly concerned with the set \(\{ w \in {\mathbb {C}}: r(w) z = 1 \}\). We first note that
Second, we observe that by taking \(d_0\) small enough, we can ensure that \(w \mapsto r(w)\) is continuous for \({\text {Re}}(w)\le d_0\) and thus
is a closed set. Third, by considering the limit along the imaginary axis, we get
Thus, for \(\left| w\right| \) sufficiently large, it holds that \(\left| r(w)\right| \le 1/r_0\).
Overall, we get that
is a compact set with empty intersection with the imaginary axis. Thus, it must have a positive distance from it. These observations and Lemma 4.2 conclude the proof. \(\square \)
Lemma 4.4
Let Assumptions 2.I and 2.II hold. For \(r_0<1\), there exists \(k_0=k_0(\omega ,r_0)>0\) such that for all \(k\le k_0\) and \(|z|\le r_0\) the problem
has a unique solution for arbitrary \({\widehat{F}}\in {\mathcal {X}}^m\) and \({\widehat{\varXi }}\in {\mathcal {M}}^m\). If \(\omega =0\) in Proposition 2.1, then there are no restrictions on k, and the results holds for all \(|z|<1\).
Proof
Assume first that \({\mathcal {S}}\in {\mathbb {C}}^{m\times m}\) is such that \(\sigma ({\mathcal {S}}) \subset \{ z\,:\, \mathrm {Re}\,z>\omega \}\) and consider the problem
Take first \({\widehat{V}}:=({\mathcal {I}}\otimes {\mathscr {E}}){\widehat{\varXi }}\) (where \({\mathscr {E}}\) is the lifting operator of Assumption 2.I) and then seek \({\widehat{W}}\in (\mathrm {dom}\,(A))^m\) satisfying
This problem is uniquely solvable by Lemma 2.4, since \(\sigma (A)\subset \{ z\,:\,\mathrm {Re}\,z\le \omega \}\) and therefore \(\sigma (A)\cap \sigma ({\mathcal {S}})=\emptyset .\) We then define \({\widehat{U}}:={\widehat{V}}+{\widehat{W}}\), which solves (4.6a, 4.6b). To see uniqueness, one observes that the difference of two solutions of (4.6a, 4.6b) solves the homogeneous problem (\(\widehat{\varXi }=0\) and \(\widehat{F}=0\)) for which uniqueness was established in Lemma 2.4.
By Corollary 4.3, the union of the spectra of \(\delta (z)\) for \(|z|\le r_0\) has a positive distance \(d(r_0)>0\) from the imaginary axis. If we take \(k_0< d(r_0)/\omega \), then \(\sigma (k^{-1}\delta (z))\subset \{ s\,:\, \mathrm {Re}\,s>\omega \}\) for all \(|z|\le r_0\) and \(k\le k_0\). When \(\omega =0\), we can take any \(k_0\). By the previous considerations this implies unique solvability. \(\square \)
Proposition 4.5
Let \(U^k=\{U^k_n\}\) and \(u^k=\{u^k_n\}\) be sequences satisfying (2.7a–2.7c) with \(u^k_0=0\). The sequence \(V^k=\{ V^k_n\}=\partial ^k U^k\) satisfies
for data \(v^k_0=0\), \(G^k=\{G^k_n\}:=\partial ^k \{ F(t_n+k{\mathbf {c}})\}\), and \(\varTheta ^k=\{ \varTheta ^k_n\}:=\partial ^k \{ \varXi (t_n+k{\mathbf {c}})\}\). Moreover,
Proof
Recall that Eqs. (2.7a) and (2.7c) are equivalent to (2.7a) and (4.3), as shown in the proof of Proposition 4.1. Moreover, the latter equations are equivalent to (4.2a–4.2c) in the Z-domain. In the present case we have \(u_0=0\). For a given square matrix \({\mathcal {P}}\in {\mathbb {C}}^{m\times m}\) and an operator C, we have
which proves that
By Proposition 4.1, Eq. (4.9a–4.9c) are equivalent to (4.7a–4.7c). Finally (4.8a) follows from (4.2a), while (4.8b) follows from (4.4) and (4.8a). \(\square \)
Proposition 4.6
Let \(U^k=\{U^k_n\}\) and \(u^k=\{u^k_n\}\) be sequences satisfying (2.7a–2.7c) with \(u^k_0=0\). The sequence \(X^k=\{X^k_n\}=(\partial ^k)^{-1} U^k\) satisfies
for data \(x^k_0:=0\), \(H^k=\{H^k_n\}:=(\partial ^k)^{-1} \{ F(t_n+k{\mathbf {c}})\}\), \(\varGamma ^k=\{\varGamma ^k_n\}:=(\partial ^k)^{-1} \{ \varXi (t_n+k{\mathbf {c}})\}\).
Proof
Follow the proof of Proposition 4.5. \(\square \)
5 Some Lemmas regarding Runge–Kutta methods
In order to shorten the statements of the results of this section, in all of them we will understand that:
-
(1)
We have an RK method with coefficients \({\mathcal {Q}}, {\mathbf {b}}, {\mathbf {c}}\) satisfying Assumption 2.II (invertibility of \({\mathcal {Q}}\) and A-stability). The method has classical order p and stage order q.
-
(2)
We have an operator A in \({\mathcal {X}}\) that is the generator of a \(C_0\)-semigroup, characterized by the quantities M and \(\omega \) of Proposition 2.1. The associated Sobolev tower \(\{ {\mathcal {X}}_\mu \}\), obtained by interpolation of \({\text {dom}}(A^\mu) \) for positive integer values of \(\mu \), will also be used.
The following lemma will be used at a key point in the arguments below.
Lemma 5.1
Let A be a linear operator in \({\mathcal {X}}\) and q be a rational function bounded at infinity whose poles are outside \(\sigma (A)\). The following properties hold:
-
(a)
The operator q(A) maps \({\text {dom}} \left(A^\ell \right) \) to \({\text {dom}} \left( A^\ell \right) \) for all \(\ell .\)
-
(b)
If \(0\notin \sigma (A)\) and we define \(p(z):=z^{-\ell } q(z)\), then \(q(A)=p(A)A^\ell \) in \({\text {dom}} \left ( A^\ell \right )\).
Proof
To prove (a), show first by induction on \(\ell \) that \((A-\lambda I)^{-1}\) maps \({\text {dom}} \left( A^\ell \right) \) into \({\text {dom}}\left( A^{\ell +1} \right)\). Using this result for each of the factors in the definition (2.3) the result follows. To prove (b) note first that p is rational, bounded at infinity, and that \(\sigma (A)\) does not intersect the set of poles of p. Using Definition 2.2, we have \(p(A)=A^{-\ell } q(A)= q(A) A^{-\ell }\), and the result follows. \(\square \)
We start by recalling some simple facts about RK methods that we will need in the sequel. Using the notation \({\mathbf {c}}^{\ell }:=(c_{1}^{\ell },\ldots , c_{m}^{\ell })^\top \), the following equalities (order conditions) hold (see e.g. [1, 25]):
Therefore,
For a stiffly accurate method we have (2.15) and therefore
The following result is well-known. We just summarize it for ease of reference later on.
Lemma 5.2
(Discrete antiderivative and RK quadrature) Let \(f:[0,T]\rightarrow {\mathcal {X}}\), \(g:=\partial ^{-1}f\), \(G^k=\{ G^k_n\}=(\partial ^k)^{-1} \{ f(t_n+k{\mathbf {c}})\}\) and \(\{ g^k_n\}\) be given by the recursion
For the errors \(d^k_n:=g(t_n)-g^k_n\), and for n such that \(nk\le T\), we have the estimates
Additionally, at the stage level we have
Proof
Follows from the fact that the Runge–Kutta method defines a quadrature formula of order p. \(\square \)
5.1 Estimates on rational functions of the operator
The following results in this section are adaptations from [1]. While they focus on the case \(\beta =0\), we present the necessary generalizations to \(\beta =\pm 1\). We will use the rational functions
Note that these rational functions are bounded at infinity and that \(r_{\ell ,\beta }(0)=0\). We will also use the vector-valued rational function
and note that \({\mathbf {g}}(0)={\mathbf {0}}\) and \(r(z)=1+\mathbf{g}(z)^\top {\mathbf {1}}\).
Lemma 5.3
The rational functions (5.5) satisfy
The estimate (5.8) is also valid for \(\beta =-1\) if the method is stiffly accurate.
Proof
For the case \(\beta =0\), the proof is given in [1, Lemma 5]. For the case \(\beta =\pm 1\), an analogous proof can be brought to fruition. Namely, one can expand the inverse in the definition of \(r_{\ell ,\beta }\) into the Neumann series and apply the order conditions (5.2a, 5.2b). For \(\beta =-1\), one has to use (5.3) for the leading term. \(\square \)
Lemma 5.4
If the RK method satisfies Assumption 2.III, then there exists a constant \(k_0>0\) depending on the RK method and on \(\omega \) such that for \(\beta \in \{0,1\}\), \(\ell \le p-\beta \), and all \(0<k\le k_0\) and all n with \(0\le n\,k \le T\), we have the estimate
with \(\rho _k(T)\) defined in (2.11). If \(\ell =p\) and \(\beta =1\), the left-hand side of (5.9) is bounded by \(C\rho _k(T)\). The constant \(C>0\) in (5.9) depends only on the Runge–Kutta method, M and \(\omega \), \(k_0\), \(\ell \), and \(\mu \), but is independent of n and k. If the Runge–Kutta method is stiffly accurate, then the estimate (5.9) also holds for \(\beta =-1\). If \(\omega =0\), then \(k_0\) can be chosen arbitrarily.
Proof
We adapt the proof of [1, Lemma 6], which only covers the case \(\beta =0\). Consider first the case \(p-\ell -\beta \ge 0\) and take any integer \(\mu \) such that \(0\le \mu \le p-\ell -\beta \). Then
By Lemma 5.3, the rational function \(r_{\ell ,\beta }\) has a zero of order \(p-\ell -\beta +1\) at \(z=0\). The rational function \((r(z)-1)z^\mu \) has a zero of order \(\mu +1\le p-\ell -\beta +1\) at \(z=0\), and all other zeros are in \({\mathbb {C}}_+\) by A-stability and Assumption 2.III. This implies that the rational function \(q_{\ell ,\beta ,\mu }\) has its poles in
Therefore, for \(k_0>0\) sufficiently small we get using (2.4):
where C depends on M, \(\omega \), \(k_0\), and the RK method. By Lemma 5.1 we have
This, (5.11), and applying (2.11) to control \(r(kA)^{n+1}\) by \(\rho _k(T)\), proves (5.9) for integer \(\mu \le p-\ell -\beta \). For larger integer values of \(\mu \), the result does not need to be proved as the maximum rate is already attained. We just have to estimate the \({\mathcal {X}}_{p-\ell -\beta }\) norm by the stronger \({\mathcal {X}}_\mu \) norm. For real values of \(\mu \), we use interpolation.
We still need to prove the result when \(p-\ell -\beta =-1,\) which can only happen when \(\ell =p\) and \(\beta =1.\) We note that \(r_{p,1}(0)=0\) and we can therefore argue as in the previous case for \(\mu =0\). \(\square \)
Lemma 5.5
If the RK method satisfies Assumption 2.III and \(k_0\) is the value given in Lemma 5.4, then
for all \(k\le k_0\) and n such that \(nk\le T\).
Proof
Since \({\mathbf {g}}(0)=\mathbf{0 }\), we can adapt the proof of Lemma 5.4 to each of the components of the vector-valued function \({\mathbf {g}}\). The key step is to show that \(\mathbf{h}(z)^\top :=(r(z)-1)^{-1} {\mathbf {g}}(z)\) is bounded at infinity and has all its poles in the set defined in (5.10) and therefore
Since the operator \(s_n(k \,A) {\mathbf {g}}(k\, A)^\top \) on the left-hand side of (5.12) can be rewritten as \((r(kA)^{n+1}-I){\mathbf {h}}(k\,A)^\top \), the bound (5.12) follows readily. \(\square \)
When dealing with Runge–Kutta methods that do not satisfy the additional Assumption 2.III, we still have the following result:
Lemma 5.6
For \(k_0>0\) taken as in Lemma 5.4, we can bound for all \(k\le k_0\)
for \(\ell \le p\), \(\beta \in \{0,1\}\), and \(\mu \ge 0\). The constant C depends on M, \(\omega \), \(k_0\), \(\mu \), and the RK method. The estimate (5.13) also holds for \(\beta =-1\) if the method is stiffly accurate. Additionally
Proof
The argument to prove (5.13) is very similar to that of Lemma 5.4. By interpolation it is clear that we just need to prove the result for any integer \(\mu \) satisfying \(0\le \mu \le p+1-\ell -\beta \). Consider then the rational function \(q_{\ell ,\beta ,\mu }(z):=z^{-\mu }r_{\ell ,\beta }(z)\), which is bounded at infinity and has all its poles in \(\sigma ({\mathcal {Q}}^{-1})\) [see (5.10)]. We can then use the same argument to prove (5.11) for this redefined new function \(q_{\ell ,\beta ,\mu }\). (Note that we do not use Assumption 2.III in this argument.) Using that \(r_{\ell ,\beta }(k\,A)=k^\mu q_{\ell ,\beta ,\mu }(k A) A^\mu \) in \({\text {dom}}\left( A^\mu \right)\), the result follows. Stiff accuracy of the method is used in the case \(\beta =-1\) when we apply Lemma 5.3, dealing with the zeros of \(r_{\ell ,-1}\).
The proof of (5.14) is a similar adaptation of the proof of Lemma 5.5. \(\square \)
5.2 Estimates on discrete convolutions
The RK error will naturally induce several types of discrete convolutions that we will need to estimate separately. In all of them we will have the structure
We first deal with the simplest cases.
Lemma 5.7
For \(nk\le T\), the sequence defined by (5.15) can be bounded by
If \(\eta _n:={\mathbf {g}}(kA)^\top {\varvec{\xi }}_n\) for \({\varvec{\xi }}_n\in {\mathcal {X}}^m\), then
Proof
Follows by writing the recurrence (5.15) as a discrete convolution. \(\square \)
The next estimate is related to the consistency error of the RK method in the sense of how the RK method approximates derivatives at the stage level. We introduce the operator
The following well-known result about \({{\varvec{D}}}^k(y;t)\) underlies the proofs of [1, Theorem 1] and [25, Theorem 2].
Lemma 5.8
If \(y\in {\mathcal {C}}^{p+1}([0,T];{\mathcal {X}})\), then
where
Proof
Follows easily from the Taylor expansion and the order conditions (5.1a, 5.1b). \(\square \)
We are almost ready for the two main lemmas of this section, the first one without Assumption 2.III and the second one with it. These results and their proofs follow [1, Theorem 1 and 2], where only the case \(\beta =0\) is covered.
Lemma 5.9
Let \(y\in {\mathcal {C}}^{p+1}([0,T];{\mathcal {X}})\cap \mathcal C^p([0,T];{\mathcal {X}}_\mu )\) for some \(\mu \ge 0\). Let the sequence \(\omega _n\) be defined by (5.15) with \(\eta _n:=k^{-1}{\mathbf {g}}(kA)^\top (k{\mathcal {Q}})^\beta {\varvec{D}}(y;t_n)\). Then there exists a constant \(k_0>0\) depending only on \(\omega \) from (2.2) and the RK method such that for \(k \le k_0\), \(\beta \in \{0,1\}\) and for \(n k \le T\)
The estimate also holds for \(\beta =-1\) if the method is stiffly accurate. If \(\omega =0\), then \(k_0\) can be chosen arbitrarily.
Proof
Introduce
and note \(\eta _n=k^{-1}e^k_\beta (t_n)\). Using Lemmas 5.6 and 5.8 [recall the definition of \(r_{j,\beta }\) in (5.5)], we can bound
By (5.18), we then have
and the result then follows from Lemma 5.7. \(\square \)
Lemma 5.10
Let \(y\in {\mathcal {C}}^{p+1}([0,T];{\mathcal {X}}_\mu )\) for some \(\mu \ge 0\). Let the RK method satisfy Assumption 2.III. Then there exists a constant \(k_0>0\) depending only on \(\omega \) from (2.2) and the RK method such that the sequence \(\omega _n\) defined in Lemma 5.9 satisfies for \(\beta \in \{0,1\}\) and \(k\le k_0\)
If the method is stiffly accurate and \(y\in \mathcal {C}^{p+2}([0,T];{\mathcal {X}})\cap {\mathcal {C}}^{p+1}([0,T];\mathcal {X}_\mu )\), then for \(\beta =-1\)
If \(\omega =0\), then \(k_0\) can be chosen arbitrarily.
Proof
We will use the function \(e^k_\beta \) defined in (5.19) and Abel’s summation by parts:
an expression involving the rational functions \(s_n\) defined in (5.6) (recall that \(s_0=1\)). We first apply Lemmas 5.4, 5.5 and 5.8 to estimate
Since
and using that \(\left\| y^{(j)}(t)-y^{(j)}(t-k)\right\| _{{\mathcal {X}}_{\mu }} \le k\max _{t-k\le \tau \le t+k}\left\| y^{(j+1)}\right\| _{{\mathcal {X}}_{\mu }}\), a computation analogous to the above bound, but using \(y-y(\cdot -k)\) as data implies
and therefore
Note that if \(\beta \in \{0,1\}\) we can make a simpler estimate for the term originating from \({\mathbf {R}}^k\), (i.e., the one containing the highest derivative) using less regularity for y by not taking advantage of the difference between \(y^{(p+1)}(t_j)\) and \(y^{(p+1)}(t_{j-1})\) and thus end up requiring less regularity. Using the estimate (5.20) for the last term in (5.21), we have thereby already derived estimates for all three terms in (5.21). \(\square \)
6 Proofs
The two different cases (with or without Assumption 2.III) will be collected by using the parameter
6.1 Proof of Theorem 3.4
Recall that u solves (2.1) with \(u(0)=0\). The functions \(\varXi \) and F are the given boundary and volume data. If \(\varGamma :=\partial ^{-1} \varXi \) and \(H:=\partial ^{-1}F\), then \(x=\partial ^{-1} u\) solves
On the other hand, \(\{ X^k_n\}=(\partial ^k)^{-1} \{ U^k_n\}\) solves by Proposition 4.6:
Before we can estimate the difference between the functions x and \(x_n^k\), we need one final lemma.
Lemma 6.1
Let x solve
Assume that for some \(\mu \ge 0\) we have
Then \(x - {\mathscr {E}}\varGamma \in {\mathcal {C}}^p([0,T];\mathcal X_{\mu +1})\).
Proof
We set \(y:=x - {\mathscr {E}}\varGamma \). By assumption we have \(y \in {{\mathcal {C}}}^p([0,T];{\mathcal {X}}_{\mu })\) and \(B\left( x- {\mathscr {E}}\varGamma \right) =0\). Since \(x \in {\text {dom}}(A_\star )\) and \({\text {range }}{{\mathscr {E}}}\subset {\text {dom}}(A_\star )\) this implies \(y(t) \in {\text {dom}}(A)\) for all \(t \in [0,T]\). We further calculate using (6.4) and \({\text {range }}{\mathscr {E}}\subseteq {\text {ker}}(I-A_\star) \):
Each of the terms is assumed in \({\mathcal {C}}^p([0,T];\mathcal X_{\mu })\), thus \(y \in {\mathcal {C}}^p([0,T];{\mathcal {X}}_{\mu +1})\). \(\square \)
We will need the sequences \(\{ \gamma _n^k\}\) and \(\{ h_n^k\}\) with the scalar parts of the computations of \(\{\varGamma ^k_n\}\) and \(\{H^k_n\}\) respectively, namely (see Lemma 2.6),
We then consider
Using (6.5a), the definition \(\varGamma ^k=(\partial ^k)^{-1} \varXi \), and (2.13), we can write
Lemma 5.2 (take \(f={\mathscr {E}}\varXi \) for the first three inequalities and \(f=F\) for the last one) proves that
The error analysis is derived by tracking the evolution of the following differences
(compare (6.2) and (6.3b) to see the vanishing boundary condition for \(E_n^k\)) and note that by (6.7a)
which shows that we only need to estimate \(e^k_n\) to prove Theorem 3.4.
We start with the observation that x solves the following equation, as can be easily derived from Eq. (6.2):
Recalling that Assumption 2.I included the hypothesis \(\mathrm {range}\,{\mathscr {E}}\subset \ker (I-A_\star )\), we have \(({\mathcal {Q}}\otimes A_\star )\varDelta ^k_n={\mathcal {Q}}\varDelta ^k_n\). Combining (6.8) and (6.3a), we get
Naive estimation of the terms \({{\varvec{D}}}^k(x,t_n)\) and \(\varDelta ^k_n - {\mathbf {1}}\delta ^k_n\) would yield convergence rates similar to Propositions 3.1 and 3.2. In order to get an increased rate, as stated in Theorem 3.4, we combine these two terms using the function \(Y(t):=x(t) - {\mathscr {E}}\varGamma (t)\). Lemma 6.1 and the assumptions of Theorem 3.4 ensure \(Y \in {\mathcal {C}}^{p+\alpha }([0,T];{\mathcal {X}}_{\mu +1}) \cap {\mathcal {C}}^{p+1}([0,T];{\mathcal {X}})\).
We can thus further simplify
This then immediately gives (recall (5.7) for the definition of \({\mathbf {g}}\))
It is easy to see from (6.2) that x satisfies
Subtracting (6.3c) from this, inserting (6.10), using that \(({\mathbf {b}}^\top \otimes A_\star )\varDelta ^k_n={\mathbf {b}}^\top \varDelta ^k_n\), and setting
we have
What is left is the careful combination of terms so that we can bound everything using Lemmas 5.7, 5.9, and 5.10 by writing
Since the above recurrence defining \(\{e^k_n\}\) is linear as a function of the right-hand side, we can estimate its norm by adding the effects of each of the terms. In the order in which they appear in the last expression, we use: Lemmas 5.9–5.10 with \(\beta =0\), but noting that \(Y(t) \in {\text {dom}}(A^{\mu +1})\); Lemmas 5.9–5.10 with \(\beta =1\); Lemma 5.7 combined with (6.7a) and (6.7d); Lemma 5.7 combined with (6.7c) and (6.7b); for the first term of \(\varphi ^k_h\) we use Lemma 5.7 combined with Lemma 5.2 with \(f:={\dot{x}}\). Finally, for the second contribution to \(\varphi ^k_h\), we use (5.4c).
Combined, these results give
If we apply Lemma 6.1 to bound the \({\mathcal {X}}_{\mu +1}\)-norm, we arrive at the stated estimate.
6.2 Proof of Theorem 3.5
This proof is very similar to the one of Theorem 3.4 but slightly simpler. We will point out the main steps of the proof. We first focus on showing the estimate for \(v-v^k\). Note that we use the simple form of \(\partial ^k\) for stiffly accurate RK methods given in Lemma 2.7. We define \(G:=\dot{F}\) and \(\varTheta :={\dot{\varXi }}\) so that \(v=\dot{u}\) satisfies
Its RK approximation
and \(\{V_n^k\}=\partial ^k \{U_n^k\}\) satisfies (see Proposition 4.5 and Lemma 2.7, where we use stiff accuracy of the RK scheme, and recall that \(\{G^k_n\}=\partial ^k \{F(t_n+k{\mathbf {c}})\}\) and \(\{\varTheta ^k_n\}=\partial ^k\{ \varXi (t_n+k{\mathbf {c}})\}\))
Let then
and [note (6.11b) and (6.11c)]
By (6.11a) and (6.12a), using that \(({\mathcal {Q}}\otimes A_\star )\varDelta _n^k={\mathcal {Q}}\varDelta ^k_n\) (assumption on the lifting) and Lemma 2.7 to represent \(G_n^k\), we have
and therefore, from (6.11c) and (6.12c)
The final term can be shown to be of order \({\mathcal {O}}(k^{p+1})\) by combining (5.17) with (5.2b) and (5.3). We then use Lemmas 5.9 and 5.10 with \(\beta =-1\) and \(\beta =0\) as well as Lemma 5.7 to bound
Finally Propositions 3.1 and 3.2 are used to bound
The estimate involving \(A_{\star }u\) can be proved as an easy corollary of the estimate on v. Since the last stage of a stiffly accurate method is the step, we have that (4.8a) implies that
and therefore
7 Maximal dissipative operators in Hilbert space
In this short section we summarize some results that show that the hypotheses on the abstract equation and its discretization are simpler for maximal dissipative operators on Hilbert spaces. These results are well-known and will be needed when applying the theory developed in the previous sections to some model problems in Sect. 8.
If A is maximal dissipative in the Hilbert space \(\mathcal X\), i.e.,
and if \(A-I:{\text {dom}}(A)\rightarrow {\mathcal {X}}\) is invertible with bounded inverse, then the constants in Proposition 2.1 can be chosen as \(M=1\) and \(\omega =0\). In this case A generates a contraction semigroup in \({\mathcal {H}}\). See [26, Sect. 1.4].
In particular, if the RK method satisfies Assumption 2.II and
then the Eq. (2.7a–2.7c) [or equivalently (2.8a–2.8c)], defining the RK approximation of (2.1a–2.1c) are uniquely solvable for any \(k>0\) (apply Lemma 2.4 with \(\mathcal S=k^{-1}{\mathcal {Q}}^{-1}\)). The following lemma gives a bound for \(\rho _k(T)\) in this specific setting.
Lemma 7.1
(Discrete stability) Let \(A\) be a linear, maximal dissipative operator on a Hilbert space \({\mathcal {H}}\). For A-stable Runge–Kutta methods and arbitrary \(k>0\), we can bound
and therefore \(\rho _k(T)\le 1\) for all k and \(T>0\).
Proof
Let \(c(z):=(z+1)/(z-1)\), and note that \(c(A)=(A+I)(A-I)^{-1}\) is well defined and since
it is clear that \(\Vert c(A)\Vert _{{\mathcal {H}} \rightarrow {\mathcal {H}}}\le 1.\) Consider now the rational function \(q:=r\circ c\). Since c maps B(0; 1) bijectively into \(\{ z:\mathrm {Re}\,z<0\}\) and r maps the latter set to B(0; 1) (this is A-stability), it follows that \(q:B(0;1)\rightarrow B(0;1)\). Since \(\sigma (c(A))\subset \overline{B(0;1)}\) and c(A) is bounded, we can define q(c(A)) and show (use a classical result of von Neumann [37, Section 4] or [30, Chapter XI, Section 154]) that \(\Vert q(c(A))\Vert _{{\mathcal {H}} \rightarrow {\mathcal {H}}}\le 1\).
Finally, using that \(c(c(z))=z\) for all z, it follows that \(r=q\circ c\). It is then an easy computation to prove that \(r(A)=q(c(A))\). (We remark that this equality can also be proved using functional calculus.) \(\square \)
In Propositions 3.1 and 3.2, if A is maximal dissipative, \(k_0\) can be chosen arbitrarily. In Lemma 5.4, if \(A\) is maximal dissipative, \(k_0\) can be chosen arbitrarily.
8 Applications
In this section \(\varOmega \) is a bounded Lipschitz open set in \({\mathbb {R}}^d\) (\(d=2\) or 3) with boundary \(\varGamma \).
We use the usual (fractional) Sobolev spaces \(H^s(\varOmega )\) for \(s\ge 0\) and introduce the space \(H^1_{\varDelta }(\varOmega ):=\{ u \in H^1(\varOmega ): \varDelta u \in L^2(\varOmega ) \}\). On the boundary \(\varGamma \), we also consider Sobolev spaces \(H^s(\varGamma )\) and their duals \(H^{-s}(\varGamma )\). Details can, for example be found in [22].
For the trace operators, we make the convention that the index \(+\) relates to exterior and − means the trace is taken from the interior of \(\varOmega \). For example, the two bounded surjective trace operators \(\gamma ^\pm :H^1({\mathbb {R}}^d\setminus \varGamma )\rightarrow H^{1/2}(\varGamma )\) denote the trace from \({\mathbb {R}}^d\setminus {\overline{\varOmega }}\) and \(\varOmega \), respectively. and we will denote \(H^{-1/2}(\varGamma )\) for the dual of the trace space. The angled bracket \(\langle \,\cdot \,,\,\cdot \,\rangle _\varGamma \) will be used for the \(H^{-1/2}(\varGamma )\times H^{1/2}(\varGamma )\) duality pairing and \((\cdot ,\cdot )_{{\mathbb {R}}^d}\) will be used for the inner product in \(L^2({\mathbb {R}}^d)\) and \(\big [L^2({\mathbb {R}}^d)\big ]^d\). We will also use the normal traces \(\gamma ^\pm _\nu : H({\text {div}},{\mathbb {R}}^d \setminus \varGamma )\rightarrow H^{-1/2}(\varGamma )\) and the normal derivative operators \(\partial _\nu ^\pm \). Here we make the convention that the normal derivative points out of \(\varOmega \) for both interior and exterior trace.
We note that the applications in this section are chosen for their simplicity. More complicated applications, also involving full discretizations by convolution quadrature and boundary elements of systems of time domain boundary integral equations can be found in [29] and [27].
8.1 Boundary integral equations and convolution quadrature
In this section, we give a very brief introduction to boundary integral equations and their discretization using convolution quadrature. In that way, we can later easily state our methods for both the heat and wave equations in a concise and unified language. We present the result mostly formally, but note that they can be made rigorous under mild assumptions on the appearing functions. This theory can be found in most monographs on boundary element methods, see e.g. [22, 32, 33] or [31].
For \(s \in {\mathbb {C}}_{+}\), we consider solutions \(u \in H^1({\mathbb {R}}^d\setminus \varGamma )\) to the Helmholtz equation
For this problem, the fundamental solution is given by
where \(H_0^{(1)}\) denotes the first kind Hankel function of order 0. Using the representation formula, u can be rewritten using only its boundary data:
where the single layer and double layer potentials are given by
and the expressions \(\llbracket \gamma u \rrbracket :=\gamma ^- u - \gamma ^+ u\) and \(\llbracket \partial _{\nu } u\rrbracket := \partial _{\nu}^- u - \partial _{\nu}^+u\) denote the jump of the trace of v and normal derivative across \(\varGamma \).
We note that both \(S(s) \lambda \) and \(D(s) \psi \) solve the Helmholtz equation for any given densities \(\lambda \in H^{-1/2}(\varGamma )\) and \(\psi \in H^{1/2}(\varGamma )\).
We will need the following four boundary integral operators:
When solving problems in the time domain, we can leverage our knowledge of the Helmholtz equation using the Laplace transform \({\mathscr {L}}\). For an operator valued analytic function F with \({\text {dom}}(F) \supset {\mathbb {C}}_{+}\), we can then define the convolution operator \(F(\partial ):={\mathscr {L}}^{-1} \circ \mathrm F \circ {\mathscr {L}}\), where \({\mathscr {L}}\) is the Laplace transform in the sense of causal distributions. (Precise definitions can be found in [31, Chapter 3] and [21]).
Given a Runge–Kutta method, it is then easy to define the convolution quadrature approximation to such operators, as was introduced in [19]. We just replace the Laplace transform by the Z-transform \(\mathcal {Z}\) and s with the function \(\delta /k\), i.e., we define:
where g denotes a sequence in the shared domain of F(s), and \(k>0\) denotes the step size. The matrix-valued function \(z \mapsto F(\frac{\delta (z)}{k})\) is defined using the Riesz-Dunford calculus, but can be computed in practice by diagonalizing the argument.
Remark 8.1
We note that our use of the notation \(\partial ^k\) and \((\partial ^k)^{-1}\) is consistent with this definition by using the functions \(F(s):=s\) and \(F(s):=s^{-1}\).
8.2 An exotic transmission problem
In this section we show how to apply Theorems 3.4 and 3.5 to a transmission problem in free space associated to the infinitesimal generator of a group of isometries (both \(\pm A\) are maximal dissipative) with some exotic transmission conditions which impose partial observation of a trace. In Sect. 8.3 we will explain how this problem is related to a boundary integral representation of a scattering problem and how the current results yield the analysis of a fully discrete method for that integral representation. We keep the presentation brief. For more details and exemplary applications we refer to [13].
Let \(Y_h\) be a closed subspace of \(H^{1/2}(\varGamma )\) (in practice it will be finite-dimensional) and consider the spaces
The condition \(\llbracket \gamma v \rrbracket \in Y_h\) is equivalent to
We then set
In \({\mathcal {X}}\) we use the natural inner product, in \({\mathcal {V}}\) we use the norm of \(H^1({\mathbb {R}}^d\setminus \varGamma )\times \mathbf{H}(\mathrm {div},{\mathbb {R}}^d\setminus \varGamma )\), and in \({\mathcal {M}}\) we use the usual norm. We will define \(A_\star :{\text {dom}}\left( A_\star \right) ={\mathcal {V}}\rightarrow {\mathcal {X}}\) and \(B:{\mathcal {V}} \rightarrow {\mathcal {M}}\) by
understanding that \(A_\star \) can also be extended to \(H^1({\mathbb {R}}^d\setminus \varGamma )\times \mathbf {H}(\mathrm {div},{\mathbb {R}}^d\setminus \varGamma )\). As we did in Assumption 2.I, we consider \({\text {dom}} (A)=\ker B =\mathrm V_h\times {\mathbf {W}}_h^0\) and define A as the restriction of \(A_\star \) to this subset.
Proposition 8.2
The operators \(\pm A\) are maximal dissipative.
Proof
The identity (8.5) shows that \(\langle A(v,\mathbf {w}),(v,{\mathbf {w}})\rangle _{{\mathcal {X}}}=0\) for all \((v,{\mathbf {w}})\in \mathrm {V}_h \times {\mathbf {W}}_h^0.\) Given \((f,{\mathbf {f}})\in \mathcal {X}\), solving the coercive problem
and defining \({\mathbf {w}}=\nabla v+{\mathbf {f}}\), we have a pair \((v,{\mathbf {w}})\in \mathrm {V}_h \times {\mathbf {W}}_h^0\) such that \((v,{\mathbf {w}})-A(v,{\mathbf {w}})=(f,{\mathbf {f}})\) and thus A is maximal dissipative. The proof that \(-A\) is maximal dissipative is similar. (Note that this is a particular case of what appears in [13].) \(\square \)
We consider the standard problem (2.1a–2.1c) with vanishing initial conditions and data \(F=0\) and \(\varXi =g:[0,\infty )\rightarrow L^2(\varGamma )\), namely, we look for \((v_h,{\mathbf {w}}_h):[0,\infty )\rightarrow {\text {dom}} \left( A_\star \right) \) such that
Uniqueness of the solution to (8.6a–8.6c) follows from Proposition 8.2. We will handle existence of a solution below. The quantities of interest are \(u_h:=\partial ^{-1}v_h\) and its Dirichlet trace \(\psi _h:=\llbracket \gamma u_h \rrbracket :\; [0,\infty ) \rightarrow Y_h\).
Proposition 8.3
There exists a linear bounded right inverse \({\mathscr {E}}:{\mathcal {M}} \rightarrow {\text {dom}} \left( A_\star \right) \) of B such that \(\mathrm {range}\,{\mathscr {E}}\subset \ker (I-A_\star )\). The norm of \({\mathscr {E}}\) is independent of the space \(Y_h\).
Proof
Given \(\xi \in {\mathcal {M}}=H^{-1/2}(\varGamma )\), we solve the coercive problem
and we set \({\mathbf {w}}:=\nabla v\).
This problem is equivalent to (note (8.5))
Since \(\left| \langle \xi , \gamma ^+ {w} \rangle _{\varGamma }\right| \lesssim \left\| \xi \right\| _{H^{-1/2}(\varGamma )} \left\| w\right\| _{H^1({\mathbb {R}}^d \setminus \varGamma )} \) it follows that the norm of the solution operator for (8.7) is independent of the space \(Y_h\). \(\square \)
Proposition 8.4
The lifting \({\mathscr {E}}\) from Proposition 8.3 is a bounded linear map \(L^2(\varGamma ) \rightarrow {\mathcal {X}}_{1/2}:=[\mathcal {X},{\text {dom}} (A)]_{1/2}\) with \(\left\| {\mathscr {E}}\lambda \right\| _{{\mathcal {X}}_{1/2}} \le C \left\| \lambda \right\| _{L^2(\varGamma )}.\) The constant C depends only of \(\varOmega \).
Proof
We will need spaces encoding homogeneous normal traces:
Let \(\lambda \in L^2(\varGamma )\) be given. By applying Theorem A.4 to the exterior and setting \(\widetilde{{\mathbf {w}}} =0 \) inside, we can construct a function \(\widetilde{{\mathbf {w}}} \in \mathbf {H}(\mathrm {div},{\mathbb {R}}^d\setminus \varGamma ) \) satisfying \(\llbracket \gamma _{\nu } \widetilde{{\mathbf {w}}}\rrbracket = \lambda \) and
Upon identifying the product of function spaces on \(\varOmega \) and \({\mathbb {R}}^d \setminus \overline {\varOmega} \) with a function space on \({\mathbb {R}}^d \setminus \overline {\varGamma} \), we have
The product of interpolation spaces equals the interpolation of product spaces (cf. Lemma A.5); we can therefore also estimate:
If we consider \((v,{\mathbf {w}}) := {\mathscr {E}}\lambda \), then \((v, {\mathbf {w}} - \widetilde{{\mathbf {w}}}) \in {\text {dom}}(A)\) by construction of the lifting. Thus we have
The continuity of \({\mathscr {E}}\) from Proposition 8.3 and (8.9) conclude the proof. \(\square \)
Proposition 8.5
If \(g\in {\mathcal {C}}^2([0,\infty );H^{-1/2}(\varGamma ))\) satisfies \(g(0)=\dot{g}(0)=0\), then (8.6a–8.6c) has a unique strong solution.
Proof
Thanks to Propositions 8.2 and 8.3, this problem fits in the abstract framework described in [13], which proves existence and uniqueness of solution to (8.6a–8.6c). \(\square \)
Propositions 8.2 and 8.3 have some consequences. First of all, Assumption 2.I holds. Secondly, assuming \(g(t) \in L^2(\varGamma )\), any solution to (2.1a–2.1c) with the above data (\(F=0\), \(\varXi =g\)) is in \(\mathcal X_{1/2}\), and therefore, solutions to (8.6a–8.6c) take values in \({\mathcal {X}}_{1/2}\) as well. Finally, if \(g\in \mathcal C^s([0,\infty ];L^2(\varGamma ))\) then \({\mathscr {E}}g\in \mathcal C^s([0,\infty ];{\mathcal {X}}_{1/2})\).
We also need a regularity result that allows us to bound time derivatives of the solution in terms of the data. The continuity condition for the \((s+2)\)-nd derivative of g in Proposition 8.6 can be relaxed to local integrability, but then the norms on the right-hand side of (8.10) have to be modified.
Proposition 8.6
If \(g\in {\mathcal {C}}^{s+2}([0,\infty );L^2(\varGamma ))\) satisfies \(g^{(\ell )}(0)=0\) for \(\ell \le s+1\), then the unique solution to (8.6a–8.6c) satisfies
-
(a)
\((v_h,{\mathbf {w}}_h)\in {\mathcal {C}}^{s+1}([0,\infty );{\mathcal {X}})\),
-
(b)
\((v_h,{\mathbf {w}}_h)\in \mathcal C^{s}([0,\infty );{\mathcal {V}})\) and \((v_h,{\mathbf {w}}_h)\in \mathcal C^{s}([0,\infty );{\mathcal {X}}_{1/2})\),
-
(c)
For all \(\ell \le s\), there exists C, independent of the choice of \(Y_h\), such that for all \(t\ge 0\)
$$\begin{aligned} \Vert (v_h^{(\ell )}(t),{\mathbf {w}}_h^{(\ell )}(t))\Vert _{{\mathcal {X}}_{1/2}} \le C\,t\,\sum _{j=\ell }^{\ell +2} \max _{\tau \le t} \Vert g^{(j)}(\tau )\Vert _{L^2(\varGamma )}. \end{aligned}$$(8.10)
Proof
This result follows from [13, Theorem 3.1]. To see item (b), we note that \((v_h,{\mathbf {w}}_h)\) is constructed by writing
with \((v^0_h(t),{\mathbf {w}}^0_h(t)) \in {\text {dom}}(A)\). The statement then follows from Proposition 8.4. \(\square \)
We now consider the RK approximation of (8.6a–8.6c) in a finite time interval [0, T], which provides pairs of stage values \((V^k_{h,n},{\mathbf {W}}^k_{h,n})\in {\mathcal {X}}^m\) and step approximations \((v^k_{h,n},{\mathbf {w}}^k_{h,n})\in {\mathcal {X}}\). We then define
with \(u^k_{h,0}=0\) (see Lemma 2.6) and \(\psi ^k_{h,n}:=\llbracket \gamma u^k_{h,n} \rrbracket \) .
Proposition 8.7
For sufficiently smooth g, with RK approximations using a method satisfying Assumption 2.II, and with \(\alpha \) given by (6.1), for \(nk\le T\) we have the estimates
and
The constants depend on \(\varGamma \) and the Runge–Kutta method, but do not depend on T or on the choice of \(Y_h\).
Proof
We will use Theorems 3.4 as well as Propositions 3.1 and 3.2. We note that \(\rho _k(T)\le 1\) by Lemma 7.1 and Proposition 8.2. Also, with the \({\mathscr {E}}\) operator of Proposition 8.3, we have
with C independent of \(Y_h\). The bound (8.12) follows from Theorem 3.4, using (8.10) and (8.14) to estimate the right-hand side. The bound
follows from Propositions 3.1 and 3.2, using (8.10) for the estimate in terms of the data. The \(H^1({\mathbb {R}}^d\setminus \varGamma )\) estimate (8.13) is then a direct consequence of (8.12) and (8.15). The estimate for \(\psi _h-\psi _h^k\) follows from the standard trace theorem. \(\square \)
8.3 Scattering
As it does not incur much difficulty, we cover both the exterior scattering problem, which is an exterior Neumann problem, as well as the interior Neumann problem. In order to do so, we define the domains \(\varOmega ^+:={\mathbb {R}}^d \setminus \overline{\varOmega }\) and \(\varOmega ^-:=\varOmega \) and distinguish the problems by adding the superscripts \(+\) or − to the functions involved. We stay in the geometric setting of the previous section. Assume that \({\mathbf {d}}\in {\mathbb {R}}^d\) is a unit vector (direction of propagation) and that \(c\in {\mathbb {R}}\) is such that \(\varOmega \subset \{{\mathbf {x}}\in {\mathbb {R}}^d\,:\, \mathbf {x}\cdot {\mathbf {d}}> c\}\). Let \(\phi :{\mathbb {R}}\rightarrow {\mathbb {R}}\) be a function such that \(\phi (r)=0\) for all \(r\ge c.\) The incident wave \(u^{\mathrm {inc}}({\mathbf {x}},t):=\phi ({\mathbf {x}}\cdot {\mathbf {d}}-t)\) propagates in the direction \({\mathbf {d}}\) at unit speed and has not reached the scatterer given by \(\varOmega \) at time \(t=0\). The data for our problem will be the function \(g:[0,T]\rightarrow L^2(\varGamma )\) given by \(g(t):=-\partial ^{\pm }_\nu u^{inc}(\cdot ,t)\).
The problems under consideration are: Find \(u^{\pm }:[0,T]\rightarrow H^1(\varOmega ^\pm )\) satisfying
so that \(\partial ^{\pm }_\nu (u^{\pm }+u^{\mathrm {inc}})=0\). (Note that we can take the trace of the normal derivative of the incident wave, since it is locally smooth.) The exterior problem (posed on \(\varOmega ^+\)) is the classical sound soft scattering problem of the incident wave \(u^{\mathrm inc}\).
A direct formulation for solving this problem is obtained by extending the function by zero to the complement of the domain of interest. That is, we solve:
with \(u^{\pm }(0)=\dot{u}^{\pm }(0)=0.\) By imposing some additional hypotheses on the growth of g (which is needed to have a well-defined distributional Laplace transform), we can represent the solution to (8.16) as \(u^{\pm }={\mp }S(\partial )g - D(\partial ) \psi ^{\pm }\), where \(\psi ^{\pm }:=\llbracket \gamma u^{\pm } \rrbracket \). Note that, to be precise with the use of weak distributional definitions, all functions have to be extended by zero to \(t<0\) (we say that they are causal) and the time interval is extended to infinity.
Taking the trace in this representation formula, the solution of (8.16) can be found by solving an equation for \(\psi ^{\pm }\) and then postprocessing with the potential operators:
and we still have that \(\psi ^{\pm }=\llbracket \gamma u^{\pm } \rrbracket \).
For simplicity of notation, we will skip the indices ± for the different functions from now on. We can equivalently write (8.16) and the equivalent (8.17) by using the variables \(v:=\dot{u}\) and \({\mathbf {w}}:=\nabla u\). We note that \(u=\partial ^{-1} v\) and \(\psi =\partial ^{-1} \llbracket \gamma v \rrbracket \). Here, \((v,{\mathbf {w}})\) solve (we restrict t to the interval [0, T] again)
that is, (8.6a–8.6c) with \(Y_h=H^{1/2}(\varGamma )\).
For the discretization, we consider a finite dimensional space \(Y_h\) and the Galerkin approximation to (8.17), so that we look for \(\psi _h:\,{\mathbb {R}}\rightarrow X_h\) causal such that
The functions \(v_h:={\dot{u}}_h\) and \({\mathbf {w}}_h:=\nabla u_h\) satisfy (8.6a–8.6c). The difference between the solutions of (8.16) and (8.18) can be studied by comparing the solutions to (8.6a–8.6c) when \(Y_h=H^{1/2}(\varGamma )\) and when \(Y_h\) is a finite dimensional space, see [13] for details. For our purposes, it is sufficient to note that we get quasi-optimal estimates for the discretization in space.
Discretization in time is performed by applying convolution quadrature to (8.18). The fully discrete solution reads
The approximations \(\psi ^k_h\) and \(u^k_h\) are then computed by the usual post-processing, i.e.,
Lemma 8.8
The sequences \(u^k_h\) and \(\psi _h^k\) computed via (8.19) coincide with the Runge–Kutta approximations to (8.6a–8.6c) and their traces respectively.
Proof
The details of the computation can be found in the appendix of [23]. The basic idea is to take the Z-transform and show that both approaches solve the matrix-valued Helmholtz problem (4.2a–4.2c). \(\square \)
This gives the following immediate corollary, representing an a priori bound for the fully discrete method:
Corollary 8.9
Let the assumptions of Proposition 8.7 hold. Then for \(u_h\) and \(\psi _h\), approximated using convolution quadrature, we can estimate:
The constants depend on \(\varGamma \) and the Runge–Kutta method, but do not depend on T or on the choice of \(Y_h\).
Remark 8.10
There is another approach for analyzing convolution quadrature methods, which is based on estimates in the Laplace domain. It can be shown that the Neumann-to-Dirichlet map, realized by the boundary integral Eq. (8.19), satisfies a bound of the form
see [20, Appendix 2]. Applying the abstract theory of [2] then implies convergence rate \(\min (q+1,p)\) for the boundary data \(\psi _h\). Modifying their proof, one can also get for \(\widehat{g} \in L^2(\varGamma )\) that
which would yield the same convergence rate as Corollary 8.20, but without insight into the dependence on the end-time T. \(\blacksquare \)
8.4 Numerical example
We solve (8.19) on a “hollow square”, as depicted in Fig. 1, and focus on the interior Neumann problem, i.e. computing \(\psi ^-=:\psi \). The geometry was chosen to be non-convex and not simply connected, in order to test if the rate observed is a general result, or if our estimates might prove sharp in some situation.
We prescribe the exact solution as a traveling wave, given by
\(s_0:=4\) is chosen so that \(\phi (0)\) is sufficiently small in the domain. We set \({\mathbf {d}}:=[\frac{\sqrt{2}}{2},\frac{\sqrt{2}}{2}]^\top \) and solve up to an end time of \(T=12\). An approximation of the \(H^{1/2}\)-error is computed via
i.e., we compare to the \(L^2\)-projection of the exact solution. Since we are interested in the convergence rate with respect to the timestep size k, we consider a fixed, but sufficiently fine mesh.
We used 3 and 5 stage Radau IIA methods, with orders (q, p) of (3, 5) and (5, 9), respectively (see [16] for their definition). While their strong damping properties are not advantageous for wave propagation problems, they nevertheless are the standard method used with convolution quadrature. This is in part due to the fact that the standard theory (see, e.g., [2]) makes some assumptions not satisfied by the Gauss methods. A more detailed analysis of the dissipation and dispersion of the Radau methods was performed in [7, Sect. 4.3], showing that higher order Runge–Kutta methods posess favorable properties compared to their low order brethren.
Our theory predicts convergence rates of 4.5 and 6.5. In Fig. 2, we observe a rate that is closer to 5 and 8. This means that (just like the standard Laplace-domain estimates) our estimates do not appear to be sharp in this case. Further investigations into the cause of this phenomenon are required. Results trying to explain this phenomenon, initially prompted by the work on this article, can be found in [24] but with a different model problem.
Performance of Radau IIA methods for the wave equation, cf. Sect. 8.4
8.5 The heat equation
In this section, as an example where our estimates turn out to be sharp, we consider a heat conduction problem and will apply Theorem 3.5 to get convergence of the boundary trace. The physical situation is a body \(\varOmega \subset {\mathbb {R}}^d\) that is held at a given temperature distribution and radiates heat into a medium \(\varOmega ^+:={\mathbb {R}}^d \setminus \overline{\varOmega }\). We make the simplifying assumption that at \(t=0\) the temperature is 0. Since the problem is posed on an unbounded domain, it is a good candidate for boundary integral equations, while being simple enough to showcase our more general results. We only briefly give the mathematical setting. More details and a more involved physical example can be found in [27]. The setting is as follows: find \(u: {\mathbb {R}}_+ \rightarrow H^1_{\varDelta }(\varOmega ^+)\) such that
It is well-known that \(\varDelta \) with homogeneous Dirichlet boundary conditions generates an analytic semigroup (see e.g. [26, Sect. 7.2]) on \(L^2({\mathbb {R}}^d \setminus \overline{\varOmega })\). The rest of our assumptions are also easily checked. We summarize:
-
(i)
\({\text {dom}}(A_{\star })=\{u \in H^1({\mathbb {R}}^d \setminus \overline{\varOmega }): \varDelta u \in L^2({\mathbb {R}}^d \setminus \overline{\varOmega }) \}\),
-
(ii)
\(B: H^1({\mathbb {R}}^d \setminus \overline{\varOmega }) \rightarrow H^{1/2}(\varGamma )=:{\mathcal {M}}, \; Bv := \gamma ^+ v\) (using the standard trace operator).
In order to derive the boundary integral formulation, we take the Laplace transform of (8.21a), giving for \(\kappa :=\sqrt{s}\):
which is Helmholtz’s equation for a complex wave number \(\kappa \). We make an ansatz of the form \(\widehat{u}=S(\kappa ) \widehat{\lambda }\) for some unknown density \(\widehat{\lambda }\), which can be determined by applying the trace operator, giving the equation \(V(\kappa )\widehat{\lambda }={\mathscr {L}}(g)\).
Transforming back, and using the definition \(V_{\kappa }(s):=V(\sqrt{s})\), we get the formulation:
The solution u can then be recovered by computing \(u=S_{\kappa }(\partial )\), where \(S_\kappa (s):=S(\sqrt{s})\).
The discrete version of this is then given by solving
It can be shown that plugging the discrete solution into the representation formula \(U^k:=S_{\kappa }(\partial ^k) \varLambda ^k\) gives back the Runge–Kutta approximation of (8.21a–8.21c). The approximations at the endpoints \(t_n=n\,k\), denoted by \(\lambda ^k\) and \(u^k\) respectively can be computed by the usual post-processing. We refer to the appendix of [23] for an analogous computation in the context of the Schrödinger equation, which easily transfers to our situation. For simplicity, we do not consider any discretization in space. A Galerkin approach could easily be included into the analysis, analogously to Sect. 8.2.
From the definition \(A:=A_{\star }|_{\ker (B)}\) we get \({\text {dom}}(A):= \big \{ u \in H^1({\mathbb {R}}^d \setminus \overline{\varOmega }): \varDelta u \in L^2({\mathbb {R}}^d \setminus \overline{\varOmega }), \; \gamma ^+ u =0 \big \}. \) We need the following analog of Proposition 8.4:
Proposition 8.11
\({\text {dom}}(A_{\star }) \subseteq [L^2({\mathbb {R}}^d \setminus \overline{\varOmega }), {\text {dom}}(A)]_{\mu ,\infty }\) for \(\mu \in [0,1/4]\).
Proof
It is easy to see that \(H^2_0({\mathbb {R}}^d \setminus \overline{\varOmega }) \subseteq {\text {dom}}(A)\).
Using the Besov spaces introduced in Appendix A, we can write, if \(\mu \le 1/4\):
where in the last step, we used [22, Theorem B.9]. \(\square \)
The convergence of our numerical method can then be analyzed quite easily using Proposition 3.2 and Theorem 3.5.
Theorem 8.12
Let p and q denote the classical and stage order of the Runge–Kutta method used. Let \(g \in {\mathcal {C}}^{p+3}([0,T], H^{1/2}(\varGamma ))\) with \(g^{(j)}(0)=0\) for \(j=0,\ldots , p+2\). Set \(\alpha :=1\) if the Runge–Kutta method satisfies Assumption 2.III and \(\alpha :=0\) otherwise. Then the following estimate holds for the post-processed approximation:
Assume that the Runge–Kutta method used for discretization is stiffly accurate. Then the following estimates hold for the \(H^1\)-norm:
with
For the density, we get:
where the rate \(r_{\lambda }\) is given by:
Proof
We first note that we can control the derivatives \(u^{(\ell )}\) by the data. This can be done completely analogous to Proposition 8.6 by the techniques of [13]. The estimates read:
For simplicity of notation, we only consider the case \(q<p-1\). All the other cases follow analogously but giving different rates when applying the abstract theory. By Proposition 8.11, we can apply Propositions 3.1 or 3.2 with \(\mu =1/4\), depending on whether we are in the setting \(\alpha =0\) or \(\alpha =1\). This gives estimate (8.23).
Applying Theorem 3.5, we get the following convergence in the graph norm of \(A_{\star }\):
Since for stiffly accurate RK-methods, \(u^k\) also satisfies the boundary conditions (it is just the last entry of the stage vector) we get from the Dirichlet-boundary conditions that \(\gamma ^+ u(t_n) = g(t_n) = \gamma ^+ u^k(t_n)\). Therefore, integration by parts and the Cauchy-Schwarz inequality give:
Estimate (8.24) then follows from (8.23) and (8.26). For the estimate (8.25) of the density, we fix \(\xi \in H^{1/2}(\varGamma )\), and let v denote a lifting to \(H^1({\mathbb {R}}^d)\). We calculate
We are still free to choose the precise lifting v. Doing so as in [31, Proposition 2.5.1], we get
The result then follows from the previous estimates. \(\square \)
Remark 8.13
Note that in the cases \(q=p-1\) with \(\alpha =1\) and \(q=p\) with \(\alpha =0\), the rates \(r_1\) and \(r_{\lambda }\) in Theorem 8.12 are sharp from what can be extracted from Theorem 3.5 and Propositions 3.1 and 3.2. Nevertheless, we expect it to be possible to extract better rates from a more explicit investigation of these limiting cases.
8.5.1 Numerical example
In order to demonstrate that the estimate (8.25) is sharp, we consider a simple model problem. Following [34], we take \(\varOmega \) to be the unit sphere and consider a right-hand side g(x, t) of the form
where \(Y^m_n\) is the spherical harmonic of degree n and order m. It is well-known that the spherical harmonics are eigenfunctions of the pertinent boundary integral operators. Most notably for us, we have
where \(j_n\) denotes the spherical Bessel functions and \(h_n^{(1)}\) is the spherical Hankel function of the first kind. Due to this relation, solving (8.22) becomes a purely one dimensional problem, i.e., we can write \(\lambda (x,t)={\widetilde{\lambda }}(t) Y^m_n(x)\) and the solution can be easily computed to very high accuracy. For our experiments we chose \(n=2\).
We compare the 3-stage and 5-stage Radau IIA methods (see [16] for their definitions). These methods have stage orders 3 and 5 respectively and both are stiffly accurate and satisfy Assumption 2.III. We therefore expect convergence rates for the density \(\lambda \) of order 3.5 and 5.5. Since the exact solution is not available, we compute the difference to an approximation with step size k/4 and use this as an approximation to the discretization error. The results can be seen in Fig. 3. We observe that the results are in good agreement with our predictions.
Convergence for the density \({\widetilde{\lambda }}\) for the heat conduction problem (cf. Sect. 8.5), comparing Radau IIA methods
Change history
10 February 2021
A Correction to this paper has been published: https://doi.org/10.1007/s42985-020-00060-w
References
Alonso-Mallo, I., Palencia, C.: Optimal orders of convergence for Runge–Kutta methods and linear, initial boundary value problems. Appl. Numer. Math. 44(1–2), 1–19 (2003)
Banjai, L., Lubich, Ch., Melenk, J.M.: Runge–Kutta convolution quadrature for operators arising in wave propagation. Numer. Math. 119(1), 1–20 (2011)
Banjai, L., Laliena, A.R., Sayas, F.-J.: Fully discrete Kirchhoff formulas with CQ-BEM. IMA J. Numer. Anal. 35(2), 859–884 (2015)
Banjai, L., Rieder, A.: Convolution quadrature for the wave equation with a nonlinear impedance boundary condition. Math. Comput. 87(312), 1783–1819 (2018)
Bergh, J., Löfström, J.: Interpolation Spaces. An Introduction. Grundlehren der Mathematischen Wissenschaften, vol. 223. Springer, Berlin, New York (1976)
Bramble, J.H., Scott, R.: Simultaneous approximation in scales of Banach spaces. Math. Comput. 32(144), 947–954 (1978)
Banjai, L., Schanz, M.: Wave propagation problems treated with convolution quadrature and BEM. In: Schanz, M. (ed.) Fast Boundary Element Methods in Engineering and Industrial Applications: Lecture Notes in Applied and Computational Mechanics, vol. 63, pp. 145–184. Springer, Heidelberg (2012)
Brenner, Ph, Thomée, V.: On rational approximations of semigroups. SIAM J. Numer. Anal. 16(4), 683–694 (1979)
Crouzeix, M.: Sur les méthodes de Runge-Kutta pour l’approximation des problèmes d’évolution. In: Computing methods in applied sciences and engineering (Second international symposium, Versailles, 1975), Part 1. Lecture notes in Econom. and Math. systems, Vol. 134. Springer, Berlin, pp. 206–223 (1976)
Chandler-Wilde, S.N., Graham, I.G., Langdon, S., Spence, E.A.: Numerical-asymptotic boundary integral methods in high-frequency acoustic scattering. Acta Numer. 21, 89–305 (2012)
Engel, K.-J., Nagel, R.: A Short Course on Operator Semigroups. Springer, New York (2006)
Henrici, P.: Applied and Computational Complex Analysis: Power Series—Integration—Conformal Mapping—Location of Zeros, vol. 1. Wiley, New York (1988). (Reprint of the 1974 original)
Hassell, M.E., Qiu, T., Sánchez-Vizuet, T., Sayas, F.-J.: A new and improved analysis of the time domain boundary integral operators for the acoustic wave equation. J. Integral Equ. Appl. 29(1), 107–136 (2017)
Hassell, M., Sayas, F.-J.: Convolution quadrature for wave simulations. Numerical Simulation in Physics and Engineering: SEMA SIMAI, vol. 9, pp. 71–159. Springer, Cham (2016)
Hassell, M.E., Sayas, F.-J.: A fully discrete BEM-FEM scheme for transient acoustic waves. Comput. Methods Appl. Mech. Eng. 309, 106–130 (2016)
Hairer, E., Wanner, G.: Solving ordinary differential equations: II. Springer Series in Computational Mathematics: Stiff and differential-algebraic problems, vol. 14, 2nd edn. Springer, Berlin (2010)
Jerison, D.S., Kenig, C.E.: The Neumann problem on Lipschitz domains. Bull. Am. Math. Soc. 4(2), 203–207 (1981)
Li, J., Melenk, J.M., Wohlmuth, B., Zou, J.: Optimal a priori estimates for higher order finite elements for elliptic interface problems. Appl. Numer. Math. 60(1–2), 19–37 (2010)
Lubich, Ch., Ostermann, A.: Runge–Kutta methods for parabolic equations and convolution quadrature. Math. Comp. 60(201), 105–131 (1993)
Laliena, A.R., Sayas, F.-J.: Theoretical aspects of the application of convolution quadrature to scattering of acoustic waves. Numer. Math. 112(4), 637–678 (2009)
Lubich, Ch.: On the multistep time discretization of linear initial-boundary value problems and their boundary integral equations. Numer. Math. 67(3), 365–389 (1994)
McLean, W.: Strongly Elliptic Systems and Boundary Integral Equations. Cambridge University Press, Cambridge (2000)
Melenk, J.M., Rieder, A.: Runge–Kutta convolution quadrature and FEM-BEM coupling for the time-dependent linear Schrödinger equation. J. Integral Equ. Appl. 29(1), 189–250 (2017)
Melenk, J.M., Rieder, A.: On superconvergence of Runge-Kutta convolution quadrature for the wave equation. (2019). arXiv:1904.00847 (in press in Numer. Math.)
Ostermann, A., Roche, M.: Runge–Kutta methods for partial differential equations and fractional orders of convergence. Math. Comp. 59(200), 403–420 (1992)
Pazy, A.: Semigroups of Linear Operators and Applications to Partial Differential Equations. Applied Mathematical Sciences, vol. 44. Springer, New York (1983)
Qiu, T., Rieder, A., Sayas, F.-J., Zhang, S.: Time-domain boundary integral equation modeling of heat transmission problems. Numer. Math. 143(1), 223–259 (2019)
Rieder, A.: Convolution quadrature and boundary element methods in wave propagtion : a time domain point of view. PhD thesis, Technische Universität Wien (2017)
Rieder, A., Sayas, F.-J., Melenk, J. M.: Time domain boundary integral equations and convolution quadrature for scattering by composite media (2020). arXiv:2010.14162
Riesz, F., Nagy, S.-B.: Functional analysis. Dover Books on Advanced Mathematics. Dover Publications Inc, New York (1990)
Sayas, F.-J.: Retarded potentials and time domain boundary integral equations. A Road Map Springer Series in Computational Mathematics, vol. 50. Springer, Cham (2016)
Sauter, S.A., Schwab, Ch.: Boundary element methods. Springer Series in Computational Mathematics: Translated and Expanded from the 2004 German Original, vol. 39. Springer, Berlin (2011)
Steinbach, O.: Numerical Approximation Methods for Elliptic Boundary Value Problems: Finite and Boundary elements, Translated from the 2003 German Original. Springer, New York (2008)
Sauter, S., Veit, A.: Retarded boundary integral equations on the sphere: exact and numerical solution. IMA J. Numer. Anal. 34(2), 675–699 (2014)
Tartar, L.: An Introduction to Sobolev Spaces and Interpolation spaces: Lecture Notes of the Unione Matematica Italiana, vol. 3. Springer, Berlin (2007)
Triebel, H.: Interpolation theory, function spaces, differential operators, 2nd edn. Johann Ambrosius Barth, Heidelberg (1995)
von Neumann, J.: Eine Spektraltheorie für allgemeine Operatoren eines unitären Raumes. Math. Nachr. 4, 258–281 (1951)
Acknowledgements
Work of A.R. was financially supported by the Austrian Science Fund (FWF) through the doctoral school “Dissipation and Dispersion in Nonlinear PDEs” (project W1245) and the research program “Taming complexity in partial differential systems” (Grant SFB F65). FJS is partially supported by NSF-DMS grant 1818867. Part of this work was developed while FJS was a Visiting Professor at TU Wien.
Funding
Open access funding provided by Austrian Science Fund (FWF).. Open access funding provided by Austrian Science Fund (FWF).
Author information
Authors and Affiliations
Corresponding author
Additional information
This article is part of the topical collection “Waves 2019 – invited papers” edited by Manfred Kaltenbacher and Jens Markus Melenk.
A Interpolation of Sobolev spaces
A Interpolation of Sobolev spaces
In this appendix we prove that in Lipschitz domains and for certain parameters \(\mu \), the spaces \([L^2(\varOmega ),H^1_0(\varOmega )]_{\mu }\) contain functions with non-vanishing boundary conditions. Such estimates are the main ingredient when determining the convergence rate of Runge–Kutta methods using the theory developed in the previous sections. For \(\mu <1/2\), it is well-known that the fractional Sobolev spaces \(H^\mu (\varOmega )=[L^2(\varOmega ),H^1(\varOmega )]_{\mu ,2}\) and \(\widetilde{H}^{\mu }(\varOmega )=[L^2(\varOmega ),H^1_0(\varOmega )]_{\mu ,2}\) coincide (see e.g. [22, Theorem 3.40] together with the results in [22, Appendix B] to identify the Sobolev spaces with the interpolation space). We prove that when interpolating using the index \(\infty \), the critical value \(\mu =1/2\) is also admissible, provided that some further regularity is provided.
In order to state our result, we need additional notation, notably we define interpolation spaces for \(q\in [1,\infty )\) as
and introduce the following Besov spaces:
For \(t >0\), we define the strip
which will play an important role in the following proofs.
Theorem A.1
Let \(\varOmega \) be either a bounded Lipschitz domain or the complement of a bounded Lipschitz domain. Fix \(\mu \in (0,1/2]\). Then
with equivalent norms. The implied constant depends on \(\varOmega \) and \(\mu \).
Proof
For simplicity, assume that \(\varOmega \) is bounded. We focus on the case \(\mu =1/2\), since the case \(0 \le \mu < 1/2\) follows from \(H^\mu (\varOmega ) = {\widetilde{H}}^\mu (\varOmega )\) (see [22, Theorem 3.40(i) and Theorem 3.33]) combined with the embeddings
Consider \(u \in B^{1/2}_{2,1}(\varOmega )\). For each \(t>0\), we select \(v(t) \in H^1(\varOmega )\) almost realizing the infimum appearing in the interpolation norm, i.e.,
By [6, Lemma], the following estimate holds for all \(t> 0\):
We also note that
We consider a smooth cutoff function \(\chi _t: \varOmega \rightarrow [0,1]\) satisfying:
We then define \(\widetilde{v}(t):=\chi _t v(t) \in H_0^1(\varOmega )\) and calculate:
where we used the fact that \(1-\chi _t\) vanishes on \(\varOmega \setminus \varOmega _{2t}\) and applied [18, Lemma 2.1] to estimate the \(L^2\)-norm there.
Similarly,
For the interpolation norm, we therefore get
If \(\varOmega \) is the exterior of a bounded Lipschitz domain, the proof applies almost verbatim as all important steps can be localized to a neighborhood of the boundary. \(\square \)
Remark A.2
The use of the second parameter \(\infty \) in the interpolation norm is crucial for Theorem A.1 to hold in the case \(\mu =1/2\). For \(L^2\)-based interpolation it is well-known that the interpolation space \(\left[ L^2(\varOmega ), H^1_0(\varOmega )\right] _{1/2,2}\) is the Lions-Magenes space \(H^{1/2}_{00}(\varOmega )\), see [35, Chapter 33], which is distinct from \(H^{1/2}(\varOmega )\).
When considering the Neumann problem in Sect. 8.3, we need to devise a lifting to a vector field with a given normal jump in \(L^2\). In general, such liftings do not have \(B^{1/2}_{2,1}\)-regularity. Thus Theorem A.1 is not applicable. Instead, we have a modified construction.
Lemma A.3
Let \(\varOmega \) be a bounded Lipschitz domain or the exterior of a bounded Lipschitz domain with boundary \(\varGamma :=\partial \varOmega \). For \(C > 0\), \(c>0\) fixed with c sufficiently small, define the non-tangential maximal function
Let \(u \in H^1(\varOmega )\) be harmonic and satisfy \(N(\nabla u) \in L^2(\varGamma )\).
Then for \(t>0\) we can bound the \(L^2\) norm on strips \(\varOmega _t\) by
Proof
We focus on a single chart in the parametrization of (a vicinity of) \(\varGamma \). Let \({\mathcal {O}} \subseteq \varOmega \) and \({\mathcal {D}} \subseteq {\mathbb {R}}^{d-1}\) be open, \({\varvec{r}} \in {\mathbb {R}}^n\), \(\varphi : {\mathcal {D}}\rightarrow {\mathbb {R}}\), \(y_0: {\mathcal {D}} \rightarrow {\mathbb {R}}\) such that we can write
By the Lipschitz assumption, we note that \(y_0(x) \lesssim C t\). Following the considerations in [10, Appendix A.4], one can see that as long as C in the definition of \(\varTheta \) is taken sufficiently large, we have that for all \(x \in {\mathcal {D}}\)
We calculate
Repeating the same calculation for all boxes needed to parametrize a neighborhood of \(\varGamma \) then concludes the proof. \(\square \)
Theorem A.4
Let \(\varOmega \subset {\mathbb {R}}^d \) be a bounded Lipschitz domain or the exterior of a bounded Lipschitz domain and write \({\mathbf {H}}_0(\mathrm {div},\varOmega ):=\{ {\varvec{w}} \in {\mathbf {H}}(\mathrm {div},\varOmega ): \gamma _{\nu } {\varvec{w}} =0\}\).
For every \(g \in L^2(\varOmega )\), there exists a function \({\varvec{w}} \in {\mathbf {H}}(\mathrm {div},\varOmega )\) such that
with an implied constant depending only on \(\varOmega \).
Proof
For simplicity, assume that \(\varOmega \) is bounded. By performing an appropriate cutoff away from \(\partial \varOmega \), all arguments can be localized.
Step 1: Consider the case \(\int _{\varGamma } g =0\). Let u be the solution of the Neumann problem
In addition to \(u \in H^1(\varOmega )\), by [17] (see also [10, Theorem A.6]), this harmonic function u also satisfies
For fixed \(t>0\) we again select a smooth cutoff function \(\chi _t\) satisfying (A.4). We set \({\varvec{w}}:=\nabla u\) and calculate using Lemma A.3:
Step 2: In the case \(\int _{\varGamma } g \ne 0\), the harmonic Neumann problem does not have a solution. Instead, we define u as the solution to
and again set \({\varvec{w}}:=\nabla u\). By construction we have \(\nabla \cdot {\varvec{w}}=-u \in H^1(\varOmega )\).
We decompose \(u=u_{0} + u_1\), where \(u_0\) solves the full-space problem
where u was extended by 0 outside of \(\varOmega \). As \(u \in L^2({\mathbb {R}}^d)\), standard regularity theory gives \(u_0 \in H^2(B)\) on any ball B and in particular \(u_0 \in H^2(\varOmega )\). In turn this yields \(\partial _{\nu } u_0 \in L^2(\partial \varOmega )\). By construction \(u_1\) then solves
As g, \(\partial _{\nu } u_0 \in L^2(\partial \varOmega )\) and \(\int _\varGamma g - \partial _\nu u_0 = 0\), we can apply Step 1 to get
Since \(\nabla u_0 \in (H^1(\varOmega ))^d \subseteq (B^{1/2}_{2,1}(\varOmega ))^d\) we can apply Theorem A.1 to get
to conclude the proof of (A.7). \(\square \)
The following lemma appears to be known in the community, see e.g. [5, Section 3.13, Exercise 4], but in order to be able to rigorously cite, we provide a short proof.
Lemma A.5
Let \(X:=(X_1, \dots , X_N)\) and \(Y:=(Y_1,\dots , Y_N)\), where \(X_j, Y_j\) are Banach spaces with continuous embedding \(Y_j \subset X_j\), and the product space carries any \(l^{p}\)-norm. Fix \(q \in [1,\infty ]\) and \(\theta \in (0,1). \) Then, the product of the interpolation spaces coincides with the interpolation of the product spaces. Namely, the following estimate holds for all \(x:=(x_1,\dots ,x_N) \in [X,Y]_{\theta ,q}\):
Proof
For \(j \in \{1,\dots ,N\}\), consider the operators \(S_j\) and \(T_j\) defined as
It is easy to see using the interpolation estimate (2.6) that
We therefore calculate:
For the opposite direction, we observe that
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article's Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article's Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Rieder, A., Sayas, FJ. & Melenk, J.M. Runge–Kutta approximation for \(C_0\)-semigroups in the graph norm with applications to time domain boundary integral equations. SN Partial Differ. Equ. Appl. 1, 49 (2020). https://doi.org/10.1007/s42985-020-00051-x
Received:
Accepted:
Published:
DOI: https://doi.org/10.1007/s42985-020-00051-x