Abstract
We prove large-scale \(C^\infty \) regularity for solutions of nonlinear elliptic equations with random coefficients, thereby obtaining a version of the statement of Hilbert’s 19th problem in the context of homogenization. The analysis proceeds by iteratively improving three statements together: (i) the regularity of the homogenized Lagrangian \(\overline{L}\), (ii) the commutation of higher-order linearization and homogenization, and (iii) large-scale \(C^{0,1}\)-type regularity for higher-order linearization errors. We consequently obtain a quantitative estimate on the scaling of linearization errors, a Liouville-type theorem describing the polynomially-growing solutions of the system of higher-order linearized equations, and an explicit (heterogenous analogue of the) Taylor series for an arbitrary solution of the nonlinear equations—with the remainder term optimally controlled. These results give a complete generalization to the nonlinear setting of the large-scale regularity theory in homogenization for linear elliptic equations.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
1.1 Motivation: Quantitative Homogenization for Nonlinear Equations
This article is concerned with nonlinear, divergence-form, uniformly elliptic equations of the form
The Lagrangian L(p, x) is assumed to be uniformly convex and regular in p and possess some mild regularity in x. Furthermore, L is a stochastic object: it is sampled by a probability measure \(\mathbb {P}\) which is statistically stationary and satisfies a unit range of dependence. This essentially means that \(x\mapsto L(\cdot ,x)\) is a random field, valued in the space of uniformly convex functions, the law of which is independent of x (or, to be precise, periodic in x; see Section 1.4 for the assumptions).
The objective is to describe the statistical behavior of the solutions of (1.1), with respect to the probability measure \(\mathbb {P}\), on large length scales. In other words, we want to understand what the solution u looks like in the case that the “macroscopic” domain U is very large relative to “microscopic” scale, which is the correlation length scale of the coefficients (taken to the unit scale).
At a qualitative level, a satisfactory characterization of the solutions, in the regime in which the ratio of these two length scales is large, is provided by the principle of homogenization. First proved in this context by Dal Maso and Modica [11, 12], it asserts roughly that a solution of (1.1) is, with probability approaching one, close in \(L^2\) (relative to its size in \(L^2\)) to a solution of a deterministic equation of the form
for an effective Lagrangian \(\overline{L}\) which is also uniformly convex and \(C^{1,1}\).
This result is of great importance, from both the theoretical and computation points of view, since the complexity of the homogenized equation (1.2) is significantly less than that of (1.1) as it is both deterministic and spatially homogeneous. It hints that the structure of (1.1) should, on large domains and with high probability, possess some of the structure of a constant coefficient equation and thus we may expect it to be more amenable to our analysis than the worst-possible heterogeneous equation of the form (1.1). In other words, since the Lagrangian L is sampled by a probability measure \(\mathbb {P}\) with nice ergodic properties, rather than given to us by the devil, can we expect its solutions to have a nicer structure? In order to answer this kind of question, we need to build a quantitative theory of homogenization.
To be of practical use, the principle of homogenization must be made quantitative. We need to have answers to questions such as:
How large does the ratio of scale separation need to be before we can be reasonably sure that solutions of (1.1) are close to those of (1.2)? In other words, what is the size of a typical error in the homogenization approximation in terms of the size of U?
Can we estimate the probability of the unlikely event that the error is large?
What is \(D_p\overline{L}\) and how can we efficiently compute it? How regular can we expect it to be? Can we efficiently compute its derivatives?
Can we describe the fluctuations of the solutions?
In this paper we show that (1.1) has a \(C^\infty \) structure. In particular, we will essentially answer the third question posed above by demonstrating that the effective Lagrangian \(\overline{L}\) is as regular as \(L(\cdot ,x)\) with estimates for its derivatives. We will identify the higher derivatives of \(\overline{L}\) as the homogenized coefficients of certain linearized equations and give quantitative homogenization estimates for these, implicitly indicating a computational method for approximating them and thus a Taylor approximation for \(\overline{L}\). Finally, we will prove large-scale \(C^{k,1}\) type estimates for solutions of (1.1), for \(k\in \mathbb {N}\) as large as can be expected from the regularity assumptions on L, a result analogous to Hilbert’s 19th problem, famously given for spatially homogeneous Lagrangians by De Giorgi and Nash. For this classical result, the main roadblock was in obtaining the continuity of solutions to linear equations, as it is completely straightforward to differentiate the equation apply the Schauder theory, which was previously known. In the case of homogenization, the situation is reversed, as it is less clear how one should “differentiate the equation” since literally doing so would produce negative powers of \(\varepsilon \), even if the coefficients were smooth. Our analysis resolves this difficulty and reveals the interplay between these three seemingly different kinds of results: (i) the regularity of \(\overline{L}\), (ii) the homogenization of linearized equations and the commutability of homogenization and linearization, and (iii) the large-scale regularity of the solutions. These three statements must be proved together, iteratively in the parameter \(k\in \mathbb {N}\) which represents the degree of regularity of \(\overline{L}\), the order of the linearized equation, and the order of the \(C^{k,1}\) estimate.
1.2 Background: Large-Scale Regularity Theory and Its Crucial Role in Quantitative Homogenization
In the last decade, beginning with the work of Gloria and Otto [16, 18], a quantitative theory of homogenization has been developed to give precise answers to questions like the ones stated in the previous subsection. Until now, most of the progress has come in the case of linear equations
which corresponds to the special case \(L(p,x) = \frac{1}{2}p\cdot \mathbf {a}(x)p\) of (1.1), where \(\mathbf {a}(x)\) is a symmetric matrix. By now there is an essentially complete quantitative theory for linear equations, and we refer to the monograph [4] and the references therein for a comprehensive presentation of this theory. Quantitative homogenization for the nonlinear equation (1.1) has a comparatively sparse literature; in fact, the only such results of which we are aware are those of [5, 6] (see also [4, Chapter 11]), our previous paper [1] and a new paper of Fischer and Neukamm [14] which was posted to arXiv as we were finishing the present article.
Quantitative homogenization is inextractably linked to regularity estimates on the solutions of the heterogeneous equation. This is not surprising when we consider that the homogenized flux \(D_p\overline{L}(\nabla u_{\mathrm{hom}})\) should be related to the spatial average (say, on some mesoscopic scale) of the heterogeneous flux \(D_pL(\nabla u(x),x)\). In order for spatial averages of the latter to converge nicely, we need to have bounds. It could be unfortunate and lead to a very slow rate of homogenization if, for instance, the flux was concentrated on sets of very small measure which percolate only on very large scales. To rule this out we need much better estimates: ideally, we would like to know that the size of the flux on small scales is the same as on large scales, which amounts to a \(W^{1,\infty }\) estimate on solutions.
Unfortunately, solutions of equations with highly oscillating coefficients do not possess very strong regularity, in general. Indeed, the best deterministic elliptic regularity estimate for solutions of (1.1), which does not degenerate as the size of the domain U becomes large, is \(C^{0,\delta }\) in terms of Hölder regularity (the De Giorgi–Nash estimate) and \(W^{1,2+\delta }\) in terms of Sobolev regularity (the Meyers estimate). The tiny exponent \(\delta >0\) in each estimate becomes small as the ellipticity ratio becomes large (see [4, Example 3.1]) and thus both estimates are far short of the desired regularity class \(W^{1,\infty } = C^{0,1}\).
One of the main insights in the quantitative theory of homogenization is that, compared to a generic (“worst-case”) L, solutions of the equation (1.1) have much better regularity if L is sampled by an ergodic probability measure \(\mathbb {P}\). This is an effect of homogenization itself: on large scales, (1.1) should be a “small perturbation” of (1.2), and therefore better regularity estimates for the former can be inherited from the latter. This is the same idea used to prove the classical Schauder estimates. In the context of homogenization, the result states that there exists a random variable \(\mathcal {X}\) (sometimes called the minimal scale) which is finite almost surely such that, for every \(\mathcal {X}< r < \frac{1}{2}R\) and solution \(u\in H^1(B_R)\) of (1.1) with \(U=B_R\), we have the estimate
Here C depends only on dimension and ellipticity and denotes the mean of a function in U. If we could send \(r\rightarrow 0\) in (1.4), it would imply that
which is a true Lipschitz estimate, same estimate in fact as holds for the homogenized equation (1.2). As (1.4) is valid only for \(r>\mathcal {X}\), it is sometimes called a “large-scale \(C^{0,1}\) estimate” or a “Lipschitz estimate down to the microscopic scale”. This estimate, first demonstrated in [6] in the stochastic setting, is a generalization of the celebrated result in the case of (non-random) periodic coefficients due to Avellaneda and Lin [7]. Of course, it then becomes very important to quantify the size of \(\mathcal {X}\). The estimate proved in [6], which is essentially optimal, states that \(\mathcal {X}\) is bounded up to “almost volume-order large deviations”: for every \(s<d\) and \(r\geqq 1\),
Here the constant C depends only on s, d, and the ellipticity. A proof of this large-scale regularity estimate together with (1.5) can be found in [4, Chapter 3] in the linear case and in [4, Chapter 11] for the nonlinear case. The right side of (1.5) represents the probability of the unlikely event that the L sampled by \(\mathbb {P}\) will be a “worst-case” L in the ball of radius r. A proof of the optimality of (1.5) can be found in [4, Section 3.6].
This large-scale regularity theory introduced in [6] was further developed in the case of (1.3) in [2, 3, 5, 15, 17] and now plays an essential role in the quantitative theory of stochastic homogenization. Whether one employs functional inequalities [13, 17] or renormalization arguments [2, 3, 19], it is a crucial ingredient in the proof of the optimal error estimates in homogenization for (1.3): see the monograph [4] and the references therein for a complete presentation of these developments.
The large-scale \(C^{0,1}\) estimate is, from one point of view, the best regularity one can expect solutions of (1.1) or (1.3) to satisfy: since the coefficients are rapidly oscillating, there is no hope for the gradient to exhibit continuity on the macroscopic scale. However, as previously shown in the periodic case in [7, 9], the solutions of the linear equation (1.3) still have a \(C^\infty \) structure. To state what we mean, let us first think of an (interior) \(C^{k,1}\) estimate not as a pointwise bound on the \((k+1)\)th derivatives of a function, but as an estimate on how well a function may be approximated on small balls by a kth order polynomial. By Taylor’s theorem, these are of course equivalent in the following sense:
where \(\mathcal {P}_k\) denotes the set of polynomials of degree at most k and and we use the notation to denote the volume-normalized \(L^2(U)\) norm. Thus the interior \(C^{k,1}\) estimate for a harmonic function can be stated in the form: for any harmonic function u in \(B_R\) and any \(r\in \left( 0,\tfrac{1}{2}R\right] \),
Moreover, the infimum on the left side may be replaced by the set of harmonic polynomials of degree at most k.
As we cannot expect a solution of (1.1) to have regularity beyond \(C^{0,1}\) in a classical (pointwise) sense, in order to make sense of a \(C^{k,1}\) estimate for the heterogeneous equation (1.3) we need to replace the set of polynomials by a heterogeneous analogue. The classical Liouville theorem says that the set of harmonic functions which grow like \(o(|x|^{k+1})\) is just the set of harmonic polynomials of degree at most k. This suggests that we should use the (random) vector space
We think of these as “\(\mathbf {a}(x)\)-harmonic polynomials.” It turns out that one can prove that, \(\mathbb {P}\)-almost surely, this set is finite dimensional and has the same dimension as the set of at most kth order harmonic polynomials. In fact, one can match any \(\mathbf {a}(x)\)-harmonic polynomial to an \({\overline{\mathbf {a}}}\)-harmonic polynomial in the highest degree, and vice versa. In close analogy to (1.6), the statement of large-scale \(C^{k,1}\) regularity is then as follows: there exists a minimal scale \(\mathcal {X}\) satisfying (1.5) such that, for any \(R > 2\mathcal {X}\) and any solution \(u\in H^1(B_R)\) of \(-\nabla \cdot \mathbf {a}\nabla u =0\), we can find \(\phi \in \mathcal {A}_k\) such that, for every \(r\in \left[ \mathcal {X},\tfrac{1}{2} R \right] \),
see [4, Theorem 3.8] for the full statement, which was first proved in the periodic setting by Avellaneda and Lin [8]. Subsequent versions of this result, which are based on the ideas of [7, 8] in their more quantitative formulation given in [6], were proved in various works [2, 15, 17], with the full statement here given in [3, 10].
In all of its various forms, higher regularity in stochastic homogenzations is based on the simple idea that solutions of the heterogeneous equation should be close to those of the homogenized equation, which should have much better regularity. In the case of the linear equation (1.3), it does not a large leap, technically or philosophically, to go from (1.4) to (1.7). Indeed, to gain control over higher derivatives, one just needs to differentiate the equation (not in the microscopic parameters, of course, but in macroscopic ones) and, luckily, since the equation is linear, this does not change the equation. Roughly speaking, the idea is analogous to bootstrapping the regularity of a constant-coefficient, linear equation by differentiating it. Therefore the estimate (1.7) is perhaps not too surprising once one has the large-scale \(C^{0,1}\) estimate in hand.
1.3 Summary of the Results Proved in this Paper
The situation is very different in the nonlinear case. When one differentiates the equation (again, in a macroscopic parameter) we get a new equation, namely the first-order linearized equation. If we want to apply a large-scale regularity result to this equation, we must first (quantitatively) homogenize it! Achieving higher-order regularity estimates requires repeatedly differentiating the equation, which lead to a hierarchy of linearized equations requiring homogenization estimates.
Let us be a bit more explicit. The gradient of the homogenized Lagrangian \(\overline{L}\) is given by the well-known formula
where \(\phi _p\) is the first-order corrector with slope \(p\in {\mathbb {R}^{d}}\), that is, it satisfies
The limit in the second line of (1.8) is to be understood in a \(\mathbb {P}\)-almost sure sense, and it is a consequence of the ergodic theorem, which states that macroscopic averages of stationary fields must converge to their expectations. The formula (1.8) says that \(D_p\overline{L}(p)\) is the flux per unit volume of the first-order corrector with slope \(p\in {\mathbb {R}^{d}}\). It naturally arises when we homogenize the nonlinear equation. We can try to show that \(\overline{L}\in C^2\) by formally differentiating (1.8), which leads to the expression
If we define the linearized coefficients around \(\ell _p + \phi _p\) by
and put \(\psi ^{(1)}_{p,e_i}:= \partial _{p_i}\phi _p(x)\), then we see that \(\psi ^{(1)}_{p,e_i}\) is the first-order corrector with slope \(e_i\) of the linear equation with coefficients \(\mathbf {a}_p\):
We call \(\psi ^{(1)}_{p,e}\) a first-order linearized corrector. Moreover, we have the formula
That is, “linearization and homogenization commute”: the Hessian of \(\overline{L}\) at p should be equal to the homogenized coefficients \({\overline{\mathbf {a}}}_p\) corresponding to the linear equation with coefficient field \(\mathbf {a}_p=D^2_pL(p+\nabla \phi _p(\cdot ),\cdot )\). This reasoning is only formal, but since the right side of the above formula for \(D^2_p\overline{L}\) is well-defined (and only needs qualitative homogenization), we should expect that it should be rather easy to confirm it rigorously. Moreover, while quantitative homogenization of the original nonlinear equation gives us a \(C^{0,1}\) estimate, we should expect that quantitative homogenization of the linearized equation should give us a \(C^{0,1}\) estimate for differences of solutions and a large-scale \(C^{1,1}\) estimate for solutions. This is indeed the case and was proved in our previous paper [1], where we were motivated by the goal of obtaining this regularity estimate for differences of solutions in anticipation of its important role in the proof of optimal quantitative homogenization estimates. Indeed, in the very recent preprint [14], Fischer and Neukamm showed that this estimate can be combined with spectral gap-type assumptions on the probability measure to obtain quantitative bounds on the first-order correctors which are optimal in the scaling of the error.
We may attempt to differentiate the formula for the homogenized Lagrangian a second time, with the ambition of obtaining a \(C^3\) estimate for \(\overline{L}\), a \(C^{2,1}\) estimate for solutions and a higher-order improvement of our \(C^{0,1}\) estimate for differences (which will be a \(C^{0,1}\) estimate for linearization errors): we get
If we define the vector field
then we see that \(\psi ^{(2)}_{p,e_i,e_j}\) is the first-order corrector with slope zero of the linear equation
and the formula for the tensor \(D^3_p\overline{L}\) becomes
the corresponding homogenized coefficient. Unlike the case for the Hessian of \(\overline{L}\), we should not expect this formula to be valid under qualitative ergodic assumptions! Indeed, the qualitative homogenization of (1.9) and thus the validity of (1.10) requires that \(\mathbf {F}_{2,p,e_i,e_j}\) belong to \(L^2\), in the sense that \(\mathbb {E}\left[ \int _{[0,1]^d} \left| \mathbf {F}_{2,p,e_i,e_j} (x) \right| ^2\,\mathrm{d}x \right] < \infty \), and due to the product of two first-order correctors we only have \(L^1\)-type integrability for \(\mathbf {F}_{2,p,e_i,e_j}\).Footnote 1 This is a serious problem which can only be fixed using the large-scale regularity theory for the first-order linearized equation, and a suitable bound on the minimal scale \(\mathcal {X}\), thereby obtaining a bound in \(L^\infty ({{\square }}_0)\) and hence \(L^4({{\square }}_0)\) for \(\nabla \psi ^{(1)}_{p,e_i}\), with at least a fourth moment in expectation. Note that this also requires some regularity of the Lagrangian L on the smallest scale.
If we differentiate the equation once more in an effort to prove that \(\overline{L}\in C^4\), we will be faced with similar difficulties, this time with the more complicated vector field
Notice that the last term has three factors of the first-order linearized correctors instead of two, and is thus “even further” from being obviously \(L^2\) than was \(\mathbf {F}^{(2)}_{p,e_i,e_j}\). Homogenizing the third-order linearized equation will therefore require large-scale regularity estimates for both the first-order and second-order linearized equations, and one can now see the situation will get worse as the order increases beyond three. Moreover, proving quantitative homogenization for these equations will also require some smoothness of the homogenized coefficients associated to the lower-order equations, due to the needs for the homogenized solutions to be smooth in quantitative two-scale expansion arguments.
This suggests a bootstrap argument for progressively and simultaneously obtaining (i) the smoothness of \(\overline{L}\); (ii) the higher-order large-scale regularity of solutions (and solutions of linearized equations); and (iii) the homogenization of the higher-order linearized equations and commutation of homogenization and higher-order linearization. The point of this paper is to formalize this idea and thereby give a proof of “Hilbert’s 19th problem for homogenization”. Here is a rough schematic of the argument, as discussed above, which comes in three distinct steps, discussed in more detail below:
Homogenization & large-scale \(C^{0,1}\) regularity for the linearized equations up to order \(\mathsf {N}\)\(\implies \)\(\overline{L} \in C^{2+\mathsf {N}}\).
\(\overline{L} \in C^{2+\mathsf {N}}\) and large-scale \(C^{0,1}\) regularity for the linearized equations up to order \(\mathsf {N}\)\(\implies \) homogenization for the linearized equations up to order \(\mathsf {N}+1\).
\(\overline{L} \in C^{2+\mathsf {N}}\), large-scale \(C^{0,1}\) regularity for the linearized equations up to order \(\mathsf {N}\) and homogenization for the linearized equations up to order \(\mathsf {N}+1\)\(\implies \) large-scale \(C^{0,1}\) regularity for the linearized equations up to order \(\mathsf {N}+1\).
The three implications above are the focus of most of the paper and their proofs are given in Sections 2–5.
Once this induction argument is completed, we consequently obtain a full \(C^{k,1}\)-type large scale regularity estimate for solutions of the original nonlinear equation, generalizing (1.7). The main question becomes what the replacement for \(\mathcal {A}_k\) should be, that is, what the “polynomials” should be. We show that these are certain solutions of the system of linearized equations (linearized around a first-order corrector) exhibiting polynomial-type growth, which we classify by providing a Liouville-type result which is part of the statement of the theorem (see the discussion between the statements of Theorems 1.5 and 1.6, below, for a definition of these spaces, which are denoted by \(\mathsf {W}_n^p\)). The resulting theorem we obtain, which is a version of the statement of Hilbert’s 19th problem in the context of homogenization, provides a very precise description of the solutions of (1.1) in terms of the first-order correctors and the correctors of a hierarchy of linear equations.
We also obtain, as a corollary, the improvement of the scaling of linearization errors—which is closely related to the regularity of solutions. To motivate this result, suppose we are given two solutions \(u,v\in H^1(B_R)\) of (1.1) in a large ball (\(R\gg 1\)) which are close to each other in the sense that
Suppose that we attempt to approximate the difference \(u-v\) by the solution \(w\in H^1(B_{R/2})\) of the linearized problem
Then we may ask the question of how small we should expect the first-order linearization error to be. The best answer that we have from deterministic elliptic regularity estimates is that there exists a small exponent \(\delta (d,\Lambda )>0\) such that
This can be easily proved for instance using the Meyers gradient \(L^{2+\delta }\) estimate, and it is sharp in the sense that it is not possible to do better than the very small exponent \(\delta \). We can say roughly that the space of solutions of (1.1) is a \(C^{1,\delta }\) manifold, but no better. Of course, if L does not depend on x, or if R is of order one and L is smooth in both variables (p, x), then one expects to have the estimate above with \(\delta =1\) and to be able to prove more precise estimates using higher-order linearized equations. In fact, this is essentially a reformulation of the statement of Hilbert’s 19th problem (indeed—see Appendix E, where we give a proof of Hilbert’s 19th problem in its classical formulation by following this line of reasoning).
In this paper we also prove a large-scale version of the quadratic response to first-order linearization in the context of homogenization, which states that (1.12) holds with \(\delta =1\) whenever R is larger than a random minimal scale. Moreover, we prove a full slate of higher-order versions of this result: see Corollary 1.4. These results roughly assert that, with probability one, the large-scale structure of solutions of (1.1) resembles that of a smooth manifold.
In the next two subsections, we state our assumptions and give the precise statements of the results discussed above.
1.4 Assumptions and Notation
In this subsection, we state the standing assumptions in force throughout the paper.
We fix following global parameters: the dimension \(d\in \mathbb {N}\) with \(d\geqq 2\), a constant \(\Lambda \in [1,\infty )\) measuring the ellipticity, an integer \(\mathsf {N}\in \mathbb {N}\) with \(\mathsf {N}\geqq 1\) measuring the smoothness of the Lagrangian, and constants \(\mathsf {M}_0, \mathsf {K}_0 \in [1,\infty )\). For short, we denote
This allows us to, for instance, denote constants C which depend on \((d,\Lambda ,\mathsf {N},\mathsf {M}_0,\mathsf {K}_0)\) by simply \(C(\mathrm {data})\) instead of \(C(d,\Lambda ,\mathsf {N},\mathsf {M}_0,\mathsf {K}_0)\).
The probability space is the set \(\Omega \) of all Lagrangians \(L:{\mathbb {R}^{d}}\times {\mathbb {R}^{d}}\rightarrow \mathbb {R}\), written as a function of \((p,x)\in {\mathbb {R}^{d}}\times {\mathbb {R}^{d}}\), satisfying the following conditions:
- (L1)
L is \(2+\mathsf {N}\) times differentiable in the variable p and, for every \(k\in \{2,\ldots ,2+\mathsf {N}\}\), the function \(D^{k}_pL\) is uniformly Lipschitz in both variables and satisfies
$$\begin{aligned} \left[ D^{k}_pL \right] _{C^{0,1}({\mathbb {R}^{d}}\times {\mathbb {R}^{d}})} \leqq \mathsf {K}_0. \end{aligned}$$(1.13)For \(k=1\), we assume that, for \(z \in \mathbb {R}^d\),
$$\begin{aligned} \left[ D_pL(z,\cdot ) \right] _{C^{0,1}({\mathbb {R}^{d}})} \leqq \mathsf {K}_0 (1+|z|). \end{aligned}$$(1.14) - (L2)
L is uniformly convex in the variable p: for every \(p\in {\mathbb {R}^{d}}\) and \(x\in {\mathbb {R}^{d}}\),
$$\begin{aligned} I_d \leqq D^2_p L(p,x) \leqq \Lambda I_d. \end{aligned}$$ - (L3)
\(D_pL(0,\cdot )\) is uniformly bounded:
$$\begin{aligned} \left\| D_pL(0,\cdot ) \right\| _{L^\infty ({\mathbb {R}^{d}})} \leqq \mathsf {M}_0. \end{aligned}$$
We define \(\Omega \) to be the set of all such Lagrangians L:
Note that \(\Omega \) depends on the fixed parameters \((d,\Lambda ,\mathsf {N},\mathsf {M_0},\mathsf {K}_0)\). It is endowed with the following family of \(\sigma \)-algebras: for each Borel subset \(U \subseteq {\mathbb {R}^{d}}\), define
The largest of these is denoted by \(\mathcal {F}:= \mathcal {F}({\mathbb {R}^{d}})\).
We assume that the law of the “canonical Lagrangian” L is a probability measure \(\mathbb {P}\) on \((\Omega ,\mathcal {F})\) satisfying the following two assumptions:
- (P1)
\(\mathbb {P}\) has a unit range of dependence: for all Borel subsets \(U,V\subseteq {\mathbb {R}^{d}}\) such that \({{\,\mathrm{dist}\,}}(U,V) \geqq 1\),
$$\begin{aligned} \mathcal {F}(U) \text{ and } \mathcal {F}(V) \text{ are } \mathbb {P}\text{-independent. } \end{aligned}$$ - (P2)
\(\mathbb {P}\) is stationary with respect to \(\mathbb {Z}^{d}\)-translations: for every \(z\in \mathbb {Z}^{d}\) and \(E\in \mathcal {F}\),
$$\begin{aligned} \mathbb {P}\left[ E \right] = \mathbb {P}\left[ T_z E \right] , \end{aligned}$$where the translation group \(\{T_z\}_{z\in \mathbb {Z}^{d}}\) acts on \(\Omega \) by \((T_zL)(p,x) = L(p,x+z)\).
The expectation with respect to \(\mathbb {P}\) is denoted by \(\mathbb {E}\).
Since we will be often concerned with measuring the stretched exponential moments of the random variables we encounter, the following notation is convenient: for every \(\sigma \in (0,\infty )\), \(\theta >0\), and random variable X on \(\Omega \), we write
This is essentially notation for an Orlicz norm on \((\Omega ,\mathbb {P})\). Some basic properties of this notation is given in [4, Appendix A].
1.5 Statement of the Main Results
We begin by introducing the higher-order linearized equations. These can be computed by hand, as we did for the second and third linearized equations in Section 1.3, but it is convenient to work with more compact formulas. Observe that, by Taylor’s formula with remainder, we have, for every \(n\in \{ 0,\ldots ,\mathsf {N}+1 \}\),
Define, for \(p,x,h_1,\ldots ,h_{\mathsf {N}} \in {\mathbb {R}^{d}}\) and \(t \in \mathbb {R}\),
Also define, for each \(m\in \{ 1,\ldots ,\mathsf {N}+1\}\) and \(p,x,h_1,\ldots ,h_{m-1} \in {\mathbb {R}^{d}}\),
Observe that \(\mathbf {F}_1\equiv 0\) by definition (that is, the right side of the first linearized equation is zero, as we have already seen).
Our first main result concerns the regularity of the effective Lagrangian \(\overline{L}\) and states that it has essentially the same regularity in p as we assumed for L.
Theorem 1.1
(Regularity of \(\overline{L}\)). For every \(\beta \in (0,1)\), the effective Lagrangian \(\overline{L}\) belongs to \(C^{2+\mathsf {N},\beta }_{\mathrm{loc}}({\mathbb {R}^{d}})\) and, for every \(\mathsf {M}\in [1,\infty )\), there exists \(C(\beta ,\mathsf {M},\mathrm {data})<\infty \) such that
In view of Theorem 1.1, we may introduce homogenized versions of the above functions. We define, for every \(p\in {\mathbb {R}^{d}}\), \(\{h_i\}_{i=1}^{\mathsf {N}} \subseteq \mathbb {R}^d\) and \(\delta \in \mathbb {R}\),
and then, for every \(m\in \{ 1,\ldots ,\mathsf {N}+1\}\) and \(\{h_i\}_{i=1}^{m-1} \subseteq \mathbb {R}^d\),
As above, we have that \(\overline{\mathbf {F}}_1\equiv 0\) by definition.
In the next theorem, we present a statement concerning the commutability of homogenization and higher-order linearizations. It generalizes [1, Theorem 1.1], which proved the result in the case \(\mathsf {N}=1\).
Theorem 1.2
(Homogenization of higher-order linearized equations).
Let \(n\in \{0,\ldots ,\mathsf {N}\}\), \(\delta \in \left( 0,\tfrac{1}{2}\right] \), \(\mathsf {M}\in [1,\infty )\), and \(U_1,\ldots ,U_{n+1} \subseteq {\mathbb {R}^{d}}\) be a sequence of bounded Lipschitz domains satisfying
There exist \(\sigma (\mathrm {data})>0\), \(\alpha \left( \{U_m\},\delta ,\mathrm {data}\right) >0\), \(C\left( \{U_m\},\mathsf {M},\delta ,\mathrm {data}\right) <\infty \), and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid. Let \(\varepsilon \in (0,1]\), \(f \in W^{1,2+\delta }(U_0)\) be such that
and, for each \(m\in \{1,\ldots , n+1\}\), fix \(g_m\in W^{1,2+\delta }(U_m)\) and let \(u^\varepsilon \in H^1(U_1)\) and the functions \(w_1^\varepsilon \in H^1(U_1),w_2^\varepsilon \in H^1(U_2),\ldots ,w_{n+1}^\varepsilon \in H^1(U_{n+1})\) satisfy, for every \(m\in \{1,\ldots ,n+1\}\), the Dirichlet problems
Finally, let \(\overline{u} \in H^1(U_1)\) and, for every \(m\in \{1,\ldots ,n+1\}\), the function \(\overline{w}_{m}\in H^1(U_m)\) satisfy the homogenized problems
Then, for every \(m\in \{1,\ldots ,n+1\}\), we have the estimate
Observe that, due to the assumed regularity of L in the spatial variable and the Schauder estimates, the vector fields \(\mathbf {F}_m(\nabla u^\varepsilon ,\nabla w^\varepsilon _1,\ldots ,\nabla w^\varepsilon _{m-1},\tfrac{x}{\varepsilon })\) on the right side of the equations for \(w_m\) in (1.19) belong to \(L^\infty (U_m)\). In particular, they belong to \(L^2(U_m)\) and therefore the Dirichlet problems in (1.19) are well-posed in the sense that the solutions \(w_m\) belong to \(H^1(U_m)\). Of course, this regularity given by the application of the Schauder estimate depends on \(\varepsilon \) (and indeed the constants blow up like a large power of \(\varepsilon ^{-1}\)) and therefore this remark is not very useful as a quantitative statement. To prove the homogenization result for the mth linearized equation, we will need to possess much better bounds on these vector fields, which amounts to better regularity on the solutions \(\nabla u, \nabla w_1,\ldots ,\nabla w_{m-1}\).
This is the reason we must prove Theorems 1.1 and 1.2 at the same time (in an induction on the order m of the linearized equation and of the regularity of \(D^2\overline{L}\)) as the following result on the large-scale regularity of solutions of the linearized equations and of the linearization errors. Its statement is a generalization of the large-scale \(C^{0,1}\) estimates for linearized equation and differences of solutions proved in [1].
As mentioned above, throughout the paper we use the following notation for volume-normalized \(L^p\) norms: for each \(p\in [1,\infty )\), \(U \subseteq {\mathbb {R}^{d}}\) with \(|U|<\infty \) and \(f\in L^p(U)\),
Theorem 1.3
(Large-scale \(C^{0,1}\) estimates for the linearized equations).
Let \(n\in \{0,\ldots ,\mathsf {N}\}\), \(q\in [2,\infty )\), and \(\mathsf {M}\in [1,\infty )\). Then there exist \(\sigma (q,\mathrm {data})>0\), a constant \(C(q,\mathsf {M},\mathrm {data}) <\infty \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}\leqq \mathcal {O}_{\sigma }\left( C \right) \) such that the following statement is valid. For \(R \in \left[ 2\mathcal {X},\infty \right) \) and \(u,v,w_1,\ldots ,w_{n+1}\in H^1(B_R)\) satisfying, for every \(m\in \{1,\ldots ,n+1\}\),
and defining, for \(m\in \{0,\ldots ,n\}\), the mth order linearization error \(\xi _m \in H^1(B_R)\) by
then we have, for every \(r \in \left[ \mathcal {X}, \tfrac{1}{2} R \right] \) and \(m\in \{0,\ldots ,n\}\), the estimates
and
The main interest in the above theorem is the case of the exponent \(q=2\). However, we must consider arbitrarily large exponents \(q\in [2,\infty )\) in order for the induction argument to work. In particular, in order to show that Theorem 1.3 for some n implies Theorem 1.2 for \(n+1\), we need to consider potentially very large q (depending on n).
As mentioned above, Theorems 1.1, 1.2 and 1.3 are proved together in an induction argument. Each of the theorems has already been proved in the case \(\mathsf {N}=0\) and \(q=2\) in our previous paper [1]. The integrability in Theorem 1.3 is upgraded to \(q \in (2,\infty )\) in Propositions 4.7 and 5.1 below for \(w_1\) and \(\xi _0\), respectively. These serve as the base case of the induction. The main induction step is comprised of the following three implications:
Regularity of \(\overline{L}\) (Section 2). We show that if Theorems 1.1, 1.2 and 1.3 are valid for some \(n\in \{0,\ldots ,\mathsf {N}-1\}\), then Theorems 1.1 is valid for \(n+1\). The argument essentially consists of differentiating the corrector equation for the nth linearized equation in the parameter p. However, the reader should not be misled into expecting a simple argument based on the implicit function theorem. Due to the lack of sufficient spatial integrability of the vector fields \(\mathbf {F}_{m}\), it is necessary to use the large-scale regularity theory (that is, the assumed validity of Theorem 1.3 for n) to complete the argument.
Homogenization of higher-order linearized equations (Section 3). We argue, for \(n\in \{0,\ldots ,\mathsf {N}-1\}\), that if Theorem 1.1 is valid for \(n+1\) and Theorems 1.2 and 1.3 are valid for n, then Theorems 1.2 is valid for \(n+1\). The regularity of \(\overline{L}\) allows us to write down the homogenized equation, while the homogenization and regularity estimates for the previous linearized equations allow us to localize the heterogeneous equation; that is, approximate it with another equation which has a finite range of dependence and bounded coefficients. This allows us to apply homogenization estimates from [4].
Large-scale \(C^{0,1}\)regularity of linearized solutions and linearization errors (Sections 4 and 5). We argue, for \(n\in \{0,\ldots ,\mathsf {N}-1\}\), that if Theorems 1.1 and 1.2 are valid for \(n+1\) and Theorem 1.3 is valid for n, then we may conclude that Theorem 1.3 is also valid for \(n+1\). Here we use the method introduced in [6] of applying a quantitative excess decay iteration, based on the “harmonic” approximation provided by the quantitative homogenization statement. This estimate controls the regularity of the \(w_m\)’s on “large” scales (that is, larger than a multiple of the microscopic scale). To obtain \(L^q\)-type integrability for \(\nabla w_m\), it is also necessary to control the small scales, and for this we apply deterministic Calderón–Zygmund-type estimates (this is our reason for assuming L possesses some small-scale spatial regularity). The estimates for the linearization errors \(\xi _{m-1}\) are then obtained as a consequence by comparing them to \(w_m\).
From a high-level point of view, the induction argument summarized above resembles the resolution of Hilbert’s 19th problem on the regularity of minimizers of integral functionals with uniformly convex and smooth integrands. The previous three theorems allow us to prove the next two results, which can be considered as resolutions of Hilbert’s 19th problem in the context of homogenization.
The first is the following result on precision of the higher-order linearization approximations which matches the one we have in the constant-coefficient case, as discussed near the end of Section 1.3:
Corollary 1.4
(Large-scale estimates of linearization errors). Fix \(n\in \{0,\ldots ,\mathsf {N} \}\), \(\mathsf {M}\in [1,\infty )\) and let \(U_0,U_1,\ldots ,U_{n} \subseteq {\mathbb {R}^{d}}\) be a sequence of bounded Lipschitz domains satisfying
There exist constants \(\sigma (\mathrm {data}) \in \left( 0,\tfrac{1}{2} \right] \), \(C(\{U_m\},\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid. Let \(r \in \left[ \mathcal {X},\infty \right) \), \(n\in \{1,\ldots ,\mathsf {N}\}\) and \(u,v\in H^1(rU_0)\) satisfy
and recursively define \(w_m \in H^1(rU_m)\), for every \(m\in \{1,\ldots ,n\}\), to be the solution of the Dirichlet problem
Then, for every \(m\in \{1,\ldots ,n\}\),
Corollary 1.4 is an easy consequence of the previous theorems stated above. Its proof is presented in Section 6.
The analysis of the linearized equations presented in the theorems above allow us to develop a higher regularity theory for solutions of the nonlinear equation on large scales, in analogy to the role of the Schauder theory in the resolution of Hilbert’s 19th problem on the regularity of solutions of nonlinear equations with smooth (or constant) coefficients. This result generalizes the large-scale \(C^{1,1}\)-type estimate proved in our previous paper [1] to higher-order regularity as well as the result in the linear case [4, Theorem 3.6].
Before giving the statement of this result, we introduce some additional notation and provide some motivational discussion. Given a domain \(U\subseteq {\mathbb {R}^{d}}\), we define
This is the set of solutions of the nonlinear equation in the domain U, which we note is a stochastic object. We next define \(\mathcal {L}_1\) to be the set of global solutions of the nonlinear equation which exhibit at most linear growth at infinity:
For each \(p\in {\mathbb {R}^{d}}\), we denote the affine function \(\ell _p\) by \(\ell _p(x):=p\cdot x\). Observe that if the difference of two elements of \(\mathcal {L}_1\) has strictly sublinear growth at infinity, it must be constant, by the \(C^{0,1}\)-type estimate for differences (the estimate (1.23) with \(m=0\)). Therefore the following theorem, which was proved in [1], gives a complete classification of \(\mathcal {L}_1\).
Theorem 1.5
(Large-scale \(C^{1,1}\)-type estimate [1, Theorem 1.3]).
Fix \(\sigma \in \left( 0,d\right) \) and \(\mathsf {M} \in [1,\infty )\). There exist \(\delta (\sigma ,d,\Lambda )\in \left( 0, \frac{1}{2} \right] \), \(C(\mathsf {M},\sigma ,\mathrm {data})<\infty \) and a random variable \(\mathcal {X}_\sigma \) which satisfies the estimate
such that the following statements are valid:
- (i)
For every \(u \in \mathcal {L}_1\) satisfying \(\limsup _{r \rightarrow \infty } \frac{1}{r} \left\| u - (u)_{B_r} \right\| _{\underline{L}^2(B_r)} \leqq \mathsf {M}\), there exist an affine function \(\ell \) such that, for every \(R\geqq \mathcal {X}_\sigma \),
$$\begin{aligned} \left\| u - \ell \right\| _{\underline{L}^2(B_R)} \leqq C R^{1-\delta } . \end{aligned}$$ - (ii)
For every \(p\in B_\mathsf {M}\), there exists \(u\in \mathcal {L}_1\) satisfying, for every \(R\geqq \mathcal {X}_\sigma \),
$$\begin{aligned} \left\| u - \ell _p \right\| _{\underline{L}^2(B_R)} \leqq C R^{1-\delta } . \end{aligned}$$ - (iii)
For every \(R\geqq \mathcal {X}_s\) and \(u\in \mathcal {L}(B_R)\) satisfying \(\frac{1}{R} \left\| u - (u)_{B_R} \right\| _{\underline{L}^2 \left( B_{R} \right) } \leqq \mathsf {M}\), there exists \(\phi \in \mathcal {L}_1\) such that, for every \(r \in \left[ \mathcal {X}_\sigma , R \right] \),
$$\begin{aligned} \left\| u - \phi \right\| _{\underline{L}^2(B_r)} \leqq C \left( \frac{r}{R} \right) ^{2} \inf _{\psi \in \mathcal {L}_1} \left\| u - \psi \right\| _{\underline{L}^2(B_R)}. \end{aligned}$$(1.28)
Statements (i) and (ii) of the above theorem, which give the characterization of \(\mathcal {L}_1\), can be considered as a first-order Liouville-type theorem. In the case of deterministic, periodic coefficient fields, this result was proved by Moser and Struwe [23], who generalized the result of Avellaneda and Lin [8] in the linear case. Part (iii) of the theorem is a quantitative version of this Liouville-type result, which we call a “large-scale \(C^{1,1}\) estimate” since it states that, on large scales, an arbitrary solution of the nonlinear equation can be approximated by an elements of \(\mathcal {L}_1\) with the same precision as harmonic functions can be approximated by affine functions. It can be compared to similar statements in the linear case (see for instance [4, 17]).
In this paper we prove a higher-order version of Theorem 1.5. We will show that, just as a harmonic function can be approximated locally by harmonic polynomials, we can approximate an arbitrary element of \(\mathcal {L}(B_R)\) by elements of a random set of functions which are the natural analogue of harmonic polynomials. In order to state this result, we must first define this space of functions.
Let us first discuss the constant-coefficient case. If \(\overline{L}\) is a smooth Lagrangian, we know from the resolution of Hilbert’s 19th problem that solutions of \(-\nabla \cdot D_p \overline{L}(\nabla \overline{u}) = 0\) are smooth and thus may be approximated by a Taylor expansion at each point. One may then ask, can we characterize the possible Taylor polynomials? In Appendix E we provide such a characterization in terms of the linearized equations. The quadratic part is an \({\overline{\mathbf {a}}}_p:=D^2\overline{L}(p)\)-harmonic polynomial and the higher-order polynomials satisfy the equations the linearized equations, involving the \(\overline{\mathbf {F}}_m\)’s as right hand sides. More precisely, for each \(p\in {\mathbb {R}^{d}}\) and \(n\in \mathbb {N}\), we set
It is not too hard to show that \(\overline{\mathsf {W}}_n^{p,\text {hom}} \subseteq \mathcal {P}^{\mathrm{hom}}_2 \times \cdots \times \mathcal {P}^{\mathrm{hom}}_{n+1}\), where \(\mathcal {P}^{\mathrm{hom}}_j\) stands for homogeneous polynomials of degree j. Indeed, we see, by Liouville’s theorem, that \(\overline{w}_1\) is an \({\overline{\mathbf {a}}}_p\)-harmonic polynomial of degree two. More importantly, according to Appendix E, we have that if \(\overline{u}\) solves \(-\nabla \cdot D_p \overline{L}(\nabla \overline{u}) = 0\) in the neighborhood of origin, and we set \(p = \nabla \overline{u}(0)\) and \(\overline{w}_m(x) := \frac{1}{m+1} \nabla ^{m+1} \overline{u}(0) \, x^{\otimes (m+1)}\), then
In particular, \(\overline{w}_m\) is a sum of a special solution of
in \(\mathcal {P}^{\mathrm{hom}}_{m+1}\) and an \({\overline{\mathbf {a}}}_p\)-harmonic polynomial in \(\mathcal {P}^{\mathrm{hom}}_{m+1}\) . For our purposes it is convenient to relax the growth condition at the origin and define simply
With this definition, we lose the homogeneity of polynomials. Following the approach in [4, Chapter 3], it is natural to define heterogeneous versions of these spaces by
that is, the tuplets of heterogeneous solutions with prescribed growth. Here \(\ell _p + \phi _p\) is the unique element of \(\mathcal {L}_1\) (up to additive constants) satisfying
In other words, \(\phi _p\) is the first-order corrector with slope p: see Lemma 2.10.
The next theorem gives a higher-order Liouville-type result which classifies the spaces \(\mathsf {W}_n^p\) and states that they may be used to approximate any solution of the nonlinear equation with the precision of a \(C^{n,1}\) estimate.
Theorem 1.6
(Large-scale regularity). Fix \(n \in \{1,\ldots ,\mathsf {N}\}\) and \(\mathsf {M} \in [1, \infty )\). There exist constants \(\sigma (n,\mathsf {M},\mathrm {data}),\delta (n,\mathrm {data})\in \left( 0, \frac{1}{2} \right] \) and a random variable \(\mathcal {X}\) satisfying the estimate
such that the following statements hold:
\(\mathrm {(i)}_n\) There exists a constant \(C(n,\mathsf {M},\mathrm {data}) <\infty \) such that, for every \(p \in B_{\mathsf {M}}\) and \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\), there exists \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^p \) such that, for every \(R\geqq \mathcal {X}\) and \(m \in \{1,\ldots ,n\}\),
$$\begin{aligned} \left\| w_m - \overline{w}_m \right\| _{\underline{L}^2(B_R)} \leqq C R^{1-\delta } \left( \frac{R}{\mathcal {X}} \right) ^{m} \sum _{i=1}^m \left( \frac{1}{\mathcal {X}} \left\| \overline{w}_i \right\| _{\underline{L}^2(B_{\mathcal {X}})}\right) ^{\frac{m}{i}} . \end{aligned}$$(1.30)\(\mathrm {(ii)}_n\) For every \(p \in B_{\mathsf {M}}\) and \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^{p}\), there exists \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\) satisfying (1.30) for every \(R\geqq \mathcal {X}\) and \(m \in \{1,\ldots ,n\}\).
\(\mathrm {(iii)}_n\) There exists \(C(n,\mathsf {M},\mathrm {data})<\infty \) such that, for every \(R\geqq \mathcal {X}\) and \(v \in \mathcal {L}(B_R)\) satisfying \( \left\| \nabla v \right\| _{\underline{L}^2 \left( B_{R} \right) } \leqq \mathsf {M}, \) there exist \(p \in B_{C}\) and \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\) such that, defining
$$\begin{aligned} \xi _k(x) := v(x) - p\cdot x - \phi _p(x) - \sum _{i=1}^{k} \frac{w_i}{i!} , \end{aligned}$$we have, for every \(r \in \left[ \mathcal {X}, \frac{1}{2} R \right] \) and \(k \in \{0,1,\ldots ,n\},\) the following estimates:
$$\begin{aligned} \sum _{i=0}^{k}\left( \left\| \nabla \xi _i \right\| _{\underline{L}^{2}(B_r)} \right) ^{\frac{k+1}{i+1}} \leqq C \left( \frac{r}{R} \right) ^{k+1} \sum _{i=0}^{k}\left( \frac{1}{R} \left\| \xi _i - (\xi _i)_{B_R} \right\| _{\underline{L}^{2}(B_R)} \right) ^{\frac{k+1}{i+1}} \end{aligned}$$(1.31)and
$$\begin{aligned} \left\| \nabla \xi _k \right\| _{\underline{L}^{2}(B_r)} \leqq C \left( \frac{r}{R} \right) ^{k+1} \frac{1}{R} \inf _{\phi \in \mathcal {L}_1}\left\| v - \phi \right\| _{\underline{L}^{2}(B_R)}. \end{aligned}$$(1.32)
The proof of Theorem 1.6 is given in Section 7.
We conclude this introduction with some comments about the possibility of generalizing our results in two different directions. First, it is not possible that our results can be extended from the scalar case to general nonlinear elliptic systems. Indeed, for even for constant-coefficient, analytic Lagrangians exhibiting uniform ellipticity and quadratic growth, there are explicit counterexamples (in dimensions larger than two) demonstrating that solutions may not even be Lipschitz: see for instance the dark side survey [22, Sections 3.3 & 3.4]. However, our arguments readily adapt in certain situations special cases (as might be expected), such as the two dimensional case or under an assumption that L has a particular form such as an Uhlenbeck-type structure (see [22, Section 4.7]). Roughly speaking, if one can identify conditions under which constant-coefficient elliptic systems have \(C^1\) regularity, and these conditions are preserved by homogenization, then our techniques should be expected to apply.
Second, one might wonder whether it is possible to develop an analogous theory (in the scalar case) for functionals exhibiting non-quadratic growth, for example heterogeneous versions of the p-Laplacian for \(p\in (1,\infty ) {\setminus } \{2\}\), under assumptions such as [22, (2.20)]. This is indeed an interesting problem which seems to be wide open. Of course, qualitative homogenization is well-known in such a setting [11, 12]. The main roadblock at the present time to establishing a quantitative theory of homogenization, as well as a large-scale regularity theory, is a better understanding of the homogenized Lagrangian. In particular, we would need to show that ellipticity and growth assumptions such as [22, (2.20)] are preserved by homogenization, without which there is no hope for higher regularity. To our knowledge this is open even in the periodic setting.
Appendices C–E of this paper contain estimates for constant-coefficient equations which are essentially known but not to our knowledge written anywhere. We also collect some auxiliary estimates and computations in Appendices A and B.
2 Regularity Estimates for the Effective Lagrangian
In this section, we suppose that \(n\in \{0,\ldots ,\mathsf {N}-1 \}\) is such that
The goal is to prove Theorem 1.1 for \(n+1\).
We proceed by constructing the linearized correctors \(\psi ^{(m)}_{p,h}\) up to \(m=n+2\) and relate the correctors of different orders to each other via differentiation in the parameter p. We show that these results allow us to improve the regularity of \(D^2\overline{L}\) up to \(C^{n+1,\beta }\) and obtain the statement of Theorem 1.1 for \(n+1\). In particular, this allows us to define the effective coefficient \(\overline{\mathbf {F}}_{n+1}\). We also give formulas for the derivatives of \(\overline{L}\) and for \(\overline{\mathbf {F}}_{m}\) in terms of the correctors, which allow us to relate them to each other and show that (1.16) holds.
2.1 The First-Order Correctors and Linearized Correctors
In this subsection we construct the linearized correctors up to order \(n+2\).
For each \(p\in {\mathbb {R}^{d}}\), we define \(\phi _p\) to be the first-order corrector of the nonlinear equation, that is, the unique solution of
The existence and uniqueness (up to additive constants) of the first-order corrector \(\phi _p\) is classical: it can be obtained from a variational argument (applied to an appropriate function space of stationary functions. Alternatively, it can be shown (following the proof given in [4, Section 3.4]) that the elements of \(\mathcal {L}_1\), which was characterized in Theorem 1.5 above (which was proved already in [1]), have stationary gradients.
We define the coefficient field \(\mathbf {a}_p(x)\) to be the coefficients for the linearized equation around the solution \(x\mapsto p\cdot x+\phi _p(x)\):
Given \(p,h\in {\mathbb {R}^{d}}\), we define the first linearized corrector\(\psi ^{(1)}_{p,h}\) to satisfy
In other words, \(\psi ^{(1)}_{p,h}\) is the first-order corrector with slope h for the equation which is the linearization around the first-order corrector \(x\mapsto p\cdot x + \phi _p(x)\). For \(m \in \{ 2,\ldots ,\mathsf {N}+2 \}\), we define the mth linearized corrector to be the unique (modulo additive constants) random field \(\psi ^{(m)}_{p,h}\) satisfying
In other words, \(\psi ^{(m)}_{p,h}\) is the corrector with slope zero for the mth linearized equation around \(x\mapsto p\cdot x + \phi _p(x)\) and \(x\mapsto h\cdot x + \psi ^{(1)}_{p,h}(x)\), \(x\mapsto \psi ^{(2)}_{p,h}(x),\ldots , x\mapsto \psi ^{(m-1)}_{p,h}(x)\). Notice that this gives us the complete collection of correctors for the latter equation, since by linearity we observe that \(\psi ^{(m)}_{p,h} + \psi ^{(1)}_{p,h'}\) is the corrector with slope \(h'\). Furthermore, by the linearity of the map \(h \mapsto h + \nabla \psi ^{(1)}_{p,h}(x)\), it is easy to see from the structure of the equations of \(\psi ^{(m)}_{p,h}\) that, for \(p,h \in \mathbb {R}^d\), and \(t \in \mathbb {R}\),
For \(p,h \in \mathbb {R}^d\), we define
By (2.4), we have that
We first show that the the problem (2.3) for the mth linearized corrector is well-posed for \(m\in \{2,\ldots ,n+1\}\). This is accomplished by checking, inductively, using our hypothesis (2.1) (and in particular the validity of Theorem 1.3 for \(m\leqq n\)), that we have appropriate estimates on the vector fields \(\mathbf {F}_m(\cdots )\) on the right side. We have to make this argument at the same time that we obtain estimates on the smoothness of the correctors \(\psi ^{(m)}_{p,h}\) as functions of p. In fact, we prove that \(\nabla \psi ^{(m+1)}_{p,h} = h\cdot D_p\nabla \psi ^{(m)}_{p,h}\), which expressed in coordinates is
We will also obtain \(C^{0,1}\)-type bounds on the linearized correctors, which together with the previous display yields good quantitative control on the smoothness of the correctors in p.
Before the main statements, let us collect a few preliminary elementary results needed in the proofs. The following lemma is well-known and can be proved by the Lax–Milgram lemma (see for instance [21, Chapter 7]).
Lemma 2.1
Let \(\mathbf {a}(\cdot )\) be a \(\mathbb {Z}^{d}\)-stationary random field valued in the symmetric matrices satisfying the ellipticity bound \(I_d \leqq \mathbf {a}\leqq \Lambda I_d\). Suppose that \(\mathbf {f}\) is a \(\mathbb {Z}^{d}\)-stationary, \({\mathbb {R}^{d}}\)-valued random field satisfying
Then there exists a unique random potential field \(\nabla z\) satisfying
and, for a constant \(C(d,\Lambda )<\infty \), the estimate
Next, a central object in our analysis is the quantity \(\mathbf {F}_m\) defined in (1.15). Fix \(m\in \{2,\ldots , \mathsf {N}+1\}\) and \(x,p,h_1,\ldots ,h_{m-1} \in \mathbb {R}^d\). One can easily read from the definition that we have
It then follows, by Young’s inequality, that there exists \(C(m,\mathrm {data})<\infty \) such that
We have, similarly, that
and, for \(k \in \{1,\ldots ,m-1\}\),
Using these we get, for \(x,p,\tilde{p}, h_1,\tilde{h}_1,\ldots , h_{m-1},\tilde{h}_{m-1} \in \mathbb {R}^d\),
Therefore, by Young’s inequality, we get, for all \(\delta >0\),
Furthermore, as we will be employing an induction argument in m, it is useful to notice that the leading term in \(\mathbf {F}_m\) has a simple form, and we have
with \(\tilde{\mathbf {F}}_2 = 0\). Moreover, as is shown in Appendix B, if, for \(p,h \in \mathbb {R}^d\), \(t \mapsto \mathbf {g}(p+th)\) is m times differentiable at \(t=0\) and noticing that since \(m \leqq \mathsf {N}+1\), \(p\mapsto L(p,x)\) is in \(C^{m+2}\), then
Our first result in this section gives direct consequences of (2.1) for estimates on the first-order correctors and linearized correctors.
Theorem 2.2
(Quantitative estimates on linearized correctors). Assume (2.1) is valid. Fix \(\mathsf {M}\in [1,\infty )\). For every \(m\in \{ 2,\ldots , n+1 \}\) and \(p,h\in {\mathbb {R}^{d}}\), there exists a function \(\psi ^{(m)}_{p,h}\) satisfying (2.3). Moreover, there exist constants \(C(\mathsf {M},\mathrm {data})<\infty \) and \(\sigma (n,d,\Lambda ) \in \left( 0,\tfrac{1}{2}\right] \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}\leqq \mathcal {O}_\sigma (C)\) such that the following statement is valid. For every \(p\in B_{\mathsf {M}}\), \(h \in \overline{B}_1\), \(m\in \{1,\ldots ,n\}\), and \(r\geqq \mathcal {X}\),
and, for every \(p\in B_{\mathsf {M}}\), \(h \in \overline{B}_1\), \(m\in \{1,\ldots ,n+1\}\) and \(r\geqq \mathcal {X}\),
Finally, for \(q \in [2,\infty )\)and \(m\in \{1,\ldots ,n+1\}\), there exist constants \(\delta (m,d,\Lambda )\in \left( 0,\tfrac{1}{2}\right] \) and \(C(q,m,\mathsf {M},\mathrm {data})<\infty \) such that, for every \(p\in B_{\mathsf {M}}\) and \(h \in \overline{B}_1\),
Proof
Set, for \(m \in \{1,\ldots ,n+1\}\),
We first collect two consequences of Theorem 1.3 assumed for n. Fix \(q \in [2,\infty )\). Theorem 1.3 implies that there is a minimal scale \(\mathcal {X}\) such that (1.22) and (1.23) are valid with \(q (n+1)\) instead of q and for every \(r \in \left[ \mathcal {X}, \tfrac{1}{2} R \right] \). Hence, for every \(r \in \left[ \mathcal {X}, \tfrac{1}{2} R \right] \) and \(m\in \{0,\ldots ,n\}\), we get the estimates
and
On the other hand, using (2.7) we obtain
In particular, it follows by (2.17) that, for \(m \in \{1,\ldots ,n+2\}\), \(R>2\mathcal {X}\) and \(r \in [\mathcal {X},\tfrac{1}{2} R)\),
Since \(\nabla \psi ^{(i)}_{p,h} \) is \(\mathbb {Z}^d\)-stationary random field, we have by the ergodic theorem, after sending \(R\rightarrow \infty \), that a.s.
Furthermore, by Lemma 2.1 and the previous display we get, for \(m \in \{1,\ldots ,n+2\}\), that
Observe that the limiting behavior of \(\xi _0\) and \(\psi ^{(1)}_{p,h}\) can be identified via their equations
respectively, where
By \(\mathbb {Z}^d\)-stationary of \(\nabla \phi _p\) and \(\nabla \phi _{p+h}\), implying \(\mathbb {Z}^d\)-stationarity of \(\mathbf {a}_p\) and \(\tilde{\mathbf {a}}_p\), we may apply Lemma 2.1 to obtain
It then follows inductively that, for \(m \in \{1,\ldots ,n+2\}\),
Using this together with the ergodic theorem and (2.17), (2.18), we obtain inductively, for \(q\in [2,\infty )\), \(m \in \{0,\ldots ,n+1\}\) and \(r \geqq \mathcal {X}\),
Now (2.16) follows by giving up a volume factor. The proof is complete. \(\quad \square \)
We next show, again using (2.1), that the corrector \(\psi ^{(n+2)}_{p,h}\) satisfies an \(L^2\)-type gradient estimate.
Lemma 2.3
Assume (2.1) is valid. Let \(\mathsf {M} \in [1,\infty )\). Suppose that \(p \in B_{\mathsf {M}}\) and \(h \in \overline{B}_1\). There exists \(\psi ^{(n+2)}_{p,h}\) satisfying (2.3) for \(m=n+2\) and a constant \(C(n,\mathsf {M},\mathrm {data})\) such that
Proof
The result follows directly by Lemma 2.1 and (2.16) using (2.7). \(\quad \square \)
Lemma 2.4
Assume (2.1) is valid. Suppose that \(p \in B_{\mathsf {M}}\) and \(h \in \overline{B}_1\). Then, for \(m\in \{1,\ldots ,n+1\}\), we have, a.s. and almost everywhere,
Moreover, for \(\beta \in (0,1)\) and \(m \in \{1,\ldots ,n+1\}\), there exists \(C(\beta ,m,\mathsf {M},\mathrm {data})<\infty \) such that, for \(t \in (-1,1)\),
For \(m \in \{1,\ldots ,n\}\), we can take \(\beta =1\) in (2.20). Finally, we have, for \(m \in \{1,\ldots ,n+2\}\), that, a.s. and almost everywhere,
Proof
Fix \(p \in \mathbb {R}^d\) and, without loss of generality, \(h \in \partial B_1\). By (2.14) we have that \(p \mapsto p + \nabla \phi _p(x)\) is \(C^n\) and \(D_p^j \nabla \phi _p(x) h^{\otimes j} = \psi _{p,h}^{(j)}\) for every \(j \in \{2,\ldots ,n\}\) almost surely for almost every x. Thus (2.20) is valid with \(\beta =1\) for \(m\in \{1,\ldots ,n-1\}\) and, by (2.13),
We denote, in short, for \(t \ne 0\),
Observe that, by (2.16), \(\mathbb {E}\left[ \left\| \nabla \zeta _{p,h,t}^{(n)} \right\| _{L^2({{\square }}_0)}^2 \right] < \infty \) for \(t\ne 0\).
Step 1. We prove that, for \(t\in (-1,1)\), \(t\ne 0\),
To show (2.23), we first claim that the difference quotient solves the equation
Indeed, then (2.23) follows by Lemma 2.1. Rewriting
we observe that the first three terms on the right are solenoidal by the equations of \(\psi _{p+th,h}^{(n)}\), \(\psi _{p,h}^{(n)}\) and \(\psi _{p,h}^{(n+1)}\), respectively, and the fourth term on the right is zero by (2.22). We thus obtain (2.24).
Step 2. We will estimate the terms on the right in (2.23) separately, and in this step we first show that, for \(t\in (-1,1)\), \(t\ne 0\),
By the triangle inequality we have
Therefore, by Hölder’s inequality and (2.16),
which is (2.25).
Step 3. We show that
We have that
and since
Step 4. We then prove that
Using decomposition (2.12), we have that
where
is a function of \(\nabla \phi _p,\nabla \psi _{p,h}^{(1)}, \ldots ,\nabla \psi _{p,h}^{(n-1)}\) for \(n \geqq 2\), and thus differentiable in p. In particular, by (2.16), for every \(q \in [2,\infty )\) there is \(\delta (q,n,\mathrm {data})>0\) and \(C(q,n,\mathsf {M},\mathrm {data})<\infty \) such that
Using the above decomposition for \(D_p \mathbf {f}_{p,h}^{n} \cdot h\), we compute
We have by (2.16) that
and therefore
The right hand side can be estimated with the aid of (2.25) to obtain (2.27).
Step 5. Conclusion. Combining (2.23) with (2.25), (2.26) and (2.27) yields (2.20) by Young’s inequality. Now, (2.20) implies (2.19) for \(m=n\). Therefore, we may replace n by \(n+1\) in Steps 1-4 above, and conclude that (2.20) is valid for \(m=n+1\) as well, which then gives (2.19) for \(m=n+1\). Using obtained formula, it is straightforward to show that (2.20) is valid for \(m=n\) with \(\beta =1\). Indeed, we notice that
from which we get (2.20) for \(m=n\) with \(\beta =1\). Finally, (2.20) implies that \(t \mapsto \nabla \phi _{p+th}\) is in \(C^{n+2,\beta }\) close to \(t=0\), and thus we have that (2.21) is valid by (2.13). The proof is complete. \(\quad \square \)
Lemma 2.5
Assume (2.1) is valid. Let \(\mathsf {M} \in [1,\infty )\) and \(p \in B_{\mathsf {M}}\). Then \(p \mapsto p + \nabla \phi _p\) is \((n+2)\) times differentiable with respect to p and, for \(q \in [2,\infty )\), there are constants \(\delta (q,n,\mathrm {data}) \in \left( 0, \tfrac{1}{2} \right] \) and \(C(q,n,\mathsf {M},\mathrm {data})<\infty \) such that, for \(m \{1,\ldots ,n+1\}\),
The proof of Lemma 2.5 relies on a general principle in polarization based on multilinear analysis, and it is formalized in the following lemma:
Lemma 2.6
Let V be a real, finite-dimensional vector space, and let \(\Phi :V^{n}\rightarrow \mathbb {R}\) be a multilinear, symmetric form, that is, for all \(v_1,\ldots ,v_n \in V\) and any permutation \(\sigma \) of \(\{1,\ldots ,n\}\), we have
For \(v\in V\), define \(\phi (v):=\Phi (v, v, \dots , v)\). Then, for \(v_{1}, \dots , v_{n}\in V\), the polarization formula
holds, where the leftmost summation is over all non-empty subsets \(A\subseteq \{1,2,\dots , n\}\) and |A| is the number of elements in A.
Proof
We show the equivalent statement that
where the sum on the left is over all non-empty subsets A of \(\{1,2,\dots , n\}\). For this, we begin by expanding each summand \(\phi \left( \sum _{j\in A}v_{j}\right) =\Phi \Big (\sum _{j\in A}v_{j},\dots , \sum _{j\in A}v_{j}\Big )\) fully, as a sum of terms of the form \(\Phi (v_{j_{1}}, \dots , v_{j_{n}})\) with \(j_1,\dots , j_n\in A\). Using the symmetry of \(\Phi \), each such term can be written as \(\Phi (v_{f(1)}, \dots , v_{f(n)})\), with non-decreasing indices \(f(1)\leqq f(2)\leqq \cdots \leqq f(n)\) in A. Denote
and, for \(f \in \mathcal {M}\),
Letting \(c_{A}(f)\) denote the number of ordered n-tuples \((j_{1}, \dots , j_{n})\) of elements of A which can be reordered to form \((f(1),\dots , f(n))\), it follows that the expression \(\sum _{A\subseteq \{1,\dots , n\}}(-1)^{n-|A|}\phi \left( \sum _{j\in A}v_{j}\right) \) can be expanded to give
Changing the order of summation gives
where the sum on the right is over all subsets A of \(\{1,2,\dots , n\}\) which contain \(\text {im}\,f\). Each such subset A can be written as \(\{f(1), \dots , f(n)\}\cup B\), for some set B, possibly empty, satisfying \(B\subseteq \{1,2,\dots , n\}{\setminus } \text {im}\,f \). Hence, as \(|A|=|\text {im}\, f|+|B|\) for B defined in this way, we can write our expression as
where
Finally, it is a well-known combinatorial fact that for every nonempty finite set S, we have
where the sum on the left is over all subsets B of S. Thus, above, we have
unless \(f(1)< \cdots < f(n)\). Therefore \(f(1)=1, f(2)=2, \dots , f(n)=n\), and in this case \(s_{f}=1\) and \(c_{\{1,\dots , n\}}(f)=n!\). It follows that
as was to be shown; this proves the polarization formula. \(\quad \square \)
Proof of Lemma 2.5
The lemma follows from Lemmas 2.3, 2.4, 2.6, and (2.16).
\(\square \)
We next prove Hölder continuity of \(p\mapsto D_p^{n+2} \phi _p\) and \(p \mapsto \mathbf {f}^{n+2}_{p,h}\).
Lemma 2.7
Assume (2.1) is valid. Let \(\mathsf {M} \in [1,\infty )\) and \(\beta \in (0,1)\). Then there is a constant \(C(\beta ,n,\mathsf {M},\mathrm {data})<\infty \) such that, for all \(p,p' \in B_{\mathsf {M}}\) and \(h \in \overline{B}_1{\setminus }\{0\}\),
Proof
Fix \(\mathsf {M} \in [1,\infty )\) and \(\beta \in (0,1)\), and fix \(p,p' \in B_{\mathsf {M}}\) and \(h \in \overline{B}_1{\setminus }\{0\}\). By Lemma 2.1 and equations of \(\psi _{p,h}^{(n+2)}\) and \(\psi _{p',h}^{(n+2)}\) we have that
Therefore, in view of Lemma 2.6, it is enough to show that
By (2.12) we may decompose \(\mathbf {f}^{n+2}_{p,h}\) as
Since \(\tilde{\mathbf {F}}_{n+2}\) is \(C^{0,1}\) in its first argument and polynomial in its last n arguments, we obtain by homogeneity, the chain rule and Lemma 2.5 that
Therefore, the leading term is the first one on the right in (2.31). Indeed, we observe by the above display that
To conclude, we need to estimate the first term on the right. To this end, we use the triangle inequality to get, for any \(\beta \in [0,1]\),
It follows, by Hölder’s inequality, that
By Lemma 2.5 and Hölder’s inequality, we get
together with
and
Consequently, combining above displays yields (2.30), finishing the proof. \(\quad \square \)
2.2 Smoothness of \(\overline{L}\)
It is a well-known and easy consequence of qualitative homogenization that \(D\overline{L}\) is given by the formula
In [1], we proved that \(\overline{L}\in C^2\) and \(D^2\overline{L}\) is given by the formula
We next generalize this result to higher derivatives of \(\overline{L}\) by essentially differentiating the previous formula many times and applying the results of the previous subsection. Moreover, we validate the statement of Theorem 1.1 for \(n+1\).
Proposition 2.8
Assume that (2.1) is valid. Then Theorem 1.1 is valid for \(n+1\). Moreover, for every \(m\in \{ 1,\ldots ,n+1\}\) and \(h\in \overline{B}_1\), we have the formula
Proof
Fix \(\mathsf {M} \in [1,\infty )\) and \(p \in B_{\mathsf {M}}\). We begin by showing (2.33). Starting from (2.32), and observing that since
we have that
This implies that
Assume then, inductively, that for some \(m \in \{3,\ldots ,n+1\}\) we have that, for all \( k \in \{2,\ldots ,m\}\),
We prove that (2.34) is valid for \(k=m+1\) as well. Differentiating with respect to p yields, using (2.19) and (2.21), that
proving the induction step. This validates (2.33).
To show the regularity of \(\overline{L}\), we first observe that Theorem 2.2 and Lemma 2.3, together with (2.34) and Lemma 2.6, yield that
Fix then \(p' \in B_{\mathsf {M}}\). Since
decomposing
and noticing that
we obtain by Lemmas 2.5 and 2.7, together with Hölder’s inequality, that
In view of Lemma 2.6 this yields
proving that \([D_p^{n+3} \overline{L}]_{C^{0,\beta }(B_{\mathsf {M}})} \leqq C\). Together with (2.35) we thus get
which is the statement of Theorem 1.1 for \(n+1\). The proof is complete. \(\quad \square \)
Remark 2.9
Since we now have that Theorem 1.1 is valid for \(n+1\), we also have that \(p \mapsto \overline{F}_{n+2}(p,\cdot ,\ldots ,\cdot )\) belongs to \(C^{0,\beta }\) for all \(\beta \in (0,1)\). In particular, for \(\beta \in (0,1)\) and \(\mathsf {M}\in [1,\infty )\), there exists \(C(n,\beta ,\mathsf {M},\mathrm {data})<\infty \) such that, for every tuplet \((h_1,\ldots ,h_{n+1}) \in \mathbb {R}^d \times \ldots \mathbb {R}^d\), we have that
2.3 Sublinearity of Correctors
By the ergodic theorem, we have that, for every \(p,h\in {\mathbb {R}^{d}}\) and \(m\in \{1,\ldots ,n+1\}\), the correctors and linearized correctors are (qualitatively) sublinear at infinity:
The assumption (2.1) allows us to give a quantitative estimate of this sublinearity.
Lemma 2.10
(Sublinearity of correctors). Assume (2.1) is valid. Let \(\mathsf {M} \in [1,\infty )\). There exist \(\alpha (\mathrm {data}),\delta (\mathrm {data}) >0\), \(C(\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}\leqq \mathcal {O}_\delta (C)\) such that, for every \(r\geqq \mathcal {X}\), \(p \in B_{\mathsf {M}}\), \(h\in \overline{B}_1\) and \(m\in \{1,\ldots ,n+1\}\),
Proof
Fix \(p \in B_\mathsf {M}\) and \(h \in \overline{B}_1\). Clearly \(x \mapsto p\cdot x + \phi _p(x)\) belongs to \(\mathcal {L}_1\), and thus the result follows from [1, Theorem 1.3] as in [4, Section 3.4]. Hence we are left to show that \(\psi _{p,h}^{(m)} \) satisfies the estimate in the statement. For \(m=1\) the result follows by [1, Theorem 5.2] and [4, Section 3.4]. We thus proceed inductively. Assume that (2.36) is valid for \(m \in \{1,\ldots ,k\}\) for some \(k \in \{1,\ldots , n\}\). We then show that it continues to hold for \(m=k+1\). Since (2.1) is valid, by taking \(\alpha \) and \(\mathcal {X}\) as in Theorem 1.2, and setting \(\mathcal {Y}:= \mathcal {X}^{2/\alpha }\), we have that if \(R:=\varepsilon ^{-1}\geqq \mathcal {Y}\), then \(\varepsilon ^\alpha \mathcal {X}\leqq R^{-\frac{\alpha }{2}}\). We relabel \(\mathcal {Y}\) to \(\mathcal {X}\) and \(\frac{\alpha }{2}\) to \(\alpha \). We further take \(\mathcal {X}\) larger, if necessary, so that [1, Proposition 4.3] is at our disposal. Suppressing both p and h from the notation, we let \(\phi _{r}\) and \(\psi _{r}^{(m)}\), for \(m \in \{1,\ldots ,n+1\}\), solve
where
Now, \(\phi _r, \psi _r^{(1)},\ldots , \psi _r^{(m-1)}\) all homogenize to zero and we get, by Theorem 1.2, that, for \(r \geqq \mathcal {X}\),
This and the induction assumption, that is that (2.36) is valid for \(m \in \{1,\ldots ,k\}\), together with Lemma 2.11 below, imply that
Combining the previous two displays yields
Now, since \(\psi _{2r}^{(k+1)}- \psi _{r}^{(k+1)}\) is \(\mathbf {a}_p\)-harmonic in \(B_r\), we have by the Lipschitz estimate [1, Proposition 4.3] that, for \(r \geqq \mathcal {X}\) and \(t \in [\mathcal {X},r]\),
Therefore, by compactness, there exists \(\hat{\psi }^{(k+1)}\) such that, for \(t \in [\mathcal {X},r]\),
Proceeding now as in [4, Section 3.4] proves that \(\nabla \hat{\psi }^{(k+1)}\) is \(\mathbb {Z}^d\)-stationary. Finally, by integration by parts we also obtain that \(\mathbb {E}\left[ \int _{{{\square }}_0} \nabla \hat{\psi }^{(k+1)} \right] = 0\) and, therefore, since \(\hat{\psi }^{(k+1)}\) solves the same equation as \(\psi ^{(k+1)}\), by the uniqueness we have that \(\hat{\psi }^{(k+1)} = \psi ^{(k+1)}\) up to a constant. The proof is hence complete by the previous display. \(\quad \square \)
Above we made use of a lemma, which roughly states that if two solutions of the systems of linearized equations are close in \(L^2\) then their gradients are also close. This lemma will also be applied repeatedly in the following sections.
Lemma 2.11
Suppose that (3.1) holds. Let \(q\in [2,\infty )\) and \(\mathsf {M}\in [1,\infty )\). There exist \(\delta (q,\mathrm {data})>0\), \(C(q,\mathsf {M},\mathrm {data}) <\infty \) and a random variable \(\mathcal {X}\leqq \mathcal {O}_{\delta }\left( C \right) \) such that the following holds. Let \(R\geqq \mathcal {X}\), \(N\leqq n+1\) and \(\left( u,w_1,\ldots ,w_{N}\right) ,\left( \tilde{u},\tilde{w}_1,\ldots ,\tilde{w}_{N} \right) \in (H^1(B_R))^{N+1}\) each be a solution of the system of the linearized equations, that is, for every \(m\in \{1,\ldots ,N\}\), we have
and the same holds with \(\left( \tilde{u},\tilde{w}_1,\ldots ,\tilde{w}_{N} \right) \) in place of \(\left( u,w_1,\ldots ,w_{N}\right) \). Then
and, denoting
we have, for every \(m\in \{1,\ldots ,N\}\),
and
Proof
Let us define \(\mathbf {a}(x):= D^2_pL(\nabla u,x)\) and \(\mathbf {f}_m(x):= \mathbf {F}_m(\nabla u,\nabla w_1,\ldots ,\nabla w_{m-1},x)\) and analogously define \(\tilde{\mathbf {a}}\) and \(\tilde{\mathbf {f}}_m\). We assume that \(R \geqq 2^{m+2}\mathcal {X}\), where \(\mathcal {X}\) is as in Theorem 1.3 for n, valid by the assumption of (3.1).
The estimate (2.38) is just the estimate for \(\xi _0\) in Theorem 1.3. It also implies
By (2.10), we have that
Using Hölder’s inequality and applying Theorem 1.3 for n, we obtain, for any \(p\in [2,\infty )\) and \(\delta >0\),
We observe that \(\zeta := w_m-\tilde{w}_m\) satisfies the equation
By Meyer’s estimate, if \(\delta >0\) is small enough, then
Combining these and using (2.41) and the validity of Theorem 1.3, we get
Taking \(\delta _0\) sufficiently small and putting \(\delta _m:=2^{-m}\delta _0\), we get by induction (using Young’s inequality and rearranging several sums) that, for every \(m\in \{1,\ldots ,N\}\),
Combining this with (2.42), we get
These imply (2.39) and (2.40) after a covering argument. \(\quad \square \)
3 Quantitative Homogenization of the Linearized Equations
In this section, we suppose that \(n\in \{0,\ldots ,\mathsf {N}-1 \}\) is such that
The goal is to prove that Theorem 1.2 is also valid for \(n+1\) in place of n. That is, we need to homogenize the \((n+2)\)th linearized equation.
In order to prove homogenization for the \((n+2)\)th linearized equation, we follow the procedure used in [1] for homogenizing the first linearized equation. We first show, using the induction hypothesis (3.1), that the coefficients \(D^2_pL\left( \nabla u^\varepsilon ,\tfrac{x}{\varepsilon }\right) \) and \(\mathbf {F}_{n+1}\left( \tfrac{x}{\varepsilon }, \nabla u^\varepsilon , \nabla w_1^\varepsilon , \ldots ,\nabla w_{n}^\varepsilon \right) \) can be approximated by random fields which are local (they satisfy a finite range of dependence condition) and locally stationary (they are stationary up to dependence on a slowly varying macroscopic variable). This then allows us to apply known quantitative homogenization results for linear elliptic equations which can be found for instance in [4].
3.1 Stationary, Finite Range Coefficient Fields
We proceed in a similar fashion as in [1, Section 3.4] by introducing approximating equations which are stationary and localized.
We fix an integer \(k\in \mathbb {N}\) which represents the scale on which we localize. Let \(v_{p,z}^{(k)}\) denote the solution of the Dirichlet problem
where \(\ell _p\) is the affine function \(\ell _p(x):=p\cdot x\). We then define, for each \(z\in 3^k\mathbb {Z}^{d}\), a coefficient field \(\tilde{\mathbf {a}}_{p,z}^{(k)}\) in \(z+{{\square }}_{k+1}\) by
and then recursively define, for each \(\Theta = (p, h_{1}, \dots , h_{n+1})\in \left( \mathbb {R}^{d}\right) ^{n+2}\), \(m\in \{1,\ldots ,n+1\}\) and \(z\in 3^k\mathbb {Z}^{d}\), the functions \(w_{m,\Theta ,z}^{(k)} \in H^1(z+(1+2^{-m}){{\square }}_{k})\) to be the solutions of the sequence Dirichlet problems
where \(\tilde{\mathbf {F}}_{m,\Theta ,z}^{(k)} \in L^2(z+(1+2^{-(m-1)}){{\square }}_k)\) is defined for \(m\in \{1,\ldots ,n+2\}\) by
Finally, we create \(3^k\mathbb {Z}^{d}\)-stationary fields by gluing the above functions together: for each \(x\in {\mathbb {R}^{d}}\), we define
Notice that \(v^{(k)}_p\) and \(w^{(k)}_{m,\Theta }\) might not be \(H^1\) functions globally, but we can nevertheless define their gradients locally in \(z+{{\square }}_k\). The \(\mathbb {R}^{d(n+3)}\)-valued random field \(\left( \nabla v^{(k)}_p, \nabla w^{(k)}_{1,\Theta },\ldots ,\nabla w^{(k)}_{n+2,\Theta } \right) \) is \(3^{k}\mathbb {Z}^{d}\)-stationary and has a range of dependence of at most \(3^k\sqrt{15+d}\), by construction. The same is also true of the corresponding coefficient fields, since these are local functions of this random field:
In the next subsection, we will apply some known quantitative homogenization estimates to the linear equation
In order to anchor these estimates, we require some deterministic bounds on the vector field \(\mathbf {F}^{(k)}_{n+2,\Theta }\) and thus on the gradients of the functions \(w^{(k)}_{m,\Theta }\) defined above. These bounds are exploding as a large power of \(3^k\), but we will eventually choose \(3^k\) to be relatively small compared with the macroscopic scale (so these large powers of \(3^k\) will be absorbed).
Lemma 3.1
Fix \(\mathsf {M}\in [1,\infty )\). There exist exponents \(\beta (\mathrm {data})>0\), \(Q(\mathrm {data})<\infty \) and a constant \(C(\mathsf {M},\mathrm {data})<\infty \) such that, for every \(k\in \mathbb {N}\), \(m\in \{ 1,\ldots ,n+2\}\) and \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) with \(|p|\leqq \mathsf {M}\), we have
Proof
By [1, Proposition A.3], there exist \(\beta (d,\Lambda )\in (0,1)\) and \(C(\mathsf {M},\mathrm {data})<\infty \) such that, for every \(z\in 3^k\mathbb {Z}^{d}\) and \(p\in B_{\mathsf {M}}\),
We deduce the existence of \(C(\mathsf {M},\mathrm {data})<\infty \) such that
We will argue by induction in \(m\in \{1,\ldots ,n+2\}\) that there exist \(Q(\mathrm {data})<\infty \) and \(C(\mathsf {M},\mathrm {data})<\infty \) such that
For \(m=1\) the claim follows from Proposition A.2 and (3.6). Suppose now that there exists \(M\in \{2,\ldots ,n+2\}\) such that the bound (3.7) holds for each \(m \in \{1,\ldots ,M-1\}\). Then we obtain that, for some \(Q(M,\mathrm {data})<\infty \),
By the Caccioppoli inequality, we get
In view of (3.6) and the previous two displays, another application of Proposition A.2 yields, after enlarging Q and C, the bound (3.7) for \(m=M\). This completes the induction argument and the proof of (3.7). The bound (3.5) immediately follows. \(\quad \square \)
By the assumed validity of Theorem 1.3 for n, we also have that, for each \(\mathsf {M},q \in [2,\infty )\) there exist \(\delta (q,\mathrm {data})>0\) and \(C(\mathsf {M},q,\mathrm {data})<\infty \) and a random variable \(\mathcal {X}= \mathcal {O}_\delta (C)\) such that, for every \(k\in \mathbb {N}\) with \(3^k\geqq \mathcal {X}\) and every \(m\in \{1,\ldots ,n+1\}\) and \(\Theta = (p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) with \(|p| \leqq \mathsf {M}\), we have
and hence, for such k and \(\Theta \) and \(m\in \{1,\ldots ,n+2\}\),
Observe that (3.5) and (3.9) together imply that, for \(\delta \) and C as above and every \(m\in \{ 1,\ldots ,n+2\}\),
We next study the continuity of \(\mathbf {a}_p^{(k)}\) and \(\mathbf {F}^{(k)}_{m,\Theta }\) in the parameter \(\Theta \).
Lemma 3.2
(Continuity of \(\mathbf {a}_p^{(k)}\) and \(\mathbf {F}^{(k)}_{m,\Theta }\) in \(\Theta \)). Fix \(q\in [2,\infty )\) and \(\mathsf {M}\in [1,\infty )\). There exist constants \(\delta (q,\mathrm {data})>0\) and \(Q(q,\mathrm {data}),C(q,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}= \mathcal {O}_\delta (C)\) such that, for every \(k\in \mathbb {N}\) with \(3^k\geqq \mathcal {X}\), \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) and \(\Theta '=(p',h_{1}',\ldots ,h_{n+1}')\in \mathbb {R}^{d(n+2)}\) with \(|p|,|p'|\leqq \mathsf {M}\),
and, for every \(m\in \{ 1,\ldots ,n+2\}\),
Proof
We take \(\mathcal {X}\) to be larger than the random variables in the statements of Lemma 2.11 and Theorems 1.2 and 1.3 for n. The bound (3.11) is then an immediate consequence of (2.38) and the obvious fact that
We then use the equation for the difference \(w^{(k)}_{m,\Theta } - w^{(k)}_{m,\Theta '}\) [see (2.43)] and then apply the result of Lemma 2.11 to obtain, for every \(m\in \{1,\ldots ,n+1\}\),
By induction we now obtain, for every \(m\in \{1,\ldots ,n+1\}\),
This implies (3.12). \(\quad \square \)
By combining (3.5) and (3.12) and using interpolation, we obtain the existence of \(\alpha (\mathrm {data})>0\), \(Q(\mathrm {data})<\infty \) and \(C(\mathrm {data})>0\) such that, with \(\mathcal {X}\) as in the statement of Lemma 3.2, then for every \(k\in \mathbb {N}\) with \(3^k\geqq \mathcal {X}\), \(m\in \{1,\ldots ,n+2\}\), and \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) and \(\Theta '=(p',h_{1}',\ldots ,h_{n+1}')\in \mathbb {R}^{d(n+2)}\) with \(|p|,|p'|\leqq \mathsf {M}\), we have
Likewise, we can use (3.6) and (3.11) to obtain
This variation of Lemma 3.2 will be needed below.
3.2 Setup of the Proof of Theorem 1.2
We are now ready to begin the proof of the implication
We fix parameters \(\mathsf {M} \in [1,\infty )\), \(\delta >0\) and \(\varepsilon \in (0,1)\), a sequence of Lipschitz domains \(U_1,U_2,\ldots ,U_{n+2} \subseteq {{\square }}_0\) satisfying
a function \(f\in W^{1,2+\delta }(U_1)\) satisfying
and a sequence of boundary conditions \(g_1\in W^{1,2+\delta }(U_1), \ldots , g_{n+2} \in W^{1,2+\delta }(U_{n+2})\). We let \(u^\varepsilon \in f+H^1_0(U_1)\) and \(w^\varepsilon _1\in g_1+H^1_0(U_1),\ldots ,w^\varepsilon _{n+2} \in g_{n+2}+H^1_0(U_{n+2})\) as well as \(\overline{u} \in f+H^1_0(U_1)\) and \(\overline{w}_1\in g_1+H^1_0(U_1),\ldots ,\overline{w}_{n+2} \in g_{n+2}+H^1_0(U_{n+2})\) be as in the statement of Theorem 1.2 for \(n+1\) in place of n.
We denote
and
We also choose \(K\in \mathbb {N}\) to be the unique positive integer satisfying \(3^{-K-1}< \varepsilon \leqq 3^{-K}\). We write
By the assumption (3.1), we only need to homogenize the linearized equation for \(m=n+2\). As we have already proved a homogenization result in [1, Theorem 1.1] for the linearized equation with zero right-hand side, it suffices to prove (1.21) for \(m=n+2\) under the assumption that the boundary condition vanishes:
We can write the equations for \(w_{n+2}^\varepsilon \) and \(\overline{w}_{n+2}\) respectively as
and
Our goal is to prove the following estimate: there exists a constant \(C(\mathsf {M},\mathrm {data})<\infty \), exponents \(\sigma (\mathrm {data})\) and \(\alpha (\mathrm {data})>0\) and random variable \(\mathcal {X}\) satisfying
as in the statement of the theorem:
To prove (3.24), we first compare the solution \(w_{n+2}\) to the solution \(\tilde{w}_{n+2}^\varepsilon \) of a second heterogeneous problem, namely
where the coefficient fields \(\tilde{\mathbf {a}}^\varepsilon \) and \(\tilde{\mathbf {F}}^\varepsilon _{n+2}\) are defined in terms of the localized, stationary approximating coefficients (introduced above in Section 3.1) by
The parameter \(k\in \mathbb {N}\) will be chosen below in such a way that \(1\ll 3^{k} \ll \varepsilon ^{-1}\). We also need to declare a second mesoscopic scale by taking \(l\in \mathbb {N}\) such that \(1\ll 3^k \ll 3^l \ll \varepsilon ^{-1}\) and, for every \(m\in \{1,\ldots ,n+1\}\),
Like k, the parameter l will be declared later in this section. For convenience we will also take a slightly smaller domain \(U_{n+3}\) than \(U_{n+2}\), which also depends on \(\varepsilon \) and l and is defined by
Thus we have (3.27) for every \(m\in \{1,\ldots ,n+2\}\). See Section 3.4 below for more on the choices of the parameters k and l.
The estimate (3.24) follows from the following two estimates, which are proved separately below (here \(\mathcal {X}\) denotes a random variable as in the statement of the theorem):
and
3.3 Estimates on Size of the \(w_m^\varepsilon \) and \(\overline{w}_m\)
To prepare for the proofs of (3.29) and (3.30), we present some preliminary bounds on the size of the \(w_m^\varepsilon \)’s. The first is a set of deterministic bounds representing the “worst-case scenario” in which nothing has homogenized up to the current scale.
Lemma 3.3
There exist exponents \(\beta (d,\Lambda )\in (0,1)\) and \(Q(\mathrm {data})<\infty \) and a constant \(C(\{ U_m \}, \mathsf {M},\mathsf {K}_0,\mathrm {data})<\infty \) such that, for every \(m\in \{1,\ldots ,n+2\}\),
Proof
By [1, Proposition A.3], there exist \(\beta (d,\Lambda )\in (0,1)\), \(Q(\mathrm {data})<\infty \) and \(C(\{ U_m\},\mathsf {M},\mathsf {K}_0,\mathrm {data})<\infty \) such that
and hence
We will argue by induction in \(m\in \{1,\ldots ,n+2\}\) that there exists \(Q(m,\mathrm {data})<\infty \) and \(C(m,\{U_k\}, \mathsf {M},\mathsf {K}_0,\mathrm {data})<\infty \) such that
For \(m=1\) the claim follows from Proposition A.2 and (3.32). Suppose now that there exists \(M\in \{2,\ldots ,n+2\}\) such that the bound (3.33) holds for each \(m \in \{1,\ldots ,M-1\}\). Then we obtain that, for some \(Q(\mathsf {M},\mathrm {data})<\infty \),
Then by the basic energy estimate, we get
In view of (3.6) and the previous two displays, another application of Proposition A.2 yields, after enlarging Q and C, the bound (3.33) for \(m=M\). This completes the induction argument and the proof of the lemma. \(\quad \square \)
The typical size of \(|\nabla w_m^\varepsilon |\) is much better than (3.31) gives.
Lemma 3.4
For each \(q\in [2,\infty )\), there exist \(\sigma (q,data)>0\), \(C(q,\mathsf {M},\mathsf {K}_0,\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that, for each \(m\in \{1,\ldots ,n+1\}\),
Proof
We argue by induction in m. Assume that (3.34) holds for \(m\in \{1,\ldots ,M-1\}\) for some \(M\leqq n+1\). By (3.1), in particular the assumed validity of Theorem 1.3 for \(m\leqq n+1\), it suffices to show that
The induction hypothesis yields that
and then the basic energy estimate yields (3.35). \(\quad \square \)
We also require bounds on the homogenized solutions \(\overline{u}\) and \(\overline{w}_1,\ldots ,\overline{w}_{n+2}\). These are consequences of elliptic regularity estimates presented in Appendix D, namely Lemma D.1, which is applicable here because of the assumption (3.1) which ensures that \(\overline{L} \in C^{3+\mathsf {N},\beta }\) for every \(\beta \in (0,1)\). We obtain the existence of \(C(\mathsf {M},\mathrm {data})<\infty \) such that, for every \(m\in \{1,\ldots ,n+1\}\),
In particular, the function \(\overline{\Theta }\) defined in (3.19) is Lipschitz continuous. By the global Meyers estimate, we also have, for some \(\delta (d,\Lambda )>0\) and \(C(\mathsf {M},\mathrm {data})<\infty \), the bound
We may also apply Lemma D.1 to get a bound on \(w_{n+2}\). In view of the merely mesoscopic gap between \(\partial U_{n+2}\) and \(U_{n+3}\), which is much smaller than the macroscopic gaps between \(\partial U_{m}\) and \(U_{m+1}\) for \(m\in \{1,\ldots ,n+1\}\), cf. (3.15) and (3.28), the estimate we obtain is, for every \(\beta \in (0,1)\),
for an exponent \(Q(\beta ,\mathrm {data})<\infty \) which can be explicitly computed rather easily (but for our purposes is not worth the bother) and a constant \(C(\beta ,\mathsf {M},\mathrm {data})<\infty \).
3.4 The Mesoscopic Parameters k and l
Here we enter into a discussion regarding the choice of the parameters k and l which specify the mesoscopic scales. Recall that \(1 \ll 3^k \ll 3^l \ll \varepsilon ^{-1}\). As we will eventually see later in this section, we will estimate the left sides of (3.29) and (3.30) by expressions of the form
where \(Q(\mathrm {data})<\infty \) is a large exponent and \(\alpha (\mathrm {data})>0\) is a small exponent. We then need to choose k and l so that the expression in parentheses is a positive power of \(\varepsilon \). We may do this as follows. First, we may assume throughout for convenience that \(Q\geqq 1 \geqq 4\alpha \). Then we may take care of the first two terms by choosing l so that \(\varepsilon 3^l\) is a “very large” mesoscale: we let l be defined so that
Then we see that, for some \(\beta >0\),
Next, we take care of the last term: since, for some \(\beta >0\),
we can make this smaller than \(\varepsilon ^{\frac{\beta }{2}}\) by taking \(\varepsilon 3^k\) to be a “very small” mesoscale. We take k so that, for \(\beta \) and Q as in the previous display,
From this we deduce that \(\left( \varepsilon 3^l \right) ^\alpha 3^{kQ} \leqq C\varepsilon ^{\frac{\beta }{2}}\). We see that this choice of k also makes the third term inside the parentheses on the right side of (3.40) smaller than a positive power of \(\varepsilon \). With these choices, we obtain that, for some \(\beta >0\),
Throughout the rest of this section, we will allow the exponents \(\alpha \in \left( 0,\tfrac{1}{4}\right] \) and \(Q\in [1,\infty )\) to vary in each occurrence, but will always depend only on \(\mathrm {data}\). Similarly, we let c and C denote positive constants which may vary in each occurrence and whose dependence will be clear from the context.
3.5 The Minimal Scales
Many of the estimates we will use in the proof of Theorem 1.2 are deterministic estimates (that is, the constants C in the estimate are not random) but which are valid only above a minimal scale \(\mathcal {X}\) which satisfies \(\mathcal {X}= \mathcal {O}_\delta (C)\) for some \(\delta (\mathrm {data})>0\) and \(C(\mathsf {M},\mathrm {data})<\infty \). This includes, for instance, the estimates of Theorems 1.1, 1.2 and 1.3 assumed to be valid by our assumption (3.1), as well as Lemmas 2.11, 3.1, 3.2 and 3.4, some estimates like (3.13) which appear in the text, and future estimates such as those of Lemmas 3.5, 3.6 and 3.7. We stress that this list is not exhaustive.
In our proofs of (3.29) and (3.30), we may always suppose that \(3^k \geqq \mathcal {X}\), where \(\mathcal {X}\) is the maximum of all of these minimal scales. In fact, we may suppose that \(3^k\) is larger than the stationary translation \(T_z\mathcal {X}\) of \(\mathcal {X}\) by any element of \(z\in \mathbb {Z}^{d}\cap \varepsilon ^{-1} U_1\). To see why, first we remark that if \(\mathcal {X}\) is any random variable satisfying \(\mathcal {X}= \mathcal {O}_\delta (C)\), then a union bounds gives that, for any \(Q<\infty \) the random variable
satisfies, for a possibly smaller \(\delta \) and larger \(C<\infty \), depending also on Q, the estimate \(\tilde{\mathcal {X}} = \mathcal {O}_{\delta /2}(C)\). Since, as explained in the previous subsection, \(3^k\) is a small but positive power of \(\varepsilon ^{-1}\), we see that by choosing Q suitably we have that \(3^k \geqq \tilde{\mathcal {X}}\) implies the validity of all of our estimates. Finally, in the event that \(3^k < \tilde{\mathcal {X}}\), we can use the deterministic bounds obtained in Lemma 3.3 and (3.38) very crudely as follows:
Then we use that
where in the previous display the exponent Q is larger in the second instance than in the first. This yields (3.29) and (3.30) in the event \(\{ 3^k > \tilde{\mathcal {X}} \}\), so we do not have to worry about this event.
Therefore, throughout the rest of this section, we let \(\mathcal {X}\) denote a minimal scale satisfying \(\mathcal {X}=\mathcal {O}_\delta (C)\) which, in addition to both \(\delta \) and C, may vary in each occurrence.
3.6 The Proof of the Local Stationary Approximation
We now turn to the proof of (3.29), which amounts to showing that the difference between the coefficient fields \((\mathbf {a}^\varepsilon ,\mathbf {F}^\varepsilon _{n+2})\) and \((\tilde{\mathbf {a}}^\varepsilon ,\tilde{\mathbf {F}}^\varepsilon _{n+2})\) is small. This is accomplished by an application of Lemma 2.11.
Lemma 3.5
There exist \(\alpha (\mathrm {data})>0\), \(Q(\mathrm {data})<\infty \) and \(C(\mathsf {M},\mathrm {data})<\infty \) and a minimal scale \(\mathcal {X}=\mathcal {O}_\delta (C)\) such that, if \(3^k\geqq \mathcal {X}\), then
Proof
Throughout \(\mathcal {X}\) denotes a random scale which may change from line to line and satisfies \(\mathcal {X}=\mathcal {O}_\delta (C)\).
For each \(z\in \varepsilon 3^l \mathbb {Z}^{d}\) with \(z+\varepsilon {{\square }}_{l+1}\subseteq U_{n+1}\), we compare \(\nabla u\) and \(\nabla w_m^\varepsilon \) to the functions \(\nabla v^{(l)}_{\nabla \overline{u}(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \) and \(\nabla w^{(l)}_{m,\overline{\Theta }(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \), using Lemma 2.11 and the assumed validity of Theorems 1.2 and 1.3 for n. The latter yields that, if \(\varepsilon ^{-1}\geqq \mathcal {X}\), then for every such z, and every \(m\in \{1,\ldots ,n+1\}\)
Likewise, if \(3^l \geqq \mathcal {X}\), then for every z as above, and every \(m\in \{1,\ldots ,n+1\}\),
Here we used (3.37) in order that the C not depend implicitly on \(\left| \nabla u(z) \right| \) and in order that the prefactor on the right side of the second line be is as it is, rather than \(\sum _{j=1}^m \left| \nabla \overline{w}_j(z) \right| ^{\frac{m}{j}}\). Using (3.37) again, we see that
Combining the three previous displays with the triangle inequality, we get, for all such m and z and provided that \(3^l\geqq \mathcal {X}\),
An application of Lemma 2.11 then yields
By \(L^p\) interpolation and the bounds (3.8) and (3.34), we obtain, for every \(q\in [2,\infty )\), an \(\mathcal {X}\) such that \(3^l\geqq \mathcal {X}\) implies that, for every m and z as above,
This estimate implies
By a very similar argument, comparing the functions \(\nabla v^{(l)}_{\nabla \overline{u}(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \) and \(\nabla w^{(l)}_{m,\overline{\Theta }(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \) to \(\nabla v^{(k)}_{\nabla \overline{u}(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \) and \(\nabla w^{(k)}_{m,\overline{\Theta }(z),\frac{z}{\varepsilon }}\left( \tfrac{\cdot }{\varepsilon }\right) \), also using Lemma 2.11 and the assumed validity of Theorems 1.2 and 1.3 for n, we obtain, for all m and z as above,
Using (3.13) and (3.14), in view of (3.37), we find an exponent \(Q(\mathrm {data})<\infty \) such that, for every m and z as above,
Combining the above estimates, we finally obtain that, for every z and m as above,
Using that these coefficient fields are bounded and that the boundary layer (that is, the set of points which lie in \(U_{n+2}\) but not in any cube of the form \(z+\varepsilon {{\square }}_{l}\) with z as above) has thickness \(O(\varepsilon 3^l)\), we get from the previous estimate that
where (in order to shorten the expressions) we denote
To complete the proof of the lemma, we observe that \(\zeta := \tilde{w}_{n+2}^\varepsilon -w^\varepsilon _{n+2}\in H^1_0(U_{n+2})\) satisfies the equation
By the basic energy estimate (test the equation for \(\zeta \) with \(\zeta \)), we obtain
The first term on the right side is already estimated above. For the second term, we use the Meyers estimate and (3.34), without forgetting (3.20), to find \(\delta (d,\Lambda )>0\) such that
Therefore, by Hölder’s inequality and the above estimate on \(\left\| \mathbf {a}^\varepsilon - \tilde{\mathbf {a}}^\varepsilon \right\| _{\underline{L}^q(z+\varepsilon {{\square }}_l)} \) with exponent \(q:=\frac{4+2\delta }{\delta } \), we obtain
This completes the proof of (3.44). \(\quad \square \)
In view of the discussion in Sections 3.4 and 3.5, Lemma 3.5 implies (3.29).
3.7 Homogenization Estimates for the Approximating Equations
To prepare for the proof of (3.29), we apply the quantitative homogenization estimates proved in [4] for the linear equation (3.4). We denote by \({\overline{\mathbf {a}}}_p^{(k)}\) and \(\overline{\mathbf {F}}^{(k)}_{n+2,\Theta }\) the homogenized coefficients.
By an application of the results of [4, Chapter 11], the solutions \(v(\cdot ,U,e,\Theta )\) of the family of Dirichlet problems
indexed over bounded Lipschitz domains \(U\subseteq {\mathbb {R}^{d}}\) and \(e\in B_1\) and \(\Theta \in \mathbb {R}^{d(n+2)}\), satisfy the following quantitative homogenization estimate: there exist \(\alpha (s,d,\Lambda )>0\), \(\beta (\mathrm {data})>0\), \(\delta (\mathrm {data})>0\) and \(Q(\mathrm {data})<\infty \) and \(C(s,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}= \mathcal {O}_\delta (C)\) such that, for every \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) with \(|p|\leqq \mathsf {M}\), \(e\in {\mathbb {R}^{d}}\) and every \(l\in \mathbb {N}\) with \(3^{l-k} \geqq \mathcal {X}\),
as well as
To obtain (3.47), we change the scale by performing a dilation \(x\mapsto 3^k \left\lceil \left( 15+d\right) ^{\frac{1}{2}} \right\rceil x\) so that the resulting coefficient fields have a unit range of dependence and are still \(\mathbb {Z}^{d}\)-stationary, cf. (3.3). We then apply [4, Lemma 11.11] to obtain (3.47), using Lemma 3.1, namely (3.5), to give a bound on the parameter \(\mathsf {K}\) in the application of the former. The reason we quote results for general nonlinear equations, despite the fact that (3.46) is linear, is because (3.46) has a nonzero right-hand side and the results for linear equations have not been formalized in that generality.
The bound (3.48) is a consequence the large-scale \(C^{0,1}\)-type estimate for the equation (3.4) (see [4, Theorem 11.13]) which, together with a routine union bound, yields (3.48) with the left side replaced by \(\sup _{z \in \mathbb {Z}^{d}\cap {{\square }}_l} \left\| \nabla v \right\| _{\underline{L}^2(z+(l-k){{\square }}_k)}\). Indeed, if we denote by \(\mathcal {X}_z\) the minimal scale in the large-scale \(C^{0,1}\) estimate, then a union bound yields (for any \(s\in (0,d)\), so in particular we can take \(s>1\)),
We obtain (3.48) by combining this bound for \(\sup _{z \in \mathbb {Z}^{d}\cap {{\square }}_l} \left\| \nabla v \right\| _{\underline{L}^2(z+(l-k){{\square }}_k)}\) with the deterministic bounds (3.5), (3.6) and an application of the Schauder estimates (see Proposition A.2).
We also need the following estimate for the difference of v’s on overlapping cubes, which can be inferred from (3.47) and the Caccioppoli inequality: there exist \(\alpha (s,d,\Lambda )>0\), \(\beta (\mathrm {data})>0\), \(\delta (\mathrm {data})>0\) and \(Q(\mathrm {data})<\infty \) and \(C(s,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}= \mathcal {O}_\delta (C)\) such that, for every \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) with \(|p|\leqq \mathsf {M}\), \(e\in {\mathbb {R}^{d}}\) and every \(l\in \mathbb {N}\) with \(3^{l-k} \geqq \mathcal {X}\),
Finally, we mention that we have the following deterministic estimate on the functions \(v(\cdot ,z+{{\square }}_{l+1},e,\Theta )\) which states that
This follows from testing the problem (3.46) with \(U=z+{{\square }}_{l+1}\) with the test function \(v(\cdot ,z+{{\square }}_{l+1},e,\Theta ) - \ell _e\in H^1_0(z+{{\square }}_{l+1})\) and then using (3.9) with \(m=n+2\).
Our next goal is to compare the homogenized coefficients \(\overline{\mathbf {F}}^{(k)}_{m,\Theta }\) for the approximating equations to the functions \(\overline{\mathbf {F}}_m (\Theta )\) defined in (1.16). In view of the results of Section 2, it is natural to proceed by comparing the vector fields \(\mathbf {F}_{m,\Theta }^{(k)}\) to \(\mathbf {f}^{(m)}_{p,h}\) defined in (2.5). This is accomplished by invoking Lemma 2.11 again.
Lemma 3.6
Fix \(q\in [2,\infty )\) and \(\mathsf {M}\in [1,\infty )\). There exist \(\delta (q,\mathrm {data}),\alpha (q,\mathrm {data})\in (0,\frac{1}{2}]\), \(C(q,\mathsf {M}, \text {data})<\infty \) such that, for every \(k\in \mathbb {N}\) with \(3^k\geqq \mathcal {X}\), \(m\in \{ 1,\ldots ,n+2\}\) and \(p,h\in {\mathbb {R}^{d}}\) with \(|p|\leqq \mathsf {M}\), we have
and, for \(\Theta :=(p,h,0,\cdots ,0)\in \mathbb {R}^{d(n+2)}\),
Proof
We take \(\mathcal {X}\) to be the maximum of the random scales in Theorem 2.2, Lemmas 2.10 and 2.11 and Theorems 1.2 and 1.3 for n, the latter two being valid by assumption (3.1). Assume \(3^k \geqq \mathcal {X}\).
By Lemma 2.10 and the assumed validity of Theorem 1.2 for n, we have
and, for \(m\in \{2,\ldots ,n+1\}\),
By Theorem 2.2 and the assumed validity of Theorem 1.3 for n, we also have that, for every \(m\in \{1,\ldots ,n+1\}\),
Using these estimates and Lemma 2.11, we obtain, for \(m\in \{2,\ldots ,n+1\}\),
The previous display holds for \(3^k\geqq \mathcal {X}\). Combining this estimate with (2.16) and (3.10) and using \(L^p\) interpolation, we obtain
The conclusion of the lemma now follows. \(\quad \square \)
We next observe that the homogenized coefficients \({\overline{\mathbf {a}}}^{(k)}_p\) and \(\overline{\mathbf {F}}^{(k)}_{n+2,\Theta }\) agree, up to a small error, with \({\overline{\mathbf {a}}}_p:=D^2\overline{L}(p)\) and \(\overline{\mathbf {F}}_{n+2}(\Theta )\). This will help us eventually to prove that the homogenized coefficients for the linearized equations are the linearized coefficients of the homogenized equation.
Lemma 3.7
Fix \(\mathsf {M}\in [1,\infty )\). There exist \(\alpha (\mathrm {data})>0\) and \(C(n,\mathsf {M},\mathrm {data})<\infty \) such that, for every \(\Theta =(p,h_1,\ldots ,h_{n+1}) \in B_\mathsf {M} \times \mathbb {R}^{d(n+1)}\),
and
Proof
We first give the argument in the case that \(\Theta =(p,h,0,\ldots ,0)\). From the definition (1.16) and Proposition 2.8, we see that \(\left( {\overline{\mathbf {a}}}_p,\overline{\mathbf {F}}_{n+2}(\Theta ) \right) \) are the homogenized coefficients for the stationary coefficient fields \(\left( \mathbf {a}_p, \mathbf {f}_{p,h}^{(n+2)} \right) \). By Lemma 3.6, and the bounds (3.5) and (2.16), we find that, for every \(q\in [2,\infty )\),
and
with the constants C depending additionally on q. The result now follows easily by the Meyers estimate in the case \(\Theta =(p,h,0,\ldots ,0)\). Indeed, by subtracting the equations and applying Hölder’s inequality and the Meyers estimate, we can show that the first-order correctors for the two linear problems are close in the \(\underline{L}^2({{\square }}_k)\) norm (with a second moment in expectation). Therefore their fluxes are also close, and taking the expectations of the fluxes gives the homogenized coefficients. This argument can also be performed in a finite volume box \({{\square }}_M\) with the result obtained after sending \(M\rightarrow \infty \). See [1, Lemma 3.2] for a similar argument.
For general \(\Theta \) the result follows by polarization (see Lemma 2.6) and the multilinear structure shared by the functions \(\Theta \mapsto \overline{\mathbf {F}}^{(k)}_{n+2,\Theta }\) and \(\Theta \mapsto \overline{\mathbf {F}}_{n+2}(\Theta )\). \(\quad \square \)
We also require the following continuity estimate for the functions \(w(\cdot ,{{\square }}_l,e,\Theta )\) in \((e,\Theta )\). Following the proof of Lemma 3.2 for \(L^2\) dependence in \(\Theta \) (we need need to do one more step of the iteration described there), using (3.47) for \(L^2\) dependence in e, and then interpolating this result with (3.48), we obtain exponents \(\alpha (\mathrm {data})>0\) and \(Q(\mathrm {data})<\infty \) and a random variable \(\mathcal {X}= \mathcal {O}_\delta (C)\) such that, for every \(e,e'\in {\mathbb {R}^{d}}\) and \(\Theta =(p,h_1,\ldots ,h_{n+1})\in \mathbb {R}^{d(n+2)}\) and \(\Theta '=(p',h_{1}',\ldots ,h_{n+1}')\in \mathbb {R}^{d(n+2)}\) with \(|p|,|p'|\leqq \mathsf {M}\), if \(3^k\geqq \mathcal {X}\), then
3.8 Homogenization of the Locally Stationary Equation
We next present the proof of (3.30), which is the final step in the proof of Theorem 1.2. The argument follows a fairly routine (albeit technical) two-scale expansion argument and requires the homogenization estimates given in Section 3.7.
The proof of (3.30)
In view of the discussion in Section 3.5, we may assume throughout that \(3^k\geqq \mathcal {X}\) where \(\mathcal {X}\) is any minimal scale described above. We begin by building an approximation of the solution \(\tilde{w}_{n+2}^\varepsilon \) of (3.25) built out of the solutions of (3.46). We select a smooth function \(\zeta \in C^\infty _c( {\mathbb {R}^{d}})\) such that, for some \(C(d)<\infty \) and each \(i\in \{1,2\}\),
We can construct such a \(\zeta \) by a suitable mollification of \(\mathbb {1}_{\varepsilon {{\square }}_{l}}\). We write \(\zeta _z:= \zeta (\cdot -z)\) for short. We next define a function \(\tilde{v}^\varepsilon _{n+2} \in H^1_0(U_{n+2})\) by
Since \(\tilde{v}_{n+2}^\varepsilon - \tilde{w}_{n+2}^\varepsilon \in H^1_0(U_{n+2})\), We have that
A preliminary goal is therefore to prove the estimate
The proof of (3.63) is essentially completed in Steps 1–5 below: see (3.72) and (3.73). After obtaining this estimate, we will prove that
Together these inequalities imply (3.30).
Denote \(v_z:= v\left( \cdot , \tfrac{z}{\varepsilon }+ {{\square }}_{l+1},\nabla \overline{w}_{n+2}(z),\overline{\Theta }(z) \right) \) for short and compute
Thus
Using the equation for \(v_z\), we find that
Therefore, we have that
We next estimate each of the five terms on the right side of the previous display.
Step 1. The estimate of the first term on the right side of (3.67). This is where we use the homogenized equation for \(\overline{w}_{n+2}\). Fix \(h\in H^1_0(U_{n+2})\) with \(\left\| h \right\| _{H^1(U_{n+2})}=1\). By the equation for \(\overline{w}_{n+2}\) and integration by parts that
Therefore we may have
For the first term on the right, we use (3.37), (3.47), (3.56) and (3.57) to obtain
For the second term, we denote
observe that \(\left| U_{n+2}{\setminus } U_{n+4} \right| \leqq C \varepsilon 3^{l}\), and compute, using the Hölder inequality, to obtain, using also (3.38),
Combining the above estimates and taking the supremum over \(h\in H^1_0(U_{n+2})\) with \(\Vert h \Vert _{H^1(U_{n+2})}=1\) yields
Step 2. The estimate of the second term on the right side of (3.67). We use (3.14), (3.37) and (3.50) to find
Step 3. The estimate of the third term on the right side of (3.67). We have by (3.13) and (3.37) that
Step 4. The estimate of the fourth term on the right side of (3.67). For convenience, extend \(v_z\) so that it is equal to the affine function \(\ell _{\nabla \overline{w}_{n+2}(z)}\) outside of the cube \(z+\varepsilon {{\square }}_{l+1}\). By the triangle inequality, we have
where here and below we use the notation \(y\sim z\) to mean that \(y+{{\square }}_{l+1}\) and \(z+{{\square }}_{l+1}\) have nonempty intersection. Using (3.61) and the fact that \(\sum _{z\in \varepsilon 3^l\mathbb {Z}^{d}} \nabla \zeta _z = 0\), we have that
where
is a boundary layer of \(U_{n+3}\) of thickness \(\varepsilon 3^l\). Observe that \(|V| \leqq C\varepsilon 3^l\). Thus by (3.47) we find that
Combining (3.61) with (3.49), (3.37) and the triangle inequality to compare the \(v_z\)’s on overlapping cubes, we find that
For the last term, we compute, using (3.39),
Putting the above inequalities together, we find that
Step 5. The estimate of the fifth term on the right side of (3.67). Recall that \(U_{n+4}\) is defined in (3.68). We have that, using (3.38) and (3.39),
Step 6. Let us summarize what we have shown so far. Substituting the results of the five previous steps into (3.67), we obtain that
This implies by (3.62) that
Therefore to obtain the lemma it suffices to prove (3.64). To prove this bound, we use the identity
which, combined with (3.47), and (3.37), yields
This is (3.64).
In view of the selection of the mesoscopic parameters k and l in Section 3.4, which gives us (3.41), the proof of (3.30) is now complete. \(\quad \square \)
4 Large-Scale \(C^{0,1}\)-Type Estimates for Linearized Equations
In the next two sections, we suppose that \(n\in \{0,\ldots ,\mathsf {N}-1 \}\) is such that
The goal is to prove that Theorem 1.3 is also valid for \(n+1\). Combined with the results of the previous two sections and an induction argument, this completes the proof of Theorems 1.1, 1.2 and 1.3.
The goal of this section is to prove the half of Theorem 1.3, namely estimate (1.22) for \(w_{n+2}\). The estimate (1.23) will be the focus of Section 5.
Both of the estimate in the conditional proof of Theorem 1.3 are obtained by “harmonic” approximation: homogenization says that on large scales the heterogeneous equations behave like the homogenized equations, and therefore we may expect the former to inherit some of the better regularity estimates of the latter. The quantitative version of the homogenization statement provided by Theorem 1.2 allows us to prove a \(C^{0,1}\)-type estimate, following a well-known excess decay argument originally introduced in [6].
4.1 Approximation of \(w_{n+2}\) by Smooth Functions
The large-scale regularity estimates are obtained by approximating the solutions of the linearized equations for the heterogeneous Lagrangian L, as well as the linearization errors, by the solutions of the linearized equations for the homogenized Lagrangian \(\overline{L}\). Since the latter possess better smoothness properties, this allows us to implement an excess decay iteration for the former.
We begin by giving a quantitative statement concerning the smoothness of solutions to the linearized equations for the homogenized operator \(\overline{L}\). This is essentially well-known, but we need a more precise statement than we could find in the literature.
We next present the statement concerning the approximating of the solutions \(w_m\) of the linearized equations for L, as well as the linearized errors \(\xi _m\), by solutions of the linearized equations for \(\overline{L}\). For the \(w_m\)’s, this is essentially a rephrasing of the assumed validity of Theorem 1.2 for \(n+1\) in place of \(\mathsf {N}\), see (4.1). For the \(\xi _m\)’s, this can be thought of as a homogenization result.
Lemma 4.1
(Smooth approximation of \(w_{n+2}\)). Assume that (4.1) holds. Fix \(\mathsf {M} \in [1,\infty )\). There exist \(\delta (n,d,\Lambda )\in (0,d)\), \(\alpha (d,\Lambda ) \in \left( 0,\tfrac{1}{2}\right] \), \(C(n,\mathsf {M},d,\Lambda )<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid. For every \(R\geqq \mathcal {X}\) and \(u,v,w_1,\ldots ,w_{n+2} \in H^1(B_R)\) satisfying, for each \(m\in \{1,\ldots ,n+2\}\),
if we let \(\overline{u},\overline{v},\overline{w}_1,\ldots ,\overline{w}_{n+2} \in H^1(B_{R/2})\) be the solutions of the Dirichlet problems
then we have, for each \(m \in \{1,\ldots ,n+2\}\), the estimate
Proof
Denote, in short, \(R_m := \frac{1}{2} (1+2^{-m}) R\) and \(r_m := \frac{1}{4} (R_m - R_{m-1})\). Since we assume (4.1), we have that Theorem 1.2 applied for \(n+1\) instead of \(\mathsf {N}\) and Theorem 1.3, assumed now for n in place of \(\mathsf {N}\), are both valid. In particular, there is \(\sigma (n,\mathrm {data}) \in (0,1)\) and \(C(n,\mathsf {M},\mathrm {data})<\infty \), a random variable \(\tilde{\mathcal {X}} \leqq \mathcal {O}_\sigma (C)\) such that, for \(R \geqq \tilde{\mathcal {X}}\) and \(m \in \{1,\ldots ,n+2\}\), Theorem 1.3 gives
and Theorem 1.2 yields
We set
Clearly \(\mathcal {X}= \mathcal {O}_{\frac{1}{2} \alpha \sigma }(C)\) and \(\mathcal {X}\geqq \tilde{\mathcal {X}}\), and if \(R \geqq \mathcal {X}\), then \( \tilde{\mathcal {X}} R^{-\alpha } \leqq R^{-\frac{1}{2} \alpha }\). In conclusion, by taking \(\alpha \) smaller if necessary, we obtain by (4.6) that, for \(m \in \{1,\ldots ,n+2\}\) and \(R \geqq \mathcal {X}\),
Furthermore, we notice that (2.7) yields
and thus we get, by (4.5) and (4.7), that
It then follows by the Meyers estimate and equation of \(w_m\) that
finishing the proof of (4.4) by (4.7). \(\quad \square \)
4.2 Excess Decay Iteration for \(w_{n+2}\)
In this subsection we conditionally prove the statement of Theorem 1.3 for \(n+1\) and for \(q=2\). The restriction on q will be removed in the next subsection. The proof is by a decay of excess iteration, following along similar lines as the argument from [6], using “harmonic” approximation. The statement we prove is summarized in the following proposition:
Proposition 4.2
Assume that (4.1) holds. Fix \(\mathsf {M} \in [1,\infty )\). Then there exist constants \(\sigma (\mathrm {data}),\alpha (d,\Lambda ) \in \left( 0,\frac{1}{2} \right] \), \(C(\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid: let \(R \in [\mathcal {X}, \infty )\) and \(u,w_1,\ldots ,w_{n}\) satisfy, for each \(m\in \{1,\ldots ,n+1\}\),
where u satisfy the normalization
Then, for \(m \in \{1,\ldots ,n+2\}\) and for every \(r \in [\mathcal {X},\tfrac{1}{2} R]\), we have
Remark 4.3
Proposition 4.2 holds for \(n=0\) by [1, Proposition 4.3] without assuming (4.1).
Proof
The proof is based on combination of harmonic approximation and decay estimates for homogenized solutions presented in Appendix D. The necessary estimate is (D.7). We take the minimal scale \(\mathcal {X}\) as the maximum of the minimal scale in Lemma 4.1 and in Theorem 1.3, which is valid with n in place of \(\mathsf {N}\), corresponding \(q =2 \), and a constant \(\mathsf {R}(\mathsf {M},\mathrm {data}) \in [1,\infty )\) to be fixed in the course of the proof. This choice, in particular, implies that there exist constants \(C(\mathsf {M},\mathrm {data})<\infty \) and \(\sigma (\mathrm {data}) \in \left( 0 ,\frac{1}{2} \right] \) such that \(\mathcal {X}\leqq \mathcal {O}_{\sigma }(C)\) and \(\mathcal {X}\geqq \mathsf {R}\).
We will prove the statement using induction in shrinking radii. Indeed, we set, for \(j \in \mathbb {N}\) and \(\theta \in (0,\frac{1}{2}]\), \(r_j := \theta ^{j} r_0 \), where \(r_0 \in [\mathcal {X}, \delta R]\) and \(\delta \in \left( 0 , \tfrac{1}{2} \right] \). Parameters \(\theta \), \(\delta \) and \(\mathsf {R}\) will all be fixed in the course of the proof. Having fixed \(\theta \), \(\delta \) and \(\mathsf {R}\), we assume that there is \(J \in \mathbb {N}\) such that \(r_{J+1} < \mathcal {X}\leqq r_J\) for some \(J \in \mathbb {N}\). If there is no such J or \(\mathcal {X}\geqq \delta R\), the result will follow simply by giving up a volume factor. Furthermore, we device the notation of this proof in such a way that it will also allow us to prove the result of the next lemma, Lemma 4.4.
We denote, in short, for \(m \in \{1,\ldots ,n+2\}\) and \(\gamma \in [0,1)\) to be fixed,
and
Theorem 1.3 implies that, for \(r \in [\mathcal {X}, \tfrac{1}{2} R]\) and \(m \in \{1,\ldots ,n+1\}\), we have that
Notice then that, by (4.12) and Poincaré’s inequality, we have, for \(r \in [\mathcal {X}, \tfrac{1}{2} R]\) and \(m\in \{1,\ldots ,n+2\}\), that
Step 1. Letting \(\overline{u}_j\) solve
we show that there exist, for \(\eta \in (0,1]\), constants \( \alpha _1(d,\Lambda ) \in (0,\frac{1}{2})\), \(\delta (\eta ,\mathsf {M},\mathrm {data})<\infty \) and \(\mathsf {R}(\eta ,\mathsf {M},\mathrm {data})<\infty \) such that, for \(j \in \{0,\ldots ,J\}\),
The parameter \(\eta \) shall be fixed later in (4.19). On the one hand, we have from [1, Theorem 2.3] that there exist constants \(\beta _1(d,\Lambda ) \in (0,1)\) and \(C(\mathsf {M},d,\Lambda )<\infty \) such that
On the other hand, by harmonic approximation ([1, Corollary 2.2]) and Lipschitz estimate for u ([1, Theorem 2.3]) we get that there exist constants \(\beta _2(d,\Lambda ) \in (0,1)\) and \(C(\mathsf {M},d,\Lambda )<\infty \) such that
Thus (4.14) follows by the triangle inequality by taking \(\alpha _1 := \tfrac{1}{2} (\beta _1 \wedge \beta _2)\), and choosing \(\delta \) small enough and \(\mathsf {R}\) large enough so that
We assume, from now on, that \(\delta \) and \(\mathsf {R}\) are such that (4.15) is valid.
Step 2. Letting \(j \in \{0,\ldots ,J\}\) and \(m\in \{1,\ldots ,n+2\}\), and \(\overline{w}_{1,j},\ldots ,\overline{w}_{m,j}\) to solve, with \(\overline{u}_j\) as in Step 2, equations
we show that then there is \(\alpha _2(\mathrm {data}) \in \left( 0,\tfrac{1}{2}\right] \) such that
This, however, is a direct consequence of (4.12) and Lemma 4.1.
Step 3. Induction assumption. Set \(\alpha := \frac{\beta }{2} (\alpha _1 \wedge \alpha _2)\), where \(\alpha _1\) and \(\alpha _2\) are as in Steps 1 and 2, respectively, and \(\beta \) comes from the \(C^{n+2,\beta }\) regularity of \(\overline{L}\). Let \(\delta _j\) be defined as
We assume inductively that, for \(j^* \in \{1,\ldots ,J\}\), \(j \in \{0,\ldots ,j^*\}\), and \(m \in \{1,\ldots ,n+2\}\), we have, for a constant \(\mathsf {C} \in [1,\infty )\) to be fixed in Step 5, that
Here \(\mathsf {D}_m \) and \(\mathsf {E}_m \) are defined in (4.10) and (4.11), respectively. Obviously (4.18) is valid for \(j=0\) by the definitions of \(\mathsf {D}_m \) and \(\mathsf {E}_m \). Fixing
we have that (4.14) implies
Using (4.16) and (4.18) as well, we obtain, for \(m \in \{1,\ldots ,n+1\}\),
provided that
We assume, from now on, that \(\mathsf {R}\) is such that both (4.15) and (4.22) are valid.
Step 4. We show that the first inequality in (4.18) continues to hold for \(j = j^*+1\). First, by the triangle inequality, (4.21) and (4.18), we have that
and
By a similar computation, using also the induction assumption (4.18),
Now, applying (D.7) we obtain by the previous three displays, (4.20) and (4.13), for each \(m \in \{1,\ldots ,n+2\}\), that
and, consequently,
Thus, choosing \(\theta \) small enough so that \(C \theta ^{\frac{\beta }{2}} \leqq \frac{1}{2}\), we obtain that the first inequality in (4.18) is valid for \(j= j^*+1\).
Step 5. The last step in the proof is to show the second inequality in (4.18) for \(j = j^*+1\). Let \(\ell _j\) be the minimizing affine function in \( \inf _{\ell \in \mathcal {P}_1} \left\| w_{m} - \ell \right\| _{\underline{L}^2(B_{r_j})}\). Then, by the triangle inequality and the first inequality in (4.18) valid for \(j \in \{0,\ldots ,j^*\}\),
Thus, summation gives that
Therefore,
By the triangle inequality we have that
and hence
Choosing \(\mathsf {C} = C\), where C is as in the above inequality, verifies the second inequality in (4.18) for \(j = j^*+1\). This finishes the proof of the induction step, and thus completes the proof. \(\quad \square \)
To show Lipschitz estimates for the linearization errors in the next section, we need a small variant of the previous proposition.
Lemma 4.4
Assume that (4.1) holds. Fix \(\mathsf {M} \in [1,\infty )\). Then there exist constants \(\alpha (\mathrm {data}),\sigma (n,\mathsf {M},\mathrm {data}), \theta (n,\mathsf {M},\mathrm {data}) \in \left( 0, \frac{1}{2} \right] \) and \(C(\sigma ,\mathsf {M},\mathrm {data})<\infty \), and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid. Let \(R \in [\mathcal {X}, \infty )\) and \(u,w_1,\ldots ,w_{n}\) satisfy, for each \(m\in \{1,\ldots ,n+1\}\),
and, for \(r \in [\mathcal {X}, \frac{1}{2} R]\),
where u satisfy the normalization
Then
Proof
The proof is a rearrangement of the argument in the proof of Proposition 4.2. Indeed, we take \(r_0=r\) and combine the first inequality of (4.23) with (4.16) and (4.9), which is valid for \(m\in \{1,\ldots ,n+1\}\). \(\quad \square \)
4.3 Improvement of Spatial Integrability
We next complete the conditional proof of (1.22) by improving the spatial integrability of (4.25) from \(L^2\) to \(L^q\) for every \(q\in [2,\infty )\). To do this, we use the estimate (4.25) to pass from the large scale R to the microscopic, random scale \(\mathcal {X}\). We then use deterministic estimates from classical elliptic regularity theory to obtain local \(L^q\) estimates in balls of radius one, picking up a volume factor—which is power of \(\mathcal {X}\)—as a price to pay. The first formalize the latter step in the following lemma:
Lemma 4.5
Assume (4.1) and the hypotheses of Theorem 1.3. Let \(\beta \in (0,1)\) and \(q\in (2,\infty )\). Then there exist \(\sigma (q,\mathrm {data}) \in (0,d)\), \(C(q,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}= \mathcal {O}_\sigma (C)\) such that, for every \(r\geqq \mathcal {X}\) and \(m\in \{1,\ldots ,n+2\}\),
Proof
In view of the assumption of (4.1) and thus the validity of Theorem 1.3 for n, we only need to prove (4.26) for \(m=n+2\). Fix \(q\in (2,\infty )\), \(r\in [2,\infty )\), \(\beta \in (0,1)\) to be selected below. The \(C^{1,\beta }\)-estimate in Proposition A.2, together with a covering argument, yields
Setting \(\mathbf {a}:=D^2_pL(\nabla u,x)\), we deduce by the assumption of (L1) that
Applying Proposition A.1 we find that, for each \(x\in B_{r/2}\) and \(q\in (2,\infty ]\),
By a covering argument, we therefore obtain
If we now take \(\mathcal {X}\) to be the maximum of the minimal scales in the statements of:
- (1)
[4, Theorem 11.13];
- (2)
Theorem 1.3 for n in place of \(\mathsf {N}\) and with a sufficiently large exponent of spatial integrability \(q'\) in place of q (which can be computed explicitly in terms of our q using the Hölder inequality, although we omit this computation)—the validity of which is given by assumption (4.1);
- (3)
Proposition 4.2;
then we have that \(\mathcal {X}= \mathcal {O}_\sigma (C)\) as stated in the lemma and that \(r \geqq \mathcal {X}\) implies the following estimates:
and
Combining these with (4.27), we obtain
This completes the proof of the lemma. \(\quad \square \)
In the next lemma we finally achieve the goal of this section, which is to show that (4.1) implies (1.22) for \(m=n+1\).
Lemma 4.6
Assume that (4.1) holds. Fix \(\mathsf {M} \in [1,\infty )\) and \(q\in (2,\infty )\). Then there exist constants \(\sigma (q,\mathrm {data}) \in (0,d)\), \(C(q,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following is valid. Suppose that \(R \in [\mathcal {X}, \infty )\) and \(u,w_1,\ldots ,w_{n} \in H^1(B_R)\) such that, for every \(m\in \{1,\ldots ,n+2\}\),
Then, for every \(r \in [\mathcal {X},\tfrac{1}{2} R]\), we have
Proof
Fix \(q\in (2,\infty )\). Select a parameter \(\theta \in (0,1)\) which will denote a mesoscopic scale. For each \(z\in {\mathbb {R}^{d}}\), we take \(\mathcal {X}_z\) to be the random variable \(\mathcal {X}\) in the statement of Proposition 4.2, centered at the point z. Define another random variable (a minimal scale) by
It is clear from Proposition 4.2 and an routine union bound argument that
Next, for every \(k\in \mathbb {N}\) and \(z\in {\mathbb {R}^{d}}\) we let \(\mathcal {Z}_{k,z}\) denote the random variable
where the supremum is taken over all \((u,w_1,\ldots ,w_{n+2})\in \left( H^1(z+{{\square }}_{k+1})\right) ^{n+3}\) satisfying, for every \(m\in \{1,\ldots ,n+2\}\),
Observe that \(\mathcal {Z}_{k,z}\) is \(\mathcal {F}(z+{{\square }}_{k+1})\)-measurable and, by Lemma 4.5 and an easy covering argument, it satisfies the estimate
Fix \(\mathsf {A}\in [1,\infty )\) and define another random variable (a minimal scale) \(\mathcal {Z}\) by
We will show below that, if \(\mathsf {A}\) is chosen sufficiently large (depending of course on the appropriate parameters) then
Assuming that (4.32) holds for the moment, let us finish the proof of the lemma. Suppose now that \(k\in \mathbb {N}\) satisfies \(\mathcal {Y} \vee \mathcal {Z} \leqq 3^k \leqq 3^{k+1}\leqq R\). Let \((u,w_1,\ldots ,w_{n+2})\in \left( H^1(B_R)\right) ^{n+3}\) satisfy (4.28). Then we have that
Note that in the third and final lines we used that \(3^k\geqq \mathcal {Y}\), that is, we used the result of Proposition 4.2. This is the desired estimate for \(\mathcal {X}= \mathcal {Y}\vee \mathcal {Z}\), and so the proof of the lemma is complete subject to the verification of (4.32).
Turning to the proof of (4.32), we notice first that it suffices to show, for \(\mathsf {A}\) sufficiently large, the existence of \(\sigma (q,\mathrm {data})>0\) and \(C(q,\mathsf {M},\mathrm {data})<\infty \) such that, for every \(k\in \mathbb {N}\),
Indeed, we can see that (4.33) implies (4.32) using a simple union bound. Fix then a parameter \(\lambda \in [1,\infty )\) and compute, using (4.31) and the Chebyshev inequality,
Using the simple large deviations bound for sums of bounded, independent random variables, we have
Here we are careful to notice that, while the collection \(\{ \mathcal {Z}_{\lceil \theta k \rceil ,z}^q \wedge \lambda : z\in 3^{\lceil \theta k \rceil }\mathbb {Z}^{d}\cap {{\square }}_{k+1} \}\) of random variables is not independent (since adjacent cubes are touching and thus not separated by a unit distance), we can partition this collection into \(3^d\) many subcollections which have an equal number of elements and each of which is independent. The large deviations estimate can then be applied to each subcollection, and then a union bound yields (4.35). Combining (4.34), (4.35) and the observation that \(\mathbb {E}\left[ \mathcal {Z}_{\lceil \theta k \rceil ,z}^q \wedge \lambda \right] \leqq \mathbb {E}\left[ \mathcal {Z}_{\lceil \theta k \rceil ,z}^q\right] \leqq C\) by (4.31), we obtain
Taking \(\lambda := 3^{\frac{d}{4}(1-\theta )k}\) and \(\mathsf {A}^q:=C+1\) yields (4.33). \(\quad \square \)
The above proof, in view of Remark 4.3, gives the following result without assuming (4.1). This, together with (5.1) below, serves as the base case for the induction.
Proposition 4.7
Let \(q\in [2,\infty )\), \(\mathsf {M}\in [1,\infty )\). Then there exist \(\sigma (q,\mathrm {data})>0\) and \(C(q,\mathsf {M},\mathrm {data}) <\infty \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}\leqq \mathcal {O}_{\sigma }\left( C \right) \) such that the following statement is valid. For \(R \in \left[ 2\mathcal {X},\infty \right) \) and \(u,w_1 \in H^1(B_R)\) satisfying \(\left\| \nabla u \right\| _{\underline{L}^2(B_R)} \leqq \mathsf {M}\) and
we have, for all \(r\in [\mathcal {X},\tfrac{1}{2} R]\),
5 Large-Scale \(C^{0,1}\)-Type Estimate for Linearization Errors
In this section we continue to suppose that \(n\in \{0,\ldots ,\mathsf {N}-1 \}\) is such that (4.1) holds. The goal is to complete the proof that Theorem 1.3 is also valid for \(n+1\). Combined with the results of the previous three sections and an induction argument, this completes the proof of Theorems 1.1, 1.2 and 1.3.
5.1 Excess Decay Iteration for \(\xi _{n+1}\)
We start by proving higher integrability for a difference of two solutions. This, together with Proposition 4.7, yields the base case for the induction.
Proposition 5.1
Fix \(\mathsf {M} \in [1,\infty )\) and \(q \in [2,\infty )\). There exist \(\alpha (\mathrm {data}),\sigma (q,\mathrm {data}) \in \left( 0,\tfrac{1}{2}\right] \), \(C(q,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following statement is vali: for every \(R\geqq \mathcal {X}\) and \(u,v \in H^1(B_R)\) satisfying, for each \(m\in \{1,\ldots ,n+1\}\),
Then, for \(m\in \{1,\ldots ,n+1\}\), \(\xi _0 = u-v\) and \(r \in [\mathcal {X}, \tfrac{1}{2} R]\), we have
Proof
On the one hand, the estimate
follows by [1, Proposition 4.2]. On the other hand, the proof of
is similar to the proof of \(L^q\)-integrability of \(w_1\) presented in Section 4.3. Indeed, noticing that \(\xi _0\) satisfies the equation
we have by the normalization in (5.1) and \(C^{1,\alpha }\) regularity of u and v that we may replace \(w_1\) with \(\xi _0\) in Lemma 4.5 applied with \(m=1\). Using this together with the Lipschitz estimate (5.3) for \(\xi _0\) as in the proof of Lemma 4.6, concludes the proof of (5.4). We omit further details. \(\quad \square \)
Proposition 5.2
Assume that (4.1) holds. Fix \(\mathsf {M} \in [1,\infty )\). There exist constants \(\alpha (n,\mathrm {data}),\sigma (n,\mathrm {data}) \in \left( 0,\tfrac{1}{2}\right] \), \(C(n,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying
such that the following statement is valid. For every \(R\geqq \mathcal {X}\) and \(u,v,w_1,\ldots ,w_{n+1} \in H^1(B_R)\) solving, for each \(m\in \{1,\ldots ,n+1\}\),
we have, for \(m\in \{1,\ldots ,n+1\}\) and \(r \in [\mathcal {X}, \tfrac{1}{2} R]\), the estimate
Proof
We start by fixing some notation. Let \(q(n,d,\Lambda )\) be as in Lemma C.1, applied for \(n+1\) instead of n. Corresponding this q, choose \(\mathcal {X}\) to be the maximum of minimum scales appearing in Proposition 5.1, Theorem 1.3 and Lemma 4.4, of which last two are valid for \(n+1\) in place of \(\mathsf {N}\) by (4.1). We also assume that \(\mathcal {X}\geqq \mathsf {R}\), by taking \(\mathcal {X}\vee \mathsf {R}\) instead of \(\mathcal {X}\), where \(\mathsf {R}\) will be fixed in the course of the proof to depend on parameters \((n,\mathsf {M},\mathrm {data})\). Chosen this way, \(\mathcal {X}\) satisfies \(\mathcal {X}\leqq \mathcal {O}_\sigma (C)\) for some constants \(\sigma (n,\mathrm {data})>0\) and \(C(n,\mathsf {M},\mathrm {data})<\infty \).
Let \(r_j = \theta ^{j} \eta R\), where \(\theta \) is as in Lemma 4.4 and \(\eta \in \left( 0,\frac{1}{2} \right] \). The constant \(\eta \), as well as \(\mathsf {R}\), will be fixed in the course of the proof, so that \(\eta \) is small and \(\mathsf {R}\) is large. We track down the dependencies on \(\eta \) and \(\mathsf {R}\) carefully and, in particular, constants denoted by C below do not depend on them. We may assume, without loss of generality, that \(\eta R \geqq \mathcal {X}\).
Set, for \(k \in \{0,1,\ldots ,n\}\),
We denote
Step 1. Induction on degree. We assume that, for fixed \(k \in \{0,\ldots ,n\}\), we have, for every \(m \in \{0,\ldots ,k\}\) and every \(r \in [\mathcal {X}, R_m]\),
Notice that if \(k=0\), then (5.8) follows by Proposition 5.1.
Step 2. Cacciopppoli estimate for \(\xi _{k+1}\). We show that, for all \(r \in [\mathcal {X}, R_{k}]\),
We first have by (C.2) that
By Theorem 1.3 and the choice of q in the beginning of the proof, we obtain, for every \(r \in \left[ \mathcal {X}, \tfrac{1}{2} R \right] \) and \(m\in \{1,\ldots ,k+1\}\), the estimates
and, by Proposition 5.1,
Combining above displays yields (5.9) in view of the induction assumption, that is, that (5.8) holds for \(m \in \{0,\ldots ,k\}\).
Step 3. We prove that there is a constant \(C<\infty \) independent of \(\eta \) and \(\mathsf {R}\) such that, for \(j \in \mathbb {N}_0\) such that \(r_j \geqq 2 (\mathcal {X}\vee \mathsf {R})\), we have
where
and \(\alpha (\mathrm {data})\) is the minimum of parameter \(\alpha \) appearing in statements of Proposition 5.1 and Lemma 4.4.
Let us fix \(j \in \mathbb {N}_0\) such that \(r_j \geqq 2 (\mathcal {X}\vee \mathsf {R})\). We argue in two cases, namely, either (5.13) or (5.15) below is valid. We prove that in both cases (5.11) follows.
Step 3.1. We first assume that
where \(\varepsilon _j \) has been fixed in (5.12). We show that this implies
Notice that this gives (5.11). To show (5.14), we have by the triangle inequality that
By the choice of \(\alpha \), we get by Proposition 5.1 that
and, by Proposition 4.2,
Combining the estimates and using (5.13), we have that
We then obtain (5.14) by the choice of \(\varepsilon _j\) in (5.12), provided that (5.13) is valid.
Step 3.2. We then assume that (5.13) is not the case, that is,
holds with \(\varepsilon _j\) defined in (5.12). We validate (5.11) also in this case. To this end, let us first observe an immediate consequence of (5.15). By Young’s inequality, we have
and
Hence we get, by the definition of \(\mathsf {E}_{k}\) in (5.7) and reabsorption, that
Let then \({w}_{m+2,j}\) solve
It follows from Lemma C.1, together with (5.2), (5.10) and (5.8), assumed inductively for \(m \in \{1,\ldots ,k\}\), that
Consequently, since \( \xi _{k+1} - w_{k+2,j }\in H_0^1(B_{r_j})\), (5.16) yields
By Lemma 4.4 we have
Combining this with (5.17) and the triangle inequality yields (5.11).
Step 4. We show that, for \(r \in [ 2 (\mathcal {X}_\sigma \vee \mathsf {R}),r_0]\) we have
We proceed inductively, and assume \(J \in \mathbb {N}_0\) is such that \(r_J \geqq 2(\mathcal {X}\vee \mathsf {R})\) and for all \(j \in \{0,\ldots ,J\}\) we assume that there exists a constant \(\mathsf {C}(d,\theta ) \in [1,\infty )\), independent of \(\eta \) and \(\mathsf {R}\), such that
This is true for \(J=0\) by the definition of \(\mathsf {E}_{k}\). We claim that it continues to hold for \(j=J+1\) as well. Combining (5.14) and (5.11) with the induction assumption we have, for \(j \in \{1,\ldots ,J\}\), that
Since \(r_J \geqq \mathsf {R}\), we may take \(\mathsf {R}\) large and \(\eta \) small enough so that
Thus, by summing and reabsorbing,
Letting \(\ell _j\) being the minimizing affine function in \( \inf _{\ell \in \mathcal {P}_1} \left\| \xi _{k+1} - \ell \right\| _{\underline{L}^2(B_{r_{j}})}\), we obtain by the above display and the triangle inequality that
By the triangle inequality again, we obtain that
and, consequently, for \(C(d,\theta ) <\infty \),
We thus obtain by the triangle inequality, together with (5.20) and the above display, that, there exists \(C(d,\theta ) <\infty \) such that
Hence we can take \(\mathsf {C} = C\), proving the induction step. Letting then J be such that \(r \in (r_{J+1},r_{J}]\), we obtain (5.19) by giving up a volume factor.
Step 5. Conclusion. To conclude, we obtain from (5.11) and (5.19) by iterating that
We hence find \(\alpha \) such that, for all \(r \in [2 (\mathcal {X}\wedge \mathsf {R} , r_0]\),
Moreover, by (5.19) and (5.9) we deduce that, for all \(r \in [2 (\mathcal {X}\wedge \mathsf {R}) , R_{k+1}]\),
Therefore, (5.8) is valid for \(m= k+1\), by giving up a volume factor. This proves the induction step and completes the proof. \(\quad \square \)
5.2 Improvement of Spatial Integrability
Following the strategy in Section 4.3, we improve the spatial integrability of (5.6) from \(L^2\) to \(L^q\) for every \(q\in [2,\infty )\). Fix \(q \in [2,\infty )\). Now, \(\xi _{n+1}\) solves
where \( \mathbf {E}_{n+1} \) satisfies the estimate (C.13) for \(\delta = q\) and \(n+1\) instead of n. Recall that both (1.22) and (1.23) are valid for \(m \in \{1,\ldots ,n\}\) with 2nq instead of q. These, together with Proposition 5.1, yields by (C.13) that, for \(R \geqq \mathcal {X}\),
Having this at our disposal, we may repeat the proof of Section 4.3 to conclude (5.6) for \(q \in [2,\infty )\).
6 Sharp Estimates of Linearization Errors
Here we show that Corollary 1.4 is a consequence of Theorems 1.1, 1.2 and 1.3.
Proof of Corollary 1.4
For each \(k\in \{0,\ldots ,n-1\}\), let \(\{ B^{(k)}_j \}\) be a finite family of balls such that
By the Vitali covering lemma, we may further assume that \(\frac{1}{3} B^{(k)}_j \cap \frac{1}{3} B^{(k)}_i=\emptyset \) whenever \(i\ne j\). Let \(\mathcal {Z}\) be the finite set consisting of the centers of these balls. The size of \(\mathcal {Z}\) depends only on the geometry of the sequence of domains \(U_0,U_1,\ldots ,U_{n}\). Let \(\mathcal {X}\) be the maximum of the random variables \(\mathcal {X}\) given in Theorem 1.3, centered at elements of \(\mathcal {Z}\), divided by the radius of the smallest ball. We assume that \(r\geqq \mathcal {X}\). This ensures the validity of Theorem 1.3 in each of the balls \(B^{(k)}_j\): that is, for every \(q\in [2,\infty )\) and \(m\in \{1,\ldots ,n\}\), we have the estimate
and hence, by the covering,
Proceeding with the proof of the corollary, we define, as usual,
Arguing by induction, let us suppose that \(m\in \mathbb {N}\) with \(m\geqq 1\) and \(\theta >0\) are such that, for every \(j \in \left\{ 0,\ldots ,m-1\right\} \) and \(q\in [2,\infty )\), there exists \(C(q,\mathrm {data})<\infty \) such that
This obviously holds for \(m=1\) and some \(\theta (d,\Lambda )>0\) by the Meyers estimate and Theorem 1.3. We will show that it must also hold for \(m+1\) and some other (possibly smaller) exponent \(\theta >0\).
Step 1. We show that
By the basic energy estimate,
where
By Theorem 1.3 (using our definition of \(\mathcal {X}\) and the fact that \(r\geqq \mathcal {X}\)) and the induction hypothesis, we have, for every \(k\in \{ 1,\ldots ,m\}\),
This completes the proof of (6.5).
Step 2. We show that
Observe that, since \(m\geqq 1\), \(\xi _m \in H^1_0(U_m)\), that is, \(\xi _m\) vanishes on \(\partial U_m\). Therefore, by Meyers estimate and Lemma C.1, in particular (C.13) with \(q=2+\frac{1}{2}\theta \) and \(\delta =\frac{1}{2}\theta \), we get
This completes the proof of (6.6).
The corollary now follows by induction. \(\quad \square \)
7 Liouville Theorems and Higher Regularity
In this section we prove Theorem 1.6 by an induction in the degree n. The initial step \(n=1\) has been already established in [1]. Indeed, \(\mathrm {(i)}_1\) and \(\mathrm {(ii)}_1\) are consequences of [1, Theorem 5.2], and \(\mathrm {(iii)}_1\) follows from Theorem 1.5 which is [1, Theorem 1.3]. Moreover, these estimates hold with optimal stochastic integrability, namely we may take any \(\sigma \in (0,d)\) for \(n=1\) (with the constant C then depending additionally on \(\sigma \)).
Throughout the section we will use the following notation: given \(p \in \mathbb {R}^d\) and \(k \in \mathbb {N}\), we denote
and
Remark 7.1
In proving Theorem 1.6\(\mathrm {(ii)}_n\) by induction in n, it will be necessary to prove a stronger statement, namely that, for every \(p \in B_{\mathsf {M}}\), \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^{p}\) and \((w_1,\ldots ,w_{n-1}) \in \mathsf {W}_{n-1}^p\) satisfying (1.30) for \(m\in \{1,\ldots ,n-1\}\), there exists \(w_n\) satisfying (1.30) for \(m=n\) such that \((w_1,\ldots ,w_{n}) \in \mathsf {W}_{n}^p\).
Proof of Theorem1.6\(\mathrm {(ii)}_n\) For fixed n, we will take \(\mathcal {X}\) as the maximum of random variables \(\mathcal {X}\) appearing in Theorem 1.2, Theorem 1.3 corresponding \(q=2+\delta \), where \(\delta \) is as in Theorem 1.2, and a deterministic constant \(\mathsf {R}(n,\mathsf {M}, d,\Lambda ) < \infty \). Clearly (1.29) holds then.
Given \(p \in B_{\mathsf {M}}\) and \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^p \), our goal is to prove that there exists a tuplet \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\) such that (1.30) holds for every \(R\geqq \mathcal {X}\) and \(k \in \{1,\ldots ,n\}\).
Step 1. Induction assumption. We proceed inductively and assume that there is a tuplet \((w_1,\ldots ,w_{n-1}) \in \mathsf {W}_{n-1}^p\) such that (1.30) is true for every \(R\geqq \mathcal {X}\) and \(m \in \{1,\ldots ,n-1\}\). The base case for the induction is valid by the results of [1], as mentioned at the beginning of this section. Our goal is therefore to construct \(w_n\) such that \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\) and (1.30) holds for every \(R\geqq \mathcal {X}\) and \(k \in \{1,\ldots ,n\}\). Recall that since \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^p\), we have, for every \(R\geqq \mathcal {X}\), that
Moreover, by the induction assumption, if the lower bound \(\mathsf {R}\) for \(\mathcal {X}\) is large enough, we deduce by (1.30) that, for \(R \geqq \mathcal {X}\) and \(m \in \{1,\ldots ,n-1\}\),
Step 1. Let \(r_j:= 2^j \mathcal {X}\) and let \(w_{n,j}\), \(j \in \mathbb {N}\), solve
We show that
Theorem 1.2 yields that, for \(m \in \{1,\ldots ,n\}\), solutions of
satisfy the estimate
By Theorem 1.3, together with (7.1) and (7.2), assumed for \(m \in \{1,\ldots ,n-1\}\), we obtain
Similarly, by Theorems 1.2 and 1.3, together with (7.2) valid for \(m \in \{1,\ldots ,n-1\}\), we get, for \(m \in \{1,\ldots ,n-1\}\), that
Consequently, again by (1.30) assumed for \(m \in \{1,\ldots ,n-1\}\), that
We denote, in short,
together with
Now, it is easy to see (cf. the proof of [1, Theorem 5.2]) that
so that, by Theorem 1.1,
Therefore, by an analogous computation to the proof of Lemma 2.11, using (7.6), we get
as well as
By testing the equation
with \(\overline{w}_{n,j} - \overline{w}_{n}\), we then obtain
Therefore, (7.4) follows by (7.5) and the above display by taking \(\delta = \alpha /2\).
Step 2. We show that there is \(w_n\) such that \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\) and \(w_n\) satisfies (1.30). Setting \(z_{n,j} := w_{n,j} - w_{n,j+1}\) we have by the triangle inequality that
Notice that \(z_{n,j}\) is \(\mathbf {a}_p\)-harmonic in \(B_{r_j}\). Thus, by [1, Theorem 5.2], we find \(\phi _{n,j} \in \mathcal {A}_{n}^p\) such that, for every \(r \in [\mathcal {X},r_j]\),
Consequently, for every \(r \in [\mathcal {X},r_j]\),
Setting then \(\tilde{w}_{n,j} := w_{n,j} - \sum _{h=1}^{j-1} \phi _{n,h}\), we have that \( \tilde{w}_{n,k} - \tilde{w}_{n,j} = \sum _{i= j}^{k-1} (z_{n,i} - \phi _{n,i}) \) and it follows that, for all \(j,k \in \mathbb {N}\), \(j < k\),
Therefore, \(\{\tilde{w}_{n,k} \}_{k=j}^\infty \) is a Cauchy sequence and, by the Caccioppoli estimate and the diagonal argument, we find \(w_n\) such that
and, for all \(j \in \mathbb {N}\),
We now use the facts that \((\overline{w}_1,\ldots ,\overline{w}_n) \in \overline{\mathsf {W}}_n^{p} \) and \(\phi _{n,j} \in \mathcal {A}_{n}^p\), together with (7.7) and (7.8), to deduce that
Combining this with (7.11) yields that \(w_n\) satisfies (1.30). Moreover, obviously \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\). The proof is complete. \(\quad \square \)
Proof of Theorem1.6\(\mathrm {(i)}_n\) Let \(\mathcal {X}\) be as in the beginning of the proof of \(\mathrm {(ii)}_n\). Fix \(R \geqq \mathcal {X}\). We proceed via induction. Assume that \((w_1,\ldots ,w_{n-1})\) satisfy (1.30), that is, we find \((\overline{w}_1,\ldots ,\overline{w}_{n-1}) \in \overline{\mathsf {W}}_{n-1}^p\) such that, for \(k \in \{1,\ldots ,n-1\}\) and \(t \geqq \mathcal {X}\),
Since \((\overline{w}_1,\ldots ,\overline{w}_{n-1}) \in \mathcal {P}_2 \times \cdots \times \mathcal {P}_{n}\), we have by the homogeneity that
We find, by Remark 7.2 below, a solution \(\overline{w} \in \mathcal {P}_{n+1}\) such that \((\overline{w}_1,\ldots ,\overline{w}_{n-1},\overline{w}) \in \overline{\mathsf {W}}_{n}^p\) and that there exists a constant \(\mathsf {C}(n,d,\Lambda )< \infty \) such that, for every \(t \geqq \mathcal {X}\),
Consequently, Remark 7.1 provides us \(\tilde{w}\) such that \((w_1,\ldots ,w_{n-1},\tilde{w}) \in \mathsf {W}_{n}^p\) and, for \(t\geqq \mathcal {X}\),
Moreover, by the equations and the growth condition at infinity, \(w_n - \tilde{w} \in \mathcal {A}_{n+1}^p\). Therefore, by [1, Theorem 5.2], there is \(q \in \overline{\mathcal {A}}_{n+1}^p\) such that, for all \(t \geqq \mathcal {X}\),
We set \(\overline{w}_n := \overline{w} + q\). Obviously, \((\overline{w}_1,\ldots ,\overline{w}_{n-1},\overline{w}_n) \in \overline{\mathsf {W}}_{n}^p\) since q is \({\overline{\mathbf {a}}}_p\)-harmonic. By (7.13) and the triangle inequality we have, for \(t\geqq \mathcal {X}\), that
and
We thus have by the triangle inequality that
proving the induction step and finishing the proof of \(\mathrm {(i)}_n\). \(\quad \square \)
Remark 7.2
We show that there is \(\overline{w}\) such that (7.13) is valid. Indeed, by letting \(\overline{v}\) solve
using the fact that \(\overline{\mathbf {F}}(p,\nabla \overline{w}_1,\ldots ,\nabla \overline{w}_{n-1}) \) is a polynomial of degree n, it is straightforward to show by homogeneity that
solves
We have the estimate
for all \(m \in \{1,\ldots ,n\}\), and by the equations of \(\overline{w}_1,\ldots ,\overline{w}_{n-1}\) we see that
Therefore, for all \(R>0\) and \(m \in \{1,\ldots ,n\}\),
Thus we have that, for \(t \geqq R\),
which yields (7.13).
We now turn to the proof of the large-scale \(C^{n,1}\) estimate.
Proof of Theorem1.6\(\mathrm {(iii)}_n\) Fix \(\mathsf {M} \in [1,\infty )\). By Theorem 1.3 there exist constants \(\sigma (n,\mathsf {M},\mathrm {data}) \in (0,1)\) and \(C(n,\mathsf {M},\mathrm {data})<\infty \) and a random variable \(\mathcal {X}\) satisfying \(\mathcal {X}\leqq \mathcal {O}_\sigma (C)\) so that the statement of Theorem 1.3 is valid with \(q = 2(n+2)\). We now divide the proof in two steps.
Step 1. Induction assumption. Assume \(\mathrm {(iii)}_{n-1}\). Consequently there is \(p \in B_{C}\) and a tuplet \((w_1,\ldots ,w_{n-1})\) such that, for \(k \in \{0,\ldots ,n-1\}\) and
we have that there exists \(C(k,\mathsf {M},\mathrm {data})\) such that, for every \(m \in \{0,\ldots ,n-1\}\) and for every \(r \in [\mathcal {X}, \frac{1}{2} (1+ 2^{-k-2})R]\),
where we denote, in short,
Our goal is to show that (7.16) continues to hold with \(m=n\) and for every \(r \in [\mathcal {X}, \frac{1}{2} (1+ 2^{-n-2})R]\). The base case \(n=1\) is valid since, by [1, Theorem 1.3], we have that there is \( p \in B_C\) such that, for all \(r \in [\mathcal {X}, \frac{3}{4} R]\),
Step 2. Construction of a special solution. We construct a solution \(\tilde{w}_n\) of
satisfying, for \(r \geqq \mathcal {X}\),
To show this, it first follows by (7.16) and the triangle inequality that, for \(m \in \{1,\ldots ,n-1\}\),
Since we have Theorem 1.3 at our disposal with \(q = 2(n+1)\), we can increase the integrability and obtain by (1.22) and (7.16) that, for \(r \in [\mathcal {X}, \frac{1}{2} (1+ \frac{3}{8} 2^{-n})R]\) and \(m \in \{1,\ldots ,n-1\}\),
Consequently, by (1.23) and (7.20), we also get, for \(m \in \{0,\ldots ,n-1\}\) and \(r \in [\mathcal {X}, \frac{1}{2} (1+\frac{3}{16} 2^{-m})R]\), that
Next, Theorem 1.6\(\mathrm {(i)}_{n-1}\) yields that we find \((\overline{w}_1,\ldots ,\overline{w}_{n-1}) \in \overline{\mathsf {W}}_{n-1}^{p}\) such that, for \(m \in \{1,\ldots ,n-1\}\),
In particular, applying this inductively, assuming that the lower bound \(\mathsf {R}\) for \(\mathcal {X}\) is such that \(C \mathsf {R}^{-\delta } \leqq \frac{1}{2}\), we deduce by (7.20) that, for \(k \in \{1,\ldots ,n-1\}\),
By Remark 7.2 we find a solution \(\overline{w}_n\) of
satisfying, for \(r \geqq \mathcal {X}\),
In view of (7.23) this yields, for \(r \geqq \mathcal {X}\), that
By Remark 7.1 we then find \(\tilde{w}_n\) solving (7.18) such that, for \(r\geqq \mathcal {X}\),
Now (7.19) follows by (7.23), (7.24) and (7.25) together with Theorem 1.3.
Step 3. We show the induction step, that is, we validate (7.16) for \(k=n\) and \(r \in [\mathcal {X}, \frac{1}{2} (1+ 2^{-n-2})R]\). Denote \(\tilde{\xi }_n := \xi _{n-1} - \frac{1}{n!} \tilde{w}_n\). We begin by deducing an estimate for \(\tilde{\xi }_n\). Appendix C tells us that \(\tilde{\xi }_n\) solves the equation
and there exist constants \(C(n,\mathsf {M},\mathrm {data})<\infty \) such that
By (7.21) and (7.20) we then obtain, for \(r \in [\mathcal {X}, \frac{1}{2} (1+\frac{3}{8} 2^{-n})R]\), that
Next, set, for \(j\in \mathbb {N}_0\), \(r_j := \frac{1}{2} \theta ^j (1+2^{-n-2})R\), where \(\theta (n,\mathsf {M},\mathrm {data}) \in \left( 0,\frac{1}{2} \right] \) will be fixed shortly. Let \(\phi _0 \in \mathcal {A}_{n+2}^p\) and, for given \(\phi _{j} \in \mathcal {A}_{n+2}^p\), let \(h_{j}\) solve
By testing and (7.27) we get
Furthermore, by [1, Theorem 5.2], there is \(\tilde{\phi }_{j+1} \in \mathcal {A}_{n+2}^p\) such that
Combining, we have by the triangle inequality, for \(\phi _{j+1} := \tilde{\phi }_{j+1} + \phi _{j} \in \mathcal {A}_{n+2}^p\), that
Choosing \(C\theta ^{1/2} = 1\), we thus arrive at
It follows by iteration that, for \(r \in [\mathcal {X}, \frac{1}{2} (1+ 2^{-n-2})R]\),
We can now revisit [4, Proof of (3.49)] and obtain that there exists \(\phi \in \mathcal {A}_{n+1}^p\) such that, taking \(w_n = \tilde{w}_n + \phi \) and \(\xi _n := \xi _{n-1} - \frac{1}{n!} w_n\), we get, for all \(r \in [\mathcal {X}, \frac{1}{2} (1+2^{-n-2})R]\),
Since \(\phi \in \mathcal {A}_{n+1}^p\), we see that \((w_1,\ldots ,w_n) \in \mathsf {W}_n^p\). Now (7.16) follows for \(k=n\) by the previous inequality together with the Caccioppoli estimate and (7.27). The proof is complete. \(\quad \square \)
Notes
To be pedantic, we actually have \(L^{1+\delta }\)-type integrability for a tiny \(\delta >0\) by the Meyers estimate, but this does not help.
References
Armstrong, S., Ferguson, S., Kuusi, T.: Homogenization, linearization and large- scale regularity for nonlinear elliptic equations. Comm. Pure Appl. Math. arXiv:1805.00467 (to appear)
Armstrong, S., Kuusi, T., Mourrat, J.-C.: Mesoscopic higher regularity and subadditivity in elliptic homogenization. Commun. Math. Phys. 347(2), 315–361, 2016
Armstrong, S., Kuusi, T., Mourrat, J.-C.: The additive structure of elliptic homogenization. Invent. Math. 208(3), 999–1154, 2017
Armstrong, S., Kuusi, T., Mourrat, J.-C.: Quantitative Stochastic Homogenization and Large-Scale Regularity, Vol. 352. Grundlehren der Mathematischen Wissenschaften [Fundamental Principles of Mathematical Sciences]. Springer, Cham, 2019
Armstrong, S.N., Mourrat, J.-C.: Lipschitz regularity for elliptic equations with random coefficients. Arch. Rational Mech. Anal. 219(1), 255–348, 2016
Armstrong, S.N., Smart, C.K.: Quantitative stochastic homogenization of convex integral functionals. Ann. Sci. Éc. Norm. Supér. (4)49(2), 423–481, 2016
Avellaneda, M., Lin, F.-H.: Compactness methods in the theory of homogenization. Commun. Pure Appl. Math. 40(6), 803–847, 1987
Avellaneda, M., Lin, F.-H.: Un théorème de Liouville pour des équations elliptiques à coefficients périodiques. C. R. Acad. Sci. Paris Sér. I Math. 309(5), 245–250, 1989
Avellaneda, M., Lin, F.-H.: \(L^p\) bounds on singular integrals in homogenization. Commun. Pure Appl. Math. 44(8–9), 897–910, 1991
Bella, P., Giunti, A., Otto, F.: Effective multipoles in random media. arXiv:1708.07672
Dal Maso, G., Modica, L.: Nonlinear stochastic homogenization. Ann. Mat. Pura Appl. 4(144), 347–389, 1986
Dal Maso, G., Modica, L.: Nonlinear stochastic homogenization and ergodic theory. J. Reine Angew. Math. 368, 28–42, 1986
Duerinckx, M., Gloria, A., Otto, F.: The structure of fluctuations in stochastic homogenization. arXiv:1602.01717
Fischer, J., Neukamm, S.: Optimal homogenization rates in stochastic homogenization of nonlinear uniformly elliptic equations and systems, August 2019. arXiv:1908.02273
Fischer, J., Otto, F.: A higher-order large-scale regularity theory for random elliptic operators. Commun. Partial Differ. Equ. 41(7), 1108–1148, 2016
Gloria, A., Neukamm, S., Otto, F.: Quantification of ergodicity in stochastic homogenization: optimal bounds via spectral gap on Glauber dynamics. Invent. Math. 199(2), 455–515, 2015
Gloria, A., Neukamm, S., Otto, F.: A regularity theory for random elliptic operators. arXiv:1409.2678
Gloria, A., Otto, F.: An optimal variance estimate in stochastic homogenization of discrete elliptic equations. Ann. Probab. 39(3), 779–856, 2011
Gloria, A., Otto, F.: The corrector in stochastic homogenization: optimal rates, stochastic integrability, and fluctuations, p. 114, May 2016. arXiv:1510.08290v3
Han, Q., Lin, F.: Elliptic Partial Differential Equations, Vol. 1. Courant Lecture Notes in Mathematics. American Mathematical Society, Providence, 1997
Jikov, V.V., Kozlov, S.M., Oleĭnik, O.A.: Homogenization of Differential Operators and Integral Functionals. Springer, Berlin, 1994. Translated from the Russian by G.A. Yosifian
Mingione, G.: Regularity of minima: an invitation to the dark side of the calculus of variations. Appl. Math. 51(4), 355–426, 2006
Moser, J., Struwe, M.: On a Liouville-type theorem for linear and nonlinear elliptic differential equations on a torus. Bol. Soc. Brasil. Mat. (N.S.)23(1–2), 1–20, 1992
Acknowledgements
Open access funding provided by University of Helsinki including Helsinki University Central Hospital. Scott Armstrong was partially supported by the NSF Grant DMS-1700329. Samuel J. Ferguson was partially supported by NSF Grants DMS-1700329 and DMS-1311833. Tuomo Kuusi was supported by the Academy of Finland and the European Research Council (ERC) under the European Union’s Horizon 2020 research and innovation programme (Grant Agreement No 818437).
Author information
Authors and Affiliations
Corresponding author
Additional information
Communicated by A. Figalli
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendices
Appendix A: Deterministic Regularity Estimates
In this first appendix, we record some deterministic regularity estimates of Schauder and Calderón–Zygmund type for linear equations with Hölder continuous coefficients. These estimates, while well-known, are not typically written with explicit dependence on the regularity of the coefficients, which is needed for our purposes in this paper.
Proposition A.1
(Calderón–Zygmund gradient \(L^q\) estimates). Let \(\beta \in (0,1]\), \(q\in [2,\infty )\) and \(\mathbf {a}\in \mathbb {R}^{d\times d}\) be a symmetric matrix with entries in \(C^{0,\beta }(B_2)\) satisfying
Suppose \(\mathbf {f} \in L^q(B_2;{\mathbb {R}^{d}})\) and \(u\in H^1(B_2)\) is a solution of
Then \(u \in W^{1,q}_{\mathrm{loc}}(B_2)\) and there exists \(C(q,d,\Lambda )<\infty \) such that
Proof
We will explain how to extract the statement of the proposition from that of [4, Proposition 7.3]. The latter asserts the existence of \(\delta _0(q,d,\Lambda )>0\) such that, for every ball \(x\in B_1\) and \(r\in \left( 0,\tfrac{1}{2}\right] \) satisfying
we have, for a constant \(C(q,d,\Lambda )<\infty \), the estimate
Since \({{\,\mathrm{osc}\,}}_{B_{2r}} \mathbf {a}\leqq (2r)^\beta \left[ \mathbf {a}\right] _{C^{0,\beta }(B_2)}\), we have the above estimate for every \(x\in B_1\) and
From this, Fubini’s theorem and Young’s inequality for convolutions, we obtain
This completes the proof. \(\quad \square \)
Proposition A.2
Let \(\beta \in (0,1)\) and \(\mathbf {a}\in \mathbb {R}^{d\times d}\) be a symmetric matrix with entries in \(C^{0,\beta }(B_2)\) satisfying
Suppose \(\mathbf {f} \in C^{0,\beta } (B_2;{\mathbb {R}^{d}})\) and \(u\in H^1(B_2)\) is a solution of
Then \(u \in C^{1,\beta }_{\mathrm{loc}}(B_2)\) and there exists \(C(\beta ,d,\Lambda )<\infty \) such that
and
Proof
We will explain how to extract the statement of the proposition from the gradient Hölder estimate found in [20, Theorem 3.13]. The latter states that, under the assumption that
there exists \(C(\beta ,d,\Lambda )<\infty \) such that
After changing the scale, we obtain the corresponding statement in \(B_r\), which asserts that, under the assumption that
there exists \(C(\beta ,d,\Lambda )<\infty \) such that
Therefore we take \(r:= \frac{1}{2}\wedge \left[ \mathbf {a}\right] _{C^{0,\beta } (B_{2})}^{-\frac{1}{\beta }}\) and apply the previous statement in every ball \(B_r(x)\) with \(x\in B_1\) to obtain
After a covering argument, we obtain
which yields the proposition. \(\quad \square \)
Appendix B: Differentiation of \(\mathbf {F}_m\)
In this appendix, we show that (2.13) holds.
Lemma B.1
Fix \(m \in \mathbb {N}\) and \(h,p \in \mathbb {R}^d\). Suppose that \(z\mapsto L(z,x)\) is \(C^{m+2}\) and \(t \mapsto \mathbf {g}( p + th)\) is m times differentiable at 0. Then
Proof
We first observe that the terms \(D_p \mathbf {g}(p) h^{\otimes 1},\ldots ,D_p^{m} \mathbf {g}(p) h^{\otimes m}\) in (B.1) are precisely the directional derivatives of \(\mathbf {g}\) in the h direction, up to mth degree. The terms in (B.1) involve derivatives of \(z\mapsto L(z,x)\) up to the \((m+2)\text {th}\) degree. Hence, we can assume by approximation, without loss of generality, that \(z \mapsto L(z,x)\) and \(p \mapsto \mathbf {g}(p)\) are polynomials, of degrees at most \(m+2\) and m, respectively. Fix \(h \in \mathbb {R}^d\) and let \(t \in \mathbb {R}\). We write
Denote
Examining the relation between the p and t derivatives of \(\mathbf {Z}(p,t)\), we find that
Set now, for fixed \(h,x\in \mathbb {R}^d\),
and, by the definition of \(\mathbf {F}_m\) in (1.15),
and similarly for \(\mathbf {F}_{m+1}\). Computing the directional derivative, recalling that we assume that \(z \mapsto L(z,x)\) is a \((m+2)\)th degree polynomial,
we get by (B.2) that
Consequently, we have
which is (B.1), concluding the proof. \(\quad \square \)
Appendix C: Linearization Errors
In this appendix we compute the equation satisfied by a higher-order linearization error and thereby obtain gradient estimates.
Lemma C.1
(Equation for the linearization error). Fix \(\mathsf {M},R \in (0,\infty )\) and \(n\in \mathbb {N}\) with \(n \geqq 2\). Assume that \(p \mapsto L(p,x)\) is \(C^{n+1,1}\) for every \(x\in \mathbb {R}^d\) and
Suppose that \(u,v,w_1,\ldots ,w_n \in H^1(B_R)\) satisfy
and, for each \(m \in \{1,\ldots ,n\}\),
where \( \mathbf {F}_m\) is defined in (1.15). Denote \(\xi _0 = v -u\) and, for each \(m \in \{1,\ldots ,n\}\),
Then there is vector field \(\mathbf {E}_{n}\) such that \(\xi _n\) solves
and there exists a constant \(C(n,\mathsf {M},d)<\infty \) such that
Furthermore, there exist constants \(C(n,\mathsf {M},d)<\infty \), \(q(n,d) \in (2,\infty )\) and \(\delta (d,\Lambda ) \in \left( 0 ,\tfrac{1}{2} \right] \) such that
Proof
Throughout the proof we use the notation \(s_k = \sum _{j=1}^k \frac{w_j}{j!}\) and \(\xi _0 = v-u\), so that \(\xi _k = \xi _0 - s_k\).
Step 1. Recalling that \(\mathbf {F}_1 = 0\), we may rewrite
where we define
By the equations of u, v and \(w_k\), we have that
It thus remains to estimate \(\mathbf {E}_{n}\).
Step 2. We show that, for \(k \in \{2,\ldots n\}\) and \(m \in \{k,\ldots ,n\}\),
where \(\mathbf {S}^{(j)}_{m}\) and \(\mathbf {E}^{(j)}_{m}\) are defined, for \(j \in \{2,\ldots ,k\}\), recursively as
and
with
Indeed, suppose that we have, for \(j \in \{1,\ldots ,k-1\}\) and \( m \in \{j,\ldots ,n\}\), that
This is obviously true for \(j=1\). We compute, for \(m \in \{k,\ldots ,n\}\),
which proves the recursive formula (C.3).
Step 3. We show that, for \(k \in \{2,\ldots , n\}\) there exists a constant \(C(n,k,d)<\infty \) such that
The statement is easy to verify by induction. Indeed, for \(m=j=2\) we have by (C.5) that
Assume then that, for \( m\in \{2,\ldots ,n-1\}\) and \(j \in \{2,\ldots ,m\}\), we have
By the definition of \(\mathbf {E}^{(j)}_{n}\), we have, for \(j \in \{2,\ldots ,n\}\), that
By (C.8), using Fubini for sums, we obtain,
This, together with (C.9), proves the induction step, and gives also (C.7).
Step 4. We show that
By the recursive formula we have, for \(k \in \{2,\ldots ,n\}\), that
and thus (C.10) follows by (C.7).
Step 5. We show that
For this, we first abbreviate
and observe that, by definition,
Second, we observe that, by induction on \(j\geqq 2\), we have
for all \(n\geqq j\). Third, by commutativity of addition, we observe that
Finally, letting \(\mathbf {F}_{m}=0\) for \(m<2\) and \(\mathbf {S}^{(j)}_{n}=0\) for \(j>n\) for notational convenience, we note that the above equation may be rewritten as
which is (C.11).
Step 6. Conclusion. We have that
Indeed, by a Taylor expansion, we see that
Applying this with \(z_0 = \nabla u\) and \(z = \nabla v - \nabla u\) gives (C.12). Combining this with the previous steps yields the desired estimate (C.1) for \(\mathbf {E}_{n}\). Finally, by the Hölder and Young inequalities, we get, for all \(q \in [2,\infty )\) and \(r \in (0,R]\),
Let \(\delta _0\) be the Meyers exponent corresponding \(\Lambda \). Let
Set also \(R_h := \frac{1}{2} (1+2^{-h}) R\). With this notation the previous display yields, by Hölder’s inequality, that
Now (C.2) follows by the Caccioppoli estimate, concluding the proof. \(\quad \square \)
Appendix D: Regularity for Constant Coefficient Linearized Equations
In this appendix we prove a lemma tracking down the regularity of a solution \((\overline{w}_1,\ldots ,\overline{w}_n)\) of the linearized system in the case that \(\overline{L}\) is a smooth, constant-coefficient Lagrangian.
Throughout we fix \(n \in \mathbb {N}_0\), \(\Lambda \in [1,\infty )\), \(\beta \in (0,1)\), and assume that \(\overline{L}\) satisfies
and for all \(\mathsf {M}_0 \in [1,\infty )\) there is \(C(\mathsf {M}_0,\beta ,d)<\infty \) such that
Lemma D.1
(Regularity of \(\overline{w}_{m}\)). Let \(\eta \in [\tfrac{1}{2},1)\), \(\mathsf {M} \in [1,\infty )\) and \(R \in (0,\infty )\). Assume that \(\overline{L}\) satisfies (D.1) and (D.2). Let \(\overline{u},\overline{w}_1,\ldots ,\overline{w}_n\) solve the equations, for \(m \in \{1,\ldots ,n+1\}\),
where \(\overline{\mathbf {F}}_m\) has been defined in (1.16) and \(\overline{u}\) satisfies
Then, for \(m \in \{1,\ldots ,n+1\}\), there exists a constant \(C(m,\eta ,\mathsf {M},\beta ,d,\Lambda )<\infty \) such that
Moreover, letting
we have, for \(m \in \{1,\ldots ,n+1\}\), that
and, for \(m \in \{1,\ldots ,n\}\) and \(k \in \{1,\ldots ,n+1-m\}\),
Notice that by (D.4) we may always take \(\delta = \mathsf {M}^{\beta }\) in (D.6). When applying the result in practice, we typically take \(\delta \) to be very small.
Proof
Fix \(m \in \{1,\ldots ,n+1\}\), \(\eta \in [\tfrac{1}{2},1)\), \(\mathsf {M} \in [1,\infty )\). Let \(\overline{u},\overline{w}_1,\ldots ,\overline{w}_n\) solve (D.3) and assume (D.4). Fix also \(\delta \) as in (D.6).
Throughout the proof we denote, for \(\theta \in (0,\infty )\),
We also denote
and, for \(j \in \mathbb {N}\),
Below we denote by C a constant depending only on parameters \((m,\eta ,\mathsf {M},\beta ,d,\Lambda )\). It may change from line to line.
Step 1. Basic properties of \(\overline{u}\). In view of [1, Proposition A.1], assumption (D.1) and normalization (D.4) imply that there exists a constant \(\mathsf {M_0}(\eta ,\mathsf {M},d,\Lambda )<\infty \) such that
Therefore, (D.2) is applicable in \(B_{\frac{1}{3} (2+\eta )R}\), and we obtain by [1, Proposition A.1] that
We also define
We have by (D.10) and (D.4) that
Notice also that, by (D.12), we have
Step 2. Induction assumption on degree m. We assume inductively that, for \(j \in \{1,\ldots ,m-1\}\) there exists a constant \(\mathsf {K}_{j}(\eta ,\mathsf {M},\mathsf {K}_{0},d,\Lambda )<\infty \) such that
and
Notice that the case \(m=2\) has been established in [1, Proposition A.1].
Throughout the next steps of the proof, we let constants C depend on parameters \((\{\mathsf {K}_{i}\}_{i=1}^{m-1}, m,\eta ,\mathsf {M},\beta ,\mathsf {K}_{0},d,\Lambda )\), and they may change from line to line.
Step 3. Bounds on \(\mathbf {f}\). We show that under induction assumptions (D.14) and (D.15), we have that
and, for \(r \in (0,r_m]\) and \(y \in B_{R_m}\), defining
we have that
To show (D.16), we have by (D.15) that
To show (D.18), using Hölder regularity of \(\overline{\mathbf {f}}_{m}\) with respect to \(\nabla \overline{u}\) variable, similarly to (2.11), gives us
Applying (D.14) and (D.6) yields that
Thus (D.18) follows by (D.12) and (D.19).
Step 4. Caccioppoli estimate. We show that under induction assumptions (D.14) and (D.15), we have that, for all \(y \in B_{R_m + r_m}\) and \(r \in (0,r_m]\),
Since \(\overline{w}_{m}\) solves the equations, for \(\mathbf {f}_{m,y,r}\) defined in (D.17) and any affine function \(\ell \),
and
we obtain (D.20) simply by testing and (D.18).
Step 5. Induction assumption on the scale. We now assume that we find \(\varepsilon \in (0,1]\), a constant \(\mathsf {C}_\varepsilon \), and \(r^* \in (0, \varepsilon r_m]\) such that
Notice that, by the Caccioppoli estimate (D.20) and (D.13), we have, for any \(\varepsilon \in (0,1]\), that
implying that (D.21) is valid for \(r^* = \varepsilon r_m\) provided that \(\mathsf {C}_\varepsilon \geqq C_\varepsilon \).
Step 6. We verify that (D.15) is true for \(j=m\). This gives us also (D.5). Actually, we prove that if (D.21) is valid for some \(r^* \in (0, \varepsilon r_m]\), then it remains valid for \(\tfrac{1}{2} r^*\) instead. This proves, by induction, that we may take any \(r^* \in (0, \varepsilon r_m]\) in (D.21). In particular, we obtain (D.15) for \(j=m\).
Fix \(y \in B_{R_m}\). Rewriting the equation of \( \overline{w}_{m}\) as before, for \(r \in \left( 0 ,r_m \right] \),
where \(\mathbf {f}_{m,y,r}\) defined in (D.17), and consequently solving
we obtain by testing and (D.18) that
In particular, we get by (D.21) for \(r\in [r^*,\varepsilon r_m]\) that
By decay estimate for harmonic functions we have for small enough \(\theta (\beta ,d,\Lambda ) \in \left( 0,\frac{1}{2}\right] \) that
Therefore, by the triangle inequality, we get
By an iteration argument we thus obtain that, for \(r \in [\theta r^*, \varepsilon r_m]\)
Letting \(r \in [\theta r^*, \varepsilon r_m]\) and \(n\in \mathbb {N}_0\) be such that \(r \in (\theta ^{n+1} \varepsilon r_m, \theta ^n \varepsilon r_m]\), we obtain by the triangle inequality that
Thus, by the previous two displays and (D.22), we obtain, for \(r \in [\theta r^*, \varepsilon r_m]\), that
We first take \(\varepsilon \) so small that \(C\varepsilon = \frac{1}{2}\) and then choose \(\mathsf {C}_\varepsilon \geqq 2C_\varepsilon \). All in all, we have proved that
which implies that (D.21) is valid for \(\frac{1}{2} r^*\) instead of \(r^*\), which was to be shown.
Step 7. We now prove that (D.14) is valid for \(j=m\), giving also (D.7). An application of the Caccioppoli estimate (D.20), together with (D.5), which was proved in Step 6 above, we have that, by giving up volume factors,
Therefore, (D.23) yields, for all \(y \in B_{\eta R}\), that
This yields, via telescoping summation as in Step 6, that
Thus, if, on the one hand, \(r = |x-y| \in (0,r_m]\), we get by the above two displays that
If, on the other hand \( |x-y|> r_m\), we get
by noticing that
and applying once more (D.20). Thus we have proved (D.7).
Step 8. We finally sketch the proof of (D.8). Since it is very similar to the above reasoning, we will omit most of the details. We prove the statement by using induction in m and in k. First, we observe that by differentiation we see that \(\partial _{x_j}^k \overline{u}\) satisfies the equation
Thus we can apply (D.5) and (D.8) for \(\overline{w}_k = \partial _{x_j}^k u\) recursively and obtain, by polarization as in Lemma 2.6, that, for every \(k \in \{1,\ldots ,n+1\}\) and \(\eta \in \left[ \tfrac{1}{2},1\right) \), there is a constant \(C(k,\eta ,\mathsf {M},\beta ,d,\Lambda )\) such that
Next, \(\overline{w}_1\) solves \(-\nabla \cdot \left( \mathbf {b}\nabla \overline{w}_1 \right) = 0\) and \(z \mapsto \mathbf {b}(z) = D^2_p\overline{L}(\nabla u(z))\) is in \(C^{n,\beta }\) by (D.24), we may differentiate the equation at most n times and obtain that \(\overline{w}_1 \in C^{n+1,\beta }\) and it is also straightforward to show that \(\overline{w}_1\) satisfies (D.8) for \(m=1\); this is just classical Schauder theory.
We than assume that (D.8) is valid for every \(m \in \{1,\ldots ,M\}\) and \(k \in \{1,\ldots ,n+1-m\}\) with some \(M \in \{1,\ldots ,n-1\}\). We then show that it continues to hold for \(m = M+1\) and \(k \in \{1,\ldots ,n+M\}\) as well. Now
Recalling that \((h_1,\ldots ,h_{m-1}) \mapsto \overline{F}_m(\nabla \overline{u},h_1,\ldots ,h_{m-1})\) is a polynomial, using (D.24) and (D.8) for \(m \in \{1,\ldots ,M\}\) and \(k \in \{1,\ldots ,n+1-m\}\), we can actually deduce that
Therefore, using (D.24) once more, we can differentiate the equation of \( \overline{w}_{M+1}\)k times and then show that \( \overline{w}_{M+1}\) satisfies (D.8). The proof is complete. \(\quad \square \)
Appendix E: \(C^\infty \) Regularity for Smooth Constant-Coefficient Lagrangians
In this section we give an alternative proof of the statement that \(C^{1,1}\) regularity implies \(C^\infty \) regularity for smooth, constant-coefficient Lagrangians. Our argument is similar to the classical argument by Schauder theory, but we keep track of the linearized equations to obtain a Taylor series with an explicit representation of the Taylor polynomials in terms of the linearized equations. We note that it is relatively simple to obtain real analyticity for solutions using this argument.
Proposition E.1
Fix \(\varepsilon \in (0,\frac{1}{2} ]\), \(\mathsf {M} \in [0,\infty )\), \(\mathsf {N} \in \mathbb {N}\), \(\sigma \in [2,\infty )\), and \(R \in (0,\infty ]\). Suppose that \(\overline{L} \in C^{\mathsf {N}+2,1}(\mathbb {R}^d)\) is uniformly convex, that is, for all \(\zeta ,\xi \in \mathbb {R}^d\),
Let \(u \in H^1(B_R)\) solve \(\nabla \cdot D\overline{L}(\nabla u) = 0\) such that \(\left\| \nabla u \right\| _{\underline{L}^2(B_R)} \leqq \mathsf {M}\). Then there exist constants \(C(\overline{L},\mathsf {M},\mathsf {N},\varepsilon ,\mathrm {data})\) and polynomials \(q_1, \ldots , q_{\mathsf {N+2}}\) such that \(q_{m+1}\) is homogeneous polynomial of degree \(m+1\) solving
and, for all \(r \in \left( 0,\frac{R}{2}\right] \),
Proof
Without loss of generality, we may take \(R=1\). By \(C^{1,1}\)-estimates, see for example [1, Proposition A.1], we have that
We set
Assume then inductively that, for \(m \in \{1,\ldots ,n\}\), there exists homogeneous polynomials \(q_m\) of degree m such that, for every \(\sigma \in [2,\infty )\) and \(\varepsilon \in (0,\tfrac{1}{2}]\), there exists a constant \(\mathsf {N}_{m,\sigma }(\varepsilon ,d,\Lambda )\) such that
and that, for \(m \in \{1,\ldots ,n-1\}\), \(q_{m+1}\) satisfies the equation
Let us denote \({\overline{\mathbf {a}}}:= D^2_p\overline{L}\left( \nabla q_1 \right) \) and, for \(m \in \{1,\ldots ,n-1\}\),
By homogeneity, we find a homogeneous polynomial \(q_{n+1}\) of degree \(n+1\) solving the equation
Notice that there is a degree of freedom in the choice of \(q_{n+1}\). Namely, the solution is unique up to an \({\overline{\mathbf {a}}}\)-harmonic polynomial of degree \(n+1\). We will fix this shortly. To draw parallels between this appendix and Appendix C, we set
Rewrite
where
By the estimate in Appendix C, we have that
Taking divergence gives us, by the equations of u, \( \overline{w}_{1},\ldots , \overline{w}_{n} \), that
Using the induction assumption we get that
Now Lemma E.2 below allows us to identify the homogeneous \({\overline{\mathbf {a}}}\)-harmonic polynomial part of \(q_{n+1}\) of degree \(n+1\) such that
and
This proves the induction step, and finishes the proof. \(\quad \square \)
Lemma E.2
Let \(n \in \mathbb {N}\) and let \(\alpha \in (n,n+1)\). Let \(\mathsf {M} \in [0,\infty )\) and \(\varepsilon \in (0,1)\). Suppose that \(\mathbf {A} \) is a constant symmetric matrix having eigenvalues on the interval \([1,\Lambda ]\). There is a constant \(C(n,\alpha ,\varepsilon ,d,\Lambda )\) such that if \(\mathbf {F} \in L^p(B_1)\) and \(u \in H^1(B_1)\) solve
and that \(\mathbf {F} \in L^p(B_1)\) satisfies, for \(r \in (0,1)\),
then there is \(\mathbf {A}\)-harmonic \(q \in \mathcal {P}_{n+1}\) such that
Proof
We proceed via harmonic approximation. Let \(v_r \in u + H_0^1(B_r)\) be A-harmonic. Denote
By Calderón–Zygmund estimates and (E.2),
Using the oscillation decay estimate
and defining
we obtain by the triangle inequality, for \(\theta >0\) such that \( C \theta ^{n+1-\alpha } = \frac{1}{2}\), that
It follows by reabsorption that, for \(r \in \left( 0,1\right) \),
In particular, letting \(\tilde{q}_t \) be the minimizing element of \(\overline{\mathcal {A}}_{n+1}\) in the definition of D(t), we get by the triangle inequality that
This allows us to identify \(q \in \overline{\mathcal {A}}_{n+1}\) such that, for \(t \in (0,r)\),
and it follows that
The proof is complete by an easy estimate \( D(1-\varepsilon ) \leqq C_\varepsilon (\left\| \nabla u \right\| _{\underline{L}^2 \left( B_{1} \right) } + \mathsf {M} )\). \(\quad \square \)
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Armstrong, S., Ferguson, S.J. & Kuusi, T. Higher-Order Linearization and Regularity in Nonlinear Homogenization. Arch Rational Mech Anal 237, 631–741 (2020). https://doi.org/10.1007/s00205-020-01519-1
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00205-020-01519-1