## 1 Introduction

Exploiting the recent advances in artificial intelligence and, in particular, in deep learning, several innovative numerical techniques have been developed in the last few years to compute numerical solutions of partial differential equations (PDEs). In such methods, the solution is approximated by a neural network that is trained by taking advantage of the knowledge of the underlying differential equation. One of the earliest models involving a neural network was described in [1]: it is based on the concept of Physics Informed Neural Networks (PINN) and it inspired further works such as e.g. [2] or [3], until the recent paper [4] which presents a very general framework for the solution of operator equations by deep neural networks.

In such papers, given an arbitrary PDE coupled with proper boundary conditions, the training of the PINN aims at finding the weights $${{\mathbf {w}}}$$ of a neural network such that the associated function $$u^{{{{\mathcal {N}}}{{\mathcal {N}}}}}({\mathbf {x}};{{\mathbf {w}}})$$ minimizes some functional of the equation residual while satisfying as much as possible the imposed boundary conditions. To do so, the neural network is trained to minimize the residual only at a finite set of collocation points and additional terms are added to the loss function in order to force the network to approximately satisfy the boundary conditions. Thanks to the good approximation properties of neural networks, formally proved e.g. in [5,6,7,8,9,10] under suitable assumptions, the PINN approach looks very promising because it is able to efficiently and accurately compute approximate solutions of arbitrary PDEs encoding their structures in the loss function.

Subsequently, the PINN paradigm has been further developed in [11] to obtain the so-called Variational Physics Informed Neural Networks (VPINN). The main differences with respect to the PINN are that the weak formulation of the PDE is exploited, the collocation points are replaced by test functions, and quadrature points are used to compute the integrals involved in the variational residuals. In such a method the solution is still approximated by a neural network, but the test functions are represented by a finite set of known functions or by a second neural network (see [12]); therefore, the technique can be seen as a Petrov-Galerkin method. The method is more flexible than the standard PINN because the integration by parts, involved in the weak formulation, decreases the required regularity of the approximate solution. Furthermore, the fact that the dataset used in the training phase consists of quadrature points significantly reduces the computational cost of the training phase. Indeed, quadrature points are, in general, much fewer than collocation points.

Combining the VPINN with the Finite Element Method (FEM), the authors of [13] developed VarNet, a VPINN that exploits the test functions of the $${\mathbb {P}}_1$$-FEM. Such a work has been then extended in [14] to consider arbitrary high-order polynomials as test functions, as in the hp version of the FEM. Although the authors of the cited works empirically observed that both PINNs and VPINNs are able to efficiently approximate the desired solution, no proof of a priori error estimates with convergence rates is provided for VPINNs. On the contrary, rigorous a posteriori error analyses are already available (see, for instance, [15]). Recently (see [16]) we derived a posteriori error estimates for the discretization setting considered in this paper.

The purpose of this paper is to investigate how the choice of piecewise polynomial test functions and quadrature formulas influence the accuracy of the resulting VPINN approximation of second-order elliptic problems. One might think that test functions of high polynomial degree are needed to get a high order of accuracy; we prove that this is not the case, actually we indicate that precisely the opposite is true: it is more convenient to keep the degree of the test functions as low as possible, while using quadrature formulas of precision as high as possible. Indeed, for sufficiently smooth solutions, the error decay rate is given by

\begin{aligned} q+2 - k_\text {test} \,, \end{aligned}

where q is the precision of the quadrature formula and $$k_\text {test}$$ is the degree of the test functions.

Using a Petrov–Galerkin framework, we derive an a priori error estimate in the energy norm between the exact solution and a suitable piecewise polynomial interpolant of the computed neural network; we assume that the architecture of the neural network is fixed and sufficiently rich, and we explore the behaviour of the error versus the size of the mesh supporting the test functions. Our analysis relies upon the validity of an inf-sup condition between the spaces of test functions and the space in which the neural network is interpolated.

Numerical experiments confirm the theoretical prediction. Interestingly, in our experiments the error between the exact solution and the computed neural network decays asymptotically with the same rate as predicted by our theory for the interpolant of the network; however, this behaviour cannot be rigorously guaranteed, since in general the minimization problem which defines the computed neural network is underdetermined, and the computed neural network may be affected by spurious components. Indeed, we show that for a problem with zero data the minimization of the loss function may yield non-vanishing neural networks. With the method proposed in this paper, we combine the efficiency of the VPINN approach with the availability of a sound and certified convergence analysis.

The paper is organized as follows. In Sect. 2 we introduce the elliptic problem we are focusing on, and we also present the way in which the Dirichlet boundary conditions are exactly imposed, which is uncommon in PINNs and VPINNs but can be generalized as in [17]. In Sect. 3 we focus on the numerical discretization; in particular, the involved neural network architecture is described in Sect. 3.1, while the problem discretization and the corresponding loss function are described in Sect. 3.2. Here we also introduce an interpolation operator $${\mathcal {I}}_H$$ applied to the neural networks. Sect. 4 is the key theoretical section: through a series of preliminary results, we formally derive the a priori error estimate, the main result being Theorem 6. In Sect. 5 we specify the parameters of the neural network used for the numerical tests and the training phase details. We also analyse the consequences of fulfilling the inf-sup condition in connection with the VPINN efficiency. Various numerical tests are presented and discussed in Sect. 6 for two-dimensional elliptic problems. Such tests empirically confirm the validity of the a priori estimate in different scenarios. Furthermore, we compare the accuracy of the proposed method with that of a standard PINN and the non-interpolated VPINN. We also analyse the relationship between the neural network hyperparameters and the VPINN accuracy, and we highlight, with numerical experiments and analytical examples, the importance of the inf-sup condition. In Sect. 7, we show that our VPINN can be easily adapted to solve a parametric nonlinear PDE, with accurate results for the whole range of parameters. Finally, in Sect. 8, we draw some conclusions and highlight the future perspective of the current work.

## 2 The Model Boundary-value Problem

Let $$\Omega \subset {\mathbb {R}}^n$$ be a bounded polygonal/polyhedral domain with boundary $$\Gamma =\partial \Omega$$, partitioned into $$\Gamma =\Gamma _D \cup \Gamma _N$$ with $$\Gamma _D \cap \Gamma _N = \emptyset$$ and $$\text {meas}_{n-1}(\Gamma _D)>0$$.

Let us consider the model elliptic boundary-value problem

\begin{aligned} {\left\{ \begin{array}{ll} Lu:=-\nabla \cdot (\mu \nabla u) + \varvec{\beta }\cdot \nabla u + \sigma u =f &{} \text {in } \Omega \,, \\ u=g &{} \text {on \ } \Gamma _D \,, \\ \mu \frac{\partial u}{\partial n} = \psi &{} \text {on } \Gamma _N \,, \end{array}\right. } \end{aligned}
(2.1)

where $$\mu , \sigma \in \mathrm{L}^\infty (\Omega )$$, $$\varvec{\beta } \in (\mathrm{W}^{1,\infty }(\Omega ))^n$$ satisfy $$\mu \ge \mu _0$$, $$\sigma - \frac{1}{2} \nabla \cdot \varvec{\beta } \ge 0$$ in $$\Omega$$ for some constant $$\mu _0>0$$, whereas $$f \in L^2(\Omega )$$, $$g={\bar{u}}_{\vert \Gamma _D}$$ for some $${\bar{u}} \in \mathrm{H}^1(\Omega )$$, and $$\psi \in \mathrm{L}^2(\Gamma _N)$$.

Define the spaces $$U=\mathrm{H}^1(\Omega )$$, $$V=\mathrm{H}^1_{0,\Gamma _D}(\Omega ):=\{v \in U : v_{\vert \Gamma _D}=0 \}$$, the bilinear form $$a:U\times V \rightarrow {\mathbb {R}}$$ and the linear form $$F:V\rightarrow {\mathbb {R}}$$ such that

\begin{aligned} a(w,v)=\int _\Omega \mu \nabla w \cdot \nabla v + \varvec{\beta }\cdot \nabla w \, v + \sigma w \, v\,, \qquad F(v)=\int _\Omega f \, v + \int _{\Gamma _N} \psi \, v \,; \end{aligned}
(2.2)

denote by $$\alpha \ge \mu _0$$ the coercivity constant of the form a, and by $$\Vert a \Vert$$, $$\Vert F \Vert$$ the continuity constants of the forms a and F. Problem (2.1) is formulated variationally as follows: Find $$u \in {\bar{u}}+V$$ such that

\begin{aligned} a(u,v)=F(v) \qquad \forall v \in V\,. \end{aligned}
(2.3)

We assume that we can represent u in the form

\begin{aligned} u= {\bar{u}}+\Phi {\tilde{u}} \,, \end{aligned}
(2.4)

for some (known) smooth function $$\Phi \in V$$ and some $${\tilde{u}} \in U$$ having the same smoothness of u. Let us introduce the affine mapping

\begin{aligned} B :U \rightarrow {\bar{u}}+V \qquad \text { such that} \quad Bw={\bar{u}}+\Phi w \end{aligned}
(2.5)

which enforces the given Dirichlet boundary condition. Then, Problem (2.3) can be equivalently formulated as follows: Find $${\tilde{u}} \in U$$ such that

\begin{aligned} a(B{\tilde{u}},v)=F(v) \qquad \forall v \in V\,. \end{aligned}
(2.6)

### Remark 1

(Enforcement of the Dirichlet conditions) The approach we follow to enforce Dirichlet boundary conditions will allow us to deal with a loss function which is built solely by the residuals of the variational equations. Other approaches are obviously possible: for instance, one could augment such loss function by a term penalizing the distance of the numerical solution from the data on $$\Gamma _D$$, or adopt a Nitsche’s type variational formulation of the boundary-value problem [18]. Both strategies involve parameters which may need a tuning, whereas in our approach the definition of the loss function is simple and natural, allowing us to focus on the performances of the neural networks.

## 3 The VPINN-based Numerical Discretization

In this section, we first introduce the class of neural networks used in this paper, then we describe the numerical discretization of the boundary-value problem (2.6), which uses neural networks to represent the discrete solution and piecewise polynomial functions to enforce the variational equations. An inf-sup stable Petrov-Galerkin formulation is introduced which guarantees stability and convergence, as indicated in Sect. 4; this is the main difference between the proposed method and other formulations, such as [11, 14].

### 3.1 Neural Networks

In this work we only use fully-connected feed-forward neural networks (named also multi-layered perceptrons), therefore the following description is focused on such a class of networks. Since we deal with a scalar equation, a neural network will be a function $$w:{\mathbb {R}}^n\rightarrow {\mathbb {R}}$$ defined as follows: for any $${\varvec{x}}\in {\mathbb {R}}^n$$, the output $$w({\varvec{x}})$$ is computed via the chain of assignments

\begin{aligned} \begin{aligned}&{\varvec{x}}_0={\varvec{x}}, \\&{\varvec{x}}_\ell = \rho ({{\mathbf {A}}}_\ell {\varvec{x}}_{\ell -1} + {{\mathbf {b}}}_\ell ), \ell = 1,...,L-1, \\&w({\varvec{x}}) = {{\mathbf {A}}}_{L} {\varvec{x}}_{L-1} + b_L\,. \end{aligned} \end{aligned}
(3.1)

Here, $${{\mathbf {A}}}_\ell \in {\mathbb {R}}^{N_\ell \times N_{\ell -1}}$$ and $${{\mathbf {b}}}_\ell \in {\mathbb {R}}^{N_\ell }$$, $$\ell =1,...,L$$, are matrices and vectors that store the network weights (with $$N_0=n$$ and $$N_L=1$$); furthermore, L is the number of layers, whereas $$\rho$$ is the (nonlinear) activation function which acts component-wise (i.e. $$\rho (\mathbf{y})=\left[ \rho (y_1),...,\rho (y_{n_y})\right]$$ for any vector $$\mathbf{y}\in {\mathbb {R}}^{n_y}$$). It can be noted from Eq. (3.1), that if $$\rho \in \mathrm{C}^k({\mathbb {R}})$$, then w inherits the same regularity because it can be seen as a composition of functions belonging to $$\mathrm{C}^k({\mathbb {R}})$$. Popular choices include the ReLU ($$k=0$$) and RePU ($$k>0$$ finite) functions, as well as the hyperbolic tangent ($$k=\infty$$) if one wants to exploit the maximum of regularity in the solution of interest.

The neural network structure $${{\mathcal {N}}\!{\mathcal {N}}}$$ is identified by fixing the number of layers L, the integers $$N_\ell$$ and the activation function $$\rho$$. The entire set of weights that parametrize the network can be logically organized into a single vector $${{\mathbf {w}}} \in {\mathbb {R}}^N$$. Thus, the neural network structure $${{\mathcal {N}}\!{\mathcal {N}}}$$ induces a mapping

\begin{aligned} {{\mathcal {F}}}^{{\mathcal {N}}\!{\mathcal {N}}} : {\mathbb {R}}^N \rightarrow \mathrm{C}^\infty ({\bar{\Omega }})\,, \qquad {{\mathbf {w}}} \mapsto {{\mathcal {F}}}^{{\mathcal {N}}\!{\mathcal {N}}} ({{\mathbf {w}}}) = w \,, \quad \text {where } w = w({\varvec{x}},{{\mathbf {w}}})\,. \end{aligned}
(3.2)

It is convenient to define the manifold

\begin{aligned} U^{{\mathcal {N}}\!{\mathcal {N}}} \subset U\,, \qquad U^{{\mathcal {N}}\!{\mathcal {N}}} = {{\mathcal {F}}}^{{\mathcal {N}}\!{\mathcal {N}}}({\mathbb {R}}^N) \end{aligned}

containing all functions that can be generated by the neural network structure $${{\mathcal {N}}\!{\mathcal {N}}}$$.

### 3.2 The VPINN Discretization

We aim at approximating the solution of Problem (2.1) by a generalized Petrov-Galerkin strategy. To this end, let us introduce a conforming, shape-regular triangulation $${{\mathcal {T}}}_h= \{ E \}$$ of $${\bar{\Omega }}$$ with meshsize $$h>0$$ and, for a fixed integer $$k_\text {test} \ge 1$$, let $$V_h \subset V$$ be the linear subspace formed by the functions which are piecewise polynomials of degree $$k_\text {test}$$ over the triangulation $${\mathcal {T}}_h$$. Furthermore, let us introduce computable approximations of the forms a and F by numerical quadratures. Precisely, for any $$E \in {{\mathcal {T}}}_h$$, let $$\{(\xi ^E_\iota ,\omega ^E_\iota ) : \iota \in I^E\}$$ be the nodes and weights of a quadrature formula of precision

\begin{aligned} q \ge 2k_\text {test} \end{aligned}
(3.3)

on E. Assume that $$\Gamma _N$$ is the union of a collection $$\partial {{\mathcal {T}}}_h (\Gamma _N)$$ of edges of elements of $${{\mathcal {T}}}_h$$; for any such edge e, let $$\{(\xi ^e_\iota ,\omega ^e_\iota ) : \iota \in I^e\}$$ be the nodes and weights of a quadrature formula of precision q on e. Then, assuming that all the data $$\mu$$, $$\varvec{\beta }$$, $$\sigma$$, f, $$\psi$$ are continuous in each element of the triangulation, we define the approximate forms

\begin{aligned} a_h(w,v)= & {} \sum _{E \in {{\mathcal {T}}}_h} \sum _{\iota \in I^E} [\mu \nabla w \cdot \nabla v + \varvec{\beta }\cdot \nabla w \, v + \sigma w v](\xi ^E_\iota ) \,\omega ^E_\iota \,, \end{aligned}
(3.4)
\begin{aligned} F_h(v)= & {} \sum _{E \in {{\mathcal {T}}}_h} \sum _{\iota \in I^E} [ f v](\xi ^E_\iota ) \,\omega ^E_\iota \quad + \sum _{e \in \partial {\mathcal {T}}_h(\Gamma _N)} \sum _{\iota \in I^e} [ \psi v](\xi ^e_\iota ) \, \omega ^e_\iota \,. \end{aligned}
(3.5)

With these ingredients at hand, we would like to approximate the solution of Problem (2.6) by some $$u^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}}$$ satisfying

\begin{aligned} a_h(Bu^{{\mathcal {N}}\!{\mathcal {N}}},v_h)=F_h(v_h) \qquad \forall v_h \in V_h\,. \end{aligned}
(3.6)

Such a problem might be ill-posed when, for computational efficiency, the dimension of the test space $$V_h$$ is chosen smaller than the dimension of the manifold $$U^{{\mathcal {N}}\!{\mathcal {N}}}$$. In this situation, we get an under-determined problem, with obvious difficulties in deriving stability estimates on some norms of the function $$Bu^{{\mathcal {N}}\!{\mathcal {N}}}$$. Actually, Problem (3.6) with zero data (i.e., zero f, g, $$\psi$$) could admit non-zero solutions (see Sect. 6.3).

To avoid these difficulties, we adopt the strategy of applying a projection (indeed, an interpolation) to the function $$Bu^{{\mathcal {N}}\!{\mathcal {N}}}$$, mapping it into a finite dimensional space of dimension comparable to that of $$V_h$$, and we limit ourselves with estimating some norm of this projection.

To be precise, let us introduce a conforming, shape-regular partition $${{\mathcal {T}}}_H=\{G\}$$ of $${\bar{\Omega }}$$, which is equal to or coarser than $${{\mathcal {T}}}_h$$ (i.e., each element $$E \in {{\mathcal {T}}}_h$$ is contained in an element $$G \in {\mathcal {T}}_H$$) but compatible with $${{\mathcal {T}}}_h$$ (i.e., its meshsize $$H>0$$ satisfies $$H\lesssim h$$). Let the integer $$k_\text {int}\ge 1$$ be defined by the condition

\begin{aligned} k_\text {int}+k_\text {test} =q + 2 \,. \end{aligned}
(3.7)

Let $$U_H \subset U$$ be the linear subspace formed by the functions which are piecewise polynomials of degree $$k_\text {int}$$ over the triangulation $${{\mathcal {T}}}_H$$, and let $$U_{H,0} = U_H \cap V$$ be the subspace of $$U_H$$ formed by the functions vanishing on $$\Gamma _D$$. Finally, let $${{\mathcal {I}}}_H : \mathrm{C}^0({\bar{\Omega }}) \rightarrow U_H$$ be an interpolation operator, satisfying the condition $${{\mathcal {I}}}_H : \mathrm{C}^0({\bar{\Omega }})\cap V \rightarrow U_{H,0}$$ as well as the following approximation properties: for all $$v \in \mathrm{H}^{k+1}(\Omega )$$, $$1 \le k \le k_\text {int}$$,

\begin{aligned} \vert v - {{\mathcal {I}}}_H v \vert _{\ell , G} \lesssim H^{k+1-\ell } \vert v\vert _{k+1,G}\,, \qquad 0 \le \ell \le k+1\,, \quad \forall G \in {{\mathcal {T}}}_H\,. \end{aligned}
(3.8)

In this framework, assuming the lifting $${\bar{u}}$$ to be continuous in $${\bar{\Omega }}$$, we replace the target Eq. (3.6) by the following ones:

\begin{aligned} a_h({{\mathcal {I}}}_H Bu^{{\mathcal {N}}{\mathcal {N}}},v_h)=F_h(v_h) \qquad \forall v_h \in V_h\,. \end{aligned}
(3.9)

In order to handle this problem with the neural network, let us introduce a basis in $$V_h$$, say $$V_h = \text {span}\{\varphi _i : i\in I_h\}$$, and for any w smooth enough let us define the residuals

\begin{aligned} r_{h,i}(w)=F_h(\varphi _i)-a_h({{\mathcal {I}}}_H Bw,\varphi _i)\,, \qquad i \in I_h\,, \end{aligned}
(3.10)

as well as the loss function

\begin{aligned} R_h^2(w) = \sum _{i \in I_h} r_{h,i}^2(w) \, \gamma _i^{-1}\,, \end{aligned}
(3.11)

where $$\gamma _i >0$$ are suitable weights. Then, we search for a global minimum of the loss function in $$U^{{\mathcal {N}}\!{\mathcal {N}}}$$, i.e., we consider the following discretization of Problem (2.6): Find $$u^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}}$$ such that

\begin{aligned} u^{{\mathcal {N}}\!{\mathcal {N}}} \in \displaystyle {\text {arg}\!\!\!\!\min _{w \in U^{{\mathcal {N}}\!{\mathcal {N}}}}}\, R_h^2(w) \,. \end{aligned}
(3.12)

Note that the solution $$u^{{\mathcal {N}}\!{\mathcal {N}}}$$ may not be unique; however, a suitable choice of the space $$U_H$$ may lead to the control of the error $$u-{{\mathcal {I}}}_H Bu^{{\mathcal {N}}\!{\mathcal {N}}}$$ in the $$H^1$$-norm, as we will see in the sequel.

### Remark 2

(Discretization without interpolation) For the sake of comparison, we will also consider the optimization problem in which no interpolation is applied to the neural network functions. In other words, the target equations are those in (3.6), which induce the following definition of loss function

\begin{aligned} {\hat{R}}_h^2(w) = \sum _{i \in I_h} {\hat{r}}_{h,i}^2(w) \, \gamma _i^{-1}\,, \qquad \text {with } \quad {\hat{r}}_{h,i}(w)=F_h(\varphi _i)-a_h(Bw,\varphi _i) \,, \end{aligned}
(3.13)

and the following minimization problem: Find $${\hat{u}}^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}}$$ such that

\begin{aligned} {\hat{u}}^{{\mathcal {N}}\!{\mathcal {N}}} \in \displaystyle {\text {arg}\!\!\!\!\min _{w \in U^{{\mathcal {N}}\!{\mathcal {N}}}}}\, {\hat{R}}_h^2(w) \,. \end{aligned}
(3.14)

Note that in this problem the triangulation $${{\mathcal {T}}}_H$$ and the space $$U_H$$ play no role. Although we will not provide a rigorous error analysis for such discretization, it will be interesting to numerically compare the behaviour of the approaches (i.e., with or without interpolation). This will be done in Sect. 6.

## 4 A Priori Error Estimates

Let $$u^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}}$$ be any solution of the minimization problem (3.12); let us set

\begin{aligned} u_H^{{{\mathcal {N}}\!{\mathcal {N}}}} = {{\mathcal {I}}}_H Bu^{{\mathcal {N}}\!{\mathcal {N}}} \in U_H\,. \end{aligned}
(4.1)

Recalling the definition (2.4) of the affine mapping B, it holds

\begin{aligned} u_H^{{{\mathcal {N}}\!{\mathcal {N}}}} = {\bar{u}}_H + u_H^{{{\mathcal {N}}\!{\mathcal {N}}},0}\,, \quad \text {with} \quad {\bar{u}}_H = {{\mathcal {I}}}_H {\bar{u}} \quad \text {and} \quad u_{H,0}^{{\mathcal {N}}\!{\mathcal {N}}} = {{\mathcal {I}}}_H (\Phi u^{{\mathcal {N}}\!{\mathcal {N}}}) \in U_{H,0}\,;\qquad \end{aligned}
(4.2)

note that $${\bar{u}}_H$$ is a discrete lifting in $$U_H$$ of the Dirichlet data g.

We aim at estimating the error between u and $$u_H^{{\mathcal {N}}\!{\mathcal {N}}}$$. To accomplish this task, we need several definitions, assumptions, and technical results.

### Definition 1

(norm-equivalence) Let us denote by $$0< c_h \le C_h$$ the constants in the norm equivalence

\begin{aligned} c_h \Vert v_h \Vert _{1, \Omega } \le \Vert {\varvec{v}} \Vert _\gamma \le C_h \Vert v_h \Vert _{1, \Omega } \qquad \forall v_h \in V_h \,, \end{aligned}
(4.3)

where $${\varvec{v}} = (v_i)_{i \in I_h}$$ is such that $$v_h = \sum _{i \in I_h} v_i \varphi _i$$, and $$\Vert {\varvec{v}} \Vert _\gamma = \left( \sum _{i \in I_h} v_i^2\gamma _i \right) ^{1/2}$$.

Next, we introduce the consistency errors due to numerical quadratures

\begin{aligned} E_h^a(w_H,v_h)= & {} a(w_H,v_h)-a_h(w_H,v_h) \qquad \forall w_H \in U_H, \ \forall v_h \in V_h \,, \end{aligned}
(4.4)
\begin{aligned} E_h^F(v_h)= & {} F(v_h)-F_h(v_h) \qquad \forall v_h \in V_h \,, \end{aligned}
(4.5)

and we provide a bound on these errors. To this end, let us assume that the quadrature rules used in the elements in $${{\mathcal {T}}}_h$$ are obtained by affine transformations from a quadrature rule $$\{({\hat{\xi }}_\iota ,{\hat{\omega }}_\iota ) : \iota \in {\hat{I}}\}$$ on a reference element $${\hat{E}} \subset {\mathbb {R}}^n$$; similarly, let us assume that the quadrature rules used in the edges on $$\partial {{\mathcal {T}}}_h(\Gamma _N)$$ are obtained by affine transformations from a quadrature rule $$\{({\check{\xi }}_\iota ,{\check{\omega }}_\iota ) : \iota \in {\check{I}}\}$$ on a reference element $${\check{e}} \subset {\mathbb {R}}^{n-1}$$.

### Assumption 1

(Data smoothness) Let us assume the following smoothness of data:

\begin{aligned} \mu , \ \sigma , \ f \in \mathrm{W}^{k,\infty }(\Omega )\,, \qquad \varvec{\beta } \in (\mathrm{W}^{k,\infty }(\Omega ))^n\,, \qquad \psi \in \mathrm{W}^{k,\infty }(\Gamma _N)\,, \end{aligned}
(4.6)

where k is an integer satisfying

\begin{aligned} 1 \le k \le k_\text {int}=q+2-k_\text {test}\,. \end{aligned}
(4.7)

Consequently, let us introduce the following notation

\begin{aligned} {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma )= & {} \Vert \mu \Vert _{\mathrm{W}^{k,\infty }(\Omega )} + \Vert \varvec{\beta } \Vert _{(\mathrm{W}^{k,\infty }(\Omega ))^n} + \Vert \sigma \Vert _{\mathrm{W}^{k,\infty }(\Omega )} \,, \end{aligned}
(4.8)
\begin{aligned} {{\mathcal {N}}}_k(f, \psi )= & {} \Vert f \Vert _{\mathrm{W}^{k,\infty }(\Omega )} + \Vert \psi \Vert _{\mathrm{W}^{k,\infty }(\Gamma _N)} \,, \end{aligned}
(4.9)
\begin{aligned} \Vert w_H \Vert _{k, {{\mathcal {T}}}_H}= & {} \left( \sum _{G \in {\mathcal {T}}_H} \Vert w_{H \vert G} \Vert _{\mathrm{H}^k(G)} \right) ^{1/2} \quad \forall w_H \in U_H\,. \end{aligned}
(4.10)

### Property 2

(approximation of the forms a and F) Under Assumption 1, it holds

\begin{aligned} \vert E_h^a(w_H,v_h)\vert\lesssim & {} \ h^k {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert w_H \Vert _{k, {{\mathcal {T}}}_H} \Vert v_h \Vert _{1,\Omega } \qquad \forall w_H \in U_H, \ \forall v_h \in V_h \,,\qquad \end{aligned}
(4.11)
\begin{aligned} \vert E_h^F(v_h)\vert\lesssim & {} \ h^k {{\mathcal {N}}}_k(f,\psi ) \Vert v_h \Vert _{1,\Omega } \qquad \forall v_h \in V_h \,, \end{aligned}
(4.12)

### Proof

Both estimates are classical in the theory of finite elements (see, e.g., [19]). As far as (4.11) is concerned, the standard proof given for the case in which the polynomial degree is the same for both arguments, i.e., $$k=k_\text {test} \ge 1$$ and $$q=2(k-1)$$, can be easily adapted to the present situation $$k+k_\text {test} \le q+2$$. In this way, one gets $$\vert E_h^a(w_H,v_h)\vert \ \lesssim \ h^k {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert w_H \Vert _{k, {{\mathcal {T}}}_h} \Vert v_h \Vert _{1,\Omega }$$, and one concludes by observing that $$\Vert w_H \Vert _{k, {{\mathcal {T}}}_h} = \Vert w_H \Vert _{k, {{\mathcal {T}}}_H}$$ since $${{\mathcal {T}}}_h$$ is a refinement of $${{\mathcal {T}}}_H$$. $$\square$$

Finally, we pose a fundamental assumption.

### Assumption 3

(inf-sup condition between $$U_{H,0}$$ and $$V_h$$) The bilinear form a satisfies an inf-sup condition with respect to the spaces $$U_{H,0}$$ and $$V_h$$, namely there exists a constant $$\alpha _\star >0$$, independent of the meshsizes h and H, such that

\begin{aligned} \alpha _\star \Vert w_H \Vert _{1,\Omega } \le \sup _{v_h \in V_h} \frac{a(w_H,v_h)}{\Vert v_h \Vert _{1,\Omega }} \qquad \forall w_H \in U_{H,0}\,. \end{aligned}
(4.13)

This assumption together with Property 2 yields the following result.

### Proposition 4

(discrete inf-sup condition between $$U_{H,0}$$ and $$V_h$$) Under Assumptions 1 and 3, for all $$h \le h_0$$ small enough the bilinear form $$a_h$$ satisfies an inf-sup condition with respect to the spaces $$U_{H,0}$$ and $$V_h$$, namely there exists a constant $${\tilde{\alpha }}_\star >0$$ such that

\begin{aligned} {\tilde{\alpha }}_\star \Vert w_H \Vert _{1,\Omega } \le \sup _{v_h \in V_h} \frac{a_h(w_H,v_h)}{\Vert v_h \Vert _{1,\Omega }} \qquad \forall w_H \in U_{H,0}\,. \end{aligned}
(4.14)

### Proof

We have $$a_h(w_H,v_h) = a(w_H,v_h) - E_h^a(w_H,v_h)$$. Using the bound (4.11) with $$k=1$$ and observing that $$\Vert w_H \Vert _{1,{{\mathcal {T}}}_H}=\Vert w_H \Vert _{1,\Omega }$$, one can find $$h_0>0$$ small enough such that, for all $$h \le h_0$$, $$\vert E_h^a(w_H,v_h)\vert \le \frac{1}{2} \alpha _\star \Vert w_H \Vert _{1,\Omega } \Vert v_h \Vert _{1,\Omega }$$, whence the result with $${\tilde{\alpha }}_\star = \frac{1}{2} \alpha _\star$$ $$\square$$

We are ready to estimate the error $$\Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }$$. Recalling the decomposition (4.2), we use the triangle inequality

\begin{aligned} \Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \le \Vert u - u_H \Vert _{1,\Omega } + \Vert u_H - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \,, \end{aligned}
(4.15)

where $$u_H$$ is a suitable element in the affine subspace $${\bar{u}}_H + U_{H,0} \subset U_H$$. Writing $$u_H = {\bar{u}}_H + u_{H,0}$$ with $$u_{H,0} \in U_{H,0}$$, one has $$u_H - u_H^{{\mathcal {N}}\!{\mathcal {N}}} = u_{H,0} - u_{H,0}^{{\mathcal {N}}\!{\mathcal {N}}} \in U_{H,0}$$; hence, we can apply (4.14) to get

\begin{aligned} \Vert u_H - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \le \frac{1}{{\tilde{\alpha }}_\star } \sup _{v_h \in V_h} \frac{a_h(u_H - u_H^{{\mathcal {N}}\!{\mathcal {N}}},v_h)}{\Vert v_h \Vert _{1,\Omega }} \,. \end{aligned}
(4.16)

Recalling the definitions (4.4) and (4.5), it holds

\begin{aligned} a_h(u_H,v_h)= & {} a(u_H,v_h) - E_h^a(u_H,v_h) \\= & {} a(u,v_h) - a(u-u_H,v_h) - E_h^a(u_H,v_h) \\= & {} F(v_h) - a(u-u_H,v_h) - E_h^a(u_H,v_h) \\= & {} F_h(v_h) +E_h^F(v_h)- a(u-u_H,v_h) - E_h^a(u_H,v_h) \,. \end{aligned}

Thus, the numerator in (4.16) is given by

\begin{aligned} a_h(u_H - u_H^{{\mathcal {N}}\!{\mathcal {N}}},v_h) = F_h(v_h) - a_h(u_H^{ {\mathcal {N}}\!{\mathcal {N}}},v_h) - a(u-u_H,v_h) - E_h^a(u_H,v_h) +E_h^F(v_h)\,. \end{aligned}

On the other hand, recalling (3.10) we have

\begin{aligned} F_h(v_h) - a_h(u_H^{{\mathcal {N}}\!{\mathcal {N}}},v_h)=F_h(v_h) - a_h({{\mathcal {I}}}_H Bu^{{\mathcal {N}}\!{\mathcal {N}}},v_h) = \sum _{i \in I_h} r_{h,i}(u^{{\mathcal {N}}\!{\mathcal {N}}}) \, v_i \,, \end{aligned}
(4.17)

hence, by (3.11) and (4.3),

\begin{aligned} \vert F_h(v_h) - a_h(u_H^{{\mathcal {N}}\!{\mathcal {N}}},v_h) \vert \le R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) \Vert {\varvec{v}} \Vert _\gamma \le C_h R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) \Vert v_h \Vert _{1, \Omega } \,. \end{aligned}

Using the bounds (4.11) and (4.12), we obtain the following inequality

\begin{aligned} \Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }&\lesssim \left( 1+\frac{1}{{\tilde{\alpha }}_\star } \right) \Bigl (\ \inf _{u_H \in {\bar{u}}_H +U_{H,0}}\left( \Vert u - u_H \Vert _{1,\Omega } + h^k {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert u_H \Vert _{k, {{\mathcal {T}}}_H} \right) \nonumber \\&\quad + C_h R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) + h^k {{\mathcal {N}}}_k(f,\psi ) \ \Bigr )\,. \end{aligned}
(4.18)

From now on, we assume that $$u \in \mathrm{H}^{k+1}(\Omega )$$. Then, assumption (3.8) yields the inequalities

\begin{aligned} \Vert u - {{\mathcal {I}}}_H u \Vert _{1,\Omega } \lesssim H^k \vert u\vert _{k+1,\Omega } \lesssim h^k \vert u\vert _{k+1,\Omega } \end{aligned}
(4.19)

and

\begin{aligned} \Vert {{\mathcal {I}}}_H u \Vert _{k, {{\mathcal {T}}}_H} \le \Vert u \Vert _{k,\Omega } + \Vert u-{{\mathcal {I}}}_H u \Vert _{k, {{\mathcal {T}}}_H} \lesssim \Vert u \Vert _{k,\Omega } + H \vert u\vert _{k+1,\Omega } \lesssim \Vert u \Vert _{k+1,\Omega } \,. \nonumber \\ \end{aligned}
(4.20)

Choosing $$u_H = {{\mathcal {I}}}_H u \in {\bar{u}}_H +U_{H,0}$$ in (4.18) and using these estimates, we arrive at the following intermediate result, which can be viewed as a mixed a priori/a posteriori error estimate.

### Lemma 4.1

Under the previous assumptions, it holds

\begin{aligned} \Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \lesssim h^k (\vert u\vert _{k+1,G} + {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert u \Vert _{k+1,\Omega } + {{\mathcal {N}}}_k(f,\psi ) ) + C_h R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) \,. \end{aligned}

Our next task will be bounding the term $$R_h(u^{{\mathcal {N}}\!{\mathcal {N}}})$$. To this end, we use the minimality condition (3.12) to get

\begin{aligned} R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) \le R_h(w^{{\mathcal {N}}\!{\mathcal {N}}}) \qquad \forall w^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}} \,. \end{aligned}
(4.21)

On the other hand, since $$R_h(w^{{\mathcal {N}}\!{\mathcal {N}}})$$ is a weighted $$\ell _2$$-norm in $${\mathbb {R}}^{\vert I_h\vert }$$, we can write

\begin{aligned} R_h(w^{{\mathcal {N}}\!{\mathcal {N}}}) = \sup _{{{\mathbf {z}}} \in {\mathbb {R}}^{\vert I_h\vert }} \frac{1}{\Vert {{\mathbf {z}}} \Vert _\gamma } \sum _{i \in I_h} r_{h,i}(w^{{\mathcal {N}}\!{\mathcal {N}}}) z_i \,, \end{aligned}

where, similarly to (4.17),

\begin{aligned} \sum _{i \in I_h} r_{h,i}(w^{{\mathcal {N}}\!{\mathcal {N}}}) z_i = F_h(z_h) - a_h({{\mathcal {I}}}_H Bw^{{\mathcal {N}}\!{\mathcal {N}}}, z_h) \qquad \text {with } z_h = \sum _{i \in I_h} z_i \varphi _i \in V_h \,. \end{aligned}

For convenience, in analogy with (4.1), let us set

\begin{aligned} w_H^{{{\mathcal {N}}\!{\mathcal {N}}}} = {{\mathcal {I}}}_H Bw^{{\mathcal {N}}\!{\mathcal {N}}} \in U_H\,. \end{aligned}
(4.22)

Thus, recalling (4.3), we obtain

\begin{aligned} R_h(w^{{\mathcal {N}}\!{\mathcal {N}}}) \le \frac{1}{c_h} \sup _{z_h \in V_h} \frac{F_h(z_h) - a_h(w_H^{{\mathcal {N}}\!{\mathcal {N}}}, z_h)}{\Vert z_h \Vert _{1, \Omega }} \qquad \forall w^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{ {\mathcal {N}}\!{\mathcal {N}}} \,. \end{aligned}
(4.23)

The numerator can be manipulated as above, using

\begin{aligned} F_h(z_h) = F(z_h) - E_h^F(z_h) = a(u,z_h) - E_h^F(z_h) \end{aligned}

and

\begin{aligned} a_h(w_H^{{\mathcal {N}}\!{\mathcal {N}}}, z_h) = a(w_H^{{\mathcal {N}}\!{\mathcal {N}}}, z_h) - E_h^a(w_H^{{\mathcal {N}}\!{\mathcal {N}}},z_h) \,, \end{aligned}

whence, using once more Property 2, we get

\begin{aligned} R_h(w^{{\mathcal {N}}\!{\mathcal {N}}}) \lesssim \frac{1}{c_h} \left( \Vert u - w_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } + h^k {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert w_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{k, {{\mathcal {T}}}_H} + h^k {{\mathcal {N}}}_k(f,\psi ) \right) \,.\qquad \quad \nonumber \\ \end{aligned}
(4.24)

In order to bound the terms containing $$w_H^{{\mathcal {N}}\!{\mathcal {N}}}$$, we introduce the quantity

\begin{aligned} e^{{{\mathcal {N}}\!{\mathcal {N}}}} = u-Bw^{{\mathcal {N}}\!{\mathcal {N}}}\,, \end{aligned}
(4.25)

which, recalling the definitions (2.4) and (2.5), can be written as

\begin{aligned} e^{{{\mathcal {N}}\!{\mathcal {N}}}} = \Phi ({\tilde{u}}-w^{{\mathcal {N}}\!{\mathcal {N}}}) \,, \end{aligned}
(4.26)

and we formulate a final assumption.

### Assumption 5

(smoothness of the solution and the neural network manifold) The solution u can be represented as in (2.4) with

\begin{aligned} {\tilde{u}} \in \mathrm{H}^{k+1}(\Omega ) \qquad \text {and} \qquad \Phi \in \mathrm{W}^{k+1,\infty }(\Omega ) \end{aligned}
(4.27)

for k satisfying (4.7). Furthermore, the manifold formed by the neural network functions satisfies the smoothness condition

\begin{aligned} U^{{\mathcal {N}}\!{\mathcal {N}}} \subset \mathrm{H}^2(\Omega )\,. \end{aligned}
(4.28)

Note that (4.27) implies in particular $$u \in \mathrm{H}^{k+1}(\Omega )$$ with the bound $$\Vert u \Vert _{k+1,\Omega } \lesssim \Vert {\tilde{u}} \Vert _{k+1,\Omega }\,\Vert \Phi \Vert _{k+1,\infty ,\Omega }$$; on the other hand, (4.28) implies $$e^{{\mathcal {N}}\!{\mathcal {N}}} \in \mathrm{H}^2(\Omega )$$. (We refer to Remark 4 for another set of assumptions on the neural network.)

Recalling (4.22) and using the identity

\begin{aligned} u-w_H^{{\mathcal {N}}\!{\mathcal {N}}} = (u-{{\mathcal {I}}}_H u) + {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} = (u-{{\mathcal {I}}}_H u) - e^{{\mathcal {N}}\!{\mathcal {N}}} + (I-{{\mathcal {I}}}_H) e^{{\mathcal {N}}\!{\mathcal {N}}} \,, \end{aligned}
(4.29)

we can write

\begin{aligned} \Vert u-w_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \lesssim \Vert u-{{\mathcal {I}}}_H u \Vert _{1,\Omega } + \Vert e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } + H \vert e^{{\mathcal {N}}\!{\mathcal {N}}}\vert _{2,\Omega } \end{aligned}
(4.30)

and, using a standard inverse inequality in $${\mathbb {P}}_k(G)$$ for any $$G \in {{\mathcal {T}}}_H$$,

\begin{aligned} \begin{aligned} \Vert w_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{k, {{\mathcal {T}}}_H}&\lesssim \Vert {{\mathcal {I}}}_H u \Vert _{k, {{\mathcal {T}}}_H} + H^{1-k} \Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } \\&\lesssim \Vert {{\mathcal {I}}}_H u \Vert _{k, {{\mathcal {T}}}_H} + H^{1-k} \left( \Vert e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega } + H \vert e^{{\mathcal {N}}\!{\mathcal {N}}}\vert _{2,\Omega } \right) \,. \end{aligned} \end{aligned}
(4.31)

Keeping into account (4.19) and (4.20), in order to conclude we need to identify a function $${\tilde{w}}^{{\mathcal {N}}\!{\mathcal {N}}} \in U^{{\mathcal {N}}\!{\mathcal {N}}}$$ for which a bound of the type

\begin{aligned} \vert e^{{\mathcal {N}}\!{\mathcal {N}}}\vert _{m,\Omega } \lesssim \vert {\tilde{u}}-{\tilde{w}}^{{\mathcal {N}}\!{\mathcal {N}}}\vert _{m,\Omega } \lesssim H^{k+1-m} \vert {\tilde{u}}\vert _{k+1,\Omega } \end{aligned}
(4.32)

holds true for $$m=1,2$$. The existence of such a function is guaranteed by one of the available results on the approximation of functions in Sobolev spaces by neural networks (see [20,  Theorem 5.1, Remark 5.2]; see also [9]), provided the number of layers L and the widths of the layers in the chosen $${{\mathcal {N}}\!{\mathcal {N}}}$$ satisfy suitable conditions depending on the target accuracy (hence, in our case depending on $$H^k$$). Indeed, suppose one is interested in using meshes with meshsize as small as $$H_{\min }$$ in the domain $$\Omega$$ (here assumed to satisfy $$\Omega \subset [0,1]^n$$ for the sake of simplicity), and let $$N\in {\mathbb {N}}$$ be such that

\begin{aligned} 3^n(1\!+\!\delta )(2(m\!+\!1))^{3m} \!\!\max \{R^m,\text {ln}^m(\beta N^{k+n+3})\}\!\dfrac{C(n,m,k,{{\tilde{u}}})}{N^{k+1-m}}\!\!\le \! H_{\min }^{k+1-m} \vert {\tilde{u}}\vert _{k+1,\Omega } \,,\quad \end{aligned}
(4.33)

where $$\delta$$, R, $$\beta$$ and C are constants not depending on N defined in [20]. Then, a function $${\tilde{w}}^{{\mathcal {N}}\!{\mathcal {N}}}$$ exists which fulfils (4.32) and is represented as a neural network with the hyperbolic tangent as activation function and two hidden layers with $$N_1$$ and $$N_2$$ neurons respectively, satisfying

\begin{aligned} N_1 \le 3 \left\lceil \frac{k+1}{2}\right\rceil \left| P_{k,n+1}\right| + n(N-1), N_2 \le 3 \left\lceil \frac{n+2}{2} \right\rceil \left| P_{n+1,n+1}\right| N^n, \end{aligned}
(4.34)

where

\begin{aligned} \left| P_{a,b}\right| = \left( \begin{array}{c} a+b-1 \\ a\end{array} \right) ,\qquad \forall a,b\in {\mathbb {N}}, \ b\ge 2. \end{aligned}

Substituting (4.32) into (4.30) and (4.31), and using inequalities (4.21) and (4.24), we arrive at the following bound on the loss $$R_h(u^{{\mathcal {N}}\!{\mathcal {N}}})$$.

### Lemma 4.2

Under the previous assumptions, it holds

\begin{aligned} R_h(u^{{\mathcal {N}}\!{\mathcal {N}}}) \lesssim \frac{1}{c_h} \left( H^k \vert u\vert _{k+1,\Omega } + H^k \vert {\tilde{u}}\vert _{k+1,\Omega } + h^k {{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) \Vert {\tilde{u}} \Vert _{k+1,\Omega } + h^k {{\mathcal {N}}}_k(f,\psi ) \right) \,. \end{aligned}

We remark that such a bound, when the involved neural network is comprised of at least two hidden layers and is such that there exists N satisfying both (4.33) and (4.34), does not depend on the network hyperparameters.

Concatenating Lemmas 4.1 and 4.2, and using once more $$H \lesssim h$$, we obtain the following a priori error estimate for the solution of Problem (3.12).

### Theorem 6

(a priori error estimate) Let $$u_H^{{\mathcal {N}}\!{\mathcal {N}}} \in U_H$$ be defined by (4.1). Under Assumptions 1, 3 and 5, for h sufficiently small it holds

\begin{aligned} \begin{aligned} \Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }&\lesssim \left( 1+\frac{C_h}{c_h} \right) h^k \big [ (1+{{\mathcal {N}}}_k(\mu , \varvec{\beta }, \sigma ) ) \Vert {\tilde{u}} \Vert _{k+1,\Omega } + {{\mathcal {N}}}_k(f,\psi ) \big ] \,. \end{aligned} \end{aligned}
(4.35)

### Remark 3

(on the equivalence constants $$c_h, C_h$$) If a classical Lagrange basis is used in (3.10), and the triangulation $${{\mathcal {T}}}_h$$ is quasi-uniform, then for constants weights $$\gamma _i=1$$ one has $$c_h \simeq h^{1-d/2}$$ and $$C_h \simeq h^{-d/2}$$, whence $${\frac{C_h}{c_h}} \simeq h^{-1}$$. On the other hand, if a hierarchical basis is used instead, then $$c_h \simeq C_h \simeq 1$$, hence, $${\frac{C_h}{c_h}} \simeq 1$$ in dimension $$d=1$$, whereas $$c_h \simeq \vert \log h\vert ^{-1}$$, $$C_h \simeq 1$$, hence, $$\frac{C_h}{c_h} \simeq \vert \log h\vert$$ in dimension $$d=2$$.

Thus, the presence of the ratio $$\frac{C_h}{c_h}$$ in (4.35), which originates from the control of the loss function, makes this estimate sub-optimal. However, our numerical experiments in Sect. 6.1 indicate that this adverse effect is not seen in practice. The reason may be related to the decay of the loss function $$R_h(u^{{\mathcal {N}}\!{\mathcal {N}}})$$, which is significantly faster than the decay of the approximation error when h is reduced, thereby compensating for the growth of ratio. See Remark 5.

### Remark 4

(low-regularity $${{\mathcal {N}}\!{\mathcal {N}}}$$) When the condition $$U^{{\mathcal {N}}\!{\mathcal {N}}} \subset \mathrm{H}^2(\Omega )$$ fails to be satisfied, as for the ReLU activation function, we may provide a different set of assumptions which still lead to an $$O(h^k)$$-error estimate as in Theorem 6. Precisely, we may assume that $${\tilde{u}} \in \mathrm{W}^{k+1,\infty }(\Omega )$$ and $$U^{{\mathcal {N}}\!{\mathcal {N}}} \subset \mathrm{W}^{1,\infty }(\Omega )$$. Then, referring to the first equality in (4.29), one has

\begin{aligned} \Vert u-w_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }\le & {} \Vert (u-{\mathcal {I}}_H u) \Vert _{1,\Omega } + \Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }\\\lesssim & {} \Vert (u-{{\mathcal {I}}}_H u) \Vert _{1,\Omega } + H^{-1}\Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{0,\Omega } \,, \end{aligned}

with

\begin{aligned} \begin{aligned} \Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{0,\Omega }^2&= \sum _{G \in {{\mathcal {T}}}_H} \Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{0,G}^2 \le \sum _{G \in {{\mathcal {T}}}_H} \Vert {{\mathcal {I}}}_H e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{\mathrm{L}^\infty (G)}^2 \vert G\vert \\&\lesssim \sum _{G \in {{\mathcal {T}}}_H} \Vert e^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{\mathrm{L}^\infty (G)}^2 \vert G\vert \lesssim H^{d} \Vert e^{ {\mathcal {N}}\!{\mathcal {N}}} \Vert _{\mathrm{L}^\infty (\Omega )}^2 \,. \end{aligned} \end{aligned}

The conclusion easily follows if $${\tilde{w}}^{{\mathcal {N}}\!{\mathcal {N}}}$$ is chosen to satisfy the error bound $$\Vert {\tilde{u}}-{\tilde{w}}^{{\mathcal {N}}\!{\mathcal {N}}}\Vert _{\mathrm{L}^\infty (\Omega )} \lesssim H^{k+1} \vert {\tilde{u}}\vert _{\mathrm{W}^{k+1,\infty }(\Omega )}$$ , which is possible according to the results in [6, 7].

## 5 Implementation Issues

As specified in Sect. 3.1, we use a fully-connected feed-forward neural network architecture, which is fixed and depends neither on the PDE nor on its discretization. For each simulation, we initialize the neural network with a completely new set of weights, which is important to show that our results are not initialization dependent. The activation function is the hyperbolic tangent. It has been proven in [20] that such neural networks with two hidden layers enjoy exponential converge properties with respect to the number of weights. Nevertheless, in order to simplify the training and enrich the space in which we seek the numerical solution, we consider five layers (namely, $$L=5$$ with the notation of Sect. 3.1) with 50 neurons each ($$N_\ell =50,$$ $$\ell =1,...,5$$). We also highlight that it is always possible to approximate the identity function with a neural network with a single layer with just one neuron. The best approximation obtainable with a neural network with more than two layers is thus more accurate than the one computable with a neural network with just two layers, because, in the worst possible case, the L-layers neural network can be obtained by combining an $$(L-2)$$-layers identity neural network with a suitable 2-layers neural network. Numerical tests have been performed to investigate the influence of the activation function on the model accuracy; we observed that all the commonly used activation functions led to equivalent results, thus we omit such a comparison from the present work. We remark that the hyperparameters $$L=5$$ and $$N_\ell =50$$ have been chosen to obtain a neural network sufficiently large to satisfy condition (4.32) on all grids used in our experiments. Numerical evidence that the neural network best approximation error is negligible when compared with other sources of error is presented in Sect. 6.2.

In order to compute the results shown in Sect. 6, we compared various state of the art optimizers to find the most efficient way to minimize the loss function. We observe that most of the momentum-based first-order methods have similar performances (the presented results are computed with the ADAM optimizer [21]), but it is convenient to use a learning rate scheduler in order to reduce the learning rate during the process. We tested both cyclical learning rate schedulers and exponential learning rate schedulers [22], the differences were very subtle and we thus choose to adopt the most common exponential learning rate scheduler and decided not to report images about such a comparison. To further reduce the loss function we also use the BFGS method and its limited-memory version: the L-BFGS method [23].

The Dirichlet boundary conditions are imposed via the mapping B defined in (2.5). The construction of the function $$\Phi$$ is particularly simple when $$\Omega$$ is a convex polygon, since in this case $$\Phi$$ can be defined as the product of the linear polynomials which vanish on each Dirichlet edge; this is precisely how we define $$\Phi$$ in the numerical examples discussed in the next session. In other geometries, one can build $$\Phi$$ either as described in [17], or by using a level-set method, or even by training an auxiliary neural network to (approximately) vanish on $$\Gamma _D$$. Similarly, in order to obtain an analytical expression of the extension $${\overline{u}}$$ of the Dirichlet data g, one can train another neural network to (approximately) match the values of g on $$\Gamma _D$$ or use a data transfinite interpolation [24].

### 5.1 VPINN Efficiency and the Inf-sup Condition

In Sect. 3 we introduced a discretization method in which the loss function is built by a piecewise polynomial interpolation of the neural network; on the other hand, we also mentioned in Remark 2 the possibility of building the loss function directly from the (non-interpolated) neural network.

From the theoretical point of view, only the former approach can be considered mathematically reliable, since the error control is based on the validity of an inf-sup condition, as detailed in Sect. 4. On the contrary, if the neural network is used without interpolation, one usually gets an under-determined system, for which the error control may be problematic. In fact, for instance, the discrete solution with zero data may not be identically zero, as documented in Sect. 6.3, which rules out uniqueness. Nonetheless, there is empirical evidence (see, e.g., [25,26,27,28]) that non-interpolated neural networks do succeed in computing accurate solutions even in complex scenarios. Actually, in the next section we will provide numerical evidence that the two approaches are always (in the considered cases) comparable in terms of rate of convergence and, when the solution is regular, smaller errors are obtained minimizing the same loss function without interpolation.

From the computational point of view, the two approaches have comparable advantages and disadvantages. Let us first consider non-interpolated VPINNs. The corresponding loss functions can be more easily implemented thanks to the existing deep-learning frameworks, which allow the direct computation of neural network derivatives via automatic differentiation [29]. One only needs to generate a mesh and the corresponding test functions, associate a quadrature rule with each element and assemble all the tensors required to efficiently compute the loss function. The main difference with the interpolated neural network approach is that, in the latter, the interpolation matrices have to be assembled too (see “Appendix A.1” for a detailed description of the construction of the interpolation operators), while automatic differentiation is not required. Depending on the problem at hand, this may be an advantage or not. Indeed, the interpolation matrices assembly may be tricky but, using them, all derivatives can be efficiently computed by matrix-vector multiplications that are much cheaper than the entire automatic differentiation procedure, especially when higher order derivatives are required. Therefore, for fast-converging optimization processes, a non-interpolated neural network approach may be efficient and can be more easily implemented; otherwise, each optimization step may be much more expensive than the analogous operation performed with an interpolated neural network. Furthermore, we observed that the training phase is faster when the neural network is interpolated because the procedure converges in fewer steps. This is probably related to the fact that the solution is sought in a significantly smaller space, that can be more easily explored during the training phase.

## 6 Numerical Results

In this section we present several numerical results concerning the VPINN discretization of Problem (2.1) in the square $$\Omega =(0,1)^2$$. We will vary the coefficients of the operator, the boundary conditions and the smoothness of the exact solution. For each test case, we vary the degree $$k_\text {test}$$ of the test functions, the order q of the quadrature rule and, correspondingly, we choose the polynomial degree k of the interpolating functions as $$k=k_\text {int}=q+2-k_\text {test}$$, according to (4.7). We only report results obtained with Gaussian rules, as Newton-Cotes formulas of the same order give comparable errors (see [30] for a larger set of numerical experiments about this and other comparisons).

The theoretical results in Sect. 4 suggest that it is convenient to maintain $$k_\text {test}$$ as low as possible; consequently, we only use piecewise linear ($$k_\text {test}=1$$) or piecewise quadratic ($$k_\text {test}=2$$) test functions. Recalling condition (3.3), we thus choose $$q=3$$ or $$q=5$$ if $$k_\text {test}=1$$, and $$q=5$$ if $$k_\text {test}=2$$.

The triangulations we use are generic Delaunay triangular meshes. In order to satisfy the discrete inf-sup condition, we choose $$\mathcal T_H$$ and $${\mathcal {T}}_h$$ as nested meshes whose meshsizes satisfy $$H=k_\text {int}h$$. A pair $$({\mathcal {T}}_H, {\mathcal {T}}_h)$$ of used meshes is represented in Fig. 1, together with the elemental refinement corresponding to $$k_{\text {int}}=4$$, $$k_{\text {int}}=5$$ and $$k_{\text {int}}=6$$.

### 6.1 Error Decays

Hereafter, we empirically confirm, with numerical experiments, the a priori error estimate established in Sect. 4. We also compare the behavior of the proposed NN with that of other NNs defined by different strategies. In the following, we denote the interpolated VPINN as IVPINN to distinguish it from the non-interpolated VPINN [14], simply denoted as VPINN, and the standard PINN [1].

In the subsequent plots, we report by blue dots the error $$\Vert u - u_H^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }$$, where $$u_H^{{\mathcal {N}}\!{\mathcal {N}}}$$ is the interpolated VPINN defined on the mesh $${\mathcal {T}}_H$$ as in (4.1), versus the size H of the mesh $${\mathcal {T}}_H$$. We also show a blue solid line and a blue dashed one: the former is the regression line fitting the blue dots (possibly ignoring the first ones); its slope in the log-log plane yields the empirical convergence rate. The latter is used as a reference, since its slope corresponds to an error decay proportional to $$h^{k_\text {int}}$$, which is the expected convergence rate of the $$H^1$$ error as indicated by Theorem 6, assuming that the ratio $$\frac{C_h}{c_h}$$ may be neglected (see Remark 5). The dashed line represents the best convergence rate we can expect from the proposed discretization scheme.

For comparison, we also report by green dots the error $$\Vert u - {\hat{u}}^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }$$, where $${\hat{u}}^{{\mathcal {N}}\!{\mathcal {N}}}$$ is the non-interpolated VPINN defined in Remark 2, and by red dots the error $$\Vert u - {\widetilde{u}}^{{\mathcal {N}}\!{\mathcal {N}}} \Vert _{1,\Omega }$$, where $${\widetilde{u}}^{{\mathcal {N}}\!{\mathcal {N}}}$$ is the standard PINN proposed in [1], with the same architecture of the used VPINNs and the loss function computed as described in [31]. To obtain a fair comparison, the regularization coefficient and the ratio between the control points inside the domain and the ones on the boundary are chosen as described in [31]. Since we are interested in convergence rates with respect to mesh refinement, but the PINN does not require any mesh, the corresponding errors are computed by training the network with the same numbers of inputs used during the training of $$u_H^{{\mathcal {N}}\!{\mathcal {N}}}$$; to be precise, the PINN is trained using the same number of collocation points as the number of interpolation nodes used by the interpolated VPINN.

Furthermore, in order to better analyze the trade-off between the model accuracy and the training efficiency and complexity, we plot the same errors versus the dataset size. Whenever these dots, possibly after a pre-asymptotic phase, sit close to their regression line, we draw it as well in green or red, respectively.

Convergence test #1: $$u\in C^\infty ({\bar{\Omega }})$$

Consider problem (2.1) with $$\Gamma _D = \{(x,y)\in \partial \Omega :x=0\text { or }x=1\}$$ and $$\Gamma _N = \partial \Omega \backslash \Gamma _D$$. Let us choose the following operator coefficients

\begin{aligned} \mu (x,y) = 2+\sin (x+2y),\qquad \beta (x,y) = \begin{bmatrix} \sqrt{x-y^2+5} \\ \sqrt{y-x^2+5} \end{bmatrix}, \qquad \sigma (x,y)=e^{\frac{x}{2}-\frac{y}{3}}+2\,, \end{aligned}

and the data f, g, and $$\psi$$ such that the exact solution is

\begin{aligned} u(x,y) = \sin (3.2x(x-y))\cos (4.3y+x)+\sin (4.6(x+2y))\cos (2.6(y-2x)). \end{aligned}

The corresponding error decays with respect to the meshsize H are shown in Fig. 2.

In Fig. 2a, where the IVPINN (blue dots) and the VPINN (green dots) are trained with $$q=3$$ and $$k_{\text {test}}=1$$, we observe that the points are distributed, possibly after an initial preasymptotic phase, along straight lines with slopes very close to $$k_{\text {int}}=4$$. We highlight that the PINN convergence is significantly noisier and that the corresponding $$H^1$$ error is, on average, about 7 times the IVPINN one. A similar phenomenon can be seen in Fig. 2b, although the finite precision of the used Tensorflow software prevents convergence to display at full for small values of H. In this test we use $$q=5$$ and $$k_{\text {test}}=1$$ and the regression lines for the IVPINN and the VPINN have slopes close to $$k_{\text {int}}=6$$, while the PINN accuracy is again much lower. Finally, the data in Fig. 2c are obtained with $$q=5$$ and $$k_{\text {test}}=2$$ and the blue regression line slope is 5.05, almost coinciding with $$k_{\text {int}}=5$$.

Such results highlight that, although the VPINN implementation is more complex than the PINN one, the former produces more accurate solutions than the latter, when the exact solution is regular.

In Fig. 3, the same error decays are expressed in terms of the number of training points, i.e., the number of neural network forward evaluations required to construct the loss function in a single epoch. Such an alternative visualization highlights that the performances of the IVPINN and the VPINN are very similar when trained with similar training sets. This is due to the fact that, since we stabilize the VPINN by projecting it on a space of continuous piecewise polynomials, we need fewer interpolation points (input data of the IVPINN) than quadrature points (input data of the VPINN) to evaluate the loss function.

### Remark 5

(on the quotient $$\frac{C_h}{c_h}$$) Theorem 6 indicates that the best possible convergence rate, when the solution is regular enough, is $$k_\text {int}$$. However, as discussed in Remark 3, the quotient $$\frac{C_h}{c_h}$$ is of order $$O(h^{-1})$$ when test functions are picked from the Lagrange basis associated with a quasi-uniform triangulation and the weights $$\gamma _i$$ are equal to 1. In this case, the term $$\left( 1+\frac{C_h}{c_h}\right)$$ in (4.35) reduces the predicted convergence by exactly one order.

On the other hand, in Fig. 2, we have shown cases where the order of convergence is optimal. Such a behavior is related to the fact that the loss $$R_h(u^{{\mathcal {N}}\!{\mathcal {N}}})$$ decays much faster than expected, in the considered cases, namely at least as $$O(h^8)$$. Therefore, when h is small enough, the term $$C_h R_h(u^{{\mathcal {N}}\!{\mathcal {N}}})$$ in Lemma 4.1 can be neglected, and the predicted convergence rate is not affected by the presence of the quotient $$\frac{C_h}{c_h}$$.

Convergence test #2: $$u\in H^{5/3-\varepsilon }(\Omega )$$

Let us now focus on a less smooth solution, whose regularity is commonly found in domains with reentrant corners. The problem is characterized by $$\Gamma _D=\partial \Omega$$, $$\mu =1$$, $$\beta =[2,3]^T$$, $$\sigma =4$$, whereas the forcing term and boundary conditions are such that the exact solution is, in polar coordinates,

\begin{aligned} u(r,\theta ) = r^\frac{2}{3} \sin \left( \frac{2}{3}\left( \theta +\frac{\pi }{2}\right) \right) . \end{aligned}

Since $$u\in H^{5/3-\varepsilon }(\Omega )$$ for any $$\varepsilon >0$$, we expect a convergence rate close to 2/3; indeed, $$k_{\text {int}}\ge 4$$ and the rate of convergence is always limited by the solution regularity as expected. The error decays are shown in Fig. 4. Notice that the IVPINN is even more stable and accurate than the VPINN trained on the same mesh (i.e., with more input data). The PINN behaves better in this test case than in the previous one, but still the accuracy is worse than the one provided by our IVPINN.

It is also interesting to analyze the behavior of the loss function and of the error during training as documented in Fig. 5, where the first 3000 epochs are performed with the ADAM optimizer, while the remaining ones with the BFGS optimizer. Such plots correspond to the loss function and the $$H^1$$ error associated with the dots marked by the black stars in Fig. 4a. It can be noted that the IVPINN and the VPINN initially converge very fast with the ADAM optimizer; eventually, after the initial phase in which both the loss and the error decrease, the error reaches a constant value despite the loss function keeps diminishing. This implies that there exist other sources of error that prevail when the loss function decays. On the other hand, using a standard PINN, one observes that the convergence of the loss and the error is much slower than for the VPINNs, and the second-order optimizer is needed to converge to an accurate solution. The average epoch execution time is approximately 0.0599 seconds for the PINN, 0.0587 seconds for the VPINN, and 0.0479 seconds for the IVPINN. Such a gain is due to the fact that the model derivatives are computed via automatic differentiation in the non-interpolated models, while the gradient of the IVPINN can be computed by a simple matrix-vector multiplication. Note that the gain increases when higher derivatives are involved in the PDE.

### 6.2 How the VPINN Dimension Affects Accuracy

We now focus on the dependence of the error on the neural network dimension. For the sake of simplicity, we fix the problem discretization and vary only the number of layers and the number of neurons in each layer, assuming that each layer contains exactly the same number of neurons. The considered domain, parameters, forcing term and boundary conditions are the ones described in Convergence test #1. The VPINN is trained with piecewise linear test functions ($$k_{\text {test}}=1$$) and quadrature rules of order $$q=3$$ on the finest mesh used to produce Fig. 2a and on the mesh associated with the blue dot close to the black star in the same figure.

We can observe, in Fig. 6, that the error is very high for small networks, but then it rapidly decreases while increasing the number of neurons in each layer, until a plateau is reached depending on the chosen problem discretization. Essentially, on both meshes, 3 layers with 10 neurons each suffice to achieve the lowest possible discretization error for the given loss function.

This analysis confirms that the error decays reported in Sect. 6.1 are all insensitive to the neural network hyper-parameters, as they have been obtained by a large neural network (5 layers with 50 neurons each). Such results validate the assumption made in Sect. 4 about the neural network, namely that its dimension – provided it is sufficiently large – does not influence the predicted convergence rate.

### 6.3 On the Importance of the Inf-sup Condition

In this section we show that the inf-sup condition, assumed in Proposition 4 to derive the a priori error estimate, is crucial in order to avoid spurious modes in the numerical solution. To prove such a claim, let us consider the simplest one-dimensional Poisson’s problem with zero forcing term and zero Dirichlet boundary conditions:

\begin{aligned} \left\{ \begin{aligned} -&u'' = 0&\text {in } \Omega \,, \\&\ \, u=0&\text {on } \Gamma _D \,, \end{aligned}\right. \end{aligned}
(6.1)

where $$\Omega =(0,1)$$ and $$\Gamma _D=\{0,1\}$$. For the sake of simplicity, we use piecewise linear test functions and quadrature rules of order $$q=3$$. Note that since both $$U^{{\mathcal {N}}\!{\mathcal {N}}}$$ and $$U_H$$ contain the exact solution $$u\equiv 0$$, it is always possible to obtain a numerical solution that is identical to the exact one (up to numerical precision).

Let us denote by $$u_\delta$$ any discrete solution defined in Sect. 3.2, namely, either a solution $$u_H^{ {\mathcal {N}}\!{\mathcal {N}}}$$ obtained by interpolated VPINNs, or a solution $${\hat{u}}^{{\mathcal {N}}\!{\mathcal {N}}}$$ obtained by non-interpolated VPINNs. These discrete solutions are represented in Fig. 7 in logarithmic scale to allow a direct comparison. In order to avoid numerical issues due to the logarithmic scale of the plot when $$u_{\delta }$$ gets close to 0, a truncation procedure is applied.

The functions $${u}_\delta$$ produced by non-interpolated networks are represented in the left plot of Fig. 7. Each one is obtained by minimizing the loss function up to machine accuracy; despite this, when the mesh is fairly coarse the discrete solution is significantly different from the null solution. Indeed, the initial weights in the training process are non-zero, and the minimization process is under-determined, thereby allowing the existence of non-zero global minima. Refining the mesh, the approximation improves up to a maximum precision imposed by the chosen network architecture and the Tensorflow deep learning framework.

Conversely, the plots in the subfigure on the right-hand side of Fig. 7, produced by interpolated networks, clearly indicate that the obtained discrete solutions are numerically zero, irrespective of the meshsize. Note that the case $$h=1/4$$ differs from the others since here the interpolation mesh $${{\mathcal {T}}}_H$$ is formed by just one element. The corresponding function $$u_H^{{\mathcal {N}}\!{\mathcal {N}}}$$ is thus differentiable everywhere and the used gradient-based optimizers are able to minimize the loss function more effectively. We highlight that, since $$q=3$$ and $$k_{\text {test}}=1$$, we have to choose $$H=4h$$ to satisfy the inf-sup condition.

To illustrate the mechanism that may lead to the onset of spurious modes, in “Appendix A.2” we provide an analytical example of a neural network which significantly differs from a PDE solution, yet it is a global minimizer of the corresponding loss function.

These results, although obtained in overly simple functional settings, show the potential existence of uncontrolled components in the discrete solutions obtained by non-interpolated neural networks. In more complex scenarios, the presence of spurious modes may be even more pronounced. In practice, as observed in Fig. 2, when the PDE solution is smooth enough non-interpolated solutions appear to be more accurate than the corresponding solutions obtained by interpolated neural networks using the same test functions; however, a rigorous analysis of NN-based discretization schemes should also cope with the presence of spurious components, which we have avoided by resorting to an inf-sup condition. We believe that these observations shed new light on the use of deep learning in numerical PDEs.

## 7 Application to Nonlinear Parametric Problems

In the previous sections, we investigated the features of the proposed VPINN discretization for the linear boundary-value problem (2.1). Hereafter, we provide an application where the nonlinear nature of neural networks can be exploited at best. It is well known that solving nonlinear PDEs by PINNs or VPINNs comes at little extra cost with respect to linear PDEs, since nonlinearities just impact the computation of the loss function. Similarly, parametric problems can be easily and efficiently handled by neural networks, even when the dependence of the solution upon the parameters is nonlinear. Indeed, it is enough to add as many inputs as the number of parameters in the definition of the network, and train it on a proper subset of the parameter space.

To illustrate the behavior of our VPINN in these situations, let us consider the following nonlinear parametric equation:

\begin{aligned} {\left\{ \begin{array}{ll} -\nabla \cdot (\mu \nabla u) + \varvec{\beta }\cdot \nabla u + \sigma \, \mathrm{e}^{-pu^2} =f &{} \text {in } \Omega \,, \\ u=g &{} \text {on } \partial \Omega \,, \end{array}\right. } \end{aligned}
(7.1)

where $$\mu , \beta , \sigma , f$$ and g are suitably smooth functions, and $$p\in \Omega _p\subset {\mathbb {R}}$$ is an additional parameter. Our goal is to train a neural network to compute the numerical solution for any given value of p in a prescribed parametric domain $$\Omega _p$$.

We fix $$\Omega =(0,1)^2$$, $$\mu =1$$, $$\beta =[2,3]^T$$, $$\sigma =4$$ and choose $$f=f(\cdot ,p)$$ and $$g= g(\cdot ,p)$$ such that the exact solution is

\begin{aligned} u(x,y;p) = \frac{\cos \left( 5\left( px+\frac{y}{2}\right) \right) }{1+p} + \left( x+\frac{y}{2}\right) ^2. \end{aligned}

To approximate such a solution in the parametric domain $$\Omega _p=[0.5,2]$$, we consider a neural network with three inputs (xyp) and we train it using the loss function:

\begin{aligned} R_h^2(w) = \sum _{p\in \Omega _p^\#}\sum _{i \in I_h} r_{h,i;p}^2(w) \, \gamma _i^{-1}\,, \end{aligned}
(7.2)

where $$\Omega _p^\#=\{p_1,...,p_{n_p}\}\subset \Omega _p$$ is a finite set of parameter values, and the residuals $$r_{h,i;p}$$ are defined as in (3.10), considering the new equation. In this numerical test $$\Omega _p^\#$$ contains $$n_p=13$$ equally spaced values $$0.5=p_1<p_2<...<p_{n_p}=2$$.

After the training phase, the neural network can be evaluated at new parameter values to analyze its accuracy. The error diagram is presented in Fig. 8: the blue line is computed as the $$H^1$$ error between the exact solution $$u=u(\cdot ,p)$$ and the corresponding numerical solution, while the red dots represent the error associated with parameters in $$\Omega _p^\#$$. Despite the small number of parameter values used during training, the model provides accurate solutions for the whole range of parameter values. Note that the error increases for larger values of p because the solution is more and more oscillating as p increases.

## 8 Conclusions

We have investigated VPINN methods for elliptic boundary-value problems, in what concerns the choice of test functions and quadrature rules. The aim was the derivation of rigorous a priori error estimates for some projection of the neural network solution. The neural network is trained using as test functions finite-element nodal Lagrange basis functions of degree $$k_\text {test}$$ on a mesh $${{\mathcal {T}}}_h$$, where Gaussian quadrature rules of order q are applied in each element of $${{\mathcal {T}}}_h$$. For a fixed neural network architecture with tanh activation function, we studied how the error in the energy norm depends upon the mesh parameter h, for different values of $$k_\text {test}$$ and q.

Error control was obtained for the finite-element interpolant of degree $$k_\text {int}=q+2-k_\text {test}$$ of the neural network on an auxiliary mesh $${{\mathcal {T}}}_H$$; such an interpolation enters also in the definition of the residuals which are minimized through the loss function. A key ingredient in the error control is the validity of an inf-sup condition between the spaces of test functions and interpolating functions. Indeed, the neural network solution might be affected by spurious modes due to the under-determined nature of the minimization problem, as we documented for a problem with zero data; instead, the onset of such modes is prevented by the adopted interpolation procedure.

Our analysis reveals that the convergence rate in the energy norm is at least of order $$q+1-k_\text {test}$$ for sufficiently smooth functions, and it increases to $$q+2-k_\text {test}$$ when the value of the loss function obtained by minimization is sufficiently small. The main message stemming from the analysis is that it is convenient to choose test functions of the lowest degree $$k_\text {test}=1$$ in order to get the highest convergence rate for a fixed quadrature rule. Furthermore, for smooth solutions the convergence rate may be arbitrarily increased by increasing the precision of the quadrature rule, although the realization of this theoretical statement is hampered in practice by the finite precision of machine arithmetics.

We also investigated the influence of the neural network hyperparameters on the overall accuracy of the discretization, and we found that a small network with few layers and neurons suffices to reach accuracies of practical interest. To stay on the safe side, we used a larger network in our experiments, thereby obtaining results that are essentially independent of the network hyperparameters.

For the sake of comparison, we also implemented a standard VPINN without projection upon piecewise polynomials, as well as a standard PINN trained with the same number of inputs as those used in training our VPINN. Interestingly, we experimentally observed that in general the error decay rate for the non-interpolated neural network solution replicates the one theoretically predicted for the interpolated network. The PINN solutions appear to be less accurate and noisier than the interpolated VPINN’s.

We have shown that interpolated VPINNs are able to efficiently solve nonlinear parametric problems without the need for additional nonlinear solvers or globalization methods, due to their intrinsic nonlinear nature. The VPINN can be trained in an off-line phase on a subset of the parameter domain and then efficiently evaluated on-line on any other parameter value. This is a key difference between the proposed method and standard numerical techniques such as FEM, even if the solution is sought in the same finite dimensional space. Indeed, the latter would require some iterative technique to handle nonlinearities, as well as some form of interpolation/extrapolation to get the solution for the whole range of parameters. All this is provided for free by the NN machinery.

Possible extensions of this work are related to the investigation of more advanced neural networks architectures to improve the method accuracy and efficiency [32, 33] or to more complex problems. Indeed, neural networks are known to be able to manage very high-dimensional problems, overcoming the so-called curse of dimensionality, therefore we expect them to be able to efficiently solve parametric PDEs with multiple parameters [33] or high-dimensional PDEs [4, 34]. Other possible applications are related, for instance, to inverse problems [35] or integration between PDEs and data [36].