Abstract
Globalization concepts for Newton-type iteration schemes are widely used when solving nonlinear problems numerically. Most of these schemes are based on a predictor/corrector step size methodology with the aim of steering an initial guess to a zero of f without switching between different attractors. In doing so, one is typically able to reduce the chaotic behavior of the classical Newton-type iteration scheme. In this note we propose a globalization methodology for general Newton-type iteration concepts which changes into a simplified Newton iteration as soon as the transformed residual of the underlying function is small enough. Based on Banach’s fixed-point theorem, we show that there exists a neighborhood around a suitable iterate \(x_{n}\) such that we can steer the iterates—without any adaptive step size control but using a simplified Newton-type iteration within this neighborhood—arbitrarily close to an exact zero of f. We further exemplify the theoretical result within a global Newton-type iteration procedure and discuss further an algorithmic realization. Our proposed scheme will be demonstrated on a low-dimensional example thereby emphasizing the advantage of this new solution procedure.
Similar content being viewed by others
Avoid common mistakes on your manuscript.
1 Introduction
For the time being, let \(U\subset {\mathbb {R}}^{n}\) be open and \(f:U\rightarrow {\mathbb {R}}^{n}\) be of class \(C^{1}(U;{\mathbb {R}}^{n})\). In this note we are interested in finding the zeros \(x\in U\) of f, i.e., we aim to solve the equation
In general—apart from trivial toy problems—the solutions \(x_{\infty }\) can only be computed numerically. Here, we focus on the following approach: For \(x\in U\) we consider the matrix-valued map \(x\mapsto \mathsf {M} (x) \in {\mathbb {R}}^{n\times n}\) and define \(\mathsf {F} (x):=-\mathsf {M} (x)^{-1}f(x)\). Supposing that \(\mathsf {M} (x)\) is invertible on a suitable subset of U, we now concentrate on the initial value problem
This initial value problem tackles the problem of finding the zeros of f from a dynamical system approach. In fact, if \(\mathsf {M} (x)\) is given by the Jacobian of f we recover the well known continuous Newton scheme formally satisfying \(f(x(t))=f(x_0)\mathrm {e}^{-t}\). For an excellent survey of the continuous Newton scheme see, e.g., [8, 14,15,16]. Indeed, supposing that a solution x(t) exists for all time \(t\ge 0\), we can try to follow the trajectory of x(t) numerically in order to end up with an approximate root for f. For an initial guess \(x_{0}\in U\) the simplest routine for solving (2) numerically is given by the forward Euler method:
For example, if we choose \(\mathsf {M} (x):=\mathsf {Id} \), the above iteration scheme is termed Piccard-Iteration. If \(\mathsf {J} _{f}(x)\) signifies the Jacobian of f at \(x\in U\), then for \(\mathsf {M} (x)=\mathsf {J} _{f}(x)\) we observe a damped Newton-method. Specifically, for \(t_{n}\equiv 1\) and \(\mathsf {M} (x)=\mathsf {J} _{f}(x)\), the iteration (3) results in the well known standard Newton method. In addition, another well established scheme is given by setting \(\mathsf {M} (x):=\mathsf {J} _{f}(x_0)\), which is also called simplified Newton method. The last choice simply freezes the information of the Jacobian throughout the whole iteration procedure. This typically reduces the computational effort in each iteration step. On the other hand, the number of iterations increases in general and the domain of convergence is reduced by this method. However, on a local level, i.e., when the initial guess \(x_0\) is supposed to be ‘sufficiently’ close to a zero of f, it is reasonable to expect that the simplified Newton method safely leads to a zero which is located next to the initial guess \(x_0\). Indeed, if the update \(\mathsf {M} (x_n)^{-1}f(x_{n})\) is small enough, we will see in Sect. 2 that there exists a unique zero for f locally that can be obtained by the following simplified Newton-type iteration scheme:
This observation is especially interesting when the computation of the matrix \(\mathsf {M} (x_{n})\) is computationally expensive—as for instance when we solve extremly large scale nonlinear problems arising from the discretization of PDE’s. Furthermore, the proposed result in this work asserts local uniqueness of the solution. Thus, one can think of steering an initial guess \(x_0\in U\) assumed to be far away of a zero for f, ‘sufficiently’ close to the root which is located next to \(x_{0}\). Having hit the domain of local uniqueness of the underlying zero we then switch from the adaptive iteration (3) to the simplified iteration scheme given in (4) without using any adaptive step-size control.
1.1 Notation
In this note we signify by \((\cdot ,\cdot )\) the standard Euclidean product of \({\mathbb {R}}^{n}\). For any x its norm is given by \(\left\| x\right\| :=\sqrt{(x,x)}\). For a matrix \(\mathsf {M} \in {\mathbb {R}}^{n\times n}\) we further use the operator norm \(\left\| \mathsf {M} \right\| :=\sup _{\left\| x\right\| =1}{\left\| \mathsf {M} x\right\| }\). By \(B_{R}(x)\) we denote the closed ball of radius R centered at \(x\in {\mathbb {R}}^{n}\). Finally, whenever the function f is differentiable, the derivative at a point \(x\in U\) is written as \(\mathsf {J} _{f}(x)\), thereby referring to the Jacobian of f at x.
1.2 Outline
This note is organized as follows: In Sect. 2 we state and prove a convergence result for a general class of simplified Newton-type iterations schemes as given in (4). Therefore we firstly discuss the assumptions that have to hold true in order to establish the proposed convergence result. In particular, we embed the local convergence result into a global—and therefore adaptive—Newton-type iteration scheme as given in (3). On that account, in Sect. 3 we finally present and discuss our adaptive strategy on a low dimensional example employing the advantage of the proposed iteration scheme. In Sect. 4 we summarize and comment our findings.
2 A convergence result
As a preparation towards the proposed main result we firstly address the assumptions that have to hold. In addition, we comment on a possible extension of the proposed result to a general Banach space framework.
2.1 Assumptions
Suppose we are given an initial value \(x_{0}\in U\) and suppose we can compute
Here, \(t_{j}\) signifies some adaptively chosen step size (see, e.g., [1, 3, 18, 19] for some highly efficient step size methodologies). Again, we notice that for \(\mathsf {M} (x)=\mathsf {J} _{f}(x)\) and \(t_{j}\equiv 1 \), the iteration (5) is simply the standard Newton method.
Let U be an open and convex subset of \({\mathbb {R}}^{n}\) and assume further that there exists an iterate \(x_{n}\in U\) such that there holds the following assumptions:
- A1.:
-
Let \(\omega \) be a positive constant. For any \(v \in U\) and for any \(z \in \{tx_n+(1-t)v|t\in [0,1]\}\) we assume that there holds the following affine covariant type Lipschitz-condition on \(\mathsf {J} _{f}\):
$$\begin{aligned} \left\| \mathsf {M} (x_n)^{-1}(\mathsf {J} _{f}(x_n)-\mathsf {J} _{f}(z))(x_n-v)\right\| \le \omega (1-t)\left\| x_n-v\right\| ^{2}. \end{aligned}$$(6) - A2.:
-
We further need \(\mathsf {M} (x_n)^{-1}\) to be a sufficiently accurate approximate of the inverse of the Jacobian \(\mathsf {J} _{f}(x_n)\) which we here quantify by the following assumption
$$\begin{aligned} \left\| \mathsf {Id} -\mathsf {M} (x_n)^{-1}\mathsf {J} _{f}(x_n)\right\| \le \kappa <1. \end{aligned}$$ - A3.:
-
For \(\alpha _{n}:=\left\| \mathsf {M} (x_{n})^{-1}f(x_{n})\right\| \) we need to assume that
$$\begin{aligned} \omega \alpha _{n} \le \frac{(1-\kappa )^2}{2}. \end{aligned}$$(7) - A4.:
-
For
$$\begin{aligned} R:=\frac{1-\kappa }{\omega }+ \sqrt{\frac{(1-\kappa )^2}{\omega ^2} -\frac{2\alpha _n}{\omega }} \end{aligned}$$(8)there holds \(B_{R}(x_{n})\subset U\).
Assumption A1 is called affine covariant type Lipschitz condition because in case of \(\mathsf {M} (x_n)=\mathsf {J} _{f}(x_n)\) the Lipschitz constant \(\omega \) is an affine invariant quantity. Indeed, for \(\mathsf {A} \in \text {Gl}(n)\) and \(\mathsf {F} (x):=\mathsf {A} f(x)\) there holds
For further details concerning affine invariance principles within the framework of Newton-type iterations schemes we refer to the excellent monograph [8] and the proposed adaptive schemes therein.
Supposing that \(\mathsf {M} (x_{n})^{-1}\) is bounded, then condition (7) in A3 also holds true whenever the residual \(\left\| f(x_{n})\right\| \) is ‘sufficiently’ small in the sense that
Thus, the proposed result implies that whenever the norm of the residual \(\left\| f(x_{n})\right\| \) is small enough, there exists a zero on a local level. This is of particular interest when solving nonlinear differential equations numerically within the context of a fully adaptive iteration scheme. More precisely, let X denote a Banach space—in most cases \(X=H_{0}^{1}(\Omega )\)—and \(X'\) its dual respectively. Then the weak formulation of a nonlinear differential equation reads as follows:
Find \(x\in X\) such that there holds
with \(\left\langle \cdot , \cdot \right\rangle _{X'\times X}\) signifying the duality pairing in \(X'\times X\).
Solving (10) within the context of an adaptive solution procedure over some finite dimensional space \(X_{h}\subset X\)—here h typically signifies the mesh-size parameter in the finite element method—one then can try to derive computational quantities \(\eta _\mathbf{h }(x_{h})\) and \(\eta _\mathbf{L }(x_{h})\) such that there holds:
Here, the quantity \(\eta _\mathbf{L }(x_{h})\) signifies an error estimate which measures the linearization error whereas \(\eta _\mathbf{h }(x_{h})\) represents the discretization error (see, e.g., [2, 4,5,6,7, 9,10,11,12,13]). Using (9) and supposing that the quantities \(\eta _\mathbf{h }(x_{h})\) and \(\eta _\mathbf{L }(x_{h})\) are small enough we obtain
i.e., the a posteriori existence of the solution is guaranteed. Indeed, the a posteriori existence in numerical computations has been addressed in detail by [17]—especially in the context of solving semilinear problems. However, although we discuss and present our adaptive scheme in view of dealing with systems of nonlinear equations over \({\mathbb {R}}^{n}\), it is noteworthy that the established convergence result also holds true within a general Banach space setting. Indeed, our convergence result can be used to realize a specialization of the recently established adaptive iterative linearized Galerkin methodology (ILG) discussed in [12, 13].
Theorem 1
Suppose that \(f \in C^{1}(U;{\mathbb {R}}^{n})\). Further assume that there holds the assumptions \( \mathbf{A1 } \& \mathbf{A2 } \& \mathbf{A3 }\) and \(\mathbf{A4 }\).
Then the map
satisfies
Proof
First of all we rewrite the function g as follows
Let \(v\in B_{R}(x_n)\). For \(t\in [0,1] \) we define the line segment \(z(t):=tx_{n}+(1-t)v \subset B_{R}(x_{n})\) and use the integral form of the mean value theorem
from where we obtain by A1&A2
Thus there holds
Employing \(\mathbf{A3} \), this last equality holds true if
\(\square \)
Let us go back to (14) in the proof. We see that the map g also satisfies
with \(r=\frac{1-\kappa }{\omega }-\sqrt{\frac{(1-\kappa )^2}{\omega ^2}-\frac{\alpha _{n}}{\omega }}\).
Next we give an existence result addressing the zeros \(u\in U\) of f.
Corollary 2
Assumptions and notations as in the preceding Theorem 1. Then, there exists a zero \(u \in B_{R}(x_{n}) \) of f.
Proof
From the proof of Theorem 1 we have that \(g(B_{R}(x_{n}))\subset B_{R}(x_{n})\). Employing Brouwer’s fixed point theorem we deduce the existence of a fixed point \(u\in B_{R}(x_{n})\) of g which is the asserted zero of f. \(\square \)
In view of the iteration procedure (4) it would be preferable if we can guarantee its convergence within the ball \(B_{R}(x_{n})\subset U\). Indeed, if g from (12) is a contraction in \(B_{R}(x_{n})\) we can conclude the existence of a unique fixed point of g which can be obtained by iterating (4). In doing so we need to strengthen the assumptions A1&A2&A3 and A4 as follows:
- B1.:
-
Let \(\omega ^{\star }\) be a positive constant. For any \(x,v \in U\) and for any \(z \in \{tx+(1-t)v|t\in [0,1]\}\) we assume that there holds the following affine covariant type Lipschitz-condition on \(\mathsf {J} _{f}\):
$$\begin{aligned} \left\| \mathsf {M} (x_n)^{-1}(\mathsf {J} _{f}(x)-\mathsf {J} _{f}(z))(x-v)\right\| \le \omega ^{\star } (1-t)\left\| x-v\right\| ^{2}. \end{aligned}$$(16) - B2.:
-
For any \(x \in U\) there holds:
$$\begin{aligned} \left\| \mathsf {Id} -\mathsf {M} (x_{n})^{-1}\mathsf {J} _{f}(x)\right\| \le \kappa ^{\star } <1. \end{aligned}$$(17) - B3.:
-
For \(\alpha _{n}=\left\| \mathsf {M} (x_{n})^{-1}f(x_{n})\right\| >0\) we need to assume that
$$\begin{aligned} \omega ^{\star } \alpha _{n} \le \frac{(1-\kappa ^{\star })^2}{2}. \end{aligned}$$(18) - B4.:
-
For
$$\begin{aligned} R^{\star }:=\frac{1-\kappa ^{\star }}{\omega ^{\star }}+ \sqrt{\frac{(1-\kappa ^{\star })^2}{{\omega ^{\star }}^2}-\frac{2\alpha _n}{\omega ^{\star }}} \end{aligned}$$(19)there holds \(B_{R^{\star }}(x_n)\subset U\).
Note that for \(x=x_{n}\) we have \(\omega =\omega ^{\star }\) and \(\kappa =\kappa ^{\star }\). Now we are ready to prove the following result:
Theorem 3
Suppose that \(f \in C^{1}(U;{\mathbb {R}}^{n})\). Further assume that there holds the assumption \( \mathbf{B1} \& \mathbf{B2} \& \mathbf{B3} \& \mathbf{B4} \). Then the map from (12) satisfies firtsly
and is a contraction on \( B_{R^{\star }}(x_{n})\).
Proof
The first assertion follows from the proof of Theorem 1 and choosing \(x=x_n\). Thus we are left to show that g is a contraction. Notice that
Thus, for \(x,y\in B_{R^{\star }}(x_{n})\) there holds:
Since \(\kappa ^{\star }+\omega ^{\star } \frac{1}{2}\left\| x-y\right\| \le \kappa ^{\star } +\frac{\omega ^{\star }}{2}R^{\star }\) and \(R^{\star }<\frac{2(1-\kappa ^{\star })}{\omega ^{\star }} \), there holds
i.e., we conclude that g is a contraction. \(\square \)
Corollary 4
Assumptions and notations as in the preceding Theorem 3. Then, for any initial value \(x_{n}\in U\) the simplified Newton-like iterates (4) remain in \(B_{R^{\star }}(x_{n})\) and converge to a unique zero \(u_{\infty }\in B_{R^{\star }}(x_{n})\) of f.
Proof
From the proof of Theorem 1 we have that for \(j\ge n \) the iterates \(u_{j+1}=g(u_{j})\) remain in \(B_{R^{\star }}(x_{n})\). Furthermore we have also shown that g is a contraction on \(B_{R^{\star }}(x_{n})\). Thus, by Banach’s fixed-point theorem we deduce that \(\lim _{j\rightarrow \infty }{g(u_{j})}=u_{\infty }\) exists, which is the unique zero of f in \(B_{R^{\star }}(x_{n})\). \(\square \)
From a computational point of view we can try to switch from the Newton-like iteration scheme (5) to a simplified Newton-like scheme
as soon as there holds \(\alpha _{n}\omega ^{\star }\le \frac{(1-\kappa ^{\star })^2}{2}\). Therefore we need to control the Lipschitz constant \(\omega ^{\star }\). In doing so, we replace the computational unavailable constant \(\omega ^{\star }\) by a quantity \(\hat{\omega }\) that we can easily compute during the iteration procedure. Henceforth, suppose we have computed \(x_{n+1}, x_{n}\). In view of (6), it is reasonable to switch to the iteration (20) whenever there holds
In addition, for \(\mathsf {M} (x_{n})=\mathsf {J} _{f}(x_n)\) and \(x\in B_{R^{\star }}(x_n)\) we observe
i.e. \(\kappa =0\).
3 Numerical experiments
3.1 Adaptive strategy
We now propose a procedure that realizes an adaptive strategy based on the previous observations. The individual computational steps are summarized in Algorithm 1.
Let us briefly comment on the proposed adaptive procedure given in Algorithm 1:
Remark 1
In steps \( 3 \& 18\) we predict a step size t such that \(t=1\) whenever the iterates are ‘close enough’ to the zero \(x_{\infty }\). Thus, the proposed procedure allows full steps whenever the iterates are ‘sufficiently’ close to \(x_{\infty }\). The computation of \(t\in (0,1]\) typically relies on a computational upper bound with respect to the distance \(\left\| x(t_{n})-x_{n}\right\| \). There exists different suggested approaches towards an effective computation of the step size t (see, e.g., [1, 3, 4, 8, 12, 18, 19]). Here we use the adaptive step size control given in [1]. This adaptive choice of the step size t consists mainly of two parts:
A prediction for the step size t and a correction of the step size whenever \(\left\| x_{n}-x(t_{n})\right\| > \tau \). Here, x(t) signifies the exact trajectory leading to the zero \(x_{\infty }\) and \(x_{n}\) is the numerical solution. Thus, the input \(\tau \) is a parameter that determines how close the iterates \(x_{n}\) tracks the exact trajectory x(t) leading to a zero of f (see also Fig. 1). For \(\tau =\infty \) there is no restriction on \(x_{n}\), i.e. Algorithm 1 reproduces the classical Newton scheme—apart from the simplified Newton scheme given in step 13. Furthermore, the adaptive scheme from [1] needs a lower bound \(t_{\text {lower}}\) for the step size \(t_{n}\) in (3). Indeed, if \(t_{n}\) degenerates to 0, the iterative scheme is not well defined in the sense that it must be classified as not convergent. However, \(\tau \) is an error tolerance used in the proposed adaptive computation of the step size t and determines the distance between the numerically computed iterates and the exact trajectory.
Remark 2
Before we can switch to the simplified Newton-type iteration scheme in step 12, we need to compute the proposed estimate \(\hat{\omega } \) for the Lipschitz constant in step 11. Evidently, this will roughly increase the complexity of Algorithm 1 by a constant factor of 2—before we switch to the simplified scheme, where we can reduce the computational effort by keeping \(\mathsf {M} (x_{0})\) fixed for the rest of the iteration procedure—.
Example 1
In this example we choose \(\mathsf {M} (x)=\mathsf {J} _{f}(x)\). Let us consider the function
Here, we identify f in its real form in \({\mathbb {R}}^{2}\), i.e., we separate the real and imaginary parts. The six zeros are given by
Note that \(\mathsf {J} _{f}\) is singular at (0, 0). Thus if we apply the classical Newton method with \(\mathsf {F} (x)=-\mathsf {J} _{f}(x)^{-1}f(x)\) in (3), the iterates close to (0, 0) cause large updates in the iteration procedure. More precisely, the application of \(\mathsf {F} (x)=-\mathsf {J} _{f}(x)^{-1}f(x)\) is a potential source for chaos near (0, 0). Before we discuss our numerical experiment, let us first consider the vector fields generated by the continuous problem (2). In Fig. 2, we depict the direction fields corresponding to \(\mathsf {F} (x)=f(x)\) (left) and \( \mathsf {F} (x)=-\mathsf {J} _{f}(x)^{-1}f(x)\) (right). We clearly see that some elements of \(Z_{f}\) are repulsive for \(\mathsf {F} (x)=f(x)\). Moreover, some elements of \(Z_{f}\) show a curl. If we now consider \( \mathsf {F} (x)=-\mathsf {J} _{f}(x)^{-1}f(x)\) the situation is completely different: All zeros are obviously attractive. In this example, we further observe that the vector direction field is divided into six different sectors, each containing exactly one element of \(Z_{f}\).
Next we visualize the domains of attraction of four different Newton-type iteration schemes. More precisely, we test the following four iteration procedures:
-
1.
The proposed procedure given in Algorithm 1, i.e., adaptive step size control—with \(\tau =0.01\) —and switching to the simplified Newton scheme which we abbreviate by AS.
-
2.
The proposed procedure given in Algorithm 1, i.e., adaptive step size control—with \(\tau =0.01\) —but without switching to the simplified Newton scheme which we abbreviate by ANS.
-
3.
The proposed procedure given in Algorithm 1, without step size control, i.e. \(\tau =\infty \), and without switching to the simplified Newton scheme which we abbreviate by NANS. This is simply the classical Newton iteration scheme.
-
4.
The proposed procedure given in Algorithm 1, without step size control—i.e. \(\tau =\infty \)—but switching to the simplified Newton scheme which we abbreviate by NAS.
In doing so, we compute the zeros of f by sampling initial values on a \(500\times 500 \) grid in the domain \([-3,3]^2\) (equally spaced). In Fig. 3, we show the fractal generated by the traditional Newton method NANS (left) as well as the corresponding plot for the combination of the classical Newton method and the simplified Newton method NAS (right). It is noteworthy that the chaotic behavior caused by the singularities of \(\mathsf {J} _{f}\) of the iteration procedure NAS is comparable to NANS.
In Fig. 4 we depict the basins of attraction for the adaptive procedure as proposed in Algorithm 1 AS (left) and the iteration procedure ANS. The chaotic behavior caused by the singularities of \(\mathsf {J} _{f}\) is clearly tamed—by both adaptive schemes AS and ANS.
Let us finally consider some performance data given in Table 1. An initial value \(x_{0}\in [-3,3]^{2}\) is called convergent if it is in fact convergent and additionally approaches the ‘correct’ zero, i.e., the zero that is located in the same exact attractor as the initial guess \(x_{0}\). Table 1 nicely demonstrates that—in contrast to the non adaptive schemes NANS and NAS—the number of convergent iterations for the adaptive procedures AS and ANS is close to \(100\%\). The second line in Table 1 shows the computational time—by sampling the computational time for all tested initial guesses \(x_{0}\in [-3,3]^{2}\)—with respect to the classical Newton iteration scheme NANS, i.e., we depict the quantity
In view of this quantity, the proposed iteration scheme AS is the clear winner compared to ANS as can be seen from line 2 in Table 1.
4 Conclusions
In this work, we have proved a convergence result for general simplified Newton-type iteration schemes under quite reasonable assumptions. In particular, we have shown that whenever the correction \(\left\| \mathsf {M} (x_{n})^{-1}f(x_{n})\right\| \) is small, then there locally exists a unique zero for the underlying map f. Since the proof of the proposed result relies on Banach’s fixed-point theorem, the theoretical result is constructive in the sense that it can be used for the numerical computation of the locally unique fixed point and therefore of the zero to be considered. Moreover, we have combined the convergence result with an adaptive root finding procedure thereby firstly taming the chaotic behavior of classical Newton-type iteration schemes and secondly reducing the computational effort due to the constant map \(x\mapsto \mathsf {M} (x_{n})\in {\mathbb {R}}^{n\times n}\)—without reducing the domain of convergence. We have tested our method on a low dimensional problem. Moreover, our experiment demonstrates empirically that the proposed scheme is indeed capable to tame the chaotic behavior of the iteration compared with the classical Newton scheme, i.e., without applying any step size control. In particular, our test example illustrate that the domains of convergence can—typically—be considerably enlarged in the sense that almost all initial guesses \(x_{0}\) are convergent to the ‘correct’ zeros, i.e., the zero which is located in the same attractor as the initial guess \(x_0\).
References
Amrein, M.: Adaptive Newton-Type Schemes Based on Projections, Tech. Report. arXiv:1809.04337v2 (2018)
Amrein, M., Melenk, J.M., Wihler, T.P.: An hp-adaptive Newton–Galerkin finite element procedure for semilinear boundary value problems. Math. Methods Appl. Sci. 40(6), 1973–1985 (2016). 13 pages, mma.4113
Amrein, M., Wihler, T.P.: An adaptive Newton-method based on a dynamical systems approach. Commun. Nonlinear Sci. Numer. Simul. 19(9), 2958–2973 (2014)
Amrein, M., Wihler, T.P.: Fully adaptive Newton–Galerkin methods for semilinear elliptic partial differential equations. SIAM, J. Sci. Comput. 37(4), A1637–A1657 (2015). 21 pages
Chaillou, A., Suri, M.: A posteriori estimation of the linearization error for strongly monotone nonlinear operators. J. Comput. Appl. Math. 205(1), 72–87 (2007)
Chaillou, A.L., Suri, M.: Computable error estimators for the approximation of nonlinear problems by linearized models. Comput. Methods Appl. Mech. Eng. 196(1–3), 210–224 (2006)
Congreve, S., Wihler, T.P.: Iterative Galerkin discretizations for strongly monotone problems. J. Comput. Appl. Math. 311, 457–472 (2017)
Deuflhard, P.: Newton Methods for Nonlinear Problems, Springer Series in Computational Mathematics. Springer, Berlin (2004)
Dörfler, W.: A robust adaptive strategy for the nonlinear Poisson equation. Computing 55(4), 289–304 (1995)
El Alaoui, L., Ern, A., Vohralík, M.: Guaranteed and robust a posteriori error estimates and balancing discretization and linearization errors for monotone nonlinear problems. Comput. Methods Appl. Mech. Eng. 200(37–40), 2782–2795 (2011)
Han, W.: A posteriori error analysis for linearization of nonlinear elliptic problems and their discretizations. Math. Methods Appl. Sci. 17(7), 487–508 (1994)
Heid, P., Wihler, T.P.: Adaptive Iterative Linearization Galerkin Methods for Nonlinear Problems, Tech. Report. arXiv:1808.04990 (2018)
Heid, P., Wihler, T.P: On the Convergence of Adaptive Iterative Linearized Galerkin Methods, Tech. Report. arXiv:1905.06682 (2019)
Neuberger, J.W.: Continuous Newton’s method for polynomials. Math. Intel. 21, 18–23 (1999)
Neuberger, J.W.: Integrated form of continuous Newton’s method. Lect. Notes Pure Applied. Math. 234, 331–336 (2003)
Neuberger, J.W.: The continuous Newton’s method, inverse functions and Nash Moser. Am. Math. Mon. 114, 432–437 (2007)
Ortner, C.: A posteriori existence in numerical computations. SIAM J. Numer. Anal. 47(4), 2550–2577 (2009)
Potschka, A.: Backward step control for global Newton-type methods. SIAM J. Numer. Anal. 54(1), 361–387 (2016)
Schneebeli, H.R., Wihler, T.P.: The Newton–Raphson method and adaptive ODE solvers. Fractals 19(1), 87–99 (2011)
Acknowledgements
Open access funding provided by ZHAW Zurich University of Applied Sciences. The author is grateful to Pascal Heid for comments on an earlier draft of this manuscript.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Amrein, M. A global Newton-type scheme based on a simplified Newton-type approach. J. Appl. Math. Comput. 65, 321–334 (2021). https://doi.org/10.1007/s12190-020-01393-w
Received:
Revised:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12190-020-01393-w