Abstract
In this work, we propose a free derivative quasi-Newton method for solving large-scale nonlinear systems of equation. We introduce a two-stage linear search direction and develop its global convergence theory. Besides, we prove that the method enjoys superlinear convergence rate. Finally, numerical experiments illustrate that the proposed method is competitive with respect to Newton-Krylov methods and other well-known methods for solving large-scale nonlinear systems of equations.
Similar content being viewed by others
1 Introduction
Let \(\,F:{\mathbb {R}}^{n}\rightarrow {\mathbb {R}}^{n}\,\) be a nonlinear continuously differentiable function, we are interested in solving the problem
especially when \(\,n\,\) is considerably large.
The most popular method for solving (1) is the Newton method [1, 2]. This method has very good convergence properties, but its greatest difficulty is having to calculate the Jacobian matrix of \(\,F, F',\,\) and evaluate it in each iteration which is, computationally, very expensive. One strategy to avoid this calculus and the evaluations is to use an approximation to \(\,F'(\varvec{x})\,\) which entails to a variety of methods known as quasi-Newton methods [3,4,5,6,7,8].
However, both Newton and quasi-Newton methods need to solve a linear system of equations in their routines; thus, if \(\,n\,\) is large, even quasi-Newton methods are expensive. It is important to say that although many quasi-Newton methods admit a cheap formula for updating the inverse of \(\,B_k,\,\) generally these formulas depend on a term that involves a fraction whose denominator could be zero or a very small quantity so in practice this could lead to an ill-conditioned matrix. To counter this, iterative Krylov methods [9,10,11] were introduced to Newtonian algorithms, which contributed to decrease the computational cost on these algorithms. This strategy consists of finding approximately the search direction, so the mentioned linear system will solve with some tolerable error.
For Newton-Krylov methods, fast convergence properties have been proved whereas, for quasi-Newton-Krylov methods, have been proved fast convergence properties if the Krylov method is the conjugate gradient method [12,13,14,15] or employed a Jacobian restart strategy [16]. The greatest difficulty of inexact quasi-Newton methods is that so far it has not been possible to prove that inexact quasi-Newton direction are descent directions for the associated merit function to (1).
Several studies are concerned with solving (1). In [17, 18], the authors proposed a different approach to solve (1). They proposed a nonmonotone spectral free derivative method. This was an ingenious proposal based on spectral gradient method and systematically uses \(\,\pm F(\varvec{x}_k)\,\) as a search direction. Due to the simplicity of this method, the computational cost is very low and although global convergence has been proved, linear or superlinear convergence has not yet been proved.
In this paper, we establish a global inexact quasi-Newton method, which is of low computational cost to solve (1). We propose a two-stage linear search procedure for obtaining descent direction and we ensure the global convergence without falling into infinite cycles. Also, under reasonable conditions, we prove fast convergence properties for the method.
This paper is organized as follows. In Section 2 we introduce the new algorithm and make some remarks. In Section 3 we develop the convergence theory of the algorithm introduced previously. Besides, we prove global convergence of the method and the linear and superlinear convergence rate. In Section 4 we present some numerical experiments that show the robustness and competitiveness of the new algorithm. Furthermore, we compare the performance of our algorithm with respect to the algorithms proposed in [17, 18]. Finally, we make some remarks in Section 5.
2 Algorithm
In this section, we describe the inexact free derivative quasi-Newton method (IFDQ) that we propose in this work.
Taking into account that one of our interests is to purpose a global algorithm to solve (1), we considered the following minimization problem to this.
where \(\,f(\varvec{x})=\frac{1}{2}\Vert F(\varvec{x})\Vert ^2\,\) is the associated merit function to (1) and \(\,\Vert \cdot \Vert \,\) is the Euclidean norm in \(\,{\mathbb {R}}^{n}.\,\)
Below we show the IFDQ algorithm which main innovation is the two-stage linear search procedure
Remark 1
So far it has not been possible to prove, in general, that the inexact quasi-Newton direction, obtained in Step 1 is a descent direction for the merit function \(\,f\,\) given in (2). For this reason, it is necessary to try both \(\,{\varvec{d}}_k\,\) and \(\,-{\varvec{d}}_k\,\) directions in the line search procedure.
Remark 2
In Step 2 we set two trial directions, \(\varvec{d}_k\) and \(-\varvec{d}_k\). Note that unless \(\nabla f^{T}\varvec{d}_k=0,\) at least one of the trial directions will be a descent direction, hence, the two-stage linear search procedure will not fall in an infinite cycle.
Remark 3
In the linear search procedure we seek a sufficient decrease of the merit function \(\,f\,\) but if the step length \(\,\alpha _k\,\) is too small then the algorithm breaks down, avoiding small steps with a poor decrease. Note that the lower bound for the step length is set by the user and can be as small as desired. So to prevent either, many breaks of the algorithm and small steps with a poor decrease we recommend to take \(\lambda =10^{-4}.\) Whit this value, the algorithm showed a good performance.
3 Convergence theory
In this section, we present the main theoretical results obtained for IFDQ algorithm. In the following lemmas and theorems we show that, under reasonable assumptions, the algorithm converges to a solution of the problem (1) and locally enjoys good convergence properties: full inexact quasi-Newton step is accepted and it can even has until superlinear converge.
The hypotheses under which we develop the convergence theory of the proposed algorithm are:
-
H1.
There exist \(\,\textbf{x}^*\in {\mathbb {R}}^{n}\,\) such that \(\,F(\textbf{x}^*)=0.\,\)
-
H2.
There exists \(\,T>0\,\) such that \(\,\Vert B_k^{-1}\Vert <T\,\) for all \(\,k\ge 0.\,\)
-
H3.
\(\,F'(\varvec{x})\,\) is a Lipschitz function, i.e., there exist \(\,L>0\,\) such that
$$\begin{aligned} \Vert F^{\prime }(\varvec{x})-F^{\prime }({\varvec{y}})\Vert \le L\Vert \varvec{x}-{\varvec{y}}\Vert \quad \forall \varvec{x},{\varvec{y}}\in {\mathbb {R}}^{n}. \end{aligned}$$ -
H4.
\(\{B_k\}\) is a sequence such that
$$\begin{aligned} \underset{k\rightarrow \infty }{\textrm{lim}}\frac{\Vert (B_{k}-F^{\prime }(\varvec{x}_k)){\varvec{d}}_k\Vert }{\Vert {\varvec{d}}_k\Vert }=0. \end{aligned}$$(4)
Previous hypotheses are classical hypotheses for quasi-Newton methods. Hypotheses H1 and H3 depend on the problem to be solved whereas H2 and H4 depends on the approximations to \(F^{\prime }(\varvec{x}_{k}).\)
An immediate consequence of H3 is that for all \(\,\varvec{x},\,{\varvec{y}}\in {\mathbb {R}}^{n},\,\)
It is important to mention that (4) in H4 is known as Dennis-Moré condition.
The first lemma of our theoretical development ensures that if the IFDQ algorithm does not break down then, it generates a sequence such that its images converge to zero.
Lemma 1
If \(\{\varvec{x}_k\}\) is a sequence generated by IFDQ algorithm then \(\underset{k \rightarrow \infty }{\textrm{lim}}\left\Vert {F}(\varvec{x}_k)\right\Vert =0\).
Proof
Let \(\{\varvec{x}_k\}\) be a sequence given by IFDQ algorithm, it follows that
By recalling that \(1-\lambda ^2<1,\) the result is established. \(\square \)
The next result is immediate since \(\,F\,\) is a continuous function.
Corollary 1
If \(\{\varvec{x}_k\}\) is a sequence generated by IFDQ algorithm and \(\varvec{x}^{*}\) is a cluster point of \(\{\varvec{x}_k\}\) then \(\varvec{x}^{*}\) is a solution of (1).
The following theorem ensures the convergence of the sequence generated by IFDQ algorithm under the assumption of non-singularity of the Jacobian matrix \(\,F'\,\) in the solution of the problem.
Theorem 1
Assume H2. Let \(\{\varvec{x}_k\}\) be a sequence generated by IFDQ algorithm. If \(\varvec{x}^{*}\in {\mathbb {R}}^{n}\) is a cluster point of \(\{\varvec{x}_k\}\) such that \(F'(\varvec{x}^{*})\) is nonsingular then \(F(\varvec{x}^{*})=0\) and \(\underset{k \rightarrow \infty }{\textrm{lim}}\varvec{x}_k=\varvec{x}^{*}\).
Proof
Let \(\{\varvec{x}_{k_{j}}\}\) be a subsequence of \(\{\varvec{x}_k\}\) such that \(\varvec{x}_{k_{j}}\rightarrow \varvec{x}^{*}\) as \(k_{j}\rightarrow \infty \). By recalling that F is a continuous function and Lemma 1, we have that \(F(\varvec{x}^{*})=0\).
On the other hand, let \(K=\left\Vert {F}'(\varvec{x}^{*})^{-1}\right\Vert \) and \(\delta >0\) small enough such that for any \({\varvec{y}}\in B(\varvec{x}^{*},\delta )\) we have that
-
i
\(\,F'({\varvec{y}})^{-1}\,\) there exists.
-
ii
\(\,\left\Vert {F}'({\varvec{y}})^{-1}\right\Vert <2K.\,\)
-
iii
\(\left\Vert {F}({\varvec{y}})-F(\varvec{x}^*)-F'(\varvec{x}^{*})({\varvec{y}}-\varvec{x}^{*})\right\Vert \le \frac{1}{2K}\left\Vert {\varvec{y}}-\varvec{x}^{*}\right\Vert .\,\)
The existence of \(\,\delta \,\) can be guaranteed thanks to Lemmas 1.1 and 1.2 in [15]. Observe that if \({\varvec{y}}\in B(\varvec{x}^{*},\delta )\) then
and
By combining the two last inequalities we can infer that
On the other hand, let \(\,\epsilon \in (0,\delta /4).\,\) Since \(\varvec{x}^{*}\) is a cluster point of \(\{\varvec{x}_{k}\}\) and \(F(\varvec{x}^{*})=0\) then, there exist \(\,k\,\) large enough such that
where \(\,M=\max \{K,T\}\,\) and \(\,T\,\) is the constant of hypothesis H2. Hence, we obtain
Now, since \(\varvec{x}_k\in S_{\epsilon }\) then \(\left\Vert \varvec{d}_{k}\right\Vert <\epsilon \). Moreover, we have that
Thus, we conclude \(\varvec{x}_{k+1}\in B(\varvec{x}^{*},\delta )\). On the other hand, since IFDQ attempts for a monotone decrease and \(\,\varvec{x}_k\in S_{\varepsilon }\,\) then
So, from (6) and using the last inequality we can infer that
Finally, from (8) and (9) we have that \(\varvec{x}_{k+1}\in S_{\epsilon },\) with which we prove that \(\,\varvec{x}_k\in S_{\epsilon }\,\) for all \(\,k\,\) large enough, and since \(\,\Vert F(\varvec{x}_k)\Vert \rightarrow 0\,\) then, from (6), \(\,\varvec{x}_k\rightarrow \varvec{x}^*\,\) as \(\,k\rightarrow \infty .\,\) \(\square \)
In the next lemma, we show that the trial directions in Step 1 remain bounded for all \(\,k.\,\)
Lemma 2
Assume H2. Let \(\{\varvec{x}_k\}\) be a sequence generated by IFDQ algorithm then, \(\left\Vert \varvec{d}_{k}\right\Vert \le 2T \left\Vert {F}(\varvec{x}_{k})\right\Vert \) and \(\underset{k\rightarrow \infty }{\textrm{lim}}\left\Vert \varvec{x}_{k+1}-\varvec{x}_{k}\right\Vert =0\).
Proof
The first part of this lemma follows from (7) and the fact that \(\,\theta _k\in (0,\,1).\,\)
On the other side, observe that
Thus, by Lemma 1 and the last inequality, we have the desired result. \(\square \)
The next theorem ensures convergence of the sequence generated by IFDQ algorithm without non-singularity condition of \(\,F'(\varvec{x}^*).\,\)
Theorem 2
Let \(\{\varvec{x}_{k}\}\) be a sequence generated by the algorithm. If \(\varvec{x}^{*}\) is an isolated cluster point of the sequence then, \(\underset{k\rightarrow \infty }{\textrm{lim}}\varvec{x}_{k}=\varvec{x}^{*}\).
Proof
Taking into account Lemmas 1 and 2, this proof follows the same ideas of the proof of theorem 3 in [18] \(\square \)
The next theorem ensures that, at least locally, the IFDQ algorithm shows good performance in the sense that the full quasi-Newton step will be accepted in the two-stage linear search procedure. In the proof we assume a weaker hypothesis than the Dennis-Moré condition. This hypothesis is related with the bounded deterioration property and ensures that at least, for all k large enough, the approximation \(B_k\) to the jacobian matrix \(F'(\varvec{x}_k)\) is bounded above by a constant.
As in the proof of Theorem 1, let \(\,M=\max \{K,T\}\,\) where \(\,T\,\) is the constant in H2.
Theorem 3
Assume hypotheses H1, H2 and H3. Let \(\,\{\varvec{x}_k\}\,\) be a sequence generated by IFDQ algorithm and \(\,\varvec{x}^*\,\) be a cluster point of the sequence. If \(\,F'(\varvec{x}^*)\,\) is nonsingular and for all \(\,k\,\) large enough \(\,\theta _k<\frac{1}{3}-\lambda \,\) and \(\Vert F'(\varvec{x}_k)-B_k\Vert <\frac{1}{24M^3}\) then \(\,\varvec{d}_k\,\) and \(\,\alpha _k=1\,\) will be accepted, in the two-stage linear search procedure.
Proof
Observe that
Thus,
so, from Lemma 2 and taking into account that for all \(\,k\,\) large enough
we have that
Now, since \(\,\left\Vert {F}(\varvec{x}_k)\right\Vert \,\) converges to zero then, for all \(\,k\,\) large enough
Hence, by (10), (11) and since \(\,\theta _k<\frac{1}{3}-\lambda \,\) then, for all k large enough
thus \(\,\alpha _k=1\,\) and \(\,{\varvec{d}}_k\,\) will be accepted. \(\square \)
The following theorem is the first theorem in which we show good convergence properties of the IFDQ algorithm. For this purpose, we assume that \(\,F'(\varvec{x}^*)\,\) is nonsingular, that \(\,\Vert F'(\varvec{x}^*)^{-1}\Vert =K\,\) and that \(\,M=\max \{K,T\}\,\) where \(\,T\,\) is the constant in H2.
Theorem 4
Under the same hypotheses of the previous theorem. If in addition
then \(\,\varvec{x}_k\rightarrow \varvec{x}^*\,\) linearly.
Proof
By using Theorem 3, we can ensure that the full quasi-Newton step will be accepted for all \(\,k\,\) large enough and by Theorem 1, \(\,\varvec{x}_k\rightarrow \varvec{x}^*\,\), so for all \(\,k\,\) large enough, \(\,F'(\varvec{x}_k)^{-1}\,\) there exist and \(\,\Vert F'(\varvec{x}_k)^{-1}\Vert \le 2M\,\) hence,
by (5) and Step 1 in the algorithm we have that
Thus, by Lemma 2,
So, by the mean value theorem,
thereby, for \(\,k\,\) large enough such that
since \(\,\theta _k<\frac{1}{12M^2},\,\) we can conclude that
where \(\,0<R<1,\,\) which completes the proof. \(\square \)
Finally, to complete our theoretical development, the next theorem ensures, under reasonable assumptions, superlinear convergence of the IFDQ algorithm.
Theorem 5
Assume H1, H2, H3 and H4. Let \(\,\{\varvec{x}_k\}\,\) be a sequence generated by IFDQ algorithm and \(\,\varvec{x}^*\,\) be a cluster point of the sequence. If \(\,F'(\varvec{x}^*)\,\) is nonsingular and \(\,\theta _k\rightarrow 0\,\) then \(\,\varvec{x}_k\rightarrow \varvec{x}^*\,\) superlinearly.
Proof
From (12) we can infer that
By Lemma 2 and the Mean Value Theorem,
The desired result follows from hypothesis H4 and the fact that \(\,\varvec{x}_k\rightarrow \varvec{x}^*\,\) and \(\,\theta _k\rightarrow 0\,\) as \(\,k\rightarrow \infty .\,\) \(\square \)
4 Numerical experiments
In this section we report the numerical results of the IFDQ algorithm when solving twenty problems. Sixteen of the problems were taken from [17] and references therein, the rest of the problems were taken from [19,20,21]. It is important to say that we did not take into account problems 13, 14, 15 and 18 of [17]. First, problems 13 and 14 include many random parameters that difficult the reproduction of the experiments. Second, the poor performance of the algorithms with problems 15 and 18 did not allow us to draw relevant conclusions.
The experiments were carried out in Matlab\(^{\circledR }\) using an Intel \(\text {Core}2^{TM}\) laptop with a RAM of 4GB. To evaluate the performance of the IFDQ algorithm, we ran experiments and compared the results with four more algorithms: SANE [17], DF-SANE [18], Ac-DFSANE [22] and NITSOL [23].
SANE and DF-SANE are spectral free derivative algorithms. The descent trial direction at each iteration of these methods is \(\,\pm F(\varvec{x}_k)\,\) and the main difference between them is the linear search.
Ac-DFSANE is an accelerated version, proposed recently, for the DF-SANE algorithm. This chooses in a very ingenious way the new iterate improving in many times the descent achieved in the linear search.
Finally, NITSOL is a practical and efficient implementation of the classical inexact Newton method with GMRES procedure to find the descent direction. This algorithm approximates derivatives by finite differences when these are not available.
For all algorithms we used \(\,\Vert F(\varvec{x}_k)\Vert <10^{-6}\,\) and \(\,k<300\,\) as stop criteria. For IFDQ algorithm we took \(\,B_0=I_n\,\) as initial approximation to \(\,F'(\varvec{x}_0); \theta _k=\frac{1}{k+2}\,\) as inexact parameter in Step 1 and \(\,\lambda =10^{-4}\,\) and \(\,\beta =0.5\,\) for the two-stage linear search procedure in Step 2. To find \(\,{\varvec{d}}_k\,\) in Step 1, we used GMRES procedure.
In the same way, in Step 3 we used the “good” Broyden update [24]. It is important to say that this update is a least change secant update, thus satisfying the well-known property of bounded deterioration [2, 25] with which Dennis and Schnabel in [2] showed that the sequence of matrices \(\,\{B_k\}\,\) satisfies the Dennis-Moré condition (4), i.e., the sequence of matrices \(\,\{B_k\}\,\) satisfies hypothesis H4.
On the other hand, SANE, DF-SANE, Ac-DFSANE and NITSOL algorithms were carried out with the same parameters as in respective references given above.
In Table 1, we show the complete list of problems with which we ran our experiments. Starting points were the same as in the respective references.
In Tables 2 and 3 we report the results obtained using the following conventions:
- F : :
-
functions of Table 1.
- Method::
-
algorithm used to solve the problems.
- n : :
-
size of the problem to solve.
- k : :
-
number of iterations required for the algorithms to solve each problem.
- Feval : :
-
number of evaluation of function at each problem.
- t : :
-
cpu time, given in seconds.
- \(**:\):
-
means nonconverge of the algorithm because it infringes the stop criteria.
The results in Tables 2 and 3 showed good performance of the IFDQ algorithm. First because IFDQ algorithm required the same or fewer iterations than its counterparts in 17.5% of the problems. Second, our algorithm converged on thirty-eight of the problems, that is, IFDQ algorithm converged on \(95\%\) of the experiments carried out.
It is important to mention that although in general when SANE, DF-SANE, Ac-DFSANE and NITSOL converged, they were faster, in terms of CPU time, than IFDQ, but in most cases, that difference was only for a few seconds. This behavior is due to the fact that SANE, DF-SANE and Ac-DFSANE used \(\,\pm F(\varvec{x}_k)\,\) as trial directions, so they do not have to solve a linear system of equation like (3) or make matrix vector products as products made for IFDQ to update \(\,B_k\,\). On the other side, NITSOL makes a single linear search, so generally requires less evaluations of function than IFDQ. In the same way, NITSOL approximate derivatives by finite differences when these are not available nevertheless we think that this could be affecting the convergence of the method since this was the one with the lowest success rate. For the above, IFDQ seems to be a competitive algorithm.
In the second and third columns of Table 4 we show the percentage of experiments in which each algorithm won in terms of CPU time and number of iterations. In last column of this table we show the percentage of success of each algorithm with the experiments.
The results show the IFDQ algorithm as an equilibrated method since it was the most successful without requiring a large number of iterations or CPU time compared to the other algorithms.
To test the global convergence of the IFDQ algorithm, we experimented with some of the problems by randomly varying the starting point. In all cases, we ran the algorithm with 500 starting points, whose components were uniformly distributed in the interval \(\,[-100,\,100].\,\) In Table 5 we show, for each experiment, the problem, the size of the problem and the success rate of the method.
The success rate of the algorithm on selected problems shows us a robust method; thus, it would be a good option for solving large-scale nonlinear system of equations.
To finish our experiments, we want to show the inner behavior of the IFDQ algorithm when solving the problem given by the Extended Rosenbrock function.
In Table 6, we show the behavior of the most important parameters in the algorithm when solving the above-mentioned problem. In this table,
it helps us to analyze the rate convergence of the algorithm.
As we can see in Table 6, \(\,RelRes\,\) converge to zero when \(\,\theta _k\,\) converge to zero, which suppose a superlinear convergence of the algorithm, as we proved in Theorem 5. On the other hand, \(\,\alpha _k=1\,\) for \(\,k>18\,\) just as we proved in Theorem 3.
5 Final remarks
In this work we proposed a new free derivative method for solving, especially, large-scale nonlinear systems of equations. This method takes inexact quasi-Newton directions to build the new iterate and, taking into account that so far it has not been possible to demonstrate that this is a descent direction and seeking to establish global convergence we proposed a two-stage linear search procedure.
For this new method, we show that it enjoys good convergence properties, this is, IFDQ method locally performs very well and, under reasonable hypotheses, has until superlinear convergence.
Numerical experiments showed that IFDQ had a performance according to expected and that this is a competitive method for solving large-scale nonlinear systems of equations.
Availability of data and material
The data obtained in this research are available.
Code Availability
This work introduce an algorithm completely available for the readers.
References
Nocedal, J., Wright, S.: Numerical Optimization. Springer Series in Operations Research, New York (2000)
Dennis, J.E., Schnabel. R.: Numerical Methods for Unconstrained Optimization and Nonlinear Equations. SIAM, Philadelphia (1996)
Pérez, R., Rocha, V.L.: Recent Applications and Numerical Implementation of Quasi-Newton Methods for Solving Nonlinear Systems of Equations. Numer Algorithms. 35, 261–285 (2004). https://doi.org/10.1023/B:NUMA.0000021762.83420.40
Martínez, J.M.: Practical quasi-Newton methods for solving nonlinear systems. J Comput Math. 124(1–2), 97–121 (2000). https://doi.org/10.1016/S0377-0427(00)00434-9
Gomes-Ruggiero, M.A., Martínez, J.M., Santos, S.A.: Solving nonsmooth equations by means of Quasi-Newton methods with globalization. Recent Advances in Noonsmooth Optimization, pp. 121 – 140. World Scientific Publisher, Singapore (1995). https://doi.org/10.1142/9789812812827_0008
Martínez, J.M.: A family of quasi-Newton methods for nonlinear equations with direct secant updates of matrix factorizations. SIAM J Numer Anal. 27(4), 1034–1049 (1990). https://doi.org/10.1137/0727061
Dennis, J.E., More, J.J.: A Characterization of Superlinear Convergence and its Application to Quasi-Newton Methods. Math Comput. 28(126), 549–560 (1974). https://doi.org/10.2307/2005926
Schubert, L.: Modification of quasi-Newton method for nonlinear equations with a sparse Jacobian. Math Comp. 24(109), 27–30 (1970). https://doi.org/10.2307/2004874
Dembo, R., Eisenstat, S., Trond, S.: Inexact Newton Methods. SIAM J Numer Anal. 19(2), 400–408 (1982). https://doi.org/10.1137/0719025
Deuflhard, P.: Global Inexact Newton Mehods for Very Large Scale Nonlinear Problems. Impact of Comptuning Science and engineering. 3(4), 366–393 (1991). https://doi.org/10.1016/0899-8248(91)90004-E
Martinez, J.M., Qi, L.: Inexact Newton Methods for solving nonsmooth equations. J Comput Appl Math. 60(1–2), 127–145 (1995). https://doi.org/10.1016/0377-0427(94)00088-I
Steighaug, T.: Quasi Newton Methods for Large Scale Nonlinear Problems. Yale University (1980)
Steihaug, T.: Damped Inexact Quasi Newton Methods. Technical Report TR81-03, Departament of Mathematical Science, Rice University (1984). https://hdl.handle.net/1911/101545
Eisenstat, S., Steihaug, T.: Local Analysis of Inexact Quasi Newton Methods. Technical Report TR82-07, Departament of Mathematical Science, Rice University (1982). https://hdl.handle.net/1911/101548
Eisenstat, S., Walker, H.: Globally Convergent Inexact Newton Methods. SIAM J Optim. 4(2), 393–422 (1994). https://doi.org/10.1137/0804022
Birgin, E., Krejić, N., Martínez, J.M.: Globally convergent inexact quasi-Newton methods for solving nonlinear systems. Numer Algorithms. 32, 249–260 (2003). https://doi.org/10.1023/A:1024013824524
La Cruz, W., Raydan, M.: Nonmonotone Spectral Methods for Large- Scale Nonlinear Systems. Optim Method Softw. 18(5), 583–599 (2003). https://doi.org/10.1080/10556780310001610493
La Cruz, W., Martínez, J.M., Raydan, M.: Spectral residual method without gradient information for solving large-scale nonlinear systems of equations. Math Comput. 75(255), 1429–1448 (2006). https://doi.org/10.1090/S0025-5718-06-01840-0
Lukšan, L., Vlček, J.: Sparse and Partially Separable Test Problems for Unconstrained and Equality Constrained Optimization. Technical Report 767, Institute of Computer Science, Academy of Sciences of the Czech Republic (1999). http://hdl.handle.net/11104/0123965
Tsoulos, I., Athanassios, S.: On locating all roots of systems of nonlinear equations inside bounded domain using global optimization methods. Nonlinear Anal-Real. 11(4), 2465–2471 (2010). https://doi.org/10.1016/j.nonrwa.2009.08.003
Zhou, W., Li, D.: Limited memory BFGS method for nonlinear monotone equations. J Comput Math. 25(1), 89 – 96 (2007). https://www.jstor.org/stable/43693347
Birgin, E., Martínez, J.M.: Secant acceleration of sequential residual methods for solving large-scale nonlinear systems of equations. SIAM J Numer Anal. 60(6), 3145–3180 (2022). https://doi.org/10.1137/20M1388024
Pernice, M., Walker, H.F.: NITSOL: a newton iterative solver for nonlinear systems. SIAM J Sci Comput. 19(1), 302–318 (1998). https://doi.org/10.1137/S1064827596303843
Broyden, C.G.: A Class of Methods for Solving Nonlinear Simultaneous Equations. Math Comput. 19(92), 577–593 (1965). https://doi.org/10.2307/2003941
Broyden, C.G., Dennis, J.E., Moré, J.J.: On the Local and Superlinear Convergence of Quasi-Newton Methods. IMA J Appl Math. 12(3), 223–245 (1973). https://doi.org/10.1093/imamat/12.3.223
Gasparo, M.: A nonmonotone hybrid method for nonlinear systems. Optim Method Softw. 13(2), 79–94 (2000). https://doi.org/10.1080/10556780008805776
Gomez-Ruggiero, M., Martínez, J.M., Moretti, A.C.: Comparing algorithms for solving sparse nonlinear systems of equations. SIAM J Sci and Stat Comput. 13(2), 459–483 (1992). https://doi.org/10.1137/0913025
Raydan, M.: The Barzilai and Borwein gradient method for the large scale unconstrained minimization problem. SIAM J opt. 7(1), 2–33 (1997). https://doi.org/10.1137/S1052623494266365
Acknowledgements
We are grateful with University of Cauca for the time allowed to this research through VRI ID 5352 project.
Funding
Open Access funding provided by Colombia Consortium. This work was supported by Universidad del Cauca given the time for the researcher through VRI ID 5352 project.
Author information
Authors and Affiliations
Contributions
All authors contributed equally to this work.
Corresponding author
Ethics declarations
Ethics approval
The authors declare that they followed all the rules of a good scientific practice.
Consent to participate
All authors approve their participation in this work.
Consent for publication
The authors approve the publication of this research.
Competing interests
The authors declare no competing interests.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the material. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://creativecommons.org/licenses/by/4.0/.
About this article
Cite this article
Arias, C.A., Gómez, C. Inexact free derivative quasi-Newton method for large-scale nonlinear system of equations. Numer Algor 94, 1103–1123 (2023). https://doi.org/10.1007/s11075-023-01529-6
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11075-023-01529-6