Abstract

For systems of nonlinear equations, a modified efficient Levenberg–Marquardt method with new LM parameters was developed by Amini et al. (2018). The convergence of the method was proved under the local error bound condition. In order to enhance this method, using nonmonotone technique, we propose a new Levenberg–Marquardt parameter in this paper. The convergence of the new Levenberg–Marquardt method is shown to be at least superlinear, and numerical experiments show that the new Levenberg–Marquardt algorithm can solve systems of nonlinear equations effectively.

1. Introduction

Consider the system of nonlinear equationswhere the function is continuously differentiable. In this paper, we assume that the solution set of (1) (denoted by ) is nonempty, with referring to the 2-norm.

Newton method is an important method to solve system (1) in [1]. At each iteration, it uses the trial stepwhere and is Jacobian matrix. When is Lipschitz continuous and nonsingular, then the convergence of this method is quadratic at the solution. However, trial step may not exist and is singular or near singular. Newton method may not be well defined. To overcome this difficulty, Levenberg–Marquardt (LM) method was created by Levenberg [2] and Marquardt [3] which uses the trial step at each iteration, whereand is a nonnegative constant. By introducing a nonnegative parameter , LM method overcomes the problem that is singular or near singular; furthermore, excessive step size is avoided. In this case, where and Jacobian matrix is nonsingular, the LM method is reduced to Newton method.

The efficiency of the LM method is affected by the parameter . For example, let , under the local error bound condition, the LM method is shown to have quadratic convergence by Yamashita and Fukushima in [4]. However, when the sequence is far away from the set , may be very big which may lead to large . It will result in a smaller LM step size, further reducing the efficiency of the algorithm. In [5], Fan used , where is updated with trust region technology in each iteration, the LM method also has quadratic convergence under some suitable conditions, and can alleviate the effect of the initial point being far away from the set .

To avoid this trouble, Amini used in [6]; when the sequence is far from the solution set and is very large, is close to , which effectively controls the range of . Umar proposed some new LM parameters , in [7]. Wang used with updated by trust region techniques from iteration to iteration in [8].

Ma introduced into LM method and used a new LM parameter in [9], where . It is noticeable that is a convex combination of and , and the quadratic convergence of this method is proved. There are numerous other various LM methods to solve (1); interested readers are referred to [1012] for related work. In order to discuss the range of parameter , inspired by [6, 8, 9], in this paper, we choose a new LM parameter as follows:where is a convex combination of and and is updated with trust region technology in each iteration.

Now, we setas the merit function for (1). We define the actual reduction and the predicted reduction of at the th iteration as follows:where is computed by (3). The following ratio is

Grippo applied the nonmonotone line search technique to Newton’s method in [13]; some authors have extended the nonmonotone techniques to trust region algorithm and proposed a lot of effective nonmonotone trust region methods in [14, 15]. And Amini proposed nonmonotone line search technique for the LM method in [6]. Numerical experiments show that the algorithm with the nonmonotone technique is more efficient than the algorithm without the nonmonotone technique. Inspired by these theories, we apply a nonmonotone strategy to LM method in this paper. Let us replace actual reduction (6) with the following actual reduction:where, and is a positive integer constant. Obviously, by this change, will be compared with the in each iteration, further leading to affect the ratio. The ratio after the change is

It can be used to decide whether the trial step is accepted and update the trust region parameter .

The paper is organized as follows. In Section 2, we present a new algorithm and then prove the global convergence of the new algorithm under some conditions. In Section 3, under the local error bound condition, the convergence of the new Levenberg–Marquardt method is shown to be at least superlinear. In Section 4, the new algorithm is an effective algorithm, which is demonstrated by numerical results. At last, we give some conclusions in Section 5.

2. The Efficient Algorithm and Global Convergence

In this section, firstly, we present the new efficient LM algorithm and then prove the global convergence of the new algorithm.

When sequence is close to the solution, the steps may be too large, so we requirein the new algorithm, where is a positive constant, and this is implemented by Step 5.

Lemma 1. For all , we have

Proof. This proof is directly derived from the important theory given by Powell in [16].
From literature [6], the following lemma can be obtained.

Lemma 2. Suppose the sequence be generated by Algorithm 1, then the sequence converges.

Input: given
Output:
Step 1. If , stop. Otherwise, set
where .
Step 2. Compute the search direction
Step 3. By (7), (8), and (10), compute .
Step 4. Set
Step 5. Choose as
Step 6. Set , and go to Step 1.

Next we present some of the assumptions needed in the following content.

Assumption 1. (a) is continuously differentiable and Lipschitz continuous; i.e., there exists a positive constant that makes(b) is Lipschitz continuous; i.e., there exists a positive constant such that

Lemma 3. If Assumption 1 holds, then we have

Proof. The proof of (17) can be found in [17]. So, we only prove (16). Using mean value theorem, there exists that makesand hence,According to the last equation, we can obtainSo (16) is true.

Theorem 1. Suppose that Assumption 1 is true. Then, Algorithm 1 terminates in finite iterations or satisfies

Proof. Assume that the theorem is incorrect; then, there exist a positive constant and a constant that makesFirstly, we prove thatSince is accepted by the algorithm, we haveThis, along with (17), (22), and Lemma 1, for all , that meansReplacing with , for all sufficiently large , there isFrom Lemma 2which together with the last inequality yields is a positive constant, soUsing Assumption 1, the last equality implies thatLet . Using induction, for all , we can show thatFor , we can have from (31) that (29) is true. Assuming that (29) is true for given , we show that (29) holds for given . Let be large enough such that . Substituting with and using (24), we obtainSimilarly, we can deduce thatTherefore (31) holds. Along with Assumption 1 we imply thatSimilarly, for any given , we have . On the one hand, for any , we haveUsing (31) and the fact that , we haveWith Assumption 1, we concludeAnd then (22) is proved. By using (22) and (25), we can deduce thatThen, it follows from (3) in Algorithm 1, (17), and (22) thatOn the other hand, by (6), (8), (17), (22), (38), and Lemma 1, we can deduce thatAnd then combined with (8) and (10), we can see thatConsidering Algorithm 1, obviously, for all large , there exists a positive constant such that , which conflicts with (22) and so Theorem 1 is true.

3. Local Convergence

Definition 1. Let be a subset of that makes . We say provides a local error bounded on for (1), where , if there exists a positive constant such that

Assumption 2. (a) is continuously differentiable, and provides a local error bound on subset for problem (1), where(b)andare both Lipschitz continuous on; that is, there exist two positive constants, that makewhich implies

Lemma 4. Suppose that Assumption 2 is true. Then, for all sufficiently large , we have the following.(1)There exists a positive constant that makes(2), where .

Proof. The proof process of (1) is the same as that of Lemma 3.2 in [6], so we are not going to prove it here and only give the proof of (2).where , obviously,We can obtain from Definition 1 and (45) and (46) thatThus we obtainThen we show the following inequalities:If , then the following holds:and if , thenThrough the above two inequalities, we haveSimilarly, if , then the following holdsand if , we haveHence, we obtainFrom Algorithm 1, (52), and (58), we havewhere . So when is sufficiently small,

Lemma 5. Suppose that Assumption 2 is true, for all sufficiently large . computed by (3) satisfies

Proof. If we set , then we have from (3) that is a minimizer of , so it follows from Algorithm 1, (12), (45), (46), and thatSo there is

Lemma 6. Suppose that Assumption 2 is true, for all sufficiently large . So, we have

Proof. Firstly, we deal with these two equations:On the other hand,We concludewhere .
Without loss of generality, for all , suppose that , and we prove the local convergence of Algorithm 1 by singular value decomposition (SVD) of .where with , , , and . And assume the SVD of are as follows:where and are two orthogonal matrixes. , , and , .
Since is Lipschitz continuous, by the theory of matrix perturbation [18], we haveSo there isSince converges to the set , then we have hold for all sufficiently large . So combined with (71), there is

Lemma 7. Suppose Assumption 2 holds; for all sufficiently large , we have(1)(2)

Proof. The proof process is similar to Lemma 7 in [9], so we omit it here.

Theorem 2. Under Assumption 2, let be a sequence generated by Algorithm 1 with trust region technique. If , then converges superlinearly to the solution. If , then sequence converges quadratically to the solution.

Proof. Using (3) and (69), we obtainFrom (47), (71), Definition 1, and Lemma 5, we haveFrom (44), (46), (74), Definition 1, and Lemma 5, we conclude thatOn the other hand, it is obvious thatIt follows from (75) and Lemma 5 that holds for all sufficiently large . So, , and this is related to (75). We deduce that if , . If , . So if , converges superlinearly to the solution; otherwise, if , converges quadratically to the solution.

4. Numerical Experiments

In Section 5, we compare the performance of Algorithm 1 with Algorithm 2.1 (writing Algorithm 2.1 as AELM) in [6] through some numerical experiments. The test function is improved by the method in [19]. The form is as follows:where has full column rank, and . It is certain thatwhere test problems are nonsingular test functions from [20]. We take , rank of as , and chooserank of as .

We set the following parameters in the algorithms: , or 1, and or 2. The algorithms are terminated when or the number of iterates exceeds . When and , Algorithm 1 is reduced to the AELM.

By numerical experiments, we find the numerical results of Algorithm 1 are the same as the numerical results of AELM in some functions. So we only list the results of other experiments in the following tables. Further, we adopt the efficiency index defined as EI in [21] to compare the performance of algorithm AELM and Algorithm 1. The results of the four experiments with are shown in Tables 1 and 2, and the results of the four experiments with are shown in Tables 3 and 4, respectively. We use six starting points , , and for each test problem, where is suggested in [20].(i)NF stands for the quantity of function calculations(ii)NJ stands for the quantity of Jacobian calculations(iii)’ indicates that the iteration number is more than (iv), where is the convergence order of algorithm.

It is shown in Table 1 that when and , the effect of Algorithm 1 is obviously better than that of AELM. Algorithm 1 wins of the numerical results while AELM wins , and of the two algorithms have the same results. The advantage of Algorithm 1 is not obvious when and . Algorithm 1 can win of the numerical results while AELM win , and of two algorithms have the same result.

Table 3 shows that when and , Algorithm 1 and AELM have the best experimental results. Algorithm 1 win of the numerical results, and of the two algorithms has the same result. The advantage of Algorithm 1 is not obvious when and . Algorithm 1 wins of the numerical results while AELM wins , and of the two algorithms has the same results.

Further, we adopt the EI and let in the experiment. Tables 2 and 4 show that when , the experimental data EI of AELM and Algorithm 1 are similar, but when , the EI of Algorithm 1 is obviously larger than that of AELM. In addition, in terms of the experimental time, except when ranking and , the execution time of Algorithm 1 is longer than that of AELM. In other cases, the execution time of Algorithm 1 is close to or less than that of AELM.

In general, it is shown that for most test problems, Algorithm 1 performs better than AELM. So it can indicate that Algorithm 1 is more efficient than AELM to solve systems of nonlinear equations.

5. Conclusion

In this paper, we propose a new LM algorithm by modifying the LM parameter for systems of nonlinear equations. Through numerical experiments, we find the calculation amounts of Algorithm 1 smaller than AELM in the case where and take some suitable value, which shows the effectiveness of the new Algorithm 1. Under some conditions, the global convergence of the new LM method is proved, and the local convergence of the new LM method is shown to be at least superlinear. Numerical results show that the new algorithm is efficient.

Data Availability

No data were used to support the findings of this study.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

Acknowledgments

This research was supported by the Natural Science Foundation of the Anhui Higher Education Institutions under Grant nos. KJ2020ZD008 and KJ2019A0604, by the Natural Science Foundation of Anhui Province under no. 2108085MF204, and by the Abroad Visiting of Excellent Young Talents in Universities of Anhui Province under Grant no. GXGWFX2019022.