Next Article in Journal
Solitons of η-Ricci–Bourguignon Type on Submanifolds in (LCS)m Manifolds
Next Article in Special Issue
An Efficient Solution of Multiplicative Differential Equations through Laguerre Polynomials
Previous Article in Journal
On the Solvatochromism of Fluorescein Sodium
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A New Adaptive Levenberg–Marquardt Method for Nonlinear Equations and Its Convergence Rate under the Hölderian Local Error Bound Condition

Department of Mathematics, Faculty of Mathematics and Statistics, Huaibei Normal University, Huaibei 235000, China
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(6), 674; https://doi.org/10.3390/sym16060674
Submission received: 29 April 2024 / Revised: 25 May 2024 / Accepted: 27 May 2024 / Published: 30 May 2024
(This article belongs to the Special Issue Computational Mathematics and Its Applications in Numerical Analysis)

Abstract

:
The Levenberg–Marquardt (LM) method is one of the most significant methods for solving nonlinear equations as well as symmetric and asymmetric linear equations. To improve the method, this paper proposes a new adaptive LM algorithm by modifying the LM parameter, combining the trust region technique and the non-monotone technique. It is interesting that the new algorithm is constantly optimized by adaptively choosing the LM parameter. To evaluate the effectiveness of the new algorithm, we conduct tests using various examples. To extend the convergence results, we prove the convergence of the new algorithm under the Hölderian local error bound condition rather than the commonly used local error bound condition. Theoretical analysis and numerical results show that the new algorithm is stable and effective.

1. Introduction

Nonlinear equations are widely used in key fields such as electricity, optics, mechanics, economic management, engineering technology, biomedicine, and alternative energy [1,2,3,4,5,6]. This paper discusses the following nonlinear equations:
f 1 ( x 1 , x 2 , , x n ) = 0 , f 2 ( x 1 , x 2 , , x n ) = 0 , f n ( x 1 , x 2 , , x n ) = 0 ,
which can be written in a vector form:
F ( x ) = 0 ,
where F ( x ) : R n R n is continuously differentiable and  x = x 1 , x 2 , , x n T . We denote the solution set of Equation (1) by X and assume that X is nonempty.
Several promising numerical methods [7,8,9,10,11] have been proposed for solving nonlinear equations. One of the classical methods to solve Equation (1) is the Gauss–Newton method, which at each iteration computes the trial step
d k = J k T J k 1 J k T F k ,
where F k = F x k , J k = F x k is the Jacobian matrix of F ( x ) at x k .
However, in the actual calculation, the trial step of the Gauss–Newton method may not be well defined when J ( x ) is singular or near-singular. To overcome this difficulty, the Levenberg–Marquardt(LM) method [10,11] was proposed. At the kth iteration, the LM method computes the trial step
d k = J k T J k + λ k I 1 J k T F k ,
where λ k 0 is the LM parameter and I R n × n is the identity matrix. The trial step of the LM method is actually a modification of the trial step of the Gauss–Newton method, where the parameter λ k is introduced to prevent the steps from being undefined or too large when J ( x ) is singular or nearly singular.
The LM method has quadratic convergence when J ( x ) is Lipschitz continuous and nonsingular at the solution of Equation (1) [12]. Nevertheless, the theoretical research shows that the condition of the nonsingularity of J ( x ) is too strong. To solve this problem, some scholars [13,14,15,16,17,18,19] have analysed the convergence of the LM method under the following local error bound condition, which is weaker than nonsingularity of J ( x ) :
c · dist x , X F ( x ) , x N x ,
where c > 0 is a positive constant, dist x , X is the distance from x to X , and N x is some neighbourhood of x X . In this paper, · is the 2-norm.
Although the local error bound condition is weaker than the nonsingularity of J ( x ) , this condition is not always satisfied with some ill-conditioned nonlinear equations in biochemical systems and certain applications. Recently, some scholars [20,21,22,23] have studied the convergence of LM method under the following Hölderian error bound condition, which is weaker than the local error bound condition:
c · dist x , X F ( x ) γ , x N x ,
where c is a positive constant and  γ ( 0 , 1 ] . Obviously, the Hölderian error bound condition (4) is a generalization of the local error bound condition (3), where the exponent γ of F ( x ) is extended to an interval ( 0 , 1 ] . In this paper, we study the convergence of the new algorithm under the Hölderian error bound condition.
The LM parameter λ k is vital to the efficiency of LM algorithms. Several scholars have done interesting research [13,14,15,16,17,18,19,21,22,23] on λ k . Yamashita and Fukushima [13] took λ k = F k 2 , although the disadvantage of choosing parameters in this way is that the value of λ k = F k 2 may be too small to be effective when the sequence { x k } is close to the solution set of Equation (1), which affects the local convergence rate. In order to solve this disadvantage and reduce the impact, Fan and Yuan [14] chose λ k = F k δ , which is a generalization of λ k = F k 2 , and extended the exponent δ of λ k = F k δ to an interval [ 1 , 2 ] . The numerical results when solving some equations showed better performance when δ = 1 ; however, the disadvantage of choosing parameters in this way is that it may make λ k = F k too large and step d k too small when { x k } is far away from the solution set, causing the sequence to move slowly to the solution set and affecting the global convergence rate. To compensate for this flaw, Fan [17] used λ k = μ k F k , where μ k is updated every iteration by the trust region technique. Numerical results showed that this change improved the performance of the algorithm. Chen and Ma [23] took λ k = θ F k δ + ( 1 θ ) J k T F k δ for θ [ 0 , 1 ] and δ [ 1 , 2 ] , finding that this improved the numerical results of the LM algorithm. Recently, Li et al. [24] proposed a new adaptive accelerated LM algorithm by choosing the LM parameter as λ k + 1 = μ k + 1 F k + 1 1 + F k + 1 , with numerical results showing that the algorithm is efficient for solving symmetric and asymmetric linear equations.
Inspired by the above literature, we take a new adaptive LM parameter to enhance the computing performance of the LM algorithm, as follows:
λ k = μ k ( θ F k δ 1 + F k δ + ( 1 θ ) F k δ ) , if F k 1 , μ k ( θ F k δ 1 + F k δ + ( 1 θ ) F k δ ) , otherwise , ( 0 θ 1 , 1 δ 2 ) ,
where μ k is updated every iteration via trust region technology. When { x k } is close to a solution set, F k is close to 0; thus, λ k is close to μ k F k δ if  δ = 1 , as used in [17]. Conversely, when { x k } is far from the solution set, the leading F k may be very large; thus, λ k will be close to μ k θ . This effectively regulates the range of λ k to prevent the LM step from becoming excessively small, thereby enhancing computational efficiency. Therefore, it seems that this choice of λ k is more effective for the LM algorithm.
The following sections outline the remaining contents of this paper. In Section 2, we propose a new algorithm with a new LM parameter in more detail and prove its global convergence. In Section 3, we analyse the convergence rate of the new algorithm. In Section 4, we present numerical results verifying that the new algorithm is effective. Finally, some key conclusions are put forward in Section 5.

2. The New Adaptive LM Algorithm and Its Global Convergence

In this section, we introduce our new adaptive algorithm and establish its global convergence.
If we define the merit function for Equation (1) as
ϕ ( x ) = F ( x ) 2 ,
then, at the kth iteration, the actual reduction of ϕ ( x ) is provided by
A r e d k = F k 2 F x k + d k 2
and the predicted reduction of ϕ ( x ) by
P r e d k = F k 2 F k + J k d k 2 ,
where d k is computed by Equation (2). The ratio of A r e d k to P r e d k is
r k = A r e d k P r e d k ,
which determines whether to accept d k and update μ k . Several studies have suggested that algorithms employing non-monotone strategies outperform those with monotone strategies [18,25,26,27,28]. To carry out the non-monotone strategy, Amini et al. [18] used the following actual reduction to replace Equation (5):
A ¯ r e d k = F l ( k ) 2 F x k + d k 2
where
F l ( k ) = max 0 j n ( k ) F k j , k = 0 , 1 , 2 , ,
n ( k ) = min N 0 , k , and N 0 is a positive integer constant. With this change, F x k + 1 is compared with max 0 j n ( k ) F k j at each iteration. To combine the non-monotone strategy with the new adaptive LM parameter, we use the following ratio:
r ^ k = A ¯ r e d k P r e d k
to replace the original role of the ratio r k in the algorithm.
Next, we present a new adaptive LM algorithm, named the ALLM algorithm (Algorithm 1).
Algorithm 1 (ALLM Algorithm)
Step 1. Given x 0 R n , N 0 > 0 , μ 0 > m > 0 , ε > 0 , 0 < p 0 p 1 p 2 < 1 . Set k : = 0 .
Step 2. If  J k T F k ε , stop. Otherwise let
λ k = μ k ( θ F k δ 1 + F k δ + ( 1 θ ) F k δ ) , if   F k 1 , μ k ( θ F k δ 1 + F k δ + ( 1 θ ) F k δ ) , otherwise , ( 0 θ 1 , 1 δ 2 ) .

Step 3. Compute d k
J k T J k + λ k I d = J k T F k ,

Step 4. Compute F l ( k ) , P r e d k and A ¯ r e d k by Equations (9), (6) and (8). Set
r ^ k = A ¯ r e d k P r e d k .

Step 5. Set
x k + 1 = x k + d k , if r ^ k p 0 x k , otherwise

Step 6. Choose μ k + 1 as
μ k + 1 = 4 μ k , if r ^ k < p 1 μ k , if r ^ k p 1 , p 2 max μ k 4 , m , otherwise .

Step 7. Set k = k + 1 and return to Step 2.
To prevent excessively large steps, we impose the following condition:
μ k m , k N
where m is a positive constant.
Lemma 1.
P r e d k J k T F k min { d k , J k T F k J k T J k } for all k N .
Proof. 
This proof comes from famous result in [29]. □
Lemma 2
([18]). Assume that sequence x k is generated by the ALLM algorithm; then, the sequence F l ( k ) converges.
Assumption 1.
(a) J ( x ) is Hölderian continuous, i.e., there exists a constant κ h j > 0 such that
J ( x ) J ( y ) κ h j x y v , x , y R n
where the exponent v ( 0 , 1 ] .
(b) J ( x ) is bounded, i.e., there exists a constant κ b j > 0 such that
J ( x ) κ b j , x R n .
It follows from Equation (16) that
F ( y ) F ( x ) J ( x ) ( y x ) κ h j 1 + v y x 1 + v .
Thus, there exists a constant κ b f > 0 that makes
F ( y ) F ( x ) κ b f y x .
Theorem 1.
Under Assumption 1, the ALLM algorithm satisfies
lim k inf J k T F k = 0 .
Proof. 
Assuming that Theorem 1 is not true, we obtain
J k T F k ϵ 0 , k k 0
where ϵ 0 is a positive constant and k 0 N .
If d k is accepted by the ALLM algorithm, then
F l ( k ) 2 F x k + d k 2 p 0 P r e d k .
Per Lemma 1, Equations (17) and (21) indicate that, for all k k 0 ,
F l ( k ) 2 F k + 1 2 p 0 J k T F k min { d k , J k T F k J k T J k } p 0 ϵ 0 min { d k , ϵ 0 κ b j 2 } .
Then, substituting k for l ( k ) 1 ,
F l ( l ( k ) 1 ) 2 F l ( k ) 2 p 0 ϵ 0 min d l ( k ) 1 , ϵ 0 κ b j 2
holds for all sufficiently large k.
Per Lemma 1, we obtain
lim k F l ( l ( k ) 1 ) 2 F l ( k ) 2 = 0 ;
thus,
lim k min d l ( k ) 1 , ϵ 0 κ b j 2 = 0 ,
as ϵ 0 κ b j 2 is a positive constant, meaning that
lim k d l ( k ) 1 = 0 .
Per Equation (19), the last equality implies that
lim k F ( x l ( k ) ) = lim k F ( x l ( k ) 1 ) .
Next, by considering the proof process of Theorem 2.4 in [18], we can prove that
lim k d k = 0 .
Along with Equations (10), (11), (17) and (21), this implies that
μ k , as k .
Next, per Equation (18), we obtain
F x k + d k F k + J k d k κ h j 1 + v d k 1 + v ,
which yields
| F ( x k + d k ) 2 F k + J k d k 2 | 2 κ h j 1 + v F k + J k d k d k 1 + v + κ h j 2 ( 1 + v ) 2 d k 2 + 2 v .
From Lemma 1 and Equations (17), (21), (22) and F k + J k d k F k F 1 , we obtain
r k 1 = A r e d k P r e d k P r e d k 2 κ h j 1 + v F k + J k d k d k 1 + v + κ h j 2 ( 1 + v ) 2 d k 2 + 2 v J k T F k min d k , J k T F k J k T J k 0 ;
thus,
lim k + r k = 1 .
Combined with Equations (6), (8), (9) and (12), we obtain
r ^ k = A ¯ r e d k P r e d k = F l ( k ) 2 F x k + d k 2 P r e d k F k 2 F k + 1 2 P r e d k = r k 1 .
In view of the ALLM algorithm, for all large k there exists a positive constant μ ¯ > m that makes μ k < μ ¯ , which contradicts Equation (23). Thus, Theorem 1 holds. □

3. Convergence Rate

This section discusses the convergence rate of the ALLM algorithm. Here, we let x k generated by ALLM algorithm lie within a neighborhood of x X and converge to the solution set X of Equation (1).
Assumption 2.
(a) F ( x ) provides a Hölderian local error bound, i.e., there exist constants c > 0 and 0 < b < 1 that make
c · dist x , X F ( x ) γ , x N x , b ,
where the exponent γ ( 0 , 1 ] , N x , b = x R n x x b .
(b) J ( x ) is Hölderian continuous, i.e., there exists a constant κ h j > 0 such that
J ( x ) J ( y ) κ h j x y v , x , y N x , b
where the exponent v ( 0 , 1 ] .
From Equation (25), we have
F ( y ) F ( x ) J ( x ) ( y x ) κ h j 1 + v y x 1 + v , x , y N x , b .
Thus,
F ( y ) F ( x ) κ b f y x , x , y N x , b ,
where κ b f is a positive constant.
Defining by x ¯ k X satisfies
x ¯ k x k = dist x k , X ,
which implies that x ¯ k is closest to x k .
Next, we discuss the important property of d k and μ k ; finally, we study the convergence rate of the ALLM algorithm using the singular value decomposition (SVD) technique. Without loss of generality, we assume that x k N ( x , b 4 ) .
Lemma 3.
Under Assumption 2, we have
(1) If F k 1 ; then, the following relationship holds:
d k c ¯ dist x k , X min { 1 , 1 + v δ 2 γ }
where c ¯ is a positive constant.
(2) If F k > 1 , then the following relationship holds:
d k c ˜ dist x k , X
where c ˜ is a positive constant.
Proof. 
(1) As x k N ( x , b 4 ) , we have
x ¯ k x x ¯ k x k + x k x b 2 ;
thus, x ¯ k N ( x , b 2 ) .
We define
φ k ( d ) = F k + J k d 2 + λ k d 2 .
It can be concluded from (11) that d k is the minimizer of φ k ( d ) . From (26) and F x ¯ k = 0 , we have
d k 2 φ k ( d k ) λ k φ k ( x ¯ k x k ) λ k = F k + J k ( x ¯ k x k ) 2 + λ k x ¯ k x k 2 λ k = F ( x ¯ k ) F k J k ( x ¯ k x k ) 2 + λ k x ¯ k x k 2 λ k 1 λ k κ h j 1 + v 2 x ¯ k x k 2 + 2 v + x ¯ k x k 2 .
If F k 1 , then F k δ 1 and 1 + F k δ 2 . In conjunction with (15) and (24), this yields
λ k = μ k θ F k δ 1 + F k δ + ( 1 θ ) F k δ m θ c δ γ x k x ¯ k δ γ 2 + ( 1 θ ) c δ γ x k x ¯ k δ γ 1 2 ( 3 m m θ ) c δ γ x k x ¯ k δ γ .
Thus,
d k 2 1 λ k κ h j 1 + v 2 x ¯ k x k 2 + 2 v + x ¯ k x k 2 2 κ h j 2 c δ γ ( 3 m m θ ) ( 1 + v ) 2 x ¯ k x k 2 + 2 v δ γ + x ¯ k x k 2 2 κ h j 2 c δ γ ( 3 m m θ ) ( 1 + v ) 2 + 1 x ¯ k x k 2 min { 1 , 1 + v δ 2 γ } .
Setting c ¯ = 2 κ h j 2 c δ γ / ( 3 m m θ ) ( 1 + v ) 2 , we obtain Equation ( 28 ) .
(2) If F k > 1 , then F k δ > 1 and 1 + F k δ 2 F k δ . Along with (19), this allows us to conclude that
λ k = μ k θ F k δ 1 + F k δ + ( 1 θ ) F k δ m θ F k δ 2 F k δ + ( 1 θ ) k b f δ x k x ¯ k δ m θ 2 + m ( 1 θ ) k b f δ x k x ¯ k δ .
Thus, there exists a constant c ˜ > 0 such that
d k 2 c ˜ 2 dist ( x k , X ) 2 .
Therefore, d k c ˜ dist ( x k , X ) . □
Lemma 4.
Under Assumption 2, we have the following:
(1) If F k 1 , v > max { 1 γ 1 , 1 γ ( 1 + v ) δ 2 1 , 1 γ γ ( 1 + v ) δ 2 } , then μ k is bounded above, i.e., there exists a positive constant M 1 such that μ k M 1 holds for all large k.
(2) If F k > 1 , v > 1 γ 1 , then μ k is bounded above, i.e., there exists a positive constant M 2 such that μ k M 2 holds for all large k.
Proof. 
(1) Considering Lemma 3.3 in [21], we can see that
r k 1 = A r e d k P r e d k P r e d k = F k + J k d k 2 F x k + d k 2 P r e d k κ h j 1 + v 2 d k 2 + 2 v + 2 κ h j 1 + v F k + J k d k d k 1 + v c 4 F k d k max { 1 γ , 1 γ ( 1 + v ) δ 2 , 1 γ γ ( 1 + v ) δ 2 + 1 } 0 ;
thus,
lim k + r k = 1 .
This, along with Equations (6), (8), (9), and (12), yields
r ^ k = A ¯ r e d k P r e d k = F l ( k ) 2 F x k + d k 2 P r e d k F k 2 F k + 1 2 P r e d k = r k 1 .
Considering the updating rule from (14), we can ascertain the existence of a positive constant M 1 > m , ensuring that μ k M 1 holds for sufficiently large k.
(2) Consider the following two cases.
Case 1: x ¯ k x k d k . Per Lemma 3 (2), Equations (24), (26) and v > 1 γ 1 , we have
F k F k + J k d k F k F k + J k x ¯ k x k c 1 γ x ¯ k x k 1 γ κ h j 1 + v x ¯ k x k 1 + v c 1 x ¯ k x k 1 γ c 2 d k 1 γ
which holds for some c 1 , c 2 > 0 .
Case 2: x ¯ k x k > d k . It follows from Equation (30) that
F k F k + J k d k F k F k + d k x ¯ k x k J k x ¯ k x k F k 1 d k x ¯ k x k F k + d k x ¯ k x k F k + J k x ¯ k x k d k x ¯ k x k F k F k + J k x ¯ k x k c 1 d k x ¯ k x k 1 γ 1 c 3 d k 1 γ
holds for some c 3 > 0 .
Therefore, from Equations (30) and (31), we have
P r e d k = F k + F k + J k d k F k F k + J k d k F k F k F k + J k d k c 4 F k d k 1 γ ,
which holds for some c 4 > 0 .
Because F k + J k d k F k , v > 1 γ 1 , from Equations (26) and (32) we have
r k 1 = A r e d k P r e d k P r e d k = F k + J k d k 2 F x k + d k 2 Pred k κ h j 1 + v 2 d k 2 + 2 v + 2 κ h j 1 + v F k + J k d k d k 1 + v c 5 F k d k 1 γ 0 ;
thus,
lim k + r k = 1 .
This, along with Equations (6), (8), (9) and (12), yields
r ^ k = A ¯ r e d k P r e d k = F l ( k ) 2 F x k + d k 2 P r e d k F k 2 F k + 1 2 P r e d k = r k 1 .
Therefore, there exists a positive constant M 2 > m such that μ k M 2 holds for sufficiently large k. □
Next, we consider SVD technology. In view of the findings provided by Behling and Iusem in [30], without loss of generality, we set rank ( J ( x ¯ ) ) = r for all x ¯ N x , b X . Suppose that the SVD of J x ¯ k is
J x ¯ k = U ¯ k Σ ¯ k V ¯ k T = U ¯ k , 1 , U ¯ k , 2 Σ ¯ k , 1 0 V ¯ k , 1 T V ¯ k , 2 T = U ¯ k , 1 Σ ¯ k , 1 V ¯ k , 1 T ,
where Σ ¯ k , 1 = diag σ ¯ k , 1 , , σ ¯ k , r > 0 .
Correspondingly,
J k = U k , 1 , U k , 2 Σ k , 1 Σ k , 2 V k , 1 T V k , 2 T = U k , 1 Σ k , 1 V k , 1 T + U k , 2 Σ k , 2 V k , 2 T ,
where Σ k , 2 = diag σ k , r + 1 , , σ k , n > 0 .
For clearness, we let
J k = U 1 Σ 1 V 1 T + U 2 Σ 2 V 2 T ,
which neglects the subscription k in U k , i , Σ k , i and V k , i .
Lemma 5
([21]). Under Assumption 2, the following relationship holds:
(1) U 1 U 1 T F k κ b f x ¯ k x k
(2) U 2 U 2 T F k 2 κ h j x ¯ k x k 1 + v .
Theorem 2.
Under the conditions of Lemma 3, we have the following:
(1) If F k 1 , then the x k generated by the ALLM algorithm converges to the solution set of Equation (1) with order min { γ ( 1 + δ ) , γ ( 1 + v ) , γ ( 1 + v ) ( 1 + v δ 2 γ ) } .
(2) If F k > 1 , then the x k generated by the ALLM algorithm converges to the solution set of Equation (1) with order γ.
Proof. 
(1) It follows from the SVD of J k that
d k = V 1 Σ 1 2 + λ k I 1 Σ 1 U 1 T F k V 2 Σ 2 2 + λ k I 1 Σ 2 U 2 T F k
and
F k + J k d k = F k U 1 Σ 1 Σ 1 2 + λ k I 1 Σ 1 U 1 T F k U 2 Σ 2 Σ 2 2 + λ k I 1 Σ 2 U 2 T F k = λ k U 1 Σ 1 2 + λ k I 1 U 1 T F k + λ k U 2 Σ 2 2 + λ k I 1 U 2 T F k .
According to the theory of matrix perturbation [31] and Equation (25), we have
diag Σ 1 Σ ¯ k , 1 , Σ 2 J k J x ¯ k κ h j x ¯ k x k v ,
which indicates
Σ 1 Σ ¯ k , 1 κ h j x ¯ k x k v , Σ 2 κ h j x ¯ k x k v .
As x k converges to X , without loss of generality, we let κ h j x ¯ k x k v σ ¯ 2 hold for all large k. From Equation (34), we have
Σ 1 2 + λ k I 1 Σ 1 2 1 σ ¯ κ h j x ¯ k x k v 2 4 σ ¯ 2 .
From Equations (34), (35), Lemma 5, and Σ 2 2 + λ k I 1 λ k 1 , we have
F k + J k d k 4 λ k κ b f σ ¯ 2 x ¯ k x k + 2 κ h j x ¯ k x k 1 + v .
If F k 1 , then F k δ 1 , while from Equation (27) and Lemma 4 we have
λ k = μ k θ F k δ 1 + F k δ + ( 1 θ ) F k δ M 1 θ F k δ + ( 1 θ ) F k δ M 1 k b f δ x k x ¯ k δ .
This, along with Equation (36), yields
F k + J k d k 4 M 1 κ b f 1 + δ σ ¯ 2 x ¯ k x k 1 + δ + 2 κ h j x ¯ k x k 1 + v 4 M 1 κ b f 1 + δ σ ¯ 2 + 2 κ h j x ¯ k x k min { 1 + δ , 1 + v } .
Letting c 5 = 4 M 1 κ b f 1 + δ σ ¯ 2 + 2 κ h j , from Equations (24), (26), (28) and (37) we obtain
c x ¯ k + 1 x k + 1 1 γ F x k + d k F k + J k d k + κ h j d k 1 + v c 5 x ¯ k x k min { 1 + δ , 1 + v } + κ h j c ¯ 1 + v x ¯ k x k min { 1 + v , ( 1 + v ) ( 1 + v δ 2 γ ) } c 5 + κ h j c ¯ 1 + v x ¯ k x k min { 1 + δ , 1 + v , ( 1 + v ) ( 1 + v δ 2 γ ) } .
Thus,
c x ¯ k + 1 x k + 1 c 5 + κ h j c ¯ 1 + v γ x ¯ k x k min { γ ( 1 + δ ) , γ ( 1 + v ) , γ ( 1 + v ) ( 1 + v δ 2 γ ) } ,
which indicates that x k converges to the solution set X of Equation (1) with convergence rate min { γ ( 1 + δ ) , γ ( 1 + v ) , γ ( 1 + v ) ( 1 + v δ 2 γ ) } .
(2) The proof of F k > 1 is similar to the proof of F k δ 1 . We obtain
c x ¯ k + 1 x k + 1 c 6 + κ h j c ¯ 1 + v γ x ¯ k x k γ ;
thus, x k converges to the solution set X of Equation (1) with order γ . □
Theorem 3.
Under Assumption 2, we have the following:
(1) If F k 1 , v > 1 γ 1 , 1 γ 1 < δ 2 γ v , then x k generated by the ALLM algorithm converges to some solution of Equation (1) with order min { γ ( 1 + δ ) , γ ( 1 + v ) } .
(2) If F k > 1 , v > 1 γ 1 , then x k generated by the ALLM algorithm converges to some solution of Equation (1) with order γ.
Proof. 
(1) If F k 1 and v δ 2 γ , then from Equation (28) we obtain
d k c ¯ x ¯ k x k .
It follows from v > 1 γ 1 and v δ 2 γ that
1 γ 1 1 γ γ ( 1 + v ) δ 2 0
and
1 γ γ ( 1 + v ) δ 2 1 γ ( 1 + v ) δ 2 1 0 .
Therefore, the conditions of Lemma 4 (1) hold. In conjunction with δ > 1 γ 1 and δ 2 γ v , this yields
min γ ( 1 + δ ) , γ ( 1 + v ) , γ ( 1 + v ) 1 + v δ 2 γ = min { γ ( 1 + δ ) , γ ( 1 + v ) } > 1 .
Thus, x k converges superlinearly to X .
For clearness,
x ¯ k x k x ¯ k + 1 x k + 1 + d k .
In view of Equations (38) and (40), we know the existence of a constant M > 0 , meaning that
x ¯ k x k M d k
holds for large k. Thus, from Equations (38), (39), (40), and (42), we have
d k + 1 O d k min { γ ( 1 + δ ) , γ ( 1 + v ) } ,
which means that the ALLM algorithm converges with order min { γ ( 1 + δ ) , γ ( 1 + v ) } .
(2) The proof of F k > 1 is similar to the proof of F k 1 . We obtain
d k + 1 O d k γ ;
thus, ALLM algorithm converges with order γ . □

4. Numerical Experiments

In this section, we verify the effectiveness of the ALLM algorithm by presenting some numerical experiments. Algorithm 1 (named the AELM algorithm) from [22] is used for comparison. All algorithms were tested in the MATLAB R2022b programming environment on a personal PC with an i7-7500U CPU and 2.7 GHz. We selected the parameters of the AELM algorithm as follows: p 0 = 10 4 , p 1 = 0.25 , p 2 = 0.75 , N 0 = 5 , μ 1 = 0.01 , m = 10 8 . We selected the parameters of the ALLM algorithm as follows: p 0 = 10 4 , p 1 = 0.25 , p 2 = 0.75 , N 0 = 5 , μ 1 = 0.01 , m = 10 8 , θ = 0 , 0.5 , 1 , δ = 1 , 2 . All algorithms were terminated when J k T F k 10 5 or when the number of iterations surpassed 1000.
Example 1.
We consider four special functions [22] to verify that the ALLM algorithm satisfies more theoretical applications. Functions 1–4 satisfy the Hölderian local error bound condition around the zero point but do not satisfy the local error bound condition. Here, the J ( x ) for Functions 3–4 are Hölderian continuous but not Lipschitz continuous, while the J ( x ) for Functions 1–2 are both Lipschitz continuous and Hölderian continuous.
Function 1
f 1 ( x ) = x 1 + 10 x 2 , f 2 ( x ) = 5 ( x 3 x 4 ) , f 3 ( x ) = x 2 2 x 3 2 , f 4 ( x ) = 10 ( x 1 x 4 ) 2 .
Initial point: x 0 = ( 3 , 1 , 0 , 1 ) T , zero point: ( 0 , 0 , 0 , 0 ) T .
Function 2
f 1 ( x ) = x 1 x 2 , f 2 ( x ) = x 1 2 + x 2 2 .
Initial point: x 0 = ( 1 , 1 ) T , zero point: ( 0 , 0 ) T .
Function 3
f 1 ( x ) = x 1 + 10 x 2 , f 2 ( x ) = x 3 x 4 , f 3 ( x ) = ( x 2 2 x 3 ) 3 2 , f 4 ( x ) = ( x 1 x 4 ) 3 2 .
Initial point: x 0 = ( 3 , 1 , 0 , 1 ) T , zero point: ( 0 , 0 , 0 , 0 ) T .
Function 4
f 1 ( x ) = x 1 + 10 x 2 , f 2 ( x ) = x 3 x 4 , f 3 ( x ) = ( x 2 2 x 3 ) 4 3 , f 4 ( x ) = ( x 1 x 4 ) 4 3 .
Initial point: x 0 = ( 3 , 1 , 0 , 1 ) T , zero point: ( 0 , 0 , 0 , 0 ) T .
We tested each function for three starting points, x 0 , 10 x 0 and 100 x 0 , to study the global convergence of the ALLM algorithm. Table 1 lists the numerical results achieved by the AELM and ALLM algorithms on the four test functions. The symbols in Table 1 have the following meanings:
  • NF: The number of function calculations.
  • NJ: The number of Jacobian calculations.
  • NT: We generally use the ‘ NT = NF + NJ n ’ to indicate the total computations.
Table 1. Numerical results of the AELM and ALLM algorithms with various choices of δ and θ .
Table 1. Numerical results of the AELM and ALLM algorithms with various choices of δ and θ .
AELM ALLM
δ = 1 δ = 2
Function n x 0 θ = 0 θ = 0.5 θ = 1 θ = 0 θ = 0.5 θ = 1
NF/NJ/NT NF/NJ/NT NF/NJ/NT NF/NJ/NT NF/NJ/NT NF/NJ/NT NF/NJ/NT
14110/10/5010/10/5010/10/5010/10/5010/10/5010/10/5010/10/50
1013/13/6513/13/6513/13/6513/13/6513/13/6513/13/6513/13/65
10016/16/8016/16/8016/16/8016/16/8016/16/8016/16/8016/16/80
2218/8/248/8/248/8/248/8/248/8/248/8/248/8/24
1011/11/3311/11/3311/11/3311/11/3311/11/3311/11/3311/11/33
10015/15/4515/15/4515/15/4515/15/4515/15/4515/15/4515/15/45
3418/8/408/8/408/8/408/8/408/8/408/8/408/8/40
1010/10/5010/10/5010/10/5010/10/5010/10/5010/10/5010/10/50
10012/12/6012/12/6012/12/6012/12/6012/12/6012/12/6012/12/60
44113/13/657/7/357/7/357/7/357/7/357/7/357/7/35
1016/16/809/9/459/9/459/9/459/9/459/9/459/9/45
10061/50/26111/11/5511/11/5511/11/5511/11/5511/11/5511/11/55
As can be seen from Table 1, the ALLM algorithm is obviously superior to the AELM algorithm for the numerical results of Function 4, while the two algorithms are the same for the numerical results of Functions 1–3.
Example 2.
We consider some singular problems which are created by the following form [32]:
F ^ ( x ) = F ( x ) J x A A T A 1 A T x x ,
where the test function F ( x ) is provided by Moré, Garbow, and Hillstrom in [33], x is the root of F ( x ) , and A R n × k has full column rank. It is clear that the Jacobian of F ^ ( x ) is
J ^ x = J x I A A T A 1 A T ,
with rank n k ( 1 k n ) and F ^ x = 0 . Similar to [33], we choose
A = [ 1 , 1 , , 1 ] T R n × 1 ,
which implies rank J ^ x = n 1 .
Next, we ran all test problems for three starting points: 10 x 0 , x 0 , x 0 , 10 x 0 , and 100 x 0 , where x 0 derives from [33]. Table 2 and Table 3 display the numerical results achieved by the algorithms for all test functions. The meanings of the symbols in Table 2 and Table 3 are as follows:
  • Iter: Number of iterations.
  • F: Final value of the norm of the function.
  • Time: CPU time in seconds.
From Table 2 and Table 3, it is evident that the ALLM algorithm generally outperforms the AELM algorithm in terms of CPU time across most test functions. Compared with the AELM algorithm, the performance of the ALLM algorithm exhibits superior performance when θ = 0 and δ = 2 , dominating approximately 90% of the CPU time results; about 4 % of the results of iterations of the two algorithms are the same. In particular, for certain test functions it can be seen that the ALLM algorithm consistently outperforms the AELM algorithm in terms of both iteration count and CPU time when the initial point is distant from the solution set. From Table 2, for the extended helical valley function, when n = 501 and the initial point is 10 x 0 , the number of iterations and the CPU time of the ALLM algorithm are better than those of the AELM algorithm. From Table 3, for the discrete boundary value function, when n = 1000 and the initial point is 10 x 0 or x 0 or x 0 or 10 x 0 or 100 x 0 , the number of iterations and CPU time of the ALLM algorithm are better than those of the AELM algorithm.
To compare the numerical performance profile of the AELM and ALLM algorithms, we chose the performance analysis method proposed by Dolan [34]. As can be seen from Figure 1, when θ = 0 and δ = 2 , the ALLM algorithm demonstrates the best performance in terms of iteration count, while when θ = 1 and δ = 1 , the performance in terms of the number of iterations for both algorithms. As can be seen from Figure 2, when θ = 0 and δ = 2 , the CPU time of the ALLM algorithm has the best performance, while when θ and δ take other values the ALLM algorithm maintains advantages in CPU time performance.
In general, the ALLM algorithm proves more effective in solving nonlinear equations compared to the AELM algorithm. In particular, when δ is larger and θ is smaller, the ALLM algorithm demonstrates superior performance. According to the needs of practical applications, the selection of λ k is continuously optimized by changing the values of δ and θ .
Table 2. Numerical results of the AELM and ALLM algorithms with δ = 1 and various choices of θ .
Table 2. Numerical results of the AELM and ALLM algorithms with δ = 1 and various choices of θ .
AELMALLM
δ = 1
Function n x 0 θ = 0 θ = 0.5 θ = 1
Iters/F/TimeIters/F/Time Iters/F/Time Iters/F/Time
Extended Rosenbrock500−1019/2.7631 × 10−7/0.3019/2.7858 × 10−7/0.2719/2.7744 × 10−7/0.2919/2.7631 × 10−7/0.28
−116/1.7341 × 10−7/0.2314/2.5509 × 10−7/0.1915/3.3852 × 10−7/0.2016/1.7341 × 10−7/0.25
117/2.2186 × 10−7/0.2417/1.9553 × 10−7/0.2517/2.0901 × 10−7/0.2217/2.2186 × 10−7/0.23
1019/3.9252 × 10−7/0.2919/3.8895 × 10−7/0.2519/3.9066 × 10−7/0.2719/3.9252 × 10−7/0.29
10023/1.3154 × 10−7/0.4223/1.3142 × 10−7/0.3023/1.3150 × 10−7/0.3423/1.3154 × 10−7/0.37
1000−1019/3.9156 × 10−7/1.6319/3.9442 × 10−7/1.5219/3.9299 × 10−7/1.5319/3.9156 × 10−7/1.64
−116/2.5034 × 10−7/1.4314/3.8204 × 10−7/1.1016/1.2047 × 10−7/1.2816/2.5034 × 10−7/1.34
117/3.1868 × 10−7/1.4617/2.8057 × 10−7/1.4017/2.9943 × 10−7/1.3617/3.1868 × 10−7/1.46
1020/1.3911 × 10−7/1.6720/1.3781 × 10−7/1.9320/1.3866 × 10−7/1.6520/1.3911 × 10−7/1.57
10023/1.8652 × 10−7/2.0623/1.8646 × 10−7/1.9023/1.8638 × 10−7/1.9023/1.8652 × 10−7/1.98
Extended Helical valley501−1042/1.3356 × 10−6/0.683/5.1316 × 10−7/0.0313/3.7044 × 10−7/0.1714/3.2573 × 10−7/0.23
−11/0.0000 × 100/0.011/0.0000 × 100/0.011/0.0000 × 100/0.011/0.0000 × 100/0.01
18/3.1758 × 10−7/0.138/1.4137 × 10−7/0.098/2.1648 × 10−7/0.108/3.1758 × 10−7/0.12
108/1.8981 × 10−9/0.128/8.0024 × 10−10/0.108/1.2568 × 10−9/0.108/1.8981 × 10−9/0.11
1008/5.9124 × 10−10/0.128/3.9747 × 10−10/0.118/4.8399 × 10−10/0.118/5.9124 × 10−10/0.12
1000−107/2.3766 × 10−13/0.536/3.0134 × 10−13/0.456/1.5143 × 10−9/0.447/2.3766 × 10−13/0.55
−11/0.0000 × 100/0.041/0.0000 × 100/0.041/0.0000 × 100/0.041/0.0000 × 100/0.04
18/1.8337 × 10−8/0.698/7.2043 × 10−9/0.688/1.1804 × 10−8/0.628/1.8337 × 10−8/0.67
108/1.6817 × 10−11/0.668/1.0758 × 10−11/0.618/1.4744 × 10−11/0.608/1.6817 × 10−11/0.65
10026/9.3824 × 10−13/2.5135/1.0739 × 10−7/3.2926/6.6675 × 10−8/2.1826/6.9835 × 10−11/2.16
Discrete boundary value500−106/3.3487 × 10−3/0.116/3.3666 × 10−3/0.076/3.3631 × 10−3/0.096/3.3487 × 10−3/0.12
−14/1.2234 × 10−3/0.064/1.2579 × 10−3/0.044/1.2417 × 10−3/0.054/1.2234 × 10−3/0.04
13/3.5633 × 10−4/0.043/3.6008 × 10−4/0.033/3.5823 × 10−4/0.033/3.5633 × 10−4/0.03
105/6.7290 × 10−3/0.085/6.7739 × 10−3/0.065/6.7614 × 10−3/0.065/6.7290 × 10−3/0.08
10012/1.3651 × 10−4/0.2313/1.3834 × 10−5/0.1612/1.5752 × 10−4/0.1712/1.3651 × 10−4/0.16
1000−106/3.6656 × 10−3/0.526/3.6804 × 10−3/0.506/3.6780 × 10−3/0.476/3.6656 × 10−3/0.48
−14/1.4253 × 10−3/0.304/1.4669 × 10−3/0.304/1.4474 × 10−3/0.304/1.4253 × 10−3/0.31
13/1.3022 × 10−4/0.223/1.3092 × 10−4/0.203/1.3058 × 10−4/0.213/1.3022 × 10−4/0.21
105/6.5900 × 10−3/0.405/6.6346 × 10−3/0.385/6.6209 × 10−3/0.355/6.5900 × 10−3/0.39
10013/9.9458 × 10−5/1.0913/1.0869 × 10−4/1.0613/1.0505 × 10−4/1.0713/9.9458 × 10−5/1.08
Discrete integral equation500−1012/1.2304 × 10−5/1.0612/1.2171 × 10−5/1.0312/1.2238 × 10−5/1.0412/1.2304 × 10−5/1.06
−19/1.5928 × 10−5/0.769/1.4153 × 10−5/0.769/1.5162 × 10−5/0.759/1.5928 × 10−5/0.76
17/1.3357 × 10−5/0.597/1.3770 × 10−5/0.587/1.3592 × 10−5/0.587/1.3357 × 10−5/0.58
1010/9.3502 × 10−6/0.868/9.0151 × 10−6/0.679/1.5419 × 10−5/0.7610/9.3502 × 10−6/0.86
10010/4.5155 × 10−9/0.9110/4.5463 × 10−9/0.8910/4.5306 × 10−9/0.8810/4.5155 × 10−9/0.91
1000−1012/1.7452 × 10−5/4.5012/1.7265 × 10−5/4.4812/1.7358 × 10−5/4.5012/1.7452 × 10−5/4.51
−110/6.0308 × 10−6/3.7110/5.2005 × 10−6/3.7310/5.6998 × 10−6/3.7010/6.0308 × 10−6/3.67
18/5.1495 × 10−6/2.868/5.3838 × 10−6/2.908/5.2754 × 10−6/2.508/5.1495 × 10−6/2.86
1010/1.4251 × 10−5/3.739/5.0675 × 10−6/3.3010/5.7297 × 10−6/3.5910/1.4251e × 10−5/3.66
10010/6.3828 × 10−9/3.8610/6.4261 × 10−9/3.8310/6.4040 × 10−9/3.8110/6.3828 × 10−9/3.83
Broyden banded500−1010/3.8446 × 10−12/0.1710/3.9166 × 10−12/0.1610/4.3882 × 10−12/0.1410/3.8446 × 10−12/0.17
−126/6.9212 × 10−6/0.5231/1.2468 × 10−5/0.5328/1.6756 × 10−5/0.5025/1.2128 × 10−5/0.50
112/1.5063 × 10−5/0.2012/1.5060 × 10−5/0.2012/1.5061 × 10−5/0.1912/1.5063 × 10−5/0.20
1018/1.7636 × 10−5/0.3318/1.7636 × 10−5/0.3018/1.7636 × 10−5/0.2818/1.7636 × 10−5/0.28
10024/1.0280 × 10−5/0.4424/1.0280 × 10−5/0.3624/1.0280 × 10−5/0.3724/1.0280 × 10−5/0.37
1000−1010/3.5499 × 10−12/0.9010/3.8124 × 10−12/0.9110/4.6220 × 10−12/0.9010/3.5499 × 10−12/0.99
−133/9.9927 × 10−6/3.3527/9.6110 × 10−6/2.5433/2.6949 × 10−5/3.0428/9.7912 × 10−6/2.62
112/2.1201 × 10−5/1.2212/2.1196 × 10−5/1.0812/2.1199 × 10−5/1.0912/2.1201 × 10−5/1.17
1018/2.4886 × 10−5/1.8218/2.4886 × 10−5/1.5918/2.4886 × 10−5/1.6818/2.4886 × 10−5/1.68
10024/1.4499 × 10−5/2.8024/1.4499 × 10−5/2.4224/1.4499 × 10−5/2.3724/1.4499 × 10−5/2.54
Table 3. Numerical results of the AELM and ALLM algorithms with δ = 2 and various choices of θ .
Table 3. Numerical results of the AELM and ALLM algorithms with δ = 2 and various choices of θ .
AELMALLM
δ = 2
Function n x 0 θ = 0 θ = 0.5 θ = 1
Iters/F/Time Iters/F/Time Iters/F/Time Iters/F/Time
Extended Rosenbrock500−1019/2.7631 × 10−7/0.3019/2.7841 × 10−7/0.2619/2.7734 × 10−7/0.2619/2.7635 × 10−7/0.30
−116/1.7341 × 10−7/0.2314/1.7729 × 10−7/0.1715/3.2343 × 10−7/0.2116/1.7288 × 10−7/0.20
117/2.2186 × 10−7/0.2417/1.8677 × 10−7/0.2517/2.0435 × 10−7/0.2517/2.2121 × 10−7/0.22
1019/3.9252 × 10−7/0.2919/3.8875 × 10−7/0.2419/3.9061 × 10−7/0.2619/3.9243 × 10−7/0.25
10023/1.3154 × 10−7/0.4223/1.3134 × 10−7/0.2923/1.3145 × 10−7/0.3323/1.3149 × 10−7/0.33
1000−1019/3.9156 × 10−7/1.6319/3.9423 × 10−7/1.5019/3.9284 × 10−7/1.5419/3.9140 × 10−7/1.56
−116/2.5034 × 10−7/1.4314/2.3878 × 10−7/1.0515/4.5237 × 10−7/1.2416/2.4969 × 10−7/1.32
117/3.1868 × 10−7/1.4617/2.7048 × 10−7/1.3317/2.9369 × 10−7/1.3617/3.1809 × 10−7/1.33
1020/1.3911 × 10−7/1.6720/1.3786 × 10−7/1.5320/1.3857 × 10−7/1.5720/1.3919e × 10−7/1.59
10023/1.8652 × 10−7/2.0623/1.8649 × 10−7/1.8323/1.8655 × 10−7/1.8023/1.8633 × 10−7/1.81
Extended Helical valley501−1042/1.3356 × 10−6/0.683/7.5853 × 10−13/0.0413/1.7643 × 10−7/0.1814/2.0341 × 10−7/0.19
−11/0.0000 × 100/0.011/0.0000 × 100/0.011/0.0000 × 100/0.011/0.0000 × 100/0.01
18/3.1758 × 10−7/0.138/1.0595 × 10−7/0.118/1.9230 × 10−7/0.118/3.2101 × 10−7/0.09
108/1.8981 × 10−9/0.128/4.7841 × 10−10/0.118/9.5278 × 10−10/0.108/1.6935 × 10−9/0.10
1008/5.9124 × 10−10/0.128/2.0055 × 10−10/0.108/3.2729 × 10−10/0.138/4.9161 × 10−10/0.13
1000−107/2.3766 × 10−13/0.535/7.4481 × 10−10/0.366/1.7998 × 10−12/0.516/4.6416 × 10−7/0.42
−11/0.0000 × 100/0.041/0.0000 × 100/0.041/0.0000 × 100/0.041/0.0000 × 100/0.04
18/1.8337 × 10−8/0.698/4.1226 × 10−9/0.598/9.1462 × 10−9/0.628/1.7753 × 10−8/0.63
108/1.6817 × 10−11/0.668/2.2703 × 10−11/0.598/3.2386 × 10−11/0.608/4.0803 × 10−11/0.64
10026/9.3824 × 10−13/2.5146/2.6770 × 10−11/3.7026/8.5493 × 10−9/2.0926/2.8466 × 10−10/2.15
Discrete boundary value500−106/3.3487 × 10−3/0.114/3.1555 × 10−3/0.044/4.2613 × 10−3/0.044/4.7919 × 10−3/0.05
−14/1.2234 × 10−3/0.063/7.2588 × 10−4/0.033/7.1034 × 10−4/0.033/6.9394 × 10−4/0.04
13/3.5633 × 10−4/0.043/4.1479 × 10−6/0.033/4.1402 × 10−6/0.043/4.1324 × 10−6/0.04
105/6.7290 × 10−3/0.084/2.4328 × 10−3/0.054/ 3.0758 × 10−3/0.054/3.4218 × 10−3/0.05
10012/1.3651 × 10−4/0.2311/4.2188 × 10−5/0.1612/1.5591 × 10−5/0.1812/1.3814 × 10−5/0.18
1000−106/3.6656 × 10−3/0.524/3.5180 × 10−3/0.294/4.5349 × 10−3/0.294/5.0429 × 10−3/0.28
−14/1.4253 × 10−3/0.303/9.3230 × 10−4/0.213/9.1090 × 10−4/0.213/8.8813 × 10−4/0.23
13/1.3022 × 10−4/0.222/2.6311 × 10−4/0.132/2.6303 × 10−4/0.132/2.6296 × 10−4/0.13
105/6.5900 × 10−3/0.404/2.5604 × 10−3/0.294/3.0932 × 10−3/0.304/3.4004 × 10−3/0.27
10013/9.9458 × 10−5/1.0911/1.2743 × 10−4/0.8811/2.1884 × 10−4/0.9311/2.1536 × 10−4/0.85
Discrete integral equation500−1012/1.2304 × 10−5/1.0612/1.2047 × 10−5/1.0312/1.2171 × 10−5/1.0512/1.2294 × 10−5/1.04
−19/1.5928 × 10−5/0.769/1.0655 × 10−5/0.759/1.2735 × 10−5/0.769/ 1.4195 × 10−5/0.76
17/1.3357 × 10−5/0.597/1.0869 × 10−5/0.577/1.0772 × 10−5/0.587/1.0633 × 10−5/0.57
1010/9.3502 × 10−6/0.869/4.8669 × 10−6/0.759/1.3758 × 10−5/0.7610/9.7578 × 10−6/0.84
10010/4.5155 × 10−9/0.9110/4.5453 × 10−9/0.8910/4.5300 × 10−9/0.9010/4.5154 × 10−9/0.88
1000−1012/1.7452 × 10−5/4.5012/1.7133 × 10−5/4.4112/1.7285 × 10−5/4.4612/1.7441 × 10−5/4.47
−110/6.0308 × 10−6/3.719/1.5092 × 10−5/3.259/1.8533 × 10−5/3.2710/5.2150 × 10−6/3.66
18/5.1495 × 10−6/2.867/1.5749 × 10−5/2.487/1.5610 × 10−5/2.507/1.5367 × 10−5/2.62
1010/1.4251 × 10−5/3.739/8.5398 × 10−6/3.2510/5.1845 × 10−6/3.6210/1.4626 × 10−5/3.71
10010/6.3828 × 10−9/3.8610/6.4246 × 10−9/3.7610/6.4031 × 10−9/3.7910/6.3825 × 10−9/3.77
Broyden banded500−1010/3.8446 × 10−12/0.1710/3.795 × 10−12/0.1610/4.3814 × 10−12/0.1710/3.8105 × 10−12/0.16
−126/6.9212 × 10−6/0.5229/6.5177 × 10−6/0.4528/1.6459 × 10−5/0.4425/1.1907 × 10−5/0.42
112/1.5063 × 10−5/0.2012/1.5059 × 10−5/0.2012/1.5061 × 10−5/0.2012/1.5063 × 10−5/0.18
1018/1.7636 × 10−5/0.3318/1.7636 × 10−5/0.3018/1.7636 × 10−5/0.3318/1.7636 × 10−5/0.31
10024/1.0280 × 10−5/0.4424/1.0280 × 10−5/0.3624/1.0280 × 10−5/0.4024/1.0280 × 10−5/0.36
1000−1010/3.5499 × 10−12/0.9010/4.5936 × 10−12/0.8610/3.810 × 10−12/0.8910/5.7143 × 10−12/0.93
−133/9.9927 × 10−6/3.3529/1.5374 × 10−5/2.7631/1.8408 × 10−5 /2.8428/9.7968 × 10−6/2.61
112/2.1201 × 10−5/1.2212/2.1194 × 10−5/1.0712/2.1198 × 10−5/1.0812/2.1201 × 10−5/1.13
1018/2.4886 × 10−5/1.8218/2.4886 × 10−5/1.6918/2.4886 × 10−5/1.6418/2.4886 × 10−5/1.65
10024/1.4499 × 10−5/2.8024/1.4499 × 10−5/2.3024/1.4499 × 10−5/2.3324/1.4499 × 10−5/2.51
Figure 1. Performance profile of AELM and ALLM based on number of iterations for example 1–10.
Figure 1. Performance profile of AELM and ALLM based on number of iterations for example 1–10.
Symmetry 16 00674 g001
Figure 2. Performance profile of AELM and ALLM based on CPU time for example 1–10.
Figure 2. Performance profile of AELM and ALLM based on CPU time for example 1–10.
Symmetry 16 00674 g002

5. Conclusions

In this paper, inspired by the Hölderian local error bound condition, we studied the convergence properties of our ALLM algorithm under different conditions. We used the new modified adaptive LM parameter and incorporated the non-monotone technique to modify the Levenberg–Marquardt algorithm. The numerical results show that our new algorithm is efficient and stable.

Author Contributions

Conceptualization, Y.H. and S.R.; methodology, Y.H. and S.R.; Software, Y.H.; validation, Y.H. and S.R.; formal analysis, Y.H. and S.R.; investigation, Y.H. and S.R.; resources, Y.H. and S.R.; data curation, Y.H. and S.R.; Writing—original draft, Y.H.; writing—review and editing, Y.H. and S.R.; visualization, Y.H.; supervision, S.R.; project administration, S.R.; funding acquisition, S.R. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Natural Science Foundation of the Anhui Higher Education Institutions of China, 2023AH050348.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank S.R. and everyone for their valuable comments and suggestions which helped us improve the quality of this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Leonov, E.; Polbin, A. Numerical Search for a Global Solution in a Two-Mode Economy Model with an Exhaustible Resource of Hydrocarbons. Math. Model. Comput. Simul. 2022, 14, 213–223. [Google Scholar] [CrossRef]
  2. Xu, D.; Bai, Z.; Jin, X.; Yang, X.; Chen, S.; Zhou, M. A mean-variance portfolio optimization approach for high-renewable energy hub. Appl. Energy 2022, 325, 119888. [Google Scholar] [CrossRef]
  3. Manzoor, Z.; Iqbal, M.S.; Hussain, S.; Ashraf, F.; Inc, M.; Tarar, M.A.; Momani, S. A study of propagation of the ultra-short femtosecond pulses in an optical fiber by using the extended generalized Riccati equation mapping method. Opt. Quantum Electron. 2023, 55, 717. [Google Scholar] [CrossRef]
  4. Vu, D.T.S.; Gharbia, I.B.; Haddou, M.; Tran, Q.H. A new approach for solving nonlinear algebraic systems with complementarity conditions. Application to compositional multiphase equilibrium problems. Math. Comput. Simul. 2021, 190, 1243–1274. [Google Scholar] [CrossRef]
  5. Maia, L.; Nornberg, G.; Pacella, F. A dynamical system approach to a class of radial weighted fully nonlinear equations. Commun. Partial Differ. Equ. 2021, 46, 573–610. [Google Scholar] [CrossRef]
  6. Vasin, V.V.; Skorik, G.G. Two-stage method for solving systems of nonlinear equations and its applications to the inverse atmospheric sounding problem. In Doklady Mathematics; Springer: New York, NY, USA, 2020; Volume 102, pp. 367–370. [Google Scholar]
  7. Luo, X.l.; Xiao, H.; Lv, J.h. Continuation Newton methods with the residual trust-region time-stepping scheme for nonlinear equations. Numer. Algorithms 2022, 89, 223–247. [Google Scholar] [CrossRef]
  8. Waziri, M.Y.; Ahmed, K. Two descent Dai-Yuan conjugate gradient methods for systems of monotone nonlinear equations. J. Sci. Comput. 2022, 90, 1–53. [Google Scholar] [CrossRef]
  9. Pes, F.; Rodriguez, G. A doubly relaxed minimal-norm Gauss–Newton method for underdetermined nonlinear least-squares problems. Appl. Numer. Math. 2022, 171, 233–248. [Google Scholar] [CrossRef]
  10. Levenberg, K. A method for the solution of certain non-linear problems in least squares. Q. Appl. Math. 1944, 2, 164–168. [Google Scholar] [CrossRef]
  11. Marquardt, D.W. An algorithm for least-squares estimation of nonlinear parameters. J. Soc. Ind. Appl. Math. 1963, 11, 431–441. [Google Scholar] [CrossRef]
  12. Nocedal, J.; Wright, S.J. Numerical Optimization; Springer: New York, NY, USA, 1999. [Google Scholar]
  13. Yamashita, N.; Fukushima, M. On the rate of convergence of the Levenberg-Marquardt method. In Topics in Numerical Analysis: With Special Emphasis on Nonlinear Problems; Springer: New York, NY, USA, 2001; pp. 239–249. [Google Scholar]
  14. Fan, J.Y.; Yuan, Y.X. On the quadratic convergence of the Levenberg-Marquardt method without nonsingularity assumption. Computing 2005, 74, 23–39. [Google Scholar] [CrossRef]
  15. Fischer, A. Local behavior of an iterative framework for generalized equations with nonisolated solutions. Math. Program. 2002, 94, 91–124. [Google Scholar] [CrossRef]
  16. Ma, C.; Jiang, L. Some research on Levenberg–Marquardt method for the nonlinear equations. Appl. Math. Comput. 2007, 184, 1032–1040. [Google Scholar] [CrossRef]
  17. Fan, J.Y. A modified Levenberg-Marquardt algorithm for singular system of nonlinear equations. J. Comput. Math. 2003, 625–636. [Google Scholar]
  18. Amini, K.; Rostami, F.; Caristi, G. An efficient Levenberg–Marquardt method with a new LM parameter for systems of nonlinear equations. Optimization 2018, 67, 637–650. [Google Scholar] [CrossRef]
  19. Rezaeiparsa, Z.; Ashrafi, A. A new adaptive Levenberg–Marquardt parameter with a nonmonotone and trust region strategies for the system of nonlinear equations. Math. Sci. 2023, 1–13. [Google Scholar] [CrossRef]
  20. Ahookhosh, M.; Aragón Artacho, F.J.; Fleming, R.M.; Vuong, P.T. Local convergence of the Levenberg–Marquardt method under Hölder metric subregularity. Adv. Comput. Math. 2019, 45, 2771–2806. [Google Scholar] [CrossRef]
  21. Wang, H.Y.; Fan, J.Y. Convergence rate of the Levenberg-Marquardt method under Hölderian local error bound. Optim. Methods Softw. 2020, 35, 767–786. [Google Scholar] [CrossRef]
  22. Zeng, M.; Zhou, G. Improved convergence results of an efficient Levenberg–Marquardt method for nonlinear equations. J. Appl. Math. Comput. 2022, 68, 3655–3671. [Google Scholar] [CrossRef]
  23. Chen, L.; Ma, Y. A modified Levenberg–Marquardt method for solving system of nonlinear equations. J. Appl. Math. Comput. 2023, 69, 2019–2040. [Google Scholar] [CrossRef]
  24. Li, R.; Cao, M.; Zhou, G. A New Adaptive Accelerated Levenberg–Marquardt Method for Solving Nonlinear Equations and Its Applications in Supply Chain Problems. Symmetry 2023, 15, 588. [Google Scholar] [CrossRef]
  25. Grippo, L.; Lampariello, F.; Lucidi, S. A nonmonotone line search technique for Newton’s method. SIAM J. Numer. Anal. 1986, 23, 707–716. [Google Scholar] [CrossRef]
  26. Ahookhosh, M.; Amini, K. A nonmonotone trust region method with adaptive radius for unconstrained optimization problems. Comput. Math. Appl. 2010, 60, 411–422. [Google Scholar] [CrossRef]
  27. Ahookhosh, M.; Amini, K. An efficient nonmonotone trust-region method for unconstrained optimization. Numer. Algorithms 2012, 59, 523–540. [Google Scholar] [CrossRef]
  28. Wang, P.; Zhu, D. A derivative-free affine scaling trust region methods based on probabilistic models with new nonmonotone line search technique for linear inequality constrained minimization without strict complementarity. Int. J. Comput. Math. 2019, 96, 663–691. [Google Scholar] [CrossRef]
  29. Powell, M.J.D. Convergence properties of a class of minimization algorithms. In Nonlinear Programming 2; Elsevier: Amsterdam, The Netherlands, 1975; pp. 1–27. [Google Scholar]
  30. Behling, R.; Iusem, A. The effect of calmness on the solution set of systems of nonlinear equations. Math. Program. 2013, 137, 155–165. [Google Scholar] [CrossRef]
  31. Stewart, G.; Sun, J. Matrix Perturbation Theory; Academic Press: San Diego, CA, USA, 1990. [Google Scholar]
  32. Schnabel, R.B.; Frank, P.D. Tensor methods for nonlinear equations. SIAM J. Numer. Anal. 1984, 21, 815–843. [Google Scholar] [CrossRef]
  33. Moré, J.J.; Garbow, B.S.; Hillstrom, K.E. Testing unconstrained optimization software. ACM Trans. Math. Softw. (TOMS) 1981, 7, 17–41. [Google Scholar] [CrossRef]
  34. Dolan, E.D.; Moré, J.J. Benchmarking optimization software with performance profiles. Math. Program. 2002, 91, 201–213. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Han, Y.; Rui, S. A New Adaptive Levenberg–Marquardt Method for Nonlinear Equations and Its Convergence Rate under the Hölderian Local Error Bound Condition. Symmetry 2024, 16, 674. https://doi.org/10.3390/sym16060674

AMA Style

Han Y, Rui S. A New Adaptive Levenberg–Marquardt Method for Nonlinear Equations and Its Convergence Rate under the Hölderian Local Error Bound Condition. Symmetry. 2024; 16(6):674. https://doi.org/10.3390/sym16060674

Chicago/Turabian Style

Han, Yang, and Shaoping Rui. 2024. "A New Adaptive Levenberg–Marquardt Method for Nonlinear Equations and Its Convergence Rate under the Hölderian Local Error Bound Condition" Symmetry 16, no. 6: 674. https://doi.org/10.3390/sym16060674

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop