Next Article in Journal
Deterministic Asynchronous Threshold-Based Opinion Dynamics in Signed Weighted Graphs
Previous Article in Journal
Generalized Stability of a General Septic Functional Equation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Simplified Algorithm for a Full-Rank Update Quasi-Newton Method

Doctoral School of Applied Informatics and Applied Mathematics, Óbuda University, Bécsi út 96/B, 1034 Budapest, Hungary
AppliedMath 2025, 5(1), 15; https://doi.org/10.3390/appliedmath5010015
Submission received: 23 December 2024 / Revised: 18 January 2025 / Accepted: 5 February 2025 / Published: 8 February 2025

Abstract

:
An efficient linearization method for solving a system of nonlinear equations was developed, showing good stability and convergence properties. It uses an unconventional and simple strategy to improve the performance of classic methods by a full-rank update of the Jacobian approximates. It can be considered both as a discretized Newton’s method or as a quasi-Newton method with a full-rank update of the Jacobian approximates. A solution to the secant equation presented earlier was based on the Wolfe-Popper procedure. The secant equation was splitted into two equations by introducing an auxiliary variable. A simplified algorithm is given in this paper for the full-rank update procedure.It directly solves the secant equation with the pseudoinverse of the Jacobian approximate matrix. Numerical examples are shown for demonstration purposes. The convergence and efficiency of the suggested method are discussed and compared with the convergence and efficiency of classic linearization methods.

1. Introduction

One of the most common ideas for finding the zero of a nonlinear function is to replace it with a series of suitably chosen linear functions, for which the zeros can easily be determined and the sequence of zeros approximate the zero of the nonlinear function. The widely used classic methods are Newton’s method and a large family of quasi-Newton methods (secant, Broyden’s, discretized Newton, Steffensen’s, …). This strategy can be called “linearization”, and such methods may be called “linearization methods”.
A mathematical model ϕ ω , x with some parameters x = x i   i = 1 , , n is constructed for an observed system that gives the observable response D = D j sampled at locations ω j   j = 1 , , m ( m > n ) to an observable external effect. It is assumed that the simulated system response ϕ ω , x is sensitive to the perturbations x i of all parameters. The n adjustable parameters x of the mathematical model are determined so that the distance
l q = f x q = ϕ ω , x D q
between the observed and the simulated system responses ( D j and ϕ ω , x ) is minimized, where l q : R m R is a q norm
l q = j = 1 m f j x q 1 q .
If q = 2 is chosen, then we obtain the Euclidean or least-squares norm
l 2 = j = 1 m f j x 2 1 2 ,
that is well computable and safely usable for a great class of problems. Then, minimizing the distance (1) leads to the problem of minimizing the least-squares norm
l 2 = f x 2 = ϕ ω , x D 2
( x R n , f : R n R m and n < m ). The minimum value of this norm is zero, but for “real life” problems, it is not reachable in most cases due to modeling inaccuracies and measurement noises. The suggested procedure gives a least-squares solution to the overdetermined system of nonlinear equations
f x = 0 ,
where the solution x * minimizes the least squares norm (4) of the nonlinear residual function
f x = ϕ ω , x D .
Least-squares minimizations are successively made within the actual iteration steps. The basic concept of solving Equation (5) is that the function f x is “linearized” by repeatedly replacing it with a linear function y p = y x p as
f x y p = f p + J p x x p
for Newton’s method, where f p = f x p is the function value, J p = J x p is the Jacobian matrix of the function f x ( p is iteration counter), and
f x y p = f p + S p x x p
for the quasi-Newton methods, where S p = S x p is the finite-difference Jacobian approximate of the function f x . Then, the nonlinear problem is solved by solving a series of linear problems
y p = 0 ,
and it follows that the new approximate x p + 1 to the solution x * can be given from Equations
J p x p = f p
(Newton’s method) and
S p x p = f p ,
(quasi-Newton methods) where the iteration stepsize is
x p = x p + 1 x p
in both cases.
The system of simultaneous multi-variable nonlinear Equation (5) can be solved by Newton’s method when the derivatives J p of f x are available analytically, and a new iterate x p + 1 can be determined. In many cases, explicit formulas for the function f x are not available and the Jacobian J p can only be approximated. The partial derivatives of the analytic Jacobian may be replaced by suitable finite-difference quotients (discretized Newton’s iteration [1,2]) with properly chosen stepsizes. For most problems, Newton’s method using analytic derivatives and Newton’s method using properly chosen divided differences are virtually indistinguishable [1]. However, the determination of the finite-difference stepsizes is not clearly defined. The suggested full-rank update procedure helps to overcome this deficit.
Let S p be a full column rank matrix (all column vectors of matrix S p are linearly independent). As n < m (there are more rows than columns), S p is not invertible, and the solution to the overdetermined system of Equation (11) does not exist. However, S p T S p is invertible and the pseudoinverse
S p + = S p T S p 1 S p T
of matrix S p is a unique left inverse which allows finding a least-squares approximate solution x ˜ to Equation (11). The pseudoinverse can be determined in different ways (e.g., rank factorization [3], singular value decomposition [4]). Singular value decomposition is a widely used technique, and it has the advantage that it gives the solution even if S p is singular (not a full column rank matrix). Matrix S p can be factorized as
S p = U p Σ p V p T ,
where U p ( mxn matrix) and V p ( nxn matrix) are orthogonal matrices with orthonormal column eigenvectors, and
Σ p = diag σ i , p
i = 1 , , n with σ i , p singular values of matrix S p and σ 1 σ . Then, the least-square solution to Equation (11) exists, and the iteration stepsize is
x p = S p + f p ,
where
S p + = V p Σ p 1 U p T
is the pseudoinverse of S p and
Σ p 1 = diag 1 σ i , p .
The pseudoinverse + gives a unique least-squares solution to Equation (11) for which the q = 2 norm (Euclidean norm) (4) of the residual Function (6) is minimum. If the rank of matrix S p is less than n (the column vectors of S p are not linearly independent and some singular values are zero), then a unique solution to Equation (11) does not exist. The spectral condition number
κ S p = σ 1 σ n 1
of matrix S p measures the linear dependency of the column vectors. If they are linearly independent (none of the singular values are zero), and the condition number κ is not much larger than one, then the matrix is well conditioned. This is a desirable situation throughout the whole iteration process. Then, the column vectors of matrix S p are in “general positions” and the Ortega–Rheinboldt condition [2] satisfies. If κ 1 , then the matrix is ill-conditioned, and the solution x p (iteration stepsize) to the Equation (11) may be sensitive to small changes in S p or f p .
Classic quasi-Newton linearization methods generate a sequence of improved iterates x p p = 0 , 1 , 2 , so that the next approximate x p + 1 to the solution x * is determined on the basis of rank-1 update S p + 1 of the Jacobian approximate S p . Such linearization update is not unique if only a single new approximate x p + 1 exists. In a single-variable case n = 1 , there are two possible updates (secant lines) from which the selection is not obvious (see details in Section 3). In the multi-variable case, only the rank-1 update is possible if one single new approximate x p + 1 exists.
The solution to Equation (11) presented in [5] and applied to the identification of physically nonlinear dynamic systems [6,7,8] was based on the Wolfe–Popper procedure [9,10]. The secant Equation (11) was splitted into two equations by introducing an auxiliary variable. The simplified solution presented in Section 4 directly solves Equation (11) with the pseudoinverse (13) of matrix S p . The simplified algorithm given in Section 5 gives an unconventional and simple strategy to improve the performance of Newton and quasi-Newton iterations for the solution of the system of nonlinear Equations (5). It can be considered both as a discretized Newton’s method or as a quasi-Newton method with the full-rank update of the Jacobian approximates.
The rate of convergence α of the classic secant method equals the golden section ratio φ ( α = φ 1.618) for a simple root in a single-variable case. The suggested method gives an additional new independent approximate x p + 1 B to the solution x*, then rank- n linearization update can be made with the classic approximate x p + 1 (notated as x p + 1 A in the following) and the suggested new approximate x p + 1 B . The results of single- and multi-variable numerical examples are given in Section 6. The efficiency of the proposed method is discussed in Section 7, and it is compared with other classic rank-one update and line-search methods on the basis of available test data. Concluding remarks are summarized in Section 8.

2. Notations

Vectors and matrices are denoted by bold-face letters. Subscripts refer to components of vectors and matrices, superscripts A and B refer to interpolation base points. Notations A and B are introduced to be able to clearly distinguish between the two new approximates x A and x B . Vectors and matrices may also be given by their general elements. ∆ refers to a difference between two elements. x and X denote unknown quantities, and f and F denote function values and matrices. t and T denote multiplier scalars and scaling transformation matrices. e , ε , and E denote approximate error, p is iteration counter, α is convergence rate, ε * is termination criterion. n is the number of unknowns, m is the number of function values, i , j , and k and are running indexes of matrix columns and rows. Superscripts S and TS refer to the traditional the secant method and to the suggested full-rank update method (T-Secant) receptively.

3. Linearization Methods

The origin of the secant method can be traced back to ancient times to the “rule of double false position” method, described in the 18th century B.C. on the Egyptian Rhind Papyrus [11], and it predates Newton’s method by more than 3000 years. It is also well known that the local convergence rate of the classic Newton’s and the classic secant methods for simple roots are quadratic and superlinear, respectively [12,13,14,15]. Weerakson and Fernando’s third-order Newton variant [16] involves the evaluation of a function value and two derivatives, similarly to Traub’s suggestion [15]. Secant method variations with improved convergence have been reported in the literature [16,17,18,19,20,21,22,23,24]. Mueller’s method [25] uses a quadratic approximation. A class of variations employs two function values and a derivative evaluation [26]. Third-order methods are obtained by a second derivative evaluation, such as Halley’s [21] method and the “super Halley” [27] method. Further improvements with higher-degree polynomial approximations are proposed by Chen [28], Kanwar [20], Zhang [29] and Wang [30].

3.1. Single-Variable Case

The zero x * of a scalar nonlinear function f x ( x f x , x R 1 , f : R 1 R 1 ) has to be determined, where
f x * = 0 .
Linearization means that f x is locally replaced by a linear function y x ( x y x , x R 1 , y : R 1 R 1 ) (secant or tangent line) as shown in Figure 1 (Left) and operations are made on the linear function y p x in the p th iteration. y p x may be defined through two points A p x p A f p A and B p x p B f p B of the nonlinear function f x , where f p A = f x p A and f p B = f x p B are function values, or through one point A p x p A f p A and by the “slope”
S p = f p x p
of the secant line y p x , where x p = x p B x p A and f p = f p B f p A are differences, or by the “slope”
S p = f x p A = df dx
at x p A of the tangent line y p x , where f is the derivative function of f x . The successive local replacement of the nonlinear function f x by a secant or tangent line y p x gives a simple and efficient numerical root-finding procedure. It follows from the condition
y p x p + 1 A = 0
that the zero
x p + 1 A = x p A f p A S p
of the linear function y p x approximates the zero x * of the nonlinear function f x , and the new point A p + 1 x p + 1 A f p + 1 A can be determined for the next iteration, as shown in Figure 1 (Right). With the iteration step length
x p A = x p + 1 A x p A
from point A p , Equation (24) can be rewritten as
S p x p A = f p A ,
that is called the secant equation, or we may also call it a “linearization” equation, as it also includes Newton’s method as a special case (tangent line). The secant-based procedure has the advantage that it does not need the calculation of function derivatives, it only uses function values and the order of asymptotic convergence is superlinear with convergence rate α S 1.618 . However, the “slope” of the updated secant line y p + 1 x can be determined in two ways from the secant (quasi-Newton) conditions as
S p + 1 x p A = f p + 1 A f p A
or
S p + 1 x p B = f p + 1 A f p B ,
where
x p B = x p + 1 A x p B
is the iteration step length from point B p . The decision is far from obvious, as shown in Figure 1 (Right). The tangent-line-based iteration (Newton’s method) has an advantage that the “slope”
S p + 1 = f x p + 1 A
of the updated tangent line is always well defined. The iteration then continues with the updated secant or tangent line y p + 1 x .

3.2. Multi-Variable Case

The scalar linearization procedure can be extended to multi-dimensions. The zero x * of a nonlinear vector-valued function
f x = f 1 x f m x
of n variables
x = x 1 x n
( x f x , x R n , f : R n R m , n m ) has to be determined, where
f x * = 0 .
Then, f x is locally replaced by an n -dimensional hyperplane
y x ( x y x , x R n , y : R n R m
n m and operations are made on the hyperplane y p x in the p th iteration. y p x may be defined through n + 1 points A p x p A f p A and B k , p x k , p B f k , p B ( k = 1 , , n ) of the nonlinear function f x , where f p A = f x p A and f k , p B = f x k , p B ( k = 1 , , n ) are function values, or by one point A p x p A f p A and by the “slope” (divided differences)
S p = F p X p 1
i = 1 , , n , j = 1 , , m , k = 1 , , n of the n -dimensional secant hyperplane y p x , where x k , p = x k , p B x p A and f k , p = f k , p B f p A are difference vectors with n and m components, respectively, and
F p = f k , p = f k , p B f p A = f 1 , 1 , p f n , 1 , p f 1 , m , p f n , m , p ,
X p = x k , p = x k , p B x p A = x 1 , 1 , p x n , 1 , p x 1 , n , p x n , n , p .
The n -dimensional tangent hyperplane y p x through point A p x p A f p A may be defined by the “slope”
S p = f k , j , p x k , i , p = f 1 , 1 , p x 1 , p f n , 1 , p x n , p f 1 , m , p x 1 , p f n , m , p x n , p ,
that corresponds to the Jacobian matrix J p of the nonlinear vector-valued function f x , and accordingly, Definition (35) corresponds to the approximation of the Jacobian matrix. It follows from the condition
y p x p + 1 A = 0
that the zero
x p + 1 A = x p A S p + f p A
of the n -dimensional hyperplane y x approximates the zero x * of the nonlinear vector-valued function f x , where . + stands for the pseudoinverse. Then, the i th element of the new approximate x p + 1 A in the p th iteration will be
x i , p + 1 A = x i , p A j = 1 m S i , j , p + f j , p A
i = 1 , , n , j = 1 , , m . With the iteration stepsize
x p A = x p + 1 A x p A = S p + f p A = j = 1 m S i , j , p + f j , p A i ,
Equation (40) can be rewritten as
S p x p A = f p A ,
which is called “secant equation” or “linearization equation” in the multi-variable case. With the new approximate x p + 1 A and with the corresponding function value f p + 1 A , an updated Jacobian or Jacobian approximate S p + 1 has to be determined. Table 1 summarizes the basic equations of the above-detailed linearization methods.
If all partial derivatives of the function f x are known, then the Jacobian update can easily be calculated. Newton’s method is one of the most widely used algorithms, with very attractive theoretical and practical properties and with some limitations. The computational costs of Newton’s method are high, since the Jacobian J p and the solution to the linear system (43) must be computed at each iteration. It is well known that the local convergence of Newton’s method is q -quadratic if the initial trial approximate x 0 is close enough to the solution x * , if J ( x * ) is nonsingular, and if J ( x ) satisfies the Lipschitz condition
J ( x ) J ( x * ) L x x *
for all x close enough to x * .
However, in many cases, the function f x is not an analytical function, the partial derivatives are not known or difficult to evaluate, and Newton’s method cannot be applied. Quasi-Newton methods are widely used for solving systems of nonlinear equations when the Jacobian is not known or difficult to determine. The Jacobian is approximated by divided differences (35), and the system of nonlinear Equations (5) is solved by repeatedly solving systems of linear Equations (43) providing the new approximates x p + 1 A .
The Jacobian approximate (35) should be updated according to the fundamental equation of the quasi-Newton methods (“quasi-Newton condition” or “secant condition”)
S p + 1 x p A = f p + 1 A f p A
for all p = 0 , 1 , 2 , . However, the condition (45) does not uniquely specify the update S p + 1 , so the iterative procedure is not well defined and further constraints are needed. Different methods offer their own specific solution, but new quasi-Newton approximate x p + 1 will never allow full-rank update of the Jacobian approximate S p + 1 (Equation (45) is an underdetermined system of m linear equations with mn unknowns). A large family of methods is available with different additional conditions or assumptions. Martinez [31] has been made a thorough survey on the family of practical quasi-Newton methods.
The partial derivatives of the Jacobian matrix may be replaced by suitable difference quotients (discretized Newton iteration; see [1,2])
f x = f j x + x k d k f j x x k = f j x x k
k = 1 , , n , j = 1 , , m with n additional function value evaluations, where d k is the k th Cartesian unit vector. However, it is difficult to choose the stepsize x . If any x k is too large, then Expression (46) can be a bad approximation to the Jacobian so the iteration converges much more slowly, if it converges at all. On the other hand, if any x k is too small, then f j x 0 , and cancellations can occur, which reduces the accuracy of the difference quotients (46) (see [32]). Another modification is the inexact-Newton approach, when the nonlinear equation is solved by an iterative linear solver (see [33,34,35]). Wolfe [10] and Popper [9] suggested the column update of matrix (36).
One of the most widely used formulas is that of Broyden [36], which makes a rank-one update for the Jacobian approximate as
S p + 1 = S p + Q p = S p + f p A S p x p A x p A 2 x p A T ,
where Q p is a rank-one matrix, and by using the Sherman–Morrison formula, the inverse Jacobian approximate update is given as
S p + 1 + = S p + + x p A S p + f p A x p A T S p + f p A x p A T S p + .
Broyden’s secant condition (45) can be rewritten as
S p x p A + Q p x p A = f p + 1 A f p A ,
and with the secant Equation (43), we have
Q p x p A = f p + 1 A .
As Q p is a rank-one matrix, then this equation has an infinitely large number of solutions (the left nullspace of matrix Q p is an n 1 dimensional vector space [37]), and the Ortega–Rheinboldt [2] condition will not be satisfied (the column vectors of matrix S p should be linearly independent and have to be “in general position” through the whole iteration process) and it may result an ill-conditioned Broyden update S p + 1 .

4. T-Secant Method

A numerical procedure has been developed for solving an overdetermined system of nonlinear Equations (5). It can be considered both as a discretized Newton’s method or as a quasi-Newton method with full-rank update of the Jacobian approximates. The suggested full-rank update procedure (“T-Secant” method [5]) provides an unconventional and simple strategy to improve the performance of quasi-Newton iterations. It is based on the classic secant linearization, which was completed with a new independent approximate x p + 1 B . The secant equation was modified by a suitably chosen nonuniform scaling transformation and a new independent approximate x p + 1 B was determined from the modified secant equation. The solution to the secant equation presented in [5] was based on the Wolfe–Popper procedure [9,10] so that the secant equation was splitted into two equations by introducing an auxiliary variable. The simplified algorithm, presented in this paper, directly solves the secant equation by using the pseudoinverse of the transformed Jacobian approximate.

4.1. Single-Variable Case

The suggested procedure uses the information on the improvement of the classic secant procedure and gives a new approximate x p + 1 B in the vicinity of the classic secant approximate x p + 1 A as follows. Given two initial approximates x p A and x p B with function values f p A and f p B , the new approximate x p + 1 A with function value f p + 1 A is known from the solution of the classic secant Equation (26). An independent new approximate
x p + 1 B = x p + 1 A + x p + 1
is planned to be determined in the vicinity of the classic secant approximate x p + 1 A . Let the ratio
t p f = f p + 1 A f p A
f p A 0 of the function value improvement and the ratio
t p x = x p + 1 B x p + 1 A x p + 1 A x p A = x p + 1 x p A
x p A 0 of the desired iteration stepsize change be defined. The single-variable secant Equation (26) is then modified as
S Tp x p A = f p A ,
where
S Tp = t p f t p x S p .
Then, the new approximate x p + 1 B can be expressed from Equation (54) as
t p f x p + 1 A x p A x p + 1 B x p + 1 A S p x p A = f p A .
After rearrangement, we obtain
x p + 1 A x p + 1 B = t p f f p A S p x p A 2
and with the secant Equation (26), it turns to
x p + 1 B = x p + 1 A t p f f p A S p x p A 2 = x p + 1 A + t p f x p A
and the desired new iteration stepsize
x p + 1 = t p f x p A
can be given. The suggested procedure can also be applied to Newton’s method (T-Newton method). Then, the “slope” S p corresponds to the derivative f of function f x .
The geometrical representation of the suggested method can be derived as follows ([5]). Let the new approximate x p + 1 B be the zero of a function z p x :
z p ( x p + 1 B ) = 0 .
Then, replacing x p + 1 B with x in Equation (58) gives the function
z p x = t p f S p x p A 2 x x p + 1 A + f p A = 0
with zero at
x = x p + 1 B .
Equation (61) is a hyperbolic function with vertical and horizontal asymptotes x p + 1 A and f p A , and its root x p + 1 B will be in the vicinity of x p + 1 A in an “appropriate distance” that is regulated by the function value f p + 1 A (see Figure 2). This virtue of the suggested procedure ensures an automatic mechanism for having the actual approximates x p A and x p B in general positions through the whole iteration process providing stable and efficient numerical performance. The suggested method’s performance is demonstrated with the results of numerical tests on test functions in Section 6.

4.2. Multi-Variable Case

Let two independent approximates x p A = x k , p A and x p B = x k , p B to the zero x * of the nonlinear vector-valued function f x be given in the p th iteration k = 1 , , n . Let the series of n approximates x k , p B be constructed by individually increment the elements x k , p A of the approximate x p A by an increment
x k , p = x k , p B x k , p A
as
x k , p B = x p A + x k , p d k
where d k is the k th Cartesian unit vector. It follows from this special construction of the approximates x k , p B that x k , i , p B x k , p A = 0 for i k and x k , i , p B x k , p A = x k , p for i = k , and matrix (37) will be a diagonal matrix:
X p = x k , p = x k , p B x p A = x 1 , p 0 0 x n , p = diag x k , p .
Let the ratios
T p F = diag t j , p F = diag f j , p + 1 A f j , p A
f j , p A 0 of the function value improvements j = 1 , , m and the ratios
T p X = diag t i , p X = diag x i , p + 1 B x i , p + 1 A x i , p + 1 A x i , p A = diag x i , p + 1 x i , p A
x i , p A 0 of the desired iteration stepsize changes i = 1 , , n be defined similarly as in the single-variable case. Let the secant Equation (43) be modified as
S Tp x p A = f p A ,
where
S Tp = T p F S p T p X 1 = diag t j , p F f k , j , p x i , p diag 1 t i , p X = t j , p F t i , p X f k , j , p x i , p ,
i = 1 , , n , j = 1 , , m , k = 1 , , n or in explicit form:
S Tp = f 1 , p + 1 A f 1 , p A 0 0 0 0 0 0 f m , p + 1 A f m , p A f 1 , 1 , p x 1 , p f 1 , n , p x n , p f 1 , m , p x 1 , p f n , m , p x n , p x 1 , p A x 1 , p + 1 0 0 0 0 0 0 x n , p A x n , p + 1 ,
and
S Tp + = T p X S p + T p F 1
is the pseudoinverse of S p . Then, Equation (68) can be rewritten as
x p A = S Tp + f p A .
Then, the i th element x i , p + 1 B of the new approximate x p + 1 B in the p th iteration will be
x i , p + 1 B = x i , p + 1 A + x i , p + 1 = x i , p + 1 A x i , p A 2 j = 1 m S i , j , p + f j , p A t j , p F ,
where t p , j F 0 , j = 1 , , m and i = 1 , , n . Let the ratios
μ i , p = j = 1 m S i , j , p + f j , p A j = 1 m S i , j , p + f j , p A t j , p F = x i , p A j = 1 m S i , j , p + f j , p A t j , p F
be introduced . By using Equation (42), the new iteration stepsize can be expressed from Equation (73) as
x i , p + 1 = x i , p + 1 B x i , p + 1 A = x i , p A 2 j = 1 m S i , j , p + f j , p A t j , p F = μ i , p x i , p A .
Table 2 summarizes the basic equations of the above-detailed multi-variable update method. The suggested procedure can also be applied to Newton’s method (T-Newton method). Then, matrix S p corresponds to the Jacobian matrix J p of function f x . The function value improvement parameter t j , p F   j = 1 , , m in Equation (74), defined as (66), has a key role in the suggested iteration process, and the absolute value of the denominator f j , p A has to be low-bounded with a pregiven value T min to avoid division by zero. When the values f j , p A tend to zero with increasing iteration counter p , then
μ i , p T min .
Figure 3 shows the effect of iteration parameter T min on the iteration process ( N = 1000 ). It is clear that too high a value of T min has a negative effect on the iteration efficiency. It can be seen in the example that a value 0.001 causes the iteration process failure, and there is almost no effect with lower values 0.0001 T min .
The suggested procedure may also be repeated without using the classic secant approximates and without updating the previously known Jacobian approximate S p as
x i , p + 1 + h = x i , p + 1 + h B x i , p + 1 + h A = j = 1 m S i , j , p + f j , p + h A 2 j = 1 m S i , j , p + f j , p + h A t j , p + h F ,
where
t j , p + h F = f j , p + 1 + h A f j , p + h A
h = 1 , until the new approximates
x i , p + 1 + h B = x i , p + 1 + h A + x i , p + 1 + h
i = 1 , , n are sufficiently improved. A numerical example is shown in Table 3 with test function
f x = x 3 2 x 5
for demonstration purpose. The results indicate linear convergence, with convergence rate α = 1 . α is the computed convergence rate, N f is the number of function value evaluations, and L is the mean convergence rate, suggested by Broyden [36] in Table 3.

5. Algorithm

Let ε * be the error bound for termination criterion, and if x * is known, then
e p A = x p A x *
is the error vector of approximate x p A in the p th iteration with elements e i , p A   i = 1 , , n . Let the error norm
ε p = e p A 2 n = i = 1 n e i , p A 2 n
be defined, where . 2 is Euclidean norm, and let the iteration be terminated when
ε p < ε *
holds. Choose T min as the lower bound for t j , p F j = 1 , m , and let f min be a lower bound for f j , p A   j = 1 , , m . Let p = 0 , and let the approximate x p A and the difference vector x p be given. Calculate the corresponding function values f p A and assure that f min < f j , 0 A   j = 1 , , m . Iteration constants f min and T min are necessary to avoid division by zero and to avoid computed values be near the numerical precision.
  • Step 1: Generate a set of n additional approximates x k , p B (Equation (64)) and evaluate function values f k , p B   k = 1 , , n . Assure that f min < f k , j , p B   j = 1 , , m .
  • Step 2 (Classic secant update method): Construct the Jacobian approximate matrix S p , determine its pseudoinverse S p + [4], calculate x p + 1 A from Equation (40) and ε p from Equation (82).
  • Step 3: If ε p < ε * , then terminate iteration, else continue with Step 4.
  • Step 4 (suggested update method): Calculate f p + 1 A (assure that f min < f j , p A ), T p F from Equation (66) and μ i from Equation (74). Let T min < t j , p F and determine x p + 1 B from Equation (75).
  • Step 5 : Continue iteration from Step 1 with p = p + 1 , x p + 1 A and x p + 1 B .
If p max is the number of necessary iterations for satisfying the termination criterion ε p < ε * and n is the number of unknowns to be determined, then the suggested update method needs n + 1 function evaluations in each iterations and altogether
N f = p max n + 1
function evaluations to reach the desired termination criterion. p max is dependent on many circumstances, such as the nature of the function f x , termination criteria ( ε * or others), the distance of the initial approximate x A from the solution x * , and from the iteration constants f min and T min .

6. Numerical Tests Results

6.1. Single-Variable Test Function

A numerical example is given with a single-variable test function
f x = cos x x ,
with root x * 0.73908513 The results of the classic secant iteration process are shown in Table 4 and Figure 4. Iterations were made with initial approximates x 0 A = 2.0 and x 0 B = 2.0   p = 0 providing f 0 A = 1.584 . The first secant approximate x 1 A = 0.416 is found as the zero of the first secant line y 0 x providing f 1 A = 1.331   p = 1 . The next iteration ( p = 2 ) will continue with approximate x 2 A = 0.442 and with f 2 A = 0.462 .
The results of the suggested iteration process are shown in Table 5 and Figure 5. Iterations were made with the same initial approximates as in the case of the classic secant iteration, and the first secant approximate x 1 A = 0.416 is found. The first T-secant approximate x 1 B = 0.915 is given as the zero of the first hyperbola function z 0 ( x ) (Figure 5, left). Iteration then goes on with approximates x 1 A = 0.416 and x 1 B = 0.915 providing f 1 A = 1.331   p = 1 , and new approximates x 2 A = 0.667 and x 2 B = 0.764 as the zeros of the second secant and hyperbola functions y 1 ( x ) and z 1 ( x ) receptively (Figure 5, left). The next iteration ( p = 2 ) will then continue with approximates x 2 A = 0.667 and x 2 B = 0.764 and with f 2 A = 0.119 and gives x 3 A = 0.7387 and x 3 B = 0.7391 (Figure 5, right).

6.2. Solution of an Inverse Problem

An example is given for an imaginary modeling problem. The modeling of systems on the basis of observed or specified system responses plays a key role in scientific research. A central problem of modeling is to fit a structured mathematical model to available data so that the values of the unknown model parameters are determined providing that the simulated data are as near to the available data as possible. A system response can generally be represented by a “curve” in a two (or three)-dimensional space and can be simulated by a computer program. Parameter identification corresponds to minimizing the distance between observed and simulated system responses. Parameter identification problems can generally be formulated as nonlinear least-squares problems so that some unknown model parameters are determined providing minimal deviation between observed and simulated system responses. Solving nonlinear least-squares problems is often a difficult task, especially if the number of unknowns is high.
A rational system under investigation produces a measurable response to a known measurable external effect. A mathematical model with n unknown parameters p 1 , , p n simulates the behavior of the system providing response to the same external effect as acted on the system during observation. Let the observed and the simulated system responses be represented by two-dimensional curves and sampled in k discrete points with coordinates x i and y i i = 1 , k . The coordinates of a synthetic observed response were generated as
x z = a + b cos z + α c cos a + b b z + α + x 0
y z = a + b sin z + α c sin a + b b z + α + y 0
(an epicycloid) with parameters x 0 = 10 , y = 8 , a = 4 , b = 2 , c = 3.5 and α = 1 and with 0 z 2 π . The parameters
p = x 0 y 0 a b c
were considered to be unknown, and the system response was simulated with initial approximate
p 0 = x 0 y 0 a b c = 8 11 3 . 5 2 . 5 3 .
The distance between observed and simulated system responses was defined for arbitrary two-dimensional curves. This definition can be extended to n dimensions making a wide range of parameter identification problems possible to formulate. The value of the defined distance is dimensionless, and it expresses the ratio of the area between the system response graphs in a normalized coordinate system to the area of a rectangle with unit area.
The distance between the observed and simulated system responses was quantified by area D p between the graphs of system responses in a normalized coordinate system. Partial area r j p was defined as the area of a quadrangle among the j 1 t h and j th division points on the system response graphs j = 1 , , m , as shown in Figure 6. These division points were selected equidistantly along system response graphs.
Thus, parameter identification problems can be formulated as solving a system of nonlinear equations
r p = 0 ,
where p r ( p ) , p R 5 , r : R 5 R m and r is residual vector with components r j j = 1 , , m . The unknown parameters p of an epicycloid (86) and (87) were determined according to the suggested update algorithm. The variation in parameters p through iterations and the observed and simulated system responses for the initial approximate are shown in Figure 6. The observed and simulated system response pairs are shown in Figure 7 7, 13, 19, 25, 31 and 43 simulations ( S ).
The results of numerical tests with different initial approximates show that the optimal solutions were not reached in all cases. The convergence was especially sensitive for the initial values of parameters a , b , and c . Several local optimal solutions were detected, as shown in Figure 8.

7. Efficiency

The efficiency of an algorithm for the solution of nonlinear equations is thoroughly discussed by Traub [15], as follows. Let β be the order of the iteration sequence such that for the approximate errors e i = x i x * , there exists a nonzero constant C (asymptotic error constant) for which
e i + 1 e i β C .
A natural measure of the information used by an algorithm is the “informational usage” d , which is defined as the number of new pieces of information (values of the function and its derivatives) required per iteration (called “horner” by Ostrowski [14]). Then, the efficiency of the algorithm within one iteration can be measured by the “informational efficiency”
EFF = β d .
An alternative definition of efficiency is
* EFF = β 1 d ,
called “efficiency index” by Ostrowski [14]. Another measure of efficiency, called “computational efficiency”, takes into account the “cost” of calculating different derivatives. The concepts of informational efficiency ( EFF ) and efficiency index ( * EFF ) do not take into account the cost of evaluating f and its derivatives, nor do they take into account the total number of pieces of information needed to achieve a certain accuracy in the root of the function. If f is composed of elementary functions, then the derivatives are also composed of elementary functions; thus, the cost of evaluating the derivatives is merely the cost of combining the elementary functions.
Broyden [36] suggested the mean convergence rate
L = 1 N f ln R x 0 A R x p max A
as a measure of the efficiency of an algorithm for solving a particular problem, where N f is the total number of function evaluations, x 0 A is the initial approximate, x p max A is the last approximate to the solution x * when the termination criterion is satisfied after p max iterations. R x is the Euclidean norm of f x .
Zang’s two-step method [29] is similar to the King–Werner method [38,39] with convergence rate 2.414 [40]. Chen’s [28] and Wang’s [30] three-step methods show asymptotic convergence order 2.732 . Chen’s [28] method requires two functions and one derivative evaluation using a second-order polynomial, proposed in [30] and shows 1.174 Ostrowski’s [14] efficiency index. Wang’s [30] method requires three functions and two derivative evaluations showing 1.101 Ostrowski index. Zhang [29] published a finite-difference-based secant method with 2.618 asymptotic convergence order. Later, Ren [40] showed that Zhang’s [29] method has only a 2.414 convergence order due to some mistakes in the derivation. Table 6 compares the efficiencies of classic (Rows 1–2: Secant, Newton) and improved algorithms (Rows 3, 5: T-Secant, T-Newton with the suggested full-rank update and from references [28,30]), the T-secant method with constant Jacobian approximate S p (row 4: TS-const. S p , see data in Table 3), and Chen’s [28] and Wang’s [30] methods.
Very limited data are available to compare the performance of the suggested update method (T-Secant) with other classic methods, especially for a large number of unknowns. Efficiency results were given by Broyden [36] for the Rosenbrock function for N = 2 . The calculated convergence rates for the two Broyden method variants [36], for the Powell’s method [41], for the adaptive coordinate descent method [42], and for the Nelder–Mead simplex method [43] were compared with the calculated values for the T-secant method in Table 7 [5]. Rows 1–5 are data from referenced papers, rows 6–8 are T-secant results with the referenced initial approximates, and rows 9–15 are calculated data for N > 2 . If the value of R x = 0 (for p max ) is zero, then the mean convergence rates ( L and L N ) are not countable (zero in the denominator). A substitute value 10 25 was used when R x = 0 (for p max ) in rows 6, 7, 10, and 13.
Results show that the mean convergence rate L (Equation (94)) for N = 2 is much higher for the T-secant method ( 5.5 6.9 ) than for the other listed methods ( 0.1 0.6 ). However, it is obvious that the mean convergence rate values decrease rapidly with increasing N values (more unknowns need more function evaluations). A modified convergence rate
L N = N L = N N f ln R x 0 A R x p max A
is suggested to use as an “ N ” independent measure of efficiency (see Table 7). The values of L and L N are at least 10 times larger for the T-secant method than for the referenced classic methods for N = 2 (see Table 7).
The efficiency measures ( L and L N ) also depend on the initial conditions (distance of the initial approximate from the optimal solution and termination criterion). Results from a large number of numerical tests indicate an average L N 7.4 with a standard deviation of around 3.7 for the T-secant method, even for large N values.

8. Conclusions

A numerical procedure has been developed for solving an overdetermined system of nonlinear Equation (5). It can be considered both as a discretized Newton method or as a quasi-Newton method with a full-rank update of the Jacobian approximates. Quasi-Newton methods are widely used for solving systems of nonlinear equations when the function derivatives (Jacobian) are not known or difficult to determine. The derivatives (Jacobian) are approximated by divided differences (35), and the system of nonlinear Equation (5) is solved by repeatedly solving systems of linear Equation (43). The new approximate x p + 1 A is determined from the classic secant equation, and the divided differences are updated according to the secant condition (45). However, the secant condition does not uniquely specify the Jacobian approximate, so the update procedure is not well defined, and further constraints are needed. Different methods offer specific update solutions.
The suggested numerical procedure (“T-Secant” method [5]) provides an unconventional and simple strategy to improve the performance of quasi-Newton iterations by a full-rank update of the Jacobian approximates. It is based on the classic secant linearization and allows a full-rank update of the Jacobian approximates. The classic secant iteration was completed with a new independent approximate x p + 1 B , that was determined from a modified secant Equation (68). A modification was made by a suitably chosen nonuniform scaling transformation. Scaling was made by the difference quotients of the classic secant function value improvements (66), the difference quotients of the desired new iteration stepsizes, and the classic secant stepsizes (67). A new independent approximate x p + 1 B was determined from the modified secant Equation (68). The solution to the secant equation presented in [5] was based on the Wolfe–Popper procedure [9,10] so that the secant equation was splitted into two equations by introducing an auxiliary variable. The simplified algorithm, presented in the paper, directly solves the secant equation by using the pseudoinverse (71) of the transformed Jacobian approximate (69).
It has been shown that the new T-secant approximate x p + 1 B will be in the vicinity of the classic secant approximate x p + 1 A if the classic secant iterates converge to the root of the nonlinear function [5]. The Jacobian approximate was then full-rank updated by constructing new divided differences (Jacobian approximates) from the classic secant and from the T-secant approximates.
It was shown that the iterative procedure possesses superquadratic asymptotic convergence property with convergence rate α = φ + 1 = φ 2 2 . 618 for simple root in a single-variable case ([5]), where φ is the golden section ratio. The suggested procedure can also be applied to Newton’s method (matrix S p corresponds to the Jacobian matrix J p ). Numerical test results indicate that the quadratic convergence ( α = 2 ) of Newton’s method increases to cubic ( α = 3 ).
The efficiency has been studied in a multi-variable case and compared with other classic rank-one update and line-search methods on the basis of available data. Results show that its efficiency is considerably better than the efficiency of other classic low-rank update methods. The suggested method’s performance was demonstrated by the results of numerical tests with widely used single- and multi-variable benchmark test functions. A Rosenbrock test function was used with up to 1000 variables [5]. The method has also been successfully applied for the solution of different “real-life” inverse problems [6,7,8] on physically nonlinear dynamic systems. Further studies can be carried out on convergence properties in the case of multiple roots and on the efficiency characteristics in cases of more benchmark test functions. The suggested procedure may be used for other applications, especially with a large number of unknowns and when a low number of function evaluations is crucial.

Funding

This research received no external funding.

Data Availability Statement

Data are contained within the article.

Acknowledgments

A considerable part of the research work has been performed between the years 1988 and 1992 at Technical University of Budapest (Hungary), at TNO-BOUW Structural Division (The Netherlands) and at Technical High-school of Lulea (Sweden). The work has been sponsored by the Technical University of Budapest (Hungary), by the Hungarian Academy of Sciences (Hungary), by TNO-BOUW (The Netherlands), by Sandvik Rock Tools (Sweden), by CP Test a/s (Denmark), and by Óbuda University (Hungary). Valuable discussions and personal support from Géza Petrasovits, György Popper, Peter Middendorp, Rikard Skov, Bengt Lundberg, Mario Martinez and Csaba J. Hegedűs are greatly appreciated.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Dennis, J.E., Jr.; Schnabel, R.B. Numerical Methods for Unconstrained Optimization and Nonlinear Equations; Prentice-Hall: Englewood Cliffs, NJ, USA, 1983. [Google Scholar]
  2. Ortega, J.M.; Rheinboldt, W.C. Iterative Solution of Nonlinear Equations in Several Variables; Academic Press: New York, NY, USA, 1970. [Google Scholar]
  3. Hegedus, C. Numerical Methods I; ELTE, Faculty of Informatics: Budapest, Hungary, 2015. [Google Scholar]
  4. Press, W.H.; Flannery, B.P.; Teukolsky, S.A.; Wetterling, W.T. Numerical Recepies; Cambridge University Press: Cambridge, UK, 1986. [Google Scholar]
  5. Berzi, P. Convergence and Stability Improvement of Quasi-Newton Methods by Full-Rank Update of the Jacobian Approximates. AppliedMath 2024, 4, 143–181. [Google Scholar] [CrossRef]
  6. Berzi, P.; Beccu, R.; Lundberg, B. Identification of a Percussive Drill Rod Joint from its Response to Stress Wave Loading. Int. J. Impact Eng. 1994, 18, 281–290. [Google Scholar] [CrossRef]
  7. Berzi, P. Pile-Soil Interaction due to Static and Dynamic Load. In Proceedings of the 13th International Conference on Soil Mechanics and Foundation Engineering, New Delhi, India, 5–10 January 1994; pp. 609–612. [Google Scholar]
  8. Berzi, P.; Popper, G. Evaluation of dynamic load test results on piles. In Proceedings of the International Symposium on Identification of Nonlinear Mechanical Systems from Dynamic Tests (Euromech 280), Ecully, France, 29–31 October 1991; pp. 121–128. [Google Scholar]
  9. Popper, G. Numerical method for least square solving of nonlinear equations. Period. Polytech. 1985, 29, 67–69. [Google Scholar]
  10. Wolfe, P. The Secant Method for Simultaneous Nonlinear Equations. Commun. ACM 1959, 2, 12–13. [Google Scholar] [CrossRef]
  11. Papakonstantinou, J.M.; Tapia, R.A. Origin and evolution of the secant method in one dimension. Am. Math. Mon. 2013, 120, 500–518. [Google Scholar] [CrossRef]
  12. Broyden, C.G.; Dennis, J.E.; Mor, J.J. On the local and superlinear convergence of quasi-Newton methods. J. Inst. Math. Appl. 1973, 12, 223–245. [Google Scholar] [CrossRef]
  13. Dennis, J.E.; Mor, J.J. A characterization of superlinear convergence and its application to quasi-Newton methods. Math. Comput. 1974, 28, 543–560. [Google Scholar] [CrossRef]
  14. Ostrowski, A.M. Solution of Equations and Systems of Equations; Academic Press: New York, NY, USA, 1966. [Google Scholar]
  15. Traub, J.F. Iterative Methods for the Solution of Equations, 1st ed.; Prentice-Hall, Inc.: Englewood Cliffs, NJ, USA, 1964. [Google Scholar]
  16. Weerakoon, S.; Fernando, T.G.I. A variant of Newton’s method with accelerated third-order convergence. Appl. Math. Lett. 2000, 13, 87–93. [Google Scholar] [CrossRef]
  17. Gerlach, J. Accelerated convergence in Newton’s method. SIAM Rev. 1994, 36, 272–276. [Google Scholar] [CrossRef]
  18. Homeier, H.H.H. On Newton-type methods with cubic convergence. J. Comput. Appl. Math. 2005, 176, 425–432. [Google Scholar] [CrossRef]
  19. Kou, J.; Li, Y.; Wang, X. Third-order modification of Newton’s method. J. Comput. Appl. Math. 2007, 205, 1–5. [Google Scholar]
  20. Kanwar, V.; Sharma, J.R.; Mamta, J. A new family of Secant-like method with super-linear convergence. Appl. Math. Comput. 2005, 171, 104–107. [Google Scholar] [CrossRef]
  21. Melman, A. Geometry and convergence of Euler’s and Halley’s methods. SIAM Rev. 1997, 39, 728–735. [Google Scholar] [CrossRef]
  22. Özban, A.Y. Some new variants of Newton’s method. Appl. Math. Letter. 2004, 17, 677–682. [Google Scholar] [CrossRef]
  23. Scavo, T.R.; Thoo, J.B. On the geometry of Halley’s method. Am. Math. Mon. 1995, 102, 417–426. [Google Scholar] [CrossRef]
  24. Shaw, S.; Mukhopadhyay, B. An improved regula falsi method for finding simple roots of nonlinear equations. Appl. Math. Comput. 2015, 254, 370–374. [Google Scholar] [CrossRef]
  25. Muller, D.E. A Method for Solving Algebraic Equations Using an Automatic Computer. Math. Tables Other Aids Comput. 1956, 10, 208–215. [Google Scholar] [CrossRef]
  26. Thukral, R. A New Secant-type method for solving nonlinear equations. Am. J. Comput. Appl. Math. 2018, 8, 32–36. [Google Scholar]
  27. Amat, S.; Busquier, S.; Gutiérrez, J.M. Geometric constructions of iterative functions to solve nonlinear equations. J. Comput. Appl. Math. 2003, 157, 197–205. [Google Scholar] [CrossRef]
  28. Chen, L.; Ma, Y. A new modified King–Werner method for solving nonlinear equations. Comput. Math. Appl. 2011, 62, 3700–3705. [Google Scholar] [CrossRef]
  29. Zhang, H.; Li, D.-S.; Liu, Y.-Z. A new method of secant-like for nonlinear equations. Commun. Nonlinear Sci. Numer. Simul. 2009, 14, 2923–2927. [Google Scholar]
  30. Wang, X.; Kou, J.; Gu, C. A new modified secant-like method for solving nonlinear equations. Comput. Math. Appl. 2010, 60, 1633–1638. [Google Scholar] [CrossRef]
  31. Martínez, J.M. Practical quasi-Newton methods for solving nonlinear systems. J. Comput. Appl. Math. 2000, 124, 97–121. [Google Scholar] [CrossRef]
  32. Stoer, J.; Bulirsch, R. Introduction to Numerical Analysis; Springer: Berlin/Heidelberg, Germany, 2002. [Google Scholar]
  33. Birgin, E.G.; Krejic, N.; Martinez, J.M. Globally convergent inexact quasi-Newton methods for solving nonlinear systems. Num. Algorithms 2003, 32, 249–260. [Google Scholar] [CrossRef]
  34. Dembo, R.S.; Eisenstat, S.C.; Steihaug, T. Inexact Newton methods. SIAM J. Numer. Anal. 1971, 19, 400–408. [Google Scholar] [CrossRef]
  35. Martinez, J.M.; Qi, L. Inexact Newton methods for solving non-smooth equations. J. Comput. Appl. Math. 1995, 60, 127–145. [Google Scholar] [CrossRef]
  36. Broyden, C.G. A class of Methods for Solving Nonlinear Simultaneous Equations. Math. Comput. Am. Math. 1965, 19, 577–593. [Google Scholar] [CrossRef]
  37. Strang, G. Introduction to Linear Algebra, revised international ed.; Wellesley-Cambridge Press: Wellesley, MA, USA, 2005. [Google Scholar]
  38. King, R.F. Tangent method for nonlinear equations. Numer. Math. 1972, 18, 298–304. [Google Scholar] [CrossRef]
  39. Werner, W. Über ein Verfarhren der Ordnung 1 + √ 2 zur Nullstellenbestimmung. Numer. Math. 1979, 32, 333–342. [Google Scholar] [CrossRef]
  40. Ren, H.; Wu, Q.; Bi, W. On convergence of a new secant-like method for solving nonlinear equations. Appl. Math. Comput. 2010, 217, 583–589. [Google Scholar] [CrossRef]
  41. Powell, M.J.D. An efficient method for finding the minimum of a function of several variables without calculating derivatives. Comput. J. 1964, 7, 155–162. [Google Scholar] [CrossRef]
  42. Loshchilov, I.; Schoenauer, M.; Sebag, M. Adaptive Coordinate Descent. In Proceedings of the Genetic and Evolutionary Computation Conference (GECCO), Dublin, Ireland, 12–16 July 2011; ACM Press: New York, NY, USA, 2011; pp. 885–892. [Google Scholar]
  43. Nelder, J.A.; Mead, R. A simplex method for function minimization. Comput. J. 1965, 7, 308–313. [Google Scholar] [CrossRef]
Figure 1. Left: linearization of a nonlinear function; Right: classic secant method.
Figure 1. Left: linearization of a nonlinear function; Right: classic secant method.
Appliedmath 05 00015 g001
Figure 2. Suggested full-rank update method (T-secant).
Figure 2. Suggested full-rank update method (T-secant).
Appliedmath 05 00015 g002
Figure 3. Effect of the iteration parameter T min on the iteration process ( N = 1000 ).
Figure 3. Effect of the iteration parameter T min on the iteration process ( N = 1000 ).
Appliedmath 05 00015 g003
Figure 4. Classic secant iterations with test Function (85) (see Table 4). ( Left : x 1 A = 0.416 is the root of y 0 ( x ) , then x 2 A = 0.442 is the root of y 1 ( x ) . Middle : x 3 A = 0.898 is the root of y 2 ( x ) and x 4 A = 0.728 is the root of y 3 ( x ) . Right : x 5 A = 0.7387 is the root of y 4 ( x ) and x 6 A = 0.739086 is the root of y 5 ( x ) ).
Figure 4. Classic secant iterations with test Function (85) (see Table 4). ( Left : x 1 A = 0.416 is the root of y 0 ( x ) , then x 2 A = 0.442 is the root of y 1 ( x ) . Middle : x 3 A = 0.898 is the root of y 2 ( x ) and x 4 A = 0.728 is the root of y 3 ( x ) . Right : x 5 A = 0.7387 is the root of y 4 ( x ) and x 6 A = 0.739086 is the root of y 5 ( x ) ).
Appliedmath 05 00015 g004
Figure 5. T-Secant iterations with test Function (85) (see Table 5) ( Left : x 1 A = 0.416 is the root of y 0 ( x ) , x 1 B = 0.915 is the root of z 0 ( x ) , then x 2 A = 0.667 is the root of y 1 ( x ) , x 2 B = 0.764 is the root of z 1 ( x ) . Right : x 3 A = 0.7387 is the root of y 2 ( x ) , x 3 B = 0.7391 is the root of z 2 ( x ) ).
Figure 5. T-Secant iterations with test Function (85) (see Table 5) ( Left : x 1 A = 0.416 is the root of y 0 ( x ) , x 1 B = 0.915 is the root of z 0 ( x ) , then x 2 A = 0.667 is the root of y 1 ( x ) , x 2 B = 0.764 is the root of z 1 ( x ) . Right : x 3 A = 0.7387 is the root of y 2 ( x ) , x 3 B = 0.7391 is the root of z 2 ( x ) ).
Appliedmath 05 00015 g005
Figure 6. Definition of the distance between system responses (left) and variations in parameters through iterations (right).
Figure 6. Definition of the distance between system responses (left) and variations in parameters through iterations (right).
Appliedmath 05 00015 g006
Figure 7. Observed and simulated system response pairs after “S” function value evaluations.
Figure 7. Observed and simulated system response pairs after “S” function value evaluations.
Appliedmath 05 00015 g007
Figure 8. Local optimal solutions.
Figure 8. Local optimal solutions.
Appliedmath 05 00015 g008
Table 1. Linearization method’s basic equations (single- and multi-variable cases).
Table 1. Linearization method’s basic equations (single- and multi-variable cases).
Single-Variable m = n = 1 Multi-Variable m n > 1 Equations
1 x p = x p B x p A X p = x 1 , 1 , p x n , 1 , p x 1 , n , p x n , n , p (37)
2 f p = f p B f p A F p = f p , 1 , 1 f p , n , 1 f p , 1 , m f p , n , m (36)
3 S p = f p x p S p = F p X p 1 (21) and (35)
4 x p + 1 A = x p A f p A S p x p + 1 A = x p A S p + f p A (24) and (40)
5 S p + 1 x p A = f p + 1 A f p A S p + 1 x p A = f p + 1 A f p A (27) and (45)
Table 2. Basic equations of classic and suggested multi-variable update methods.
Table 2. Basic equations of classic and suggested multi-variable update methods.
Classic Secant MethodSuggested Update MethodEquations
1 x k , p B = x p A + x k , p d k (64)
2 X p = x k , p X p = diag x k , p (37) and (65)
3 F p = f k , p B f p A F p = f k , p B f p A (36)
4 T p X = diag t i , p X = diag x i , p + 1 x i , p A (67)
5 T p F = diag t j , p F = diag f j , p + 1 A f j , p A (66)
6 S p = F p X p 1 S Tp = T p F S p T p X 1 = t j , p F f k , j , p t i , p X x i , p (35) and (69)
7 S p x p A = f p A S Tp x p A = f p A (43) and (68)
8 x p A = S p + f p A x p A = S Tp + f p A (42) and (72)
9 x i , p + 1 A = x i , p A j = 1 m S i , j , p + f j , p A x i , p + 1 B = x i , p + 1 A x i , p A 2 j = 1 m S i , j , p + f j , p A t j , p F (41) and (73)
10 x i , p + 1 B = x i , p B μ i , p = j = 1 m S i , j , p + f j , p A j = 1 m S i , j , p + f j , p A t j , p F (74)
11 x i , p + 1 = x i , p + 1 B x i , p + 1 A x i , p + 1 = μ i , p x i , p A (75)
Table 3. T-Secant iteration with constant Jacobian approximate S p (Equation (77)).
Table 3. T-Secant iteration with constant Jacobian approximate S p (Equation (77)).
h x p + h A x p + h B Δ x p + h A t p + h f Δ x p + 1 + h e p + 1 + h A α p + h T N f L
01.7002.0000.421−0.0850.036 2.6 × 10 2 21.23
12.1212.156−0.036−0.359−0.013 2.2 × 10 2 40.87
22.0852.0720.013−0.3420.0044 7.6 × 10 3 1.0860.76
32.0982.102−0.0044−0.348−0.0015 2.7 × 10 3 0.9780.70
42.0932.0920.0015−0.3460.00053 9.2 × 10 4 1.01100.67
52.09492.0955−0.00053−0.347−0.00018 3.2 × 10 4 0.997120.65
62.09442.09420.00018−0.3460.000063 1.1 × 10 4 1.001140.63
72.09462.0947−0.000063−0.3460.000022 3.8 × 10 5 0.9997160.62
Table 4. Classic secant iterations with test Function (85) (see Figure 4).
Table 4. Classic secant iterations with test Function (85) (see Figure 4).
p 0123456
x p A 2.000 2.000 0.416 0.442 0.898 0.7279 0.7387
x p 4.000 2.416 0.858 0.456 0.170 0.011 3.6 × 10 4
f p A 1.584 2.416 1.331 0.462 0.275 0.019 6.1 × 10 4
x p + 1 A 0.416 0.442 0.898 0.728 0.7387 0.739086 0.739085
e p + 1 A 1.155 0.297 0.159 0.011 0.0004 9.1 × 10 7 7.3 × 10 11
α S 0.113 15.5 0.459 4.249 1.292
f p + 1 A 1.331 0.462 0.275 0.019 6.1 × 10 4 1.5 × 10 6 1.2 × 10 10
Table 5. T-Secant iterations with test Function (85) (see Figure 5).
Table 5. T-Secant iterations with test Function (85) (see Figure 5).
p 01234
x p A 2.000 0.416 0.6668 0.7387 0.7390851328
x p 4.000 1.331 0.0968 4.1 × 10 4 3.9 × 10 10
f p A 1.584 1.331 0.1190 6.7 × 10 4 6.5 × 10 10
x p + 1 A 0.416 0.667 0.7387 0.7390851328 0.7390851332
e p + 1 A 1.155 0.072 4.0 × 10 4 3.8 × 10 10
α TS 3.210 1.874 2.668
f p + 1 A 1.331 0.119 6.7 × 10 4 6.5 × 10 10
t p F 0.840 0.089 0.0057 9.6 × 10 7
x p + 1 B 0.915 0.764 0.7391 0.7390851332
Table 6. Efficiencies of classic and improved algorithms.
Table 6. Efficiencies of classic and improved algorithms.
Method d β EFF  [15] * EFF  [14] L max  [36]
1Secant11.618…1.618…1.618…4.0
2Newton22.01.01.414…3.0
3T-Secant2 2.618…   1.309…   1.618…   4.5 
4TS-const. S p 2 1.0   0.5   1.0   0.6 
5T-Newton33.01.01.442…3.0
6Chen [28]31.618…0.539…1.173…
7Wang [30]51.618…0.323…1.101…
Table 7. Calculated values of the mean convergence rates.
Table 7. Calculated values of the mean convergence rates.
N Method R x 0 A R x p max A p max N f L L N
12Broyden 1. [36]4.92 4.7 E 10 -590.3910.78
22Broyden 2. [36]4.92 2.6 E 10 -390.6071.22
32Powell [41]4.92 7.0 E 10 -1510.1500.30
42ACD [42]130.1 1.0 E 10 -3250.0860.17
52Nelder–Mead [43]2.00 1.4 E 10 -1850.1270.25
62T-secant [36,41]4.92 1.0 E 25 396.57313.15
72T-secant [42]130.1 1.0 E 25 396.93713.87
82T-secant [43]2.00 6.7 E 15 265.55611.11
93T-secant72.72 1.4 E 14 5201.8095.43
103 32.47 1.0 E 25 4163.81511.45
115 93.53 1.3 E 14 8480.7603.80
125 7.19 5.9 E 14 4241.3516.76
1310 202.6 1.0 E 25 141540.4084.08
14200 92.78 9.0 E 15 1020100.0428.44
151000 212.4 3.6 E 13 660060.0065.66
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Berzi, P. A Simplified Algorithm for a Full-Rank Update Quasi-Newton Method. AppliedMath 2025, 5, 15. https://doi.org/10.3390/appliedmath5010015

AMA Style

Berzi P. A Simplified Algorithm for a Full-Rank Update Quasi-Newton Method. AppliedMath. 2025; 5(1):15. https://doi.org/10.3390/appliedmath5010015

Chicago/Turabian Style

Berzi, Peter. 2025. "A Simplified Algorithm for a Full-Rank Update Quasi-Newton Method" AppliedMath 5, no. 1: 15. https://doi.org/10.3390/appliedmath5010015

APA Style

Berzi, P. (2025). A Simplified Algorithm for a Full-Rank Update Quasi-Newton Method. AppliedMath, 5(1), 15. https://doi.org/10.3390/appliedmath5010015

Article Metrics

Back to TopTop