1. Introductory Notes
Consider a set of square algebraic nonlinear problems as [
1,
2]:
wherein
and
,
are coordinate functions. Assume that
is an enough differentiable function of
within a convex open set denoted by
. Now we first revisit some pioneering iterative methods to address (
1). Newton’s method (NM), widely used for such problems, is formulated as follows [
3]:
This approach achieves second-order convergence, provided that the starting vector
is adequately near to the true root
. To overcome certain constraints associated with the NM, alternative solvers [
4,
5] have been developed, including Steffensen’s method (SM), which operates without the need for derivative calculations [
6]. The formulation of the SM for addressing nonlinear systems is outlined in [
7]:
which utilizes the divided difference operator (DDO). The first-order DDO of
H for high-dimensional knots
and
is defined via (
):
More generally, the DDO for
H on
can be defined as [
7]:
that satisfies
Using
, the first-order DDO can also be given by [
8]:
However the definitions (
4)–(
6) mainly yield in dense matrices for the representation of the DDO, which restrict the applicability of the SM for tackling (
1) to some extent.
The author in [
9] provided another procedure to compute the DDO in a similar way as follows:
with
. Traub (TM) introduced an improvement to NM with local cubic convergence [
9]:
Another well-known and effective method for resolving (
1) is the fourth-order Jarratt technique (JM) [
10,
11], which is described as:
An improvement of (
3) was furnished in [
12,
13] as:
where
The distinction in (
10), as opposed to (
3), lies in its utilization of two iterations and consequently two
-D function assessments, enabling it to achieve a rate superior to quadratic. The core concept is to stabilize the DDO during each cycle and subsequently augment the substeps to maximize order enhancement and contribute to the computational efficiency index (CEI) of the solver.
This manuscript is motivated by the objective of creating a multi-step fast iterative solver that improves both accuracy and efficiency in addressing nonlinear equation sets via (
10). By eliminating the reliance on Fréchet derivatives, our proposed method based on an extension over TM seeks to alleviate the complexities and computational demands associated with derivative-involved approaches, thereby advancing the domain of this field. The primary objective is to furnish a higher-order derivative-free Steffensen-type solver capable of addressing nonlinear systems, encompassing both complex and real solutions. Our intention is to enhance computational efficiency by reducing the frequency of matrix inversions and functional evaluations, in accordance with the principles of numerical analysis. In this work, functional evaluations refer to both function and derivative evaluations, which differ from the concept typically used in the Calculus of Variations.
This article is constructed as comes next.
Section 2 explores the memorization technique utilized in the Steffensen-type scheme for addressing nonlinear sets of equations.
Section 3 formulates a multistep approach comprising several substeps to achieve rapid convergence while minimizing the number of LU decompositions using a with memory structure to accelerate the convergence as much as possible.
Section 4 presents an error analysis and assesses the rate of convergence. Subsequently,
Section 5 examines the CEI of various methods, focusing on the number of functional assessments and the flops-type index. Furthermore,
Section 6 demonstrates the applicability and advantages of the proposed method with memory through its application to several problems. Finally,
Section 7 offers concluding remarks.
2. With Memorization of the Iterative Methods
In this context, our objective is to enhance the CEI of (
10) without adding additional steps or further DDOs for every iterate; see [
14]. To achieve this, we utilize the concept of memory-based methods, which suggest that the speed of convergence and overall efficiency of iterative techniques can be enhanced by retaining and utilizing previously calculated function values.
Noting that the error equation for (
10) (the notations here will be pointed out further in
Section 3):
contains a term expressed as follows:
Here the non-zero scalar
in (
13) significantly influences both the convergence domain and the enhancement of the convergence speed. When addressing a nonlinear set of problems, and since
is unknown, one could approximate
to bring the entire relation in (
13) close to zero. Thus, we can express this as
where
represents an estimation of the root (for each iteration).
It is crucial to elaborate on how one estimates the matrix
,
by utilizing certain approximations of
derived from the available data [
15].
To enhance the performance of (
10) through the principle of memory-based methods [
13], we take into account the following iterative expression without memory shown as PM1:
This solver reads the following error equation:
Without loss of generality, we focus on
the scalar case to analyze the dynamical behavior of the iterative methods in the complex plane, rather than extending to the multi-dimensional case for (
19). Visualizing the fractal attraction basins of iterative methods for polynomial equations of various degrees in the complex plane is critical for several reasons [
16,
17], particularly when shading the plot based on the number of iterations required for convergence. In this context, different polynomial roots correspond to distinct regions of attraction. By mapping these basins, one can illustrate where initial guesses converge to specific roots. This step is essential in our work, demonstrating how the use of memory and small free parameter values can expand the convergence radii, thereby enlarging the region for selecting initial approximations.
Shading the plot based on the number of iterations needed for convergence offers insights into the solver’s effectiveness. Regions where the method converges rapidly can be identified as more stable or efficient, whereas areas requiring more iterations (or failing to converge) suggest potential inefficiencies or instability. Such analyses are illustrated in
Figure 1,
Figure 2,
Figure 3 and
Figure 4 over the domain
, with a maximum iteration count of 150 and a tolerance of
for the residual as the stopping criterion. They reveal that the higher the order is for Steffensen-type methods, the larger the attraction basin is. Note that PM1 and PM2 both are Steffensen-type methods. In these new methods, the number of iterations to get the root is lower than SM; due to this, they have lighter and fewer shaded areas in their attraction basins. Moreover, the convergence radii could be enlarged by selecting small values for the free non-zero scalar
. Thus, memorization will not only contribute to a higher-efficiency index but also to larger attraction basins, which means higher stability for such a solver in contrast to the Steffensen-type method without memory.
Figure 1,
Figure 2,
Figure 3 and
Figure 4 also reveal that by observing how the number of iterations varies across the complex plane, one can assess the convergence properties of any methods. The fractal boundaries highlight regions where small changes in the initial guess can yield drastically different outcomes (i.e., converging to different roots or diverging). This sensitivity is crucial to understand, especially when implementing these methods in practical applications where precision of the initial guess might be limited. They are used later in the paper by implying to select a small value for the free nonzero parameter (carried forward the relation (
37)) and how memorization can enhance the convergence domain.
3. Derivation of the Scheme
To facilitate the implementation of the memory-based scheme (
15), we will first examine
and
Consequently, we now present the subsequent scheme (PM2) as our main contribution, (
,
,
):
The error equation of this solver with memory will be given in the next section.
It is well known [
15] when
represents a convex nonempty domain. Then, assume that
H is three-times Fréchet smooth over
D, and
for any distinct points
(where
). Additionally, let the starting vector
and the zero
be in close proximity to each other. By defining
and setting
, finally we can derive the equation below
The relation (
20) will be used later in
Section 4 of this work.
To implement (
19), it is essential to resolve a series of linear algebraic sets of equations. This entails performing a new LU factorization at each iteration, without leveraging any information from preceding steps. Nevertheless, a substantial body of the literature exists regarding the recycling of such information to derive updated preconditioners for iterative solvers [
18]. The advantage of (
19) lies in the fact that all linear systems share the identical coefficient matrix. Consequently, a single LU factorization suffices; by retaining this decomposition, it can be applied to several distinct right-hand side parts to obtain the resolution vectors for each sub-cycle of (
19).
The solution of the nonlinear equation systems that we consider here are in
as stated in
Section 1. The roots that we are seeking for are assumed to be simple zeros. Both real and complex roots can be obtained by the discussed methods (if existed). In fact, by choosing a suitable complex initial guess, a complex root (if existed), can be obtained.
4. Convergence Order
Here, we furnish a theoretical analysis of the convergence speed of the iterative scheme presented in (
19). Before introducing the main contribution, we represent the
-dimensional Taylor expansion.
The rate at which the iteration without memory PM1 converges is determined via
-dimensional Taylor expansions. Let
denote the error at the
-th iterate. As noted in [
19]:
this error equation implies
G is a
p-linear functional, where
and
p is the speed. Additionally, we have:
Assume that
is sufficiently differentiable in the Fréchet sense in
D. Following [
10], the
-th derivative of
H at
,
, is the
-linear functional, i.e.,
so that
. For
located in a vicinity of the solution
of (
1), the Taylor expansion can be formulated as [
10]:
where
. It follows that
, as
and
. Additionally, for
, we have:
wherein
I shows the unit matrix. Additionally,
.
Theorem 1. Let in (1) be adequately Fréchet differentiable at every point in D and that at . Additionally, let be continuous and nonsingular at θ. Next, produced by (19) with memory with a selection of an appropriate starting value has 4.23607 R-convergence order. Proof. For proving the convergence speed, we consider (
24) and (
25) to write
In the context of the scheme (
15), when operating without memory (i.e., PM1) and applying (
24)–(
26), we ultimately derive
Now by considering the with memorization in (
19), we shall express (
27) in its asymptotic form as follows:
where ∼ shows for the error equation without the asymptotical term. A variety of symbolic computations conducted by considering that the coefficients of the error terms in our
-D scenario are all matrices, and applying (
20), along with the understanding that multiplication does not allow for commutativity, we can deduce
Consequently, one arrives at
By imposing (
29) and (
30) into (
28), we obtain:
This demonstrates that
, where the convergence
R-order is expressed as
The proof is concluded. □
Before ending this section, it is pointed out that with a simple change the structure of PM2, it is possible to provide another iteration scheme with memory of a similar kind as follows (PM3), (this time:
,
):
If we aim to further improve the convergence order, two possible approaches can be considered. First, we could construct a solver by incorporating additional subsets and then apply the memorization procedure. Second, one might explore faster methods to accelerate the free parameter using alternative types of interpolation, which, however, falls beyond the scope of this paper.
5. Computational Efficiency Comparisons
For the proposed solver with memory, PM2 (or equivalently PM3), only a single LU factorization is required, followed by matrix-vector multiplications, which enhances computational efficiency by eliminating the need to compute multiple matrix inverses in each iteration. The CEI for iterative solvers is defined as follows [
7]:
where
represents the total computational cost and
p signifies the convergence speed, considering the quantity of functional evaluations. The cost of computing each scalar function is considered a unit, while the costs associated with other calculations are expressed as multiples of this unit cost. To evaluate the CEI for PM2, we first outline the number of functional evaluations (cost) required for
-dimensional functions, as detailed below (excluding the indicator
):
In : evaluations.
In the DDO: evaluations.
Furthermore, we consider the costs to solve two triangular systems, all quantified in floating-point operations (flops). The flops necessary for executing the LU procedure amount to
, while resolving the two related triangular systems requires approximately
flops. Noting that, here, we assume that the cost for one functional evaluation is roughly equal to the cost for one flop. The findings displayed in
Table 1 demonstrate that for varying
, the CEI of our method surpasses that of competing approaches. Comparisons of different derivative-free Steffensen-type solvers with and without memory based on various choices of
are illustrated in
Figure 5 and
Figure 6, confirming the superiority and improvement in both classic and flops-type efficiency indices of PM relative to its main competitors.
In real-world applications, there is a trade-off between eliminating Fréchet-derivatives and increasing the method’s overall computational complexity. Generally speaking, it relies on the specific problem, which leads to a nonlinear system of equations. However, by eliminating the Fréchet derivative, we make the solver derivative-free, which is suitable for problems, in which the derivative is not available. Moreover, the concept of memorization can be accompanied in methods without Fréchet derivatives to increase the computational complexity.
It is also remarked that in the absence of Fréchet derivatives, one might ask that what specific mechanisms ensure that the proposed Steffensen-type technique remains robust over a variety of problem sets? To tackle this, it is stated that in the absence of Fréchet derivatives, the convergence radius mainly reduced tremendously, which is one weak point of Steffensen-type techniques at first sight. However, this can simply be improved by choosing very small values for the free non-zero parameter (as seen in the attraction basins of
Section 2), as well as imposing the with memorization concept.
6. Numerical Aspects
The objective of this section is to facilitate the implementation of our proposed scheme, PM2. The computations were executed using Mathematica 13.3 [
20] in standard machine precision to handle the round-off errors. The linear systems were resolved employing LU decomposition via
. All computational examples were conducted in a uniform environment. We adopt the stopping criterion as follows:
A significant challenge in implementing (
19) involves the integration of
, which is no longer constant and must be characterized as a matrix. Here,
is delineated follows:
In fact due to
Figure 1,
Figure 2,
Figure 3 and
Figure 4 and the results discussed previously in [
13], other small values for the free parameter can also be used which can yield in different choices such as
or
or
In fact, for implementation of method with memory, here, for the first iterate, we start with the without memory version of the method and after that, from the second iterate, the information of the previous step can be used to update the parameters and thus with memorization of the scheme can be done. The selection of
has a direct impact on the entire process, influencing the speed at which convergence is achieved. Here, (
37) aligns with the dynamic investigations of Steffensen-type solvers with memory, where larger basins of attraction occur when the free parameter is near zero.
To validate the analytical convergence rate in the computational experiments, we determine the numeric speed of convergence using [
7]
Example 1 ([
14]).
We investigate a nonlinear system , which possesses a complex root, as detailed belowin which the precise solution is presented up to eight floating points as The computational evidence and the numerical speed, denoted as
, are detailed in
Table 2. We utilized 1000 fixed floating points, with the initial value set as
. The choice of this is based on
[
14]. Rather than
and (
37), no other parameters should be chosen and the method works based on satisfying (
35). Additionally, the residual norm is expressed using the
notation. The numerical pieces of evidence seen in
Table 2 support the observations in
Figure 1,
Figure 2,
Figure 3 and
Figure 4, discussing that the smaller the choice of the free parameter for PM2 would results in arriving at the convergence phase faster than consider larger values for this parameter.
Example 2 ([
21]).
In this test, we investigate the nonlinear systems extracted through the computational resolution of the following Partial Differential Equation (PDE)where the coefficient of diffusion is . If we take into account , then the computational resolution is represented by:at the grid points on a equidistant mesh. Let and show the number of steps in spatial and temporal variables, respectively. The parameters are set as follows [21]: , , , and . The parameters have been chosen so as to obtain a unique and well-defined numerical solution and stay away from stiffness or irregularity for the solution of PDEs. To address this, we may utilize the first-order backward FD method for the first differentiation in temporal
:
wherein
k denotes the time step size. For the spatial terms of the PDE, we apply the second-order central difference FD method as follows:
and
where
h represents the spatial step size along
z. Following this discretization and imposing the boundary conditions will lead a to a set of nonlinear equations that must be tackled iteratively.
The numerical solutions using PM2 are illustrated in
Figure 7, and
Table 3 provides the comparative evidence for this case in double precision when the residual of the numerical obtained solution is less than
. We set
, i.e., 21 equally spaced nodes in both space and time direction and after imposing the initial and boundary conditions, we derive to a set of nonlinear system of the dimension
, where the initial vector
. Along space, we have used central three-point second-order FD approximations, and along time, we have employed first order forward FD approximations to discretize the problem.
For an alternative set of parameters, specifically
, representing 31 equally spaced points in both the spatial and temporal directions, we obtained a nonlinear system of dimension
after applying the initial and boundary conditions. The initial vector is again defined as
.
Figure 8 presents the numerical simulations for this setup, further demonstrating the effectiveness of PM2 and the concept of memorization.
The numerical performance of the proposed solver is demonstrated through root-finding for various nonlinear equations. Results from the numerical tests clearly indicate that the solver achieves convergence in fewer iterations and with higher accuracy per iteration.
We finish this section by pointing out the following matter. Another concern may arise regarding how the method handles convergence issues in nonlinear systems with multiple solutions or when the initial guess significantly influences convergence behavior. To address this, the focus of this article is on simple roots, not multiple roots; while the methods can be applied to find multiple zeros, their orders of convergence drop significantly in such cases. In fact, if a nonlinear system has multiple roots (whether the multiplicity is known or unknown), specialized solvers designed for multiple roots, along with appropriate initial guesses, must be developed to maximize efficiency.