1. Introductory Notes
The matrix sign function (MSF) is a fundamental concept in numerical linear algebra [
1], with widespread applications in solving Lyapunov equations, algebraic Riccati equations, and computing spectral decompositions [
2]. Given a square matrix
without purely imaginary eigenvalues, the MSF, denoted as
, is formally defined via its Jordan form representation [
3]:
where
is the Jordan decomposition of
A,
is the identity matrix of the dimension
,
is the identity matrix of the dimension
, and the block diagonal matrix
contains eigenvalues partitioned into those with positive and negative real parts. In addition, the MSF satisfies the nonlinear matrix equation [
4]
which provides a fundamental framework for designing iterative methods,
I is the identity matrix of appropriate dimension and
.
Since its appearance, the MSF has been employed across various domains of scientific computing [
5], graph theory, control theory [
6], and linear algebra algorithms [
7]. In the realm of differential equations, particularly those modeling physical phenomena, the MSF assists in decomposing matrices. This decomposition simplifies the solution process, making it easier to analyze and solve complex systems of differential equations.
The numerical computation of
has been extensively studied, leading to various iterative techniques, including Newton’s scheme, the Newton–Schultz local solver, Halley’s scheme, and the Padé family of iterations [
8]. While classical schemes such as Newton’s iteration
exhibit quadratic convergence, recent research has focused on developing higher-order methods to achieve superior efficiency. This paper proposes a novel iterative approach based on high-order rational approximations, improving the Padé iterations [
9].
The iterative computation of the MSF has undergone significant advancements, particularly with the development of rational approximations and multipoint iteration strategies. The foundational Newton iteration provides a straightforward approach with global convergence for matrices without imaginary-axis eigenvalues [
4]. However, its quadratic convergence rate limits its efficiency for large-scale applications. The related numerical approaches are discussed in [
10,
11].
To accelerate convergence, Padé approximations have been incorporated into iterative schemes. Kenney and Laub [
12] introduced Padé iterations, which utilize rational approximations of the function
to construct the following recurrence relation:
where
and
denote the
-Padé approximants. The choice of parameters influences convergence behavior, with specific cases such as
and
exhibiting global convergence.
Beyond the Padé iterations, high-order iterative methods have been explored to further enhance computational efficiency. Methods derived from root-finding schemes in scalar nonlinear equations, such as Halley’s method,
demonstrate cubic convergence.
A key advancement in this direction involves designing high-order iterations using conformal mappings and stability-enhancing transformations. For instance, the recent work of Jung and Chun [
13] proposes a three-point iterative framework where weight functions are tuned to optimize the stability region and convergence order. Their approach constructs an iteration scheme conformally equivalent to an eighth-order polynomial, effectively balancing convergence speed and numerical stability.
Despite these advancements, further improvements are needed to construct iterative methods with both high-order convergence and computational robustness [
14]. Motivated by these challenges [
15], this paper introduces a sixth-order iterative method based on refined rational approximations and weight functions. Unlike classical methods such as Newton’s iteration, which exhibits quadratic convergence, the proposed iterative scheme achieves a higher convergence rate of order six, leading to improved computational efficiency. The methodology is constructed by extending a high-order nonlinear solver for simple roots to the matrix setting, providing a systematic framework for deriving efficient iterative schemes for computing the MSF. The global convergence behavior of the scheme is further supported by an analysis of attraction basins, offering insights into its applicability across a broad spectrum of matrices. These contributions collectively distinguish the approach from existing methods and establish its effectiveness in computing the MSF.
The structure of the remainder of this study is organized as follows.
Section 2 explores the benefits of a high-order scheme and introduces a novel approach for solving scalar nonlinear equations, which is subsequently generalized to the matrix setting. An analytical investigation establishes the sixth-order convergence of the proposed solver. Next, in
Section 3, we investigate the global convergence behavior of the proposed method using regions of attraction basins, alongside an assessment of its stability properties.
Section 4 presents numerical experiments that substantiate the theoretical findings and further illustrate the effectiveness of the proposed methodology. Finally,
Section 5 offers a comprehensive conclusion, summarizing the key contributions of this work.
2. An Accelerated Sixth-Order Procedure
We now turn our attention to the scalar analogue of Equation (
2), which is given by
In this context,
corresponds to the scalar version of (
2), whose solutions are
.
To enhance most classical approaches, such as (
3)–(
5), (for more details, see [
16]), we propose a refined multi-step iteration scheme based on rational approximations and weight functions, formulated as follows:
The structure in (
6) has been designed intentionally for two reasons. The first one is to obtain a new method (which does not belong to the Padé family of methods). The second one is efficient in computing the MSF with global convergence behavior, as will be discussed later in this work.
Theorem 1. Let be a simple root of the adequately smooth function . If the starting approximation is chosen to be sufficiently close to θ, then the procedure (6) tends to θ with sixth-order accuracy. Proof. Let
be a simple zero of the function
f. Given that
f possesses sufficient smoothness, we perform a Taylor series expansion of
and its derivative
around
, yielding the following expressions:
as well as
Here, the notation
denotes the error term, and the coefficients
are given by
By utilizing Equations (
7) and (
8), we derive the following expression:
Substituting Equation (
9) into
as defined in (
6), we obtain
Utilizing Equations (
7)–(
10), we obtain
By substituting (
11) into Equation (
6) and performing Taylor expansion with some simplifications, we deduce that
Now, by performing a Taylor expansion of
about
and incorporating Equation (
6), we arrive at
Finally, using Equations (
7) and (
13), we obtain the refined expression
Using (
14) we can determine that (
6) describes the following final equation for the errors:
This completes the proof by showing the sixth order of convergence for the error equation. □
The iterative scheme outlined in (
6) can now be applied to solve (
2). Proceeding with this approach leads to
having the starting matrix
In a similar manner, the reciprocal formulation corresponding to Equation (
16) is obtained by employing the following systematic approach:
The computational efficiency of iterative methods for determining the MSF depends largely on the order of convergence. Classical methods such as Newton’s iteration (
3) exhibit quadratic convergence, which, while sufficient for moderate-sized problems, becomes computationally expensive for large-scale complex matrices. High-order methods are designed to achieve faster convergence rates, reducing the number of iterations needed to achieve a given accuracy. This is especially useful when dealing with complex matrices, where each iteration involves matrix–matrix operations such as multiplications and inversions. By employing high-order iterations, the total computational cost can be reduced, making the method more practical for applications in numerical linear algebra.
Another crucial advantage of high-order methods is their enhanced numerical stability. Lower-order methods, such as Newton’s iteration, often require stabilization techniques such as scaling and squaring to maintain numerical accuracy, especially for matrices with eigenvalues close to the imaginary axis. High-order methods naturally mitigate these issues by achieving rapid convergence within fewer iterations, thereby reducing error accumulation and improving robustness. Moreover, the flexibility of high-order schemes allows for the incorporation of adaptive techniques, such as dynamically selecting iteration parameters based on the spectral properties of the matrix. This adaptability makes high-order methods particularly attractive for solving ill-conditioned problems and ensures reliable performance across a broad spectrum of computational tasks.
Theorem 2. Assume serves as a suitable initial guess and A is an invertible matrix. Under these assumptions, the iterative scheme described by (18) (or equivalently (16)) converges to M, achieving a sixth rate of convergence. Proof. It is recalled that the decomposition of the matrix
A is carried out using an invertible matrix
Z of identical dimensions, in conjunction with the Jordan block matrix
J. This leads to the following factorization:
By leveraging this decomposition and conducting a meticulous structural analysis of the solver, an iterative scheme is formulated for calculating the eigenvalues (
). This iterative process transitions from iteration
k to the subsequent iteration
as follows:
where
From a theoretical standpoint, and upon performing appropriate simplifications, the iterative scheme delineated in (
20) reveals that the eigenvalues asymptotically converge toward the limiting values
. More precisely, this convergence behavior is mathematically characterized by the following expression:
Equation (
22) encapsulates the asymptotic tendency of the eigenvalues to cluster around
as the iterative process advances. With each successive iteration, the eigenvalues exhibit an increasingly tight convergence toward these limiting values. Having established the theoretical foundation for the method’s convergence, we now shift our focus toward analyzing its rate of convergence. To facilitate this investigation, we proceed as follows:
Utilizing Equation (
23) and recognizing that
represents a rational function of
A, while simultaneously ensuring that
maintains commutativity with
M in the same manner as
A, the following expression can be formulated
Using (
24) and 2-norm, it is possible to determine that:
This finding underscores that the iterative scheme attains a convergence rate of order six, contingent upon the selection of a suitably chosen initial matrix, such as the one specified in (
17), as the starting point. □
Compared to existing rational approximation-based methods, our approach integrates a high-order nonlinear solver into the iterative framework, enhancing both stability and convergence properties. Moreover, while lower-order methods often require stabilization techniques such as scaling and squaring to handle matrices with eigenvalues near the imaginary axis, the proposed method naturally mitigates such issues due to its rapid convergence and inherent numerical robustness.
3. Attraction Basins and Stability
Understanding the convergence behavior of iterative methods is crucial when computing the MSF. One effective way to visualize and analyze this behavior is through attraction basins, which depict the regions in the complex plane where different initial guesses lead to convergence to particular solutions. Attraction basins provide insights into the stability, efficiency, and robustness of numerical methods [
17].
Iterative methods for computing the MSF rely on successive approximations, where the choice of the starting value significantly affects the convergence trajectory. By plotting attraction basins, we can carry out the following steps:
Identify regions where the method converges rapidly.
Detect unstable zones where divergence or slow convergence occurs.
Compare different iterative schemes in terms of their efficiency.
Numerical stability is a crucial factor when selecting an iterative method. As analyzed in [
18], attraction basin plots provide valuable insights into the robustness of a method against perturbations in the initial guess, the presence of fractal-like structures that indicate chaotic behavior in iterative dynamics, and the impact of scaling techniques on improving convergence regions.
A critical advantage of drawing attraction basins is their role in benchmarking different iterative methods. Methods with larger, well-connected basins typically offer superior convergence properties, while those with fragmented or irregular basins may suffer from instability.
Drawing attraction basins is a powerful tool for analyzing iterative methods that can be used to compute the MSF. These visualizations help assess convergence behavior, stability, and efficiency, guiding the selection and refinement of high-order iterative schemes. As we develop new iterative methods, attraction basin analysis remains essential in ensuring their practical effectiveness.
In this work, the techniques presented in (
16) and (
18) have been introduced with the specific aim of enhancing the attraction regions associated with such schemes in the context of solving the equation
. To provide a more thorough understanding, we proceed by investigating how the presented solvers show global convergence properties and demonstrate improved radii of convergence. This is achieved by illustrating their corresponding attraction regions within the area:
while resolving
.
For this purpose, the complex plane is discretized into a grid of nodes, and the behavior of each point is evaluated by using it as an initial value. This allows us to determine whether the iteration converges or diverges. In cases of convergence, the points are shaded based upon the quantity of iterates they undergo, with the convergence condition being satisfied when
Figure 1 depicts the basins of attraction for (
16) and (
18). The results reveal that, for the methods (
16) and (
18), the convergence radii are large and global.
A similar spirit of logic as the one already discussed in [
19] indicates the stability of the scheme.
Theorem 3. Let A be an invertible matrix. Based on Equation (18), the sequence , initialized with , exhibits stability in the asymptotic sense. Proof. To initiate the analysis, we consider a perturbed evaluation of
at the
iteration within the framework of the numerical solver. Further elaboration on this perturbation approach can be found in [
20]. To systematically examine the effect of perturbations, the following iterative relation is formulated for each computational step:
At this stage, we take into account the assumption that
for all
, which is true under the framework of a first-order error analysis, provided that
remains sufficiently small. Based on this premise, and after performing a series of algebraic simplifications, the following inequality is derived:
Thus, the sequence
generated by the method in (
16) is stable. □
4. Numerical Examples
Here, we perform an assessment of the performance of the proposed iterative solvers by evaluating their effectiveness across a diverse range of problem types. The entire implementation process has been carried out using Wolfram (see [
21,
22]). A systematic approach has been adopted to address several computational aspects, including the precise detection of convergence. For the sake of clarity and coherence, the testing procedure is categorized into two distinct groups, following a methodology similar to that employed in [
23]: the first category comprises tests involving real derived values, while the second focuses on complex matrices.
The iterative methods considered in the comparative analysis include the method given in Equation (
3), referred to as NM2; the approach defined by (
5), labeled as HM3; the procedure outlined in (
16), indicated as PM61; the iteration specified in (
18), designated as PM62; and the Zaka Ullah et al. method of fourth order defined by ZUM4 [
14]:
For all Newton-type iterative schemes considered in this comparison, the initial matrix
is chosen in accordance with the specification given in (
17). The computational error at each iteration is evaluated using the following formulation:
where
represents the predefined convergence threshold, serving as the stopping criterion for the iterative process.
Example 1. A set of twelve randomly generated real matrices is obtained by employing the random seed command . Following their generation, the corresponding MSFs are computed and examined to facilitate a comparative analysis. These matrices are constructed within the numerical range and encompass dimensions varying from up to . All computations are carried out under .
Table 1 and
Table 2 showcase the numerical results corresponding to Example 1, providing substantial evidence of the efficacy of the methods introduced in this study. Of particular note, the method PM61 enhances computational efficiency by reducing the total number of iterates needed to calculate the MSF. This improvement is reflected in a marked decrease in the average CPU time, measured in seconds, across 12 randomly produced matrices of different sizes. This is implemented to compare the overall cost for computing of the proposed scheme with the existing ones.
Example 2. Within this numerical investigation, the MSF is computed for a set of eight randomly generated complex matrices. The evaluation is conducted while adhering to . The implementation of these random matrices is illustrated in the Mathematica 13.3 code snippet presented below
SeedRandom[789];
nu = 8;
Table[A[n1] = RandomComplex[{-100 - 100 I,
100 + 100 I}, {150 n1, 150 n1}];, {n1, nu}];
Table 3 and
Table 4 furnish computational comparisons for Example 2, reinforcing the efficacy of the furnished solver to determine the MSF for eight randomly produced complex matrices. Consistent numerical experiments conducted across a diverse set of related test cases further corroborate these findings. Among the evaluated methods, the PM61 algorithm exhibits superior efficiency and robustness, outperforming its counterparts in terms of computational accuracy and convergence behavior.
The numerical experiments conducted here provide substantial evidence in support of the superior performance of the proposed iterative methods, particularly PM61, in computing the MSF. The comparative assessment across multiple test cases demonstrates that PM61 exhibits a reduction in iteration count, as observed in
Table 1 and
Table 3. For instance, in Example 1, the average number of iterations needed for PM61 to achieve convergence is 8.08, which is markedly lower than NM2 (21.25), HM3 (13.66), and ZUM4 (9.75). This efficiency advantage remains consistent in Example 2, where PM61 requires an average of 8.50 iterations compared to NM2 (22.62) and HM3 (14.37). Such a reduction in iteration count is crucial when dealing with large-scale matrices, as it directly translates to fewer matrix–matrix operations, thereby minimizing computational complexity and memory usage.
Beyond the iteration count, another critical metric in evaluating iterative solvers is their CPU execution time, as shown in
Table 2 and
Table 4. The PM61 method consistently outperforms NM2, HM3, and ZUM4 in terms of computational speed. In Example 1, PM61 achieves an average execution time of 1.329 s, which is 29.7% faster than NM2 (1.889 s) and 11.2% faster than ZUM4 (1.481 s). Similar trends are observed in Example 2, where PM61 exhibits an average runtime of 4.456 s, making it significantly more efficient than NM2 (6.026 s) and HM3 (5.393 s). The improved efficiency of PM61 is attributed to its higher-order convergence rate, which reduces the number of matrix computations required per iteration. Furthermore, the method demonstrates superior scalability, handling larger matrices (e.g.,
) with competitive execution times compared to alternative schemes.
A key advantage of PM61 lies in its numerical stability and robustness, particularly when applied to real and complex matrices of varying dimensions. Unlike lower-order methods such as NM2, which may require additional stabilization techniques (e.g., scaling and squaring), PM61 converges reliably across different problem instances without exhibiting erratic behavior. This robustness is evident in the consistent performance metrics recorded in both Example 1 and Example 2. Moreover, the method maintains a balanced trade-off between iteration count and per-iteration computational cost, ensuring that the efficiency gains are realized across a broad spectrum of matrix sizes.
In summary, the numerical analysis underscores the effectiveness of PM61 in computing the MSF. The method excels in minimizing the number of iterations, reducing computational time, and enhancing numerical stability, making it a highly favorable choice for large-scale applications.