1. Introduction
For more than a century, great scientific attention has been attracted by the class of strictly diagonally dominant (SDD) matrices, as well as its numerous generalizations. The reason probably lies in the fact that the knowledge of this class has been successfully “translated” into various other areas of applied linear algebra. Let us mention only two such situations: the famous Geršgorin theorem [
1] is equivalent to the nonsingularity result for SDD matrices [
2], while an infinity norm estimation for the inverse of an SDD matrix can be obtained by Varah’s theorem [
3]. Both of them are simple and elegant, which makes them attractive from an application point of view. There are more interesting results concerning SDD matrices, like the following: the Schur complement of an SDD matrix is SDD itself [
4]; the error bound for linear complementarity problems can be easily calculated for the class of SDD matrices [
5]; regarding (nonlinear) dynamical systems, (local) stability is a direct consequence of the (Jacobian) matrix being an SDD matrix with negative diagonal entries, etc. Recently, more results related to the SDD class, also simple and elegant, have been obtained in the localization of the
-pseudospectrum of an arbitrary matrix [
6].
Thanks to such intensive and numerous benefits, many generalizations of the SDD class have been developed in various directions, all of which have been gathered together under one umbrella class of matrices, called H-matrices.
There are two equivalent definitions of H-matrices, also known in the literature under the name generalized diagonally dominant (GDD) matrices. But, first of all, let us recall two well-known classes, called strictly diagonally dominant (SDD) and M-matrices, both of which can serve as a starting point for the definition of GDD, that is, H-matrices.
A matrix
is called strictly diagonally dominant (SDD) if
where
A matrix
is called an M-matrix if it has the following sign pattern
it is nonsingular, and
(elementwise).
Definition 1. A matrix is called generalized diagonally dominant (GDD) if there exists a nonsingular diagonal matrix X, such that is an SDD matrix.
Definition 2. A matrix is called an H-matrix if its comparison matrix , defined withis an M-matrix. According to [
7], a matrix is an H-matrix if and only if it is a GDD matrix. This means that the class of H-matrices and the class of GDD matrices are the same.
However, checking if a matrix is an H-matrix is computationally very demanding, so a compromise solution is to find as many new subclasses of H-matrices as possible, described by practically usable, i.e., easily verifiable, conditions on matrix elements.
Subclasses that have been discovered up to now are numerous; let us mention just a few: doubly strictly diagonally dominant (DSDD), also known under the name Ostrowski matrices [
8,
9,
10], Dashnic–Zusmanovich matrices [
11], Dashnic–Zusmanovich-type matrices [
12], S-SDD or CKV matrices [
13,
14], CKV-type matrices [
15], Nekrasov matrices [
16,
17,
18], SDD
1 matrices [
19,
20], etc.
All of them have been widely used in applications: the localization of the spectrum of an arbitrary matrix, localization of the pseudospectrum of an arbitrary matrix, infinity norm estimations of the inverse, Schur complement properties, error bound for linear complementarity problems, etc. For an exhaustive list of references, see [
21].
However, like in (
2), all of these subclasses are defined by conditions depending on sums (or part of sums) of off-diagonal entries (which can be linked to the absolute vector norm, or the maximum matrix norm). What will happen if we choose to rely on the Euclidean vector norm, as well as the Frobenius matrix norm, instead? We will show that classes obtained in such a way remain inside the H-matrix class.
Vector norms that we will use throughout the paper are Euclidean
and maximum
norms,
defined for all
. As for matrix norms, we will mainly work with the Frobenius matrix norm, defined as
which is consistent with the Euclidean vector norm, rather than the induced Euclidean matrix norm, defined by
, where
denotes the spectral radius of
A. The induced maximum matrix norm is defined by
2. SDDF Class
Throughout the paper, we will split matrix
A as
If
is a matrix with nonzero diagonal entries, then
D is a nonsingular matrix, and the condition (
1) can be equivalently rewritten as
Motivated by this fact, we will define the following class.
Definition 3. For a given matrix with nonzero diagonal entries, we will say that it is an SDDF matrix ifwhere denotes the Frobenius matrix norm. As a start, we will prove the following theorem.
Theorem 1. Every SDDF matrix is an H-matrix.
Proof. Since
, it follows that
is a nonsingular matrix, so
is nonsingular, too. In order to prove that
A is an H-matrix, we will show that its comparison matrix
is an M-matrix. Let us show that
. Suppose, on the contrary, that there exists an eigenvalue
of
, such that
, then matrix
will satisfy condition (
3)
and consequently it will be nonsingular. But, this contradicts the fact that
is an eigenvalue of
. Hence,
, and the rest of the proof is similar to that in [
15], (proof of Theorem 6):
Hence, we conclude that is an M-matrix. □
Remark 1. At the very beginning, we emphasize the relationship between SDD and SDDF classes. The following three matricesshow that SDD and SDDF classes stand in a general position, meaning that no class is a subset of the other, and they have a nonempty intersection. This is illustrated by Figure 1. Indeed, is not an SDD matrix, because of the last two rows, while , so it is an SDDF matrix. Here and hereafter, we will assume that every matrix is split as , where is the diagonal part of . We have taken this particular example from a mathematical model representing energy flow in food webs, given by the generalized Lotka–Volterra equations. We will use it onwards as a good illustration of the possible benefits of new classes in real applications. The precise definition of the matrix based on empirical data is explained in [22], where, following the research by Moore and de Ruiter in [23] and references therein, authors considered a food web of n functional groups of living species with a pool of non-living organic matter, whose energy (as a common currency of the biomass, usually measured as the level of carbon or nitrogen) flow is approximately driven by the generalized Lotka–Volterra equations. On the other hand, and are chosen to be as simple as possible, just to verify the above Figure. Both matrices are SDD, while, for and .
The immediate consequences of Theorem 1 are the following two theorems.
Theorem 2. Let (D is the diagonal part of A). If λ is an eigenvalue of A, different from all diagonal entries , then, Proof. Suppose that there exists an eigenvalue , such that . Then, is an SDDF matrix, and hence is nonsingular, which is an obvious contradiction. □
Theorem 3. If A is an SDDF matrix, then Proof. From
, it follows that
is a nonsingular matrix, and
Finally, from
we have
□
Obviously, the upper bound given in (
5) can be treated as an upper bound for the Euclidean matrix norm (which we denote simply by
), since
3. -SDDF Class
In this section, we will recall only one special subclass of H-matrices (known under the name S-SDD class), and define the analogon of this class referring to the Frobenius norm.
For a given
S,
, matrix
A satisfying
where
, is called an
S-SDD matrix.
For the sake of transparency, here we will suppose that
and consequently
so that we can represent matrix
A as
where
D is the diagonal part of
A (as before), and the dimension of
is
k. Obviously, the relation with the previous splitting
, where
is the following:
Note that all diagonal entries of and are equal to zero.
Definition 4. A matrix with nonzero diagonal entries, satisfying the following two conditions,is called an S-SDDF matrix. Theorem 4. Every S-SDDF matrix is nonsingular; moreover, every one is an H-matrix.
Proof. Suppose, on the contrary, that there exists a vector
, such that
. Then, for all
it holds that
where
represents the Euclidean norm,
,
, and
Hence,
and, consequently,
From (
10) and (
11), it follows that
If
and
, this is a contradiction with (
9). If one of these vectors is a zero vector, say
, then
, and from (
10) we again obtain a contradiction, this time with (
8). Hence,
A can not be singular.
In order to prove that
A is an H-matrix, we will use, again, the comparison matrix
, and show that it is a nonsingular M-matrix. Again, with the splitting
, where
, we will show that
. Suppose, on the contrary, that there exists an eigenvalue
of
, such that
. Then matrix
will satisfy conditions (
8) and (
9), since:
and
Consequently, matrix
will be nonsingular, but this contradicts the fact that
is an eigenvalue of
. Hence,
, and
meaning that
is an M-matrix. □
Remark 2. Let us show that for an arbitrary , the SDDF class is a subset of the S-SDDF class. Suppose that A is an SDDF matrix, i.e., . Take an arbitrary subset S of indices (without loss of generality, suppose that , for some ). In order to prove that A is an S-SDDF matrix, we have to prove thatwhere If we denotethenso the first condition in (12) is obvious. The second condition, also holds true, because of the following simple algebraic inequalities:which, together with (13), give Remark 3. If one allows for a subset S to be equal to the whole set of indices N, i.e., if , condition (9) vanishes in this particular case, (8) becomes (3). In this sense, SDDF matrices could be considered S-SDDF matrices, for the special choice . Remark 4. Similarly as in the SDDF case, we can immediately conclude that S-SDDF and S-SDD classes (for the same subset S) stand in a general position.
3.1. Application 1: Eigenvalue Localization
Whenever a new subclass of H-matrices is defined, a corresponding new eigenvalue localization result can be formulated, just like the famous Geršgorin theorem [
1], which is the eigenvalue localization result corresponding to the SDD class. For more details about these relations, see [
2].
Hence, an immediate consequence of Theorem 4 is the following one. Like before, we supppose that
, for some
and represent matrix
as
where
D is the diagonal part of
A, and the dimension of
is
k.
Theorem 5. Let , whereand Then, all eigenvalues of A different from diagonal entries belong to .
Proof. Suppose that there exists an eigenvalue
, such that
. Then,
Here, (
14) and (15) mean that
is an
S-SDD
F matrix, and hence is nonsingular, which is an obvious contradiction. □
Remark 5. Let us denote, for all , Remark 6. For matrices with equal diagonal entries, for all , the localization set becomes a union of one disk and one Cartesian oval. Indeed, in this particular case, A is represented asandso thatmeaning that the eigenvalue localization set is the union of one circle and one Cartesian oval. It means that, for example, checking if (open left half-plane), i.e., getting an answer about (local) stability, requires plotting only one circle and one Cartesian oval. Obviously, for matrices with different diagonal entries, plotting and can be computationaly demanding. Nevertheless, there are still important benefits of the newly introduced class, as we shall see in the following subsections.
3.2. Application 2: Stability
Stability is a very important property in the dynamical analysis of linear, as well as nonlinear, systems. The classical approach is based on the Routh–Hurwitz criterion, and implies knowledge of the eigenvalue position—whether all are situated in the open left half-plane, or not. However, it requires solving the characteristic equation, which becomes too expensive for higher-order systems. On the other hand, the Geršgorin theorem localizes the eigenvalues in the union of circles, called the Geršgorin set, so it is easy to find the rightmost point of the Geršgorin set, and from this information conclude if the whole Geršgorin set (hence all eigenvalues) are in the open left half-plane.
However, the Geršgorin set can be too wide, giving no answer to if the eigenvalues are situated in the open left half-plane, or not. Fortunately, a similar reasoning is also valid in the case of the so-called Minimal Geršgorin set, for more information see [
2]. Namely, it also contains all eigenvalues; therefore, if it is situated in the open left half-plane, all eigenvalues will be also in the half-plane, and the observed dynamical system will be stable. From the very tight relation between H-matrices and the Minimal Geršgorin set, we already know that the following theorem holds true, but, nevertheless, we will state it here in the form of Lemma and present a brief proof, for readers who are not familiar with the Minimal Geršgorin set.
Lemma 1. If A is an H-matrix with negative diagonal entries, then , where denotes the spectrum of matrix A.
Proof. If
A is an H-matrix, then there exists a positive diagonal matrix
, such that
is strictly diagonally dominant, i.e.,
Since all diagonal entries of
A are negative, this can be rewritten as
If
is an eigenvalue of
A, it is also the eigenvalue of
and belongs to at least one of the Geršgorin circles for
, i.e., there exists
such that
But then,
which means that
belongs to the circle centered in
, with a radius less than
. Hence,
□
Remark 7. For those familiar with the Minimal Geršgorin set, the proof follows directly from the fact that the Minimal Geršgorin set is contained in if A is an H-matrix with negative diagonal entries.
As a consequence, we directly obtain the following corollary.
Corollary 1. If A is an S-SDDF matrix with negative diagonal entries, then
As an illustration, consider matrix , which is generated from the generalized Lotka–Volterra equations modeling energy flow in complex food webs. As we have already seen, this matrix is an SDDF matrix, which is sufficient to conclude that the whole spectrum of belongs to the open left-half complex plane, meaning that the corresponding dynamical system is (locally) stable.
Sometimes, small perturbations in measuring biomasses can obtain matrix from the SDDF class. In such a situation, there is a good chance that the perturbed matrix will remain in the S-SDDF class for at least one S, which is still enough to conclude that the spectrum of belongs to the open left-half complex plane, meaning that the corresponding dynamical system is (locally) stable.
Let us explain the reason for emphasizing this application more precisely. Our illustrative matrix comes from a model of energy flow in soil food webs (for a detailed treatment of this problem, see [
23]). To model the fluxes in carbon and nitrogen within the soil food web, the Lotka–Volterra predator–prey system of nonlinear differential equations is used as a basis, and this concept of energetic food webs was developed mainly to discuss the relationship between the complexity and stability of soil ecosystems, see [
24,
25]. In it, the community matrix (the Jacobian of the nonlinear differential system at the equilibrium point) and its properties lie at the center of attention. The question above all others is asymptotic stability (from which the analysis proceeds to robust stability, distance to instability, etc.) Since all diagonal elements of such community matrices are negative, stability is ensured knowing that the community matrix belongs to the class of H-matrices. As we have already pointed out, due to computational costs, we never check this property by definition, but check whether the community matrix belongs to some subclass of H-matrices. Until now, all known subclasses were based on the maximum matrix norm, which can realistically be interpreted as the conditions of trophic influences within one functional group. Due to uncertainties (measurement errors, stochastic fluctuations, etc.) in empirical data, these conditions are often violated. If we instead use the Frobenius norm, an approach that treats trophic influences between all functional groups together, we will obtain a condition more robust to the aforementioned uncertainties. In addition, in the community matrices we are talking about, the off-diagonal elements of one particular row are very often disproportionately larger than the others, which favors the use of the Frobenius matrix norm, rather than the maximum one.
3.3. Application 3: Norm Bounds for the Inverse
Whenever a new subclass of H-matrices is discovered, it is fruitful to try to find a new upper bound for the norm of its inverse, as has been done for SDD matrices by the well-known Varah’s result [
3]. This is important for at least two reasons: first, we are able to estimate the norm of the inverse for the new matrices, which we could not do before, and second, we can improve some already-known estimates for the classes belonging to this new matrix class.
When it comes to the Euclidean norm, one possible way to estimate the norm of an inverse matrix is given in [
26]. This approach is based on using the block structure of the observed matrix and it is efficient when the diagonal blocks are easily invertible. If this is not the case, we can estimate the norm of the inverse matrix using our newly introduced class.
Of course, it is possible to use all estimations obtained for the maximum norm, multiplied by . Needless to say that, in the case of large dimensions n, multiplying by makes the obtained estimate meaningless.
Theorem 6. Let D be the diagonal part of A, and , whereIf A is an S-SDDF matrix, for , thenwhere Proof. Let
A be an
S-SDD
F matrix. Then, it is a nonsingular matrix, so
D and
T are nonsingular matrices, too. According to conditions (
8) and (
9), we have
and, as a consequence,
, as well. Condition
means that
is an SDD
F matrix, and from Theorem 3 we have
is an SDD
F matrix, too, and
Recall, now, the Schur complement, see [
27], for matrix
T and denote it by
:
Obviously,
with
so, similarly to before, we obtain
According to the Banachiewicz inversion formula, see [
27], the inverse of
T can be represented in the block form as
Finally, from
we have
□
As an illustration, consider, again, our
matrix from ecological modeling. Up to now, we were able to estimate maximum norm of this matrix, using the fact that it is an
S-SDD matrix for
, and applying the estimate from [
28]:
where
This will give us the following estimation
In the Euclidean (and Frobenius) norm, we have
Using our new subclasses of H-matrices, for bounding the Frobenius (and Euclidean) norm of the inverse of
, we can use both (
5)
and (
16), for
:
since
is an SDD
F matrix, so it is an
S-SDD
F matrix for all
. The exact values are
and
, so this illustrative example can serve as a justification for introducing the new
S-SDD
F class of matrices.
Remark 8. It is worth mentioning that an upper bound for the inverse in the Frobenius norm can be used for bounding the smallest singular value of a given matrix. Namely, it is well known that for nonsingular matrices Sincewe immediately obtain In the case of matrix , this looks like Remark 9. Conditioning is an important property of the matrix. Roughly speaking, the condition number of a matrix A is the rate at which the solution x of a linear system will change with respect to a change in b. In the arbitrary consistent matrix norm , it is defined as In the Euclidean matrix norm, it iswhere and are maximal and minimal singular values of A, respectively. Thanks to Theorem 6, with the same notations, we have an upper bound for the condition number of an S-SDDF matrix A: In the case of matrix , since , we obtain In a special case, when matrix A has all diagonal entries equal, we have , for some scalar α, so that Since , the condition number in the Euclidean norm, as well as in the Frobenius norm, in this particular case can be estimated with Remark 10. A more precise analysis of the distance to instability, which is based on the localization of the pseudospectrum, can be performed in a similar way to what was carried out in [6], but this goes beyond the scope of this paper. 4. Conclusions
It is well known that the class of H-matrices, i.e., generalized diagonally dominant (GDD) matrices, stands as a cornerstone for numerous fields of applications of numerical linear algebra. For example, dynamical systems arising in different scientific areas exploit the knowledge of H-matrix theory in the following sense: being able to confirm that the Jacobian matrix governing the dynamics of a given dynamical system belongs to H-matrix class with negative diagonal entries is enough to ensure the system’s local stability. Finding out new subclasses of non-singular H-matrices, defined by computationally undemanding criteria, is very important, since this makes the process of H-matrix class membership confirmation easier.
Instead of focusing on numerous subclasses of H-matrices closely connected to the absolute vector norm and the maximum matrix norm, this paper offers a contribution to the introduction of novel subclasses depending on the Euclidean vector norm and the Frobenius matrix norm, whilst addressing their potential benefits in matrix spectrum localization, the stability of dynamical systems, and the Frobenius norm estimation of the matrix inverse.
An illustrative example is taken from ecological modeling, where the off-diagonal elements of one specific row are very often disproportionately greater than those of the others, a situation which favours using the Frobenius matrix norm rather than the maximum one.