1. Introduction
Systems of equations, particularly
are essential tools in linear algebra and have widespread applications in diverse scientific and engineering disciplines. These equations often appear in various domains, such as control theory, optimization, image processing, system identification, and robotics [
1,
2,
3,
4,
5,
6,
7]. Specifically, the matrix system
can represent the state-space model of a dynamic system, where
A and
B correspond to system transformations,
X represents the system state, and
C and
D are the output matrices [
8]. Solving these equations provides the system’s state at a given time. In signal processing, particularly in filter design and signal reconstruction, the filter matrix
X transforms input signals
A to output
C, ensuring the transformed signal interacts correctly with
B to produce output
D [
9]. This concept extends to image processing, where matrices
A and
B represent operations (e.g., encryption), and
C and
D are the original and transformed images. Solving for
X gives the required transformation. In robotics and computer vision, this matrix system arises in rigid body transformations. Dual quaternions represent 3D transformations, where
A and
B may represent rotation and translation, and
X is the transformation matrix [
10]. This system is vital for solving inverse kinematics problems, such as determining the joint parameters of a robotic arm.
Given the wide range of applications of the system (
1), it has been extensively studied and possesses a wealth of results. This paper aims to summarize the theoretical results related to the matrix equation system (
1), mainly focusing on the conditions and corresponding expressions for the existence of general solutions, least squares solutions, and minimum norm solutions. The paper also highlights generalized inverse methods and matrix decomposition methods in real and complex fields, as well as special solving methods for certain algebraic structures, such as Hilbert
-modules, Hilbert spaces, rings, dual numbers, quaternions, split quaternions, and dual quaternions. The special solutions of the system (
1) introduced in this article across various algebraic structures and their corresponding solution methods are shown in
Figure 1.
The most widely used and earliest approach for solving system (
1) is based on generalized inverses or inner inverses. For special forms of solutions, such as Hermitian, non-negative definite, maximal and minimal rank solutions, and generalized (anti-)reflexive solutions, this class of method provides a rich theoretical framework. On the other hand, matrix decomposition is a powerful tool for solving more complex special forms of solutions. Due to the different forms resulting from various matrix decompositions, these special forms can be used to construct corresponding special solutions. Related research covers symmetric, mirror-symmetric, bi-(skew-)symmetric, and orthogonal solutions over the real numbers, as well as unitary, (semi-)positive definite, generalized reflexive, generalized conjugate, and Hamiltonian solutions over the complex numbers.
For certain special algebraic structures, there are specialized solving methods. For example, in Hilbert
-modules, Hilbert spaces, and rings, inner inverses are widely used. For quaternions, dual numbers, and dual quaternions, generalized inverses can also be applied to solve the system (
1). However, in the case of split quaternions, matrix representations are the more widely used approach for solving the system. Additionally, some researchers have discussed the use of determinants to express the form of solutions for quaternion systems. This paper also provides examples of applying the system (
1) to dual quaternion matrices and dual split quaternion tensors for the encryption and decryption of color images and videos.
The remainder of the paper is organized as follows.
Section 2 introduces generalized inverse methods for solving the general solution, Hermitian and non-negative definite solutions, maximal and minimal rank solutions, and generalized reflexive solutions. The study of system (
1) in Hilbert
-modules, Hilbert spaces, and rings is presented in
Section 3.
Section 4 discusses eigenvalue decomposition, singular value decomposition, and generalized singular value decomposition of matrices, along with research conclusions for some special solutions of system (
1).
Section 5 and
Section 6 focus on the studies of dual numbers and quaternions, respectively.
Section 7 introduces examples of using system (
1) in the encryption and decryption of color images and videos. Finally,
Section 8 summarizes the content of the paper.
For convenience in the narration of this paper, the following notations are used uniformly. Symbols , , , , and represent the real number field, the complex number field, the set of matrices over the real numbers, the set of complex vectors with n elements, and the set of matrices over the complex numbers, respectively. O and I denote appropriately sized zero matrices and identity matrices. For an arbitrary matrix, A, , , and represent the conjugate, transpose, and conjugate transpose of A, respectively. For an matrix A over the real numbers, complex numbers, or quaternions, represents the rank of A and expresses the range (column space) of A. For a complex square matrix A, it is (semi-)positive definite if and only if, for every , we have . For two complex square matrices A and B of the same size, we say that in the Löwner partial ordering if is (semi-)positive definite. The symbols and denote the numbers of positive and negative eigenvalues of a Hermitian complex matrix A, counted with multiplicities. Note that, for Hermitian non-negative definite matrix A, is the matrix satisfying . The mentioned below represents the Frobenius-norm of a matrix.
2. The Generalized Inverse Methods for Solving (1)
Since 1954, Penrose has described a generalization of the inverse of non-singular matrices through the unique solution of a system of four matrix equations [
11]. This area has since attracted considerable attention.
For
, there exists a unique
satisfying the following system:
where
is called the general inverse or the Moore–Penrose inverse of
A. In the following discussion, we denote the symbols
and
.
Penrose proposed the necessary and sufficient conditions for the matrix system , along with an expression for its solutions in terms of the general inverse.
Theorem 1 (General solutions using the Moore–Penrose inverse for (
1) over
. [
11])
. Let . The matrix system (
1)
is solvable if and only if the equations and are consistent, and the condition holds, or equivalently,Under these conditions, the general solution is given by Remark 1. The concept of the general inverse and Theorem 1 can also be extended to von Neumann regular rings, particularly to the quaternion algebra [12]. Later, the concept of the g-inverse of a complex matrix was introduced by Rao and Mitra [
13]. For
, if the matrix
satisfies
then
is defined as the g-inverse of
A.
Remark 2. The g-inverse of a complex matrix is not necessarily unique.
The g-inverse can also be used to represent the solution of linear matrix systems. Theorem 1 can be restated in terms of the g-inverse as follows.
Theorem 2 (General solutions using the g-inverse for (
1) over
. [
14])
. Let . The matrix system (
1)
is solvable if and only ifWhen (
1)
is consistent, the general solution is expressed aswhere is arbitrary. Subsequently, researchers have explored a range of special solutions to system (
1) using generalized inverses, including Hermitian solutions, non-negative solutions, maximal and minimal rank solutions, and general (anti-)reflexive solutions, as well as real non-negative and real positive solutions, among others [
14,
15,
16,
17,
18,
19,
20,
21,
22,
23,
24,
25,
26,
27,
28,
29,
30].
The earliest research on the Hermitian and non-negative definite solutions of system (
1) was conducted by Mitra et al. [
15], followed by their subsequent work on the possible minimal rank of the solutions [
16]. Mitra’s focus on matrix equation systems continued, and in 1990 he extended the study to a more general form of the system
[
14]. After 2000, research on system (
1) became more in-depth: Peng and other scholars investigated the (anti-)reflexive solutions of the system [
17,
18,
19,
20,
21,
22,
23], while Liu et al. focused on the least squares solutions and the rank of the solutions, exploring the ranks of matrix blocks and the corresponding conditions using block matrix formulations [
24,
25]. Due to the unique properties of Hermitian matrices, Wang et al. examined the existence conditions and expressions for Hermitian solutions to system (
1) that satisfy various inequality constraints, as well as the ranks and inertia indices of these solutions [
26,
27,
28]. Additionally, some scholars have focused on bi-(skew-)symmetric solutions and reducible solutions [
29,
30].
2.1. Hermitian, Nonnegative Solutions
Hermitian and non-negative matrices are among the most widely applied special types of matrices, and their associated properties have been thoroughly studied. A complex square matrix is called Hermitian if .
In 1976, Khatri and Mitra considered the necessary and sufficient conditions for the existence of Hermitian and non-negative definite solutions to system (
1) and provided expressions for the solutions when they exist. The main results are stated in Theorem 3.
Theorem 3 (Hermitian and non-negative solutions for (
1) over
. [
15])
. Let and such that the system (
1)
is solvable. Define The system (
1)
has Hermitian solutions if and only if M is Hermitian. Under this condition, a general Hermitian solution is given bywhere is an arbitrary Hermitian matrix. The system (
1)
has non-negative definite solutions if and only if M is non-negative definite and . Under this condition, general non-negative definite solutions have the form ofwhere is an arbitrary non-negative definite matrix. Based on Theorem 3, the following theorem considers the solvability conditions and explicit expressions for the Hermitian solutions to the system (
1) with inequality constraints:
and
for given
and Hermitian
.
Theorem 4 (Hermitian solutions for (
3) and (
4) over
. [
28])
. Given and Hermitian . Assume The system (
3)
has Hermitian solutions if and only if is Hermitian, and At this point, the Hermitian solution can be expressed aswhere is an arbitrary non-negative definite Hermitian matrix, is arbitrary. The system (
4)
has Hermitian non-negative definite solutions if and only if T is a non-negative Hermitian matrix, and At this point, the Hermitian non-negative definite solution can be described aswherewith arbitrary , and non-negative definite Hermitian . Remark 3. In Theorem 4, selecting and can yield the Hermitian non-negative solutions to (
1)
. Remark 4. Theorems 3 and 4 are derived by converting the system of equations into a single matrix equation, making the form more concise. However, this approach increases the size of the matrix and requires the computation of the generalized inverse of block matrices.
In [
28], the authors consider the maximal rank and inertia of the Hermitian solution to (
3) using matrix decomposition methods, which will be introduced in the next section.
Additionally, Ke and Ma [
31] have supplemented the results for the symmetric solutions to system (
1) over
.
Theorem 5 (Symmetric solutions for (
1) over
. [
31])
. Given and . Denote , and . The system (
1)
has a symmetric solution if and only if the system of matrix equationshas a solution . In this case, the symmetric solution to (
1)
is given byOr in an equivalent way, equationshold. At this point, the symmetric solution of (
1)
can be expressed aswhere is an arbitrary matrix. 2.2. Maximal and Minimal Rank Solutions with Inequality Constrain
Through the expression of the solution to the system (
1) given by generalized inverses, the case of the rank of the solution can be further studied.
In 1984, Mitra obtained the minimal possible rank solutions to the system (
1).
Theorem 6 (Minimal possible rank solutions for (
1) over
. [
16])
. Let such that (
1)
is consistent. Assume without loss of generality that Let X be a solution of the matrix system (
1)
. Then,Additionally, if and only if
Decades later, Liu extended Theorem 6, considering the maximal and minimal ranks of the general solutions and the least squares solutions for the system (
1).
Theorem 7 (Maximal and minimal rank solutions for (
1) over
. [
24])
. For , the system (
1)
is consistent with a general solution . Then, the maximal and minimal ranks of X are given by The least squares solution for the system (
1) can be expressed as shown in [
24], which also provides the conditions for the uniqueness of the least squares solution and the expression for the solution when it is unique.
Theorem 8 (Least squares solutions for (
1) over
. [
24])
. Assume that , . The necessary and sufficient condition for (
1)
to have a least squares solution is If (
5)
holds, then the general least squares solution of (
1)
is expressed aswhere is arbitrary. The least squares solution of (
1)
is unique if and only if In this case, the unique least squares solution is Additionally, the maximal and minimal ranks of the least squares solutions for the system (
1) are considered based on Theorem 8.
Theorem 9 (Maximal and minimal least squares solutions for (
1) over
. [
24])
. For . If the system (
1)
has a general least squares solution , then the maximal and minimal ranks of X are given by Liu also presented a set of formulas for the maximal and minimal ranks of the submatrices in a general solution
X to the system (
1) in [
25].
Suppose that
X is a general solution of the system (
1), and let
X be partitioned into a
block form:
In this case, the system (
1) can be rewritten as
where
and
, with
. Adopt the following notations for the collections of submatrices
, and
as
The submatrices
can be rewritten as the form of
Substituting the general solution (
2) gives the general expressions for
, as follows:
where
.
Liu [
25] summarized the possible range of ranks for the solution to (
1) as follows:
Theorem 10 (Maximal and minimal rank solutions using block matrices for (
1) over
. [
25])
. Let , such that the matrix system (
1)
has a general solution. Then, In addition, by using the ranks of matrix blocks, the necessary and sufficient conditions for the uniqueness of the solution to system (
1) are given in block matrix form.
Theorem 11 (Unique conditions of general solutions for (
1) over
. [
25])
. Suppose that matrix system (
6)
has a solution. Then, the following statements hold. The block in a general solution to (
6)
is unique if and only if The block in a general solution to (
6)
is unique if and only if The block in a general solution to (
6)
is unique if and only if The block in a general solution to (
6)
is unique if and only if Theorem 12 (General solutions using block matrices for (
1) over
. [
25])
. Let , such that the system (
6)
has a solution. Consider in (
7)
as four independent matrix sets. Then, The four submatrices in (
7)
are independent. Specifically, for any choice of , the corresponding matrix is a solution of (
7)
if and only if Furthermore, Wang et al. first considered the extremal inertias and ranks of
and
, where
P and
Q are Hermitian and
X is a solution of (
1). They also derived the necessary and sufficient conditions for special cases such as unitary solvability, contraction solvability, and the left and right minimal solutions to the system (
1) [
26].
For
,
A is called a unitary matrix if and only if
. Let
H be a given set consisting of some matrices in
, and we say that
is minimal (maximal) if
(or
) for every
. Denote
A solution X is called left (right) minimal or maximal if () is the minimal or maximal matrix of the set (). When , X is called a contraction matrix. Furthermore, if , X is called a strict contraction matrix.
The main findings of [
26] on the extremal inertias and ranks of
and
are summarized below:
Theorem 13 (Extreme rank and inertia of
for
X satisfying (
1) over
. [
26])
. Let , , and . Suppose that (
1)
has a solution. Denote the set of all solutions to (
1)
by S. Then, where where In Theorem 13, selecting
P as the identity matrix can derive the necessary and sufficient conditions for (
1) to have some special solutions, which are presented in the following corollary.
Corollary 1 (Unitary and (strict) contraction solutions for (
1) over
. [
26])
. Let , and . Suppose that the system (
1)
has general solutions. Then, (
1)
has unitary matrix solutions if and only if The system (
1)
has strict contraction matrix solutions if and only if The system (
1)
has contraction matrix solutions when and The left (right) minimal and maximal solutions to (
1) are discussed as follows.
Theorem 14 (Left (right) minimal and maximal solutions for (
1) over
. [
26])
. Let , and , with and . There exists a solution X to (
1)
such that X is the left minimal solution if and only if Under this circumstance, the left minimal solution is
There exists a solution X to (
1)
such that X is the right minimal solution if and only if Under these circumstances, the right minimal solution is
There does not exist a right or left maximal solution X to (
1)
. In a similar manner, Yao derived the maximal and minimal ranks and inertias of
, where
X satisfies the system (
1), with
P and
Q being Hermitian.
Theorem 15 (Extreme rank and inertia of
for
X satisfying (
1) over
. [
27])
. For , , and Hermitian , assume that (
1)
has a solution. Denote the set of all solutions to (
1)
by S. Then, Theorem 15 can derive the positive definiteness of in the following corollary.
Corollary 2 (Positive definiteness of
for
X satisfying (
1) over
. [
27])
. Let , , and Hermitian . Assume that the system (
1)
is consistent. Let and be as defined in Theorem 15
. Then, we have the following statements. The system (
1)
has a solution such that if and only if The system (
1)
has a solution such that if and only if The system (
1)
has a solution such that precisely when The system (
1)
has a solution such that precisely when All general solutions of (
1)
satisfy if and only if All general solutions of (
1)
satisfy if and only if All general solutions of (
1)
satisfy precisely whenor All general solutions of (
1)
satisfy precisely whenor The system (
1)
has a solution such that if and only if At the end of this section, based on Theorem 4, we present the maximal rank and inertia of the Hermitian solutions to (
3), which has an additional inequality constraint
, where
and Hermitian
are given matrices.
Theorem 16 (Maximal rank and inertia of the Hermitian solutions of (
3) over
. [
28])
. Let , and Hermitian . Assume thatDenote the set of all Hermitian solutions to the system (
3)
by S. The maximal rank of iswhere The maximal inertia of is 2.3. Generalized (Anti-)Reflexive Solutions
The general (anti-)reflexive matrices have wide applications in fields such as engineering and science [
32]. A matrix
is called (anti-)reflexive with respect to the nontrivial generalized reflection matrix
P if
, or equivalently
, where
, and
P is the nontrivial generalized reflection matrix satisfying
.
is called generalized (anti-)reflexive if
, where
G is a given unitary matrix of order
n, or equivalently
, where
P and
Q are nontrivial generalized reflection matrices.
Qiu et al. considered the (anti-)reflexive solutions to system (
1), presenting the related results for the (anti-)reflexive solutions and the generalized (anti-)reflexive solutions, along with the corresponding least norm solutions.
Theorem 17 ((Anti-)reflexive solutions for (
1) over
. [
20])
. For given , , and the nontrivial generalized reflection matrix , let The system has an (anti-)reflexive solution if and only iffor . In the meantime, the (anti-)reflexive solution is given bywith (anti-)reflexive . The least norm (anti-)reflexive solution is expressed as The relevant conclusions for generalized (anti-)reflexive solutions are presented below.
Theorem 18 (Generalized (anti-)reflexive solutions for (
1) over
. [
20,
21,
22,
23])
. For given , , and the nontrivial generalized reflection matrix , let The system (
1)
has a generalized (anti-)reflexive solution if and only iffor . In the meantime, the generalized (anti-)reflexive solution is given bywhere is generalized (anti-)reflexive. The least norm generalized (anti-)reflexive solution is expressed as Let symbols
and
represent the set of all generalized reflexive and anti-reflexive solutions of the system (
1), respectively. For given
, when we select
in (
8), then (
8) is the unique solution of approximation problem
. If
then (
8) is the unique solution of
.
Theorems 17 and 18 have presented the generalized (anti-)reflexive solutions of (
1). For generalized (anti-)reflexive solutions and other more complex forms, the theoretical results are extensive. Additionally, matrix decomposition is a more widely used method for solving this special type of solution, which will be introduced in the next section.
2.4. Re-nnd, Re-pd Solutions
For a matrix , the real part of A is defined as . A matrix A is referred to as real non-negative definite (Re-nnd) if is positive semi-definite, and A is called real positive definite (Re-pd) if is positive definite.
Between 2011 and 2014, Xiong, Qin, and Liu explored the Re-nnd and Re-pd solutions to the system (
1) [
29,
30].
Theorem 19 (Re-nnd solutions for (
1) over
. [
30])
. For , suppose that each equation in (
1)
has a Re-nnd solution. If the system (
1)
has a solution, then there exists a Re-nnd solution if and only if all the solutions are Re-nnd solutions if and only if or . Theorem 20 (Re-pd solutions for (
1) over
. [
30])
. For , assume that each equation in (
1)
has a Re-pd solution. If the system (
1)
has a solution, then there exists a Re-pd solution.
all the solutions are Re-pd solutions if and only if Remark 5 ([
29])
. When the system (
1)
has a Re-nnd (Re-pd) solution, one of the Re-nnd (Re-pd) solutions is given byfor some Re-nnd (Re-pd) matrix . In this section, we introduced the generalized inverse method for solving the system (
1), along with the conclusions regarding special solutions, such as Hermitian, non-negative, generalized (anti-)reflexive, Re-nnd, and Re-pd solutions. Additionally, we focused on the maximal and minimal ranks and inertias of the solutions to system (
1), the conditions under which these extremal values are achieved, and the expression of solutions that satisfy inequality constraints.
The generalized inverse methods were first proposed for solving the system (
1) and are currently the most widely used approach. These methods often provide a more concise expression for the solutions. However, there are many special types of solution that cannot be represented solely using generalized inverses. In
Section 4, a deeper exploration of these special solutions will be presented using matrix decomposition techniques.
3. The System (1) over Hilbert Spaces, Hilbert -Modules, and Rings
The research mentioned above on matrices can be extended to more general cases, such as Hilbert spaces, Hilbert
-modules, and rings. However, there are several limitations when extending to these cases, leading to fewer studies compared to those on matrices. These studies primarily focus on the Hermitian and positive cases, with some scholars also investigating reducible solutions [
33,
34,
35,
36,
37,
38,
39].
Dajić and Koliha were the first to study the system (
1) for bounded linear operators between Hilbert spaces with the restriction that
A and
B have closed ranges. They provided conditions for the existence of general, Hermitian, and positive solutions and obtained formulas for the general form of these solutions [
33]. Later, they extended these results from rings to rectangular matrices and operators between complex Hilbert spaces via embedding [
34]. In 2008, Xu considered conditions under Hilbert
-modules [
35]. In 2016, (anti-)reflexive solutions over rings were examined using the inner inverse [
36]. In 2021, Radenković et al. reconsidered the system (
1) in the context of Hilbert
-modules using orthogonally complemented projections, providing alternative expressions [
37]. Subsequently, Zhang et al. discussed positive and real positive solutions to (
1) over Hilbert spaces using the reduced solution to the system
and
, where the ranges of
A and
C may not be closed [
38,
39].
A Hilbert space is a complete inner-product space. A Hilbert -module is a natural generalization of a Hilbert space, obtained by replacing the field of scalars with a -algebra. Since finite-dimensional spaces, Hilbert spaces, and -algebras can all be regarded as Hilbert -modules, matrix equations can be studied in a unified manner within the framework of Hilbert -modules. The scope of rings is even broader. In the following statements, “ring” refers to an associative ring R with a unit element .
Hereinafter, we introduce the notations and definitions used in this section:
Let H, K, and L denote complex Hilbert spaces, represent the set of all bounded linear operators between H and K, and be the set of all bounded linear operators over H. For , let , , and represent the range, the null space, and the closure of the range of the operator A, respectively. An operator is said to be regular if there exists an operator such that . is referred to as the inner inverse of A. It is well known that A is regular if and only if A has a closed range. M is a closed subspace of H, and denotes the orthogonal projection onto M.
The Hilbert -module is analogous to a Hilbert space, except that its inner product is not scalar-valued, but takes values in a -algebra. Therefore, we continue to use the same notation for the Hilbert module as is used for Hilbert spaces. On Hilbert -modules, a closed submodule M of H is said to be orthogonally complemented in H if , where . In this case, the projection from H onto M is denoted by .
For an arbitrary ring R with involution , an element is Hermitian if . If there exists such that , then a is said to be regular (or inner invertible), and b is called the inner inverse of a, denoted as .
Initially, the general solution of (
1) over rings is presented.
Theorem 21 (General solutions for (
1) over a ring. [
34,
36])
. Let such that a and b are regular elements. Then, the following statements are equivalent. There exists a solution of the system of equations .
and .
Moreover, if or is satisfied, then any solution of can be expressed asfor any . Remark 6. In [34], the authors further extended the results from elements in the ring to matrices over R, as well as to bounded linear operators between complex Banach or Hilbert spaces by constructing an embedding. Next, the expression for the general solution in Hilbert -modules is provided through orthogonal complements.
Theorem 22 (General solutions for (
1) over Hilbert
-modules. [
37])
. Let be Hilbert -modules, , such that and are orthogonally complemented. Then, the system (
1)
has a general solution if and only ifIn such a case, the general solution has the form ofwhere is arbitrary. Remark 7. Actually, and being orthogonally complemented implies that A and B are regular and have closed ranges. Additionally, Theorem 22 uses the Moore–Penrose inverse in place of the inner inverse. The definition of the Moore–Penrose inverse is similar to that in matrices, and therefore will not be elaborated further.
In 2023, Zhang et al. extended this result to infinite-dimensional Hilbert spaces without the requirement that the corresponding operators A and B have closed ranges, using reduced matrices.
Theorem 23 (General solutions for (
1) over a Hilbert space. [
38])
. Let , and . Then, the system (
1)
has a solution if and only if , , and . In this case, the general solution can be represented bywhere F is the reduced solution of , H is the reduced solution of , and is arbitrary. Specifically, if and are closed, the general solution can be represented by The situations of Hermitian solutions, positive solutions, real positive solutions, and reflexive solutions will be introduced in the following sections.
3.1. Hermitian Solutions
The Hermitian solution of (
1) over Hilbert space was first studied by Dajić and Koliha.
Theorem 24 (Hermitian solutions for (
1) over a Hilbert space. [
33,
34])
. Let , , and the operators A and B have closed ranges. Assume have a closed range, . Let , , and represent the inner inverses of A, B, and M, respectively. Then, the system (
1)
has a Hermitian solution if and only ifand and are Hermitian. The general Hermitian solution is given bywhere is Hermitian. Remark 8. In [34], the authors presented an alternative form of (
9)
and (
10)
:andwhere , , and is Hermitian. Remark 9. Theorem 24 holds true in -Hilbert modules and rings with involution as well [34,35]. By using the projection operator, Theorem 24 can be restated in another form over Hilbert -modules.
Theorem 25 (Hermitian solutions for (
1) over Hilbert
-modules. [
37])
. Let be Hilbert -modules, and , such that and are orthogonally complemented. Let , and assume that is orthogonally complemented. Then, the system (
1)
has a Hermitian solution if and only ifand and are Hermitian, where . In this case, the Hermitian solution has the formwhere and is an arbitrary Hermitian matrix. 3.2. Positive Solutions
For the cases where the system (
1) has a positive solution over Hilbert space, two different descriptions are presented as follows:
Theorem 26 (Positive solutions for (
1) over a Hilbert space. [
33])
. Let , ,Assume that and Q are regular. The system (
1)
has a positive solution if and only if Q is positive and . The general positive solution is given bywhere is an arbitrary positive matrix. Theorem 27 (Positive solutions for (
1) over a Hilbert space. [
33,
35])
. If , , , and , such that A, B, M, and are regular, . Then, (
1)
has a positive solution if and only ifand and F are positive. The general positive solution is given bywhere is positive. Remark 10. Dajić and Koliha extended Theorems 26 and 27 to strongly ∗
-reducing rings with involution, which is an extension of the -Hilbert modules [34]. Zhang et al. presented the existence and the general form of the positive solutions of (
1) without the restriction on the closed range over Hilbert space.
Theorem 28 (Positive solutions using projection operators for (
1) over a Hilbert space. [
38])
. Let and D be operators in . Let and . The system (
1)
has positive solutions if and only if the following conditions hold. , , and .
and .
, , and .
In which G, H, and K are the reduced solutions of , , and , respectively. Than, the positive solution is given byfor any positive operator , where L is the reduced solution of . The conclusions for the Hilbert -modules can be directly obtained using the projection operator.
Theorem 29 (Positive solutions using projection operators for (
1) over Hilbert
- modules. [
37])
. Let be Hilbert -modules, , such that and are orthogonally complemented. Denote , , , and . Assume that is orthogonally complemented, , along with is positive. Then, the system (
1)
has a positive solution if and only ifand is positive. In such a case, the general solution has the formwherewith is arbitrary positive. 3.3. Re-pd Solutions
The general form of the real positive solutions of (
1)
without the restriction on the closed range over Hilbert spaces was also provided by Zhang et al.
Theorem 30 (Re-pd solutions for (
1) over a Hilbert space. [
39])
. Let , , and . The system (
1)
has a real positive solution if the following conditions are satisfied. , , and .
, are real positive operators.
, where F is the reduced solution of .
In this case, one of the real positive solutions can be represented aswhere H is the reduced solution of , is arbitrary Re-pd. 3.4. (Anti-)Reflexive Solutions
Načevska used algebraic methods in rings with involution to obtain a generalization of (anti-)reflexive solutions over complex matrices [
36].
An element is said to be a generalized reflection element if and . For being generalized reflection elements, an element is called a generalized reflexive element (with respect to w and v) if , denoted by , and x is called a generalized anti-reflexive element (with respect to w and v) if , denoted by .
For
, these elements can be decomposed using projections as follows [
36]:
Theorem 31 (Reflexive solutions for (
1) over a ring)
. [36] Let and (
11)
hold, such that and are regular elements. Then, the following statements are equivalent. There exists a solution of the system (
1)
. , and .
Moreover, if or is satisfied, then any generalized reflexive solution of (
1)
can be expressed aswherefor given by (
12)
and arbitrary . Theorem 32 (Anti-reflexive solutions for (
1) over a ring. [
36])
. Let and (
11)
hold such that are regular elements. Then, the following statements are equivalent. There exists a solution of the system (
1)
. and .
Moreover, if or is satisfied, then any generalized anti-reflexive solution of (
1)
can be expressed aswherefor given by (
12)
and arbitrary . This section mainly introduces the system (
1) over Hilbert spaces, Hilbert
-modules, and rings, focusing on tools such as inner inverses, project operators, orthogonal complements, and reducibility.
4. Matrix Decomposition Methods for Solving (1)
Matrix decomposition techniques, such as eigenvalue decomposition (EVD), singular value decomposition (SVD), QR decomposition, LU decomposition, and others, play a crucial role in solving matrix equations. They are especially important in finding special solutions for system (
1), such as mirror-symmetric, skew-symmetric, orthogonal symmetric, unitary,
-reflexive, (Hermitian)
R-conjugate,
-conjugate, and Hamiltonian solutions, as well as the corresponding least squares solutions. These types of solution, which are difficult to calculate using generalized inverses alone, can be efficiently computed using matrix decompositions. This section will introduce the related methods and results.
We introduce the eigenvalue decomposition (EVD). Let
A be an
matrix with
n linearly independent eigenvectors
for
. Then,
A can be factored as
where
Q is the square
matrix whose
i-th column is the eigenvector
of
A, and
is the diagonal matrix whose diagonal elements are the corresponding eigenvalues, with
. However, it is important to note that not all square matrices are diagonalizable. In such cases, a more practical form of decomposition is the singular value decomposition.
The singular value decomposition (SVD) of a given
matrix
A of rank
k is
where
U and
V are unitary or orthogonal matrices and
, with
.
In 1981, Paige and Saunders extended the
B-singular value decomposition in [
40] to the generalized singular value decomposition (GSVD) for two matrices with the same number of columns. This is a powerful tool for solving equations. The GSVD can be expressed as the following lemma.
Lemma 1 (The generalized singular value decomposition over
. [
41])
. Let and be two matrices with the same number of columns. Denote There exist unitary matrices U and V, and a nonsingular matrix Q, such thatwherewith , , and for . Remark 11. It is important to note that the GSVD has various representations, which cannot be exhaustively listed. In the subsequent discussion, alternative forms of the GSVD will be presented.
Next, we present the necessary and sufficient conditions for the system (
1) and the expression of the general solutions through the SVD.
Theorem 33 (General solutions using the SVD for (
1) over
)
. For , and , the SVD of A and C expressed aswhere and . DenoteThen, the system (
1)
is consistent if and only if are zero matrices and . The general solution can be expressed aswhere is arbitrary. Remark 12. Theorem 33
is a special form of the system , which is considered by the GSVD in [42]. 4.1. Various Symmetric Solutions
In this section, we consider the least squares forms of various symmetric solutions to the system (
1) over
, including
-mirror(skew-)symmetric solutions, symmetric solutions, and bi-(anti-)symmetric solutions. The existence conditions and expressions for symmetric solutions in subspaces are also be discussed.
In 2006, Li et al. considered the least squares
-mirror-symmetric solutions through the SVD of matrices over
[
43].
A
-mirror matrix
is defined by
where
is the
k-square backward identity matrix with ones along the secondary diagonal and zeros elsewhere. A matrix
is called a
-mirror-symmetric matrix if and only if
We denote the set of all -mirror(skew-)symmetric matrices by .
The results about least squares -mirror(skew-)symmetric solutions are as follows.
Theorem 34 (Least squares
-mirror(skew-)symmetric solutions for (
1) over
. [
43])
. Let andDenotewhere Denote the SVDs of aswith Then, the general solution for the problem can be expressed aswhere , and is arbitrary. Then, has a solution . Moreover, the general solution can be given bywherewhere , , , , , , and are arbitrary. Let represent the solution set of with . For a given , has a unique solution, . Moreover, can be expressed aswherewith Φ
and being given in . In 2010, Yuan considered the least squares solutions of the linear equation system (
1) with a different form of Theorem 34
and the least squares symmetric solutions [
44].
Theorem 35 (Least squares symmetric solutions for (
1) over
. [
44])
. Assume that , and . Let the SVDs of the matrices be given bywhere are all orthogonal matrices and the partitions are compatible with the sizes of and . The least squares solution set of (
1)
can be expressed aswhere is an arbitrary matrix. The unique least norm least squares solution can be expressed as Consider the condition . Let the EVD of the matrix be given bywhere is an orthogonal matrix and the partition is compatible with the size of . Then, the least squares symmetric solution set of (
1)
can be expressed aswhere and is an arbitrary symmetric matrix. The unique least norm least squares symmetric solution can be expressed as In 2014, Ke and Ma derived the generalized bi-(skew-)symmetric solutions of the system (
1) with the corresponding least squares solution [
31].
For a symmetric orthogonal matrix , a matrix is called a generalized bi-(skew-)symmetric matrix if and only if .
Theorem 36 (Least squares bi-(skew-)symmetric solutions for (
1) over
. [
31])
. Assume that , , and is a symmetric orthogonal matrix. Let P be decomposed aswhere U is a symmetric orthogonal matrix. Let the partitions of , , , and bewith , , , , respectively. Denote that , , , , , , , . Then, (
1)
has bi-symmetric solutions if and only if equationshold. Under such circumstance, the bi-symmetric solutions can be expressed aswhere Let , , , . Then, (
1)
has bi-skew-symmetric solutions if and only ifhold. Under such circumstance, the bi-skew-symmetric solutions can be expressed aswhere Additionally, let the SVDs of and bewhere , , , and are all orthogonal matrices and the partitions are compatible with the sizes of The least squares bi-symmetric solutions of (
1)
can be expressed aswherewhere , , , , , , and are arbitrary symmetric matrices. The least squares bi-skew-symmetric solutions of (
1)
can be expressed aswherewith , , , , , and being an arbitrary matrix. Hu and Yuan have considered the symmetric solutions of (
1) on a subspace
[
45]. Let
be the set of all
symmetric matrices on subspace
, where
The necessary and sufficient conditions for the system (
1) to have a solution in
and also an expression for the solution
X are obtained. Additionally, the associated optimal approximation problem to a given matrix
is discussed, and the optimal solution is elucidated.
Theorem 37 (
solutions for (
1) over
. [
45])
. Given and . Assume that the SVD of G is given bywhere , , , are orthogonal matrices with and . Let The system (
1)
is solvable over if and only if In which cases, the solution set can be expressed aswhere are arbitrary matrices with . For , letwhere , . The optimal problem has the unique solution admittingwhere , and are given by (
13)
, with , and Z is determined by solving the unique solution of in [45]. 4.2. Various Orthogonal Solutions
This section introduces various orthogonal solutions to the system (
1) over
. Wang et al. extended the conditions for various symmetric solutions to the equations
or
to the system (
1), providing a series of conclusions. Qiu et al. constructed special matrices and used their EVDs to derive the corresponding conclusions [
46,
47].
Wang et al. have considered the orthogonality, (skew-)symmetric orthogonality, and least squares (skew-)symmetric orthogonality solutions, as well as the necessary and sufficient conditions for (
1) to have these solutions and their corresponding expressions, respectively [
46].
Theorem 38 (Orthogonal solutions for (
1) over
. [
46])
. Given and . Suppose the GSVD of A and C iswhere are orthogonal, . Denotewhere and . Let the GSVD of and bewhere are orthogonal, is diagonal. Then, the system (1) has orthogonal solutions if and only if In which case, the orthogonal solutions can be expressed aswhereare orthogonal, and is an arbitrary orthogonal matrix. Let the GSVD of B and D bewhere and are orthogonal, Partitionwhere . Assume the GSVD of and iswhere are orthogonal and is diagonal. Then, the system (
1)
has orthogonal solutions if and only if In which case, the orthogonal solutions can be expressed aswherewith arbitrary orthogonal . Remark 13. Some researchers have also investigated the correction of the coefficient matrices when the system (
1)
is inconsistent under orthogonal constraints [48], specifically focusing on the optimization problemsubject to Theorem 39 (Symmetric orthogonal solutions for (
1) over
. [
46])
. Given . Notations are defined in Theorem 38
. Let the symmetric orthogonal solutions of the matrix equation be described as inwhere and are orthogonal and is an arbitrary symmetric orthogonal matrix. Partitionwhere . Then, the system (
1)
has symmetric orthogonal solutions if and only if In which case, the solutions can be expressed aswhereand is arbitrary symmetric orthogonal. Let the symmetric orthogonal solutions of the matrix equation be described aswhere are orthogonal and is symmetric orthogonal. Partitionwhere Then, the system (
1)
has symmetric orthogonal solutions if and only if In which case, the solutions can be expressed aswhereand is an arbitrary symmetric orthogonal matrix. Theorem 40 (Skew-symmetric orthogonal solutions for (
1) over
. [
46])
. Given . Suppose the matrix equation has skew-symmetric orthogonal solutions with the formwhere is arbitrary skew-symmetric orthogonal, is orthogonal. Partitionwhere . Then, the system (
1)
has skew-symmetric orthogonal solutions if and only if In which case, the solutions can be expressed aswhereand is an arbitrary skew-symmetric orthogonal matrix. Suppose the matrix equation has skew-symmetric orthogonal solutions with the formwhere is arbitrary skew-symmetric orthogonal, are orthogonal. Partitionwhere . Then, the system (
1)
has skew-symmetric orthogonal solutions if and only if In which case, the solutions can be expressed aswhereand is an arbitrary skew-symmetric orthogonal matrix. Theorem 41 (Least squares (skew-)symmetric orthogonality solutions for (
1) over
. [
46])
. Given and , denoteLet the EVDs of T and N bewith , . Then, the least squares symmetric orthogonal solutions of the system (
1)
can be expressed aswith and being an arbitrary symmetric orthogonal matrix. Then, the least squares skew-symmetric orthogonal solutions of the system (
1)
can be expressed aswith , for and is an arbitrary skew-symmetric orthogonal matrix. Remark 14. Theorems 38–41 actually treat the system as an extension of the single equation or . The proof of these theorems is based on the perspective that one of the equations has a corresponding solution.
Qiu et al. consider the least squares orthogonality, symmetric orthogonality, symmetric idempotence, and their corresponding
P-commuting matrix solutions of (
1).
Theorem 42 (Least squares orthogonal, symmetric orthogonal, and symmetric idempotence solutions for (
1) over
. [
47])
. Assume that , . Let . Denote the SVD of iswhere are orthogonal matrices, Then, the least squares orthogonal solutions to (
1)
satisfieswhere is arbitrary orthogonal. Denote and with is an orthogonal matrix. Let the EVD of be given bywhere is an orthogonal matrix, Then, the least squares symmetric orthogonal solutions to (
1)
are expressed aswhere is arbitrary symmetric orthogonal. Denote and . Let the EVD of be given bywhere Then, the least squares symmetric idempotent solutions to (
1)
are expressed aswith arbitrary symmetric idempotent . Additionally, we generalize the corresponding
P-commuting constraints, where
is a given symmetric matrix. Let the EVD of
P be
where
and
is an orthogonal matrix. A matrix
X commutes with
P, (i.e.
), if and only if
where
for
.
Theorem 43 (Least squares orthogonality, symmetric orthogonality, and symmetric idempotence solutions commuting with
P for (
1) over
. [
47])
. Assume that , . Suppose that with . Partition the matrix conforming to (
14)
, where . Let the SVD of the matrix bewhere . Then, the least squares orthogonal solutions commuting with P to (
1)
arewhere satisfieswith arbitrary orthogonal . Denote by with , and partition the matrix conforming to (
14)
, where . Let the EVD of the matrix bewhere is orthogonal, . The least squares symmetric orthogonal solutions commuting with P to (
1)
arewhere satisfieswhere is an arbitrary symmetric orthogonal matrix with order . Denote by with , and partition the matrix conforming to (
14)
, where . Let the EVD of the matrix ‘bewhere is an orthogonal matrix, . The least squares symmetric idempotent solutions commuting with P to (
1)
arewhere satisfieswith arbitrary symmetric idempotent . 4.3. Unitary Solutions
This section presents the solvability conditions for the system (
1) with the constraint
. These conditions are derived by applying the EVD and SVD of matrices. The general solutions to these matrix equations are also provided. Furthermore, the associated optimal approximation problems for the given matrices are discussed, and the optimal approximate solutions are derived [
49].
Theorem 44 (Unitary solutions for (
1) over
. [
49])
. Suppose that , , , with , and the SVD of is given bywhere , , , with , . Let the matrices and , , be given by and the SVD of bewhere , , , with and . Then, (
1)
is solvable for unitary matrices if and only if In this case, let the EVD of be given bywhere , , , . The unitary solution set of (
1)
is Let and partition as in Let the SVD of bewhere , , , with and . If the conditions (
15)
are satisfied, then the solution of is given bywherewith arbitrary unitary . 4.4. Re-nnd and Re-pd Solutions and Inequality Constrains
This section introduces the relevant conclusions regarding the solutions of the system (
1) with inequality constraints, as well as the Re-pd and Re-nnd solutions, using the GSVD [
50,
51].
Recently, Liao et al. considered the system (
1) with the inequality constraint
.
Theorem 45 (General solutions with
constraint for (
1) over
. [
50])
. Given matrices , , , , , and . Let and . The EVDs of and can be given bywhere , , and , , are unitary matrices. The GSVD of and iswhere is a nonsingular matrix, are unitary matrices, and and with Partition into The matrix inequality subject to (
1)
has a general solution if and only if the conditionshold. In this case, the general solution of subject to (
1)
can be expressed aswhere W iswith , arbitrary and , arbitrary anti-Hermitian , arbitrary Hermitian non-negative definite . The matrix inequality subject to (
1)
has a general solution if and only if the conditionshold. In this case, the general solution of subject to (
1)
can be expressed aswherewith , arbitrary and , arbitrary anti-Hermitian , arbitrary Hermitian non-negative definite . Yuan et al. expanded upon the above research by deriving necessary and sufficient conditions for the system (
1) to have Re-nnd and Re-pd solutions. Additionally, explicit representations of the general Re-nnd and Re-pd solutions are provided when the stated conditions are satisfied.
Theorem 46 (Re-pd and Re-nnd solutions for (
1) over
. [
51])
. For given matrices , , and , denote , , and by , L, G and J, respectively. Suppose that the EVDs of G, and can be given bywhere , , , , and . The GSVD of the matrices and iswhere is a nonsingular matrix and , are unitary matrices, andwith , and the partition of the matrix is the form of The system (
1)
has a Re-nnd solution if and only if In this case, the general Re-nnd solution of (
1)
can be expressed aswhere , Θ, Ψ,
H, and are given bywith arbitrary , satisfying , arbitrary Hermitian non-negative definite and arbitrary contraction . The system (
1)
has a Re-pd solution if and only if In this case, the general Re-pd solution of (
1)
can be expressed aswhere , Θ, Ψ, H, , and are, respectively, given bywith arbitrary , satisfying , arbitrary Hermitian positive definite and arbitrary strict contraction . 4.5. Different Types of Reflexive Solutions
Some scholars considered the (anti-)reflexive solutions of the system (
1) and presented the following theorem.
Any nontrivial generalized reflection matrix
can be expressed in the form
where
, with
.
Theorem 47 ((Anti-)reflexive solutions for (
1) over
. [
17,
18,
19])
. Let , and the nontrivial generalized reflection matrix be known and be unknown. The EVD of P is given by (
16)
. Let The system (
1)
has an (anti-)reflexive solution with respect to a nontrivial generalized reflection matrix P if and only if In this case, the reflexive solution X with respect to P can be expressed asand the anti-reflexive solution X with respect to P can be expressed aswherewhere and are arbitrary. For a given matrix , let Symbols and represent the set of all reflexive and anti-reflexive solutions of the system . Then, the approximation problem has a unique solution,where The approximation problem has a unique solutionwhere Zhou and Yang considered the existence conditions of the (anti)-Hermitian reflexive solutions, which added the (anti)-Hermitian constraints in the reflexive solutions [
52].
Theorem 48 ((Anti-)Hermitian reflexive solutions for (
1) over
. [
52])
. Let , and the nontrivial generalized reflection matrix be known. The EVD of P is given by (
16)
. Denotewhere Let Then, the system (
1)
has Hermitian reflexive solutions in if and only if Moreover, the general Hermitian reflexive solution can be expressed aswhere areandwith arbitrary Hermitian and . Then, the system (
1)
has anti-Hermitian reflexive solutions in if and only if Moreover, the general anti-Hermitian reflexive solution can be expressed aswhere areandwith arbitrary anti-Hermitian and . Given matrix . Letwhere If (
1)
has Hermitian reflexive solutions, then the optimize problem has a unique Hermitian reflexive solutions of (
1)
, which can be represented aswherewith given by (
17)
. For given matrix , is partitioned as (
19)
. If (
1)
has anti-Hermitian reflexive solutions in, then the optimize problem has a unique anti-Hermitian reflexive solutions of (
1)
, which can be represented aswherewith given by (
18)
. Zhou et al. also considered the least squares (anti)-Hermitian reflexive solutions [
53].
Theorem 49 (Least squares (anti-)Hermitian reflexive solutions for (
1) over
. [
53])
. Let , and the nontrivial generalized reflection matrix be known. The EVD of P is given by (
16)
. Denote, and are given by the SVDs of , , :where The least squares Hermitian reflexive solutions of the system with respect to P can be expressed aswith being Hermitian, given bywhere and are arbitrary Hermitian matrices, . The least squares anti-Hermitian reflexive solutions of the system (
1)
with respect to P can be expressed aswith being anti-Hermitian, given bywhere and are arbitrary anti-Hermitian matrices. Given . Letwhere and denotewhere The optimization problem has a unique solution , which is the least squares Hermitian reflexive solution of (
1)
and can be represented aswhere are Hermitian, with The optimization problem has a unique solution , which is the least squares anti-Hermitian reflexive solution of (
1)
and can be represented aswhere are anti-Hermitian, with Dong and Wang have presented the system of matrix equations (
1) subject to
-reflexive and anti-reflexive constraints by converting it into two simpler cases:
and
. They provide the solvability conditions, the general solution to this system, and the least squares solution when (
1) is inconsistent [
54].
Let
and
be Hermitian and
-potent matrices, that is,
and
. A matrix
is called
-(anti-)reflexive if
(or
). For
and
to be Hermitian, they are
-potent matrices if and only if
P and
Q are idempotent (i.e.,
,
) when
k is odd, or tripotent (i.e.,
,
) when
k is even. Moreover, there exist
and
such that
if
k is odd, and
if
k is even, where
and
.
Theorem 50 (
-(anti-)reflexive solutions for (
1) over
. [
54])
. Given , , , . Let and be Hermitian and -potent with . For U and V are given in (
20)
, let , , , be defined inwhere , , , and . Then, we have the following results. The system (
1)
is consistent for -reflexive X if and only if In this case, the general solution iswhere is arbitrary. Let be the set of all -reflexive solutions to (
1)
and E be a given matrix in . Partitionwith . Then,has an only solution , which can be expressed as Assume that the SVDs of , is expressed aswhere , , , and are unitary matrices, , , , , , , , . Then, the least norm least squares solution can be expressed aswhere , and is an arbitrary matrix. Theorem 51 (
-(anti-)reflexive solutions for (
1) over
. [
54])
. Given , , , , and are Hermitian and -potent with . For U and V are given in (
20)
, let , , , be defined inwhere , , , , , , , and . Then, the system (
1)
is consistent for -reflexive X if and only if In this case, the general solution iswhere ,, and , are arbitrary with suitable orders. Let be the set of all -reflexive solutions to (
1)
and let E be a given matrix in . Partitionwith , . Then, has a unique solution , which can be expressed aswhere and . Remark 15. The -reflexive least squares problem can be reduced similarly to Theorem 50(c); hence, the conclusion is omitted.
4.6. Different Types of Conjugate Solutions
Chang et al. have presented the
-conjugate solution to the linear equation system (
1) [
55]. A matrix
is called an
R-conjugate matrix if it satisfies
, where
R is a nontrivial involution (i.e.
,
). A matrix
is called an
-conjugate matrix if it satisfies
, where
R and
S are nontrivial involutions. The sets of
R-conjugate and
-conjugate matrices are denoted by
and
, respectively. For nontrivial involution matrices
and
, there exists
Denote
,
. The results for the solutions in
and
to the system (
1) are presented below.
Theorem 52 (
-conjugate solutions for (
1) over
. [
55])
. Given , nontrivial involutions R and S. Suppose that and , whereLetwhere . Denote Assume that the SVDs of and arewherewith and . Then, the system (
1)
has a solution in if and only if In which case, the general -conjugate solution to (
1)
can be represented aswhere is arbitrary. For a given , let . Denote the -conjugate solution set of (
1)
is . If is nonempty, then the approximation problem has a unique solution is the form of When (
1)
does not have an -conjugate solution, then the least squares solution of (
1)
can be expressed aswhere and is an arbitrary matrix. The unique least squares least norm solution of (
1)
is Two years later, Chang et al. extended the results in [
55] to consider the Hermitian
R-conjugate solutions. They provided the necessary and sufficient conditions for the existence of the Hermitian
R-conjugate solution to the system of complex matrix equations
and
and presented an expression for the Hermitian
R-conjugate solution to this system when the solvability conditions are satisfied. In addition, the solution to an optimal approximation problem was obtained. Furthermore, the least squares Hermitian
R-conjugate solution with the least norm for this system was also considered [
56].
Theorem 53 (Hermitian
R-conjugate solutions for (
1) over
. [
56])
. For given , and . Given a nontrivial symmetric involution matrix , which can be expressed aswhere and satisfying . Denote . Let , and , whereAssume that the SVD of iswhere and are orthogonal matrices, . The system (
1)
has a Hermitian R-conjugate solution in if and only if In that case, (
1)
has the general Hermitian R-conjugate solutionwhere G is a arbitrary symmetric matrix. For , let . The system (
1)
has Hermitian R-conjugate solutions, then the optimal approximation problem has a unique Hermitian R-reflexive solution of (
1)
as The least squares Hermitian R-conjugate solution of (
1)
can be expressed aswhere is an arbitrary symmetric matrix. The least norm least squares Hermitian R-conjugate solution of (
1)
can be expressed as 4.7. Conjugate Class Solutions
Recall that matrices are in the same *congruence class if there exists a nonsingular matrix such that .
Zheng first considered the *congruence class of the solutions of (
1) [
57].
Theorem 54 (*congruence class solutions for (
1) over
. [
57])
. Let , . The GSVD of the matrices A and B is given bywhere , are unitary matrices, is a nonsingular matrix, andare block matrices with positive diagonal matrices , . , . Block and into suitable size as the form ofThe system (
1)
is solvable if and only if The general form of a solution of (
1)
iswhere and are an arbitrary. Obviously, X is *congruent to Later, Zhang presented the following result.
Theorem 55 (*congruence class solutions for (
1) over
. [
58])
. Let and . Assume that the GSVD of A and can be expressed aswhere and are unitary matrices, is nonsingular matrix, where with , , and , . Denote The system (
1)
has a solution in if and only if In that case, the general solutions of (
1)
arewhere , , , and are arbitrary. For arbitrary , , , and , there exists a solution in of (
1)
, which is *congruent to There exists a minimum norm solution in of (
1)
, which is *congruent to Remark 16. Theorem 55 differs from Theorem 54 in both the approach to decomposition and the way solutions are expressed. Specifically, Theorem 55 provides an extended formulation that generalizes the results in Theorem 54, offering a more comprehensive method for decomposing the matrix equations and presenting solutions in a broader context.
The following theorem shows the corresponding least squares and least squares least norm solutions through GSVD.
Theorem 56 (Least squares *congruence class solutions for (
1) over
. [
58])
. Let , , and . There exists a unitary matrix and nonsingular matrices and , such that the GSVD of matrix pair is given aswhere arewith , Denote the suitable block matrices with the forms The least square solutions to (
1)
arewhere , , , , , and are arbitrary, , , . For arbitrary , , , , , and , there exists a least square solution in of (
1)
, which is *congruent towhere , , . There exists a minimum norm least square solution in of (
1)
, which is *congruent towhere , , . 4.8. (Anti-)Hermitian (Anti-)Hamiltonian Solutions
Hamiltonian matrices play a crucial role in various engineering applications, particularly in solving Riccati equations. Yu et al. studied four extended Hamiltonian solutions of the system (
1) [
59].
Table 1 outlines the definitions of anti-symmetric orthogonal matrices and (anti-)Hermitian generalized (anti-)Hamiltonian matrices, with
representing a non-trivial anti-symmetric orthogonal matrix and
denoting the (anti-)Hermitian generalized (anti-)Hamiltonian matrix.
For
, the EVD of
J can be expressed as
where
is a unitary matrix [
59].
Next, we present the necessary and sufficient conditions for the (anti-)Hermitian generalized (anti-)Hamiltonian solutions to the system (
1) along with corresponding expressions. Additionally, for a given
, we consider the optimization problem
, where
X satisfies (
1).
Theorem 57 (
solutions for (
1) over
. [
59])
. Given , let the decomposition of be (
22)
. Partitionwhere and . Denote Then, the system (
1)
has a solution if and only ifin which case the Hermitian generalized Hamiltonian solution to (
1)
can be expressed aswhereand is arbitrary. For a given , let Assume that the system (
1)
has a solution . Then, the optimization problem has a unique solution of (
1)
if and only ifin which case the unique solution X can be expressed aswhere Let the SVDs of and be given bywhere Then, the least squares Hermitian generalized Hamiltonian solution to (
1)
can be described aswherewith and arbitrary . Theorem 58 (
solutions for (
1) over
. [
59])
. Given , let the decomposition of be (
22)
. The matrices , and respectively have the partitions as in (
23)
. DenoteLet the SVDs of and bewhere Setwhere Then, (
1)
has a solution if and only ifin which case the Hermitian generalized anti-Hamiltonian solution to (
1)
can be described aswhere and are arbitrary Hermitian matrices. For a given and the system (
1)
has a solution , letwhere are Hermitian. Then, the optimization problem has a unique solution of (
1)
aswhere Theorem 59 (
solutions for (
1) over
. [
59])
. Given , let the decomposition of be (
22)
. The matrices , and respectively have the partitions as in (
23)
. DenoteLet the SVDs of and be, respectively,where are unitary and Setwhere Then, (
1)
has a solution if and only ifin which case the anti-Hermitian generalized Hamiltonian solution to (
1)
can be described aswhereand are arbitrary anti-Hermitian. For a given , letwhere are anti-Hermitian. Assume that the system (
1)
has a solution . Then, the optimization problem has a unique solution of (
1)
, satisfyingwhere Theorem 60 (
solutions for (
1) over
. [
59])
. Given , let the decomposition of be (
22)
. The matrices , and respectively have the partitions as in (
23)
. Denote Then, (
1)
has a solution if and only ifin which case the anti-Hermitian generalized anti-Hamiltonian solution to (
1)
can be expressed aswhereand is arbitrary. For a given , let If the system (
1)
has a solution , the optimization problem has a unique solution of (
1)
if and only ifin which case the unique solution X can be expressed aswhere Let the SVDs of and be as given inwhere Then, the least squares Hermitian generalized Hamiltonian solution to (
1)
can be described aswherewith and is arbitrary. This section introduces matrix decomposition methods for solving special solutions of system (
1), including various symmetric solutions, orthogonal solutions over the real field, unitary solutions over the complex field, inequality-constrained solutions, real-positive definite and real-semi-positive definite solutions, reflexive solutions, various conjugate solutions, and Hamiltonian-type solutions.
5. The System (1) over Dual Numbers
In 1873, Clifford introduced dual numbers for studying non-Euclidean geometry [
60]. The set of dual numbers is typically denoted by
For two dual numbers and , the arithmetic operations for dual numbers are defined as follows:
Equality: .
Addition: .
Multiplication: .
A matrix whose elements are dual numbers is called a dual matrix. Specifically, the set of all
real dual matrices is given by
The operational rules for dual matrices follow those of dual numbers. Dual matrices have significant applications in kinematic analysis and robotics. The solution of systems of linear dual equations is a crucial task in various fields, such as synthesis problems and sensor calibration [
61].
Recently, the existence of general solutions and corresponding expressions, along with minimal norm solutions for (
1) over dual numbers has been investigated. We present the following theorem.
Theorem 61 (General solutions for (
1) over
. [
62])
. Assume that dual matrices , , and , where , , , (). Suppose that the SVDs of the matrices and arewhere , , , , , , , with , , and . Let the partitions of the matrices , , , , and be given by Then, (
1)
is solvable over dual numbers if and only if In this case, the solution set of (
1)
over dual numbers can be expressed aswhereand , are arbitrary matrices. If the conditions (
24)
are satisfied, then the solution of with being the dual number solution of (
1)
is given by wherewhere is satisfyingwhere Remark 17. In 2024, Fan presented an alternative form of Theorem 61 using the Moore–Penrose inverse instead of block matrices, which also requires the SVD form of and [63]. In fact, Theorem 61, when applied with the SVD, provides the specific form of the Moore–Penrose inverse, which may be more efficient in practical computations. In this section, we introduced the solution of the system (
1) in terms of dual quaternions. A more general form involving dual quaternions will be discussed in the next section.
6. The System (1) over Quaternions
Since Hamilton’s discovery of quaternions in 1843 [
64], they have become a widely used tool for representing concepts across algebra, analysis, topology, and physics. Additionally, quaternion matrices have garnered significant attention in fields such as computer science, quantum physics, signal processing, and color image processing [
65,
66].
In 1849, Cockle introduced split quaternions [
67]. The algebra of split quaternions is a four-dimensional Clifford algebra that is associative and noncommutative, but it has zero divisors, nilpotent elements, and nontrivial idempotents. As a result, the algebraic structure of split quaternions, denoted as
, is more complex than that of real quaternions,
. Despite this complexity, the unique algebraic properties of split quaternions make them a valuable tool in quantum mechanics and geometry [
68,
69].
In 1873, Clifford extended the concepts of dual numbers and dual quaternions [
60]. Dual quaternions have since become widely used in robot kinematics and unmanned aerial vehicle formation control due to their ability to represent the motion of rigid bodies in 3D space [
70,
71,
72,
73]. Similarly, the dual split quaternion can also be defined.
The system (
1) over quaternion algebra, split quaternion algebra, and dual quaternion algebra has also been the focus of several scholars. Comparatively, since the algebraic structure of quaternions is well-understood, and the definitions of generalized inverses and rank have been extended to quaternion matrices, the system (
1) has been thoroughly studied over quaternions. Since 2005, when Wang first proposed the general solution to the extended form of the system
of the system (
1), relevant results for the bi-symmetric, centro-symmetric, symmetric and skew-antisymmetric,
-reflexive solutions and reducible solutions have been successively presented [
74,
75,
76,
77]. The study of the split quaternion matrix equation typically relies on the real or complex representation of the split quaternion, or vectorization operators. However, recent work by Jiang on split quaternion matrix SVD and generalized inverses has enabled the consideration of more diverse approaches [
78,
79]. Dual quaternions are more intricate, and only Xie has explored the system (
1) over dual quaternions [
80]. Yang et al. also considered the results of the system (
1) over dual split quaternion tensors [
81]. The following details are introduced.
6.1. The System (1) over Quaternions
Denote the set of all real quaternions by
where
i,
j, and
k are the quaternion units. For
,
is the conjugate of
a. The set of all
quaternion matrices is denoted by
. For a quaternion matrix
, the transpose conjugate of
A is expressed as
. The Moore–Penrose inverse of
A is denoted as
, satisfying the same equations in the definition of complex Moore–Penrose. The two orthogonal projectors
and
are defined as
The rank of
A, denoted by
, is defined as the dimension of
, where
is the column right space of
A [
82,
83].
Using the results of the system (
25) and the properties of the rank of quaternion matrix equations, the general solution of system (
1) over quaternions is presented below.
Theorem 62 (General solutions for (
1) over
. [
74])
. For given , , and , then there exists three conditions, one of which is equivalent to (
1)
is consistence over ,
,
.
In this case, the general solution of (
1)
can be in the form ofwhere Y is an arbitrary matrix over with appropriate order. Based on Theorem 62, Kyrchei investigated the row–column determinant expression of the solution of system (
1) over quaternions [
84].
Let denote the symmetric group on . For a quaternion matrix , the row and column determinants are defined as follows:
Row determinant: The
i-th row determinant of
for all
is defined as
where
and
for all
and
.
Colimn determinant: The
j-th column determinant of
for all
is defined as
where
and
for all
and
.
For , let and with . The collection of strictly increasing sequences of k integers chosen from is denoted by For a fixed and , let Assume . Let be a principal submatrix of A whose rows and columns are indexed by . If is Hermitian, then denotes the corresponding principal minor of . Let be the j-th column and be the i-th row of A. Suppose denotes the matrix obtained from A by replacing its j-th column with the column b, and denotes the matrix obtained from A by replacing its i-th row with the row b.
Theorem 63 (General solutions using row and column determinants for (
1) over
. [
84])
. Let , , , , , , . Denote and . Assume that and . Quaternion matrix as the solution of (
1)
has the following determinantal representation. If and , then If and , then If and , then If and , then The following presents some special forms of symmetric solutions over quaternions and related results.
Table 2 outlines the definitions of several kinds of symmetric matrices, where
,
,
, and
is the conjugate of the quaternion
.
It is worth noting that centrosymmetric, symmetric and skew-antisymmetric, and -(skew)symmetric matrices do not necessarily need to be square.
Next, we will sequentially present the conclusions regarding the above special solutions of the system (
1) over quaternions.
Theorem 64 (Bisymmetric solutions for (
1) over
. [
74])
. Let , , and , where when and otherwise. Denotewhen , orwhen There exists block matriceswhere and when ; , , and when . Let ,Then, the system (
1)
has a bisymmetric solution over quaternions if and only ifin which case, the general bisymmetric solution can be expressed aswherewhere is an arbitrary matrix over with compatible dimension. Remark 18. In 2015, Yuan et al. considered the least squares η-bi-Hermitian solution for another linear system [85]. Theorem 65 (Cetrosymmetric solutions for (
1) over
. [
74])
. For and , denoteThen, the system (
1)
has a centrosymmetric solution if and only if In that case, the centrosymmetric solution can be expressed aswherewith arbitrary . Theorem 66 (Symmetric and skew-antisymmetric solutions for (
1) over
. [
75])
. Let , , , where when and otherwise. Denotewhen , orwhen There exists block matriceswhere and when ; , , and when . LetThen, the system (
1)
has symmetric and skew-antisymmetric solutions if and only if In which case, the general symmetric and skew-antisymmetric solution can be expressed aswherewith W is an arbitrary matrix over with compatible dimension. Theorem 67 (
-(skew-)symmetric solution for (
1) over
. [
76])
. Let , , , , and satisfying The EVDs of P and Q can be written as the form ofwhere U and V are invertible. Denotewhere Then, the system (
1)
has a -(skew-)symmetric solution if and only ifor equivalently The general -symmetric solution of (
1)
can be expressed aswhere , are arbitrary matrices over with appropriate sizes. The general -skew-symmetric solution of (
1)
can be expressed aswhere , are arbitrary matrices over with appropriate sizes. Subsequently, we introduce the extreme rank
-(skew-)symmetric solutions of the system (
1) over quaternions.
Theorem 68 (Extreme rank
-(skew-)symmetric solutions for (
1) over
. [
76])
. Suppose that the system (
1)
has a -(skew-)symmetric solutions X and is the set of all -(skew-)symmetric solutions of (
1)
. Denote The maximal rank of is The corresponding general expression of X iswhereand is chosen such that . The minimal rank of is The corresponding general expression of X iswherefor and is an arbitrary quaternion matrix with appropriate sizes. The maximal rank of the -skewsymmetric solution of (
1)
is The general expression of X attaining the maximal rank can be expressed aswhereand is chosen such that . The minimal rank of the -skewsymmetric solution of (
1)
isor The general expression of X attaining the minimal rank can be expressed aswherefor , and is an arbitrary quaternion matrix with appropriate size. At the end of this section, we introduce the reducible solution of the system (
1) over quaternions.
A matrix
is called reducible if there exists a permutation matrix
K such that
where
and
are square matrices of order at least 1 over
. Moreover, if the order of
is
k (
), we call
A to be
k-reducible with respect to the permutation matrix
K.
Theorem 69 (Reducible solutions for (
1) over
. [
77])
. Let , be known, unknown, be a permutation matrix, . Denotewhere Assume that M, N, P, Q, E, F, G, S, and T are defined asThen, the system (
1)
has a k-reducible solution with the permutation matrix K if and only if one of the following two statements holds. In that case, the k-reducible solution X of system (
1)
with respect to K can be expressed aswherewith , , , , , , , and being arbitrary matrices over with appropriate sizes. Remark 19. Due to the definition of reducible matrices, considering the reducible solution of the system (
1)
is actually equivalent to considering the general solution of a more complex system The proof process of Theorem 69 follows this approach as well.
Remark 20. In fact, solving the reducible solution over quaternions is closely related in form to solving the general solution over dual quaternions.
6.2. The System (1) over Split Quaternions
The set of real quaternions form a noncommutative division algebra. In 1849, Cockle introduced split quaternions:
where
Split quaternions have zero factors, which gives a more complex algebraic structure than . Solving the split quaternion matrix equation mainly relies on real representation and complex representation, with the real representation having better structure-preserving properties and performing better in numerical examples.
Si et al. designed several real representations of the split quaternion matrix to establish sufficient and necessary conditions for the existence of the general,
-(anti-)conjugate, and
-(anti-)Hermitian solutions. Further, they derived expressions of the corresponding solutions when the system is solvable [
86].
For any matrix
, it can be represented uniquely as
, where
. The three corresponding
-conjugates (
) are defined as
Let
be the usual conjugate transpose of
A. Then, the three other
-conjugate transposes (
) of
A are defined as follows:
For
,
,
A is called
-(anti-)Hermitian if
[
87].
Given , , , we define the following four real representations of A:
Using the real representations above, reference [
86] presents the following results.
Theorem 70 (General solutions for (
1) over
. [
86])
. Consider matrices , , , and . Then there exist two equivalent statements for the system (
1)
that has a solution . The system of real matrix equationshas a solution .
If the system (
28)
is consistent, thenwherefor arbitrary . Theorem 71 (
-conjugate solutions for (
1) over
. [
86])
. Let , , , and . Then, the following statements are equivalent: The system of split quaternion matrix equations (
1)
has a solution , . The system of real matrix equations (
28)
has a generalized (anti-)reflexive solution . hold, wherewhen ;when ;when . If the system (
1)
is consistent, thenwhere with arbitrary is a generalized (anti-)reflexive matrix. Theorem 72 (
-Hermitian solutions for (
1) over
. [
86])
. If , , the following statements are equivalent: The system of split quaternion matrix equations (
1)
has a solution , . The system of real matrix equationhas a (-)symmetric solution . and is a symmetric matrix, where In this case, the general η-(anti-)Hermitian solution to the system (
1)
can be expressed aswhereand is an arbitrary matrix, Remark 21. More complex linear matrices or even tensor equations can be solved by means of complex representations or semi-tensor products of split quaternions, as detailed in references [88,89,90,91]. 6.3. The System (1) over Dual Quaternions
The collection of dual quaternions is expressed as
where
and
represent the standard part and the infinitesimal part of
c, respectively [
92]. We denote
as the set of all
matrices over
.
For the general solution of the system (
1) over dual quaternions, Xie et al. recently presented the following theorem.
Theorem 73 (General solutions for (
1) over
. [
80])
. Let , , and be given. SetThen, the system (
1)
is consistent if and only ifor equivalently,and equations hold. In such circumstances, the general solution of the system (
1)
can be expressed as , whereand are arbitrary. 6.4. The System (1) over Dual Split Quaternions
Yang et al. studied the system (
1) over the dual split quaternion tensor and provided the general solution as well as the existence conditions and expressions for the
-Hermitian solution [
81].
For a multidimensional array tensor
with
entries, the general inverse of
can also be extended from the general inverse of a matrix [
93], denoted as
. Let
represent the sets of the order
M tensors with
dimensions over the split quaternion algebra
. The identity tensor
has all zero entries, except for the elements
.
denotes the zero tensor whose elements are all zero. Define
The sets of dual split quaternion and dual split quaternion tensors are represented as follows [
94]:
Let
and
. Then, we can define the Einstein product of tensors
and
through the operation
as
The real representations of the split quaternion tensor are of the same form as (
27), with the only difference being the use of tensor notation.
For the system
the following conclusions hold.
Theorem 74 (General solutions for (
29) over
. [
81])
. Suppose that , , , and . DenoteThen, the following descriptions are equivalent: The system of dual split quaternion tensor equation (
29)
is solvable. The system of tensor equations:is consistent. Based on these circumstances, the general solution of the system (
29)
can be represented as , wherewith arbitrary . Remark 22. The (η-)Hermitian solutions for the system (29) can be derived by selecting different types of real representations of split quaternion tensors [81]. We present the general solution of the system (
1) over quaternions, including the determinant expression for the general solution. It also covers bi-symmetric solutions, centrosymmetric solutions, symmetric and skew-symmetric solutions,
-(skew-)symmetric solutions, extreme rank
-(skew-)symmetric solutions, and reducible solutions. Additionally, the general solution over split quaternions,
-(anti-)conjugate solutions, and
-(anti-)Hermitian solutions are discussed. The general solution over dual quaternion matrices and split dual quaternion tensors are also examined. The existence conditions and corresponding expressions for these solutions are provided.
7. Applications
The system (
1) has broad applications across various fields. This section focuses on its use in encrypting and decrypting color images and videos.
In image processing, the system (
1) can be applied to various tasks, such as image transformation, filtering, and reconstruction. Matrix equations are used to model the transformation or processing of an image, where
A and
B represent certain image transformations,
X is the unknown matrix to be solved, and
C and
D represent the image before and after processing, or certain features of the image.
In color images, a pure imaginary quaternion can represent the three color channels—red, green, and blue—using , thus effectively representing a pixel. By utilizing quaternion matrices or dual quaternion matrices, which can represent even more information, we can process color images in a highly efficient manner. This approach allows us to simultaneously process multiple color channels of the image.
We present two examples of using the system (
1). The first example involves using dual quaternions for encrypting and decrypting images, as shown in
Figure 2 [
80]. The original, encrypted, and decrypted images are displayed in
Figure 3.
The other example demonstrates the application of the dual split quaternion tensor equations in color video processing [
81]. The basic framework is the same as in image processing, but the tensor, as a high-dimensional matrix, can directly represent video. The results for several frames are shown here, as illustrated in
Figure 4.
These two examples demonstrate that by using the solution method of the system (
1), color images and videos can be effectively encrypted and decrypted, ensuring the security and accuracy of the communication process.
8. Conclusions
This paper presents a comprehensive review of the system (
1), emphasizing its essential role in a wide range of applications. The discussion includes generalized inverse methods for obtaining both general and specialized solutions, such as Hermitian solutions, non-negative definite solutions, and maximal and minimal rank solutions. The theory is further extended to more advanced algebraic structures, including Hilbert spaces, Hilbert
-modules, and general rings, where specialized solving techniques can be applied. Matrix decomposition methods, such as eigenvalue decomposition, singular value decomposition, and generalized singular value decomposition, are explored for their effectiveness in solving the linear matrix equation systems. Additionally, the paper addresses solutions within specialized algebraic structures like dual numbers and various quaternions. At the end, examples of applications of the system (
1) in color image and video processing are presented. This review aims to comprehensively summarize the research on various solutions to the system (
1) across different algebraic structures. However, the differing research perspectives and the vast amount of literature may have resulted in some references being overlooked. Nonetheless, this does not detract from the primary value of the survey.
Future research may focus on addressing the computational challenges associated with large-scale matrix systems, as generalized inverses and matrix decomposition techniques can be computationally intensive. Therefore, finding numerical solutions to the system (
1) is an important research direction. Inspired by [
95], leveraging neural networks and other methods to explore these solutions could be a promising approach. Moreover, despite the widespread use of tensors in many fields due to their high-dimensional properties, exploration of the system (
1) within the tensor framework has been relatively restricted. Consequently, continuing to study the system (
1) within the context of tensors presents an exciting opportunity for future research. Lastly, it is worth noting that, given the extensive applications of dual quaternions, studying various special solutions to the system (
1) in the context of dual quaternions, dual generalized commutative quaternions, and dual split quaternions—such as minimum norm solutions, Hermitian solutions, and reflexive solutions—presents an important development direction that warrants future attention. These developments are expected to further promote the in-depth application of the system (
1) in areas such as control theory, optimization, image processing, system identification, and robotics.