Next Article in Journal
An Eye Tracking Study on Symmetry and Golden Ratio in Abstract Art
Previous Article in Journal
Unleashing the Power of Decoders: Temporal Knowledge Graph Extrapolation with Householder Transformation
Previous Article in Special Issue
A Genetic Algorithm for Three-Dimensional Discrete Tomography
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

The BiCG Algorithm for Solving the Minimal Frobenius Norm Solution of Generalized Sylvester Tensor Equation over the Quaternions

by
Mengyan Xie
1,2,
Qing-Wen Wang
2,3,* and
Yang Zhang
4
1
College of Information Technology, Shanghai Ocean University, Shanghai 201306, China
2
Department of Mathematics and Newtouch Center for Mathematics, Shanghai University, Shanghai 200444, China
3
Collaborative Innovation Center for the Marine Artificial Intelligence, Shanghai 200444, China
4
Department of Mathematics, University of Manitoba, Winnipeg, MB R3T 2N2, Canada
*
Author to whom correspondence should be addressed.
Symmetry 2024, 16(9), 1167; https://doi.org/10.3390/sym16091167
Submission received: 5 August 2024 / Revised: 31 August 2024 / Accepted: 3 September 2024 / Published: 6 September 2024
(This article belongs to the Special Issue Feature Papers in Mathematics Section)

Abstract

:
In this paper, we develop an effective iterative algorithm to solve a generalized Sylvester tensor equation over quaternions which includes several well-studied matrix/tensor equations as special cases. We discuss the convergence of this algorithm within a finite number of iterations, assuming negligible round-off errors for any initial tensor. Moreover, we demonstrate the unique minimal Frobenius norm solution achievable by selecting specific types of initial tensors. Additionally, numerical examples are presented to illustrate the practicality and validity of our proposed algorithm. These examples include demonstrating the algorithm’s effectiveness in addressing three-dimensional microscopic heat transport and color video restoration problems.

1. Introduction

An order N tensor A = a i 1 i N 1 i j I j ( j = 1 , , N ) over a field F is a multidimensional array with I 1 I 2 I N entries in F , where N is a positive integer [1,2]. The set of all such N tensors is denoted by F I 1 × × I N . Over the past few decades, there has been extensive research on tensors, which has been driven by their diverse applications in fields such as physics, computer vision, data mining, and more (see, e.g., [1,3,4,5,6,7,8,9,10,11,12,13,14,15,16,17,18,19,20,21,22,23]).
In this paper, we examine the conditions under which certain tensor equations over quaternions have solutions. This is motivated by the recent research on tensor equations as well as a long research history of solving matrix equations, which are briefly outlined as follows. It is well known that the following Sylvester matrix equation
A X + Y B = C
and its generalized forms have been widely investigated and found numerous applications in many areas.
During the past few decades, many methods have been developed for solving Sylvester-type matrix equations over the quaternion algebra. For example, Kyrchei [24] provided explicit determinantal representation formulas for the solutions to Equation (1). Heyouni et al. [25] presented the SGl-CMRH method and preconditioned framework of this method to solve matrix Equation (1) when X = Y . Zhang [26] investigated the general system of generalized Sylvester quaternion matrix equations. Ahmadi-Asl and Beik [27,28,29] developed efficient iterative algorithms for solving various quaternion matrix equations. Song [30] investigated the general solution to a system of quaternion matrix equation by Cramer’s rule. Wang et al. [31] investigated the solvability of a system of constrained two-sided coupled generalized Sylvester quaternion matrix equations. Zhang et al. [32] derived specific least-squares solutions of the quaternion matrix equation A X B + C X D = E . Meanwhile, Huang et al. [33] applied the modified conjugate gradient method to address the generalized coupled Sylvester conjugate matrix equations.
Quaternions provide more versatility and flexibility than real and complex numbers, especially when dealing with multidimensional problems. This unique property has attracted growing interest among scholars, leading to numerous valuable achievements in quaternion-related research (see, e.g., [34,35,36,37,38,39,40]). The tensor equation is a natural extension of the matrix equation.
In this paper, we examine the following generalized Sylvester tensor equation over H :
X × A ( 1 ) 1 + X × A ( 2 ) 2 + + X × N A ( N ) + Y × B ( 1 ) 1 + Y × B ( 2 ) 2 + + Y × N B ( N ) = C ,
where the tensors A ( n ) , B ( n ) H I n × I n ( n = 1 , 2 , , N ) , and C H I 1 × × I N are given, and the tensors X , Y H I 1 × × I N are unknown. The n-mode product of a tensor X H I 1 × × I N with a matrix A H I n × I n is defined as
X × A n i 1 i n 1 j i n + 1 i N = i n = 1 I n a j i n x i 1 i n 1 i n i n + 1 i N .
Observe that the n-mode product can also be represented using unfolded quaternion tensors:
Y = X × n A Y [ n ] = A X [ n ] ,
where X [ n ] is the mode-n unfolding of X [1]. We address the following problems related to (2):
Problem 1.1: Given the tensor C H I 1 × I 2 × × I N , and the matrices A ( n ) , B ( n ) H I n × I n ( n = 1 , 2 , , N ) , find the tensors X ˜ , Y ˜ H I 1 × I 2 × × I N such that
k = 1 N X ˜ × A ( k ) k + Y ˜ × B ( k ) k C = min X k = 1 N X × k A ( k ) + Y × k B ( k ) C .
Problem 1.2: Let S X Y denote the solution set of Problem 1.1. For given tensors X 0 , Y 0 H I 1 × I 2 × × I N , find the tensors X ˇ and Y ˇ H I 1 × I 2 × × I N such that
X ˇ X 0 + Y ˇ Y 0 = min X S L X X 0 + Y Y 0 .
It is worth emphasizing that the tensor Equation (2) includes several well-studied matrix/tensor equations as special cases. For example, if X and Y in (2) are order 2 tensors, i.e., matrices, then Equation (2) can be reduced to the following extended Sylvester matrix equation
A ( 1 ) X + X ( A ( 2 ) ) T + B ( 1 ) Y + Y ( B ( 2 ) ) T = C .
In the case of B ( n ) = 0 ( n = 1 , 2 , N ) , Equation (2) becomes the following equation
X × 1 A ( 1 ) + X × A ( 2 ) 2 + + X × N A ( N ) = C ,
which has been the subject of extensive research in recent years. For instance, Saberi et al. [41,42] investigated the SGMRES-BTF method and SGCRO-BTF method to solve Equation (3) over R . Wang et al. [43] proposed the conjugate gradient least-squares method to solve Equation (3) over H . Zhang and Wang [44] introduced the tensor formulations of the bi-conjugate gradient (BiCG-BTF) and bi-conjugate residual (BiCR-BTF) methods for solving tensor Equation (3) in the real number field R . Chen and Lu [45] explored a projection method using a Kronecker product preconditioner to solve Equation (3) over R . Karimi and Dehghan [46] introduced the tensor formulation of the global least-squares method for approximating solutions to (3). Additionally, Najafi-Kalyani et al. [47] developed several iterative algorithms based on the global Hessenberg process in their tensor forms to address Equation (3). Considering Equation (3) over R and N = 3 , that is,
X × 1 A ( 1 ) + X × A ( 2 ) 2 + X × 3 A ( 3 ) = C .
It has been shown that Equation (4) plays an important role in finite difference [48], thermal radiation [11], information retrieval [15], finite elements [49], and microscopic heat transport problem [18]. Therefore, our study of Equation (2) will provide a unified treatment for these matrix/tensor equations.
The remainder of this paper is structured as follows. In Section 2, we review key definitions and notations and prove several lemmas related to transforming Equation (2). In Section 3, we develop the BiCG iterative algorithm for solving the quaternion tensor Equation (2) and prove that our algorithm is correct. We also demonstrate that the minimal Frobenius norm solution can be achieved by selecting specific types of initial tensors. Section 4 provides numerical examples to illustrate the effectiveness and applications of the proposed algorithm. Finally, we summarize our contributions in Section 5.

2. Preliminaries

First, we review some notations and definitions. For two complex matrices U = u i j C m × n and V = v i j C p × q , the notation U V = u i j V C m p × n q represents the Kronecker product of U and V.
The operator vec ( · ) is defined as: for a matrix A and a tensor X ,
vec ( A ) = a 1 T , a 2 T , , a n T T , and vec ( X ) = vec ( X [ 1 ] ) ,
respectively, where a k is the kth column of A and X [ 1 ] is the mode-1 unfolding of the tensor X . The inner product of two tensors X , Y H I 1 × × I N is defined as follows:
X , Y = i 1 = 1 I 1 i 2 = 1 I 2 i N = 1 I N x i 1 i 2 i N y ¯ i 1 i 2 i N ,
where y ¯ i 1 i 2 i N represents the quaternion conjugate of y i 1 i 2 i N . If X , Y = 0 , then we say that tensors X and Y are orthogonal. The Frobenius norm of tensor X is of the form: X = X , X .
For any X H I 1 × × I N , it is well known that X can be uniquely represented as X = X 1 + X 2 i + X 3 j + X 4 k , where X i R I 1 × × I N , i = 1 , 2 , 3 , 4 . Next, we define n-mode operators for X i .
Let A ( n ) = A 1 ( n ) + A 2 ( n ) i + A 3 ( n ) j + A 4 ( n ) k ,   B ( n ) = B 1 ( n ) + B 2 ( n ) i + B 3 ( n ) j + B 4 ( n ) k H I n × I n , where A i ( n ) ,   B i ( n ) R I n × I n , i = 1 , 2 , 3 , 4 . For W R I 1 × × I N , we define
L A i ( n ) ( W ) = W × A i ( 1 ) 1 + W × 2 A i ( 2 ) + + W × N A i ( N ) , i = 1 , 2 , 3 , 4 , L B i ( n ) ( W ) = W × B i ( 1 ) 1 + W × 2 B i ( 2 ) + + W × N B i ( N ) , i = 1 , 2 , 3 , 4 .
Next, replacing W in the above equations by X i ’s, we define the following notations:
Γ 1 [ X 1 , X 2 , X 3 , X 4 ] = L A 1 ( n ) X 1 L A 2 ( n ) X 2 L A 3 ( n ) X 3 L A 4 ( n ) X 4 , Γ 2 [ X 1 , X 2 , X 3 , X 4 ] = L A 2 ( n ) X 1 + L A 1 ( n ) X 2 + L A 4 ( n ) X 3 L A 3 ( n ) X 4 , Γ 3 [ X 1 , X 2 , X 3 , X 4 ] = L A 3 ( n ) X 1 L A 4 ( n ) X 2 + L A 1 ( n ) X 3 + L A 2 ( n ) X 4 , Γ 4 [ X 1 , X 2 , X 3 , X 4 ] = L A 4 ( n ) X 1 + L A 3 ( n ) X 2 L A 2 ( n ) X 3 + L A 1 ( n ) X 4 ,
Φ 1 [ X 1 , X 2 , X 3 , X 4 ] = L B 1 ( n ) X 1 L B 2 ( n ) X 2 L B 3 ( n ) X 3 L B 4 ( n ) X 4 , Φ 2 [ X 1 , X 2 , X 3 , X 4 ] = L B 2 ( n ) X 1 + L B 1 ( n ) X 2 + L B 4 ( n ) X 3 L B 3 ( n ) X 4 , Φ 3 [ X 1 , X 2 , X 3 , X 4 ] = L B 3 ( n ) X 1 L B 4 ( n ) X 2 + L B 1 ( n ) X 3 + L B 2 ( n ) X 4 , Φ 4 [ X 1 , X 2 , X 3 , X 4 ] = L B 4 ( n ) X 1 + L B 3 ( n ) X 2 L B 2 ( n ) X 3 + L B 1 ( n ) X 4 ,
The following lemma establishes that the quaternion tensor Equation (2) can be reduced to a system of four real tensor equations.
Lemma 1.
In Equation (2), we assume that A ( n ) = A 1 ( n ) + A 2 ( n ) i + A 3 ( n ) j + A 4 ( n ) k , B ( n ) = B 1 ( n ) + B 2 ( n ) i + B 3 ( n ) j + B 4 ( n ) k H I n × I n , n = 1 , 2 , , N , and C =   C 1 + C 2 i + C 3 j + C 4 k , X = X 1 + X 2 i + X 3 j + X 4 k , Y = Y 1 + Y 2 i + Y 3 j + Y 4 k H I 1 × × I N . Thus, the quaternion Sylvester tensor Equation (2) can be expressed as the following system of real tensor equations
Γ 1 [ X 1 , X 2 , X 3 , X 4 ] + Φ 1 [ Y 1 , Y 2 , Y 3 , Y 4 ] = C 1 , Γ 2 [ X 1 , X 2 , X 3 , X 4 ] + Φ 2 [ Y 1 , Y 2 , Y 3 , Y 4 ] = C 2 , Γ 3 [ X 1 , X 2 , X 3 , X 4 ] + Φ 3 [ Y 1 , Y 2 , Y 3 , Y 4 ] = C 3 , Γ 4 [ X 1 , X 2 , X 3 , X 4 ] + Φ 4 [ Y 1 , Y 2 , Y 3 , Y 4 ] = C 4 ,
where Γ i [ X 1 , X 2 , X 3 , X 4 ] , Φ i [ Y 1 , Y 2 , Y 3 , Y 4 ] ( i = 1 , 2 , 3 , 4 ) are defined by (5) and (6). Furthermore, the system of real tensor Equation (7) is equivalent to the following linear system
[ M A , M B ] z = c ,
where
M A = Kro L A 1 ( n ) Kro L A 2 ( n ) Kro L A 3 ( n ) Kro L A 4 ( n ) Kro L A 2 ( n ) + Kro L A 1 ( n ) + Kro L A 4 ( n ) Kro L A 3 ( n ) Kro L A 3 ( n ) Kro L A 4 ( n ) + Kro L A 1 ( n ) + Kro L A 2 ( n ) Kro L A 4 ( n ) + Kro L A 3 ( n ) Kro L A 2 ( n ) + Kro L A 1 ( n ) ,
M B = Kro L B 1 ( n ) Kro L B 2 ( n ) Kro L B 3 ( n ) Kro L B 4 ( n ) Kro L B 2 ( n ) + Kro L B 1 ( n ) + Kro L B 4 ( n ) Kro L B 3 ( n ) Kro L B 3 ( n ) Kro L B 4 ( n ) + Kro L B 1 ( n ) + Kro L B 2 ( n ) Kro L B 4 ( n ) + Kro L B 3 ( n ) Kro L B 2 ( n ) + Kro L B 1 ( n ) ,
z = vec X 1 vec X 2 vec X 3 vec X 4 vec Y 1 vec Y 2 vec Y 3 vec Y 4 , c = vec C 1 vec C 2 vec C 3 vec C 4 ,
Kro L A i ( n ) = n = 1 N I I N I I n + 1 A i ( n ) I I n 1 I I 1 , i = 1 , 2 , 3 , 4 , Kro L B i ( n ) = n = 1 N I I N I I n + 1 B i ( n ) I I n 1 I I 1 , i = 1 , 2 , 3 , 4 ,
and I ( n ) denotes the identity matrix of size n.
Proof of Lemma 1. 
We apply the definition of n-mode product of the quaternion tensor for (2).
n = 1 N X × n A ( n ) + Y × n B ( n ) = n = 1 N X 1 + X 2 i + X 3 j + X 4 k × n A 1 ( n ) + A 2 ( n ) i + A 3 ( n ) j + A 4 ( n ) k + Y 1 + Y 2 i + Y 3 j + Y 4 k × n B 1 ( n ) + B 2 ( n ) i + B 3 ( n ) j + B 4 ( n ) k = n = 1 N X 1 × n A 1 ( n ) X 2 × n A 2 ( n ) X 3 × n A 3 ( n ) X 4 × n A 4 ( n ) + Y 1 × n B 1 ( n ) Y 2 × n B 2 ( n ) Y 3 × n B 3 ( n ) Y 4 × n B 4 ( n ) + n = 1 N X 1 × n A 2 ( n ) + X 2 × n A 1 ( n ) + X 3 × n A 4 ( n ) X 4 × n A 3 ( n ) + Y 1 × n B 2 ( n ) + Y 2 × n B 1 ( n ) + Y 3 × n B 4 ( n ) Y 4 × n B 3 ( n ) i + n = 1 N X 1 × n A 3 ( n ) X 2 × n A 4 ( n ) + X 3 × n A 1 ( n ) + X 4 × n A 2 ( n ) + Y 1 × n B 3 ( n ) Y 2 × n B 4 ( n ) + Y 3 × n B 1 ( n ) + Y 4 × n B 2 ( n ) j + n = 1 N X 1 × n A 4 ( n ) + X 2 × n A 3 ( n ) X 3 × n A 2 ( n ) + X 4 × n A 1 ( n ) + Y 1 × n B 4 ( n ) + Y 2 × n B 3 ( n ) Y 3 × n B 2 ( n ) + Y 4 × n B 1 ( n ) k = C 1 + C 2 i + C 3 j + C 4 k .
By the definitions of Γ i and Φ i , Equation (7) holds. To show (8), we make use of operator “vec” to Γ 1 [ X 1 , X 2 , X 3 , X 4 ] and Φ 1 [ Y 1 , Y 2 , Y 3 , Y 4 ] , that is,
vec Γ 1 [ X 1 , X 2 , X 3 , X 4 ] = vec L A 1 ( n ) X 1 L A 2 ( n ) X 2 L A 3 ( n ) X 3 L A 4 ( n ) X 4 = vec L A 1 ( n ) X 1 vec L A 2 ( n ) X 2 vec L A 3 ( n ) X 3 vec L A 4 ( n ) X 4 = Kro L A 1 ( n ) vec ( X 1 ) Kro L A 2 ( n ) vec ( X 2 ) Kro L A 3 ( n ) vec ( X 3 ) Kro L A 4 ( n ) vec ( X 4 ) ,
vec Φ 1 [ Y 1 , Y 2 , Y 3 , Y 4 ] = vec L B 1 ( n ) Y 1 L B 2 ( n ) Y 2 L B 3 ( n ) Y 3 L B 4 ( n ) Y 4 = vec L B 1 ( n ) Y 1 vec L B 2 ( n ) Y 2 vec L B 3 ( n ) Y 3 vec L B 4 ( n ) Y 4 = Kro L B 1 ( n ) vec ( Y 1 ) Kro L B 2 ( n ) vec ( Y 2 ) Kro L B 3 ( n ) vec ( Y 3 ) Kro L B 4 ( n ) vec ( Y 4 ) .
Similarly, we have the following results for the rest of Γ i ’s and Φ i ’s:
vec Γ 2 [ X 1 , X 2 , X 3 , X 4 ] = Kro L A 2 ( n ) vec ( X 1 ) + Kro L A 1 ( n ) vec ( X 2 ) + Kro L A 4 ( n ) vec ( X 3 ) Kro L A 3 ( n ) vec ( X 4 ) ,
vec Φ 2 [ Y 1 , Y 2 , Y 3 , Y 4 ] = Kro L B 2 ( n ) vec ( Y 1 ) + Kro L B 1 ( n ) vec ( Y 2 ) + Kro L B 4 ( n ) vec ( Y 3 ) Kro L B 3 ( n ) vec ( Y 4 ) ,
vec Γ 3 [ X 1 , X 2 , X 3 , X 4 ] = Kro L A 3 ( n ) vec ( X 1 ) Kro L A 4 ( n ) vec ( X 2 ) + Kro L A 1 ( n ) vec ( X 3 ) + Kro L A 2 ( n ) vec ( X 4 ) ,
vec Φ 3 [ Y 1 , Y 2 , Y 3 , Y 4 ] = Kro L B 3 ( n ) vec ( Y 1 ) Kro L B 4 ( n ) vec ( Y 2 ) + Kro L B 1 ( n ) vec ( Y 3 ) + Kro L B 2 ( n ) vec ( Y 4 ) ,
vec Γ 4 [ X 1 , X 2 , X 3 , X 4 ] = Kro L A 4 ( n ) vec ( X 1 ) + Kro L A 3 ( n ) vec ( X 2 ) Kro L A 2 ( n ) vec ( X 3 ) + Kro L A 1 ( n ) vec ( X 4 ) ,
vec Φ 4 [ Y 1 , Y 2 , Y 3 , Y 4 ] = Kro L B 4 ( n ) vec ( Y 1 ) + Kro L B 3 ( n ) vec ( Y 2 ) Kro L B 2 ( n ) vec ( Y 3 ) + Kro L B 1 ( n ) vec ( Y 4 ) .
By writing up above equations, we obtain the system (8).    □
Lemma 2
([8,50]). Assume that A R m × n , b R m , and the linear matrix equation A x = b has a solution x ˜ R A . Then, x ˜ is the unique solution with the minimum norm for the equation A x = b .
From Lemmas 1 and 2, it is straightforward to observe that the uniqueness of the solution to Equation (2) can be characterized as follows:
Theorem 1. 
The tensor Equation (2) has the unique solution with a minimal Frobenius norm if and only if the matrix Equation (8) has a solution z ˜ R [ M A , M B ] . In this case, z ˜ is the unique solution with a minimum norm of matrix Equation (8).
Given fixed matrices A ( n ) R I n × I n , n = 1 , 2 , , N , we define the following linear operators
L A ( n ) ( X ) = X × A ( 1 ) 1 + X × A ( 2 ) 2 + + X × N A ( N ) , for   any X R I 1 × I 2 × × I N .
Using the property X , Y × n A ( n ) = X × n ( A ( n ) ) T , Y in [10], the following lemma can be easily proven.
Lemma 3.
Let A ( n ) R I n × I n , n = 1 , 2 , , N , X , Y R I 1 × I 2 × × I N . Then
L A ( n ) ( X ) , Y = X , L A ( n ) * ( Y ) ,
where
L A ( n ) * ( Y ) = Y × 1 A ( 1 ) T + Y × A ( 2 ) T 2 + + Y × A ( N ) T N .
Clearly, L A ( n ) defined above is a linear mapping. The following lemma provides the uniqueness of the dual mapping for these kinds of linear mappings.
Lemma 4
([43]). Let N be a linear mapping from tensor space R I 1 × × I N to tensor space R J 1 × × J N . For any tensors X R I 1 × × I N and Y R J 1 × × J N , there exists a unique linear mapping M from tensor space R J 1 × × J N to tensor space R I 1 × × I N such that
N ( X ) , Y = X , M ( Y ) .
Finally, we use linear operators L and L * to describe the inner products involving Γ i and Φ i , which we will use in the next sections.
Lemma 5.
Let Γ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , Φ i [ Z 1 , Z 2 , Z 3 , Z 4 ] ( i = 1 , 2 , 3 , 4 ) be defined by (5) and (6), W i R I 1 × × I N ( i = 1 , 2 , 3 , 4 ) . Then
i = 1 4 Γ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , W i = i = 1 4 Z i , Γ i * [ W 1 , W 2 , W 3 , W 4 ] , i = 1 4 Φ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , W i = i = 1 4 Z i , Φ i * [ W 1 , W 2 , W 3 , W 4 ] ,
where
Γ 1 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L A 1 ( n ) * Z 1 + L A 2 ( n ) * Z 2 + L A 3 ( n ) * Z 3 + L A 4 ( n ) * Z 4 , Γ 2 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L A 2 ( n ) * Z 1 + L A 1 ( n ) * Z 2 L A 4 ( n ) * Z 3 + L A 3 ( n ) * Z 4 , Γ 3 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L A 3 ( n ) * Z 1 + L A 4 ( n ) * Z 2 + L A 1 ( n ) * Z 3 L A 2 ( n ) * Z 4 , Γ 4 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L A 4 ( n ) * Z 1 L A 3 ( n ) * Z 2 + L A 2 ( n ) * Z 3 + L A 1 ( n ) * Z 4 ,
Φ 1 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L B 1 ( n ) * Z 1 + L B 2 ( n ) * Z 2 + L B 3 ( n ) * Z 3 + L B 4 ( n ) * Z 4 , Φ 2 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L B 2 ( n ) * Z 1 + L B 1 ( n ) * Z 2 L B 4 ( n ) * Z 3 + L B 3 ( n ) * Z 4 , Φ 3 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L B 3 ( n ) * Z 1 + L B 4 ( n ) * Z 2 + L B 1 ( n ) * Z 3 L B 2 ( n ) * Z 4 , Φ 4 * [ Z 1 , Z 2 , Z 3 , Z 4 ] = L B 4 ( n ) * Z 1 L B 3 ( n ) * Z 2 + L B 2 ( n ) * Z 3 + L B 1 ( n ) * Z 4 ,
L A i ( n ) * ( X ) = X × ( A i ( 1 ) ) T 1 + X × 2 ( A i ( 2 ) ) T + + X × N ( A i ( N ) ) T , i = 1 , 2 , 3 , 4 , L B i ( n ) * ( X ) = X × ( B i ( 1 ) ) T 1 + X × 2 ( B i ( 2 ) ) T + + X × N ( B i ( N ) ) T , i = 1 , 2 , 3 , 4 .
Proof of Lemma 5. 
For the first part of the equalities, we divide i = 1 4 Γ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , W i into 4 parts by i and then apply Lemma 3 to each part, that is,
Γ 1 [ Z 1 , Z 2 , Z 3 , Z 4 ] , W 1 = L A 1 ( n ) Z 1 , W 1 L A 2 ( n ) Z 2 , W 1 L A 3 ( n ) Z 3 , W 1 L A 4 ( n ) Z 4 , W 1 = Z 1 , L A 1 ( n ) * W 1 Z 2 , L A 2 ( n ) * W 1 Z 3 , L A 3 ( n ) * W 1 Z 4 , L A 4 ( n ) * W 1 ,
Γ 2 [ Z 1 , Z 2 , Z 3 , Z 4 ] , W 2 = L A 2 ( n ) Z 1 , W 2 + L A 1 ( n ) Z 2 , W 2 + L A 4 ( n ) Z 3 , W 2 L A 3 ( n ) Z 4 , W 2 = Z 1 , L A 2 ( n ) * W 2 + Z 2 , L A 1 ( n ) * W 1 + Z 3 , L A 3 ( n ) * W 2 Z 4 , L A 4 ( n ) * W 2 ,
Γ 3 [ Z 1 , Z 2 , Z 3 , Z 4 ] , W 3 = L A 3 ( n ) Z 1 , W 3 L A 4 ( n ) Z 2 , W 3 + L A 1 ( n ) Z 3 , W 3 + L A 2 ( n ) Z 4 , W 3 = Z 1 , L A 3 ( n ) * W 3 Z 2 , L A 4 ( n ) * W 3 + Z 3 , L A 1 ( n ) * W 2 + Z 4 , L A 2 ( n ) * W 3 ,
Γ 4 [ Z 1 , Z 2 , Z 3 , Z 4 ] , W 4 = L A 4 ( n ) Z 1 , W 4 + L A 3 ( n ) Z 2 , W 4 L A 2 ( n ) Z 3 , W 4 + L A 1 ( n ) Z 4 , W 4 = Z 1 , L A 4 ( n ) * W 4 + Z 2 , L A 3 ( n ) * W 4 Z 3 , L A 2 ( n ) * W 4 + Z 4 , L A 1 ( n ) * W 4 .
By adding up the above four parts, we have
i = 1 4 Γ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , W i = i = 1 4 Z i , Γ i * [ W 1 , W 2 , W 3 , W 4 ] ,
where Γ i * [ W 1 , W 2 , W 3 , W 4 ] ’s are defined by (11). Using a similar process, we can obtain the second equality
i = 1 4 Φ i [ Z 1 , Z 2 , Z 3 , Z 4 ] , W i = i = 1 4 Z i , Φ i * [ W 1 , W 2 , W 3 , W 4 ] ,
where Φ i * [ W 1 , W 2 , W 3 , W 4 ] ’s are defined by (12).    □

3. An Iterative Algorithm for Solving the Problems 1.1 and 1.2

The purpose of this section is to propose an iterative algorithm for obtaining the solution of Sylvester tensor Equation (2). As it is well known that the classical bi-conjugate gradient (BiCG) methods for solving nonsymmetric linear systems of equations are feasible and efficient, one may refer to [44,51,52,53,54]. We extend the BiCG method using the tensor format (BTF) for solving Equation (2) and discuss its convergence. Clearly, the tensor Equations (2) and (8) have the same solution from Lemma 1. However, the size of [ M A , M B ] in Equation (8) is usually too large to save computation time and memory space. Beik et al. [55] demonstrated that algorithms using tensor formats generally outperform their classical counterparts in terms of efficiency. Inspired by these issues, we propose the following least-squares algorithm, formulated in tensor format, for solving tensor Equation (2):
Note that Γ i , Φ i and Γ i * , Φ i * are defined by (5), (6) and (11), (12), respectively. Next, we discuss some bi-orthogonality properties of Algorithm 1.
Algorithm 1 BiCG-BTF method for solving Equation (2).
Input:   A i ( n ) , B i ( n ) R I n × I n , C i R I 1 × × I N , n = 1 , 2 , , N ; i = 1 , 2 , 3 , 4 .
Output: The norm i = 1 4 R i ( · ) and the solutions X i ( · ) , Y i ( · ) , i = 1 , 2 , 3 , 4 .
Initialization:  X i ( 1 ) , Y i ( 1 ) R I 1 × × I N , i = 1 , 2 , 3 , 4 ;  
(i)
Compute
R i ( 1 ) : = C i Γ i [ X 1 ( 1 ) , X 2 ( 1 ) , X 3 ( 1 ) , X 4 ( 1 ) ] Φ i [ Y 1 ( 1 ) , Y 2 ( 1 ) , Y 3 ( 1 ) , Y 4 ( 1 ) ] , i = 1 , 2 , 3 , 4 ;
Set R i * ( 1 ) : = R i ( 1 ) ; P i ( 1 ) : = R i * ( 1 ) ; P i * ( 1 ) : = P i ( 1 ) ;
Compute the norm: R n o r m : = i = 1 4 R i ( 1 ) , i = 1 , 2 , 3 , 4 ;
Set k : = 1 ;
(ii)
If R n o r m = 0 , then stop;
(iii)
Otherwise, compute
Q i x ( k ) : = Γ i [ P 1 ( k ) , P 2 ( k ) , P 3 ( k ) , P 4 ( k ) ] , i = 1 , 2 , 3 , 4 ;
Q i y ( k ) : = Φ i [ P 1 ( k ) , P 2 ( k ) , P 3 ( k ) , P 4 ( k ) ] , i = 1 , 2 , 3 , 4 ;
α ( k ) : = i = 1 4 R i ( k ) , R i * ( k ) / i = 1 4 P i * ( k ) , Q i x ( k ) + Q i y ( k ) ;
X i ( k + 1 ) : = X i ( k ) + α ( k ) P i ( k ) , i = 1 , 2 , 3 , 4 ;
Y i ( k + 1 ) : = Y i ( k ) + α ( k ) P i ( k ) , i = 1 , 2 , 3 , 4 ;
R i ( k + 1 ) : = R i ( k ) α ( k ) ( Q i x ( k ) + Q i y ( k ) ) , i = 1 , 2 , 3 , 4 ;
Q i x * ( k ) : = Γ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ] , i = 1 , 2 , 3 , 4 ;
Q i y * ( k ) : = Φ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ] , i = 1 , 2 , 3 , 4 ;
R i * ( k + 1 ) : = R i * ( k ) α ( k ) ( Q i x * ( k ) + Q i y * ( k ) ) , i = 1 , 2 , 3 , 4 ;
β ( k ) = i = 1 4 R i ( k + 1 ) , R i * ( k + 1 ) / i = 1 4 R i ( k ) , R i * ( k ) ;
P i ( k + 1 ) : = R i ( k + 1 ) + β ( k ) P i ( k ) , i = 1 , 2 , 3 , 4 ;
P i * ( k + 1 ) : = R i * ( k + 1 ) + β ( k ) P i * ( k ) , i = 1 , 2 , 3 , 4 ;
R n o r m = i = 1 4 R i ( k ) , i = 1 , 2 , 3 , 4 ;
(iv)
Set k : = k + 1 , go to (ii);
Theorem 2. 
Assume that iterative sequences { R i ( k ) } , { R i * ( k ) } , { P i ( k ) } , { P i * ( k ) } { Q i x ( k ) } and { Q i y ( k ) } ( i = 1 , 2 , 3 , 4 ) are generated by Algorithm 1. Thus, we obtain
i = 1 4 R i ( l ) , R i * ( m ) = 0 , l m ,
i = 1 4 Q i x ( l ) + Q i y ( l ) , P i * ( m ) = 0 , l m ,
i = 1 4 R i ( l ) , P i * ( m ) = 0 , l > m .
Proof of Theorem 2. 
We apply mathematics induction on k. Let us consider 1 m < l k first.
When k = 2 , the conclusion holds, as the following calculations show:
i = 1 4 R i ( 2 ) , R i * ( 1 ) = i = 1 4 R i ( 1 ) α 1 Q i x ( 1 ) + Q i y ( 1 ) , R i * ( 1 ) = i = 1 4 R i ( 1 ) , R i * ( 1 ) i = 1 4 R i ( 1 ) , R i * ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) = 0 ,
and
i = 1 4 Q i x ( 2 ) + Q i y ( 2 ) , P i * ( 1 ) = i = 1 4 Γ i [ P 1 ( 2 ) , P 2 ( 2 ) , P 3 ( 2 ) , P 4 ( 2 ) ] , P i * ( 1 ) + i = 1 4 Φ i [ P 1 ( 2 ) , P 2 ( 2 ) , P 3 ( 2 ) , P 4 ( 2 ) ] , P i * ( 1 ) = i = 1 4 P i ( 2 ) , Γ i * [ P 1 * ( 1 ) , P 2 * ( 1 ) , P 3 * ( 1 ) , P 4 * ( 1 ) ] + i = 1 4 P i ( 2 ) , Φ i * [ P 1 * ( 1 ) , P 2 * ( 1 ) , P 3 * ( 1 ) , P 4 * ( 1 ) ] = i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + β ( 1 ) P i ( 1 ) , Q i x * ( 1 ) + i = 1 4 R i ( 2 ) , Q i y * ( 1 ) + β ( 1 ) P i ( 1 ) , Q i y * ( 1 ) = i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + i = 1 4 R i ( 2 ) , R i * ( 1 ) α ( 1 ) ( Q i x * ( 1 ) + Q i y * ( 1 ) ) i = 1 4 R i ( 1 ) , R i * ( 1 ) P i ( 1 ) , Q i x * ( 1 )
+ i = 1 4 R i ( 2 ) , Q i y * ( 1 ) + i = 1 4 R i ( 2 ) , R i * ( 1 ) α ( 1 ) ( Q i x * ( 1 ) + Q i y * ( 1 ) ) i = 1 4 R i ( 1 ) , R i * ( 1 ) P i ( 1 ) , Q i y * ( 1 ) = i = 1 4 R i ( 2 ) , Q i x * ( 1 ) i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + Q i y * ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) P i ( 1 ) , Q i x * ( 1 ) + i = 1 4 R i ( 2 ) , Q i y * ( 1 ) i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + Q i y * ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) P i ( 1 ) , Q i y * ( 1 ) = i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + Q i y * ( 1 ) i = 1 4 R i ( 2 ) , Q i x * ( 1 ) + Q i y * ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) i = 1 4 P i * ( 1 ) , Q i x ( 1 ) + Q i y ( 1 ) = 0 .
It has been found i = 1 4 R i ( 2 ) , P i * ( 1 ) = 0 clearly. Now, assume that (13) and (14) hold for 1 m < l k ( k > 2 ). Then,
i = 1 4 R i ( k + 1 ) , P i * ( m ) = i = 1 4 R i ( k ) α ( k ) Q i x ( k ) + Q i y ( k ) , P i * ( m ) = i = 1 4 R i ( k ) , P i * ( m ) α ( k ) Q i x ( k ) + Q i y ( k ) , P i * ( m ) = 0 ,
and
i = 1 4 R i ( k + 1 ) , P i * ( k ) = i = 1 4 R i ( k ) α ( k ) Q i x ( k ) + Q i y ( k ) , P i * ( k ) = i = 1 4 R i ( k ) , P i * ( k ) α ( k ) Q i x ( k ) + Q i y ( k ) , P i * ( k ) = i = 1 4 R i ( k ) , R i * ( k ) + β ( k 1 ) R i ( k ) , P i * ( k 1 ) i = 1 4 R i ( k ) , R i * ( k ) i = 1 4 P i * ( k ) , Q i x ( k ) + Q i y ( k ) i = 1 4 P i * ( k ) , Q i x ( k ) + Q i y ( k ) = 0 .
The equality (15) holds for all l > m . Next, we will prove that Equations (13) and (14) hold for all l > m .
i = 1 4 R i ( k + 1 ) , R i * ( m ) = i = 1 4 R i ( k + 1 ) , P i * ( m ) β ( m 1 ) P i * ( m 1 ) = i = 1 4 R i ( k + 1 ) , P i * ( m ) β ( m 1 ) R i ( k + 1 ) , P i * ( m 1 ) = 0 ,
i = 1 4 R i ( k + 1 ) , R i * ( k ) = i = 1 4 R i ( k + 1 ) , P i * ( m ) β ( m 1 ) P i * ( m 1 ) = i = 1 4 R i ( k + 1 ) , P i * ( m ) β ( m 1 ) R i ( k + 1 ) , P i * ( m 1 ) = 0 ,
and
i = 1 4 Q i x ( k + 1 ) + Q i y ( k + 1 ) , P i * ( m ) = i = 1 4 Q i x ( k + 1 ) , P i * ( m ) + i = 1 4 Q i y ( k + 1 ) , P i * ( m ) = i = 1 4 R i ( k + 1 ) , Γ i * [ P 1 * ( m ) , P 2 * ( m ) , P 3 * ( m ) , P 4 * ( m ) ] + β ( k ) P i ( k ) , Γ i * [ P 1 * ( m ) , P 2 * ( m ) , P 3 * ( m ) , P 4 * ( m ) ] + i = 1 4 R i ( k + 1 ) , Φ i * [ P 1 * ( m ) , P 2 * ( m ) , P 3 * ( m ) , P 4 * ( m ) ] + β ( k ) P i ( k ) , Φ i * [ P 1 * ( m ) , P 2 * ( m ) , P 3 * ( m ) , P 4 * ( m ) ] = i = 1 4 R i ( k + 1 ) , Q i x * ( m ) + β ( k ) P i ( k ) , Q i x * ( m ) + R i ( k + 1 ) , Q i y * ( m ) + β ( k ) P i ( k ) , Q i y * ( m ) = i = 1 4 R i ( k + 1 ) , 1 α ( m ) ( R i * ( m + 1 ) R i ( m ) ) + β ( k ) P i ( k ) , 1 α ( m ) ( R i * ( m + 1 ) R i ( m ) ) = 0 ,
i = 1 4 Q i x ( k + 1 ) + Q i y ( k + 1 ) , P i * ( k ) = i = 1 4 Q i x ( k + 1 ) , P i * ( k ) + i = 1 4 Q i y ( k + 1 ) , P i * ( k ) = i = 1 4 R i ( k + 1 ) , Γ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ] + β ( k ) P i ( k ) , Γ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ] + i = 1 4 R i ( k + 1 ) , Φ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ]
+ β ( k ) P i ( k ) , Φ i * [ P 1 * ( k ) , P 2 * ( k ) , P 3 * ( k ) , P 4 * ( k ) ] = i = 1 4 R i ( k + 1 ) , Q i x * ( k ) + β ( k ) P i ( k ) , Q i x * ( k ) + R i ( k + 1 ) , Q i y * ( k ) + β ( k ) P i ( k ) , Q i y * ( k ) = i = 1 4 R i ( k + 1 ) , Q i x * ( k ) + Q i y * ( k ) i = 1 4 R i ( k + 1 ) , Q i x * ( k ) + Q i y * ( k ) i = 1 4 P i ( k ) , Q i x * ( k ) + Q i y * ( k ) + P i ( k ) , Q i x * ( k ) + Q i y * ( k ) = 0 .
Similarly, Equations (13) and (14) also hold for the case 1 m < l k . Therefore, the facts illustrate that (13) and (14) are satisfied for l m . □
Corollary 1. 
Assume the conditions in Theorem 2 are satisfied. Then,
i = 1 4 R i ( k ) , P i * ( k ) = i = 1 4 R i ( k ) , R i * ( k ) ,
i = 1 4 Q i x ( k ) + Q i y ( k ) , R i * ( k ) = i = 1 4 Q i x ( k ) + Q i y ( k ) , P i * ( k ) .
Proof of Corollary 1. 
From Algorithm 1 and Theorem 2, we have
i = 1 4 R i ( k ) , P i * ( k ) = i = 1 4 R i ( k ) , R i * ( k ) + β ( k 1 ) P i * ( k 1 ) = i = 1 4 R i ( k ) , R i * ( k ) ,
and
i = 1 4 Q i x ( k ) + Q i y ( k ) , R i * ( k ) = i = 1 4 Q i x ( k ) + Q i y ( k ) , P i * ( k ) β ( k 1 ) P i * ( k 1 ) = i = 1 4 Q i x ( k ) + Q i y ( k ) , P i * ( k ) .
Theorem 3. 
Let tensor sequences { X i ( k ) } , { Y i ( k ) } ( i = 1 , 2 , 3 , 4 ) be generated by Algorithm 1. If Algorithm 1 does not break down, then the tensor sequences
{ [ X ( k ) , Y ( k ) ] X ( k ) = X 1 ( k ) + X 2 ( k ) i + X 3 ( k ) j + X 4 ( k ) k , Y ( k ) = Y 1 ( k ) + Y 2 ( k ) i + Y 3 ( k ) j + Y 4 ( k ) k , k = 1 , 2 , }
converge to the solution of Equation (2) within a finite iteration steps in the absence of round-off errors.
Proof of Theorem 3. 
We will prove that there exists a k 4 S N such that R i ( k ) = 0 . By contradiction, assume that R i ( k ) 0 , i = 1 , 2 , 3 , 4 , for all k 4 S N , and thus we can compute R i ( 4 S N + 1 ) . Suppose that R i ( 1 ) , R i ( 2 ) , , R i ( 4 S N ) is a dependent sequence, then there exist real numbers λ i , 1 , , λ i , 4 S N , not all zero, such that
λ i , 1 R i ( 1 ) + + λ i , 4 S N R i ( 4 S N ) = 0 ,
for i = 1 , 2 , 3 , 4 . Then
i = 1 4 R i ( l ) , 0 = i = 1 4 R i ( l ) , λ i , 1 R i ( 1 ) + + λ i , 4 S N R i ( 4 S N ) = i = 1 4 λ i , l R i ( l ) , R i ( l ) = 0 ,
which implies i = 1 4 R i ( l ) , R i ( l ) = 0 . This is a contradiction, since we cannot calculate R i ( 4 S N + 1 ) in this case. Therefore, there must exist a k 4 S N such that R i ( k ) = 0 , that is, the exact solution to the tensor Equation (2) can be determined by Algorithm 1 within a finite number of iterations, assuming no round-off errors. □
In the following theorem, we show that if we choose special kinds of initial tensor, then Algorithm 1 can yield the unique minimal Frobenius norm solution of the tensor Equation (2).
Theorem 4. 
By selecting the initial tensors as
X j ( 1 ) = Γ j * [ H 1 ( 1 ) , H 2 ( 1 ) , H 3 ( 1 ) , H 4 ( 1 ) ] , Y j ( 1 ) = Φ j * [ H 1 ( 1 ) , H 2 ( 1 ) , H 3 ( 1 ) , H 4 ( 1 ) ] ,
where Γ j * , Φ j * ( j = 1 , 2 , 3 , 4 ) are defined by (11) and (12), H i ( 1 ) R I 1 × I 2 × × I N , ( i = 1 , 2 , 3 , 4 ) are arbitrary tensors (for instance, we set X j ( 1 ) = O , Y j ( 1 ) = O , j = 1 , 2 , 3 , 4 , then the solution group X j ˜ , Y j ˜ ( j = 1 , 2 , 3 , 4 ) obtained from Algorithm 1 represents the unique minimal Frobenius norm solution of the tensor Equations (2).
Proof of Theorem 4. 
By selecting the initial tensors as specified in (18), it can be easily verified that the tensors X j ˜ and Y j ˜ ( j = 1 , 2 , 3 , 4 ) obtained from Algorithm 1 will have the following form:
X j ˜ = Γ j * [ H 1 , H 2 , H 3 , H 4 ] , Y j ˜ = Φ j * [ H 1 , H 2 , H 3 , H 4 ] ,
where tensors H i R I 1 × I 2 × × I N ( i = 1 , 2 , 3 , 4 ) . Now, we show that X ˜ = X 1 ˜ + X 2 ˜ i + X 3 ˜ j + X 4 ˜ k , Y ˜ = Y 1 ˜ + Y 2 ˜ i + Y 3 ˜ j + Y 4 ˜ k is the unique solution of the tensor equation with the minimal Frobenius norm (2). Let
z ˜ = vec X 1 ˜ vec X 2 ˜ vec X 3 ˜ vec X 4 ˜ vec Y 1 ˜ vec Y 2 ˜ vec Y 3 ˜ vec Y 4 ˜ = Kro L A 1 ( n ) T + Kro L A 2 ( n ) T + Kro L A 3 ( n ) T + Kro L A 4 ( n ) T Kro L A 2 ( n ) T + Kro L A 1 ( n ) T Kro L A 4 ( n ) T + Kro L A 3 ( n ) T Kro L A 3 ( n ) T + Kro L A 4 ( n ) T + Kro L A 1 ( n ) T Kro L A 2 ( n ) T Kro L A 4 ( n ) T Kro L A 3 ( n ) T + Kro L A 2 ( n ) T + Kro L A 1 ( n ) T Kro L B 1 ( n ) T + Kro L B 2 ( n ) T + Kro L B 3 ( n ) T + Kro L B 4 ( n ) T Kro L B 2 ( n ) T + Kro L B 1 ( n ) T Kro L B 4 ( n ) T + Kro L B 3 ( n ) T Kro L B 3 ( n ) T + Kro L B 4 ( n ) T + Kro L B 1 ( n ) T Kro L B 2 ( n ) T Kro L B 4 ( n ) T Kro L B 3 ( n ) T + Kro L B 2 ( n ) T + Kro L B 1 ( n ) T
× vec H 1 ˜ vec H 2 ˜ vec H 3 ˜ vec H 4 ˜ , = Kro L A 1 ( n ) Kro L A 2 ( n ) Kro L A 3 ( n ) Kro L A 4 ( n ) Kro L A 2 ( n ) Kro L A 1 ( n ) Kro L A 4 ( n ) Kro L A 3 ( n ) Kro L A 3 ( n ) Kro L A 4 ( n ) Kro L A 1 ( n ) Kro L A 2 ( n ) Kro L A 4 ( n ) Kro L A 3 ( n ) Kro L A 2 ( n ) Kro L A 1 ( n ) Kro L B 1 ( n ) Kro L B 2 ( n ) Kro L B 3 ( n ) Kro L B 4 ( n ) Kro L B 2 ( n ) Kro L B 1 ( n ) Kro L B 4 ( n ) Kro L B 3 ( n ) Kro L B 3 ( n ) Kro L B 4 ( n ) Kro L B 1 ( n ) Kro L B 2 ( n ) Kro L B 4 ( n ) Kro L B 3 ( n ) Kro L B 2 ( n ) Kro L B 1 ( n ) T × vec H 1 ˜ vec H 2 ˜ vec H 3 ˜ vec H 4 ˜ , = [ M A , M B ] T × vec H 1 ˜ vec H 2 ˜ vec H 3 ˜ vec H 4 ˜ R ( [ M A , M B ] T ) ,
where Kro L A i ( n ) and Kro L B i ( n ) are defined by (9). According to Theorem 1, we can conclude that X ˜ = X 1 ˜ + X 2 ˜ i + X 3 ˜ j + X 4 ˜ k , Y ˜ = Y 1 ˜ + Y 2 ˜ i + Y 3 ˜ j + Y 4 ˜ k produced by Algorithm 1; this solution is the unique minimal Frobenius norm solution for the tensor Equation (2). □
Now, we solve Problem 1.2. If the tensor Equation (2) is consistent, the solution pair set S X Y for Problem 1.1 is non-empty, for given tensors X ¯ , Y ¯ H I 1 × I 2 × × I N , we have
min X , Y H 1 × I 2 × × I N k = 1 N X × k A ( k ) + Y × k B ( k ) C = min X , Y H I 1 × I 2 × × I N k = 1 N X X ¯ × k A ( k ) + Y Y ¯ × k B ( k ) C k = 1 N X ¯ × k A ( k ) Y ¯ × k B ( k ) .
Let X ˜ = X X ¯ , Y ˜ = Y Y ¯ , and C ˜ = C k = 1 N X ¯ × k A ( k ) Y ¯ × k B ( k ) , then solving the tensor nearness Problem 1.2 is equivalent to first finding the solution with the minimal Frobenius norm for the tensor equation
k = 1 N X ˜ × k A ( k ) + Y ˜ × k B ( k ) = C .
By applying Algorithm 1 and setting the initial tensors as X ˜ = X 1 ˜ + X 2 ˜ i + X 3 ˜ j + X 4 ˜ k , Y ˜ = Y 1 ˜ + Y 2 ˜ i + Y 3 ˜ j + Y 4 ˜ k , where X j ( 1 ) = Γ j * [ H 1 ( 1 ) , H 2 ( 1 ) , H 3 ( 1 ) , H 4 ( 1 ) ] ,   Y j ( 1 ) = Φ j * [ H 1 ( 1 ) , H 2 ( 1 ) , H 3 ( 1 ) , H 4 ( 1 ) ] ( j = 1 , 2 , 3 , 4 ) , where H i ( 1 ) R I 1 × I 2 × × I N ( i = 1 , 2 , 3 , 4 ) are arbitrary tensors (in particular, we take X j ( 1 ) = O , Y j ( 1 ) = O , j = 1 , 2 , 3 , 4 , we can derive the unique solution with the minimal Frobenius norm X ˜ * , Y ˜ * of Equation (19). Once the above tensors X ˜ * , Y ˜ * are obtained, the unique solution X ˇ , Y ˇ of Problem 1.2 can be determined. In this situation, X ˇ and Y ˇ can be represented as X ˇ = X ˜ * + X ¯ , Y ˇ = Y ˜ * + Y ¯ .

4. Numerical Examples

In this section, we will give some numerical examples to support the efficiency and applications of Algorithm 1. The codes in our computation are written in Matlab R2018a with 2.3 GHz central processing unit (Intel(R) Core(TM) i5), 8 GB memory. Moreover, we implemented all the operations based on tensor toolbox (version 3.2.1) proposed by Bader and Kolda [1]. For all of the examples, the iterations begin with the initial values X i = Y i = 0 , i = 1 , 2 , 3 , 4 in Algorithm 1 with a stopping criterion of Res 10 5 and the number of iteration steps exceeding 2000. We describe some notations that appear in the following examples in Table 1.
Example 1. 
We consider the tensor Equation (2) under the following conditions:
A ( 1 ) = 2 + 4 i + 4 j + 5 k 1 i + j + 2 k 2 i 2 j + k 2 + j 3 + 2 i + 5 j + k 1 2 j + k 2 + 2 i + j 1 2 i + 2 j k 3 + 4 i + 4 j + k , A ( 2 ) = 3 + 4 j 1 + j 1 2 j 2 + j 4 + 3 j j j 1 2 + j , A ( 3 ) = 2 i + 3 j i 1 + j 2 i j 5 i + 3 j i 2 i j i j 5 i + 4 j , B ( 1 ) = 3 + 3 i 2 + i 2 i 2 + i 4 + 4 i 2 2 i 2 + 2 i 2 6 + 4 i , B ( 2 ) = 3 i + 5 j 2 i 2 i + 2 j 0 2 i + 3 j j 2 i + 2 j 2 i 2 j 6 i + 6 j , B ( 3 ) = 1 + 3 i + 3 k 1 2 i i + 2 k i 2 + 4 i + 2 k 2 + 2 i 2 + i + 2 k 2 + i 2 k 5 + 4 i + 6 k , C ( : , : , 1 ) = 3 11 i + 31 j + 19 k 6 8 i + j + 28 k 2 6 i + 20 j + 24 k 4 + 36 j + 8 k 7 + 3 i + 27 j + 17 k 3 + 5 i + 25 j + 13 k 24 2 i + 52 j + 22 k 27 + i + 43 j + 31 k 23 + 3 i + 41 j + 27 k , C ( : , : , 2 ) = 11 13 i + 31 j + 13 k 14 10 i + 22 j + 22 k 10 8 i + 20 j + 18 k 12 2 i + 36 j + 2 k 15 + i + 27 j + 11 k 11 + 3 i + 25 j + 7 k 32 4 i + 52 j + 16 k 35 i + 43 j + 25 k 31 + i + 41 j + 21 k ,
C ( : , : , 3 ) = 9 9 i + 45 j + 25 k 12 6 i + 36 j + 34 k 8 4 i + 34 j + 30 k 10 + 2 i + 50 j + 14 k 13 + 5 i + 41 j + 23 k 9 + 7 i + 39 j + 19 k 30 + 66 j + 28 k 33 + 3 i + 57 j + 3 k 29 + 5 i + 55 j + 33 k .
Applying for Algorithm 1, the IT is 46, the CPU time is 3.9496 s and the Res is 1.2885 × 10 6 . Figure 1 illustrates that Algorithm 1 is feasible.
Example 2 
(Test matrices from [28]). We examine the quaternion tensor Equation (2) under the condition that
A ( m ) = A m 1 + A m 2 i + A m 3 j + A m 4 k , B ( m ) = B m 1 + B m 2 i + B m 3 j + B m 4 k , m = 1 , 2 , 3 ,
where
A 11 = t r i u ( h i l b ( n ) ) , A 12 = t r i u ( o n e s ( n , n ) ) , A 13 = e y e ( n ) , A 14 = o n e s ( n ) , A 21 = z e r o s ( n ) , A 22 = z e r o s ( n ) , A 23 = t r i d i a g ( 1 , 2 , 1 , n ) , A 24 = z e r o s ( n ) , A 31 = z e r o s ( n ) , A 32 = t r i d i a g ( 0.5 , 6 , 0.5 , n ) , A 33 = e y e ( n ) , A 34 = z e r o s ( n ) ,
B 11 = e y e ( n ) , B 12 = o n e s ( n ) , B 13 = z e r o s ( n ) , B 14 = z e r o s ( n ) , B 21 = z e r o s ( n ) , B 22 = t r i d i a g ( 0.5 , 6 , 0.5 , n ) , B 23 = e y e ( n ) , B 24 = z e r o s ( n ) , B 31 = t r i d i a g ( 0.5 , 6 , 0.5 , n ) , B 32 = e y e ( n ) , B 33 = z e r o s ( n ) , B 34 = o n e s ( n ) .
C = t e n r a n d ( n , n , n ) + t e n r a n d ( n , n , n ) i + t e n r a n d ( n , n , n ) j + t e n r a n d ( n , n , n ) k .
Choosingthe initial tensor X = Y = 0 , we illustrate the convergence curves of Algorithm 1 for various values of n in Figure 2. In Table 2, for n = 20 , n = 40 and n = 60 , we provide the CPU time incurred, residual norms after a finite number of steps, and the relative errors of the approximate solutions obtained using Algorithm 1.
Example 3. 
We consider the solution of the following convection–diffusion equation over the quaternion algebra [56]
v Δ u + c T u = f in Γ = [ 0 , 1 ] × [ 0 , 1 ] × [ 0 , 1 ] u = 0 on Γ .
Based on a standard finite difference discretization on a uniform grid for the diffusion term and a second-order convergent scheme (Fromm’s scheme) for the convection term with the mesh-size h = 1 l + 1 , we solve the quaternion tensor Equation (3) with
A ( n ) = A 1 ( n ) + A 2 ( n ) i + A 3 ( n ) j + A 4 ( n ) k , n = 1 , 2 , 3 ,
where
A i ( n ) = v i ( n ) h 2 2 1 1 2 1 1 2 1 1 2 l × l + c i ( n ) 4 h 3 5 1 1 3 5 1 1 3 5 1 3 l × l ,
and i = 1 , 2 , 3 , 4 . The right-hand side tensor C is constructed so that the exact solution to Equation (3) is X * = X 1 * + X 2 * i + X 3 * j + X 4 * k = t e n o n e s ( l , l , l ) + t e n o n e s ( l , l , l ) i + t e n o n e s ( l , l , l ) j + t e n o n e s ( l , l , l ) k .
We consider two cases in order to compare Algorithm 1 with the CGLS algorithm in [43]. In case I, we choose different v i ( n ) and c i ( n ) to obtain the results. In Table 3, we set
v i ( 1 ) = 1 , c i ( 1 ) = 1 ; v i ( 2 ) = 1 , c i ( 2 ) = 2 ; v i ( 3 ) = 1 , c i ( 2 ) = 3 , i = 1 , 2 , 3 , 4
to obtain A ( n ) ( n = 1 , 2 , 3 ) with the same real part and imaginary part. In Table 4, we set
v i ( 1 ) = 1 , c 1 ( 1 ) = 1 , c 2 ( 1 ) = 1 , c 3 ( 1 ) = 0 , c 4 ( 1 ) = 1 ; v i ( 2 ) = 0.1 , c 1 ( 2 ) = 1 , c 2 ( 2 ) = 1 , c 3 ( 2 ) = 1 , c 4 ( 2 ) = 0 ; v i ( 3 ) = 0.01 , c 1 ( 3 ) = 1 , c 2 ( 3 ) = 1 , c 3 ( 3 ) = 0 , c 4 ( 3 ) = 0 , i = 1 , 2 , 3 , 4
to obtain A ( n ) ( n = 1 , 2 , 3 ) with different real parts and imaginary parts.
In case II, we set c i ( 1 ) = 1 ; c i ( 2 ) = 2 ; c i ( 2 ) = 3 , i = 1 , 2 , 3 , 4 , we apply Algorithm 1 and the CGLS algorithm in [43] with v i ( j ) = 10 3 , 1 , 100 , i = 1 , 2 , 3 , 4 , j = 1 , 2 , 3 for grid 10 × 10 × 10 . The relative errors of approximate solution i = 1 i = 4 X i ( k ) X i * / X i * computed by these methods are shown in Table 5.
The previous results show that Algorithm 1 has faster convergent rates than the CGLS algorithm in [43] as p increases.
Example 4. 
We employ Algorithm 1 to compare its performance with the CGLS algorithm in restoring a color video comprising a sequence of RGB images (slices). The video, titled ‘Rhinos,’ originates from Matlab and is saved in AVI format. Each frontal slice of this color video is represented by a pure quaternion matrix with dimensions of 240 × 320 pixels. For C = C ^ + N = X × 1 A , we consider X to represent the orginal colour video, A as the blurred matrix, and N as a noise tensor. When N = 0 , C is referred to as the blurred and noise-free color video. In this scenario, we select a blurred matrix A = A 1 A 2 R 240 × 320 , where A 1 = a i j ( 1 ) 1 i , j 16 and A 2 = a i j ( 2 ) 1 i 15 , 1 j 20 are Toeplitz matrices with entries defined as follows:
a i j ( 1 ) = 1 σ 2 π exp ( i j ) 2 2 σ 2 , | i j | r , 1 , otherwise . ; a i j ( 2 ) = 1 2 s 1 , | i j | s , 0 , otherwise .
We denote X restored as the resulting restored color video. The performance of the algorithm is evaluated using the peak signal-to-noise ratio (PSNR), which is measured in decibels (dB):
PSNR ( X ) = 10 log 10 I 1 I 2 d 2 X X restored 2 ,
where d denotes the maximum possible pixel value of the image. RE ( X ) represents the relative error, which is defined as
RE ( X ) = X X restored X .
In this case, we use d = 255 and set the variance σ = 1 . Table 6 shows the peak signal-to-noise ratio (PSNR) and the relative error (RE) for Algorithm 1 and the CGLS algorithm with various parameters. As indicated, the PSNR and the relative error of our algorithms are significantly superior to those of CGLS. For the case where r = 6 and s = 6 with slice No. 7 of the color video, we illustrate the original image, blurred image, and the restored image by CGLS and Algorithm 1 in Figure 3. This figure demonstrates that our algorithm can effectively restore blurred and noise-free color video with high quality.

5. Conclusions

The main goal of this paper is to solve the generalized quaternion tensor Equation (2). We hereby develop a BiCG iterative algorithm in tensor format to efficiently solve Equation (2), and we also prove the convergence of our proposed method. Moreover, we demonstrate that the solution with a minimal Frobenius norm can be achieved by initializing specific types of tensors. We provide several examples to effectively illustrate the effectiveness of our algorithm. Furthermore, our algorithm is successfully applied to the restoration of color videos. This contribution significantly advances the current understanding of quaternion tensor equations by introducing a practical iterative approach.

Author Contributions

Conceptualization, M.X. and Q.-W.W.; methodology, M.X.; software, M.X.; validation, M.X., Q.-W.W. and Y.Z.; formal analysis, M.X.; investigation, M.X.; resources, Q.-W.W.; data curation, M.X.; writing—original draft preparation, M.X.; writing—review and editing, Y.Z.; visualization, M.X.; supervision, Q.-W.W.; project administration, M.X., Q.-W.W. and Y.Z.; funding acquisition, M.X., Q.-W.W. and Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

The first author is supported by the Natural Science Foundation of China under Grant No. 12301028 and Startup Foundation for Young Teachers of Shanghai Ocean University. The second author is supported by the Natural Science Foundation of China under Grant No. 12371023. The third author is supported by the Canada NSERC under Grant No. RGPIN-2020-06746.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank the editor and reviewers for their valuable suggestions and comments.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  2. Qi, L.; Luo, Z. Tensor Analysis: Spectral Theory and Special Tensors; SIAM: Philadelphia, PA, USA, 2017. [Google Scholar]
  3. Beik, F.P.; Jbilou, K.; Najafi-Kalyani, M.; Reichel, L. Golub–Kahan bidiagonalization for ill-conditioned tensor equations with applications. Numer. Algorithms 2020, 84, 1535–1563. [Google Scholar] [CrossRef]
  4. Duan, X.F.; Zhang, Y.S.; Wang, Q.W. An efficient iterative method for solving a class of constrained tensor least squares problem. Appl. Numer. Math. 2024, 196, 104–117. [Google Scholar] [CrossRef]
  5. Guan, Y.; Chu, D. Numerical computation for orthogonal low-rank approximation of tensors. SIAM J. Matrix Anal. Appl. 2019, 40, 1047–1065. [Google Scholar] [CrossRef]
  6. Guan, Y.; Chu, M.T.; Chu, D. Convergence analysis of an SVD-based algorithm for the best rank-1 tensor approximation. Linear Algebra Appl. 2018, 555, 53–69. [Google Scholar] [CrossRef]
  7. Guan, Y.; Chu, M.T.; Chu, D. SVD-based algorithms for the best rank-1 approximation of a symmetric tensor. SIAM J. Matrix Anal. 2018, 39, 1095–1115. [Google Scholar] [CrossRef]
  8. Hu, J.; Ke, Y.; Ma, C. Efficient iterative method for generalized Sylvester quaternion tensor equation. Comput. Appl. Math. 2023, 42, 237. [Google Scholar] [CrossRef]
  9. Ke, Y. Finite iterative algorithm for the complex generalized Sylvester tensor equations. J. Appl. Anal. Comput. 2020, 10, 972–985. [Google Scholar] [CrossRef]
  10. Kolda, T.G. Multilinear Operators for Higher-Order Decompositions; Sandia National Laboratory (SNL): Albuquerque, NM, USA; Livermore, CA, USA, 2006. [Google Scholar]
  11. Li, B.W.; Tian, S.; Sun, Y.S.; Hu, Z.M. Schur-decomposition for 3D matrix equations and its application in solving radiative discrete ordinates equations discretized by Chebyshev collocation spectral method. J. Comput. Phys. 2010, 229, 1198–1212. [Google Scholar] [CrossRef]
  12. Li, T.; Wang, Q.-W.; Duan, X.-F. Numerical algorithms for solving discrete Lyapunov tensor equation. J. Comput. Appl. Math. 2020, 370, 112676. [Google Scholar] [CrossRef]
  13. Li, T.; Wang, Q.-W.; Zhang, X.-F. Gradient based iterative methods for solving symmetric tensor equations. Numer. Linear Algebra Appl. 2022, 29, e2414. [Google Scholar] [CrossRef]
  14. Li, T.; Wang, Q.-W.; Zhang, X.-F. A Modified conjugate residual method and nearest kronecker product preconditioner for the generalized coupled Sylvester tensor equations. Mathematics 2022, 10, 1730. [Google Scholar] [CrossRef]
  15. Li, X.; Ng, M.K. Solving sparse non-negative tensor equations: Algorithms and applications. Front. Math. China 2015, 10, 649–680. [Google Scholar] [CrossRef]
  16. Liang, Y.; Silva, S.D.; Zhang, Y. The tensor rank problem over the quaternions. Linear Algebra Appl. 2021, 620, 37–60. [Google Scholar] [CrossRef]
  17. Lv, C.; Ma, C. A modified CG algorithm for solving generalized coupled Sylvester tensor equations. Appl. Math. Comput. 2020, 365, 124699. [Google Scholar] [CrossRef]
  18. Malek, A.; Momeni-Masuleh, S.H. A mixed collocation–finite difference method for 3D microscopic heat transport problems. J. Comput. Appl. Math. 2008, 217, 137–147. [Google Scholar] [CrossRef]
  19. Qi, L. Eigenvalues of a real supersymmetric tensor. J. Symb. Comput. 2005, 40, 1302–1324. [Google Scholar] [CrossRef]
  20. Qi, L. Symmetric nonnegative tensors and copositive tensors. Linear Algebra Appl. 2013, 439, 228–238. [Google Scholar] [CrossRef]
  21. Qi, L.; Chen, H.; Chen, Y. Tensor Eigenvalues and Their Applications; Springer: Berlin/Heidelberg, Germany, 2018. [Google Scholar]
  22. Zhang, X.-F.; Li, T.; Ou, Y.-G. Iterative solutions of generalized Sylvester quaternion tensor equations. Linear Multilinear Algebra 2024, 72, 1259–1278. [Google Scholar] [CrossRef]
  23. Zhang, X.-F.; Wang, Q.-W. On RGI Algorithms for Solving Sylvester Tensor Equations. Taiwan. J. Math. 2022, 26, 501–519. [Google Scholar] [CrossRef]
  24. Kyrchei, I. Cramer’s rules for Sylvester quaternion matrix equation and its special cases. Adv. Appl. Clifford Algebras 2018, 28, 1–26. [Google Scholar] [CrossRef]
  25. Heyouni, M.; Saberi-Movahed, F.; Tajaddini, A. On global Hessenberg based methods for solving Sylvester matrix equations. Comput. Math. Appl. 2019, 77, 77–92. [Google Scholar] [CrossRef]
  26. Zhang, X. A system of generalized Sylvester quaternion matrix equations and its applications. Appl. Math. Comput. 2016, 273, 74–81. [Google Scholar] [CrossRef]
  27. Beik, F.; Ahmadi-Asl, S. An iterative algorithmfor η-(anti)-Hermitian least-squares solutions of quaternion matrix equations. Electron. J. Linear Algebra 2015, 30, 372–401. [Google Scholar] [CrossRef]
  28. Ahmadi-Asl, S.; Beik, F.P.A. An efficient iterative algorithm for quaternionic least-squares problems over the generalized η-(anti-)bi-Hermitian matrices. Linear Multilinear Algebra 2017, 65, 1743–1769. [Google Scholar] [CrossRef]
  29. Ahmadi-Asl, S.; Beik, F.P.A. Iterative algorithms for least-squares solutions of a quaternion matrix equation. J. Appl. Math. Comput. 2017, 53, 95–127. [Google Scholar] [CrossRef]
  30. Song, G.; Wang, Q.-W.; Yu, S. Cramer’s rule for a system of quaternion matrix equations with applications. Appl. Math. Comput. 2018, 336, 490–499. [Google Scholar] [CrossRef]
  31. Wang, Q.-W.; He, Z.-H.; Zhang, Y. Constrained two-sided coupled Sylvester-type quaternion matrix equations. Automatica 2019, 101, 207–213. [Google Scholar] [CrossRef]
  32. Zhang, F.; Mu, W.; Li, Y.; Zhao, J. Special least squares solutions of the quaternion matrix equation AXB + CXD = E. Comput. Math. Appl. 2016, 72, 1426–1435. [Google Scholar] [CrossRef]
  33. Huang, N.; Ma, C.-F. Modified conjugate gradient method for obtaining the minimum-norm solution of the generalized coupled Sylvester-conjugate matrix equations. Appl. Math. Model. 2016, 40, 1260–1275. [Google Scholar] [CrossRef]
  34. Gao, Z.-H.; Wang, Q.-W.; Xie, L. The (anti-)η-Hermitian solution to a novel system of matrix equations over the split quaternion algebra. Math. Meth. Appl. Sci. 2024, 1–18. [Google Scholar] [CrossRef]
  35. He, Z.-H.; Wang, X.-X.; Zhao, Y.-F. Eigenvalues of quaternion tensors with applications to color video processing. J. Sci. Comput. 2023, 94, 1. [Google Scholar] [CrossRef]
  36. Jia, Z.; Wei, M.; Zhao, M.X.; Chen, Y. A new real structure-preserving quaternion QR algorithm. J. Comput. Appl. Math. 2018, 343, 26–48. [Google Scholar] [CrossRef]
  37. Li, Y.; Wei, M.; Zhang, F.; Zhao, J. Real structure-preserving algorithms of Householder based transformations for quaternion matrices. J. Comput. Appl. Math. 2016, 305, 82–91. [Google Scholar] [CrossRef]
  38. Mehany, M.S.; Wang, Q.-W.; Liu, L. A System of Sylvester-like quaternion tensor equations with an application. Front. Math. 2024, 19, 749–768. [Google Scholar] [CrossRef]
  39. Xie, M.; Wang, Q.-W. Reducible solution to a quaternion tensor equation. Front. Math. China 2020, 15, 1047–1070. [Google Scholar] [CrossRef]
  40. Xie, M.Y.; Wang, Q.W.; He, Z.H.; Saad, M.M. A system of Sylvester-type quaternion matrix equations with ten variables. Acta Math. Sin. (Engl. Ser.) 2022, 38, 1399–1420. [Google Scholar] [CrossRef]
  41. Saberi-Movahed, F.; Tajaddini, A.; Heyouni, M.; Elbouyahyaoui, L. Some iterative approaches for Sylvester tensor equations, Part I: A tensor format of truncated Loose Simpler GMRES. Appl. Numer. Math. 2022, 172, 428–445. [Google Scholar] [CrossRef]
  42. Saberi-Movahed, F.; Tajaddini, A.; Heyouni, M.; Elbouyahyaoui, L. Some iterative approaches for Sylvester tensor equations, Part II: A tensor format of Simpler variant of GCRO-based methods. Appl. Numer. Math. 2022, 172, 413–427. [Google Scholar] [CrossRef]
  43. Wang, Q.-W.; Xu, X.; Duan, X. Least squares solution of the quaternion Sylvester tensor equation. Linear Multilinear Algebra 2021, 69, 104–130. [Google Scholar] [CrossRef]
  44. Zhang, X.-F.; Wang, Q.-W. Developing iterative algorithms to solve Sylvester tensor equations. Appl. Math. Comput. 2021, 409, 126403. [Google Scholar] [CrossRef]
  45. Chen, Z.; Lu, L. A projection method and Kronecker product preconditioner for solving Sylvester tensor equations. Sci. China Math. 2012, 55, 1281–1292. [Google Scholar] [CrossRef]
  46. Karimi, S.; Dehghan, M. Global least squares method based on tensor form to solve linear systems in Kronecker format. Trans. Inst. Measure. Control 2018, 40, 2378–2386. [Google Scholar] [CrossRef]
  47. Najafi-Kalyani, M.; Beik, F.P.A.; Jbilou, K. On global iterative schemes based on Hessenberg process for (ill-posed) Sylvester tensor equations. J. Comput. Appl. Math. 2020, 373, 112216. [Google Scholar] [CrossRef]
  48. Bai, Z.-Z.; Golub, G.H.; Ng, M.K. Hermitian and skew-Hermitian splitting methods for non-Hermitian positive definite linear systems. SIAM J. Matrix Anal. Appl. 2003, 24, 603–626. [Google Scholar] [CrossRef]
  49. Grasedyck, L. Existence and computation of low Kronecker-rank approximations for large linear systems of tensor product structure. Computing 2004, 72, 247–265. [Google Scholar] [CrossRef]
  50. Peng, Y.; Hu, X.; Zhang, L. An iteration method for the symmetric solutions and the optimal approximation solution of the matrix equation AXB = C. Appl. Math. Comput. 2005, 160, 763–777. [Google Scholar] [CrossRef]
  51. Bank, R.E.; Chan, T.-F. An analysis of the composite step biconjugate gradient method. Numer. Math. 1993, 66, 295–319. [Google Scholar] [CrossRef]
  52. Bank, R.E.; Chan, T.-F. A composite step bi-conjugate gradient algorithm for nonsymmetric linear systems. Numer. Algorithms 1994, 7, 1–16. [Google Scholar] [CrossRef]
  53. Freund, R.W.; Golub, G.H.; Nachtigal, N.M. Iterative solution of linear systems. Acta Numer. 1992, 1, 44. [Google Scholar] [CrossRef]
  54. Hajarian, M. Developing Bi-CG and Bi-CR methods to solve generalized Sylvester-transpose matrix equations. Int. J. Auto. Comput. 2014, 11, 25–29. [Google Scholar] [CrossRef]
  55. Beik, F.; Panjeh, F.; Movahed, F.; Ahmadi-Asl, S. On the Krylov subspace methods based on tensor format for positive definite Sylvester tensor equations. Numer. Linear Algebra Appl. 2016, 23, 444–466. [Google Scholar] [CrossRef]
  56. Ballani, J.; Grasedyck, L. A projection method to solve linear systems in tensor format. Numer. Linear Algebra Appl. 2013, 20, 27–43. [Google Scholar] [CrossRef]
Figure 1. Convergence history of Example 1.
Figure 1. Convergence history of Example 1.
Symmetry 16 01167 g001
Figure 2. Convergence history of Example 2.
Figure 2. Convergence history of Example 2.
Symmetry 16 01167 g002
Figure 3. The restored color video for the case with r = 6 and s = 6 , using slice No. 7.
Figure 3. The restored color video for the case with r = 6 and s = 6 , using slice No. 7.
Symmetry 16 01167 g003
Table 1. Some denotations in numerical examples.
Table 1. Some denotations in numerical examples.
ITThe Number of Iterations
CPU timeThe CPU time elapsed, measured in seconds
Res i = 1 4 R i ( k ) , R i ( k ) is the residual at kth iteration
t e n r a n d ( n , n , n ) The third-order n × n × n tensor with pseudo-random values
sampled from a uniform distribution over the unit interval
t r i u ( h i l b ( n ) ) The upper triangular part of n × n Hilbert matrix
t r i u ( o n e s ( n , n ) ) The upper triangular part of n × n matrix with all 1
e y e ( n ) Identity matrix with size n × n
z e r o s ( n ) Zero matrix with size n × n
t r i d i a g ( a , b , c , n ) The n × n tridiagonal matrix with a , b , c
Table 2. Numerical results for Example 2.
Table 2. Numerical results for Example 2.
nITCPU TimeRes
20828.3272 7.7182 × 10 6
4015624.8187 5.1233 × 10 6
6028891.4387 7.6898 × 10 6
Table 3. CPU time (IT) for Example 3 with parameter setup in (20).
Table 3. CPU time (IT) for Example 3 with parameter setup in (20).
   l = 10l = 25l = 30
Algorithm 125.6884(320)140.8233(1270)202.7709(1580)
CGLS [43]15.5685(219)144.3166(1266)230.7340(1832)
Table 4. CPU time (IT) for Example 3 with parameter setup in (21).
Table 4. CPU time (IT) for Example 3 with parameter setup in (21).
   l = 10l = 15l = 20
Algorithm 145.4157(576)104.4370(1095)163.6110(1700)
CGLS [43]44.0701(579)118.2419(1296)230.7978(2298)
Table 5. The relative errors of the solution (IT) for Example 3.
Table 5. The relative errors of the solution (IT) for Example 3.
    v i ( 1 ) = 10 3 v i ( 2 ) = 1 v i ( 3 ) = 100
Algorithm 1 5.6210 × 10 9 (248) 6.0527 × 10 9 (188) 4.1925 × 10 9 (248)
CGLS [43] 9.9300 × 10 9 (106) 9.7035 × 10 9 (178) 8.9552 × 10 9 (167)
Table 6. The numerical results for Example 4.
Table 6. The numerical results for Example 4.
Algorithm([r,s])Algorithm 1(PSNR/RE)CGLS(PSNR/RE)
[3,3]38.6181(0.0235)13.8404(0.3769)
[6,6]37.3721(0.0300)14.0529(0.3694)
[8,8]33.8073(0.0338)14.7958(0.3551)
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xie, M.; Wang, Q.-W.; Zhang, Y. The BiCG Algorithm for Solving the Minimal Frobenius Norm Solution of Generalized Sylvester Tensor Equation over the Quaternions. Symmetry 2024, 16, 1167. https://doi.org/10.3390/sym16091167

AMA Style

Xie M, Wang Q-W, Zhang Y. The BiCG Algorithm for Solving the Minimal Frobenius Norm Solution of Generalized Sylvester Tensor Equation over the Quaternions. Symmetry. 2024; 16(9):1167. https://doi.org/10.3390/sym16091167

Chicago/Turabian Style

Xie, Mengyan, Qing-Wen Wang, and Yang Zhang. 2024. "The BiCG Algorithm for Solving the Minimal Frobenius Norm Solution of Generalized Sylvester Tensor Equation over the Quaternions" Symmetry 16, no. 9: 1167. https://doi.org/10.3390/sym16091167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop