Next Article in Journal
Quaternionic Shape Operator and Rotation Matrix on Ruled Surfaces
Next Article in Special Issue
Green Matrices, Minors and Hadamard Products
Previous Article in Journal
Review of Quaternion Differential Equations: Historical Development, Applications, and Future Direction
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tensor Eigenvalue and SVD from the Viewpoint of Linear Transformation

1
School of Mathematics and Statistics, Liaoning University, Shenyang 110036, China
2
School of Mathematical Sciences, Dalian University of Technology, Dalian 116024, China
3
Business School, Dalian University of Foreign Language, Dalian 116044, China
*
Author to whom correspondence should be addressed.
Axioms 2023, 12(5), 485; https://doi.org/10.3390/axioms12050485
Submission received: 6 April 2023 / Revised: 11 May 2023 / Accepted: 16 May 2023 / Published: 17 May 2023
(This article belongs to the Special Issue Advances in Linear Algebra with Applications)

Abstract

:
A linear transformation from vector space to another vector space can be represented as a matrix. This close relationship between the matrix and the linear transformation is helpful for the study of matrices. In this paper, the tensor is regarded as a generalization of the matrix from the viewpoint of the linear transformation instead of the quadratic form in matrix theory; we discuss some operations and present some definitions and theorems related to tensors. For example, we provide the definitions of the triangular form and the eigenvalue of a tensor, and the theorems of the tensor QR decomposition and the tensor singular value decomposition. Furthermore, we explain the significance of our definitions and their differences from existing definitions.

1. Introduction

Unfolding is an important approach in tensor research, and one of the common methods of unfolding is mode-n matricization. Based on this, the mode-n product and n-rank are defined. This helps to extend the concepts of the eigenvalue and the singular value decomposition of matrices to tensors. For example, the eigenvalues of a real supersymmetric tensor are presented in [1], a multilinear singular value decomposition (HOSVD) is presented in [2], and the restricted singular value decomposition (RSVD) for three quaternion tensors is presented in [3]. The singular values of the general tensor are introduced in [4]. In [5,6], the singular values of a tensor are connected with its symmetric embedding. In [7], the authors presented the definition of the singular value of a real rectangular tensor and discussed its properties. For a tensor T R I × J × I × J , exploiting the matricization of the tensor, a singular value decomposition is presented in [8]. On the other hand, in matrix theory, many studies of matrices are inseparable from linear transformations. Due to the one-to-one correspondence between matrices and linear transformations, we always believe that it is more natural to provide definitions and theorems related to tensors from the viewpoint of linear transformation. This fact also proves that when we think from this perspective, some unresolved tensor problems have already been solved. Similar studies have been considered in [9]. We provide some supplements to the article presented in [9]; for example, we give a more general definition of the multiplication between tensor and tensor and present the form of a triangular tensor by proposing tensor QR decomposition. In addition, in matrix theory, the one-dimensional linear subspace made up of the eigenvectors of one eigenvalue of a linear transformation is stable under the action of this linear transformation. However, this property cannot be generalized for the existing tensor eigenvalue and the corresponding eigenvector. From this point of view, we present a new concept of the eigenvalue and the corresponding eigenvector of an even-order tensor. Then, the concepts of the eigenvalue decomposition, singular value, and singular value decomposition are naturally obtained.
This paper is organized as follows. In Section 2, regarding the tensor as a linear transformation, we present some definitions. For example, we present the identity tensor and a new multiplication for two tensors, and further explain the partitioning of the indices of a tensor. In Section 3, starting from the orthogonalization process of a set of tensors, we present the QR decomposition of a tensor and the form of a triangular tensor. In Section 4, motivated by the invariant eigenspace, we derive the definitions of the tensor eigenvalue and the corresponding tensor eigenvector, and the spectral decomposition theorem of a Hermitian tensor. The singular value decomposition of a tensor is defined in Section 5, and numerical examples are used to illustrate the advantages of our singular value decomposition in tensor compression. The application of the tensor singular value decomposition is described in Section 6.

2. Basic Definitions

For the convenience of writing, we illustrate the symbols to be used, which are similar to those in [9]. The set of integers { 1 , , p } will be abbreviated to the symbol [ [ p ] ] henceforth. Let ( α , β ) denote a partitioning of the set [ [ p ] ] , where α = { α 1 , , α m } and β = { β 1 , , β n } are two disjointed nonempty subsets that satisfy α β = [ [ p ] ] . Additionally, let α denote the number of elements in the set α . Under a given partitioning ( α , β ) , an element in the tensor T will be marked as τ [ I | J ] ( α , β ) , where I : = ( i 1 , , i m ) , J : = ( j 1 , , j n ) and i k [ [ I α k ] ] , j l [ [ I β l ] ] for all 1 k m and 1 l n .
Given a partitioning ( α , β ) , an order-p tensor T C I 1 × × I p can be regarded as a linear map
T β : C I β 1 × × I β n C I α 1 × × I α m .
In particular, an identity map is a linear transformation
E β : C I β 1 × × I β n C I β 1 × × I β n
that can map each tensor to itself, in other words, for arbitrary A C I β 1 × × I β n , E β ( A ) = A . Because the relationship between a tensor and the corresponding linear transformation depends on the partitioning ( α , β ) , in order to avoid confusion, we note the tensor discussed under the partitioning ( α , β ) as T β α . When T and ( α , β ) are known, we can use a command in MATLAB to determine T β α directly [10],
T_alpha_beta = permute(T,[alpha,beta]),
where [alpha,beta] indicates the order of the subscripts to be accessed when performing identification. In fact, for a fixed partitioning ( α , β ) , the tensor can be restored as a matrix. Precisely speaking, the element τ i 1 , , i p of an order-p tensor T C I 1 × I 2 × × I p is saved at location ( i , j ) of a matrix, where
i = ( i α m 1 ) I α m 1 I α 1 + + ( i α 2 1 ) I α 1 + i α 1 , j = ( i β n 1 ) I β n 1 I β 1 + + ( i β 2 1 ) I β 1 + i β 1 .
This storage method can also be regarded as a special case of blocking in [11].
Definition 1 
([11]). The set M = { m ( 1 ) , , m ( d ) } is a blocking for T C I 1 × × I p if m ( k ) = [ m 1 ( k ) , , m b k ( k ) ] is a vector of positive integers that sums to I k for k = 1 , , p .
When m ( α k ) = [ 1 , , 1 ] is a vector with ones in all elements and of length I α k for all α k α , and m ( β k ) = [ I β k ] for all β k β , the blocking M is the row blocking of T β α under the partitioning ( α , β ) . Similarly, we can obtain the column blockings. Specifically, the entry τ i 1 , , i p of an order-p tensor T C I 1 × I 2 × × I p can also be saved at the location
( i p 1 ) I p 1 I p 2 I 1 + ( i p 1 1 ) I p 2 I 1 + + ( i 2 1 ) I 1 + i 1
of the linear array. In the subsequent discussion, we shall comply with the rules in (1) or (2) whenever we want to unfold a tensor to a matrix or a vector.
There are many ways to define multiplication between tensors [12]. Based on the relationship between a tensor T and the corresponding map T β , in [9], the authors defined the tensor multiplication between the tensor T and any lower-order tensor A C I β 1 × × I β n as follows
T β ( A ) : = T β A = j 1 = 1 I β 1 j n = 1 I β n τ [ I | j 1 , , j n ] a j 1 , , j n C I α 1 × I α m .
The adjoint of T β , denoted by T β * , is a linear transformation
T β * : C I α 1 × × I α m C I β 1 × × I β n
such that the Lagrange identity [13]
T β ( A ) , B = A , T β * ( B )
is satisfied for all A C I β 1 × × I β n and B C I α 1 × × I α m . The tensor representation of the linear transformation T β * is ( T β α ) * , which is the conjugate transposition of T β α . See [9] for more details.
Similar to the derivation of the multiplication β in (3), by means of linear transformation, we define another multiplication between two tensors G and T when there exists ( γ , η ) and ( α , β ) satisfying η = α and I η i = I α i , i = 1 , , α . This definition is a generalization of the existing tensor contraction definition.
Definition 2. 
Suppose the linear transformations corresponding to G and T are
G η : C I α 1 × × I α m C I γ 1 × × I γ q , T β : C I β 1 × × I β n C I α 1 × × I α m ;
then, the composition of T and G is a linear transformation
G η T β : C I β 1 × × I β n C I γ 1 × × I γ q .
The rule of action is
( G η T β ( A ) ) K = ( G η ( T β ( A ) ) K = I g [ K | I ] J τ [ I | J ] a J = J I g [ K | I ] τ [ I | J ] a J ,
where A C I β 1 × × I β n , K [ [ I γ 1 ] ] × × [ [ I γ q ] ] . The corresponding tensor form is H = G η γ T β α , where H C I γ 1 × × I γ q × I β 1 × × I β n , and
h [ K | J ] = I g [ K | I ] τ [ I | J ] .
From this, we can present the following theorem:
Theorem 1. 
Let T β and G η be the linear transformations defined in (4) and let T β α and G η γ be their tensor representations. Denote the adjoints of T β and G η by T β * and G η * and the conjugate transposes of T β α and G η γ by ( T β α ) * and ( G η γ ) * , respectively. Then,
( G η T β ) * = T β * G η * , ( G η γ T β α ) * = ( T β α ) * ( G η γ ) * .
Proof. 
For any A C I γ 1 × × I γ q , B C I β 1 × × I β n , we have
T β * G η * ( A ) , B = T β * ( G η * ( A ) ) , B = G η * ( A ) , T β ( B ) = A , G η ( T β ( B ) ) = A , G η T β ( B ) = ( G η T β ) * ( A ) , B .
From the arbitrariness of A and B, we can obtain ( G η T β ) * = T β * G η * . Because the tensors G η γ and T β α are the tensor representations of the linear transformations G η and T β , respectively, the equation ( G η γ T β α ) * = ( T β α ) * ( G η γ ) * can be established immediately. □

3. QR Decomposition of the Tensor

QR decomposition is a fundamental tool in matrix theory and plays an important role in the design of algorithms. In this section, we aim to present the tensor QR decomposition of a given tensor from the perspective of linear transformation and describe the form of the triangular tensor based on it.
For the given tensor T, the partitioning ( α , β ) and the multi-index J [ [ I β 1 ] ] × × [ [ I β n ] ] , the process of orthogonalizing the tensors τ [ : | J ] is similar to the Gram–Schmidt orthogonalization process of the column vectors of a matrix. In this orthogonalization process, the tensor T and the tensors τ [ : | J ] play the roles of the matrix and column vectors, respectively. Suppose the only way to obtain J k J τ [ : | J ] = 0 is for all the scalars k J to be zero. The orthogonalization process of the tensors τ [ : | J ] then works as follows:
p [ : | 1 , 1 , , 1 ] = τ [ : | 1 , 1 , , 1 ] , q [ 1 , 1 , , 1 ] = p [ : | 1 , 1 , , 1 ] / p [ : | 1 , 1 , , 1 ] , p [ : | 2 , 1 , , 1 ] = τ [ : | 2 , 1 , , 1 ] τ [ : | 2 , 1 , , 1 ] , q [ 1 , 1 , , 1 ] q [ 1 , 1 , , 1 ] , q [ 2 , 1 , , 1 ] = p [ : | 2 , 1 , , 1 ] / p [ : | 2 , 1 , , 1 ] ,
where U = U , U 1 / 2 . In general, we have
p [ : | l 1 , l 2 , , l n ] = τ [ : | l 1 , l 2 , , l n ] k 1 = 1 I β 1 k n 1 = 1 I β n 1 k n = 1 l n 1 r k 1 , k 2 , , k n q [ k 1 , k 2 , , k n ] k 1 = 1 I β 1 k n 2 = 1 I β n 2 k n 1 = 1 l n 1 1 r k 1 , , k n 1 , l n q [ k 1 , , k n 1 , l n ] k 1 = 1 l 1 1 r k 1 , l 2 , , l n q [ k 1 , l 2 , , l n ] , q [ l 1 , l 2 , , l n ] = p [ : | l 1 , l 2 , , l n ] / p [ : | l 1 , l 2 , , l n ] ,
where r s 1 , s 2 , , s n = τ [ : | l 1 , l 2 , , l n ] , q [ s 1 , s 2 , , s n ] . Equation (6) can be rewritten as
τ [ : | l 1 , l 2 , , l n ] = k 1 = 1 I β 1 k n 1 = 1 I β n 1 k n = 1 l n 1 r k 1 , k 2 , , k n q [ k 1 , k 2 , , k n ] + k 1 = 1 I β 1 k n 2 = 1 I β n 2 k n 1 = 1 l n 1 1 r k 1 , , k n 1 , l n q [ k 1 , , k n 1 , l n ] + + k 1 = 1 l 1 1 r k 1 , l 2 , , l n q [ k 1 , l 2 , , l n ] + r l 1 , l 2 , , l n q [ l 1 , l 2 , , l n ] ,
where r l 1 , l 2 , , l n = p [ : | l 1 , l 2 , , l n ] . This expression leads us to obtain a tensor decomposition
T β α = Q R ,
where Q C I α 1 × × I α m × I β 1 × I β n satisfies
Q [ : | k 1 , k 2 , , k n ] = q [ k 1 , k 2 , , k n ] ,
R C I β 1 × I β n × I β 1 × I β n satisfies
R [ t 1 , t 2 , , t n | l 1 , l 2 , , l n ] = r t 1 , t 2 , , t n , ( t 1 , t 2 , , t n ) ( l 1 , l 2 , , l n ) , 0 , otherwise ,
and ( t 1 , t 2 , , t n ) ( l 1 , l 2 , , l n ) means
( t n 1 ) I β n 1 I β 1 + + ( t 2 1 ) I β 1 + t 1 ( l n 1 ) I β n 1 I β 1 + + ( l 2 1 ) I β 1 + l 1 .
Example 1. 
Consider the order-4 tensor T C 2 × 3 × 2 × 3 with the partitioning ( { 1 , 2 } , { 3 , 4 } ) . Applying the orthogonalization process, we can obtain the QR decomposition of T. The tensors Q , R C 2 × 3 × 2 × 3 can be laid out, respectively, as the "matrix" of 2 × 3 blocks of 2 × 3 matrices
Q = q 1111 q 1112 q 1113 q 1121 q 1122 q 1123 q 1211 q 1212 q 1213 q 1221 q 1222 q 1223 q 1311 q 1312 q 1313 q 1321 q 1322 q 1323 q 2111 q 2112 q 2113 q 2121 q 2122 q 2123 q 2211 q 2212 q 2213 q 2221 q 2222 q 2223 q 2311 q 2312 q 2313 q 2321 q 2322 q 2323
and
R = r 1111 r 1112 r 1113 r 1121 r 1122 r 1123 0 r 1212 r 1213 0 r 1222 r 1223 0 0 r 1313 0 0 r 1323 0 r 2112 r 2113 r 2121 r 2122 r 2123 0 0 r 2213 0 r 2222 r 2223 0 0 0 0 0 r 2323 .
When the tensor is an order-2 tensor, the QR decomposition of the tensor degenerates into the QR decomposition of the matrix. This is also in line with the fact that tensors are higher-order generalizations of matrices.
Based on the above QR decomposition, we can obtain the following definitions of triangular tensors, which are generalizations of those in matrix theory and different from those in [14].
Definition 3. 
For a given tensor T C I 1 × × I p and a partitioning ( α , β ) , we call T upper triangular (or, lower triangular, diagonal) under the partitioning ( α , β ) if τ [ i α 1 , , i α m | j β 1 , , j β n ] = 0 whenever
( i α 1 , , i α m ) > ( j β 1 , , j β n ) . ( or < , )
The elements τ [ i α 1 , , i α m | j β 1 , , j β n ] satisfying
( i α 1 , , i α m ) = ( j β 1 , , j β n )
are the diagonal elements, where ( i α 1 , , i α m ) > ( or < , = ) ( j β 1 , , j β n ) means the index ( i α 1 , , i α m , j β 1 , , j β n ) satisfies
( i α m 1 ) I α m 1 I α 1 + + ( i α 2 1 ) I α 1 + i α 1 > ( or < , = ) ( j β n 1 ) I β n 1 I β 1 + + ( j β 2 1 ) I β 1 + j β 1 .
Example 2. 
For a tensor T C 4 × 3 × 10 with the partitioning ( { 1 , 2 } , { 3 } ) , we mark its diagonal elements and upper triangular elements at the first row of Figure 1 for observation convenience, where the blue dots represent diagonal elements, and the red dots represent the strictly upper triangular elements. Additionally, we mark the upper triangular elements and the strictly upper triangular elements defined by [14] at the second row of Figure 1, where the red dots represent the strictly upper triangular elements, and the green and red dots together form the upper triangle elements. The slices in the figure are lateral slices [12], and the indices of the elements of T are denoted as ( i 1 , i 2 , i 3 ) .

4. Eigenvalue of the Tensor

The eigeninformation of a matrix is a fundamental concept in the field of matrix analysis and plays an important role in many practical applications. Therefore, it is necessary to consider the definition and existence of the eigeninformation of a tensor.
In linear algebra, if there exists a constant λ and a nonzero vector x such that A x = λ x , we call λ the eigenvalue of A, and call x the eigenvector associated with λ . The equation A x = λ x is bilinear and implies that the one-dimensional linear subspace expanded by eigenvectors associated with the same eigenvalue remains stable under the corresponding linear transformation. However, H-eigenvalues and Z-eigenvalues do not keep this property. With the benefit of the linear transformation, we present a new definition of the eigeninformation of an even-order tensor.
Firstly, we propose a new concept of square tensors from the viewpoint of an identity map, which is different from the definitions in [9,12].
Definition 4. 
For the tensor T C I 1 × × I 2 m , if there exists a partitioning ( α , β ) satisfying α = β = m and I α k = I β k , k = 1 , , m , then we call the tensor T a square tensor under the partitioning ( α , β ) .
Definition 5. 
Suppose T C I 1 × × I 2 m is a square tensor under the partitioning ( α , β ) , then we have a transformation T β : C I β 1 × × I β m C I α 1 × × I α m . We call λ C an ( α , β ) -eigenvalue of T, or an ( α , β ) -eigenvalue of T β , if when partnered with a nonzero tensor X C I β 1 × × I β m it satisfies
T β ( X ) = T β X = λ X ,
and the tensor X is called an ( α , β ) -eigentensor of T associated with the eigenvalue λ.
This definition is similar to that in [15], which conforms to our original intention of proposing the new definition of the tensor eigenvalue. The linear invariance of matrix eigensubspace is generalized.
Theorem 2. 
Suppose X 1 , X 2 , , X r are all eigentensors associated with the eigenvalue λ of the tensor T and k 1 , k 2 , , k r are complex numbers. Then, k 1 X 1 + k 2 X 2 + + k r X r is also the eigentensor associated with the eigenvalue λ of the tensor T.
Proof. 
Because X 1 , X 2 , , X r are all eigentensors associated with the eigenvalue λ , T β X i = λ X i , i = 1 , 2 , , r . Then,
T β ( k 1 X 1 + k 2 X 2 + + k r X r ) = k 1 T β X 1 + k 2 T β X 2 + + k r T β X r = k 1 λ X 1 + k 2 λ X 2 + + k r λ X r = λ ( k 1 X 1 + k 2 X 2 + + k r X r ) .
Thus, the theorem has been proven. □
It is worth noting that for an order-4 tensor, if the elements of T satisfy T i j k l = T i k j l , the eigenvalues of T under the partitioning ( { 1 , 2 } , { 3 , 4 } ) and the partitioning ( { 1 , 3 } , { 2 , 4 } ) are the same. Extending the conclusion to the general situation, we obtain that if T C I 1 × × I 2 m is square under the partitioning ( α , β ) and ( ξ , ζ ) and satisfies T β α = T ζ ξ , then the eigenvalues of T under the partitioning ( α , β ) and the partitioning ( ξ , ζ ) are the same. Importantly, if T is a supersymmetric tensor, the eigenvalues under arbitrary partitioning ( α , β ) that satisfy α = β are the same.
For a square tensor T C I 1 × × I 2 m under the partitioning ( α , β ) , if its corresponding transformation satisfies T β = T β * , we call the transformation a self-adjoint operator and call the tensor a Hermitian tensor under the partitioning ( α , β ) [9]. Based on this concept, and similar to the spectral theorem of a symmetric matrix, we present and prove the following lemma and theorem.
Lemma 1. 
Suppose the tensor T C I 1 × × I 2 m is Hermitian under the partitioning ( α , β ) , then the ( α , β ) -eigentensors corresponding to different ( α , β ) -eigenvalues of T are orthogonal.
Proof. 
Let λ 1 and λ 2 be two different ( α , β ) -eigenvalues of T, where X 1 and X 2 are ( α , β ) -eigentensors corresponding to λ 1 and λ 2 , respectively. Then, T β ( X 1 ) = T β X 1 = λ 1 X 1 , T β ( X 2 ) = T β X 2 = λ 2 X 2 . Therefore,
λ 1 X 1 , X 2 = T β ( X 1 ) , X 2 = X 1 , T β * ( X 2 ) = X 1 , T β ( X 2 ) = λ 2 X 1 , X 2 .
Because λ 1 λ 2 , X 1 , X 2 = 0 . □
Theorem 3. 
Suppose the tensor T C I 1 × × I 2 m is Hermitian under the partitioning ( α , β ) . Then, it has the decomposition
T β α = U Λ U * ,
where U C I α 1 × × I α m × I α 1 × × I α m is a unitary tensor [9], and Λ C I α 1 × × I α m × I α 1 × × I α m is a diagonal tensor whose diagonal elements are the eigenvalues under the partitioning ( α , β ) of T.
Proof. 
Let λ i 1 , i 2 , , i m , i 1 , i 2 , , i m be the ( α , β ) -eigenvalues of T, and let u [ : | i 1 , i 2 , , i m ] be the corresponding ( α , β ) -eigentensors, where i k [ [ I α k ] ] , k = 1 , , m . Then,
T β α u [ : | i 1 , i 2 , , i m ] = λ i 1 , i 2 , , i m , i 1 , i 2 , , i m u [ : | i 1 , i 2 , , i m ] .
Therefore,
U * T β α U j 1 , j 2 , , j m , i 1 , i 2 , , i m = u [ j 1 , j 2 , , j m | : ] * T β α u [ : | i 1 , i 2 , , i m ] = λ i 1 , i 2 , , i m , i 1 , i 2 , , i m u [ : | i 1 , i 2 , , i m ] , u [ : | j 1 , j 2 , , j m ] .
The results from Lemma 1 demonstrate that, when i k = j k , k = 1 , m , the above equation is equal to λ i 1 , i 2 , , i m , i 1 , i 2 , , i m ; otherwise, it is equal to 0. Furthermore, it can be concluded that U * T β α U = Λ ; thus, the theorem has been proven. □

5. Singular Value Decomposition of the Tensor

Based on the eigenvalue of a tensor in Section 4, we want to offer a new definition of tensor singular value decomposition, which is a generalization of the decomposition in [8]. To sufficiently demonstrate the feasibility of this decomposition, we first propose the following theorem and definition.
Theorem 4. 
All the eigenvalues of ( T β α ) * T β α or T β * T β are non-negative.
Proof. 
Suppose that ( λ , X ) is an arbitrary eigenpair of ( T β α ) * T β α , so it is also an eigenpair of T β * T β . Then, we can obtain
T β ( X ) , T β ( X ) = X , T β * ( T β ( X ) ) = X , T β * T β ( X ) = X , λ X = λ X , X .
Because T β ( X ) , T β ( X ) 0 and X , X > 0 , we have λ 0 . From the arbitrariness of the eigenpair, we can see that all the eigenvalues of ( T β α ) * T β α are non-negative. □
Definition 6. 
For a tensor T C I 1 × × I p with the partitioning ( α , β ) , we call the non-negative square root of an eigenvalue of ( T β α ) * T β α C I β 1 × × I β n × I β 1 × × I β n a singular value under the partitioning ( α , β ) of the tensor T.
Theorem 5. 
Each tensor T C I 1 × × I p with the partitioning ( α , β ) has a singular value decomposition under the partitioning (SVDUP)
T β α = U Σ V * ,
in which U C I α 1 × × I α m × I α 1 × × I α m and V C I β 1 × × I β n × I β 1 × × I β n are the unitary tensors, and Σ C I α 1 × × I α m × I β 1 × × I β n is a diagonal tensor whose diagonal elements are the singular values under the partitioning ( α , β ) of T.
Proof. 
Without loss of generality, we only prove the case when k = 1 α I α k k = 1 β I β k . ( T β α ) * T β α C I β 1 × × I β n × I β 1 × × I β n is a Hermitian tensor. From Theorem 3, it can be seen that there is a unitary tensor V C I β 1 × × I β n × I β 1 × × I β n and diagonal tensor Λ C I β 1 × × I β n × I β 1 × × I β n satisfying
( T β α ) * T β α = V Λ V * .
Let Σ 1 C I β 1 × × I β n × I β 1 × × I β n be a diagonal tensor whose diagonal elements are the square roots of the diagonal elements of Λ . Let Σ C I α 1 × × I α m × I β 1 × × I β n be the tensor by adding or deleting zeros in the tensor Σ 1 . U 1 C I α 1 × × I α m × I β 1 × × I β n is a tensor satisfying the linear equation U 1 Σ 1 = T β α V , that is, U 1 * T β α V = Σ 1 . Finally, we use the appropriate tensors that constitute the unitary tensor U C I α 1 × × I α m × I α 1 × × I α m together with U 1 . □
When p = 2 , T is a second-order tensor, in decomposition (9), U and V are matrices. It is clear that SVDUP is a generalization of matrix singular value decomposition in order. Therefore, we unify the concepts of a tensor singular value and unique left and right singular tensors. For the singular value decomposition of a matrix, the core matrix is a diagonal matrix, but the core tensor of HOSVD [2] is not. Additionally, it is impossible to transform higher-order tensors into a pseudodiagonal tensor by performing orthogonal transformations [2]. From this perspective, our definition would be a better way to generalize the singular value decomposition of matrices. In addition, to demonstrate the advantages of our SVDUP method in data compression, we also provide the following example.
Example 3. 
We randomly generate tensors using the command “create_problem” in the Tensor Toolbox, and compress them using truncated SVDUP and truncated HOSVD, respectively. Then, we compare the accuracy while ensuring that the compression ratio is almost the same in advance. In order to guarantee the generality of the experiment, we test 1000 data points of each size. The metric [16]
Δ ϕ = 1 T T ^ F T T ˜ F ,
where T ^ and T ˜ are the approximations of T computed by truncated SVDUP and truncated HOSVD, respectively, is introduced to describe the accuracy of the approximations. Δ ϕ > 0 means T ^ can provide a more accurate approximation. The compression ratio is equal to the number of elements contained in the approximate tensor divided by the number of elements in the original tensor. p represents order, and I represents dimension of each order.
Table 1 and Table 2 illustrate that as the tensor scale increases, our T-SVDUP can obtain higher accuracy approximations than T-HOSVD under a similar compression ratio.

6. Application

In both academic research and practical applications, there are many examples of mapping elements in one high-dimensional space to another. For example, a color image is mapped to a blurred image in 3D space during image processing, one discrete matrix is mapped to another in the process of the discretization of Poisson’s equation, and so on. The existing processing method is to stretch data into vectors, but the structure of the data is destroyed this way. In this section, we take two-dimensional image deblurring as an example to describe the application of SVDUP without the destruction of data structure.
In order to calculate the ( i , j ) pixels of the blurred pixel matrix Y, we should rotate the PSF (point spread function) array 180 degrees and match it with the pixels in the source pixel matrix X by placing the center of the rotated PSF array on the ( i , j ) pixels of X [17]. Then, the products of the corresponding components are summed to be y i j . For example, let F = ( f i j ) 3 × 3 , X = ( x i j ) 3 × 3 and Y = ( y i j ) 3 × 3 be the PSF array, the source pixel matrix and the blurred pixel matrix, respectively. Then,
y 22 = f 33 x 11 + f 32 x 12 + f 31 x 13 + f 23 x 21 + f 22 x 22 + f 21 x 23 + f 13 x 31 + f 12 x 32 + f 11 x 33 .
Particularly, if we assume zero boundary conditions, the elements y i j at the border of Y are given by
y 11 = f 22 x 11 + f 21 x 12 + f 12 x 21 + f 11 x 22 , y 21 = f 32 x 11 + f 31 x 12 + f 22 x 21 + f 21 x 22 + f 12 x 31 + f 11 x 32 , y 31 = f 32 x 21 + f 31 x 22 + f 22 x 31 + f 21 x 32 , y 12 = f 23 x 11 + f 22 x 12 + f 21 x 13 + f 13 x 21 + f 12 x 22 + f 11 x 23 , y 33 = f 33 x 22 + f 32 x 23 + f 23 x 32 + f 22 x 33 .
We assume T { 3 , 4 } : R 3 × 3 R 3 × 3 to be the map from the source pixel matrix to the blurred pixel matrix. Then, the element of its tensor representation T is
T i j k l = f 2 k + i , 2 l + j , 1 k i , l j 1 , 0 , otherwise .
Then, the transformation can be summarized as Y = T { 3 , 4 } X . The image deblurring problem is recovering the source pixel matrix X from the blurred pixel matrix Y = T { 3 , 4 } X + Θ , where Θ is a kind of noise.
We use the SVDUP T { 3 , 4 } { 1 , 2 } = U Σ V * to analyze this problem. The solution can be written as
X = V Σ 1 U * B = i = 1 I 1 × I 2 U · i , B σ i V · i ,
where X R I 1 × I 2 , B = Y Θ , U · i = U : , : , l , j , V · i = V : , : , l , j , satisfying i = l + I 1 ( j 1 ) , and σ i is the singular value satisfying σ 1 σ 2 σ I 1 × I 2 0 . In general, for the blurring tensor, σ I 1 × I 2 0 , and the number σ 1 / σ I 1 × I 2 is large. The approach that is frequently used to dampen the effects of small singular values is to discard all SVDUP components with small singular values, that is, constructing the approximation
X k = i = 1 k U · i , B σ i V · i , k < I 1 × I 2 ,
where σ j ( j k + 1 ) is smaller than the given tolerance. This method is called the truncated SVDUP, which is a particular case of the spectral filtering methods
X f i l t = i = 1 I 1 × I 2 ϕ i U · i , B σ i V · i ,
where the filter factors ϕ i are chosen such that ϕ i 1 for large singular values, and ϕ i 0 for small singular values [17]. To some extent, this method can compensate for the errors caused by the truncation of singular values.
Suppose the original image is blurred with a circular averaging blur kernel, and perturbed by Gaussian white noise. Figure 2 shows the image recovered using truncated SVDUP with different truncations. All of our computations were performed in MATLAB 2016a running on a PC Intel(R) Core(TM) i7-7500 of 2.7 GHZ CPU and 8 GB of RAM.

7. Conclusions

In this paper, from the perspective of linear transformation, we present some basic concepts and provide three tensor decompositions: QR decomposition, spectral decomposition and singular value decomposition. These concepts and decompositions are different from the existing results and are more natural generalizations of the corresponding matrix theory. The numerical results also show the effectiveness of our new decomposition method.

Author Contributions

Conceptualization, B.Y. and B.D.; methodology, X.Z. and B.D.; software, X.Z.; writing—original draft preparation, X.Z and B.D.; writing—review and editing, X.Z. and Y.Y.; funding acquisition, Y.Y., B.Y., B.D. and X.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China (Grant Nos. 11871136, 11971092 and 11801382), Natural Science Foundation of Liaoning Province, China (Grant Nos. 2020-MS-278 and 2023-MS-142), Scientific Research Foundation of Education Department of Liaoning Province, China (Grant Nos. 2020JYT04, LJKZ0096 and LJKMZ20220452).

Data Availability Statement

Data sharing not applicable to this paper, as no data sets were generated or analyzed during the current study.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Qi, L. Eigenvalues of a real supersymmetric tensor. J. Symbolic Comput. 2005, 40, 1302–1324. [Google Scholar] [CrossRef]
  2. De Lathauwer, L.; De Moor, B.; Vandewalle, J. A multilinear singular value decomposition. SIAM J. Matrix Anal. Appl. 2000, 21, 1253–1278. [Google Scholar] [CrossRef]
  3. Chen, W.J.; Yu, S.W. RSVD for Three Quaternion Tensors with Applications in Color Video Watermark Processing. Axioms 2023, 12, 232. [Google Scholar] [CrossRef]
  4. Lim, L.H. Singular values and eigenvalues of tensors: A variational approach. In Proceedings of the IEEE International Workshop on Computational Advances in Multi-Sensor Adaptive Processing, Puerto Vallarta, Mexico, 13–15 December 2005; pp. 129–132. [Google Scholar]
  5. Chen, Z.; Lu, L. A tensor singular values and its symmetric embedding eigenvalues. J. Comput. Appl. Math. 2013, 250, 217–228. [Google Scholar] [CrossRef]
  6. Ragnarsson, S.; Van Loan, C.F. Block tensors and symmetric embeddings. Linear Algebra Appl. 2013, 438, 853–874. [Google Scholar] [CrossRef]
  7. Chang, K.; Qi, L.; Zhou, G. Singular values of a real rectangular tensor. J. Math. Anal. Appl. 2010, 370, 284–294. [Google Scholar] [CrossRef]
  8. Brazell, M.; Li, N.; Navasca, C.; Tamon, C. Solving multilinear systems via tensor inversion. SIAM J. Matrix Anal. Appl. 2013, 34, 542–570. [Google Scholar] [CrossRef]
  9. Guan, Y.; Chu, M.T.; Chu, D. SVD-based algorithms for the best rank-1 approximation of a symmetric tensor. SIAM J. Matrix Anal. Appl. 2018, 39, 1095–1115. [Google Scholar] [CrossRef]
  10. Li, L.; Victoria, B. MATLAB User Manual; MathWorks: Natick, MA, USA, 1999. [Google Scholar]
  11. Ragnarsson, S.; Van Loan, C.F. Block tensor unfoldings. SIAM J. Matrix Anal. Appl. 2012, 33, 149–169. [Google Scholar] [CrossRef]
  12. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  13. Marchuk, G.I. Construction of adjoint operators in non-linear problems of mathematical physics. Sb. Math. 1998, 189, 1505–1516. [Google Scholar] [CrossRef]
  14. Ding, W.; Wei, Y. Solving multi-linear systems with M-tensors. J. Sci. Comput. 2016, 68, 689–715. [Google Scholar] [CrossRef]
  15. Cui, L.B.; Chen, C.; Wen, L.; Ng, M.K. An eigenvalue problem for even order tensors with its applications. Linear Multilinear Algebr. Int. J. Publ. Artic. Rev. Probl. 2016, 64, 602–621. [Google Scholar] [CrossRef]
  16. Silva, A.P.D.; Comon, P.; Almeida, A.L.F.D. A Finite Algorithm to Compute Rank-1 Tensor Approximations. IEEE Signal Process. Lett. 2016, 23, 959–963. [Google Scholar] [CrossRef]
  17. Hansen, P.C.; Nagy, J.G.; O’Leary, D.P. Deblurring Images: Matrices, Spectra, and Filtering; SIAM: Philadelphia, PA, USA, 2006. [Google Scholar]
Figure 1. Upper triangular elements under different definitions.
Figure 1. Upper triangular elements under different definitions.
Axioms 12 00485 g001
Figure 2. Deblurring with different truncations.
Figure 2. Deblurring with different truncations.
Axioms 12 00485 g002
Table 1. Comparison between T-SVDUP and T-HOSVD with I = 5 .
Table 1. Comparison between T-SVDUP and T-HOSVD with I = 5 .
Order Compression Ratio Δ ϕ > 0 Δ ϕ < 0
p = 4 9 % 6 994
p = 6 5 % 872 128
p = 8 1.7 % 1000 0
Table 2. Comparison between T-SVDUP and T-HOSVD with p = 6 .
Table 2. Comparison between T-SVDUP and T-HOSVD with p = 6 .
Dimension Compression Ratio Δ ϕ > 0 Δ ϕ < 0
I = 5 5 % 872 128
I = 6 9 % 1000 0
I = 7 14 % 1000 0
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhao, X.; Dong, B.; Yu, B.; Yu, Y. Tensor Eigenvalue and SVD from the Viewpoint of Linear Transformation. Axioms 2023, 12, 485. https://doi.org/10.3390/axioms12050485

AMA Style

Zhao X, Dong B, Yu B, Yu Y. Tensor Eigenvalue and SVD from the Viewpoint of Linear Transformation. Axioms. 2023; 12(5):485. https://doi.org/10.3390/axioms12050485

Chicago/Turabian Style

Zhao, Xinzhu, Bo Dong, Bo Yu, and Yan Yu. 2023. "Tensor Eigenvalue and SVD from the Viewpoint of Linear Transformation" Axioms 12, no. 5: 485. https://doi.org/10.3390/axioms12050485

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop