Next Article in Journal
Influence of Numerical Aperture on Molten Area Formation in Fusion Micro-Welding of Glass by Picosecond Pulsed Laser
Next Article in Special Issue
Image Recovery with Data Missing in the Presence of Salt-and-Pepper Noise
Previous Article in Journal
Theoretical Investigation of a Simple Design of Triple-Band Terahertz Metamaterial Absorber for High-Q Sensing
Previous Article in Special Issue
Fast Cylinder Shape Matching Using Random Sample Consensus in Large Scale Point Cloud
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Tensor Robust Principal Component Analysis via Non-Convex Low Rank Approximation

1
School of Automation, Guangdong University of Technology, Guangzhou 510006, China
2
Department of Mathematics, Southern Illinois University, Carbondale, IL 62901, USA
3
Department of Computer Science, Southern Illinois University, Carbondale, IL 62901, USA
4
School of Mathematical Sciences, South China Normal University, Guangzhou 510631, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2019, 9(7), 1411; https://doi.org/10.3390/app9071411
Submission received: 25 January 2019 / Revised: 25 March 2019 / Accepted: 29 March 2019 / Published: 3 April 2019
(This article belongs to the Special Issue Machine Learning and Compressed Sensing in Image Reconstruction)

Abstract

:
Tensor Robust Principal Component Analysis (TRPCA) plays a critical role in handling high multi-dimensional data sets, aiming to recover the low-rank and sparse components both accurately and efficiently. In this paper, different from current approach, we developed a new t-Gamma tensor quasi-norm as a non-convex regularization to approximate the low-rank component. Compared to various convex regularization, this new configuration not only can better capture the tensor rank but also provides a simplified approach. An optimization process is conducted via tensor singular decomposition and an efficient augmented Lagrange multiplier algorithm is established. Extensive experimental results demonstrate that our new approach outperforms current state-of-the-art algorithms in terms of accuracy and efficiency.

1. Introduction

An array of numbers arranged on a regular grid with a variable number of axes is described as a tensor. As the dimensionality reaches beyond 3D, conventional data representations such as vector-based and matrix-based become insufficient, and highly dimensional data sets are usually formulated as tensors [1,2,3].
Robust Principal Component Analysis (RPCA) [4,5,6,7,8], aiming to recover the low-rank and sparse matrices both accurately and efficiently, has been widely studied in data compressed sensing and computer vision. It is of great use when we try to solve the problem of subspace learning or Principal Component Analysis (PCA) in the presence of outliers [9]. Especially RPCA is a very useful tool in foreground detection, which is the first step in video surveillance system to detect moving objects [10,11]. Tensor Robust Principal Component Analysis (TRPCA) extends the RPCA from matrix to the tensor case, that is, to recover a low-rank tensor and a sparse (noise entries) tensor.
Tensor rank [12] essentially characterizes the intrinsic "dimension” or “degree of freedom” for highly dimensional data sets. In order to make the rank characteristic be more transparent, one can make use of tensor Singular Value Decomposition (t-SVD) developed recently by Kilmer and Martin [13], rendering the tensor rank approximation approachable. The key for their framework is the well-defined t-product that leads to the 3rd-order tensor space to be an inner product space so that approximation becomes more effective such as in image processing [14,15]. t-SVD allows to define tensor tubal-rank, which appears to be much more approachable such as in the study of image analysis than previous rank definitions [16,17,18]. The characteristic of a low tensor tubal-rank implies that a 3rd-order tensor can be significantly compressed (Theorem 4.3 of [13]), which is particularly important in applications.
Mathematically, TRPCA based on t-SVD [19,20] aims to recover the low tubal rank component L 0 and sparse component E 0 from noisy observations X = L 0 + E 0 of size N 1 × N 2 × N 3 by convex optimization
min L , E L T N N + λ E 1
s . t . X = L + E .
λ is a Lagrange multiplier, E is sparse if the tensor 1 norm E 1 is small. It is well-known that the t-SVD-based tensor nuclear norm (TNN, III.B of [21]) has been proven to be the tightest convex relaxation to 1 -norm of the tensor multi-rank (Theorem 2.4.1 in [15] or Theorem 3.1 in [19]). However, TNN treats the singular values in t-SVD equally to pursue the convexity of the objective function, while the singular values associated with a practical image may emphasize its different properties and should be handled differently, in particular, when some singular values are very large. Due to the gap between the tensor rank function and its lower convex envelop, the adoption of tensor nuclear norm may lead to the approximation of the corresponding tensor tubal-rank being inaccurate. Therefore, a new development of tensor tubal-rank approximation becomes necessary to reflect these fundamental characteristics.
To make the the necessity of a new tensor tubal-rank approximation instead of TNN more transparent, we consider the the matrix RPCA [4] first. Now the goal becomes to recover a low-rank matrix L from highly corrupted observations X = L + S , by minimizing L * + λ S 1 . Here L * is the summation of the singular values of L. It is easy to see that
1 + σ 2 σ 1 + + σ N σ 1 = 1 σ 1 L * rank ( L ) .
Hence, when σ 1 is very large, for example, and the rest are relatively small, we may have 1 σ 1 L * < < rank ( L ) . Then the minimization of L * will attempt to minimize σ 1 that leads to over-penalizes σ 1 , and consequently it generates undesirable solution L with small σ 1 . Recall that our goal is to find a low-rank L, whose largest singular value is not necessarily to be small. In addition, it is worth mentioning that in practice the matrix RPCA problem is very sensitive to small perturbations and this problem can be addressed by relaxing the equality constraints. The authors in [22,23] propose an alternative optimization approach that appears to be suitable to deal with robustness issues in the “Sparse Plus Low-rank” decomposition problem. It has long been noticed that the nuclear norm is not always satisfactory [24], since it over-penalizes large singular values, and consequently may only find a much biased solution. Since the the nuclear norm is the tightest convex relaxation of the rank function [25], a further improvement under the convex setting in general is limited, if it is still possible, in particular when their gap is not very small. In order to alleviate the above problem, the Gamma quasi-norm for matrices is introduced in [26] in RPCA, which results in a tighter approximation of the rank function. For L R m × n and the singular values of L, denoted by σ 1 ( L ) , , σ min { m , n } ( L ) , the Gamma quasi-norm of L is defined as
| | L | | γ = i = 1 min { m , n } ( 1 + γ ) σ i ( L ) γ + σ i ( L ) , γ > 0 .
It is not difficult to see
( 1 + γ ) σ i ( L ) γ + σ i ( L ) < 1 + γ γ + M σ i ( L ) ,
when σ i ( L ) > M > 0 and | | L | | γ ( 1 + γ ) rank ( L ) . So it overcomes the imbalanced penalization by different singular values in convex nuclear norm. In addition, when γ goes to 0, the limit of the Gamma quasi-norm is exactly the matrix rank. The Gamma quasi-norm has nice algebraic properties, i.e., unitarily invariant and positive definiteness. Thanks to the unitarily invariance [27], the Gamma quasi-norm minimization problem for A R m × n :
min Z R m × n | | Z | | γ + μ 2 Z - A F 2
can be solved very effectively (Theorem 1 of [26]). Moreover, experimental results in [26] show that Gamma quasi-norma-based RPCA outperforms several other convex/non-convex norma-based RPCA in both accuracy and efficiency.
In this paper, inspired by the success of the non-convex rank approximation [26], we proposed tensor Gamma quasi-norm via t-SVD to seek a closer approximation and to alleviate the above-mentioned limitations resulted from the tensor nuclear norm, which has not been used for TRPCA yet in literature.
The main contributions of this paper are illustrated as follows:
  • We develop new non-convex tensor rank approximation functions that can be used as desirable regularization for the optimization process, which appears to be more effective than current existing approaches in literature.  
  • An efficient algorithm based on the alternating direction method of multipliers (ADMM) is developed to solve the equivalent optimization problem in the Fourier domain, which is also suitable for general TRPCA problem. In addition, we provided the convergent analysis of the augmented Lagrange multiplier based optimization algorithm.  
  • Extensive evaluation of the proposed approach on several benchmark data sets has been conducted, and detailed comparison with most recent approaches are provided. Experimental results demonstrate that our proposed approach yields superior performance for image recovery and video background modeling.
The rest of the paper is organized as follows. In Section 2, we roughly describe some related works. Then Section 3 provides factorization strategies for 3rd-order tensors and its relationship to the corresponding tensor approximation problem. And then we present the main results on t-Gamma quasi-norm of 3rd-order tensors. Section 4 introduces our non-convex TRPCA framework and the convergent analysis of the augmented Lagrange multiplier based optimization algorithm. Experimental analysis and results are shown in Section 5 to verify the proposed algorithm. Finally, the paper ends with concluding remarks in Section 6.

2. Related Works

Recently, many authors observed the limitations of t-SVD-based TRPCA [15,19]. It was pointed out in [28] that the classical TRPCA methods based on t-SVD fails to utilize the low rank structure in the third mode. To further exploit the low-rank structures in multiway data, they defined a new TNN that extends TNN with core matrix and proposed a creative algorithm to deal with TRPCA problems. In addition, the authors of [29] presented that there are some weak points in the existing low-rank approximation approaches, for example, people need to predefine the rank values or fail to consider local information of data (e.g., spatial or spectral smooth structure). So they proposed a novel TNN-based low-rank approximation with total variation regularization, which uses the TNN to encode the global low-rank prior of tensor data and the total variation regularization to preserve the spatial-spectral continuity in a unified framework.
It has been noticed by others that TNN is not always satisfactory. For instance, to seek a closer approximation of tensor rank, Kong et al. in [30] introduced t-Schatten-p tensor quasi-norm to replace the TNN. The t-Schatten-p norm is convex when p 1 , but non-convex when 0 < p < 1 . In addition, they proved multi-Schatten-p norm surrogate theorem to convert the non-convex t-Schatten-p tensor quasi-norm for 0 < p < 1 into a weighted sum of convex functions. Then based on the multi-Schatten-p norm surrogate, they proposed a novel convex tensor RPCA model.
Besides above, a recent paper [31] developed randomized algorithms for tensor low-rank approximation and decomposition, together with tensor multiplication. Their proposed tubal focused algorithms employ a small number of lateral and/or horizontal slices of the underlying 3rd order tensor, that come with relative error guarantees for the quality of the obtained solutions.
Besides the above-mentioned methods based on t-SVD, there are many other TRPCA methods mainly applied CP (CANDECOMP/PARAFAC) or Tucker decompositions methods in their optimization models. We need to mention that recently Driggs et al. [32] studied TRPCA using a non-convex formulation of the CP based tensor atomic-norm and identified a class of local minima of this non-convex program that are globally optimal.
Presently, background subtraction become more and more important for visual surveillance systems and several subspace learning algorithms based on matrix and tensor tools have been used to perform the background modeling of the scenes. Recently, Sobral et al. [33] developed an incremental tensor subspace learning that uses only a small part of the entire data and updates the low-rank model incrementally when new data arrives and in [34] the authors proposed an online stochastic framework for tensor decomposition of multispectral video sequences. In particular, it is worth mentioning that Bouwmans et al. [35] gave a systematic and complete review of the "low-rank plus sparse matrix decomposition" framework of separating moving objects from the background.
However, all those above mentioned t-SVD-based TRPCA algorithms solve the TRPCA optimization problem under convex regularization. In addition, most analyses of ADMM algorithms are in the convex setting [36]. In our work, we explore the TRPCA method based on the non-convex t-Gamma quasi-norm of 3rd-order tensors. Due to the non-convex nature of our TRPCA model, the convergent analysis of algorithm is difficult to justify. In addition, the analysis of non-convex augmented Lagrangians is not easy, but this compels us to make efforts of clarity and pedagogy. Also it should be noted that there are few works on t-SVD-based TRPCA achieve background modeling. Thus experiments are provided in our paper on two computer vision applications, which are image recovery and video background modeling.

3. Notations and Definitions

Throughout this paper, all boldface Euler script letters ( X ) denote tensors, uppercase characters (X) denote matrices, and lowercase letters (a) denote scalars. The field of real numbers is denoted as R . We denote N i , i = 1 , 2 , 3 as the index set associated with a given tensor. For 3rd-order tensors X = [ a i j k ] R N 1 × N 2 × N 3 , we use X ( i , : , : ) , X ( : , i , : ) , X ( : , : , i ) to denote the i-th horizontal, lateral, and frontal slice, respectively. For simplicity, we denote the frontal slice X ( : , : , i ) as X i . The inner product between X and Y of size N 1 × N 2 × N 3 is defined as X , Y = i = 1 N 3 tr ( X i T Y i ) . The Frobenius norm of a tensor is defined as X F = i , j , k a i j k 2 . The  1 norm of a tensor is defined as X 1 = i , j , k a i j k . In this paper, the adopted basic definitions and notations follow the references [13,21,37] for the purpose of an easier comparison.
Definition 1
(Transpose p. 10 of [13]). For X R N 1 × N 2 × N 3 , the transpose tensor X T is an N 2 × N 1 × N 3 tensor obtained by transposing each frontal slice of X and then reversing the order of the transposed frontal slices 2 through N 3 .
Similar to matrices whose elements can be grouped by rows or by columns, higher dimensional tensors can be viewed forming by slices along the third mode. For X = [ X 1 | | X N 3 ] R N 1 × N 2 × N 3 with N 1 × N 2 frontal slices X 1 , , X N 3 . Then the block circulant is created from the frontal slices in a form of
bcirc ( X ) = X 1 X N 3 X N 3 - 1 X 2 X 2 X 1 X N 3 X 3 X N 3 X N 3 - 1 X 2 X 1 ,
where X i = X ( : , : , i ) for i = 1 , , N 3 . It is important to notice that the block circulant matrix bcirc ( X ) keeps the order of frontal slices of X in an appropriate way, and thus it better maintains X ’s structure in terms of the frontal direction, compared with the direct tensor unfolding along the third mode.
Definition 2
(t-product p. 11 of [13]). Let X R N 1 × N 2 × N 3 , we define the matvec and fold operators as m a t v e c ( X ) = X 1 ; X 2 ; ; X N 3 is an N 1 N 3 × N 2 matrix and f o l d ( m a t v e c ( X ) ) = X .
Let A R N 1 × N 4 × N 3 , and B R N 4 × N 2 × N 3 . Then the t-product of 3rd-order tensors is defined as
A * B = f o l d ( b c i r c ( A ) m a t v e c ( B ) )
which is a tensor of size N 1 × N 2 × N 3 .
Definition 3
(Identity, orthogonal and diagonal tensor p.10 of [13]). The identity tensor I R N 1 × N 1 × N 3 is a tensor whose first frontal slice is the N 1 × N 1 identity matrix and all other frontal slices are zero. A tensor Q R N 1 × N 1 × N 3 is orthogonal if Q T * Q = Q * Q T = I , where * is the t-product. A tensor is called f-diagonal if each of its frontal slices is diagonal matrix.
It is well known that circulant matrices can be diagonalized by the FFT (21.11 of [38]), similarly, block-circulant matrices can also block-diagonalized as
F ˜ N 3 I N 1 · bcirc ( X ) · F ˜ N 3 * I N 2 = bdiag ( X ^ ) = X ^ 1 X ^ N 3 ,
where F ˜ N 3 = 1 N 3 F N 3 , F N 3 denotes the N 3 × N 3 fast Fourier transform matrix, F N 3 * denotes its conjugate, and the notation ⊗ denotes the standard Kronecker product. In addition, X = A * B if and only if X ^ n 3 = A ^ n 3 B ^ n 3 for 1 n 3 N 3 (p.4 of [19]).
Since F ˜ N 3 I N 1 , F ˜ N 3 I N 2 are unitary matrices, we have
bdiag ( X ^ ) F 2 = tr bcirc ( X ) T bcirc ( X ) = N 3 | | X | | F 2 .
Then we have X F = 1 N 3 bdiag ( X ^ ) F .
Definition 4
(Tensor multi-rank p. 10 of [37]). For tensor X R N 1 × N 2 × N 3 with N 1 × N 2 frontal slices X 1 , , X N 3 , the multi-rank of X is a N 3 -vector [ r a n k ( X ^ 1 ) , , r a n k ( X ^ N 3 ) ] , consisting of the ranks of all the X ^ 1 , , X ^ N 3 .
Please note that the matrix SVD [39] can be performed on each frontal slice of X ^ , The matrix X ^ n 3 has the following Singular Value Decomposition (SVD) X ^ n 3 = U ^ n 3 Σ ^ n 3 V ^ n 3 T , U ^ n 3 O ( N 1 ) , V ^ n 3 O ( N 2 ) (here O ( N i ) stands for the set of orthogonal N i -by- N i matrices) and Σ ^ n 3 is an N 1 × N 2 diagonal matrix with diagonal entries σ 1 ( X ^ n 3 ) , , σ min { N 1 , N 2 } ( X ^ n 3 ) in descending order.
Definition 5
(t-SVD, Theorem 4.1 in [13]). The tensor Singular Value Decomposition (t-SVD) of X is given by X = U * Σ * V T , where U and V are orthogonal tensors of size N 1 × N 1 × N 3 and N 2 × N 2 × N 3 respectively. Σ is an f-diagonal tensor of size N 1 × N 2 × N 3 , and * denotes the t-product.
Definition 6
(Tensor tubal rank III.B of [21]). The tensor tubal rank is defined as the number of nonzero singular tubes of f-diagonal tensor Σ, which is the maximum of the number of nonzero { Σ ^ n , n , n 3 , 1 n min ( N 1 , N 2 ) } for 1 n 3 N 3 .
Definition 7
(Tensor nuclear norm III.B of [21]). The nuclear norm for a given 3rd-order tensor X R N 1 × N 2 × N 3 is defined to be
X T N N = 1 N 3 n 3 = 1 N 3 | | X ^ n 3 | | * ,
which is equal to
n = 1 min ( N 1 , N 2 ) n 3 = 1 N 3 1 N 3 Σ ^ n , n , n 3 .
Here X ^ n 3 is the n 3 -th frontal slice of X ^ = fft ( X , [ ] , 3 ) = [ X ^ 1 | | X ^ N 3 ] R N 1 × N 2 × N 3 obtained by applying FFT on X ^ along the third mode.
In this paper, to overcome the disadvantages of the tensor nuclear norm, we propose smooth but more effective, though being non-convex, t-SVD-based non-convex rank approximation to the data sets represented by 3rd-order tensor. This will make the larger singular values be shrunk less in order to preserve their characteristics. Moreover, under our framework, we establish explicit solutions to the corresponding non-convex tensor norm minimization problems that is not available in current literature.
Definition 8
(Gamma quasi-norm for matrix [26]). Let X R m × n and the singular values of X, denoted by σ 1 ( X ) , , σ min { m , n } ( X ) . We could introduce the Gamma norm of X, which is defined as | | X | | γ = i = 1 min m , n ( 1 + γ ) σ i ( X ) γ + σ i ( X ) , γ > 0 .
Definition 9
(t-Gamma tensor quasi-norm). The Gamma tensor quasi-norm for a given 3rd-order tensor X R N 1 × N 2 × N 3 is defined to be
| | X | | t - γ = 1 N 3 n 3 = 1 N 3 | | X ^ n 3 | | γ ,
which is equal to
1 N 3 n 3 = 1 N 3 n = 1 min { N 1 , N 2 } ( 1 + γ ) Σ ^ n , n , n 3 γ + Σ ^ n , n , n 3 .
Here X ^ n 3 is the n 3 -th frontal slice of X ^ = fft ( X , [ ] , 3 ) = [ X ^ 1 | | X ^ N 3 ] R N 1 × N 2 × N 3 obtained by applying FFT on X ^ along the third mode.
Remark 1.
It is not hard to see the t-Gamma tensor quasi-norm is unitarily invariant, that is | | X | | t - γ = | | U * X * V T | | t - γ , where U and V are orthogonal tensors of size N 1 × N 1 × N 3 and N 2 × N 2 × N 3 respectively. In addition,
lim γ 0 | | X | | t - γ = tubal rank of X ; lim γ | | X | | t - γ = | | X | | T N N .
Lemma 1
(Theorem 1 of [26]). Let A = U Σ A V T be the SVD of A R m × n and Σ A = d i a g ( σ A ) are singular values of A. Let F ( Z ) = f σ ( Z ) be a unitary invariant function and μ > 0 . Then an optimal solution to the following problem
min Z F ( Z ) + μ 2 Z - A F 2 ,
is Z * = U Σ Z * V T , where Σ Z * = d i a g ( σ * ) and σ * = prox f , μ ( σ A ) . Here prox f , μ ( σ A ) is the Moreau-Yosida operator [40], defined as
prox f , μ ( σ A ) : = arg min σ 0 f ( σ ) + μ 2 σ - σ A 2 2 .
Now let us consider the tensor t-Gamma quasi-norm minimization problem:
Theorem 1
(t-Gamma tensor rank minimization). Let us consider a 3rd-order tensor X R N 1 × N 2 × N 3 , which has t-SVD X = U * Σ * V T as in Definition 5. If Z * minimizes
1 2 X - Z F 2 + τ Z t - γ ,
then
Z * = U * Ω Z * V T ,
where Ω Z R N 1 × N 2 × N 3 is a f-diagonal tensor. For 1 n min ( N 1 , N 2 ) , 1 n 3 N 3 , let Σ ^ = [ Σ ^ n , n , n 3 ] = fft ( Σ , [ ] , 3 ) , Ω ^ Z = [ Ω ^ n , n , n 3 ] = fft ( Ω , [ ] , 3 ) , then Ω ^ n , n , n 3 is the limit point of Fixed Point Iteration
Ω ^ n , n , n 3 ( k + 1 ) = ( Σ ^ n , n , n 3 - τ γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ( k ) ) 2 ) + .
Remark 2.
Theorem 1 provides the connection of the optimal solution between the original domain and the Fourier domain. More importantly, it shows that the minimization problem under our framework is tractable. After the original tensor X is transformed to X ^ in the Fourier domain, an equivalent minimization problem could be solved by using Lemma 1 for each slice X ^ n 3 , 1 n 3 N 3 in the Fourier domain. Then the optimal solution Z * in the original domain can be obtained by the inverse Fourier transform Z ^ * Z * , whose detailed approach can be found from the proof in Theorem 1.
(Proof of Theorem 1)
Let f ( Z ) = 1 2 X - Z F 2 + τ | | Z | | t - γ , then it can be reformulated as matrix objective function in Fourier domain:
f ^ ( Z ^ ) = 1 2 bdiag ( X ^ - Z ^ ) F 2 + τ bdiag ( Z ^ ) γ = n 3 = 1 N 3 1 2 X ^ n 3 - Z ^ n 3 F 2 + τ Z ^ n 3 γ .
Without loss of generality, for each diagonal block Z ^ n 3 of Z ^ , 1 n 3 N 3 , let us consider
Z ^ n 3 * = arg min Z ^ n 3 1 2 X ^ n 3 - Z ^ n 3 F 2 + τ Z ^ n 3 γ .
From Lemma 1, Z ^ n 3 * = U ^ X , n 3 Ω ^ Z , n 3 V ^ X , n 3 T , and
Ω ^ n , n , n 3 = arg min ω > 0 1 2 ( ω - Σ ^ n , n , n 3 ) 2 + τ ( 1 + γ ) ω γ + ω .
Following the DC programming algorithm used in Theorem 1 of [26] (similar method has been used in Section III of [41]), we can solve it efficiently using a local minimization method.
Let f ( ω ) = ( 1 + γ ) ω γ + ω , it can be approximated using first order Taylor expansion (with Big “O” truncation error) as
f ( ω ) = f ( Ω ^ n , n , n 3 ( k ) ) + γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ( k ) ) 2 ( ω - Ω ^ n , n , n 3 ( k ) ) + O ( ( ω - Ω ^ n , n , n 3 ( k ) ) 2 ) .
Now we consider
Ω ^ n , n , n 3 ( k + 1 ) = arg min ω > 0 1 2 ( ω - Σ ^ n , n , n 3 ) 2 + τ γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ( k ) ) 2 | ω | ,
and it admits a closed-form solution
Ω ^ n , n , n 3 ( k + 1 ) = ( Σ ^ n , n , n 3 - τ γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ( k ) ) 2 ) + ,
where ( x ) + = max { x , 0 } . Here Ω ^ n , n , n 3 ( k + 1 ) is the solution obtained in the ( k + 1 ) -th iteration. Let
g ( Ω ^ n , n , n 3 ) = ( Σ ^ n , n , n 3 - τ γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ) 2 ) + , 0 Ω ^ n , n , n 3 < Σ ^ n , n , n 3 ,
it can be easily seen that | g ( Ω ^ n , n , n 3 ) |   <   1 , when 2 τ ( 1 + γ ) < γ 2 . The Fixed Point Iteration Ω ^ n , n , n 3 ( k + 1 ) = g ( Ω ^ n , n , n 3 ( k ) ) converge to the local minimum Ω ^ n , n , n 3 after several iterations (Chapter 1 of [42]), and the proof is thus completed. ☐

4. Tensor Robust Principal Component Analysis with Non-Convex Regularization

A TRPCA model using non-convex regularization can be formulated as
min L , E L t - γ + λ E 1 s . t . X = L + E .
Its augmented Lagrangian function is L β ( L , E , M ) = L t - γ + λ E 1 + β 2 X - L - E - M β F 2 + C , where M is the Lagrangian multiplier, β is the penalty parameter for the violation of the linear constraints, and C is a constant.
Then, the problem arg min L , E , M L β ( L , E , M ) can be updated as:
L k + 1 = arg min L X - E k - M k β - L F 2 + 2 β L t - γ , E k + 1 = S λ β X - L k + 1 + M k β , M k + 1 = M k + β ( X - L k + 1 - E k + 1 ) ,
where the tensor non-negative soft-thresholding operator E v ( · ) [43] is defined as
S v ( B ) = B ¯
with
b ¯ n 1 n 2 n 3 = s g n ( b n 1 n 2 n 3 ) ( | b n 1 n 2 n 3 | - v ) , | b n 1 n 2 n 3 |   >   v , 0 , | b n 1 n 2 n 3 |     v .
Algorithm 1 shows the pseudocode for the proposed non-convex tensor robust component analysis method.
Algorithm 1 Solve the non-convex TRPCA model (5) by ADMM.
Input: The observed tensor X R N 1 × N 2 × N 3 , the set of index of observed entries Ω , parameter λ , stopping criterion ϵ .
Initialize: L ω > 0 0 = X ω > 0 , L Ω C 0 = rand ( N 1 × N 2 × N 3 ) Ω C , E 0 = M 0 = zeros ( N 1 × N 2 × N 3 ) , β = 10 - 4 , β m a x = 10 10 , η = 1.2 .
1:
while not converged do
2:
L k + 1 U X - E k - M k β k * Ω L * V X - E k - M k β k T using the local minimum in Theorem 1 with τ = 1 β k   
3:
E k + 1 S λ β k X - L k + 1 + M k β k   
4:
M k + 1 M k + β k ( X - L k + 1 - E k + 1 )   
5:
β k + 1 min { β k η , β m a x }   
6:
Check the convergence conditions L k + 1 - L k ϵ , E k + 1 - E k ϵ , L k + 1 + E k + 1 - X ϵ   
7:
end while  
Output: The low rank tensor L and the sparse tensor E
Theorem 2
(Convergent Analysis). Suppose that P k = { L k , E k , M k } is generated from the proposed Algorithm 1, then { P k } is bounded. In addition, the accumulation point { L * , E * , M * } of { P k } k = 1 + is a KKT stationary point, which satisfy the KKT conditions M * | | L | | t - γ , L * + E * = X , M * | | E | | 1 .
(Proof of Theorem 2)
E k + 1 satisfies the first-order necessary local optimality condition,
0 E L β k L k + 1 , E , M k | E k + 1 = λ E 1 | E k + 1 + M k + β k L k + 1 - X + E k + 1 = λ E 1 | E k + 1 + M k + 1 .
For E 1 case, since | | E | | 1 is nonsmooth at E i j k = 0 , we redefine subgradient E 1 i j k = 0 if E i j k = 0 . Then 0 E 1 F 2 N 1 N 2 N 3 , hence ( λ E 1 ) | E k + 1 is bounded. Thus { M k } is bounded.
With some algebra, we have the following equality
L β k L k , E k , M k = L β k - 1 L k , E k , M k - 1 + β k - β k - 1 2 L k - X + E k F 2 + t r [ ( M k - M k - 1 ) ( L k - X + E k ) ] = L β k - 1 L k , E k , M k - 1 + β k + β k - 1 2 ( β k - 1 ) 2 M k - M k - 1 F 2 .
Then, L β k L k + 1 , E k + 1 , M k L β k ( L k + 1 , E k , M k ) L β k ( L k , E k , M k ) L β k - 1 L k , E k , M k - 1 + β k + β k - 1 2 ( β k - 1 ) 2 M k - M k - 1 F 2 .
By induction, we could obtain
L β k ( L k + 1 , E k + 1 , M k ) L β 0 L 1 , E 1 , M 0 + i = 1 k β i + β i - 1 2 ( β i - 1 ) 2 M i - M i - 1 F 2 .
Since β i + β i - 1 2 ( β i - 1 ) 2 M i - M i - 1 F 2 is bounded, it is not hard to see L β k L k + 1 , E k + 1 , M k is upper bounded. Meanwhile,
L β k L k + 1 , E k + 1 , M k + 1 2 β k M k F 2 = | | L k + 1 | | t - γ + λ E k + 1 1 + β k 2 L k + 1 - X + E k + 1 + M k β k F 2 .
Since each term on the right-hand side is bounded, E k + 1 is bounded. By the last term on the right-hand, L k + 1 is bounded. Therefore, { L k } and { E k } are both bounded.
From Bolzano-Weierstrass theorem [39], there must be at least one accumulation point of the sequence { P k } k = 1 + . We denote one of the points P * = { L * , E * , M * } . Without loss of generality, we assume { P k } k = 1 + converge to P * .
Since
( M k + 1 - M k ) / β k = L k + 1 + E k + 1 - X ,
we have
lim k ( L k + 1 + E k + 1 - X ) = lim k ( M k + 1 - M k ) / β k = 0 .
Then L * + E * = X is obtained.
Since L k + 1 is optimally obtained by minimizing L β k L , E k , M k according to its definition, we have
0 L t - γ | L k + 1 + M k + β k L k + 1 + E k - X .
From Theorem 3 of [26], we know
L ^ n 3 | | L ^ n 3 | | γ = U ^ L , n 3 D i a g ( γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ) 2 ) V ^ L , n 3 T
and γ ( 1 + γ ) ( γ + Ω ^ n , n , n 3 ) 2 1 + γ γ L n 3 ^ | | L ^ n 3 | | γ is bounded.
Therefore,
L t - γ L ^ = [ | | L ^ 1 | | γ L ^ 1 | | | | L ^ N 3 | | γ L ^ N 3 ]
is bounded.
From the fact that L ^ is equivalent to Tucker product (for its definition see Chapter 3 of [44])
L ^ = L × 3 F ˜ N 3
(see the proof of Proposition 3 of [30] and Remark 2.3. of [45]) and using the chain rule [39], we can deduce that
L L t - γ = L t - γ L ^ × 3 F ˜ N 3 *
is bounded.
In addition, it is not hard to see
0 L t - γ | L k + 1 + M k + 1 - β k E k + 1 - E k
Please note that β k < β m a x = 10 10 in Algorithm 1, then k , β k E k + 1 - E k 0 - M * L * t - γ .
Similarly, since E k + 1 is the minimum of the subproblem L β k L k , E , M k , we have
0 | | E | | 1 | E k + 1 + M k + β k L k + E k + 1 - X .
Thus M k + 1 M * | | E * | | 1 . Therefore L * , E * , M * satisfies the the Karush-Kuhn-Tuker (KKT) conditions of the Lagrange function L β ( L , E , M ) (Chapter 9 of [46]). We thus complete the proof. ☐

5. Experiments

In this section, we conduct several applications to demonstrate the performance of our algorithm. All experiments are implemented in Matlab 2017b on Ubuntu 16.04 LTS with 12 cores, 2.40 GHz CPU and 64 GB RAM.

5.1. Image Recovery

In this part, we focus on noisy color image recovery. A color image with size N 1 × N 2 and 3 channel, i.e., red, green and blue, can be regarded as a third order tensor X N 1 × N 2 × N 3 and each frontal slice of X is the corresponding channel of the color image. Please note that each channel of the color image is usually not a low rank matrix, but it is observed that the top singular values dominate the main information [19,47]. Thus we can reconstruct the noisy image in low rank structure.
We perform the experiments on Berkeley Segmentation Dataset [48], which contains 200 color image for training. For each image, we randomly set 10 % of pixels to the random values in [0, 255], that is, we corrupt the image on 3 channels simultaneously which is more challenging than make corruptions on only one channel. We compare our t-Gamma algorithms with RPCA [4], Gamma [26], SNN [49] and TRPCA [20], in which RPCA minimizes a weighted combination of the nuclear norm and of the l 1 norm in matrix form, Gamma replaces the nuclear norm by matrix Gamma quasi-norm, SNN uses the sum of the nuclear norm for all mode-i matricization of the given tensor, and TRPCA extends the RPCA from the matrix to tensor case. For RPCA and Gamma, we conduct the algorithms on each channel, i.e., regarding each channel of the image as a matrix, and then combine the channel recovery result into a color image. The parameter λ in RPCA and Gamma is set to λ = 1 max { N 1 , N 2 } as suggested in [4]. However, when the parameter γ in Gamma is set to 0.01 as stated in [26], it cannot perform well. We empirically set the γ = min { N 1 , N 2 } in Gamma and t-Gamma. For SNN, we set λ = [ 15 , 15 , 1.5 ] as suggested in [20]. The parameters λ in TRPCA and t-Gamma is set to λ = 1 max { N 1 , N 2 } N 3 . By using the Peak Signal to Noise Ratio (PSNR) to evaluate the recovery results and PSNR is defined as
P S N R = 10 log 10 M 2 1 N 1 N 2 N 3 X ^ - M F 2 .
In Figure 1, we give the comparison of PSNR value on 200 images and each subfigure contains 50 results. The sample recovery performance are shown in Figure 2. From Figure 1, we can observe that the proposed t-Gamma algorithm perform better on almost every image and can achieve around 2 dB improvement in many cases. Table 1 shows the average recovery performance obtained by these five algorithms on Berkeley Segmentation Dataset. In Figure 2 we can visually conclude that our algorithm yields a better result than the compare algorithms which can illustrate the merit of the our approach.

5.2. Video Background Modeling

5.2.1. Qualitative Experiments

In this part, we consider the background modeling problem, an essential task in video surveillance, which aims to separate the foreground objects from the background. The frames of background in video surveillance are highly correlated and thus it is reasonable to consider it as a low rank tensor. While the objects in the foreground usually only occupy a small part of the image in some frames and hence can regard it as the sparse part.
To solve this problem, we test our algorithm on four example color videos, Hall, MovedObject, Escalator and Lobby http://perception.i2r.a-star.edu.sg/bk_model/bk_index.html, and compare it with RPCA [4], Gamma [26], SNN [49] and TRPCA [20]. We extract N 2 = 200 frames from these four videos. Suppose the size of frames in the given video is h × w , we stack each frame in each color channel as a column vector of size N 1 × 1 , where N 1 = h × w , and then all column vectors into a matrix of size 3 N 1 × N 2 for RPCA and Gamma, and into a tensor of size N 1 × N 2 × 3 for SNN, TRPCA and t-Gamma. The parameter l a m b d a is set to λ = 1 max { 3 N 1 , N 2 } for RPCA and Gamma, and λ = 1 max { N 1 , N 2 } 3 for SNN, TRPCA and t-Gamma. The parameter γ in Gamma and t-Gamma is set to γ = 1 min { 3 N 1 , N 2 } and γ = 3 min { N 1 , N 2 } respectively. As suggested in [20], λ in SNN is set to [ 200 , 2 , 20 ] .
The background separated results are depicted in Figure 3, Figure 4, Figure 5 and Figure 6. In Figure 3 and Figure 4, we can see that our method could reconstruct the background with fewer ghosting effects and more completely. From Figure 5 and Figure 6, it could be observed that though all the method separate the background from the foreground effectively, the proposed algorithm are more robust to the dynamic foreground and illumination variation, e.g., only Gamma and t-Gamma can remove the steps of moving escalator in Figure 5, while the method Gamma cannot adapt to change of illumination in Figure 6.

5.2.2. Quantitative Experiments

Since the datasets in Section 5.2.1 do not include the ground truth background for each frame of the video, we perform the algorithms on the ChangeDetection.net(CDNet) dataset2014 [50] which consists of visual surveillance and smart environments. However, only part of frames are accompanied with ground truth background and thus we only take out some of frames of the video for testing. That is, we select the frames from 900 to 1000 of the dataset highway, pedestrians and PETS2006 in the baseline category. Hence the size of the video will be h × w × 101 , where h × w is the resolution of each frame, and 101 is the number of frames. We compare our algorithm with RPCA [4], Gamma [26], SNN [49] and TRPCA [20]. The parameters setting is conducted from Section 5.2.1.
To evaluate the performance of the experiment results, we use the three metrics for the quantitative measure, that is, Recall, Precision and F-Measure. They are defined as:
Recall = TP TP + FN ,
Precision = TP TP + FP ,
F-Measure = 2 Precision · Recall Precision + Recall ,
where TP, FP and FN are true positives numbers, false positives numbers and false negatives numbers respectively. We primarily use F-Measure for overall evaluation since it combines the metrics Recall and Precision and also makes a balance between Recall and Precision which can be viewed as a good indicators for comparison.
From Table 2, we can see that the proposed algorithm achieves better segmentation performance among the compare methods with respect to the F-Measure. Some of the frames of the segmentation results are shown in Figure 7, it can be observed that among the three dataset in CDNet our method can usually get the clearer and more accurate background segmentation. These quantitative comparison illustrates the superior of t-Gamma tensor quasi-norm as the surrogate of the rank of tensor.

6. Concluding Remarks

In this paper, we conduct the TRPCA with a new t-Gamma tensor quasi-norm used as the regularization for tensor rank function. The t-Gamma tensor quasi-norm is developed via t-SVD in the Fourier domain, which simplifies the required optimization process. Compared with the convex tensor nuclear norm, our regularization approach appears to be more effective in approximating the tensor tubal rank. We incorporate the alternating direction method of multipliers with the t-Gamma tensor quasi-norm to efficiently solve the non-convex TRPCA problem, yielding an approachable framework for non-convex optimization. Numerical experimental results indicate that our approach outperforms the existing methods in distinguishable ways.

Author Contributions

Conceptualization, M.Y.; Data curation, Q.L.; Formal analysis, M.Y.; Funding acquisition, S.C.; Investigation, M.Y. and M.X.; Methodology, Q.L., M.Y. and M.X.; Project administration, S.C., W.L. and M.X.; Resources, M.X.; Software, Q.L.; Supervision, S.C., W.L. and M.X.; Validation, S.C.; Visualization, Q.L.; Writing—original draft, M.Y.; Writing—review & editing, M.X.

Funding

This research was funded by the National Science Foundation Division of Mathematical Sciences of the United States (Grant No. 140928), the National Natural Science Foundation of China (Grant Nos. 11671158, U181164, 61201392), the Natural Science Foundation of Guangdong Province, China (Grant No. 2015A030313497) and the Science and Technology Planning Project of Guangdong Province, China (Grant Nos. 2017B090909004, 2017B010124003, 2017B090909001).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Landsberg, J. Tensors: Geometry and Applications; Graduate Studies in Mathematics; American Mathematical Society: Providence, RI, USA, 2012. [Google Scholar]
  2. Bro, R. Multi-Way Analysis in the Food Industry: Models, Algorithms, and Applications. Ph.D. Thesis, University of Amsterdam (NL), Amsterdam, The Netherlands, 1998. [Google Scholar]
  3. Yang, M.; Li, W.; Xiao, M. On identifiability of higher order block term tensor decompositions of rank Lr⊗ rank-1. Linear Multilinear Algebra 2018, 1–23. [Google Scholar] [CrossRef]
  4. Candès, E.J.; Li, X.; Ma, Y.; Wright, J. Robust principal component analysis? J. ACM 2011, 58, 11. [Google Scholar] [CrossRef]
  5. Wright, J.; Ganesh, A.; Rao, S.; Peng, Y.; Ma, Y. Robust principal component analysis: Exact recovery of corrupted low-rank matrices via convex optimization. In Proceedings of the Advances in Neural Information Processing Systems, Vancouver, BC, USA, 7–10 December 2009; pp. 2080–2088. [Google Scholar]
  6. Kang, Z.; Pan, H.; Hoi, S.C.H.; Xu, Z. Robust Graph Learning From Noisy Data. IEEE Trans. Cybern. 2019, 1–11. [Google Scholar] [CrossRef]
  7. Peng, C.; Kang, Z.; Cai, S.; Cheng, Q. Integrate and Conquer: Double-Sided Two-Dimensional k-Means Via Integrating of Projection and Manifold Construction. ACM Trans. Intell. Syst. Technol. 2018, 9, 57. [Google Scholar] [CrossRef]
  8. Sun, P.; Qin, J. Speech enhancement via two-stage dual tree complex wavelet packet transform with a speech presence probability estimator. J. Acoust. Soc. Am. 2017, 141, 808–817. [Google Scholar] [CrossRef] [PubMed]
  9. Vaswani, N.; Bouwmans, T.; Javed, S.; Narayanamurthy, P. Robust subspace learning: Robust PCA, robust subspace tracking, and robust subspace recovery. IEEE Signal Process. Mag. 2018, 35, 32–55. [Google Scholar] [CrossRef]
  10. Bouwmans, T.; Zahzah, E.H. Robust PCA via principal component pursuit: A review for a comparative evaluation in video surveillance. Comput. Vis. Image Underst. 2014, 122, 22–34. [Google Scholar] [CrossRef]
  11. Bouwmans, T.; Javed, S.; Zhang, H.; Lin, Z.; Otazo, R. On the applications of robust PCA in image and video processing. Proc. IEEE 2018, 106, 1427–1457. [Google Scholar] [CrossRef]
  12. Cichocki, A.; Lee, N.; Oseledets, I.; Phan, A.H.; Zhao, Q.; Mandic, D.P. Tensor networks for dimensionality reduction and large-scale optimization: Part 1 low-rank tensor decompositions. Found. Trends® Mach. Learn. 2016, 9, 249–429. [Google Scholar] [CrossRef]
  13. Kilmer, M.E.; Martin, C.D. Factorization strategies for third-order tensors. Linear Algebra Appl. 2011, 435, 641–658. [Google Scholar] [CrossRef]
  14. Hao, N.; Kilmer, M.E.; Braman, K.; Hoover, R.C. Facial recognition using tensor-tensor decompositions. SIAM J. Imaging Sci. 2013, 6, 437–463. [Google Scholar] [CrossRef]
  15. Zhang, Z.; Ely, G.; Aeron, S.; Hao, N.; Kilmer, M. Novel methods for multilinear data completion and de-noising based on tensor-SVD. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 3842–3849. [Google Scholar]
  16. Signoretto, M.; De Lathauwer, L.; Suykens, J.A. Nuclear norms for tensors and their use for convex multilinear estimation. Linear Algebra Appl. 2010, 43. Available online: ftp://ftp.esat.kuleuven.be/sista/signoretto/Signoretto_nucTensors2.pdf (accessed on 3 April 2019).
  17. Cichocki, A. Era of big data processing: A new approach via tensor networks and tensor decompositions. arXiv, 2014; arXiv:1403.2048. [Google Scholar]
  18. Grasedyck, L.; Kressner, D.; Tobler, C. A literature survey of low-rank tensor approximation techniques. GAMM-Mitteilungen 2013, 36, 53–78. [Google Scholar] [CrossRef]
  19. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis: Exact Recovery of Corrupted Low-Rank Tensors via Convex Optimization. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5249–5257. [Google Scholar]
  20. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis with A New Tensor Nuclear Norm. arXiv, 2018; arXiv:1804.03728. [Google Scholar] [CrossRef] [PubMed]
  21. Semerci, O.; Hao, N.; Kilmer, M.E.; Miller, E.L. Tensor-based formulation and nuclear norm regularization for multienergy computed tomography. IEEE Trans. Image Process. 2014, 23, 1678–1693. [Google Scholar] [CrossRef]
  22. Ciccone, V.; Ferrante, A.; Zorzi, M. Robust identification of “sparse plus low-rank” graphical models: An optimization approach. In Proceedings of the 2018 IEEE Conference on Decision and Control (CDC), Miami Beach, FL, USA, 17–19 December 2018; pp. 2241–2246. [Google Scholar]
  23. Ciccone, V.; Ferrante, A.; Zorzi, M. Factor Models with Real Data: A Robust Estimation of the Number of Factors. IEEE Trans. Autom. Control 2018. [Google Scholar] [CrossRef]
  24. Fazel, S.M. Matrix Rank Minimization with Applications. Ph.D. Thesis, Stanford University, Stanford, CA, USA, 2003. [Google Scholar]
  25. Recht, B.; Fazel, M.; Parrilo, P.A. Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 2010, 52, 471–501. [Google Scholar] [CrossRef]
  26. Kang, Z.; Peng, C.; Cheng, Q. Robust PCA Via Nonconvex Rank Approximation. In Proceedings of the 2015 IEEE International Conference on Data Mining (ICDM), Atlantic City, NJ, USA, 14–17 November 2015; pp. 211–220. [Google Scholar]
  27. Lewis, A.S.; Sendov, H.S. Nonsmooth analysis of singular values. Part I: Theory. Set-Valued Anal. 2005, 13, 213–241. [Google Scholar] [CrossRef]
  28. Liu, Y.; Chen, L.; Zhu, C. Improved Robust Tensor Principal Component Analysis via Low-Rank Core Matrix. IEEE J. Sel. Top. Signal Process. 2018, 12, 1378–1389. [Google Scholar] [CrossRef]
  29. Chen, Y.; Wang, S.; Zhou, Y. Tensor Nuclear Norm-Based Low-Rank Approximation With Total Variation Regularization. IEEE J. Sel. Top. Signal Process. 2018, 12, 1364–1377. [Google Scholar] [CrossRef]
  30. Kong, H.; Xie, X.; Lin, Z. t-Schatten-p Norm for Low-Rank Tensor Recovery. IEEE J. Sel. Top. Signal Process. 2018, 12, 1405–1419. [Google Scholar] [CrossRef]
  31. Tarzanagh, D.A.; Michailidis, G. Fast Monte Carlo Algorithms for Tensor Operations. arXiv, 2017; arXiv:1704.04362. [Google Scholar]
  32. Driggs, D.; Becker, S.; Boyd-Graber, J. Tensor Robust Principal Component Analysis: Better recovery with atomic norm regularization. arXiv, 2019; arXiv:1901.10991. [Google Scholar]
  33. Sobral, A.; Baker, C.G.; Bouwmans, T.; Zahzah, E.H. Incremental and multi-feature tensor subspace learning applied for background modeling and subtraction. In Proceedings of the International Conference Image Analysis and Recognition, Vilamoura, Portugal, 22–24 October 2014; pp. 94–103. [Google Scholar]
  34. Sobral, A.; Javed, S.; Ki Jung, S.; Bouwmans, T.; Zahzah, E.H. Online stochastic tensor decomposition for background subtraction in multispectral video sequences. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Santiago, Chile, 7–13 December 2015; pp. 106–113. [Google Scholar]
  35. Bouwmans, T.; Sobral, A.; Javed, S.; Jung, S.K.; Zahzah, E.H. Decomposition into low-rank plus additive matrices for background/foreground separation: A review for a comparative evaluation with a large-scale dataset. Comput. Sci. Rev. 2017, 23, 1–71. [Google Scholar] [CrossRef]
  36. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternating direction method of multipliers. Found. Trends® Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  37. Kilmer, M.E.; Braman, K.; Hao, N.; Hoover, R.C. Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging. SIAM J. Matrix Anal. Appl. 2013, 34, 148–172. [Google Scholar] [CrossRef]
  38. Tyrtyshnikov, E.E. A Brief Introduction to Numerical Analysis; Springer Science & Business Media: New York, NY, USA, 2012. [Google Scholar]
  39. Marc Peter Deisenroth, A.A.F.; Ong, C.S. Mathematics for Machine Learning; Cambridge University Press: Cambridge, UK, 2019. [Google Scholar]
  40. Moreau, J.J. Proximité et dualité dans un espace hilbertien. Bull. Soc. Math. France 1965, 93, 273–299. [Google Scholar] [CrossRef]
  41. Dong, W.; Shi, G.; Li, X.; Ma, Y.; Huang, F. Compressive Sensing via Nonlocal Low-Rank Regularization. IEEE Trans. Image Process. 2014, 23, 3618–3632. [Google Scholar] [CrossRef] [PubMed]
  42. Zhang, G.; Lin, Y. Lectures in Functional Analysis; Peking University Press: Beijing, China, 1987. [Google Scholar]
  43. Donoho, D.L. De-noising by soft-thresholding. IEEE Trans. Inf. Theory 1995, 41, 613–627. [Google Scholar] [CrossRef]
  44. Lu, H.; Plataniotis, K.; Venetsanopoulos, A. Multilinear Subspace Learning: Dimensionality Reduction of Multidimensional Data; Chapman & Hall/CRC Machine Learning & Pattern Recognition Series; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  45. Zhang, Z.; Aeron, S. Exact Tensor Completion Using t-SVD. IEEE Trans. Signal Process. 2017, 65, 1511–1526. [Google Scholar] [CrossRef]
  46. Luenberger, D.G. Optimization by Vector Space Methods; John Wiley & Sons: New York, NY, USA, 1997. [Google Scholar]
  47. Liu, J.; Musialski, P.; Wonka, P.; Ye, J. Tensor Completion for Estimating Missing Values in Visual Data. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 208–220. [Google Scholar] [CrossRef] [PubMed]
  48. Martin, D.; Fowlkes, C.; Tal, D.; Malik, J. A database of human segmented natural images and its application to evaluating segmentation algorithms and measuring ecological statistics. In Proceedings of the Eighth IEEE International Conference on Computer Vision ( ICCV 2001), Vancouver, BC, Canada, 7–14 July 2001; Volume 2, pp. 416–423. [Google Scholar]
  49. Huang, B.; Mu, C.; Goldfarb, D.; Wright, J. Provable models for robust low-rank tensor completion. Pac. J. Optim. 2015, 11, 339–364. [Google Scholar]
  50. Goyette, N.; Jodoin, P.-M.; Porikli, F.; Konrad, J.; Ishwar, P. Changedetection.net: A new change detection benchmark dataset, 2012. In Proceedings of the IEEE Workshop on Change Detection (CDW-2012) at CVPR-2012, Providence, RI, USA, 16–21 June 2012. [Google Scholar]
Figure 1. Comparison of the PSNR values with RPCA [4], Gamma [26], SNN [49], TRPCA [20] and t-Gamma.
Figure 1. Comparison of the PSNR values with RPCA [4], Gamma [26], SNN [49], TRPCA [20] and t-Gamma.
Applsci 09 01411 g001
Figure 2. Recovery results on 5 sample images. The first two columns are original and corrupted images reps. The third to seventh columns are recovery images obtained by RPCA, Gamma, SNN, TRPCA and t-Gamma resp. (a) Original; (b) Corrupted; (c) RPCA [4]; (d) Gamma [26]; (e) SNN [49]; (f) TRPCA [20]; (g) t-Gamma.
Figure 2. Recovery results on 5 sample images. The first two columns are original and corrupted images reps. The third to seventh columns are recovery images obtained by RPCA, Gamma, SNN, TRPCA and t-Gamma resp. (a) Original; (b) Corrupted; (c) RPCA [4]; (d) Gamma [26]; (e) SNN [49]; (f) TRPCA [20]; (g) t-Gamma.
Applsci 09 01411 g002
Figure 3. Background modeling: two example frames from Hall sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Figure 3. Background modeling: two example frames from Hall sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Applsci 09 01411 g003
Figure 4. Background modeling: two example frames from MovedObject sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Figure 4. Background modeling: two example frames from MovedObject sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Applsci 09 01411 g004
Figure 5. Background modeling: two example frames from Escalator sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Figure 5. Background modeling: two example frames from Escalator sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Applsci 09 01411 g005
Figure 6. Background modeling: two example frames from Lobby sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Figure 6. Background modeling: two example frames from Lobby sequences are shown. (a) Original; (b) RPCA [4]; (c) Gamma [26]; (d) SNN [49]; (e) TRPCA [20]; (f) t-Gamma.
Applsci 09 01411 g006
Figure 7. Visual results for CDNet dataset. Row 1 and row 2 are samples of highway; row 3 and row 4 are samples of pedestrians; row 5 and row 6 are samples of PETS2006. (a) Original; (b) Ground True; (c) RPCA [4]; (d) Gamma [26]; (e) SNN [49]; (f) TRPCA [20]; (g) t-Gamma
Figure 7. Visual results for CDNet dataset. Row 1 and row 2 are samples of highway; row 3 and row 4 are samples of pedestrians; row 5 and row 6 are samples of PETS2006. (a) Original; (b) Ground True; (c) RPCA [4]; (d) Gamma [26]; (e) SNN [49]; (f) TRPCA [20]; (g) t-Gamma
Applsci 09 01411 g007
Table 1. Average PSNR obtained by five Algorithms on 200 images.
Table 1. Average PSNR obtained by five Algorithms on 200 images.
AlgorithmRPCAGammaSNNTRPCAt-Gamma
Average PSNR25.584327.076227.632929.115730.7896
Table 2. Quantitative Evaluation for algorithm on CDNet Datasets
Table 2. Quantitative Evaluation for algorithm on CDNet Datasets
DatasetRPCAGammaSNNTNNt-Gamma
RecallPrecisionF-MeasureRecallPrecisionF-MeasureRecallPrecisionF-MeasureRecallPrecisionF-MeasureRecallPrecisionF-Measure
highway0.86420.91090.88690.88370.90670.8950.69820.90410.78790.82350.9020.8610.88350.9090.8961
pedestrians0.99830.94190.96920.99820.92750.96160.99550.92850.96080.99830.92580.96070.99820.93090.9634
PETS20060.81510.71310.76070.83260.76780.79890.2880.69740.40770.80330.68530.73960.83010.78050.8045

Share and Cite

MDPI and ACS Style

Cai, S.; Luo, Q.; Yang, M.; Li, W.; Xiao, M. Tensor Robust Principal Component Analysis via Non-Convex Low Rank Approximation. Appl. Sci. 2019, 9, 1411. https://doi.org/10.3390/app9071411

AMA Style

Cai S, Luo Q, Yang M, Li W, Xiao M. Tensor Robust Principal Component Analysis via Non-Convex Low Rank Approximation. Applied Sciences. 2019; 9(7):1411. https://doi.org/10.3390/app9071411

Chicago/Turabian Style

Cai, Shuting, Qilun Luo, Ming Yang, Wen Li, and Mingqing Xiao. 2019. "Tensor Robust Principal Component Analysis via Non-Convex Low Rank Approximation" Applied Sciences 9, no. 7: 1411. https://doi.org/10.3390/app9071411

APA Style

Cai, S., Luo, Q., Yang, M., Li, W., & Xiao, M. (2019). Tensor Robust Principal Component Analysis via Non-Convex Low Rank Approximation. Applied Sciences, 9(7), 1411. https://doi.org/10.3390/app9071411

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop