Next Article in Journal
Quantitative 31P NMR Spectroscopy: Principles, Methodologies, and Applications in Phosphorus-Containing Compound Analysis
Previous Article in Journal
Application of Enhanced Weighted Least Squares with Dark Background Image Fusion for Inhomogeneity Noise Removal in Brain Tumor Hyperspectral Images
Previous Article in Special Issue
Context-Encoder-Based Image Inpainting for Ancient Chinese Silk
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low Tensor Rank Constrained Image Inpainting Using a Novel Arrangement Scheme

1
School of Space Information, Space Engineering University, Beijing 101400, China
2
China Academy of Space Technology (Xi’an), Xi’an 710100, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(1), 322; https://doi.org/10.3390/app15010322
Submission received: 21 October 2023 / Revised: 30 November 2024 / Accepted: 27 December 2024 / Published: 31 December 2024
(This article belongs to the Special Issue AI-Based Image Processing: 2nd Edition)

Abstract

:
Employing low tensor rank decomposition in image inpainting has attracted increasing attention. This study exploited novel tensor arrangement schemes to transform an image (a low-order tensor) to a higher-order tensor without changing the total number of pixels. The developed arrangement schemes enhanced the low rankness of images under three tensor decomposition methods: matrix SVD, tensor train (TT) decomposition, and tensor singular value decomposition (t-SVD). By exploiting the schemes, we solved the image inpainting problem with three low-rank constrained models that use the matrix rank, TT rank, and tubal rank as constrained priors. The tensor tubal rank and tensor train multi-rank were developed from t-SVD and TT decomposition, respectively. Then, ADMM algorithms were efficiently exploited for solving the three models. Experimental results demonstrate that our methods are effective for image inpainting and superior to numerous close methods.

1. Introduction

Image inpainting refers to the process of completing missing entries or restoring damaged regions of an image. It is a typical ill-posed inverse problem, generally solved by exploiting the image priors [1,2], such as smoothness, sparsity, and low rankness. In recent years, tensor analysis, including tensor low-rank decomposition and tensor completion, has attracted increasing attention [3,4,5,6]. A color image itself is an order-3 tensor, or it can be used to construct a high-order (greater than 3) tensor; then, the image inpainting problem becomes a tensor completion problem. A tensor is more challenging to analyze than a matrix due to the complicated nature of higher-order arrays [7]. Low tensor rank can be constrained to recover the missing pixels. The effectiveness relies on the tensor rank. The lower the tensor rank, the better the recovery results. Thus, finding ways to decrease the tensor rank is essential in the tensor completion problem. Unlike the matrix rank, the definition of tensor rank is not unique and is related to the tensor decomposition scheme.
Low tensor rank completion methods can be categorized according to the tensor decomposition frameworks they use [8]. The traditional tensor decomposition tools include CANDECOMP/PARAFAC (CP) and Tucker decomposition [8,9]. Recently proposed decomposition frameworks include tensor singular value decomposition (t-SVD) [10,11,12], tensor train (TT) decomposition [13,14], tensor tree (TTR) decomposition [6,15], tensor ring (TR) decomposition [16], fully connected tensor network (FCTN) decomposition [17], etc. The CP rank is hard to estimate. Tucker rank is multi-rank, whose elements are the ranks of mode-n matrices that are highly unbalanced. TT rank is also multi-rank, whose elements are the ranks of TT matrices. For a high-order tensor, TT matrices are more balanced than the mode-n matrices. As matrix rank minimization is only efficient when the matrix is balanced, TT decomposition is more suitable for describing global information of high-order tensors than Tucker decomposition. T-SVD defines the tubal rank of the high-order tensor, which can be easily estimated according to a fast Fourier-based method. The tubal rank has been shown to be more efficient than the matrix rank and Tucker multi-rank in video applications [18,19,20]. TTR rank is essentially equivalent to Tucker multi-rank. It has to be mentioned that both TT and TR decompose the Nth-order tensor into a list of third-order core tensors. Among them, TR decomposition is an improvement of TT decomposition, and TR rank is also a type of multi-rank. TR connects multiple core tensors in a circular manner, but TR rank only constrains the factor relationship between two adjacent core tensors. FCTN decomposition can flexibly characterize the correlation of multiple core tensor factors.
Many popular tensor completion methods have applied the traditional CP or Tucker decomposition on color image inpainting. Some recent works have exploited the sparse Tucker core tensor and nonnegative Tucker factor matrices for image restoration [7,21,22]. Some works have constrained the low rankness of the mode-n matrix caused by the decomposition of a color image for inpainting [23,24]. As low tensor rank constraints cannot fully capture the local smooth and global sparsity priors of tensors, some works have combined Tucker and total variation (TV). The SPCTV (smooth PARAFAC tensor completion and total variation) method [25] uses the PD (PARAFAC decomposition, a derivation of Tucker decomposition) framework and constrains the TV (total variation) on every factor matrix of PD. Some works have combined the constraints of the low rankness of every mode-n matrix and the TV regularization on every mode-n matrix for color image inpainting [26,27]. Some works have proposed data restoration methods based on Bayesian tensor completion [28,29,30,31].
The aforementioned methods all take the color image as an order-3 tensor directly and do not deeply explore the potential low-rank decomposition prior to a color image. The RNC-FCTN method [32] proposes a robust tensor completion method based on FCTN, but it is limited to high-dimensional data processing, such as hyperspectral video, color video completion, and video background subtraction. The CFN-RTC method [33] proposes a robust tensor completion method that combines the capped Frobenius norm and TR decomposition and adds dimensions through simple data reshaping for processing image or video data. As TT decomposition is efficient for higher-order tensors, the TMac-TTKA method [34] first uses the ket augmentation (KA) scheme to permute the image to high-order data, then applies the optimal models by enforcing low TT rankness. The KA scheme is proven to be efficient for improving the accuracy of color image/video inpainting and dynamic MR image reconstruction in TT rank-based completion methods [34,35,36,37]. As far as we know, the KA scheme is the only method used to permute data into high-order data.
This study aims to explore the potential low-rank structure of images and find an efficient way to apply SVD, t-SVD, and TT decomposition in image inpainting problems. The contributions of our work are summarized as follows:
  • First, we developed a novel rearrangement, named the quarter arrangement (QA) scheme, for permuting the image into three flexible forms of data. The first flexible QA scheme can permute an image into an unfolding matrix (with a low matrix rank structure). The second and the third flexible QA schemes can permute the color image into a balanced 3-order form of data (with low tubal rank structure) and a higher-order form of data (with low TT rank structure), respectively. Because the developed schemes are designed to exploit the internal structure similarity of the original data as much as possible, the rearranged data have the corresponding low-rank structure.
  • Second, based on the above QA scheme, we developed three image inpainting models that exploit the unfolding matrix rank, tensor tubal rank, and TT multi-rank of the rearranged data to solve the image inpainting problem.
  • Lastly, three efficient ADMM algorithms were developed to solve the above three models. Compared with numerous close image inpainting methods, the experimental results demonstrated the superior performance of our methods.
An abbreviation index is shown in Table 1. The remainder of this paper is organized as follows. In Section 2, related work is described. In Section 3, we introduce the proposed methods. Section 4 presents the experimental results and analyses. The conclusion is given in Section 5.

2. Related Work

In this section, a brief introduction to the KA scheme, t-SVD decomposition, and tensor train decomposition is provided. Notations and definitions are summarized in Table 2.

2.1. Ket Augmentation

The ket augmentation (KA) scheme was originally introduced by Latorre in [38] for casting a grayscale image into the real ket state of a Hilbert space. Bengua et al. [34] used KA to reshape a low-order tensor, e.g., a color image, to a higher-order tensor and proved that KA is efficient in improving the accuracy of the recovered image in TT-based completion.
For example, through the KA scheme, an 8 × 8 matrix can be turned into a 3-order tensor of size 4 × 4 × 4, which is achieved by dividing the matrix into several blocks and rearranging the blocks row by row. In addition, the KA scheme can turn a 3-order tensor size of x N × y N × N 3 into an N + 1 -order tensor with a size of x y × x y × : : : × x y × N 3 . For more details refer to [37].

2.2. T-SVD Decomposition

Definition 1.
t-product [39]. For A R n 1 × n 2 × n 3 and B R n 2 × n 4 × n 3 , the t-product A B = C is a tensor of size n 1 × n 4 × n 3 , where C   ( i ,   j ,   : ) is given by k = 1 n 2 A   ( i ,   k ,   : ) B   ( k ,   j ,   : ) , denotes the circular convolution between the two vectors, and i = 1 ,   2 , ,   n 1 , j = 1 ,   2 , ,   n 4 .
The t-SVD of A R n 1 × n 2 × n 3 is given by
A = U S V T
where U and V are orthogonal tensors of size n 1 × n 1 × n 3 and n 2 × n 2 × n 3 , respectively. S is a rectangular f-diagonal tensor of size n 1 × n 2 × n 3 , * denotes t-product [39], and T denotes the tensor transpose defined in [39].
The tensor rank defined in t-SVD is tensor tubal rank, which is the number of nonzero singular tubes in S . Ref. [40] proposed the fast Fourier-based method to calculate the tubal rank and used the tensor nuclear norm (TNN) as the convex relaxation of the tensor tubal rank.
A T N N = b l o c k d i a g ( A ¯ ) *
where A ¯   =   f f t   ( A ,   [   ]   ,   3 ) is the tensor obtained by applying the 1D FFT along the third dimension of A ,   * denotes the nuclear norm, and b l o c k d i a g ( A ¯ ) = A ¯ ( 1 ) A ¯ ( 2 ) A ¯ ( I 3 ) .

2.3. Tensor Train Decomposition

A ( i 1 , i 2 , i n , i N ) = U 1 ( : , i 1 , : ) U 2 ( : , i 2 , : ) U n ( : , i n , : ) U N ( : , i N , : )
Given a tensor A R I 1 × I 2 × I N , tensor train (TT) decomposition [13,14] can decompose it to N order-3 tensors U n R S n × I n × S n + 1 , n = 1 , , N . The tensor rank defined in TT decomposition is multi-rank, i.e., ( S 1 , S 2 , , S N + 1 ) , which is combined with the second-dimensional size of each U n .
The widely used way to find TT rank is to estimate the rank of each TT matrix [41] as the element of ( S 1 , S 2 , , S N + 1 ) . The TT matrix A [ n ] ( n = 1 , , N 1 ) with rank S n is the mode- 1 , 2 , , n matricization of the tensor with the size of m × h , where m = l = 1 n I l , h = l = n + 1 N I l .

3. Methods

3.1. Quarter Arrangement

To explore the more efficient low-rank structure of an image, we propose a novel rearrangement scheme, named the quarter arrangement (QA) scheme, to turn a color image into other forms of data. The QA scheme can maintain the internal similarity of the original image in the rearranged data.
The basic QA scheme: For example, as shown in Figure 1a, M is a 2D matrix ( 8 × 8 ). First, the entries of M are extracted for every other row and column to obtain four smaller matrices. Each smaller matrix has a size of 4 × 4 . Then, these four smaller matrices are placed along the third dimension in a designed order. Lastly, a 3D tensor of size 4 × 4 × 4 is obtained from the 8 × 8 matrix M without changing the total number of entries. The entries in the four smaller matrices are labeled as the MATLAB notation ( : , : , 1 ) , ( : , : , 2 ) , ( : , : , 3 ) and ( : , : , 4 ) . If M is smooth (most images satisfy), the four smaller matrices are similar in structure due to the adjacent entries.
Applying the basic QA scheme on the single Lena image, the Lena image can be divided into 4 smaller Lena images. As shown in Figure 1b, the four smaller Lena images are similar to each other. In Figure 1c, the pixel values curves of the four smaller images have overlapped into one curve. It can be said that the similarity of the local image structure is mostly maintained by the basic QA scheme.
Under this basic QA scheme, three flexible QA schemes are proposed for permuting the image into three flexible forms of data. The three flexible QA schemes can enhance the low rankness for an image by matrix SVD, tensor train decomposition, and tensor-SVD. Then, by exploiting the flexible QA schemes, three low rank-constrained methods that use the TT rank, tubal rank, and matrix rank as constrained priors are exploited for image inpainting.
The three flexible QA schemes and methods are described in detail in the following three sections. By using the flexible QA scheme, tensor decomposition models based on low rank can be used not only for a color image or video completion but also for grayscale image operations.

3.2. Method 1: The Low Unfolding Matrix Rank-Based Method

The unfolding method is widely used to permute order-3 video or dynamic magnetic resonance images into an unfolding matrix and then exploit the low rankness of this matrix for data reconstruction [23,42]. The unfolding matrix has a low-rank structure because of the similarity of every slightly changed slice along the time dimension.
We try to dig out the potential low unfolding-matrix rankness of a color image by a flexible QA scheme and call this scheme the first flexible QA scheme.
Take a 256 × 256 × 3 Lena image as an example, as shown in Figure 2a. First, the image is permuted into the 3-order tensor size of 32 × 32 × 192 using the basic QA scheme. Then, similar slices of this 3-order are unfolded. Lastly, the balanced (the context of ‘balanced’ is that the size changes from the unbalanced 256 × 256 × 3 to the more balanced size of 1024 × 192) unfolding matrix size of 1024 × 192 is obtained. As the slices (32 × 32) in the 3-order tensor are similar, the unfolding matrix is low rank, as shown in Figure 2b. In practice, the size of the designed unfolding matrix should be balanced so that the minimization of the unfolding matrix rank is efficient.
We exploit the low unfolding matrix rank in image inpainting and give the low unfolding matrix rank-based model as follows:
min X Μ ( Φ 1 X ) * subject   to X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω
where X denotes the image to be recovered; Φ 1 denotes the operator of permuting the image into a suitable 3-order tensor by multiple basic QA schemes; Μ denotes the operator of the unfolding process, which unfolds every slice along the third dimension of the 3-order tensor Φ 1 X ; Ω is the position without painting; and Y is the painted image with damaged entries at the positions Ω .
To reduce the computational complexity, in model (1), the following SVD-free approach [43,44] is exploited to constrain the low rankness of the unfolding matrix Μ ( Φ 1 X ) instead of the nuclear norm:
min U   V H = Μ ( Φ 1 X ) 1 2 U F 2 + V F 2 = Μ ( Φ 1 X ) *
As total variation (TV) has been proved as an effective constraint of smooth prior [45,46], model (1) is incorporated with 2D TV to exploit the local smooth priors of visual image data. Then, the image inpainting model (1) turns into the following:
min U ,   V ,   X 1 2 U F 2 + V F 2 + β 2 X T V subject   to     X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω ,   Μ Φ 1 X = U V H
where β is the regularization parameter.
The algorithm of alternating direction method of multipliers (ADMM) is used to solve the low unfolding matrix rank and TV-based image inpainting model (3). Firstly, an auxiliary variable Z = D X is introduced, where D is the finite difference operator, and then (3) is rewritten as the unconstrained convex optimization problem (4):
min U ,   V ,   Λ ,   L ,   Z ,   X τ Ω ( X ) + 1 2 U F 2 + V F 2 + ρ 1 2 Μ Φ 1 X U V H + Λ F 2     + β 2 Z 1 + β ρ 2 2 D X Z + L F 2
where τ Ω ( X ) denotes the indicator function:
τ Ω ( X ) = 0 ,   X Ω ,   otherwise  
L and Λ are the Lagrangian multipliers for variables Z and U V H , respectively. The regularization parameter β is used to balance the low rankness and sparsity constraints (i.e., TV). The penalty parameters ρ 1 > 0 and ρ 2 > 0 generally affect the convergence of the algorithm. By applying ADMM, each sub-problem is performed at each iteration t as follows:
X t = arg min X τ Ω ( X ) + ρ 1 2 Μ Φ 1 X U t 1 V ( t 1 ) H + Λ t 1 F 2
U t = arg min U U F 2 + ρ 1 2 Μ Φ 1 X t U V ( t 1 ) H + Λ t 1 F 2
V t = arg min V V F 2 + ρ 1 2 Μ Φ 1 X t U t V H + Λ t 1 F 2
Z t = arg min Z Z 1 + ρ 2 2 D X t Z + L t 1 F 2
Λ t = Λ t 1 + Μ Φ 1 X t U t V ( t ) H
L t = L t 1 + D X t Z t
The initial U and V can be determined by solving the following optimization problem using the LMaFit method [47]:
min U , V , X U   V H Μ Φ 1 X F 2     s u b j e c t   t o X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω
The whole algorithm for solving the model (3) is shown in Algorithm 1.
Algorithm 1. The algorithm for solving the model (3)
Input: Y ,   Ω ,   ρ 1 , β ,   ρ 2 , maximum number of iteration t max , convergence condition η t o l .
Initialization: initial U ( 0 ) , V ( 0 ) by solving the matrix completion problem (11), Λ ( 0 ) , L ( 0 ) , Z ( 0 ) , t = 0.
While  t < t max and η < η max  do
        The first flexible QA scheme: Turn an image into an order-N tensor Φ 1 X , then unfold it.
        Solve (5)–(10) for X * , where * represents the optimal solution.
        Update η t + 1 = X n t + 1 ( : ) X n t ( : ) F X n t ( : ) F , t = t + 1 .
End while
Output: X * .

3.3. Method 2: The Low Tubal Rank-Based Method

Tensor-SVD decomposition has been efficiently used in video image completion and dynamic MR image reconstruction problems [4,16,48,49]. As the color image is highly unbalanced in the size of three dimensions, which is not suitable for the low tubal rank constraint, the second flexible QA scheme is exploited to dig out the potential low tubal rank prior information.
Considering that tubal rank minimizations are more efficient for the balanced tensor [10], we first turn the unbalanced image into the balanced order-3 data using the second flexible QA scheme.
Take the color image size of 256 × 256 × 3 as an example, as shown in Figure 3a. The order-4 tensor size of 128 × 128 × 4 × 3 is obtained by the basic QA schemes, and by multiplying the basic QA schemes, the order-4 tensor size of 64 × 64 × 4 × 4 × 3 is obtained. Lastly, the order-4 tensor is reshaped into the balanced order-3 tensor size of 64 × 64 × 48. Here, the context of ‘balanced’ is that the size changes from the unbalanced 256 × 256 × 3 to the more balanced size of 64 × 64 × 48. In practice, the size of the designed order-3 tensor should be as balanced as possible. We call the above the second flexible QA scheme.
Figure 3b shows the low tubal rankness of the balanced order-3 data (with the size of n 1 × n 2 × n 3 = 64 × 64 × 48 here) by plotting δ j , which is defined as follows:
δ j = 1 n 3 i = 1 n 3 T ( i ,   i ,   j ) , i = 1 ,   2 ,   ,   min ( n 1 ,   n 2 )
Then, TNN is used to enforce the tensor tubal rank in the image inpainting model as follows:
min X Φ 2 X T N N s u b j e c t   t o     X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω
where Φ 2 denotes the operator of permuting the color image into a more ‘balanced’ order-3 tensor using the second flexible QA scheme. Combining the low tubal rank and sparsity, we introduce auxiliary variables B = Φ 2 X and Z = D X , then rewrite (12) as the following unconstrained convex optimization problem:
min X τ Ω ( X ) + i I 3 B ¯ ( i ) * + ρ 2 Φ 2 X B + Λ F 2     + β 2 Z 1 + β ρ 2 2 D X Z + L F 2
where I 3 is the third size of the 3-order tensor Φ 2 X . We apply the algorithm by ADMM to solve model (13), as shown in Algorithm 2.
Algorithm 2. The algorithm for solving the model (13)
Input:  Y ,   Ω ,   ρ 1   , β ,   ρ 2 , the maximum number of iteration t max , convergence condition η t o l .
Initialization:  Λ ( 0 ) , L ( 0 ) , B ( 0 ) , Z ( 0 ) , t = 0.
While  t < t max and η < η max  do
        QA scheme: Turn an image into the balanced order-3 tensor Φ 2 X .
        Update X t = arg min X τ Ω ( X ) + ρ 2 Φ 2 X B t 1 + Λ t 1 F 2
        Update B ¯ ( i ) t = arg min Z ¯ ( i )   B ¯ ( i ) * + ρ 2 Φ 2 X t B + Λ t 1 F 2 , i = 1 ,   ,   I 3
        Update Z t = arg min Z ¯ ( i )   Z 1 + ρ 2 D X t Z + L t 1 F 2
        Update Λ t = Λ t 1 + Φ 2 X t B t , L t = L t 1 + D X t Z t
        Update η t + 1 = X n t + 1 ( : ) X n t ( : ) F X n t ( : ) F , t = t + 1 .
End while
Output:  X * .

3.4. Method 3: The Low TT Rank-Based Method

TT decomposition works better on higher-order tensors than Tucker decomposition. To fulfill TT decomposition efficiently, the third flexible QA scheme is exploited to permute the 3-order image into a higher-order tensor. Based on the basic QA scheme, high-order tensors can be obtained flexibly.
The third flexible QA scheme is shown below. Take a 16 × 16 matrix as an example, as shown in Figure 4a. First, a 2-order matrix is turned into a 3D tensor via the basic QA scheme, and the basic QA is repeated to obtain the final 4D tensor with the size of 4 × 4 × 4 × 4 . The entry comes from the ith smaller matrix of the first basic QA scheme, and the jth smaller matrix of the second basic QA scheme is labeled as the MATLAB notation ( : , : , i , j ) . By analogy, the third flexible QA scheme can permute a matrix with the size of 4 P × 4 Q to order- min { P , Q } tensor with the size of 4 × 4 × × 4 min { P , Q } . An RGB image with the size of 4 P × 4 Q × 3 can be permuted into an order- min { P , Q } + 1 tensor with the size of 4 × 4 × × 4 × 3 1 + min { P , Q } . The third flexible QA scheme should ensure that the designed tensor has a higher order.
We permute the Lena image into a high-order tensor via the third flexible QA scheme, and then obtain the TT matrices of the augmented tensor. We name these TT matrices as QA-TT matrices, and their singular values are shown in Figure 4b, which demonstrates the low TT rankness of the rearranged tensor.
Then, we enforce the low TT rankness to improve the inpainting accuracy. The third model is as follows:
min X n = 1 N α n T n Φ 3 X * s u b j e c t   t o     X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω
where Φ 3 stands for the third flexible QA used to permute image X into a high-dimensional tensor. We name the tensor obtained by the third flexible QA scheme as a QA tensor. T n is the operator that converts a tensor into the nth TT matrix, n = 1 ,   2 ,   ,   N . The order of QA tensor is N . The inverse operators corresponding to Φ and T n are Φ 1 and T n 1 , respectively. The weight α n is as follows:
α n = θ n n = 1 N 1 θ n with   θ n = min ( l = 1 n I l , l = k + 1 N I l )
where I 1 × I 2 × × I N is the size of the QA tensor.
Combining the low TT rank and sparsity constraints, we introduce auxiliary variables Z = D X and U n V n H = T n Φ 3 X and rewrite (14) as the following unconstrained convex optimization problem for all n = 1 ,   , N 1 :
min U n , V n , Λ n , L , Z , X τ Ω ( X ) + 1 2 n = 1 N 1 α n ( U n F 2 + V n F 2 ) + β 2 Z 1 + ρ 1 2 n = 1 N 1 α n T n Φ 3 X U n V n H + Λ n F 2 ) + β ρ 2 2 D X Z + L F 2
By applying ADMM, each sub-problem is performed at each iteration t. Lastly, we obtain X by X * = n = 1 N 1 α n X n , where X n * represents the optimal solution of the nth subproblem. The whole algorithm for solving the model (16) is shown in Algorithm 3.
Algorithm 3. The algorithm for solving the model (16)
Input:   Y ,   Ω ,   β ,   ρ 1   , ρ 2 , the maximum number of iteration t max , convergence condition η t o l .
Initialization:  U n ( 0 ) , V n ( 0 ) by the LMaFit method [47];  Λ n ( 0 ) , L n ( 0 ) , Z ( 0 ) .
For n = 1 to N − 1 do
   t = 0.
   While  t < t max and η < η max  do
        QA scheme: permute image to order-N tensor Φ 3 X .
        Update X n t = arg min X τ Ω ( X ) + ρ 1 2 T n Φ 3 X U n t 1 V n ( t 1 ) H + Λ n t 1 F 2
        Update U n t = arg min U n U n F 2 + ρ 1 2 T n Φ 3 X t U n V n ( t 1 ) H + Λ n t 1 F 2
        Update V n t = arg min V n V n F 2 + ρ 1 2 T n Φ 3 X t U n t V n H + Λ n t 1 F 2
        Update Z t = arg min Z Z 1 + ρ 2 2 D X t Z + L t 1 F 2
        Update Λ n t = Λ n t 1 + T n Φ 3 X t U n t V n ( t ) H , L t = L t 1 + D X t Z t
        Update η t + 1 = X n t + 1 ( : ) X n t ( : ) F X n t ( : ) F , t = t + 1 .
   End while
End for
Output:  X * = n = 1 N 1 α n X n * .

4. Experimental Results and Analyses

We applied the above methods 1–3 to solve image inpainting problems. For simplicity, methods 1–3, which only exploit low unfolding matrix rank, low tensor tubal rank, and low tensor train rankness, are denoted as UfoldingLR, TTLR, and tSVDLR methods, respectively. The methods that enhance the low rank and total variation constraints simultaneously are denoted as UnfoldingLRTV, tSVDLRTV, and TTLRTV methods, respectively. The low matrix rank completion method, which is solved by model (17) and the ADMM algorithm, is denoted as the MatrixLR method.
min X r a n k ( X ) subject   to X ( i ,   j ) = Y ( i ,   j ) ,   ( i ,   j ) Ω
The method that only exploits sparsity in the gradient domain and is solved by the ADMM algorithm is denoted as the TV method. In addition, the following close methods were applied for comparison; some of their codes are available online.
STDC: This method exploits the images into three-factor matrices and one core tensor for image inpainting [7,21,22] (http://doi.ieeecomputersociety.org/10.1109/TPAMI.2013.164, accessed on 12 December 2019).
HaLRTC: This method constrains the low rankness of the three mode-n matrices caused by the decomposition of a color image for inpainting, which is solved by the ADMM [23,24] (https://www.cs.rochester.edu/u/jliu/publications.html, accessed on 18 December 2019).
SPCTV (https://sites.google.com/site/yokotatsuya/home/software, accessed on 12 January 2020). The smooth PARAFAC tensor completion and total variation method [25] uses the PD (PARAFAC decomposition, a derivation of Tucker decomposition) framework and constrains the TV on every factor matrix of PD.
LRTV: This method combines the constraints of the low rankness of every mode-n matrix and the TV regularization on every mode-n matrix for color image inpainting [26,27] (https://sites.google.com/site/yokotatsuya/home/software/lrtv_pds, accessed on 16 April 2021).
FBCP: Inpainting methods based on Bayesian tensor completion [28,29,30,31] (https://github.com/qbzhao/BCPF, accessed on 10 March 2021).
TTC and TTCTV: The tensor completion method with tensor train and with total variation regularized tensor trains, respectively [50] (https://github.com/IRENEKO/TTC, accessed on 9 March 2021).
CFN-RTC: This robust tensor completion method combines capped Frobenius norm and TR decomposition and adds dimensions through simple data reshaping for processing image [33] (https://github.com/Li-X-P/Code-of-Robust-Tensor-Completion, accessed on 10 March 2021).
RNC-FCTN: This robust tensor completion method is based on FCTN, but it is limited to high-dimensional data processing [32] (https://faculty.swjtu.edu.cn/zhengyubang/zh_CN/zdylm/839966/list/, accessed on 10 March 2021). Therefore, to adapt to their code, we rearranged the images into four-dimensional space.
All simulations were carried out on Windows 10 and MATLAB R2019a running on a PC with an Intel Core i7 CPU 2.8 GHz and 16 GB of memory. For a fair comparison, every method was conducted with its optimal parameters to ensure every method had the best performance. The reconstruction quality was quantified using the peak signal-to-noise ratio (PSNR) and structural similarity (SSIM) (http://www.ece.uwaterloo.ca/~z70wang/research/ssim/, accessed on 10 April 2017) [41]. The original color images (from the standard image database) and missing patterns used in the experiments are shown in Figure 5. The images used in this section are from the Berkeley Segmentation dataset (https://www2.eecs.berkeley.edu/Research/Projects/CS/vision/bsds/, accessed on 21 April 2024) and USC-SIPI image database (http://sipi.usc.edu/database/database.php, accessed on 17 March 2023).
We set the maximum number of iterations t max = 100 and convergence condition η t o l = 10 6 in all our methods (UnfoldingLRTV, tSVDLRTV, and TTLRTV). The pixel range of all the images was normalized to 0-1. In UnfoldingLR, tSVDLR, and TTLR methods, we set   ρ 1   = 0.04, 0.002, and 0.6, respectively. In UnfoldingLRTV, Tsvdlrtv, and TTLRTV methods, the parameter set ( ρ 1   ,   β ,     ρ 2 ) was (0.4, 0.004, 2), (0.6, 0.07, 0.1), and (0.7, 0.03, 0.1), respectively.

4.1. Analyses of the Three Flexible QA Schemes

Next, we used the first, second, and third flexible QA schemes. The PSNRs (dB)/SSIMs of the UnfoldingLR, tSVDLR, and TTLR methods with and without the QA scheme are shown in Table 3. The red numerical values correspond to the worst results. It can be seen that, without the QA scheme, Lena and Airplane could not be recovered. The UnfoldingLR, tSVDLR, and TTLR methods with the QA scheme had better numerical results than those without the QA scheme. In the low matrix rank completion method (i.e., MatrixLR), no QA scheme was applied, i.e., the color image was dealt with as three-channel matrices directly.
Due to the support of the QA scheme, the low tensor rank-based methods (TTLR, tSVDLR, and UnfoldingLR) with the QA scheme provided better results than the traditional low matrix rank completion method (i.e., MatrixLR method). Therefore, the QA scheme can be successfully used as the first step to explore the low tensor rank prior to an image.
The KA scheme and the third flexible QA scheme can both rearrange an image into a high-order tensor. However, our QA scheme is different from the KA scheme used in [23]. The KA scheme can maintain the local block similarity of the image, while the third flexible QA scheme can use adjacent pixels to maintain the global similarity of the image. We conducted a comparison of KA and the third flexible QA scheme under the corresponding TMac-TTKA [23] and TTLR methods. As shown in Figure 6, the small blocks are obvious in the recovered images obtained by the TMac-TTKA method. The images recovered by the TTLR method preserve more details without the obvious blocks.

4.2. Analyses of the Methods Exploiting Both Low Rankness and Sparsity

We analyzed the recovery results of the methods exploiting low rankness and sparsity. Figure 7, Figure 8 and Figure 9 show visual comparisons of the eleven methods for recovering the House, Lena, and Baboon images, respectively. Table 4 shows the PSNR (dB)/SSIM results of the nine methods for recovering different color images under different missing patterns. Figure 10 depicts the PSNR curves of the inpainting results using the different methods; the missing ratio ranges from 10% to 70% under a random missing pattern.
As shown in Figure 7, Figure 8, Figure 9 and Figure 10 and Table 4, compared to the numerous close STDC, HaLRTC, FBCP, TMac-TTKA, SPCTV, and LRTV methods, the UnfoldingLRTV, tSVDLRTV, and TTLRTV methods had superior performance for both visual and quantity results. The SPCTV and LRTV methods also enhanced the low rankness and sparsity simultaneously, but the results were worse than our methods.
Table 5 shows the PSNR (dB)/SSIM results of the eight methods. The MatrixLR method only constrains the low matrix rank; the TV method only exploits the TV prior; the UnfoldingLR, tSVDLR, and TTLR methods only constrain the low unfolding matrix rank, low tubal rank, and low TT rank, respectively; and the UnfoldingLRTV, tSVDLRTV, and TTLRTV methods combine both sparsity and low tensor rankness. As shown in Table 5, the combination of sparsity and low tensor rankness constraints could yield better inpainting results than enforcing sparsity or low rankness alone. The TTLR method was more efficient than the MatrixLR and TV methods. The results of the tSVDLR method and the TTLR method were comparable. The UnfoldingLR method provided the best results among the TTLR, tSVDLR, TuckerLR, MatrixLR, and TV methods. UnfoldingLRTV, tSVDLRTV, and TTLRTV methods had improved numerical results compared to the corresponding UnfoldingLR, tSVDLR, and TTLR methods, which demonstrates that TV prior is efficient in improving the accuracy of low rank-based inpainting methods.
The visual and numerical PSNR (dB)/SSIM comparisons of our methods for recovering the pepper image under 80% random missing patterns are shown in Figure 11. In the first row of Figure 11, the methods only exploit low-rank constraints. As shown in the color box, there are small blocky errors in the recovered image, these are caused by the QA scheme. This phenomenon can be solved by combining the constraints of low rank and sparsity (TV), as shown in the second row of Figure 11.
All in all, due to the support of the QA scheme and the efficient TV prior, the low tensor rank-based methods (UnfoldingLRTV, tSVDLRTV, and TTLRTV) were superior to other close low tensor rank-based methods. The UnfoldingLRTV method provided the best results among all the methods examined.

4.3. Analyses of TTLR and TTLRTV Methods

The TT-based methods (i.e., TTLR and TTLRTV) were analyzed in detail. As TT rank is multi-rank, how does every TT matrix rank affect the final result? We answer this question with the below experimental results.
We conducted the experiments on recovering House, Lena, and Airplane images with a size of 256 × 256 × 3. The random missing patterns had four missing ratios: 10%, 30%, 50%, and 70%, respectively. We labeled the eight TT matrices as k = 1, 2, …, 8. The PSNR (dB) results of X n * (the optimal solution of the nth subproblem, which exploits the nth TT matrix rank), in the TTLR and TTLRTV methods are shown in Figure 12.
From Figure 12, it can be seen that the PSNR (dB) results of each subproblem is steeply different for the TTLR method, which demonstrates that each TT matrix rank contributes different PSNR results. As there are no rules to find which TT matrix rank meets the best PSNR result, each solution of X n * should be combined to obtain the final X * , i.e., X * = n = 1 N 1 α n X n * . Comparing the PSNR curves of TTLR and TTLRTV methods in Figure 12, the PSNR (dB) results of each subproblem is slightly different for the TTLRTV method, which demonstrates that the combination of TT and TV can make the PSNR more balanced among all k.

4.4. Missing Ratio of 90%

To further test the effectiveness of our method under 90% missing entries, we conducted comparative experiments with the state-of-the-art TTC and TCTV methods [50] based on tensor TT decomposition, the CFN-RTC method [33] based on TR decomposition, and the RNC-FCTN method [32] based on FCTN decomposition. For fairness, we directly used the codes and images provided by the TTC and TTCTV methods in [50] and https://github.com/IRENEKO/TTC, where the related parameters are also set by their code. We directly used the codes provided by the CFN-RTC and RNC-FCTN methods, set the optimal parameters such as tensor rank, and arranged the input images according to the size in their codes. The original image is shown in Figure 13. At a random 90% missing entries, the PSNR and SSIM of the observed image were 5.7036 dB and 3.56%, respectively. Comparing ten methods, including TTC, TTCTV, CFN-RTC, RNC-FCTN, TTLR, UnfoldingLR, TTLRTV, tSVDLR, tSVDLRTV, and UnfoldingLRTV, the visual results are shown in Figure 13, and the numerical indicators are shown in Table 6.
Experiments showed that, under high missing rates, there was significant visual pollution in the recovered images of the ten methods. However, compared to the original polluted image, there was a significant improvement, and there were differences in the recovered details among the ten methods. The CFN-RTC, TTC, TTCTV, and tSVDL methods had significant grid-like pollution in repairing images, while the TTLR and UnfoldingLR methods had significant block-like pollution. The RNC-FCTN method had the worst visual effect, but this does not mean that this method has no advantage in processing high-dimensional videos or hyperspectral video (HSV) data. We only forced it to process image data in this experiment, and the results showed that it had a certain effect. The TTLRTV, tSVDLRTV, and UnfoldingLRTV methods had better visual effects than the other methods, but there was significant pulse-like noise in the details.
In terms of numerical indicators, the TTC and TTCTV methods could only improve PSNR and SSIM by about 3 dB and 14%, respectively, which is clearly not advantageous compared to our five methods. Our five methods could improve PSNR and SSIM by more than 13 dB and 40%, respectively. The CFN-RTC method could also achieve close improvements.
The RTC-FCTN method is based on FCTN decomposition, while the CFN-RTC method is based on TR decomposition. FCTN is an improvement of TR decomposition, but the experimental results of the two methods were different. The RNC-FCTN method was directly used to process high-dimensional data, while in the CFN-RTC method experiment, the authors first simply reshaped low-dimensional data into high-dimensional data, which enabled the TR decomposition tool to be applied to low-dimensional data processing.
Analyzing the reasons for the above experimental results, our method may greatly benefit from using more flexible data rearrangement schemes, which make low-rank approximation more sufficient and are based on more effective distributed iterative solving algorithms. Moreover, the TTC and TTCTV methods are not based on the inpainting model established by decomposing tensors into multiple TT matrix forms but rather on directly decomposing tensors into less precise forms of multiple core tensors.

4.5. Runtime and Complexity Analysis

From a high-dimensional curse perspective, converting an image to a higher-order tensor can result in increased complexity, which inevitably leads to a longer runtime. We compared our methods (UnfoldingLRTV, tSVDLRTV, and TTLRTV) with the traditional MatrixLR method and the close STDC, HaLRTC, FBCP, SPCQV, and LRTV methods in terms of running time, as shown in Table 7.
UnfoldingLRTV methods: In the first step, the QA scheme was used to decompose a single image into several small graphs. Because of the similarity of these small graphs, the QA tensor could be reduced to a matrix with a low-rank structure in an unfolding way. Ignoring TV constraints, the unfoldingLRTV method only needed to solve the low-rank matrix completion problem of an unfolding matrix, so the running time was similar to the traditional MatrixLR method, and the accuracy was higher than the traditional MatrixLR method.
The tSVDLRTV methods: As the color image was highly unbalanced in the size of the three dimensions, which is not suitable for the low tubal rank constraint, we used the QA scheme to rearrange an image into a third-order tensor with a more balanced size of every dimension. Then, TNN was used to constrain the low tubal rank of the rearranged tensor. Due to the fast Fourier scheme, it is necessary to perform a low-rank matrix constraint on each frontal slice after the threeh more efficient low t-dimensional Fourier transform. At this time, the SVD decomposition process will increase the time consumption.
TTLRTV methods: TT multi-rank is the combination of the rank of each TT matrix. The TTLRTV method essentially completes the same data amount N-1 times, where N is the order of the QA tensor. Therefore, although the TITRTV method is effective, it is necessarily more computationally expensive than the low matrix rank completion method.
In summary, among the three methods, the UnfoldingLRTV method achieved the best performance both in accuracy and runtime. The TTLRTV method reached better accuracy, but it was time-consuming. The tSVDLRTV method showed moderate performance both in runtime and accuracy.
All in all, the above three methods can exploit the potential low-rank prior of an image and can be successfully used for image inpainting problems, demonstrating that the three flexible QA schemes are perfect for exploring the low-rank prior of an image.

5. Conclusions

To effectively explore the potential of low tensor rank prior to an image, a rearrangement scheme (QA) was first exploited for permuting the color image (3-order) into three flexible rearrangement forms (with a more efficient low tensor rank structure). Based on the scheme, three optimization models exploiting the low unfolding matrix rank, low tensor tubal rank, and low TT multi-rank were developed to improve the accuracy in image inpainting. Combined with TV constraints, efficient ADMM algorithms were developed to solve these three optimization models. The experimental results demonstrate that our low tensor rank-based methods are effective for image inpainting and are superior to the low matrix rank completion method and numerous close methods. The low tensor rank constraint is effective for image inpainting, which is mainly due to the support of the QA scheme.

Author Contributions

Conceptualization, S.M.; methodology, S.M.; investigation, S.M. and Y.F.; resources, S.M. and S.F.; writing—original draft preparation, S.M. and S.F.; writing—review and editing, S.F., W.Y., and Y.F.; supervision, L.L. and W.Y.; funding acquisition, S.M. and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Key Laboratory of Science and Technology on Space Microwave, No. HTKJ2021KL504012; supported by the Science and Technology Innovation Cultivation Fund of Space Engineering University, No. KJCX-2021-17; and supported by the Information Security Laboratory of National Defense Research and Experiment, No.2020XXAQ02.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pendu, M.; Jiang, X.; Guillemot, C. Light field inpainting propagation via low rank matrix completion. IEEE Trans. Image Process. 2018, 27, 1989–1993. [Google Scholar] [CrossRef] [PubMed]
  2. Yu, Y.; Peng, J.; Yue, S. A new nonconvex approach to low-rank matrix completion with application to image inpainting, Multidim. Syst. Signal Process. 2018, 30, 145–174. [Google Scholar]
  3. Gong, X.; Chen, W.; Chen, J. A low-rank tensor dictionary learning method for hyperspectral image denoising. IEEE Trans. Signal Process. 2020, 68, 1168–1180. [Google Scholar] [CrossRef]
  4. Su, X.; Ge, H.; Liu, Z.; Shen, Y. Low-Rank tensor completion based on nonconvex regularization. Signal Process. 2023, 212, 109157. [Google Scholar] [CrossRef]
  5. Ma, S.; Ai, J.; Du, H.; Fang, L.; Mei, W. Recovering low-rank tensor from limited coefficients in any ortho-normal basis using tensor-singular value decomposition. IET Signal Process. 2021, 19, 162–181. [Google Scholar] [CrossRef]
  6. Liu, Y.; Long, Z.; Zhu, C. Image completion using low tensor tree rank and total variation minimization. IEEE Trans. Multimedia. 2018, 21, 338–350. [Google Scholar] [CrossRef]
  7. Gong, W.; Huang, Z.; Yang, L. Accurate regularized Tucker decomposition for image restoration. Appl. Math. Model. 2023, 123, 75–86. [Google Scholar] [CrossRef]
  8. Long, Z.; Liu, Y.; Chen, L.; Zhu, C. Low rank tensor completion for multiway visual data. Signal Process. 2019, 155, 301–316. [Google Scholar] [CrossRef]
  9. Kolda, T.; Bader, B. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  10. Kilmer, M.; Braman, K.; Hao, N.; Hoover, R.C. Third-order tensors as operators on matrices: A theoretical and computational framework with applications in imaging. SIAM J. Matrix Anal. Appl. 2013, 34, 148–172. [Google Scholar] [CrossRef]
  11. Semerci, O.; Hao, N.; Kilmer, M.; Miller, E.L. Tensor-based formulation and nuclear norm regularization for multienergy computed tomography. IEEE Trans. Image Process. 2014, 23, 1678–1693. [Google Scholar] [CrossRef] [PubMed]
  12. Zhou, P.; Lu, C.; Lin, Z.; Zhang, C. Tensor factorization for low-rank tensor completion. IEEE Trans. Image Process. 2018, 3, 1152–1163. [Google Scholar] [CrossRef] [PubMed]
  13. Oseledets, I.; Tyrtyshnikov, E. TT-cross approximation for multidimensional arrays. Linear Algebra Appl. 2010, 432, 70–88. [Google Scholar] [CrossRef]
  14. Oseledets, I. Tensor-train decomposition. SIAM J. Sci. Comput. 2011, 33, 2295–2317. [Google Scholar] [CrossRef]
  15. Hackbusch, W.; Kuhn, S. A new scheme for the tensor representation. J. Fourier Anal. Appl. 2009, 15, 706–722. [Google Scholar] [CrossRef]
  16. Zhao, Q.; Zhou, G.; Xie, S.; Zhang, L.; Cichocki, A. Tensor ring decomposition. arXiv 2016, arXiv:1606.05535. [Google Scholar]
  17. Zheng, Y.; Huang, T.; Zhao, X.; Zhao, Q.; Jiang, T. Fully-connected tensor network decomposition and its application to higher-order tensor completion. Proc. AAAI Conf. Artif. Intell. 2021, 35, 11071–11078. [Google Scholar] [CrossRef]
  18. Zhang, Z.; Aeron, S. Exact tensor completion using T-SVD. IEEE Trans. Signal Process. 2017, 65, 1511–1526. [Google Scholar] [CrossRef]
  19. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor robust principal component analysis with a new tensor nuclear norm. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 925–938. [Google Scholar] [CrossRef]
  20. Du, S.; Xiao, Q.; Shi, Y.; Cucchiara, R.; Ma, Y. Unifying tensor factorization and tensor nuclear norm approaches for low-rank tensor completion. Neurocomput 2021, 458, 204–218. [Google Scholar] [CrossRef]
  21. Chen, Y.; Hsu, C.; Liao, H. Simultaneous tensor decomposition and completion using factor priors. IEEE Trans. Pattern Anal. Mach. Intell. 2014, 36, 577–591. [Google Scholar] [CrossRef] [PubMed]
  22. Xue, J.; Zhao, Y.; Liao, W.; Chan, J.C.-W.; Kong, S.G. Enhanced sparsity prior model for low-rank tensor completion. IEEE Trans. Neural. Netw. Learn. Syst. 2019, 31, 4567–4581. [Google Scholar] [CrossRef]
  23. Liu, J.; Musialski, P.; Wonka, P.; Ye, J. Tensor completion for estimating missing values in visual data. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 208–220. [Google Scholar] [CrossRef] [PubMed]
  24. Qin, M.; Li, Z.; Chen, S.; Guan, Q.; Zheng, J. Low-Rank Tensor Completion and Total Variation Minimization for Color Image Inpainting. IEEE Access 2020, 8, 53049–53061. [Google Scholar] [CrossRef]
  25. Yokota, T.; Zhao, Q.; Cichocki, A. Smooth PARAFAC decomposition for tensor completion. IEEE Trans. Signal Process. 2016, 64, 5423–5436. [Google Scholar] [CrossRef]
  26. Yokota, T.; Hontani, H. Simultaneous visual data completion and denoising based on tensor rank and total variation minimization and its primal-dual splitting algorithm. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3732–3740. [Google Scholar]
  27. Li, L.; Jiang, F.; Shen, R. Total Variation Regularized Reweighted Low-rank Tensor Completion for Color Image Inpainting. In Proceedings of the 25th IEEE International Conference on Image Processing (ICIP), Athens, Greece, 7–10 October 2018; pp. 2152–2156. [Google Scholar]
  28. Zhao, Q.; Zhang, L.; Cichocki, A. Bayesian CP factorization of incomplete tensors with automatic rank determination. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1751–1763. [Google Scholar] [CrossRef]
  29. Wang, X.; Philip, L.; Yang, W.; Su, J. Bayesian robust tensor completion via CP decomposition. Pattern Recognit. Lett. 2022, 163, 121–128. [Google Scholar] [CrossRef]
  30. Zhu, Y.; Wang, W.; Yu, G.; Wang, J.; Tang, L. A Bayesian robust CP decomposition approach for missing traffic data imputation. Multimed Tools Appl. 2022, 81, 33171–33184. [Google Scholar] [CrossRef]
  31. Cui, G.; Zhu, L.; Gui, L.; Zhao, Q.; Zhang, J.; Cao, J. Multidimensional clinical data denoising via Bayesian CP factorization. Sci. China Technol. 2020, 63, 249–254. [Google Scholar] [CrossRef]
  32. Liu, Y.; Zhao, X.; Song, G.; Zheng, Y.; Ng, M.K.; Huang, T. Fully-connected tensor network decomposition for robust tensor completion problem. Inverse Probl. Imaging 2024, 18, 208–238. [Google Scholar] [CrossRef]
  33. Li, X.P.; Wang, Z.; Shi, Z.; So, H.C.; Sidiropoulos, N.D. Robust tensor completion via capped Frobenius norm. IEEE Trans. Neural Netw. Learn. Syst. 2024, 35, 9700–9712. [Google Scholar] [CrossRef] [PubMed]
  34. Bengua, J.; Phien, H.; Tuan, H.; Do, M.N. Efficient tensor completion for color image and video recovery: Low-Rank Tensor Train. IEEE Trans. Image Process. 2017, 26, 1057–7149. [Google Scholar] [CrossRef]
  35. Ma, S.; Du, H.; Hu, J.; Wen, X.; Mei, W. Image inpainting exploiting tensor train and total variation. In Proceedings of the 12th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI), Suzhou, China, 19–21 October 2019; pp. 1–5. [Google Scholar]
  36. Ma, S. Video inpainting exploiting tensor train and sparsity in frequency domain. In Proceedings of the IEEE 6th International Conference on Signal and Image Processing (ICSIP), Nanjing, China, 22–24 October 2021; pp. 1–5. [Google Scholar]
  37. Ma, S.; Du, H.; Mei, W. Dynamic MR image reconstruction from highly undersampled (k, t)-space data exploiting low tensor train rank and sparse prior. IEEE Access 2020, 8, 28690–28703. [Google Scholar] [CrossRef]
  38. Latorre, J. Image Compression and Entanglement. Available online: https://arxiv.org/abs/quant-ph/0510031 (accessed on 3 March 2018).
  39. Kilmer, M.; Martin, C. Factorization strategies for third-order tensors. Linear Algebra Appl. 2011, 435, 641–658. [Google Scholar] [CrossRef]
  40. Martin, C.; Shafer, R.; Larue, B. An order-p tensor factorization with applications in imaging. SIAM J. Sci. Comput. 2013, 35, A474–A490. [Google Scholar] [CrossRef]
  41. Oseledets, I. Compact matrix form of the d-dimensional tensor decomposition. In Proceedings of the International Symposium on Nonlinear Theory and Its Applications, Sapporo, Japan, 19–21 October 2009. [Google Scholar]
  42. Lingala, S.; Hu, Y.; Dibella, E.; Jacob, M. Accelerated dynamic MRI exploiting sparsity and low-rank structure: K-t SLR. IEEE Trans. Med. Imaging. 2011, 30, 1042–1054. [Google Scholar] [CrossRef] [PubMed]
  43. Recht, B.; Fazel, M.; Parrilo, P. Guaranteed minimum-rank solutions of linear matrix equations via nuclear norm minimization. SIAM Rev. 2010, 52, 471–501. [Google Scholar] [CrossRef]
  44. Signoretto, M.; Cevher, V.; Suykens, J. An SVD-free approach to a class of structured low rank matrix optimization problems with application to system identification. In Proceedings of the IEEE Conference on Decision and Control (CDC), Firenze, Italy, 10–13 December 2013. no. EPFL-CONF-184990. [Google Scholar]
  45. Liu, H.; Xiong, R.; Zhang, X.; Zhang, Y. Nonlocal gradient sparsity regularization for image restoration. IEEE Trans. Circ. Syst. Vid. 2017, 27, 1909–1921. [Google Scholar] [CrossRef]
  46. Feng, X.; Li, H.; Li, J.; Du, Q. Hyperspectral Unmixing Using Sparsity-Constrained Deep Nonnegative Matrix Factorization with Total Variation. IEEE Trans. Geosci. Remote Sens. 2018, 56, 6245–6257. [Google Scholar] [CrossRef]
  47. Wen, Z.; Yin, W.; Zhang, Y. Solving a low-rank factorization model for matrix completion by a nonlinear successive over-relaxation algorithm. Math. Program. Comput. 2012, 4, 333–361. [Google Scholar] [CrossRef]
  48. Ai, J.; Ma, S.; Du, H.; Fang, L. Dynamic MRI Reconstruction Using Tensor-SVD. In Proceedings of the 14th IEEE International Conference on Signal Processing, Beijing, China, 12–16 August 2018; pp. 1114–1118. [Google Scholar]
  49. Wang, Z.; Bovik, A.; Sheikh, H.; Simoncelli, E. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  50. Ko, C.; Batselier, K.; Yu, W.; Wong, N. Fast and Accurate Tensor Completion with Total Variation Regularized Tensor Trains. IEEE Trans. Image Process. 2020, 29, 6918–6931. [Google Scholar] [CrossRef]
Figure 1. (a) Examples of the basic QA scheme. Using the basic QA scheme, the matrix M size of 8 × 8 can be turned into an order-3 tensor with a size of 4 × 4 × 4. (b) Using the basic QA scheme, the Lena image can be divided into 4 small Lena images. (c) The four-pixel values curves of the four smaller Lena images have overlapped into one curve.
Figure 1. (a) Examples of the basic QA scheme. Using the basic QA scheme, the matrix M size of 8 × 8 can be turned into an order-3 tensor with a size of 4 × 4 × 4. (b) Using the basic QA scheme, the Lena image can be divided into 4 small Lena images. (c) The four-pixel values curves of the four smaller Lena images have overlapped into one curve.
Applsci 15 00322 g001
Figure 2. (a) The first flexible QA scheme to obtain the unfolding matrix. Take the Lena RGB image as an example. First, the image size of 256 × 256 × 3 is permuted to order-3 tensor with the size of 32 × 32 × 192 using the basic QA scheme. Then, this order-3 tensor is reshaped into the unfolding matrix of size 1024 × 192. (b) The singular values of this unfolding matrix.
Figure 2. (a) The first flexible QA scheme to obtain the unfolding matrix. Take the Lena RGB image as an example. First, the image size of 256 × 256 × 3 is permuted to order-3 tensor with the size of 32 × 32 × 192 using the basic QA scheme. Then, this order-3 tensor is reshaped into the unfolding matrix of size 1024 × 192. (b) The singular values of this unfolding matrix.
Applsci 15 00322 g002
Figure 3. (a) The second flexible QA scheme to permute the image into a balanced order-3 tensor. Take the Lena image size of 256 × 256 × 3 as an example. The balanced order-3 tensor size of 64 × 64 × 48 is obtained by multiple QA schemes. This balanced order-3 tensor is more suitable for the t-SVD decomposition than the original image size of 256 × 256 × 3. (b) The low tubal rankness of the balanced order-3 tensor.
Figure 3. (a) The second flexible QA scheme to permute the image into a balanced order-3 tensor. Take the Lena image size of 256 × 256 × 3 as an example. The balanced order-3 tensor size of 64 × 64 × 48 is obtained by multiple QA schemes. This balanced order-3 tensor is more suitable for the t-SVD decomposition than the original image size of 256 × 256 × 3. (b) The low tubal rankness of the balanced order-3 tensor.
Applsci 15 00322 g003
Figure 4. (a) Examples of the third flexible QA scheme. Using the third flexible QA, the matrix size of 16 × 16 can be permuted into an order-4 tensor size of 4 × 4 × 4 × 4. (b) Singular values of TT matrices. We permute the order-3 Lena RGB image size of 256 × 256 × 3 into an 8-order tensor with the size of 4 × 4 × 4 × 4 × 4 × 4 × 4 × 4 × 3 by the third flexible QA scheme. Then, eight different TT matrices are obtained from this higher-order tensor. We labeled those TT matrices as k = 1, 2, …, 8.
Figure 4. (a) Examples of the third flexible QA scheme. Using the third flexible QA, the matrix size of 16 × 16 can be permuted into an order-4 tensor size of 4 × 4 × 4 × 4. (b) Singular values of TT matrices. We permute the order-3 Lena RGB image size of 256 × 256 × 3 into an 8-order tensor with the size of 4 × 4 × 4 × 4 × 4 × 4 × 4 × 4 × 3 by the third flexible QA scheme. Then, eight different TT matrices are obtained from this higher-order tensor. We labeled those TT matrices as k = 1, 2, …, 8.
Applsci 15 00322 g004
Figure 5. Original color images and missing patterns.
Figure 5. Original color images and missing patterns.
Applsci 15 00322 g005
Figure 6. Comparison of KA and QA schemes under the corresponding TMac-TTKA method [23] and our TTLR method. The first row lists the painted images with a random missing pattern, and the missing ratio is 80%. The second row lists the recovered images by the TMac-TTKA method. The last row lists the recovered images by the LRTT method.
Figure 6. Comparison of KA and QA schemes under the corresponding TMac-TTKA method [23] and our TTLR method. The first row lists the painted images with a random missing pattern, and the missing ratio is 80%. The second row lists the recovered images by the TMac-TTKA method. The last row lists the recovered images by the LRTT method.
Applsci 15 00322 g006
Figure 7. The missing patterns and inpainting results of House image solved by different methods.
Figure 7. The missing patterns and inpainting results of House image solved by different methods.
Applsci 15 00322 g007
Figure 8. The missing patterns and inpainting results of Lena image solved by different methods.
Figure 8. The missing patterns and inpainting results of Lena image solved by different methods.
Applsci 15 00322 g008
Figure 9. The missing patterns and inpainting results of Baboon image solved by different methods.
Figure 9. The missing patterns and inpainting results of Baboon image solved by different methods.
Applsci 15 00322 g009
Figure 10. The PSNR curves of the inpainting results of the nine methods; the missing ratio ranges from 10% to 70% under a random missing pattern.
Figure 10. The PSNR curves of the inpainting results of the nine methods; the missing ratio ranges from 10% to 70% under a random missing pattern.
Applsci 15 00322 g010
Figure 11. The visual and numerical PSNR (dB)/SSIM comparisons of our methods for recovering the pepper image under 80% random missing patterns.
Figure 11. The visual and numerical PSNR (dB)/SSIM comparisons of our methods for recovering the pepper image under 80% random missing patterns.
Applsci 15 00322 g011
Figure 12. The PSNR (dB) results of X n * (in Algorithm 3, the optimal solution of the nth subproblem exploits the nth TT matrix rank) using the TTLR method and the TTLRTV method. We permuted the image size of 256 × 256 × 3 to an order-9 tensor by the QA scheme. Then, the TT matrices of this order-9 tensor were labelled as k = 1, 2, …, 8. We used the random missing patterns with four missing ratios: 10%, 30%, 50%, and 70% respectively. The tested color images for the PSNR curves in (ac) are House, Lena and Airplane images, respectively.
Figure 12. The PSNR (dB) results of X n * (in Algorithm 3, the optimal solution of the nth subproblem exploits the nth TT matrix rank) using the TTLR method and the TTLRTV method. We permuted the image size of 256 × 256 × 3 to an order-9 tensor by the QA scheme. Then, the TT matrices of this order-9 tensor were labelled as k = 1, 2, …, 8. We used the random missing patterns with four missing ratios: 10%, 30%, 50%, and 70% respectively. The tested color images for the PSNR curves in (ac) are House, Lena and Airplane images, respectively.
Applsci 15 00322 g012
Figure 13. The 90% missing patterns and inpainting results solved by the ten methods.
Figure 13. The 90% missing patterns and inpainting results solved by the ten methods.
Applsci 15 00322 g013
Table 1. An abbreviation index.
Table 1. An abbreviation index.
AbbreviationsFull Terms
t-SVDtensor singular value decomposition
TTtensor train
TVtotal variation
KAket augmentation
QAquarter arrangement
Table 2. Notations and definitions.
Table 2. Notations and definitions.
SymbolsNotations and Definitions
fiberA vector defined by fixing every index but one of a tensor.
sliceA matrix defined by fixing all but two indices of a tensor.
A   ( : ,   : ,   k ) The   k t h frontal slice of a 3-order tensor A .
A ( n ) Mode-n matrix, resulting from unfolding tensor A by reshaping its mode-n fibers to the columns of A ( n ) .
f-diagonal tensorOrder-3 tensor A is called f-diagonal if each frontal slice A   ( : ,   : ,   k ) is a diagonal matrix [10].
orthogonal tensorTensor A with the size of n × n × n 3 is called orthogonal tensor if A A H = I , where I stands for identity tensor if the first frontal slice I ( 1 ) is the n × n identity matrix and all other frontal slices I ( k ) ( k = 1 ,     2 , ,   n 3 ) are zero.
Table 3. PSNR (dB)/SSIM and SSIM of the seven methods without rearrangement and with rearrangement.
Table 3. PSNR (dB)/SSIM and SSIM of the seven methods without rearrangement and with rearrangement.
MethodsPSNR (dB)/SSIM of Different Color Images Under Different Missing Patterns
HouseLenaAirplaneBoats
Random 50%LinesRandom lineRandom 80%
Without rearrangementMatrixLR9.38/0.897013.34/0.58507.118/0.130819.18/0.5680
TTLR28.61/0.87113.34/0.5857.11/0.13019.25/0.519
tSVDLR32.30/0.93213.34/0.5857.11/0.13021.60/0.707
UnfoldingLR7.83/0.09313.34/0.5857.11/0.1306.32/0.102
With rearrangementTTLR30.21/0.925131.79/0.955925.77/0.879621.44/0.7144
tSVDLR29.79/0.898931.20/0.956118.91/0.838621.34/0.6879
UnfoldingLR32.58/0.941633.45/0.977128.75/0.946423.46/0.8139
Table 4. PSNR (dB)/SSIM and SSIM of the nine methods.
Table 4. PSNR (dB)/SSIM and SSIM of the nine methods.
No.MethodsPSNR (dB)/SSIM of Different Color Images Under Different Missing Patterns
HousePeppersLenaAirplaneBaboonBoats
Random 50%TextLinesRandom LineBlocksRandom 80%
Other methods1STDC32.04/0.930033.61/0.981328.56/0.899523.49/0.775627.01/0.929321.88/0.7340
2HaLRTC32.07/0.942325.84/0.949613.34/0.585019.94/0.633428.04/0.939720.56/0.6858
3FBCP26.41/0.8701NAN14.56/0.524210.25/0.195418.71/0.554620.91/0.6947
4TMac-TTKA23.18/0.811329.47/0.968129.93/0.946220.82/0.752128.04/0.94298.83/0.1229
5SPCTV29.56/0.913323.38/0.915416.02/0.610718.58/0.689424.21/0.914420.98/0.7254
6LRTV30.93/0.938236.98/0.994534.07/0.972426.82/0.922827.10/0.931921.62/0.7541
Our methods1TTLRTV33.02/0.957937.27/0.994534.94/0.982328.82/0.956129.46/0.955922.37/0.7487
2tSVDLRTV32.20/0.955037.49/0.995034.70/0.981828.03/0.950729.56/0.957422.86/0.8021
3UnfoldingLRTV35.61/0.968937.72/0.995234.87/0.982129.55/0.963929.59/0.955625.43/0.8863
Table 5. PSNR (dB)/SSIM and SSIM of the eight methods.
Table 5. PSNR (dB)/SSIM and SSIM of the eight methods.
No.MethodsPSNR (dB)/SSIM of Different Color Images Under Different Missing Patterns
HousePeppersLenaAirplaneBaboonBoats
Random 50%TextLinesRandom LineBlocksRandom 80%
1MatrixLR9.38/0.897033.23/0.981413.34/0.58507.118/0.130827.62/0.934319.18/0.5680
2TV29.70/0.881634.14/0.991329.21/0.910722.85/0.846323.18/0.906620.32/0.6103
3TTLR30.21/0.925134.86/0.989231.79/0.955925.77/0.879625.42/0.923921.44/0.7144
4tSVDLR29.79/0.898933.86/0.984031.20/0.956118.91/0.838628.03/0.937321.34/0.6879
5UnfoldingLR32.58/0.941636.86/0.993833.45/0.977128.75/0.946422.22/0.923823.46/0.8139
6TTLRTV33.02/0.957937.27/0.994534.94/0.982328.82/0.956129.46/0.955922.37/0.7487
7tSVDLRTV32.20/0.955037.49/0.995034.70/0.981828.03/0.950729.56/0.957422.86/0.8021
8UnfoldingLRTV35.61/0.968937.72/0.995234.87/0.982129.55/0.963929.59/0.955625.43/0.8863
Table 6. Under 90% missing, the PSNR (dB)/SSIM of the ten methods, the observed PSNR (dB)/SSIM was 5.7/0.0356.
Table 6. Under 90% missing, the PSNR (dB)/SSIM of the ten methods, the observed PSNR (dB)/SSIM was 5.7/0.0356.
MethodsCFN-RTCTTCTTLRtSVDLRUnfoldingLR
PSNR19.658.7019.0520.5419.44
SSIM0.45070.17450.46880.48620.4455
MethodsRNC-FCTNTTCTVTTLRTVtSVDLRTVUnfoldingLRTV
PSNR8.798.7521.4622.1121.65
SSIM0.09290.18170.54830.60070.5706
Table 7. Runtimes(s) of the different methods.
Table 7. Runtimes(s) of the different methods.
MethodsRuntime (s)
HouseLenaAirplaneBoats
Random 50%LinesRandom LinesRandom 80%
MratrixLR4.950.170.165.01
STDC5.435.135.175.16
HaLRTC8.000.880.846.84
FBCP188.3286.45132.09219.33
SPCTV19.2516.3716.0317.69
LRTV19.0820.1721.0421.05
TTLRTV145.5143.2142.6142.3
tSVDLRTV15.2315.0715.1715.14
UnfoldingLRTV9.498.538.698.72
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ma, S.; Fan, Y.; Fang, S.; Yang, W.; Li, L. Low Tensor Rank Constrained Image Inpainting Using a Novel Arrangement Scheme. Appl. Sci. 2025, 15, 322. https://doi.org/10.3390/app15010322

AMA Style

Ma S, Fan Y, Fang S, Yang W, Li L. Low Tensor Rank Constrained Image Inpainting Using a Novel Arrangement Scheme. Applied Sciences. 2025; 15(1):322. https://doi.org/10.3390/app15010322

Chicago/Turabian Style

Ma, Shuli, Youchen Fan, Shengliang Fang, Weichao Yang, and Li Li. 2025. "Low Tensor Rank Constrained Image Inpainting Using a Novel Arrangement Scheme" Applied Sciences 15, no. 1: 322. https://doi.org/10.3390/app15010322

APA Style

Ma, S., Fan, Y., Fang, S., Yang, W., & Li, L. (2025). Low Tensor Rank Constrained Image Inpainting Using a Novel Arrangement Scheme. Applied Sciences, 15(1), 322. https://doi.org/10.3390/app15010322

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop