Next Article in Journal
Adaptive Incremental Genetic Algorithm for Task Scheduling in Cloud Environments
Previous Article in Journal
On the Second-Degree Exterior Derivation of Kahler Modules on XY
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Image Denoising via Improved Dictionary Learning with Global Structure and Local Similarity Preservations

1
School of Automation, Guangdong University of Technology, Guangzhou 510006, China
2
School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu 611731, China
3
Department of Computer Science, Southern Illinois University-Carbondale, Carbondale, IL 62901, USA
4
College of Computer Science and Technology, Qingdao University, Qingdao 266071, China
5
Department of Mathematics, Southern Illinois University-Carbondale, Carbondale, IL 62901, USA
*
Author to whom correspondence should be addressed.
Symmetry 2018, 10(5), 167; https://doi.org/10.3390/sym10050167
Submission received: 4 April 2018 / Revised: 14 May 2018 / Accepted: 14 May 2018 / Published: 16 May 2018

Abstract

:
We proposed a new efficient image denoising scheme, which leads to four important contributions. The first is to integrate both reconstruction and learning based approaches into a single model so that we are able to benefit advantages from both approaches simultaneously. The second is to handle both multiplicative and additive noise removal problems. The third is that the proposed approach introduces a sparse term to reduce non-Gaussian outliers from multiplicative noise and uses a Laplacian Schatten norm to capture the global structure information. In addition, the image is represented by preserving the intrinsic local similarity via a sparse coding method, which allows our model to incorporate both global and local information from the image. Finally, we propose a new method that combines Method of Optimal Directions (MOD) with Approximate K-SVD (AK-SVD) for dictionary learning. Extensive experimental results show that the proposed scheme is competitive against some of the state-of-the-art denoising algorithms.

1. Introduction

While images are widely used in various fields, they are usually contaminated by noise during acquisition, transmission and compression. Consequently, real-life images are often degraded with noise and there is often a need for image denoising techniques. Image denoising is known to be ill-posed problem in image processing and computer vision. Theoretically, it is hard to guarantee the recovery of a distored image since image denoising is a highly under-constrained problem. For instance, medical images are usually affected by a combination of impulsive, additive or multiplicative noise [1] and it is hard to identify the type and model the noise in real world problems [2]. Images with high resolutions are desirable in many applications, e.g., object recognition [3], face clustering [4,5], and image segmentation in medical and biological science [6]. Hence, denoising is a critical step for improving the visual quality of images [7]. Denoising methods developed so far have focused one of the two forms of noise, additive and multiplicative. Though a plethora of noise removal techniques have appeared in recent years, image denoising for real-life noise still remains an important challenge [8].
A number of denoising techniques have been developed to address this problem. For example, pixel level filtering methods and patch based filtering methods such as Gaussian filtering, total variation (TV) [9], non-local means (NLM) [7], block-matching 3D filtering (BM3D) [10], and low-rank regularization [11] have provided improved image quality with image details well recovered. Among them, the classic TV method makes use of Laplacian or hyper-Laplacian models for image filtering, where they assume that natural image gradients usually exhibit heavy-tailed distributions [12,13,14]. For instance, the Hessian-Schatten approach has been proposed in [15], which maintains the advantages of TV while eliminating the staircase effect by not penalizing first-order derivatives. The patch-based filtering methods group similar image patches together and then recover their common structures. For instance, BM3D usually requires expensive pair-wise patch comparisons. Its basic idea is to get a sparse representation in the transformed domain. It first groups similar 2D patches of the image into 3D data arrays. A highly sparse representation is obtained through 3D transformation and shrinkage. Through this procedure, the finest details shared by grouped patches are captured while the essential, unique features of each individual patch are preserved. This algorithm obtains outstanding denoising performance; however, it requires many implementation tricks [16]. Though observed effective for slightly noisy image, the performance of the above-mentioned methods is far from satisfying by the over-smooth effect, due to the reason of significantly degraded accuracy in patch matching. Besides such single-image based methods, learning based methods have been developed by integrating natural image priors, such as neural network training [16], maximizing expected patch log-likehood (EPLL) [17], and fields of experts [18].
Sparse and redundant representation modeling [19] has recently received extensive research attention and found quite successful applications in signal and image processing. The most common framework for image denoising is formulated with an energy to minimize the following [20]:
min x R N 1 2 y x 2 2 + λ ψ ( x ) ,
where ψ : R N R is a regularization function, and the quadratic “data-fitting” term ensures that the estimated x is close to the noisy observation y. In general, it is difficult to find a good regularization function ψ , and, in fact, it is probably one of the most important research topics in image processing nowadays [21].
Sparse signal representation has been shown successful [20]. It describes that a signal can be approximated as a linear combination of as few as possible atoms from a given dictionary. More precisely, a target signal y R N can be described as y Φ ω , where Φ R N × M is an overcomplete dictionary if M > N and ω is a vector containing the representation coefficient of y. We are interested in seeking the sparsest solution ω , i.e., the one with the fewest nonzero entries. The solution can be obtained by solving the following problem:
min ω R M 1 2 y Φ ω 2 2 + λ ψ ( ω ) .
Here, a typical choice for the regularization term ψ ( ω ) might be the L 0 -norm of ω that counts the number of nonzero elements of ω . Exact determination of sparsest representations is known to be an non-deterministic polynomial (NP)-hard problem. Thus, a number of algorithms have been proposed to provide the sparsest approximation of a signal, including Orthogonal Matching Pursuit (OMP) and Basis Pursuit (BP) [21]. BP relaxes the L 0 -penalty by replacing it with L 1 -penalty [21]. Dictionary design employed for sparse decomposition of a signal is also an important problem. Basically, the dictionaries can be classified into two categories: non-adaptive dictionaries and adaptive dictionaries [21].
Although current methods have been shown successful, they are often designed for a specific type of noise removel problem. Unfortunately, it is usually hard to have a perfect knowledge of the noise in real world problems. To address this problem, in this paper, we propose a new image denoising method that is capable of removing both additive and multiplicative noise. Moreover, our new method integrates both learning-based and recosntruction-based parts, allowing them to mutually enhance along with the optimization procedure, leading to a powerful denoising capability. For optimization, we use a two-stage optimization strategy, which divides our objective function into convex and non-convex parts. Specifically, for the non-convex part, we embark from the framework of K-SVD denoising and improved it based on [22]; for the convex part, we use an alternating optimization and gradient descent method similar to the one used in [23].
We summarize the contributions of this paper as follows:
  • We developed an image denoising approach that processes advantages of both reconstruction-based and learning-based methods. A practical two-stage optimization solution is proposed for the implementation.
  • We introduced a sparse term to reduce the multiplicative noise approximately to additive noise. Consequently, our method is capable of removing both additive and multiplicative noise from a noisy image.
  • We used the Laplacian Schatten norm to capture the edge information and preserve small details that may be potentially ignored by learning based methods. Hence, both global and local information can be preserved in our model for image denoising.
  • We established a new method that combines Method of Optimal Directions (MOD) with Approximate K-SVD (AK-SVD) for dictionary learning.
The rest of the paper is organized as follows. In Section 2, we introduce our proposed method and develop a two-stage optimization procedure. We conduct extensive experiments and show the results in Section 3 to verify the effectiveness of the proposed method. Finally, we conclude our work in Section 4.

2. Proposed Method

In this section, we discuss the proposed method. First, we present the formulation of the proposed model. Then, we develop a practical yet effective optimization strategy for the proposed method.

2.1. Formulation

A visually meaningful image usually contains global structures such as edges, contours, textures and smooth regions. These structures constitute the visual contents and can be captured with an aid of a high-pass filter. At the same time, local image patches usually have high self similarities. Any image patch could be sparsely represented as a linear combination of the others. The local similarity is revealed via a learned dictionary. In the global structure part, it contains a fidelity term, a low rank term, and a sparse term while, in the local similarity part, it contains a patch based term and constraint. We use one formulation, which consists of two parts: one part is designed for reconstruction of global structure while the other one for preservation of local similarity.
  • Global Structure Reconstruction: High pass filter emphasizes fine details of an image by effectively enhancing contents that are of high intensity gradient in the image. After high pass filtering, clean image contains the high frequency contents that represent global structures while low frequency contents are eliminated, making the filtered image of low rank. However, since noise usually has high-frequency components too, it may still remain together with the structural information after high-pass filtering. For each pixel, noise usually does not depend on neighboring pixels while the pixels on the global structure such as edges and textures have correlations with their neighbouring pixels. To differentiate noise and structural pixels, we consider minimizing the rank of high-pass filtered image. As Schatten norm can effectively approximate the rank [24], we use the Schatten norm of high-pass filtered image to capture the underlying structures.
    Let X R n 1 × n 2 be a matrix with singular value decomoposition (SVD) X = U Σ V T where U R n 1 × n 1 and V R n 2 × n 2 are unitary matrices consisting of singular vectors of X, and  Σ R n 1 × n 2 is a rectangular diagonal matrix consisting of singular values of X. Then, the Schatten p-norm ( S p norm) of X is defined as
    X S p = k = 1 min { n 1 , n 2 } σ k p 1 p ,
    where p 1 is the order of Schatten norm and σ k is the kth singular value of X. The family of Schatten norms include three common matrix norms, including the nuclear norm ( p = 1 ), the Frobenius norm ( p = 2 ) and the spectral norm ( p = ).
    In this paper, to high-pass filter the image, we adopt an 8-neighborhoods Laplacian operator defined as
    L = 1 9 1 1 1 1 8 1 1 1 1 .
    This Laplacian filter captures 8-directional connectedness of each pixel and thus the structures of the image as well. By filtering the image with such Laplacian filter, we can obtain a low-rank filtered image L X containing the global structures of the image. Hence, it is desireable to minimize the rank of L X to ensure the low-rankness of the global structures. To achieve this goal, we propose to adopt the above defined Hessian Schatten-p norm as rank approximation, and by minimizing LS p = L X S p , the global structures of the image can be well preserved.
    Because multiplicative noise is image content dependent, it may remain mixed with the clean image after minimizing Laplacian Schatten norm of the noisy image. To alleviate the effect of multiplicative noise, we introduce a sparse matrix S that may as well capture the outliers in the case of additive noise. In a now-standard way, we minimize the 1-norm of S to obtain the sparsity. In summary, our model is as follows:
    Y = X + S + E ,
    where Y is the noisy image, X is the clean image, S denotes a matrix containing globally sparse noise, and E is the remaining noise matrix. For convenience of optimization, we use Frobenius norm as a loss function to measure the strength of E. Combining them together, we formulate an objective function to preserve the global structure as follows:
    min X , S Y X S F 2 + λ 1 L X S p p + λ 2 S 1 ,
    where · 1 represents the matrix l 1 norm, and λ 1 , λ 2 are balancing parameters.
  • Local Similarity Preservation: We define the local similarity of an image using its patches with a size of n × n pixels. We define an operator R i that extracts the ith patch from X and orders it as a column vector, i.e., x i = R i X R n × 1 . To preserve the local similarity of image patches we exploit the dictionary learning. Define a dictionary Φ R n × k , where k > n is the number of dictionary basis. Each column of Φ is a basis, i.e., Φ = [ φ 1 , , φ n ] and the dictionary is redundant. The local similarity suggests that every patch x in the clean image may be sparsely represented over this dictionary. The sparse representation vector is obtained by solving the following constrained minimization problem:
    ω * = arg min ω ω 0 s . t . Φ ω x 2 2 ϵ
    or alternatively by MAP estimator
    ω * = arg min ω Φ ω x 2 2 s . t . ω 0 T
    where ω * is the sparse representation vector of patch φ , · 0 represents the l 0 norm and ϵ and T are parameters that control the error of the sparse coding and the sparsity of representation.
For reconstruction of global structures and preservation of local similarities, the unified image denoising needs to solve the following optimization problem:
min X , S , Φ , ω 1 , , ω M Y X S F 2 + λ 1 L X S p p + λ 2 S 1 + μ i = 1 N R i X Φ ω i 2 2 s . t . ω i 0 T .
It is seen that the above model has incorporated both global and local information in the first three terms and the last term, respectively, to recover the original image. We will develop an effective optimization scheme in the remainder of this section.

2.2. Practical Solution for Optimization

In the following, for the ease of notation, we define Ω to be the matrix of which each column is ω i , i.e., Ω = [ ω 1 , , ω M ] . It is noted that the first three terms of Equation (7) are pixel based while the last and the constraint are patch based. Due to this fact, it is difficult to directly optimize the overall objective function. Inspired by [24], we use a two-stage approach to find a local optimal solution in which we do optimization over pixel and patch based terms separately.
We decompose the objective function into two parts, each of which contains only pixel-wise operation or patch-wise operation:
F ( X , S , Φ , Ω ) = G ( X , S ) + μ H ( X , Φ , Ω )
where
G ( X , S ) = Y X S F 2 + λ 1 L X S p p + λ 2 S 1 ,
and
H ( X , Φ , Ω ) = i = 1 N R i X Φ ω i 2 2 s . t . ω i 0 T .
Basically, the updating rules of the two-stage strategy is given by alternatively applying Equations (10) and (11)
X 1 t + 1 arg min X , S G ( X , S ) + μ X X 2 t F 2 ,
X 2 t + 1 arg min X , Φ , Ω H ( X , Φ , Ω ) + 1 μ X X 1 t + 1 F 2 .
It is important to notice that the terms μ X X 2 t F 2 and 1 μ X X 1 t + 1 F 2 are critical because they represent the connection between the two stages. For simpler notation, we define two modified functions:
G ¯ ( X , S ) = G ( X , S ) + μ X X 2 t F 2 ,
H ¯ ( X , Φ , Ω ) = H ( X , Φ , Ω ) + 1 μ X X 1 t + 1 F 2 .

2.2.1. Global Structure Reconstruction Stage

For the first-stage optimization, we use an alternating strategy to optimize the function with respect to X and S by fixing one and updating another. For the initialization of X 2 0 , there are two choices: (1) for the very first iteration, we only do optimization over function G instead of G ¯ and later do optimization over the modified one; and (2) we initialize X 2 0 = Y . The second choice is potentially more computationally expensive because it forces X 2 0 to be the noisy image Y. In our experiments, we use the first approach. Since at each alternating step the objective function is convex, we make use of gradient descent method for the optimization. By the fact that Z S p p = Tr Z T Z p / 2 , the Laplacian Schatten norm term can be reformulated as
L X S p p = Tr L X T L X p / 2 .
When p = 1 , the Schatten norm is non-smooth. In addition, the sparse term is non-smooth. For the two non-smooth terms, we use two different approaches to obtain the gradient. On one hand, we introduce a small smoothing parameter δ in the above equation to get a smoothed approximation:
L X S p p = Tr L X T L X + δ 2 I p / 2 ,
where I R N 2 × N 2 is the identity matrix and δ . Thus, G could be reformulated by replacing the Schatten norm with a smoothed trace norm in Equation (8):
G ( X ) = Y X S F 2 + λ 1 Tr L X T L X + δ 2 I p / 2 + λ 2 S 1 .
Using the alternating optimization strategy, we get the updating rule of X and S in the first stage as follows:
S t ( s + 1 ) arg min S Y X t ( s ) S F 2 + λ 2 S 1 ,
X t ( s + 1 ) arg min X Y X S t ( s + 1 ) F 2 + λ 1 Tr [ ( L X ) T ( L X ) + δ 2 I ] p / 2 + μ X X 2 t F 2 ,
where t denotes the iteration of the outer optimization, s represents the iteration of the inner alternating optimization in the first stage, and X t ( s ) and S t ( s ) denote the values of X and S at the tth outer and sth inner iteration, respectively.
Notice that G ¯ ( X , S ) is not differentiable with respect to S at zeros. On the other hand, we adopt a sub-gradient when taking a derivative with respect to S, i.e.,
Δ S G ¯ = 2 ( S + X t ( s ) Y ) + λ 2 S S 1
where S S 1 is the sub-differential matrix defined as:
S i , j S 1 = s g n ( S i , j ) , S i , j 0 , 0 , S i , j = 0 .
The Laplacian Schatten norm term has been smoothed by using δ , so it is straightforward to take the derivative of G with respect to X:
Δ X G ¯ = 2 ( X + S t ( s + 1 ) Y ) + 2 ( X X 2 t ) + λ 1 p 2 ( L T L X ) [ ( L X ) T ( L X ) + δ 2 I ] p / 2 1 .
Now, using gradient descent method, S and X are updated alternatively until convergence, i.e., S t ( s , r ) r S t ( s + 1 ) and X t ( s , r ) r X t ( s + 1 ) by the follows:
S t ( s , r + 1 ) S t ( s , r ) d S Δ S G ¯ S | X t ( s , r ) , X 2 t ,
X t ( s , r + 1 ) X t ( s , r ) d X Δ X G ¯ X | S t ( s , r + 1 ) , X 2 t ,
where r denotes the iteration of gradient descent optimization.
Proposition 1.
The above updating rules including Equations (10), (11), (17)–(20) are convergent.
The proof of the proposition is in the Appendix A.

2.2.2. Dictionary Learning Stage

In the second stage, we optimize the following function:
H ¯ ( X , Φ , Ω ) = 1 μ X X 1 t F 2 + i = 1 N R i X Φ ω i 2 2 s . t . ω i 0 T .
There are three variables in this objective function: the underlying clean image X , the sparse coefficient matrix Ω and the underlying dictionary Φ . The underlying dictionary Φ is initialized with a 2D separable Discrete Cosine Transform(DCT) dictionary of size L × K . First, we produce a 1D-DCT matrix Φ 1 D of size L × K . Each atom of the matrix Φ 1 D can be obtained by φ k 1 D = c o s ( ( i 1 ) ( k 1 ) π / 11 ) , i = 1 , 2 , . . . L , k = 1 , 2 , . . . , K . Then, we use a Kronecker-product Φ = Φ 1 D Φ 1 D to initialize the dictionary Φ .
We adopt also the alternating optimization strategy for this stage. First, by fixing Φ and X, we update Ω . Then, by fixing Ω and X, we update Φ . Finally, by fixing Φ and Ω , we update X. We repeat these three steps a given number of times. In the first step, we aim to find optimal ω i by solving:
ω i = arg min ω i R i X Φ ω i 2 2 s . t . ω i 0 T .
This problem is equivalent to the following with a proper value of the penalty parameter τ :
ω i = arg min ω i R i X Φ ω i 2 2 + τ ω i 0 .
The main task of this stage is thus to solve a set of l 0 minimization problems. The exact solution of l 0 minimization is very difficult and has been proven to be NP hard. Because of this fact, matching pursuit algorithms such as basis pursuit (BP) [25], matching pursuit (MP) [26], orthogonal matching pursuit (OMP) [27], and the focal underdetermined system solver (FOCUSS) [28] are widely considered to obtain the approximate solutions of sparse representation [21]. MP and OMP greedily select the dictionary atoms sequentially. BP suggests an approximation of the sparse representation by replacing l 0 -norm with l 1 -norm. FOCUSS is similar to BP, which replaces l 0 -norm by l p -norm with 0 < p < 1 instead of l 1 -norm. This generalization measure approximates the true sparsity better when p < 1 , but the overall problem becomes nonconvex. However, convergence is not always guaranteed using the above methods. Besides the matching pursuit methods, the message passing algorithm (MPA) [19] is able to directly solve l 0 minimization problem. MPA is designed to solve l p problem with p 0 . When the problem is convex, i.e., p 1 , MPA gives global optimum and when the problem is nonconvex, i.e., 0 p < 1 , MPA finds a local minimum. Besides K-SVD, the idea of obtaining a sparse representation for a set of training image patches by learning a dictionary has been studied in a series of works during the recent years. Although we have convergence guarantee from MPA, in this paper, we adopt an OMP method as in [22] because MPA takes more time than OMP, which is effective enough in our problem. After the sparse representation matrix Ω is fixed, we adopt AK-SVD algorithm in [22] to update the dictionary Φ column by column. The AK-SVD proposes an iterative algorithm that handle the task effectively. Finally, given the coefficient matrix Ω , we then update X by solving:
X ^ = arg min X 1 μ X X 1 t F 2 + i = 1 N R i X Φ ω i 2 2 .
For this quadratic function, there is a close-form solution, which can be obtained by setting its first-order derivative to zero:
2 μ ( X X 1 t ) + 2 i = 1 N ( R i T R i X R i T Φ ω i ) = 0 ,
leading to
1 μ + i = 1 N R i T R i X = 1 μ X i t + i = 1 N R i T Φ ω i .
Hence, it is clear to see the closed-from solution of X:
X ^ = 1 μ I + i = 1 N R i T R i 1 1 μ X 1 t + i = 1 N R i T Φ ω i .
This expression implies that the clean image is obtained by averaging the denoised patches with some relaxation by averaging the patches of X 1 t , regarded as an original noised image input in the learning based stage. By introducing the additional term in Equation (13), we can directly use AK-SVD in [22] to solve Equation (21). It is flexible and can work well with OMP.
Now, following [22], we use X as an initial value to update the dictionary, which is to solve
min Φ i = 1 N R i X Φ ω i 2 2 s . t . φ j = 1 , 1 < j < M .
MOD is an appealing dictionary training algorithm [21]. A significant advantage of this method is its simple way for updating the dictionary. The MOD algorithm involves two stages described above. Assume that we fix Φ and aim to find the representations coefficient vectors ω i to build the matrix Ω by using OMP. We define the errors e i = x i Φ ω i and evaluate the overall representation mean square error using a Frobenius norm, which is given by
E F 2 = [ e 1 , e 2 , , e N ] F 2 = i = 1 N R i X Φ ω i 2 2 .
Once the sparse coding task is done, we fix X and search for an update to Φ to minimize the above error, which is (using pseudo-inverse)
Φ = X Ω T ( Ω Ω T ) 1 .
The K-SVD algorithm [21,29] takes a different update rule for the dictionary, in which the atoms in Φ are updated sequentially. Moreover, the K-SVD updates each atom along with the coefficients in Φ that multiply it using singular value decomposition (SVD) [30]. As described above, this problem leads to a matrix rank-l approximation [30] given by
min Φ X Φ Ω F 2 = min φ j 0 ( X j j 0 φ j ω j ¯ ) φ j 0 ω j 0 ¯ F 2 = min φ j 0 E j 0 φ j 0 ω j 0 ¯ 2 2 s . t . φ j 0 2 2 = 1 ,
where φ j denotes the j-th atom in Φ , ω j ¯ denotes the j-th coefficients row in Ω , E j 0 = X j j 0 φ j ω j ¯ is a known pre-computed error matrix without the j 0 -th atom. φ j 0 is the updated atom, and ω j 0 ¯ is the new coefficients row in Φ . The optimal solution can be directly obtained via performing an SVD operation.
In practice, it is difficult to obtain the exact solution of Label (31), as performing SVD for atom updating leads to its computational burden, especially in high dimensions. Therefore, Rubinstein [31] proposed a new algorithm to provide an approximate solution rather than the exact one. The resulting algorithm is known as the Approximate K-SVD (AK-SVD). The AK-SVD perform a single iteration of alternate optimization over the atom φ j 0 and the coeffcients row ω j 0 ¯ , which is given by
φ j 0 = E j 0 ω j 0 ¯ T , ω j 0 ¯ = φ j 0 T E j 0 .
This process is simple and also quite intuitive. It is important that this process not only finally converges to the optimum, but also provides an approximate solution, which effectively minimizes the error as defined in Label (31). The main contribution of the AK-SVD method is that it avoids the use of the SVD to find alternative d j 0 and ω j 0 ¯ .
Smith [32] puts forward an idea that applying multiple dictionary update cycles via the MOD or K-SVD approach can effectively minimize the representation error. In this paper, following [22], we derive a new method for dictionary learning based on multiple dictionary update cycles. We call this method as MOD-AK-SVD to distinguish it from the above reported algorithms.
Our objective is to find an update of Φ and X such that the supports in Φ remain intact. To achieve this, the dictionary update stage is divided into two optimization process. Minimizing i = 1 N R i X Φ ω i 2 2 over Φ with a fixed X and getting the results of formula (30). Next, we minimize i = 1 N R i X Φ ω i 2 2 over Φ and X keeping the support in Φ intact. By defining t j = { i : ω j ¯ ( i ) 0 } , ω j ¯ ( t j ) denotes the non-zeros coefficients in ω j ¯ . Our problem becomes
min φ j 0 , ω j 0 ¯ E j 0 ( : , t j 0 ) φ j 0 ω j 0 ¯ ( t j 0 ) 2 2 s . t . φ j 0 2 2 = 1 .
Applying alternating minimization, formula (33) leads to the following solutions:
φ j 0 = E j 0 ( : , t j 0 ) ω j 0 ¯ T ( t j 0 ) , ω j 0 ¯ ( t j 0 ) = φ j 0 T E j 0 ( : , t j 0 ) .
At the first stage, we update φ j 0 with a fixed ω j 0 ¯ ( t j 0 ) , and, at the second stage, we allow only the existing non-zeros coefficients ω j 0 ¯ ( t j 0 ) to update using the previously updated φ j 0 .
Performing a few alternations between Label (30) and Label (34) can better approximate the overall solution of Label (21).
The detailed parameter setting is described in next Section. In summary, the above practical solution is listed in Algorithm 1. We name our algorithm the Laplacian Schatten p-norm and Learning Algorithm (LSLA-p). In our work, we consider the cases when p = 1 and p = 2 , namely, LSLA-1 and LSLA-2. The empirical value of parameters in Algorithm 1 was shown in Table 1.
Algorithm 1 The Laplacian Schatten p-norm and Learning Algorithm (LSLA-p).
Require: Noisy Image: Y;
    Penalty parameter: λ 1 , λ 2 , μ ;
    Smoothing parameter: δ ;
    Stopping tolerence: ϵ 1 , ϵ 2 ;
    Clearn Image X;
1:
Initialize Φ , X 2 0 = 0 , S = 0
2:
t = 0 ; a = 1; j = 1 ; Δ 1 = ϵ 1 + 1 ; Δ 2 = ϵ 2 + 1
3:
repeat
4:
s = 0
5:
while Δ 2 ϵ 2   &   s s max do
6:
S t s + 1 arg min S Y X t s S F 2 + λ 2 S 1 (17)
7:
X t s + 1 arg min X Y X S t s + 1 F 2 + λ 1 T r [ ( L X ) T ( L X ) + δ 2 I ] p / 2 + μ X X 2 t F 2 (18)
8:
Δ 2 = min X t s + 1 X t s F 2 , S t s + 1 S t s F 2
9:
s = s + 1
10:
  end while
11:
   X 1 t + 1 = X t ( s )
12:
   X 2 t + 1 arg min X 1 μ X X 1 t + 1 F 2 + i = 1 N R i X Φ ω i 2 2 s . t . ω i 0 T (21)
13:
   Δ 1 = min X 1 t + 1 X 1 t F 2 , X 2 t + 1 X 2 t F 2
14:
   t = t + 1
15:
until Δ 1 ϵ 1 or t t max
16:
return X = X 2 t
17:
Sparse Coding Stage: Ω = OMP( X , Φ , k 0 )
18:
Update Dictionary Φ
19:
repeat
20:
   X = 1 μ I + i = 1 N R i T R i 1 1 μ X 1 t + i = 1 N R i T Φ ω i (27)
21:
   E = X Φ Ω
22:
  repeat
23:
   E j = E + φ j ω j ¯
24:
   φ j = E j ( : , t j ) ω j ¯ T ( t j ) where t j = { i : ω j ¯ ( i ) 0 }
25:
   ω j ¯ ( t j ) = φ j T E j ( : , t j )
26:
   E = E j φ j ω j ¯
   j++
27:
until j = K
  a++
28:
until a = A

3. Experiments

In this section, we conduct extensive experiments to verify the effectiveness of the proposed method. Particularly, we present the parameter settings in the first subsection and discuss the experimental results in the second subsection.

3.1. Parameter Setting

We tune the parameters in two parts: the reconstruction based part and the learning based part, which are described independently.
For the reconstruction part, usually λ 1 and λ 2 are selected from a set of values with λ 1 { 2 , 3 , 5 , 8 , 10 , 15 , 20 , 25 , 30 } and λ 2 { 0 . 01 , 0 . 02 , 0 . 05 , 0 . 08 , 0 . 1 } . Large values such as { 100 , 150 , 200 , 250 , 300 , 400 } for λ 1 and { 0 . 5 , 1 , 2 , 3 , 5 } for λ 2 are also used for a small number of images. As a common stragety for unsupervised learning methods [4], in the experiments, we use all combinations of parameters from the above sets and report the best performance. As will be clearer in the later section, our method has comparable performance with a broad range of parameters values. Parameter μ appears in both of the two stages and the setting will be mentioned later. The smoothing parameter δ is set to be 0 for LSLA-2, since the L 2 S 2 = L X S 2 2 is smooth. For LSLA-1, we test a set of values of δ . It reveals that very small δ would possibly cause numerical issues and leads to poor performance in both Peak Signal to Noise Ratio(PSNR) and Structural SIMilarity index(SSIM). Empirically, a good choice for δ is around 0 . 1 and we set δ to be 0 . 1 or 0 . 12 , i.e., δ 2 to be 0 . 01 or 0 . 014 , depending on the image with the purpose of high PSNR and SSIM.
For the dictionary learning-based part, in our work, the required parameters for this algorithm are the penalty parameter 1 / μ , the noise level σ , the patch size of dictionary L and the number of atoms in the dictionary K. The number of atoms is set to be K = 4 L 2 , where 4 is a redundancy parameter. The patch sizes in our experiments are 8, 10, 12 and 14. Usually, based on our experience, small patch size would cause an over smooth effect, and a larger patch size would increase the basis, which leads to more computation. Based on our results, finally, we use 12 × 12 patches to balance the two effects. The penalty parameter τ , i.e., 1 / μ is related to the noise level σ . In fact, it has been revealed that empirically a good relationship that leads to the best results is τ = 1 / μ = 30 / σ , i.e., μ = σ / 30 . Here, it is natural to assume that, with the iteration number of two stages increasing, the level of the remaining noise would decrease. This implies that each time when we apply K-SVD algorithm, the input σ should change to meet the variation of noise level, and thus the best parameter μ should also change with it. We initialize an estimate of noise level σ as input of K-SVD and reduce it by multiplying 0 . 9 every time after the first stage process and keep μ constant naturally. By our experience, although we do not always satisfy μ = σ / 30 , the influence is not noticeable. The reasons for this may include the following: (1) the remaining noise is small and the result is not sensitive to parameters of a given set of values in our experiments; (2) the number of alternating steps between the first and second stages is small and thus it avoids the “bad” effects to accumulate to a remarkable extent. Depending on the noise levels and types, we assign σ to be 3, 4, 5 for additive noise with noise level from low to high and 4, 6, 8 for multiplicative noise with level from low to high respectively, which empirically shows to be reasonable.
In our practical optimization method, by introducing the additional terms in Equations (12) and (13), at each stage, we make use of information from another. Based on this, it is not necessary to process the first and second stages with the same times. Learning based stage would potentially have over smoothing effects and ignore fine details in the resulting image, and, as mentioned before, the first stage would recapture the lost details to ensure the image fine details. Thus, in our experiments, we start with the first stage and end with the first stage.

3.2. Performance and Analysis

We use six standard test images, including Face, Kids, Wall, Abdomen, Nimes, and Fields with different noise types and levels to evaluate the performance of our proposed method. Face, Kids, Wall and Abdomen images are used for additive noise experiments; Fields and Nimes are used for multiplicative noise experiments. Face, Kids, and Wall images are of size 255 × 255 . Among the rest images, Nimes and Fields are of size 512 × 512 and Abdomen is of size 360 × 540 . In our work, we evaluate our performance with two criterion: peak signal-to-noise ratio (PSNR) and structure similarity (SSIM). PSNR is defined as 10 log 10 M 2 M S E , where M denotes the maximum intensity of the underlying image and M S E = 1 n 1 × n 2 i = 1 n 1 j = 1 n 2 ( X i , j X ^ i , j ) 2 is the mean squared error between the denoised image X ^ and the noiseless image X. SSIM is defined as ( 2 μ X μ X ^ + c 1 ) ( σ X , X ^ + c 2 ) / [ ( μ X 2 μ X ^ 2 + c 1 ) ( σ X 2 σ X ^ 2 + c 2 ) ] , where μ X , μ X ^ , σ X 2 , σ X ^ 2 , and σ X , X ^ denote the average of X, the average of X ^ , the variance of X and the variance of X ^ , respectively. c 1 and c 2 are two variables to stabilize the division with weak denominator. In Table 2, we provide the additive noise image restoration results in comparison with Block-matching three dimension (BM3D) [10], Hessian Schatten-norm (HS 1 ) [33], Expected Patch Log Likelihood (EPLL) [17], and Total Variation (TV) [34] for a set of four images with different noise levels. In Table 3, we list the results of our proposed method and two other methods including (multiplicative image denoising by augmented Lagrangian (MIDAL) [35] and AA [36] for a set of two images degraded by different levels of multiplicative noise.
From Table 2, it is noted that our method has the best performance in PSNR in all cases and 11 out of 12 cases in SSIM with significant improvements. In terms of PSNR, our proposed method usually outperforms BM3D by around 1–2 dB. In addition, it is observed that our methods have improved SSIM by around 0.05 on Wall and Abdomen images. Overall, LSLA-1 and LSLA-2 produce the best results although Figure 1 shows that LSLA-1 has some dark peaks in the uniform regions (liver and kidney). To visually evaluate the performance, we show some visual results. Figure 1, Figure 2, Figure 3 and Figure 4 show some examples of the resulting images by different methods. Visually, BM3D results in very clean images, but many fine details are eliminated. To better show visual effects of the methods, we enlarge some local patches in Figure 3 and Figure 4. It is observed that indeed BM3D has an over smoothing effect and the major details are missing, whereas our methods keep such details. Moreover, it is observed that the brightness of different regions has changed in the images produced by BM3D, whereas our method shows similar brightness to the original ones. Our method shows similar results to TV and HS 1 with more smoothing effects in the smooth region.
For multiplicative noise removal, it is seen that the proposed method has the best performance in all tests, except in one case when compared with MIDAL in SSIM. This observation, again, verifies that the proposed method is indeed effective for both additive and multiplicative noise removal. To visually evalute the results, we show some examples of the resulting images in Figure 5 and Figure 6. It is observed that the proposed method captures the edges well from the images, while MIDAL fails when the noise is strong. In the smooth regions, our method keeps the fine details such as the gradual change well, while MIDAL makes such gradual change difficult to distinguish. In addition, the intensities shown in the denoised images by our method appears to be much closer to the clean ones than AA and MIDAL because the contrast between the darkness and brightness is more like that in the clean images. Such observations have confirmed the effectiveness of the proposed method. At the same time, it can be seen that multiplicative noise is a much harder problem.

3.3. Parameter Sensitivity

For image denoising problem, it is hard to learn or theoretically analyze or the optimal parameters. To better investigate the performance of the proposed method, in this subsection, we present how the parameters affect the denoising performance. We have used the combination of parameters { λ 1 , λ 2 } selected from the set { 2 , 5 , 10 , 15 , 20 } × { 0 . 01 , 0 . 05 , 0 . 1 , 0 . 2 , 0 . 3 } . Without loss of generality, we test our method on two degraded images and report the results in Figure 7 where all combinations of parameters are used from the above set. It is seen that our method has comparablely high performance with a broad range of parameter values on both images, which implies the insensitivity of our method to the parameters. This observation indeed ensures the potential of our method in real world applications.

3.4. Time Comparison and Analysis

In this subsection, we test the time needed for the methods in comparison. Our simulations were performed in a MATLAB R2010b environment (MathWorks, Natick, Massachusetts, USA) on a Windows 7 operating system (Microsoft, Redmond, Washington, USA) with 2.60 GHz CPU and 4 GB RAM. Without loss of generality, we test the methods on images as given in Table 2, where average time costs are reported in Table 4 for each image and method. It is observed that the proposed methods need more time than others except EPLL. Considering that our methods have superior performance in denoising results, such cost in time is fairly acceptable.
Moreover, we investigate the time cost for each stage of our algorithm. Without loss of generality, we report the results on Face and Kids images in Figure 8. It is seen that the second stage costs roughly 60% of the overall time on average. As the major step for time cost, it should be noted that this stage involves K-SVD, which generally is time-consuming. In this paper, we aim at proposing a new image denoising method and the way to speed up our method is not in the scope of our current work, which will be considered in the future.

3.5. Discussion

From the above experiments, it is seen that the proposed methods have superior performance to state-of-the-art algorithms both quantatively and visually. Quantatively, the proposed methods improve PSNR and SSIM significantly while visually they keep fine details of the images when other methods fail. Though the proposed method has slower speed than BM3D, TV, etc., it is noted that our method is comparable to some state-of-the-art algorithms such as EPLL, yet with superior performance. Hence, it is convincing to claim the stronger applicability of the proposed method to real world applications, such as hyperspectal image denoising, biomedical image denoising, or preprocessing of noisy image data for recognition, etc. Possible reasons for slower speed of the proposed method may be the need of matrix inverse operations and sparse coding, which generally have high cost. It is possible to speed up the proposed method with approximation techniques for matrix inverse or with more efficient sparse coding technique. This may be considered as a further line of research.

4. Conclusions

This paper presents an image denoising method that can be applied to both additive and multiplicative noise. The proposed method is designed to capture global structures and preserve local similarities simultaneously. This method produces promising results in terms of PSNR, SSIM and visual quality. The advantages of this novel method include the following: (1) for additive noise, our method outperforms or shows comparable results to TV , HS 1 and BM3D methods either in terms of SSIM or PSNR; (2) for multiplicative noise, our method has performance superior to AA and MIDAL algorithms either in SSIM or PSNR; and (3) our method captures structures and keeps fine details well.
There are several future research directions. We are further exploring other optimization strategies for more effective convergence and further improvement. We are also considering transformation based method. Transformation and learning based model might potentially lead to more promising results.

Author Contributions

S.C., M.X. and C.P. conceived and designed the experiments; C.P., M.Y. and Z.K. performed the experiments; M.Y. and C.P. analyzed the data; X.X. contributed reagents/materials/analysis tools; S.C., M.Y. and C.P. wrote the paper.

Acknowledgments

This work was supported by the National Natural Science Foundation of China (61201392), and the Science and Technology Planning Project of Guangdong Province, China, (No. 2017B090909004).

Conflicts of Interest

The authors declare no conflict of interest. The founding sponsors had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, and in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
MDPIMultidisciplinary Digital Publishing Institute
DOAJDirectory of open access journals
TLAThree Letter Acronym
LDLinear Dichroism

Appendix A

Proof. 
Because G ¯ ( X , S ) is convex, in the tth outer and kth inner iteration, G ¯ S | X t ( k ) , X 2 t and G ¯ X | S t ( k ) , X 2 t are both convex. By the gradient descent method,
G ¯ S t ( k + 1 ) , X t ( k ) | X 2 t < G ¯ S , X t ( k ) | X 2 t
and
G ¯ S t ( k + 1 ) , X t ( k + 1 ) | X 2 t < G ¯ S t ( k + 1 ) , X | X 2 t .
Regardless of X 2 t , we may get the inequality sequence:
G ¯ S t ( k + 1 ) , X t ( k + 1 ) < G ¯ S t ( k + 1 ) , X t ( k ) < G ¯ S t ( k ) , X t ( k ) .
Because G ¯ S t ( k ) , X t ( k ) is a positive sequence, G ¯ S t ( k ) , X t ( k ) converges. ☐

References

  1. Jabarullah, B.M.; Saxena, S.; Babu, D.K. Survey on Noise Removal in Digital Images. IOSR J. Comput. Eng. 2012, 6, 45–51. [Google Scholar] [CrossRef]
  2. Chouzenoux, E.; Jezierska, A.; Pesquet, J.C.; Talbot, H. A Convex Approach for Image Restoration with Exact Poisson-Gaussian Likelihood. SIAM J. Imaging Sci. 2015, 8, 2662–2682. [Google Scholar] [CrossRef]
  3. Peng, C.; Cheng, J.; Cheng, Q. A Supervised Learning Model for High-Dimensional and Large-Scale Data. ACM Trans. Intell. Syst. Technol. 2017, 8, 30. [Google Scholar] [CrossRef]
  4. Peng, C.; Kang, Z.; Cheng, Q. Subspace Clustering via Variance Regularized Ridge Regression. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 682–691. [Google Scholar]
  5. Kang, Z.; Peng, C.; Cheng, Q. Kernel-driven similarity learning. Neurocomputing 2017, 267, 210–219. [Google Scholar] [CrossRef]
  6. Pham, D.L.; Xu, C.; Prince, J.L. Current methods in medical image segmentation. Annu. Rev. Biomed. Eng. 2000, 2, 315–337. [Google Scholar] [CrossRef] [PubMed]
  7. Buades, A.; Coll, B.; Morel, J.M. A Non-Local Algorithm for Image Denoising. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; pp. 60–65. [Google Scholar]
  8. Chatterjee, P.; Milanfar, P. Is denoising dead? IEEE Trans. Image Process. 2010, 19, 895–911. [Google Scholar] [CrossRef] [PubMed]
  9. Rudin, L.I.; Osher, S.; Fatemi, E. Nonlinear total variation based noise removal algorithms. Phys. D Nonlinear Phenom. 1992, 60, 259–268. [Google Scholar] [CrossRef]
  10. Dabov, K.; Foi, A.; Katkovnik, V.; Egiazarian, K. Image Denoising by Sparse 3-D Transform-Domain Collaborative Filtering. IEEE Trans. Image Process. 2007, 16, 2080–2095. [Google Scholar] [CrossRef] [PubMed]
  11. Dong, W.; Shi, G.; Li, X. Nonlocal Image Restoration With Bilateral Variance Estimation: A Low-Rank Approach. IEEE Trans. Image Process. 2013, 22, 700–711. [Google Scholar] [CrossRef] [PubMed]
  12. Zuo, W.; Zhang, L.; Song, C.; Zhang, D. Texture Enhanced Image Denoising via Gradient Histogram Preservation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 1203–1210. [Google Scholar]
  13. Fergus, R.; Singh, B.; Hertzmann, A.; Roweis, S.T.; Freeman, W.T. Removing camera shake from a single photograph. ACM Trans. Graph. 2006, 25, 787–794. [Google Scholar] [CrossRef]
  14. Han, Y.; Xu, C.; Baciu, G.; Li, M.; Islam, M.R. Cartoon and texture decomposition-based color transfer for fabric images. IEEE Trans. Multimed. 2017, 19, 80–92. [Google Scholar] [CrossRef]
  15. Lefkimmiatis, S.; Ward, J.P.; Unser, M. Hessian Schatten-Norm Regularization for Linear Inverse Problems. IEEE Trans. Image Process. 2013, 22, 1873–1888. [Google Scholar] [CrossRef] [PubMed]
  16. Harmeling, S. Image Denoising: Can Plain Neural Networks Compete with BM3D? In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Providence, RI, USA, 16–21 June 2012; pp. 2392–2399. [Google Scholar]
  17. Zoran, D.; Weiss, Y. From Learning Models of Natural Image Patches to Whole Image Restoration. In Proceedings of the IEEE International Conference on Computer Vision, Barcelona, Spain, 6–13 November 2011; pp. 479–486. [Google Scholar]
  18. Roth, S.; Black, M.J. Fields of Experts: A Framework for Learning Image Priors. Int. J. Comput. Vis. 2009, 82, 205. [Google Scholar] [CrossRef]
  19. Elad, M. Sparse and Redundant Representations: From Theory to Applications in Signal and Image Processing; Springer: Berlin/Heidelberg, Germany, 2010. [Google Scholar]
  20. Elad, M.; Aharon, M. Image denoising via sparse and redundant representations over learned dictionaries. IEEE Trans. Image Process. 2006, 15, 3736–3745. [Google Scholar] [CrossRef] [PubMed]
  21. Mairal, J.; Bach, F.; Ponce, J. Sparse modeling for image and vision processing. Found. Trends Comput. Graph. Vis. 2014, 8, 85–283. [Google Scholar] [CrossRef] [Green Version]
  22. Cai, S.; Weng, S.; Luo, B.; Hu, D.; Yu, S.; Xu, S. A Dictionary-Learning Algorithm based on Method of Optimal Directions and Approximate K-SVD. In Proceedings of the 35th Chinese Control Conference (CCC), Chengdu, China, 27–29 July 2016; pp. 6957–6961. [Google Scholar]
  23. Lin, Z.; Liu, R.; Su, Z. Linearized alternating direction method with adaptive penalty for low-rank representation. In Proceedings of the Advances in Neural Information Processing Systems, Granada, Spain, 12–14 December 2011; pp. 612–620. [Google Scholar]
  24. Yu, J.; Gao, X.; Tao, D.; Li, X.; Zhang, K. A unified learning framework for single image super-resolution. IEEE Trans. Neural Netw. Learn. Syst. 2014, 25, 780–792. [Google Scholar] [PubMed]
  25. Chen, S.S.; Donoho, D.L.; Saunders, M.A. Atomic Decomposition by Basis Pursuit; Society for Industrial and Applied Mathematics: Philadelphia, PA, USA, 1998. [Google Scholar]
  26. Cotter, S.F.; Rao, B.D. Sparse channel estimation via matching pursuit with application to equalization. IEEE Trans. Wirel. Commun. 2002, 50, 374–377. [Google Scholar] [CrossRef]
  27. Tropp, J.A.; Gilbert, A.C. Signal Recovery from Random Measurements via Orthogonal Matching Pursuit. IEEE Trans. Inf. Theory 2007, 53, 4655–4666. [Google Scholar] [CrossRef]
  28. Gorodnitsky, I.F.; Rao, B.D. Sparse signal reconstruction from limited data using FOCUSS: A re-weighted minimum norm algorithm. IEEE Trans. Signal Process. 2002, 45, 600–616. [Google Scholar] [CrossRef]
  29. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  30. Golub, G.H.; Van Loan, C.F. Matrix Computations; JHU Press: Baltimore, MD, USA, 2012; Volume 3. [Google Scholar]
  31. Rubinstein, R.; Zibulevsky, M.; Elad, M. Efficient Implementation of the K-SVD Algorithm using Batch Orthogonal Matching Pursuit. Cs Tech. 2008, 40, 1–15. [Google Scholar]
  32. Smith, L.N.; Elad, M. Improving dictionary learning: Multiple dictionary updates and coefficient reuse. IEEE Signal Process. Lett. 2013, 20, 79–82. [Google Scholar] [CrossRef]
  33. Lefkimmiatis, S.; Unser, M. Poisson image reconstruction with Hessian Schatten-norm regularization. IEEE Trans. Image Process. 2013, 22, 4314–4327. [Google Scholar] [CrossRef] [PubMed]
  34. Combettes, P.L.; Pesquet, J.C. Image restoration subject to a total variation constraint. IEEE Trans. Image Process. 2004, 13, 1213–1222. [Google Scholar] [CrossRef] [PubMed]
  35. Bioucas-Dias, J.M.; Figueiredo, M.A. Multiplicative noise removal using variable splitting and constrained optimization. IEEE Trans. Image Process. 2010, 19, 1720–1730. [Google Scholar] [CrossRef] [PubMed]
  36. Aubert, G.; Aujol, J.F. A variational approach to removing multiplicative noise. SIAM J. Appl. Math. 2008, 68, 925–946. [Google Scholar] [CrossRef]
Figure 1. Results on the abdomeng image degraded by Gaussian noise of level = 0.1.
Figure 1. Results on the abdomeng image degraded by Gaussian noise of level = 0.1.
Symmetry 10 00167 g001
Figure 2. Results on the kids image degraded by Gaussian noise of level = 0.07.
Figure 2. Results on the kids image degraded by Gaussian noise of level = 0.07.
Symmetry 10 00167 g002
Figure 3. Results on the face image degraded by Gaussian noise of level = 0.05.
Figure 3. Results on the face image degraded by Gaussian noise of level = 0.05.
Symmetry 10 00167 g003
Figure 4. Results on the wall image degraded by Gaussian noise of level = 0.1.
Figure 4. Results on the wall image degraded by Gaussian noise of level = 0.1.
Symmetry 10 00167 g004
Figure 5. Results on the Fields image degraded by multiplicative noise of level L = 1.
Figure 5. Results on the Fields image degraded by multiplicative noise of level L = 1.
Symmetry 10 00167 g005
Figure 6. Results on the Nimes image degraded by multiplicative noise of level L = 4.
Figure 6. Results on the Nimes image degraded by multiplicative noise of level L = 4.
Symmetry 10 00167 g006
Figure 7. Example of denoising performance changes with respect to parameters. From left to right are results on images of Face with Gaussian noise of level 0.05 and Kids with Gaussian noise of level 0.04.
Figure 7. Example of denoising performance changes with respect to parameters. From left to right are results on images of Face with Gaussian noise of level 0.05 and Kids with Gaussian noise of level 0.04.
Symmetry 10 00167 g007
Figure 8. (a) Face image degraded by Gaussian noise of level 0.05; (b) Kids image degraded by Gaussian noise of level 0.04. Example of time cost on two stages using different combinations of parameters.
Figure 8. (a) Face image degraded by Gaussian noise of level 0.05; (b) Kids image degraded by Gaussian noise of level 0.04. Example of time cost on two stages using different combinations of parameters.
Symmetry 10 00167 g008
Table 1. Empirical value of parameters.
Table 1. Empirical value of parameters.
ParameterSymbolEmpirical Value for p = 1 Empirical Value for p = 2
Penalty parameter λ 1 1010
Penalty parameter λ 2 0.10.1
Penalty parameter μ σ / 30 σ / 30
Smoothing parameter δ 0.120
Stopping tolerence ϵ 1 0.0010.001
Stopping tolerence ϵ 2 0.0010.001
Table 2. Comparison of different methods, in PSNR and SSIM, for additive noise with different noise levels.
Table 2. Comparison of different methods, in PSNR and SSIM, for additive noise with different noise levels.
ImageFaceKids
σ = 0.05 σ = 0.07 σ = 0.10 σ = 0.04 σ = 0.07 σ = 0.10
MethodTVPSNR21.4720.4419.0223.0621.0319.49
SSIM0.69150.64350.61310.67900.59320.5667
HS 1 PSNR22.1320.9219.4224.0321.6019.96
SSIM0.74170.69920.66160.74090.67060.6333
EPLLPSNR22.0220.8519.3024.0221.6219.39
SSIM0.73200.70300.66360.75310.68170.6366
BM3DPSNR22.8020.7620.0524.5222.0820.40
SSIM0.75360.66790.67650.76030.68820.6458
LSLA-2PSNR23.2522.5020.9524.6923.0323.68
SSIM0.76790.73960.68510.75550.70520.6578
LSLA-1PSNR23.4822.0521.1924.5923.2922.45
SSIM0.76940.72170.69120.74230.70630.6825
ImageWallAbdomen
σ = 0.05 σ = 0.07 σ = 0.10 σ = 0.04 σ = 0.07 σ = 0.10
MethodTVPSNR20.7018.1916.8022.5720.0618.50
SSIM0.65210.56010.49780.55790.49400.4697
HS 1 PSNR21.3318.5417.0323.2920.5218.77
SSIM0.70430.59750.54600.63840.55920.5300
EPLLPSNR21.3618.3816.7623.5120.6418.84
SSIM0.72540.62540.56980.65170.59150.5440
BM3DPSNR21.9719.0417.4224.1421.2619.50
SSIM0.74210.64100.58380.67000.60260.5603
LSLA-2PSNR22.2820.1119.2225.0622.6821.47
SSIM0.75980.67300.64770.75300.66800.6237
LSLA-1PSNR22.5120.3119.1224.9722.7221.37
SSIM0.76750.67360.63110.74620.66630.6096
Table 3. Comparison of different methods for multiplicative noise with different noise levels.
Table 3. Comparison of different methods for multiplicative noise with different noise levels.
ImageNoise LevelMethod
AAMIDALLSLA-2
PSNRSSIMPSNRSSIMPSNRSSIM
Nimes L = 1 22.400.537822.680.504123.640.5942
L = 4 25.590.757225.360.753726.500.7757
L = 10 27.530.851127.880.891028.510.8625
Fields L = 1 24.380.336925.130.338025.270.3505
L = 4 26.430.423027.400.402427.460.4622
L = 10 26.770.446428.270.537128.640.5421
Table 4. Time comparison on Gaussian noise removal.
Table 4. Time comparison on Gaussian noise removal.
AlgorithmTime (s)
FaceKidsWallAbdomen
BM3D1.02841.03361.10083.7333
HS116.945418.084217.532437.0296
EPLL146.244378.3126146.561502.4728
TV0.66960.68410.65381.3947
LSLA2124.036185.6624122.9288404.6401
LSLA1169.4185139.1726178.3549438.5715

Share and Cite

MDPI and ACS Style

Cai, S.; Kang, Z.; Yang, M.; Xiong, X.; Peng, C.; Xiao, M. Image Denoising via Improved Dictionary Learning with Global Structure and Local Similarity Preservations. Symmetry 2018, 10, 167. https://doi.org/10.3390/sym10050167

AMA Style

Cai S, Kang Z, Yang M, Xiong X, Peng C, Xiao M. Image Denoising via Improved Dictionary Learning with Global Structure and Local Similarity Preservations. Symmetry. 2018; 10(5):167. https://doi.org/10.3390/sym10050167

Chicago/Turabian Style

Cai, Shuting, Zhao Kang, Ming Yang, Xiaoming Xiong, Chong Peng, and Mingqing Xiao. 2018. "Image Denoising via Improved Dictionary Learning with Global Structure and Local Similarity Preservations" Symmetry 10, no. 5: 167. https://doi.org/10.3390/sym10050167

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop