This section provides a detailed introduction of OGS-FOTVR model. Additionally, it discusses the numerical solution of the model by combining it with the MM algorithm within the framework of the ADMM algorithm.
3.1. Model Formulation
The texture and intricate details within an image often exhibit non-local similarities. However, the domain of first-order gradient variation only encompasses the local attributes of the pixels, thereby rendering the OGS-TV model ineffective in accurately reconstructing the sophisticated texture information within the image. To compensate for these limitations, it is essential to explore a gradient domain that is better suited for retaining the texture information in the image. As illustrated in
Figure 1, the FO differential at a given point exhibits non-local properties, meaning that the differential of this point is collectively influenced by the information from multiple surrounding points. Thus, in contrast to the conventional differential form, the FO differential is more congruent with the real-world scenarios when it comes to representing image texture information. Furthermore, the FO differential has a relatively subdued amplification scale for high-frequency information. This attribute is beneficial as it effectively precludes the misidentification of edge information as noise, thereby preventing its removal, and concurrently conserves the edges and contours of the image.
In light of the aforementioned analysis, the OGS-FOTVR regularization term has been devised to incorporate the rich texture and complex details inherent in the image, thereby augmenting the quality of the reconstructed image. The structure of the proposed OGS-FOTVR is presented below:
where
u and
f denote the restored and observed images, respectively.
operates as the regularization parameter, while FO differentiation is denoted by
. The function
is the overlapping group function as defined by Equation (
2). The term
signifies the data fidelity term, which quantifies the similarity between the restored image
u and the observed image
f. The component
represents the overlapping group sparse FO variational regularization term and serves to characterize the prior information of the restored image. It is crucial to highlight that when
, certain low-frequency components escalate non-linearly, while specific high-frequency components diminish non-linearly. Under these circumstances, the frequencies of textures and noises converge, exacerbating the challenge of differentiating between texture and noise. This can lead to the concurrent elimination of texture and noise information. Conversely, when
, some low-frequency components decline non-linearly, while certain high-frequency components increase in a non-linear fashion. In this case, the frequency difference between textures and noises is amplified, making it easier to distinguish between them. Although larger values of
lead to better preservation of image texture information, they also tend to misclassify some edge and contour information in the image as noise, resulting in blurring and reduced visual quality of the restored image. Therefore, in this paper, we set
.
Furthermore, it can be observed that is the overlapping group sparse regularizer defined in the fractional gradient transform domain. Firstly, due to the non-local characteristics of the fractional gradient, the OGS-FOTVR model can effectively suppress the staircase effect and restore more complex texture information in the image. Secondly, by utilizing the overlapping group to measure the sparsity of the fractional gradient variation domain, the model can preserve edge information in the image. Therefore, the OGS-FOTVR model not only mitigates the staircase effect but also achieves a balance between texture and edge restoration.
3.2. Numerical Algorithm
Since the OGS-FOTVR model is a large-scale variational minimization problem, the ADMM algorithm is used to decompose it into several sub-problems for solving. To achieve variable separation, an auxiliary variable
v is introduced for variable substitution, transforming the original unconstrained problem (8) into the following constrained problem:
The Equation (
10) is characterized as non-convex, primarily due to the inclusion of the function
. In this context,
signifies a higher-order gradient of the variable
u. The function
is typically a non-linear operation applied to this gradient, such as an absolute value or a square, which is employed to enhance certain attributes in the solution, such as sparsity or smoothness. The initial term,
, is a convex function, as it is a squared Euclidean norm (which is inherently convex) scaled by a positive constant,
. However, the overall problem is rendered non-convex due to the incorporation of the non-convex term,
. Non-convex optimization problems generally pose a greater challenge than their convex counterparts [
33], as they can exhibit multiple local minima, and the solution can become ensnared in these local minima, thereby failing to locate the global minimum. Despite this, numerous practical problems are non-convex, and a variety of strategies have been developed to address them [
6,
34]. Furthermore, by utilizing the augmented Lagrangian method, the aforementioned constrained problem is converted into an unconstrained problem, specifically constructing an augmented Lagrangian function
:
where
is the Lagrange multiplier and
is the parameter for the quadratic penalty term. Furthermore, the augmented Lagrangian function
is obtained using the augmented Lagrangian method, which converts the aforementioned constrained problem into an unconstrained problem. The goal is to find the saddle point of
using the alternating direction iterative algorithm, which involves iteratively solving for the minimum values of
u and
v while maximizing the value of
. By initializing the parameters appropriately, the original problem can be transformed into separate coupled sub-problems for solution.
Up to this point, the solution to model (
11) is decomposed into a series of sub-problems via ADMM algorithm. Subsequent sections will individually tackle the solutions for these distinct sub-problems.
(1) Sub-problem for u: firstly, the sub-problem for
u can be written as follows:
By applying the variational Euler-Lagrange equation, we can determine that the optimal solution
u for sub-problem (
11) must satisfy the following necessary condition:
where
I represents a matrix that is the same size as
u, and every element at each position within the matrix is set to 1. It is evident that Equation (
12) incorporates a 2D convolution operation, which implies that the optimal solution
u cannot be directly extracted from it. This convolution operation is within the spatial domain that can be converted into a direct multiplication operation when this process is executed within the frequency domain. This transformation simplifies the process significantly. Consequently, the numerical solution to the aforementioned equation can be readily obtained by employing the Fourier transform and its corresponding inverse transform as follows:
where
and
represent the operators for fast Fourier transform (FFT) and fast inverse Fourier transform (IFFT), respectively. In (
14),
is discrete gradient operator with periodic boundary conditions,
and
are the x-derivative and
y-derivative located at the
i-th pixel
.
present first order difference operator, it can be stated as;
. For a comprehensive understanding of image reconstruction, specifically in the context of either integer-order or FO in the Fourier domain, we direct the reader’s attention to references [
34,
35,
36].
To analyze the computational cost of the Equation (
14), we can break down the different operations involved and estimate their complexity. The main operations in (
14) are the FFT and the IFFT. We also consider the cost of the other arithmetic operations present in (
14). At each iteration, the complexity of FFT and IFFT for an n-point transform is
, respectively [
37,
38]. The complexity of applying the gradient operator depends on the specific implementation and the size of the grid points but is usually
in the context of fast algorithms. Now, consider the full equation and analyze its complexity as follows:
Applying :
Applying the discrete gradient operator and its adjoint: or
Subtracting :
Adding :
Dividing by :
Adding and :
Constructing the denominator :
Applying and for the numerator and denominator:
Since the numerator and denominator each involve FFT and IFFT, the overall complexity for computing is + or . In general, the dominant term in the computational cost will be the FFT and IFFT operations , especially for large values of n. The other operations are typically linear and do not significantly affect the overall complexity when compared to the FFT and IFFT operations.
(2) Sub-problem for v: it is evident that the sub-problem for
v is a problem of overlapping group sparse regularization, we have
To facilitate subsequent writing and calculations, it is referred to as:
Due to the complex structural form of this type of problem, the MM algorithm [
39] is adopted to solve the above minimization problem. According to the mean value inequality, we have
Based on this, the function
at point
m can be written as:
According to the inequality (16), it can be proven that
and
hold. For the convenience of computation, after simplification,
can be written in the following form:
where
C is a constant that does not depend on
v,
is a diagonal matrix with its diagonal elements as follows:
where
,
can be computed using two-dimensional convolution. So far, one alternative function
of
in Equation (
14) can be expressed as follows:
Upon observing the above equation, it is not difficult to notice that the following expressions hold:
and
. This means that the function
satisfies the prerequisite conditions of the MM algorithm. Therefore, in order to minimize
according to the MM algorithm, we initialize
and then repeatedly minimize the alternative function
, i.e.,
where
. According to the Euler-Lagrange equation, the numerical solution of sub-problem for
v can be obtained as:
(3) Updating Lagrange Multiplier : Finally, according to the ADMM algorithm, the updating form of the Lagrange multiplier is given by:
Therefore, through the aforementioned discussion, the proposed OGS-FOTVR model (8) has been solved using numerical algorithms. The specific solution process is shown in Algorithm 2.
Algorithm 2 The solution of the proposed OGS-FOTVR is presented in a step-by-step manner |
Input: the observed image f, and set parameters , and the overlapping group size K.
- 1:
Step (1) initialize , and , and determine penalty parameter and maximum iteration count N; - 2:
while k = 0 to number of iterations to do - 3:
Step (2) Solve for using Equation ( 14). - 4:
Step (3) Solve for by combining Equations (21) and (22). - 5:
Step (4) Update the Lagrange multiplier using Equation ( 23). - 6:
Step (5) If , update and return to Step 2, otherwise, output the result. - 7:
return
|
3.3. Convergence Analysis
The convergence of the proposed method, which employs the ADMM, is a crucial aspect to consider. The ADMM algorithm is known for its robustness and efficiency in solving large-scale optimization problems, particularly those involving separable structures, as in our case. The convergence of the ADMM algorithm has been extensively studied and proven in the literature [
8,
30,
40,
41]. Moreover, the convergence of the ADMM algorithm in the context of non-convex problems has been studied by Hong et al. [
42]. They showed that if the objective function is proper, lower semi-continuous, and the set of optimal solutions is non-empty, then the sequence generated by the ADMM algorithm converges to an optimal solution. In contrast to convex optimization problems where convergence is typically demonstrable, non-convex optimization problems pose a considerably greater challenge due to the necessity of taking various assumptions into account. With regard to the OGS-FOTVR approach, the primary problem is subdivided into several sub-problems, each constituting an optimization problem in its own. These sub-problems are iterative methods, which can be analyzed individually, thus enabling us to provide insights on the convergence characteristics of the OGS-FOTVR. The iterative process involves finding the minimum values of
u and
v while maximizing the value of
, which is the Lagrange multiplier. This process continues until the algorithm converges to a solution that satisfies the constraints of the problem. The convergence of each sub-problem is also guaranteed. For the
u sub-problem, the Euler-Lagrange equation is used to find the optimal solution, which is then obtained numerically using the Fourier transform and its inverse [
34,
36]. For the
v sub-problem, the MM algorithm is employed. The MM algorithm is a well-established method for solving optimization problems, and its convergence has been proven under mild conditions [
39]. Finally, the Lagrange multiplier
is updated according to the ADMM algorithm, which ensures the convergence of the overall method. The updating of
continues until the difference between
and
is sufficiently small, indicating that the solution has converged. The relative error is employed as the stopping criterion. Specifically, the algorithm is designed to stop via this equation, i.e.,
. Where
and
indicate the restored images at the current iteration and the previous iteration, respectively. In order to illustrate the convergence analysis, we have graphically represented it in terms of relative error, PSNR and SSIM values in relation to the optimization iterations for OGS-FOTVR, as shown in
Figure 2. It is evident that as the number of iterations increases, the relative error demonstrates a decreasing trend and ultimately converges. Moreover, with each iteration of Algorithm 2, there is a noticeable increase in both the PSNR and SSIM values, which eventually reach a point of convergence. This observed pattern provides a clear affirmation of the convergent nature of the proposed OGS-FOTVR, demonstrating its effectiveness and reliability in achieving a stable solution.
In short, the convergence of our proposed method is guaranteed by the properties of the ADMM algorithm and the MM algorithm, as well as the specific structures of the sub-problems. This ensures that our method can effectively and reliably solve the large-scale variational minimization problem posed by the OGS-FOTVR model.