Next Article in Journal
Enhanced Algorithm for the Detection of Preimpact Fall for Wearable Airbags
Previous Article in Journal
A Non-Touchscreen Tactile Wearable Interface as an Alternative to Touchscreen-Based Wearable Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Single Remote Sensing Image Super-Resolution with an Adaptive Joint Constraint Model

1
College of Electronics and Information Engineering, Sichuan University, Chengdu 610065, China
2
Key Laboratory of Wireless Power Transmission of Ministry of Education, College of Electronics and Information Engineering, Sichuan University, Chengdu 610065, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(5), 1276; https://doi.org/10.3390/s20051276
Submission received: 10 February 2020 / Revised: 23 February 2020 / Accepted: 25 February 2020 / Published: 26 February 2020
(This article belongs to the Section Remote Sensors)

Abstract

:
Remote sensing images have been widely used in many applications. However, the resolution of the obtained remote sensing images may not meet the increasing demands for some applications. In general, the sparse representation-based super-resolution (SR) method is one of the most popular methods to solve this issue. However, traditional sparse representation SR methods do not fully exploit the complementary constraints of images. Therefore, they cannot accurately reconstruct the unknown HR images. To address this issue, we propose a novel adaptive joint constraint (AJC) based on sparse representation for the single remote sensing image SR. First, we construct a nonlocal constraint by using the nonlocal self-similarity. Second, we propose a local structure filter according to the local gradient of the image and then construct a local constraint. Next, the nonlocal and local constraints are introduced into the sparse representation-based SR framework. Finally, the parameters of the joint constraint model are selected adaptively according to the level of image noise. We utilize the alternate iteration algorithm to tackle the minimization problem in AJC. Experimental results show that the proposed method achieves good SR performance in preserving image details and significantly improves the objective evaluation indices.

1. Introduction

With the constant development of the remote sensing technology in recent years, many remote sensing image applications have been proposed, such as fine-grained classification [1,2], detailed land monitoring [3], and target recognition [4,5]. The performance of these applications is closely related to the image quality. However, the resolution of the remote sensing image is largely affected by the spatial resolution of the optical sensor. To improve the quality of remote sensing images, many super-resolution (SR) methods [6,7,8,9,10] have been proposed. They can be divided into multi-frame [11,12,13,14,15,16,17] and single-image SR methods [18,19,20,21,22,23].
The multi-frame SR methods aim to recover a high-resolution (HR) image from multiple low-resolution (LR) frames. A multi-frame SR method based on frequency domain was proposed in [13]. Though this method can be implemented fast, it leads to serious visual artifacts [12]. To address the above issue, many spatial domain-based multi-frame SR methods have been proposed [15,16]. Farsiu et al. [15] proposed an iterative method using l 1 -norm in the fidelity and regularization terms. Patanavijitt and Jitapunkult [16] proposed a stochastic regularization term according to the Bayesian maximum a posteriori (MAP) estimation. Huang et al. [14] proposed a bidirectional recurrent convolutional network for multi-frame images SR. Although the spatial domain-based multi-frame SR methods can achieve good SR performance, they either consume quite a lot of memory or take a large amount of running time [15]. Moreover, in some cases, only one image is available instead of multi-images, which is more challenging. Therefore, single-image SR methods are significant in practical applications [23].
Generally, the problem of single-image SR can be formulated as:
y = D B x + n ,
where y is the observed LR image, x denotes the original HR image, B represents the blurring operator, D denotes the downsampling operator, and n is the additive Gaussian white noise with standard deviation of σ n . Single-image SR methods can be roughly divided into interpolation-based prior-based methods. Interpolation-based methods [24,25] exploit adjacent pixels to estimate the unknown pixel. The interpolation-based methods have low computational complexity and can recover degraded images fast, but they may generate severe staircase artifacts and smooth out image details.
Many prior-based methods have been used to address the single-image SR problem. Tai et al. [26] exploited edge prior and single-image detail synthesis to enhance the resolution of the LR image. Some researchers have proposed example learning-based methods by exploiting the priors according to the similarity of the image. Yang et al. [22] proposed a joint sparse dictionary (HR and LR dictionaries) to reconstruct the HR image. This method assumes that the sparse vector of the HR patch is the same as the corresponding LR patch. Timofte et al. [27] proposed a joint sparse dictionary-based anchored neighborhood regression (ANR) approach, where regressors are anchored to the dictionary atoms. A joint dictionary-based statistical prediction model that assesses the statistical relationships between the sparse coefficients of LR and HR patches was introduced by Peleg and Elad [28]. NPDDR [29] trained a joint dictionary to deal with the lack of HR component and employed improved nonlocal self-similarity and local kernel to constrain the optimization process. Hou et al. [7] proposed a global joint dictionary model for the single remote sensing image SR, including the local constraint within each patch and the nonlocal constraint between HR and LR images. A single dictionary-based adaptive sparse representation scheme was presented in [30]. Dong et al. [31] proposed a single dictionary-based method, which restores the HR image by using the nonlocal self-similarity of the sparse vector. Compared with the joint dictionary methods, the single dictionary methods do not require external examples to train. Although the above methods enhance the resolution of the LR image, they are prone to generate some artifacts. Recently, deep learning technologies have been introduced into SR. For example, Dong et al. [32] proposed a deep convolutional neural network for the SR methods. Haut et al. [6] exploited a novel convolutional generator model to improve the quality of remote sensing images. Li et al. [33] applied spatial-spectral group sparsity and spectral mixture analysis to the SR problem. Tai et al. [34] proposed a deep recursive residual network(DRRN) that contains 52 convolutional layers. A generative adversarial network (GAN) for image SR methods was introduced in [35].
However, deep learning-based methods take a great deal of time to train models by using external datasets. Furthermore, they require retraining for each different degradation, which costs more training time. In contrast, the single dictionary-based sparse representation SR methods require neither external datasets nor retraining for each degradation. However, traditional single dictionary-based sparse representation SR methods ignore some complementary constraints of the image itself. Previous works show that sparse representation-based SR methods are of great significance in improving the quality of remote sensing images [7,8]. Based on the above reasons, we focus on improving the single dictionary-based sparse representation SR methods to enhance the quality of the reconstructed HR image.
Inspired by [7], we consider introducing a joint constraint for the single dictionary-based sparse representation SR methods. In the references [7,8], authors pointed out that the sparse representation model is beneficial to improve the quality of the remote sensing image. However, this kind of method does not consider the continuity and correlation of an image’s local neighborhood, and the edge preservation ability should be further improved. To address this issue, we propose a local structure constraint based on the local gradient of the image, which can further improve the edge-preserving ability. For the sparse representation-based SR methods, it is usually assumed that the sparse coefficients corresponding to the LR image and the HR image are equal. However, due to the degradation of the LR image such as blurring and down-sampling, there is a gap between the LR sparse coefficient and the HR sparse coefficient, i.e., sparse noise. In addition, image artifacts tend to appear in the reconstructed images. Since the sparse noise and image artifacts can be well suppressed by exploiting these overlapping nonlocal similar patches, as shown in [7,31], we use the nonlocal self-similarity of the image to construct a nonlocal constraint. By using the nonlocal constraint and the proposed local structure constraint, we further propose a novel adaptive joint constraint (AJC) for the single remote sensing image SR. In the reconstruction phase, according to the level of image noise, we adaptively select regularized parameters for the proposed AJC model to obtain better reconstructed images. Our main contributions are summarized as follows:
(1) Since human visual system is more sensitive to image edges, we propose a local structure filter based on the local gradient of the remote sensing image and then construct a local structural prior.
(2) The fusion of the complementary local and nonlocal priors can achieve higher SR performance. Based on this, we combine the nonlocal self-simialrity prior and the local structural prior and then propose a novel adaptive joint constraint (AJC) for the single remote sensing image SR.
(3) To further improve the proposed AJC SR method, the regularization parameters are selected adaptively according to the level of remote sensing image noise.
The organization of this paper is as follows. Section 1 is the introduction. Section 2 introduces the related work about the sparse coding-based single image SR. In Section 3, the proposed AJC algorithm is described in detail. Section 4 introduces the basic parameters setting and shows experiment results. Section 5 summarizes our work.

2. Related Work

The single-image SR reconstruction methods aim to transform the LR image into the HR image. To achieve this goal, sparse priors have been proposed [22,27,30,31,36,37]. Researchers have found that the image can be sparsely represented as x Θ α , where α is a sparse vector composed of coefficients, and Θ denotes an over-complete dictionary, such as a discrete cosine transform (DCT) or a wavelet dictionary [30,38]. The sparse vector α can be evaluated via the following formula:
α ^ = arg min α { x - Θ α 2 2 + ι α 0 } ,
where ι represents a regularization parameter, x is a HR image, Θ denotes the dictionary of x , and α represents the sparse vector of x . Since l 0 -minimization is an NP-hard problem, in general, l 1 -minimization is used to approximate the l 0 -minimization. In the single-image SR problem, we want to restore the missing details of the LR image y to enhance the spatial resolution. For this problem, we can address it by the following equation:
α ^ y = arg min α { y - D B Θ α 2 2 + λ ˜ α 1 } ,
where α ^ y denotes the estimated sparse vector of the LR image y , and λ ˜ is a regularization parameter of the sparse term. Since the sparse vector of the HR patch is the same as the corresponding LR patch [22], we can get the reconstructed HR image by x ^ Θ α ^ y .

3. Proposed Method

To remove artifacts and improve the quality of reconstructed HR images, we propose the AJC model that includes the nonlocal and local priors. The proposed framework is shown in Figure 1, the nonlocal prior is constructed by exploiting the nonlocal self-similarity of the sparse vector, and the local prior can be obtained according to the proposed local structure filter. The proposed AJC model can be formulated as:
α ^ y = arg min α y - DB Θ α 2 2 + η J N L α + γ J L x ,
where the first term is the data term, J N L α denotes the nonlocal constraint, J L x represents the local prior, η is a parameter to balance the data term and the nonlocal constraint, and γ is a regularization parameter used to keep balance between the data term and the local structural constraint. For the subsequent reconstruction process, we first need to use the self-learning method to train the compact dictionary. Then, the nonlocal and local priors are constructed. Finally, we use the iterative shrinkage algorithm [39] to solve the SR reconstruction problem.

3.1. The Compact Dictionary Learning

In sparse representation, it is important to select a suitable dictionary for the image. In this paper, we learn a compact dictionary for the image according to [31]. It is worth noting that our dictionary learning uses a LR image as the input of training. The compact dictionary learning first needs to extract the HR image patches X ( For the first reconstruction, the HR image was reconstructed from the corresponding LR image by Bicubic.). Considering that human vision is sensitive to edges, we remove extremely smooth image patches in X and express them in X h . Furthermore, to further supplement the high frequency details of the images, we learn the dictionary in the high-frequency feature space of patches. Specifically, we adopt a rotationally symmetric Laplacian of Gaussian filter (size 5 × 5 with standard deviation 0 . 5 ), which can improve the average PSNR of the test images by 0.11dB in SR application. The high-frequency feature of X h is denoted by C = [ C 1 , C 2 , , C N ] , where N represents the number of patches. Then, we adopt the K-means algorithm to divide C into K clusters ( H = [ H 1 , H 2 , , H K ] ) . Next, the compact dictionary Θ can be learned from H . Specially, since the PCA approach can reduce dimension and realize de-correlation, we apply the PCA algorithm to each cluster H k ( k = 1 , 2 , , K ) . After PCA, the sub-dictionary Θ k corresponding to each cluster H k can be obtained according to [31]. Finally, the compact dictionary Θ corresponding to the reconstructed image x ^ can be constructed by Θ = [ Θ 1 , Θ 2 , , Θ K ] . After learning the dictionary Θ , we apply the dictionary to the proposed AJC model for obtaining the reconstructed HR image x ^ . It should be noted that the estimated x ^ will be used to update the dictionary Θ and K will be updated in the dictionary learning phase.

3.2. The Joint Priors

3.2.1. The Nonlocal Self-Similarity Prior

The remote sensing image has a large number of repetitive structures. That means, for a given image patch, we can find its similar patches in other parts of the image. We illustrate the nonlocal self-similarity of the remote sensing image by using the “airplane” image in Figure 1. The red patch is a target patch, its nonlocal similar patches can be found in the image, such as the blue, green, black patches. Thus, we can use the nonlocal self-similarity of the remote sensing image to guide the image SR reconstruction. According to [31], the sparse vector also has the nonlocal self-similarity. Thus, we can exploit the nonlocal self-similarity of the sparse vector to guide the image SR reconstruction, where the nonlocal prior can be formulated as:
J N L α = α - β 1 ,
where α = [ α 1 ,   ,   α i ,   ,   α N ] , β = [ β 1 ,   ,   β i ,   ,   β N ] , and β i is the estimated α i by using the nonlocal self-similarity of α i .
In view of the above, the closer the patches are, the greater the similarity weight will be. This can be achieved by gaussian weighting. Thus, we can estimate β by the nonlocal gaussian weighting of the nonlocal similar sparse vector α . For each patch x i , β i can be computed by the following equation:
β i = x i s C o exp ( - x i - x i s 2 2 / ϵ 2 ) x i s C o exp ( - x i - x i s 2 2 / ϵ 2 ) α i s ,
where C o is the collection of similar patches of x i , x i s denotes the s-th similar patch of x i , ϵ 2 is a constant that controls the degree of decay, and α i s represents the s-th similar vector of α i .

3.2.2. The Local Structure Prior

However, the previous prior term can not sufficiently constrain the local structures of remote sensing images. Therefore, we propose a local prior as a complementary prior. The methods [7,22,40] exploit local features of the images to constrain the solution of SR problem. JIDM [7] and SRSC [22] employed local image patches as local features. K-SVD [40] ensured that mean square error (MSE) of each pixel neither reduction nor change in each iteration, which constrains the local structures. However, it is not sufficient for these methods to preserve edge energy while taking into account the relationship between adjacent pixels. To address the above issue, we use the local gradient of the image to solve the local structures-preserving problem in the SR reconstruction. The image gradient is sensitive to the image edge. The large gradients correspond to sharp edges of the image, while the small gradients correspond to smooth regions of the image. The objective function based on the local gradient of the image can be expressed as:
O = arg min O p = 1 Q ( Z p - O p ) 2 + ω [ a h , p ( h , p ( O ) ) 2 + a v , p ( v , p ( O ) ) 2 ] ,
where p is the pixel location in the image, Z R m × n ( m × n represents the size of the image) denotes the clean image, O R m × n represents the smoother image of Z , Q ( Q = m × n ) is the number of pixels in the image, a h , p and a v , p are weight vectors, ( h , p ( O ) ) 2 and ( v , p ( O ) ) 2 are the square of gradients along the horizontal and the vertical directions at the p-th position of the image O , respectively. In Equation (7), the first term is the data term to ensure that the image O is close to the image Z , the second term is a regularization term that makes the gradient of O as small as possible (except where the image has significant gradients) and constrains the edge energy, and the parameter ω balances the data term and the regularization term.
By using matrix notations, we can rewrite Equation (7) as:
O = arg min O ( Z - O ) T ( Z - O ) + ω ( O T S h T A h S h O + O T S v T A v S v O ) ,
where A h , A v R Q × Q ,
A h = a h , 1 0 0 0 a h , 2 0 0 0 a h , Q , A v = a v , 1 0 0 0 a v , 2 0 0 0 a v , Q .
S h and S v R Q × Q are discrete difference operations along the h and v directions, respectively, which are defined as follows:
S h c r = 1 if c = r - 1 if c = r 0 + m 0 if others , S v c r = 1 if c = r - 1 if c = r 0 + 1 0 if others ,
where c and r are the column and row numbers of S h and S v , respectively, c , r =   1 ,   2 ,   ,   Q , and r 0 is the value of r in the case of r = c . Minimizing Equation (8), we can acquire the following formula:
( I + ω M ) O = Z ,
where M = S h T A h S h + S v T A v S v , I denotes an identity matrix, and we call E = ( I + ω M ) the local structure filter. Inspired by [41], we calculate weight vectors a h , p and a v , p by:
a h , p = [ h , p ( O ) + ϵ 1 ] - 1 , a v , p = [ v , p ( O ) + ϵ 1 ] - 1 ,
where ϵ 1 is a constant for mathematical stability, and ϵ 1 = 0 . 0001 . Since O is unknown, a h , p and a v , p are estimated by exploiting the image Z . The calculation of the local structure filter E is summarized in Algorithm 1.
Algorithm 1 The calculation of the local structure filter E.
Input: Z
Output: E
1:
set parameters ϵ 1 , ω and compute a h , p , a v , p by Equation (12);
2:
calculate matrices A h ,   A v via Equation (9);
3:
calculate difference matrices S h ,   S v via Equation (10);
4:
compute the symmetrical matrix M by M = S h T A h S h + S v T A v S v ;
5:
obtain the local structure filter E by E = ( I + ω M ) ;
The proposed local structure filter E has strong local correlation because it takes into account the local gradient of the image. Moreover, remote sensing images are usually terrain and targets images, such as airplane, harbor, parking lot, river, and mountain. These images have local correlation. To describe the local correlation of remote sensing images, we construct a local prior for remote sensing images. In order to explore the local property, we perform statistical analysis on a large number of remote sensing images. Since E can describe the local continuity of the image, we analyze the statistic property of ψ = Ex - x . For the image “Building” as an example, the distribution function curves are shown in Figure 2. We can find that the probability density of ψ is close to Gaussian distribution (note that the value of each entry in ψ is normalized to [−1, 1] in this test). According to the results, we can obtain the following formula:
F ( ψ ) = 1 2 π σ h exp ( - ψ 2 2 σ h 2 ) ,
where F ( ψ ) is the the probability density of ψ , and σ h denotes a standard deviation. Then the local constraint J L x can be constructed according to ψ in the MAP estimate, and the mathematical expression is as follows:
J L x = Ex - x 2 2 .

3.3. Regularized Parameter Settings

Proper parameters are conducive to improve the performance of SR reconstruction, so we adopt the parametric adaptive method. Since the noise level is related to the regularization parameters [23], in this paper, η and γ are both selected adaptively according to the noise level. The noise level can be calculated by [42]. Let ξ = α - β , and assume that ξ and ψ are independent. The MAP estimation considering both ξ and ψ is a multi-prior estimation. Multi-prior estimation methods have been widely used in image processing [43,44]. The MAP estimation of ξ and ψ can be expressed as:
ξ , ψ = arg max ξ , ψ log F ( y | ξ ) + log F ( ξ ) + log F ( y | ψ ) + log F ( ψ ) .
Since E is a pre-calculated matrix and ψ = Ex - x , ψ is only related to x . We can obtain F ( y | ψ ) = F ( y | x ) .
F ( y | ξ ) and F ( y | x ) are generally characterized by the Gaussian distribution [43,44]:
F ( y | ξ ) = F ( y | α , β ) = 1 2 π σ n exp - y - DB Θ α 2 2 2 σ n 2 ,
F ( y | x ) = 1 2 π σ n exp - y - DBx 2 2 2 σ n 2 ,
where ξ and β are assumed to be independent according to [31]. Since Θ is pre-calculated and x = Θ α , F ( y | ξ ) = F ( y | x ) .
F ( ξ ) can be obtained according to [31,45]:
F ( ξ ) = i g 1 2 σ i , g exp - 2 ξ i , g 1 σ i , g ,
where ξ i , g is the g-th entry of ξ i , and σ i , g represents a standard deviation of the error ξ i , g corresponding to the Laplace distribution.
Substituting Equations (13), (16), (17) and (18) into Equation (15), and ignoring constant terms, we can get:
α y = arg min α y - DB Θ α 2 2 + 2 σ n 2 i g α i , g - β i , g 1 σ i , g + σ n 2 Ex - x 2 2 2 σ h 2 ,
where α i , g denotes the g-th entry of α i , and β i , g represents the g-th entry of β i .
Comparing Equation (4) with Equation (19), we can obtain parameters η and γ . Furthermore, to ensure mathematical stability, we add additional parameters ϵ η and ϵ γ in the calculation of η and γ , respectively.
η = i g 2 σ n 2 + ϵ η σ i , g , γ = ( σ n 2 + ϵ γ ) / 2 σ h 2 ,
where σ h denotes the evaluated noise level from x ^ during each iteration.
Finally, we apply the iterative shrinkage algorithm [39] to the AJC model, and the details of solution are given in Algorithm 2.
Algorithm 2 Details of solution.
1:
initialization
(a)
set the initializing parameter;
(b)
get initialized x ^ by Bicubic of the input LR image;
(c)
estimate the noise level σ n of the input LR image;
2:
outer loop: iterate on l o = 1 , 2 , , l o m a x
(a)
update the dictionary Θ ;
(b)
inner loop: iterate on l = 1 , 2 , , L ;
(1)
obtain the noise level σ h of the reconstructed HR image x ^ l ;
(2)
estimate η and γ by Equation (20);
(3)
calculate the filter E by using Algorithm 1, where x ^ l as the input;
(4)
compute x ^ ( l + 1 / 2 ) = x ^ l + δ B T D T ( y - DB x ^ l ) + γ ( E - I ) T ( E x ^ l - x ^ l ) ;
(5)
estimate α i ( l + 1 ) according to α i ( l + 1 ) = S ς ( Λ i l - β l ) + β l , where ς = λ 1 l / u , u is an auxiliary parameter, and Λ i l = Θ T B T D T ( y - DB Θ Θ k T R i x ^ ( l + 1 / 2 ) ) u + Θ k T R i x ^ ( l + 1 / 2 ) , where Θ k denotes the corresponding sub-dictionary for the patch R i x ^ ( l + 1 / 2 ) ;
(6)
if mod ( l , L 0 ) = 0 , update β i , where L 0 is constant;
(7)
reconstruct the HR image x ^ ( l + 1 ) according to
x ^ ( l + 1 ) = Θ α ^ ( l + 1 ) = ( i = 1 N R i T R i ) - 1 i = 1 N ( R i T Θ α ^ i ( l + 1 ) ) ,
where R i denotes the extraction matrix of x ^ i ;
end of inner loop;
end of outer loop.
3:
obtain the reconstructed HR image x ^ .

4. Experimental Results

4.1. Experimental Setting

In this subsection, we show minutely the datasets and parameters settings. Test images are shown in Figure 3. These images come from three different remote sensing image datasets: UCMLUD [46], USC-SIPI-aerials [47] and NWPU-RESISC45 [48]. UCMLUD contains 21 classes of remote images, and each class contains 100 images with size 256 × 256 . The spatial resolution for this dataset is 0.3 m/pixel. The spatial resolution for USC-SIPI-aerials is 1 m/pixel. NWPU-RESISC45 contains more kinds of images than UCMLUD. Specifically, NWPU-RESISC45 contains 45 categories, and each class has 700 images with size 256 × 256 . The spatial resolution of this database ranges from 0.2 m/pixel to 30 m/pixel.
Furthermore, the basic parameters settings in the proposed AJC method are as follows: the number of inner loop J is 160, the number of outer loop L is 5, ω is set to 0 . 0001 , and δ = 7 . In addition, to generate the test LR images, first, the HR image is blurred by the Gaussian Blur Kernel with size 7 × 7 and standard deviation 1.6. Then, we downsample the blurred image by a scale factor of 3 [7,49].

4.2. Parameters Setting

The size ( τ × τ ) of patch and the number (K) of clustering have an important impact on the SR performance. Too few clusters will eliminate the gaps between classes. Too many clusters will make the dictionary lose its representativeness and reliability. So we need to find an optimal K by [30]. Specifically, we first divide the training patches into K clusters, and then merge the classes containing a few image patches into the nearest neighboring classes. We analyze the impact of τ and K 0 on peak signal-to-noise ratio (PSNR) for all the test images, where K 0 denotes the predefined clustering number of K, and the results are shown in Table 1. The average PSNR varies with the patch size. In the case of the patch size of 5, the average PSNRs of the test images are superior. In the case of the same patch size and K 0 larger than 10, PSNRs are close. This phenomenon shows the robustness of the method to find an optimal K [30]. Furthermore, the larger the number of clustering, the more time it takes. In order to obtain a higher PSNR at a reasonable time, we set τ to 5 and K 0 to 50.

4.3. Comparison with Different Traditional Methods

To demonstrate the SR performance of the proposed AJC algorithm, we compare it with other SR methods, including Bicubic, SRSC [22], ASDS [30], MSEPLL [43], NARM [50] and LANR-NLM [51]. PSNR, structural similarity index (SSIM) [52] and erreur relative globale adimensionnelle de synthèse (ERGAS) [53] are used as the objective evaluation indices. The result with higher PSNR/SSIM and smaller ERGAS means the quality of the reconstructed image is better.

4.3.1. Noiseless Remote Sensing Images

In the noiseless case, the reconstruction results using different methods are shown in Table 2. For “Island”, the LANR-NLM method acquires better objective evaluation indices. However, considering all the test images, our algorithm has superior objective indices. Specifically, the average PSNR, SSIM and ERGAS are 31.90 dB, 0.8876 and 2.2912, respectively. In addition, a graph is drawn in Figure 4 for the results provided in Table 2. From Figure 4, it can be intuitively observed that our method has better objective indices than other methods. To intuitively show the visual quality of the reconstructed image, we compare the visual results as shown in Figure 5 and Figure 6. The SR performance of Bicubic interpolation is the worst. NARM produces smoother images. As shown in Figure 5f and Figure 6f, the MSEPLL method also smoothes many details of the image. In Figure 5, compare with other SR reconstructed methods, the proposed method reconstructs the HR image with fewer artifacts and clearer edges.
In Figure 6, ASDS and LANR-NLM tend to smooth out image details to some extent. As shown in Figure 6h, the image reconstructed using the proposed method has a clearer white line than others.

4.3.2. Noisy Remote Sensing Images

To demonstrate the effectiveness of our method in the noisy case, we add noise with a standard deviation of 5 to the degraded image. The results are shown in Table 3.
Compared with other methods, the ASDS method achieves better SR results on the “Aerial” image. However, the average PSNR, SSIM and ERGAS of our algorithm are the highest among these SR methods, which are 29.58 dB, 0.8144 and 2.8257, respectively. In addition, to more intuitively reflect the performance of our method, we draw a graph for the results provided in Table 3, as shown in Figure 4. Furthermore, we give visual effects as shown in Figure 7 and Figure 8. We can find that the proposed method also has better SR performance in suppressing image noise and preserving details and edges.
Consequently, according to the experiments on test images, the proposed method can achieve the best SR results in both noiseless and noisy cases.

4.4. Comparison with Different Deep Learning Methods

To reasonably compare with deep learning methods, we chose the classic deep learning methods SRCNN [32] and SRGAN [35], and the recently proposed remote sensing images method LGCnet [54]. For these methods, we retrain and fine tune their models, and then use these models for SR reconstruction. The training data come from three remote sensing image datasets: UCMLUD [46], USC-SIPI-aerials [47] and NWPU-RESISC45 [48]. All the training images are first blurred and then down-sampled to obtain the low-resolution images. Next, both the obtained low-resolution images and their original versions are collected as training pairs to train SRCNN [32], SRGAN [35], and LGCnet [54]. When the training error stops decreasing, we reduce their initial learning rates to fine-tuned their models to achieve their best performance. The results are shown in Table 4. For “Runway”, the resolution reconstructed by the LGCnet method is higher. However, for all the test images, our method is superior to SRCNN, SRGAN and LGCnet, and the average PSNR/SSIM gains over SRCNN, SRGAN and LGCnet are 0.84 dB/0.0215, 1.64 dB/0.0778 and 0.2 dB/0.0048, respectively. For ERGAS, our method is 0.0394 better than SRGAN on the test images. Our average ERGAS is 0.04 better than LGCnet’s average ERGAS and 0.1984 better than the average ERGAS of SRCNN. In order to show that our method does not need a lot of external data compared with the deep learning methods, we provide the training time and the number of training images, as shown in Table 5. Compared with other methods, the proposed approach can train with only one image, and saves lots of training time. The experiment results show that our method has better SR performance in noiseless cases.

4.5. Comparison with Different Methods on Datasets

To verify our method performance, we use different methods to reconstruct the test databases, including the “Airplane” and “Storage-tank” sub-databases from the UCMLUD, and the “Island” sub-database from the NWPU-RESISC45. The results are shown in Table 6. In Table 6, in addition to our approach, ASDS is the best traditional method, LGCnet is superior to SRCNN in deep learning methods. For the “Airplane” sub-database, our method is superior to the ASDS and LGCnet methods, and the average PSNR/SSIM gains over ASDS and LGCnet are 0.81 dB/0.0106 and 0.32 dB/0.0026, respectively. Meanwhile, our method has a better ERGAS than other methods, it is 1.7553. For the “Storage-tank” sub-database, the PSNR/SSIM of our algorithm is 0.56 dB/0.0119 higher than ASDS. Our ERGAS is 0.1194 better than the ERGAS of ASDS. Compare with LGCnet, although LGCnet has better SSIM and ERGAS, our approach gets the larger PSNR. For the “Island” sub-database, the PSNR/SSIM of our method is 0.34 dB/0.0046 superior to ASDS, and the PSNR/SSIM of LGCnet is 0.08 dB/0.0019 inferior to ours. The results show that the images reconstructed using our method has better performance.
The previous experimental analysis shows the comprehensive performance of our proposed method. Specifically, for images that satisfy the joint constraint (i.e., images with repeated structures and many edges), the proposed method has superior performance. For images with highly complex texture, the proposed method is not effective enough. In our paper, the proposed approach is based on local gradient constraint and nonlocal similarity. Because images with repeated structures and many edges can perfectly satisfy these two features, they can acquire superior reconstruction performance. Thus, in most cases, this approach performs better than many other super-resolution methods. However, for images with highly complex texture, these two features are not reliable. If we use these two features for reconstruction, the reconstructed image will not be very good. In fact, many SR approaches do not work well under such extreme conditions.

4.6. The Effectiveness of Joint Constraint

The image itself comprises repetitive structures, i.e., self-similarity. Researchers often exploit the nonlocal self-similarity of images for the SR reconstruction. However, it is not enough to consider only the nonlocal self-similarity to improve the resolution of the input image. To address the above problem, we construct a local constraint according to the proposed filter. Considering the joint constraint, the quality of the reconstructed image will be greatly improved. In order to demonstrate the performance of the joint constraint, we compare the SR performance for the joint constraint and the single nonlocal constraint. The results are shown in Table 7. For all the test images, the method for the joint constraint achieves superior objective qualities. Specifically, the average PSNR and SSIM gains of the joint constraint over the single nonlocal constraint are 0.38 dB and 0.0071, respectively, and the average ERGAS of the joint constraint is 0.097 better than the single nonlocal constraint. Experiment results demonstrate that the complementary joint constraint can effectively improve the image quality.

4.7. The Effectiveness of Adaptive Parameters

In the classic sparse coding problem, the choice of regularization parameters is very important. However, the parameters of most methods are fixed. In this paper, we adaptively select the parameters according to the noise level. To demonstrate the effectiveness of our adaptive parameters, we compare the results of the fixed parameters with those of the adaptive parameters as shown in Figure 9. In the case of fixed parameters, λ 1 and λ 2 are set to 0.33 and 0.001, respectively. In Figure 9, the maximum PSNR gain is 0.36 dB, and the corresponding image is “Runway”. At the same time, the image with the largest SSIM gain is also “Runway”. Specifically, it is 0.0061. For “Runway”, the ERGAS with adaptive parameters is 0.0408 better than that with fixed parameters. Experiment results indicate that the adaptive parameters are beneficial to improve the SR performance.

4.8. Complexity Analysis

Our method takes major cost on three part: the sub-dictionaries learning, the calculation of local structural filter, and the nonlocal search. In the sub-dictionaries learning, the core procedure involves the clustering of K-means. Its computational complexity is O ( L τ 2 m n K ) . The calculation of local structural filter E takes O ( L J m n ) . The nonlocal search is related to the patch size, the search window size b × b and the number of the images. It takes O ( m n τ 2 b 2 ) . Take the “Airplane” with size 256 × 256 as a test image, the LANR-NLM method takes 40.82 s, LGCnet takes 0.74s, NARM and ASDS take 68.02 s and 107.92 s, respectively, the MSEPLL and our methods take 215.88 s and 297.91 s, respectively. Therein, traditional SR methods are implemented in MATLAB 2014a on a computer with Intel(R) Core(TM) i7-7700K CPU @ 4.20GHZ 4.20GHZ, 16.0GB RAM and 64-bit Windows 7 operating system. Our method takes a little longer time than others. This is because the proposed method learns the dictionary online for the input image and performs adaptive filtering for each iteration of the HR image. However, our method does not need external training examples, which saves a large amount of training time. In conclusion, our approach achieves the best visual and objective quality with reasonable running time.

5. Conclusions

In this paper, we propose a novel SR scheme based on sparse representation for the single remote sensing image. First, we use the single-dictionary method to learn the compact dictionary that exploits only the unique information of remote sensing image itself. Compared with the double-dictionary method and the deep learning method, this method has an advantage in the absence of external samples. Second, we propose a local structure filter based on the local gradient of image, and then a local structure prior is constructed. After that, the joint prior is constructed, including the local structure prior and the nonlocal self-similarity prior, which can effectively improve the fine structures recovery ability. Finally, the reconstructed HR images can be obtained by using the iterative shrinkage algorithm. The results show that the proposed local structure prior shows superior edge-preserving performance and the complementary prior constructed is more conducive in improving the SR performance. Compared with other methods, the HR images reconstructed by our scheme have better visual quality and higher objective evaluation indices. In the future, we will extend the proposed method to other image processing applications.

Author Contributions

L.F. wrote this manuscript. L.F., C.R., X.H., X.W. and Z.W. contributed to the writing, direction, and content, and revised the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 61801316, the National Postdoctoral Program for Innovative Talents of China under Grant BX201700163, and the Chengdu Industrial Cluster Collaborative Innovation Project under grant no. 2016-XT00-00015-GX.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Lu, X.; Zheng, X.; Yuan, Y. Remote sensing scene classification by unsupervised representation learning. IEEE Trans. Geosci. Remote. Sens. 2017, 55, 5148–5157. [Google Scholar] [CrossRef]
  2. Zhu, Q.; Zhong, Y.; Wu, S.; Zhang, L.; Li, D. Scene Classification Based on the Sparse Homogeneous–Heterogeneous Topic Feature Model. IEEE Trans. Geosci. Remote. Sens. 2018, 56, 2689–2703. [Google Scholar] [CrossRef]
  3. Belward, A.S.; Skøien, J.O. Who launched what, when and why; trends in global land-cover observation capacity from civilian earth observation satellites. ISPRS J.Photogramm. Remote. Sens. 2015, 103, 115–128. [Google Scholar] [CrossRef]
  4. Duan, H.; Gan, L. Elitist chemical reaction optimization for contour-based target recognition in aerial images. IEEE Trans. Geosci. Remote. Sens. 2015, 53, 2845–2859. [Google Scholar] [CrossRef]
  5. Pei, J.; Huang, Y.; Huo, W.; Zhang, Y.; Yang, J.; Yeo, T.S. SAR Automatic Target Recognition Based on Multiview Deep Learning Framework. IEEE Trans. Geosci. Remote. Sens. 2018, 56, 2196–2210. [Google Scholar] [CrossRef]
  6. Haut, J.M.; Fernandez-Beltran, R.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Pla, F. A new deep generative network for unsupervised remote sensing single-image super-resolution. IEEE Trans. Geosci. Remote. Sens. 2018, 56, 6792–6810. [Google Scholar] [CrossRef]
  7. Hou, B.; Zhou, K.; Jiao, L. Adaptive Super-Resolution for Remote Sensing Images Based on Sparse Representation With Global Joint Dictionary Model. IEEE Trans. Geosci. Remote. Sens. 2018, 56, 2312–2327. [Google Scholar] [CrossRef]
  8. Feng, R.; Zhong, Y.; Xu, X.; Zhang, L. Adaptive sparse subpixel mapping with a total variation model for remote sensing imagery. IEEE Trans. Geosci. Remote. Sens. 2016, 54, 2855–2872. [Google Scholar] [CrossRef]
  9. Li, F.; Jia, X.; Fraser, D.; Lambert, A. Super resolution for remote sensing images based on a universal hidden Markov tree model. IEEE Trans. Geosci. Remote. Sens. 2010, 48, 1270–1278. [Google Scholar]
  10. Sadjadi, F.A. Seeing through degraded visual environment. In Proceedings of the Automatic Target Recognition XXII. International Society for Optics and Photonics, Baltimore, MD, USA, 11 May 2012; Volume 8391, p. 839106. [Google Scholar]
  11. Chen, J.; Nunez-Yanez, J.; Achim, A. Video super-resolution using low rank matrix completion. In Proceedings of the 2013 IEEE International Conference on Image Processing, Melbourne, Australia, 15–18 September 2013; pp. 1376–1380. [Google Scholar]
  12. Liu, X.; Chen, L.; Wang, W.; Zhao, J. Robust Multi-Frame Super-Resolution Based on Spatially Weighted Half-Quadratic Estimation and Adaptive BTV Regularization. IEEE Trans. Image Process. 2018, 27, 4971–4986. [Google Scholar] [CrossRef]
  13. Tsai, R. Multiframe image restoration and registration. Adv. Comput. Vis. Image Process. 1984, 1, 317–339. [Google Scholar]
  14. Huang, Y.; Wang, W.; Wang, L. Bidirectional recurrent convolutional networks for multi-frame super-resolution. In Proceedings of the Advances in Neural Information Processing Systems, Montreal, QC, Canada, 7–12 December 2015; pp. 235–243. [Google Scholar]
  15. Farsiu, S.; Robinson, M.D.; Elad, M.; Milanfar, P. Fast and robust multiframe super resolution. IEEE Trans. Image Process. 2004, 13, 1327–1344. [Google Scholar] [CrossRef] [PubMed]
  16. Patanavijit, V.; Jitapunkul, S. A robust iterative multiframe super-resolution reconstruction using a Huber Bayesian approach with Huber-Tikhonov regularization. In Proceedings of the International Symposium on Intelligent Signal Processing and Communications, Tottori, Japan, 12–15 December 2006; pp. 13–16. [Google Scholar]
  17. Wang, W.; Ren, C.; He, X.; Chen, H.; Qing, L. Video Super-Resolution via Residual Learning. IEEE Access 2018, 6, 23767–23777. [Google Scholar] [CrossRef]
  18. Ren, C.; He, X.; Nguyen, T.Q. Single image super-resolution via adaptive high-dimensional non-local total variation and adaptive geometric feature. IEEE Trans. Image Process. 2017, 26, 90–106. [Google Scholar] [CrossRef] [PubMed]
  19. Ren, C.; He, X.; Nguyen, T.Q. Adjusted Non-Local Regression and Directional Smoothness for Image Restoration. IEEE Trans. Multimed. 2019, 21, 731–745. [Google Scholar] [CrossRef]
  20. Ren, C.; He, X.; Teng, Q.; Wu, Y.; Nguyen, T.Q. Single image super-resolution using local geometric duality and non-local similarity. IEEE Trans. Image Process. 2016, 25, 2168–2183. [Google Scholar] [CrossRef] [PubMed]
  21. Ren, C.; He, X.; Pu, Y. Nonlocal Similarity Modeling and Deep CNN Gradient Prior for Super Resolution. IEEE Signal Process. Lett. 2018, 25, 916–920. [Google Scholar] [CrossRef]
  22. Yang, J.; Wright, J.; Huang, T.S.; Ma, Y. Image super-resolution via sparse representation. IEEE Trans. Image Process. 2010, 19, 2861–2873. [Google Scholar] [CrossRef]
  23. Huang, S.; Sun, J.; Yang, Y.; Fang, Y.; Lin, P.; Que, Y. Robust Single-Image Super-Resolution Based on Adaptive Edge-Preserving Smoothing Regularization. IEEE Trans. Image Process. 2018, 27, 2650–2663. [Google Scholar] [CrossRef]
  24. Li, X.; Orchard, M.T. New edge-directed interpolation. IEEE Trans. Image Process. 2001, 10, 1521–1527. [Google Scholar]
  25. Zhang, Y.; Fan, Q.; Bao, F.; Liu, Y.; Zhang, C. Single-Image Super-Resolution Based on Rational Fractal Interpolation. IEEE Trans. Image Process. 2018, 27, 3782–3797. [Google Scholar] [PubMed]
  26. Tai, Y.W.; Liu, S.; Brown, M.S.; Lin, S. Super resolution using edge prior and single image detail synthesis. In Proceedings of the 2010 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Francisco, CA, USA, 13–18 June 2010; pp. 2400–2407. [Google Scholar]
  27. Timofte, R.; De Smet, V.; Van Gool, L. Anchored neighborhood regression for fast example-based super-resolution. In Proceedings of the IEEE International Conference on Computer Vision, Sydney, NSW, Australia, 1–8 December 2013; pp. 1920–1927. [Google Scholar]
  28. Peleg, T.; Elad, M. A statistical prediction model based on sparse representations for single image super-resolution. IEEE Trans. Image Process. 2014, 23, 2569–2582. [Google Scholar] [CrossRef]
  29. Gou, S.; Liu, S.; Yang, S.; Jiao, L. Remote sensing image super-resolution reconstruction based on nonlocal pairwise dictionaries and double regularization. IEEE J. Sel. Top. Appl. Earth Obs. Remote. Sens. 2014, 7, 4784–4792. [Google Scholar] [CrossRef]
  30. Dong, W.; Zhang, L.; Shi, G.; Wu, X. Image deblurring and super-resolution by adaptive sparse domain selection and adaptive regularization. IEEE Trans. Image Process. 2011, 20, 1838–1857. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  31. Dong, W.; Zhang, L.; Shi, G.; Li, X. Nonlocally centralized sparse representation for image restoration. IEEE Trans. Image Process. 2013, 22, 1620–1630. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  32. Dong, C.; Loy, C.C.; He, K.; Tang, X. Image super-resolution using deep convolutional networks. IEEE Trans. Pattern Anal. Mach. Intell. 2016, 38, 295–307. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  33. Li, J.; Yuan, Q.; Shen, H.; Meng, X.; Zhang, L. Hyperspectral Image Super-Resolution by Spectral Mixture Analysis and Spatial-Spectral Group Sparsity. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1250–1254. [Google Scholar] [CrossRef]
  34. Tai, Y.; Yang, J.; Liu, X. Image super-resolution via deep recursive residual network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 3147–3155. [Google Scholar]
  35. Ledig, C.; Theis, L.; Huszár, F.; Caballero, J.; Cunningham, A.; Acosta, A.; Aitken, A.; Tejani, A.; Totz, J.; Wang, Z.; et al. Photo-realistic single image super-resolution using a generative adversarial network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 4681–4690. [Google Scholar]
  36. Mairal, J.; Bach, F.; Ponce, J.; Sapiro, G.; Zisserman, A. Non-local sparse models for image restoration. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2272–2279. [Google Scholar]
  37. Zeyde, R.; Elad, M.; Protter, M. On single image scale-up using sparse-representations. In Proceedings of the International Conference on Curves and Surfaces, Avignon, France, 24–30 June 2010; pp. 711–730. [Google Scholar]
  38. Vera-Candeas, P.; Ruiz-Reyes, N.; Rosa-Zurera, M.; Martinez-Munoz, D.; López-Ferreras, F. Transient modeling by matching pursuits with a wavelet dictionary for parametric audio coding. IEEE Signal Process. Lett. 2004, 11, 349–352. [Google Scholar] [CrossRef]
  39. Daubechies, I.; Defrise, M.; De Mol, C. An iterative thresholding algorithm for linear inverse problems with a sparsity constraint. Commun. Pure Appl. Math. 2004, 57, 1413–1457. [Google Scholar] [CrossRef] [Green Version]
  40. Aharon, M.; Elad, M.; Bruckstein, A. K-SVD: An algorithm for designing overcomplete dictionaries for sparse representation. IEEE Trans. Signal Process. 2006, 54, 4311–4322. [Google Scholar] [CrossRef]
  41. Lischinski, D.; Farbman, Z.; Uyttendaele, M.; Szeliski, R. Interactive local adjustment of tonal values. ACM Trans. Graph. (TOG) 2006, 25, 646–653. [Google Scholar] [CrossRef]
  42. Liu, X.; Tanaka, M.; Okutomi, M. Single-image noise level estimation for blind denoising. IEEE Trans. Image Process. 2013, 22, 5226–5237. [Google Scholar] [CrossRef] [PubMed]
  43. Papyan, V.; Elad, M. Multi-scale patch-based image restoration. IEEE Trans. Image Process. 2016, 25, 249–261. [Google Scholar] [CrossRef] [PubMed]
  44. Zhang, K.; Gao, X.; Tao, D.; Li, X. Single image super-resolution with non-local means and steering kernel regression. IEEE Trans. Image Process. 2012, 21, 4544–4556. [Google Scholar] [CrossRef]
  45. Li, Y.; Dong, W.; Shi, G.; Xie, X. Learning parametric distributions for image super-resolution: Where patch matching meets sparse coding. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 450–458. [Google Scholar]
  46. Yang, Y.; Newsam, S. Bag-of-visual-words and spatial extensions for land-use classification. In Proceedings of the 18th SIGSPATIAL International Conference on Advances in Geographic Information Systems, San Jose, CA, USA, 2–5 November 2010; pp. 270–279. [Google Scholar]
  47. The USC-SIPI Image Database. Available online: http://sipi.usc.edu/database/database.php (accessed on 26 February 2020).
  48. Cheng, G.; Han, J.; Lu, X. Remote sensing image scene classification: Benchmark and state of the art. Proc. IEEE 2017, 105, 1865–1883. [Google Scholar] [CrossRef] [Green Version]
  49. Li, F.; Xin, L.; Guo, Y.; Gao, D.; Kong, X.; Jia, X. Super-resolution for GaoFen-4 remote sensing images. IEEE Geosci. Remote Sens. Lett. 2017, 15, 28–32. [Google Scholar] [CrossRef]
  50. Dong, W.; Zhang, L.; Lukac, R.; Shi, G. Sparse representation based image interpolation with nonlocal autoregressive modeling. IEEE Trans. Image Process. 2013, 22, 1382–1394. [Google Scholar] [CrossRef] [Green Version]
  51. Jiang, J.; Ma, X.; Chen, C.; Lu, T.; Wang, Z.; Ma, J. Single image super-resolution via locally regularized anchored neighborhood regression and nonlocal means. IEEE Trans. Multimed. 2017, 19, 15–26. [Google Scholar] [CrossRef]
  52. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error measurement to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [Green Version]
  53. Veganzones, M.A.; Simoes, M.; Licciardi, G.; Yokoya, N.; Bioucas-Dias, J.M.; Chanussot, J. Hyperspectral super-resolution of locally low rank images from complementary multisource data. IEEE Trans. Image Process. 2016, 25, 274–288. [Google Scholar] [CrossRef] [Green Version]
  54. Lei, S.; Shi, Z.; Zou, Z. Super-resolution for remote sensing images via local–global combined network. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1243–1247. [Google Scholar] [CrossRef]
Figure 1. The proposed framework.
Figure 1. The proposed framework.
Sensors 20 01276 g001
Figure 2. The statistic property of ψ .
Figure 2. The statistic property of ψ .
Sensors 20 01276 g002
Figure 3. Experimental test images, including the following images: aerial, airplane, residential, parking-lot, terrace, farmland, mountain, industrial-area, river, building, meadow, island, runaway, storage-tank, harbor.
Figure 3. Experimental test images, including the following images: aerial, airplane, residential, parking-lot, terrace, farmland, mountain, industrial-area, river, building, meadow, island, runaway, storage-tank, harbor.
Sensors 20 01276 g003
Figure 4. The average objective indices for the test images in the case of noise = 0 and noise = 5.
Figure 4. The average objective indices for the test images in the case of noise = 0 and noise = 5.
Sensors 20 01276 g004
Figure 5. Visual comparisons of the proposed method and other methods on the aerial image (noise = 0). (a) original image. (b) Bicubic interpolation (PSNR:24.67, SSIM:0.6537, ERGAS:3.9962). (c) SRSC (PSNR:25.10, SSIM:0.6925, ERGAS:3.8693). (d) ASDS (PSNR:29.24, SSIM:0.8242, ERGAS:2.3713). (e) NARM (PSNR:25.97, SSIM:0.6757, ERGAS:3.6780). (f) MSEPLL (PSNR:29.09, SSIM:0.8138, ERGAS:2.4803). (g) LANR-NLM (PSNR:29.03, SSIM:0.8220, ERGAS:2.4152). (h) Ours (PSNR:29.61, SSIM:0.8355, ERGAS:2.2616).
Figure 5. Visual comparisons of the proposed method and other methods on the aerial image (noise = 0). (a) original image. (b) Bicubic interpolation (PSNR:24.67, SSIM:0.6537, ERGAS:3.9962). (c) SRSC (PSNR:25.10, SSIM:0.6925, ERGAS:3.8693). (d) ASDS (PSNR:29.24, SSIM:0.8242, ERGAS:2.3713). (e) NARM (PSNR:25.97, SSIM:0.6757, ERGAS:3.6780). (f) MSEPLL (PSNR:29.09, SSIM:0.8138, ERGAS:2.4803). (g) LANR-NLM (PSNR:29.03, SSIM:0.8220, ERGAS:2.4152). (h) Ours (PSNR:29.61, SSIM:0.8355, ERGAS:2.2616).
Sensors 20 01276 g005
Figure 6. Visual comparisons of the proposed method and other methods on the airplane image (noise = 0). (a) original image. (b) Bicubic interpolation (PSNR:26.27, SSIM:0.7989, ERGAS:2.6212). (c) SRSC (PSNR:26.95, SSIM:0.8205, ERGAS:2.4746). (d) ASDS (PSNR:30.95, SSIM:0.8949, ERGAS:1.4973). (e) NARM (PSNR:27.20, SSIM:0.8108, ERGAS:2.6506). (f) MSEPLL (PSNR:31.45, SSIM:0.8944, ERGAS:1.5388). (g) LANR-NLM (PSNR:30.69, SSIM:0.8919, ERGAS:1.5706). (h) Ours (PSNR:32.19, SSIM:0.9078, ERGAS:1.3277).
Figure 6. Visual comparisons of the proposed method and other methods on the airplane image (noise = 0). (a) original image. (b) Bicubic interpolation (PSNR:26.27, SSIM:0.7989, ERGAS:2.6212). (c) SRSC (PSNR:26.95, SSIM:0.8205, ERGAS:2.4746). (d) ASDS (PSNR:30.95, SSIM:0.8949, ERGAS:1.4973). (e) NARM (PSNR:27.20, SSIM:0.8108, ERGAS:2.6506). (f) MSEPLL (PSNR:31.45, SSIM:0.8944, ERGAS:1.5388). (g) LANR-NLM (PSNR:30.69, SSIM:0.8919, ERGAS:1.5706). (h) Ours (PSNR:32.19, SSIM:0.9078, ERGAS:1.3277).
Sensors 20 01276 g006
Figure 7. The reconstruction for the meadow of noisy image with different methods (noise = 5). (a) original image. (b) bicubic interpolation (PSNR:32.19, SSIM:0.7914, ERGAS:2.6744). (c) ASDS (PSNR:33.39, SSIM:0.8274, ERGAS:2.3056). (d) NARM (PSNR:32.56, SSIM:0.7993, ERGAS:2.7643). (e) MSEPLL (PSNR:33.42, SSIM:0.8260, ERGAS:/2.2176). (f) ours (PSNR:33.65, SSIM:0.8302, ERGAS:2.2605).
Figure 7. The reconstruction for the meadow of noisy image with different methods (noise = 5). (a) original image. (b) bicubic interpolation (PSNR:32.19, SSIM:0.7914, ERGAS:2.6744). (c) ASDS (PSNR:33.39, SSIM:0.8274, ERGAS:2.3056). (d) NARM (PSNR:32.56, SSIM:0.7993, ERGAS:2.7643). (e) MSEPLL (PSNR:33.42, SSIM:0.8260, ERGAS:/2.2176). (f) ours (PSNR:33.65, SSIM:0.8302, ERGAS:2.2605).
Sensors 20 01276 g007
Figure 8. The reconstruction for the residential of noisy image with different methods (noise = 5). (a) original image. (b) bicubic interpolation (PSNR:22.05, SSIM:0.6359, ERGAS:6.6547). (c) ASDS (PSNR:26.16, SSIM:0.7958, ERGAS:4.1301). (d) NARM (PSNR:23.61, SSIM:0.6981, ERGAS:5.6835). (e) MSEPLL (PSNR:26.21, SSIM:0.8071, ERGAS:4.0817). (f) ours (PSNR:26.47, SSIM:0.8209, ERGAS:3.9976).
Figure 8. The reconstruction for the residential of noisy image with different methods (noise = 5). (a) original image. (b) bicubic interpolation (PSNR:22.05, SSIM:0.6359, ERGAS:6.6547). (c) ASDS (PSNR:26.16, SSIM:0.7958, ERGAS:4.1301). (d) NARM (PSNR:23.61, SSIM:0.6981, ERGAS:5.6835). (e) MSEPLL (PSNR:26.21, SSIM:0.8071, ERGAS:4.0817). (f) ours (PSNR:26.47, SSIM:0.8209, ERGAS:3.9976).
Sensors 20 01276 g008
Figure 9. Peak Signal-to-noise Ratio(PSNR) and Structural Similarity Index(SSIM) gains of the method with adaptive parameters over the method with fixed parameters. (a) PSNR gains. (b) SSIM gains.
Figure 9. Peak Signal-to-noise Ratio(PSNR) and Structural Similarity Index(SSIM) gains of the method with adaptive parameters over the method with fixed parameters. (a) PSNR gains. (b) SSIM gains.
Sensors 20 01276 g009
Table 1. Average PSNR (Peak Signal-to-noise Ratio) of Test Images with Different Parameters (Scale Factor = 3, Noise = 0).
Table 1. Average PSNR (Peak Signal-to-noise Ratio) of Test Images with Different Parameters (Scale Factor = 3, Noise = 0).
τ K 0
1030507090110
331.1731.2231.2131.2131.2131.21
531.8431.8631.9031.8731.8731.86
731.6831.7731.7831.7731.7831.78
931.4931.5731.5931.5831.5931.59
Table 2. PSNR, SSIM (Structural Similarity Index) and ERGAS (Erreur Relative Globale Adimensionnelle De synthèse) Results of Reconstruction for Test Images (Scale Factor = 3, Noise = 0).
Table 2. PSNR, SSIM (Structural Similarity Index) and ERGAS (Erreur Relative Globale Adimensionnelle De synthèse) Results of Reconstruction for Test Images (Scale Factor = 3, Noise = 0).
ImageBicubicSRSCASDSNARMMSEPLLLANR-NLMProposed Method
Aerial24.6725.1029.2425.9729.0929.0329.61
0.65370.69250.82420.67570.81380.82200.8355
3.99623.86932.37133.67802.48032.41522.2616
Airplane26.2726.9530.9527.2031.4530.6932.19
0.79890.82050.89490.81080.89440.89190.9078
2.62122.47461.49732.65061.53881.57061.3277
Building22.3723.0628.3823.8028.3327.7329.69
0.70160.75340.89580.74780.88130.87360.9295
4.11243.90172.06493.72222.29192.19731.7704
Farmland28.3528.7233.5029.8433.7433.3934.41
0.78740.80800.89680.80210.89400.89570.9088
2.70812.64181.48122.58451.46991.51641.3482
Residential22.1222.3527.2623.6927.3426.6928.13
0.64900.67970.86120.70790.86320.84210.8866
6.60386.43143.62635.67363.62073.90713.3038
Harbor19.4619.8022.4620.3322.3022.2022.74
0.58390.62820.77740.62500.75750.75560.7954
9.75459.54906.96179.03847.09547.18176.6975
Industrial-area24.3624.7728.2025.3428.0227.8028.77
0.58260.62870.79510.62250.78700.78180.8154
4.17384.07082.67683.91012.78032.79632.5078
Island42.0542.4644.5541.7644.0445.5445.28
0.96510.96780.97750.96110.97550.98210.9798
0.63970.62820.47301.24500.47270.42990.4444
Meadow32.8133.1435.6632.8935.7235.8235.94
0.82310.83910.89600.81340.90050.90160.8994
2.48632.44101.76472.67571.69461.76571.7365
Mountain28.6528.9234.4930.2634.5634.3834.81
0.74100.76770.89960.76600.89750.90220.9063
3.26983.23371.68073.01631.74131.68661.6089
Parking-lot20.4420.9424.7321.6824.5124.1925.06
0.57410.63240.80190.62120.78050.78310.8225
7.90697.65704.90417.10715.14415.17084.6497
River28.4128.8931.8128.7931.5331.3132.09
0.74590.77610.87100.74920.86510.86350.8780
3.68943.55372.49263.76402.50602.65062.4179
Runway27.0827.4431.9428.4731.5630.9332.49
0.77650.79880.88610.80780.87260.86960.9031
3.23693.17141.84892.99071.90892.07701.7356
Storage-tank26.2826.7131.8827.7731.9431.0833.11
0.81430.83490.92240.83530.92640.91120.9367
3.28703.19181.67672.99331.57351.89491.4944
Terrace28.3228.9332.1529.2731.5832.1334.20
0.70660.75260.85720.74850.84640.86050.9086
2.09332.01331.32152.24631.38961.34781.0632
Average26.7827.3631.1527.8031.0530.8631.90
0.72690.76340.87050.75300.86370.86240.8876
4.03863.92572.45613.81972.51392.57392.2912
Table 3. PSNR/SSIM/ERGAS Results of Reconstruction for Test Images (Scale Factor = 3, Noise = 5).
Table 3. PSNR/SSIM/ERGAS Results of Reconstruction for Test Images (Scale Factor = 3, Noise = 5).
ImageBicubicASDSNARMMSEPLLProposed Method
Aerial24.57/0.6385/4.042327.92/0.7557/2.756325.87/0.6670/3.699827.76/0.7388/2.867727.84/0.7490/2.7710
Airplane26.13/0.7693/2.665829.57/0.8361/1.763727.06/0.7967/2.622529.93/0.8277/1.703430.24/0.8456/1.6599
Building22.31/0.6817/4.139726.96/0.8113/2.433823.75/0.7405/3.733626.82/0.8093/2.600227.27/0.8479/2.3386
Farmland28.09/0.7613/2.789331.35/0.8210/1.899729.67/0.7899/2.607831.66/0.8249/1.844631.88/0.8379/1.8029
Residential22.05/0.6359/6.654726.16/0.7958/4.130123.61/0.6981/5.683526.21/0.8071/4.081726.46/0.8207/4.0046
Harbor19.44/0.5679/9.777121.99/0.7214/7.344220.31/0.6178/8.999521.82/0.7063/7.447222.02/0.7297/7.2726
Industrial-area24.26/0.5704/4.219926.96/0.7195/3.085725.21/0.6119/3.956626.84/0.7092/3.159027.08/0.7228/3.0452
Island38.37/0.9164/0.985139.64/0.9420/0.851840.32/0.9453/1.313537.91/0.9068/1.015739.20/0.9339/0.8952
Meadow32.19/0.7914/2.674433.39/0.8274/2.305632.56/0.7993/2.764333.42/0.8260/2.217633.65/0.8303/2.2597
Mountain28.37/0.7233/3.374931.65/0.8168/2.324330.02/0.7533/3.081931.81/0.8196/2.344232.01/0.8293/2.2192
Parking-lot20.39/0.5593/7.950323.92/0.7187/5.373221.60/0.6098/7.144923.82/0.7098/5.519723.76/0.7283/5.3984
River28.16/0.7222/3.796230.13/0.7950/3.024828.64/0.7375/3.790629.92/0.7881/3.006430.19/0.7969/3.0082
Runway26.90/0.7451/3.304729.77/0.8200/2.366728.44/0.7974/2.994229.55/0.7980/2.386230.47/0.8431/2.1896
Storage-tank26.12/0.7854/3.347730.20/0.8618/2.045927.64/0.8227/2.992030.22/0.8662/1.929830.50/0.8766/2.0174
Terrace28.07/0.6842/2.153730.07/0.7717/1.684829.10/0.7374/2.237329.79/0.7688/1.713931.19/0.8245/1.5030
Average26.36/0.7035/4.125129.31/0.8009/2.892727.59/0.7417/3.841529.16/0.7938/2.922529.58/0.8144/2.8257
Table 4. PSNR/SSIM/ERGAS Results for Test Images (Scale Factor = 3, Noise = 0).
Table 4. PSNR/SSIM/ERGAS Results for Test Images (Scale Factor = 3, Noise = 0).
ImageBicubicSRCNNLGCnetSRGANProposed Method
Aerial24.67/0.6537/3.996229.04/0.8182/2.412829.36/0.8287/2.329827.89/0.7873/2.392129.61/0.8355/2.2616
Airplane26.27/0.7989/2.621231.22/0.8960/1.476932.00/0.9033/1.355230.46/0.8806/1.369432.19/0.9078/1.3277
Building22.37/0.7016/4.112427.83/0.8742/2.171528.95/0.9121/1.926228.02/0.9045/1.866129.69/0.9295/1.7704
Farmland28.35/0.7874/2.708133.56/0.8968/1.486734.03/0.9043/1.409432.51/0.8905/1.426534.41/0.9088/1.3482
Residential22.12/0.6490/6.603827.15/0.8563/3.705927.71/0.8757/3.470026.56/0.8712/3.424828.13/0.8866/3.3038
Harbor19.46/0.5839/9.754522.42/0.7738/6.992422.67/0.7920/6.754521.58/0.7891/6.586022.74/0.7954/6.6975
Industrial-area24.36/0.5826/4.173828.05/0.7878/2.717028.41/0.8043/2.615927.03/0.8063/2.609828.77/0.8154/2.5078
Island42.05/0.9651/0.639745.04/0.9797/0.455745.05/0.9789/0.456142.61/0.9702/0.559745.28/0.9798/0.4444
Meadow32.81/0.8231/2.486335.78/0.8989/1.773835.87/0.8990/1.752434.35/0.8845/1.782135.94/0.8994/1.7365
Mountain28.65/0.7410/3.269834.50/0.9012/1.664934.53/0.9034/1.661233.06/0.8939/1.689034.81/0.9063/1.6089
Parking-lot20.44/0.5741/7.906924.97/0.7937/4.727525.46/0.8204/4.441324.27/0.8102/4.387825.06/0.8225/4.6497
River28.41/0.7459/3.689431.51/0.8665/2.590532.01/0.8756/2.439630.73/0.8680/2.435132.09/0.8780/2.4179
Runway27.08/0.7765/3.236930.99/0.8661/2.062332.59/0.9052/1.716030.67/0.8716/1.808932.49/0.9031/1.7356
Storage-tank26.28/0.8143/3.287031.48/0.9165/1.807732.62/0.9300/1.583631.09/0.9188/1.573833.11/0.9367/1.4944
Terrace28.32/0.7066/2.093332.45/0.8654/1.299034.26/0.9087/1.055933.05/0.9023/1.048334.20/0.9086/1.0632
Average26.78/0.7269/4.038631.06/0.8661/2.489631.70/0.8828/2.331230.26/0.8098/2.330631.90/0.8876/2.2912
Table 5. Comparison for the number of training images and training time.
Table 5. Comparison for the number of training images and training time.
SRCNNLGCnetSRGANProposed Method
The number of training images2145214542901
Training time2 days6 h8 h60 s
Table 6. PSNR/SSIM/ERGAS Results on Databases (Scale Factor = 3, Noise = 0).
Table 6. PSNR/SSIM/ERGAS Results on Databases (Scale Factor = 3, Noise = 0).
DatasetAirplaneStorage-TankIsland
Bicubic25.05/0.7351/3.758625.43/0.7148/4.108533.49/0.8812/2.6453
ASDS30.82/0.8665/1.919230.04/0.8474/2.586637.78/0.9357/1.4686
SRCNN30.77/0.8640/1.922530.05/0.8469/2.542937.87/0.9373/1.4911
LANR-NLM30.29/0.8580/2.086129.63/0.8388/2.689637.77/0.9376/1.5116
LGCnet31.31/0.8745/1.794830.55/0.8622/2.356038.04/0.9384/1.4303
Proposed method31.63/0.8771/1.755330.60/0.8593/2.467238.12/0.9403/1.4403
Table 7. PSNR/SSIM/ERGAS Results of Nonlocal and Joint constraints (Scale Factor = 3, Noise = 0).
Table 7. PSNR/SSIM/ERGAS Results of Nonlocal and Joint constraints (Scale Factor = 3, Noise = 0).
ImageNonlocal ConstraintJoint Constraint
Aerial29.50/0.8342/2.291729.61/0.8355/2.2616
Airplane31.44/0.9026/1.446432.19/0.9078/1.3277
Building27.93/0.9014/2.168829.69/0.9295/1.7704
Farmland33.65/0.8993/1.472534.41/0.9088/1.3482
Residential28.06/0.8834/3.332528.13/0.8866/3.3038
Harbor22.18/0.7614/7.138722.74/0.7954/6.6975
Industrial-area28.72/0.8129/2.521528.77/0.8154/2.5078
Island44.71/0.9758/0.474645.28/0.9798/0.4444
Meadow35.94/0.8996/1.736435.94/0.8994/1.7365
Mountain34.80/0.9060/1.610134.81/0.9063/1.6089
Parking-lot24.95/0.8188/4.711325.06/0.8225/4.6497
River32.07/0.8774/2.422132.09/0.8780/2.4179
Runway31.67/0.8908/1.908532.49/0.9031/1.7356
Storage-tank32.92/0.9354/1.527833.11/0.9367/1.4944
Terrace34.22/0.9091/1.060334.20/0.9086/1.0632
Average31.52/0.8805/2.388231.90/0.8876/2.2912

Share and Cite

MDPI and ACS Style

Fu, L.; Ren, C.; He, X.; Wu, X.; Wang, Z. Single Remote Sensing Image Super-Resolution with an Adaptive Joint Constraint Model. Sensors 2020, 20, 1276. https://doi.org/10.3390/s20051276

AMA Style

Fu L, Ren C, He X, Wu X, Wang Z. Single Remote Sensing Image Super-Resolution with an Adaptive Joint Constraint Model. Sensors. 2020; 20(5):1276. https://doi.org/10.3390/s20051276

Chicago/Turabian Style

Fu, Lingli, Chao Ren, Xiaohai He, Xiaohong Wu, and Zhengyong Wang. 2020. "Single Remote Sensing Image Super-Resolution with an Adaptive Joint Constraint Model" Sensors 20, no. 5: 1276. https://doi.org/10.3390/s20051276

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop