Next Article in Journal
Data Center Traffic Prediction Algorithms and Resource Scheduling
Previous Article in Journal
Garage-Fabricated, Ultrasensitive Capacitive Humidity Sensor Based on Tissue Paper
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Blind Deblurring of Remote-Sensing Single Images Based on Feature Alignment

1
Aerospace Information Research Institute, Chinese Academy of Sciences, No.9 Dengzhuang South Road, Haidian District, Beijing 100094, China
2
School of Optoelectronics, University of Chinese Academy of Sciences, No.19(A) Yuquan Road, Shijingshan District, Beijing 100049, China
3
Department of Key Laboratory of Computational Optical Imagine Technology, CAS, No.9 Dengzhuang South Road, Haidian District, Beijing 100094, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(20), 7894; https://doi.org/10.3390/s22207894
Submission received: 24 August 2022 / Revised: 3 October 2022 / Accepted: 14 October 2022 / Published: 17 October 2022
(This article belongs to the Section Remote Sensors)

Abstract

:
Motion blur recovery is a common method in the field of remote sensing image processing that can effectively improve the accuracy of detection and recognition. Among the existing motion blur recovery methods, the algorithms based on deep learning do not rely on a priori knowledge and, thus, have better generalizability. However, the existing deep learning algorithms usually suffer from feature misalignment, resulting in a high probability of missing details or errors in the recovered images. This paper proposes an end-to-end generative adversarial network (SDD-GAN) for single-image motion deblurring to address this problem and to optimize the recovery of blurred remote sensing images. Firstly, this paper applies a feature alignment module (FAFM) in the generator to learn the offset between feature maps to adjust the position of each sample in the convolution kernel and to align the feature maps according to the context; secondly, a feature importance selection module is introduced in the generator to adaptively filter the feature maps in the spatial and channel domains, preserving reliable details in the feature maps and improving the performance of the algorithm. In addition, this paper constructs a self-constructed remote sensing dataset (RSDATA) based on the mechanism of image blurring caused by the high-speed orbital motion of satellites. Comparative experiments are conducted on self-built remote sensing datasets and public datasets as well as on real remote sensing blurred images taken by an in-orbit satellite (CX-6(02)). The results show that the algorithm in this paper outperforms the comparison algorithm in terms of both quantitative evaluation and visual effects.

1. Introduction

Relying on the advantages of the spatial location of spacecraft platforms, such as satellites and space stations, the use of space optical cameras to observe the Earth has the advantages of a wide range of detection, fast data acquisition, a large amount of information, and less restrictions induced by geographical and natural conditions. It is an important way to obtain information for global change monitoring, regional surveillance, etc. It has been applied in many fields, such as in land and mineral resources management and monitoring, traffic and road network safety monitoring, early warning systems geological disasters, national defense system construction, and cosmic origin research [1]. However, as the demand for the geometric resolution of aerospace optical camera images continues to rise, capturing high-quality images has also encountered great challenges, as highlighted by the problem of avoiding image quality degradation due to image shift blur at high speeds and random vibrations [2].
To reduce the blurring of image shift caused by the high-speed space motion of platforms, such as satellites and space stations, they are usually compensated using a motion compensation mechanism. To ensure that the relative motion of the imaging sensor is less than one-half of an image element or less during the exposure time, this compensation method increases the development cost of the spacecraft. At decimeter resolution, motion compensation accuracy is required to a higher degree, and the corresponding motion compensation mechanism becomes complex [3]. Furthermore, spacecraft vibrations in space cannot be avoided, and micro-vibrations can cause complex imaging blurring when reaching decimeter resolution. High-precision vibration isolation techniques and devices [4] are very complex, costly, and take up substantial weight and space, reducing the cost-effectiveness ratio of the satellite. Therefore, the use of data processing to repair motion blur in remote sensing images can significantly reduce the requirements for satellite motion compensation mechanisms and vibration isolation mechanisms or may take on some of the functions of image shift compensation, improving the efficiency ratio of the spacecraft, which has important application value.
The common formula [5] for nonuniform fuzzy models is as follows:
B = S k ( M ) + N
where B is the blurred image, S is the corresponding clear image, k(M) denotes the unknown blur kernel, M denotes a sparse matrix with each row containing a local fuzzy kernel, ∗ is the discrete convolution, and N is the noise. Deblurring methods are generally divided into two categories: non-blind image deblurring and blind image deblurring. Non-blind deblurring algorithms assume that the blur kernel is known or pre-estimated, whereas blind deblurring algorithms estimate both the blur kernel and the clear image or output the clear image directly. Early conventional deblurring models used natural image priors to design the corresponding algorithms [6,7,8]; however, when dealing with images with null-variant properties, most a priori models do not capture the complex blurring variations in real images well. To ameliorate this problem, a number of deep learning-based deblurring algorithms have been proposed, with neural networks themselves offering unique advantages that are unmatched by traditional algorithms:
  • They generate images directly from the network without estimating fuzzy kernels, avoiding the problem of error superposition during the estimation of fuzzy kernels;
  • There is no need to consider whether the motion blur has a spatially varying property, extending the applicability of the algorithm.
Therefore, from the initial stages of the development of neural network algorithms, researchers have applied neural networks to the problem of processing image deblurring with good results [9].
The recent introduction of generative adversarial networks (GANs) [10] has greatly facilitated the development of deep learning, which has led to breakthroughs in several areas of image processing. In particular, in the field of image restoration, GANs often produce sharper and more plausible textures than feedforward encoders [11]. DeblurGAN [5] introduces GAN to the deblurring task, treating image deblurring as a special image-to-image conversion task, where the model can recover clear and visually appealing images of good quality from both synthetic and real blurred images. DeblurGAN-v2 [11] builds on the success of DeblurGAN by constructing a new cGAN framework. A feature pyramid network (FPN) with multiscale feature fusion was introduced in the generator, global (image) and local (patch) scales were introduced in the discriminator, and RaGAN-LS loss was proposed to improve the deblurring performance and inference efficiency.
However, due to the unlearnable nature of upsampling operations commonly used in FPN (e.g., nearest neighbor interpolation) and the reuse of downsampling and upsampling, there is a problem of feature misalignment between downsampling and upsampling. The unaligned features in turn adversely affect learning in subsequent layers, resulting in the poor recovery of the final blurred image, especially in terms of detail. One approach to solving feature alignment is to perform alignment by explicitly estimating the optical flow field between a reference frame and its neighboring frames [12]. A further approach is to implement implicit motion compensation through circular contrast [13] or deformable convolution [14]. Spatially varying blurred images presents a significant challenge to existing alignment algorithms. For stream-based methods, in particular, stream estimation and image-level warping strategies are very time-consuming and prone to artifacts [14]. In the presence of large image motion, it is difficult to compensate for motion explicitly or implicitly within a single resolution range.
This paper aims to provide another substantial improvement to the cGAN-based motion deblurring task. To address the problem of feature misalignment, feature selection and feature alignment modules are introduced to improve the structure of the generator network, called SDD-GAN, which improves the deblurring performance of the model.
The innovations in this paper are summarized below.
  • Feature importance selection module
Remote sensing images are rich in content, and, to obtain useful features, this paper proposes FISM, which constructs feature attention modules in the spatial and channel domains, explicitly models the interdependence between features, emphasizes lower-level features with rich spatial details, obtains reliable detail information in remote sensing images, achieving the purpose of feature importance selection, and helps to recover details in remote sensing blurred images.
2.
Feature alignment fusion module
Multiscale feature extraction and fusion can effectively deal with the complex and variable motion blur of remote sensing images, but multiscale fusion can result in feature misalignment. This paper proposes FAFM, which learns to align and fuse the feature map of the upsampled channel with the corresponding reference feature map of another channel by adjusting each sampling position in the convolution kernel using the learned offsets. FAFM can better aggregate feature information from different spatiotemporal locations for blurred image recovery.
3.
Color feature loss function
This paper introduces a color loss function to minimize the error between the clear image and the reconstructed deblurred image, thereby forcing the generator to produce an image with the same color distribution in detail as the clear image. The perceptual loss function was also optimized, and the perceptual model VGG19 was pre-trained on ImageNet [15] followed by further training on the RSDATA dataset and the dichotomous GoPro dataset so that the VGG19 network could better obtain the desired image features.
4.
RSDATA dataset
In this paper, according to the imaging mechanism of remote sensing satellites, the types of blurring that may occur and the offset range of pixels in the blurred images are analyzed by detailed calculations. Accordingly, a self-built remote sensing dataset RSDATA is constructed, with the raw data coming from the publicly available datasets UCAS-AOD and DOTA for remote sensing. On the test set, the algorithm in this paper outperformed all comparative algorithms in terms of quantitative evaluation and visual effects, which demonstrates the important role and promise of feature alignment in the field of remote sensing image deblurring.

2. Related Work

2.1. Image Deblurring-Based Deep Learning

Deep learning methods have achieved great success in several image processing tasks including image deblurring. Sun et al. [16] were the first to introduce convolutional neural networks (CNNs) into the field of image deblurring, significantly changing the poor results of fully connected network-based image deblurring algorithms. DeepDeblur [17] proposed a multiscale neural network to iteratively restore clear images, introducing the idea of generative adversarial networks [10]. DeblurGAN [5] proposed an image deblurring algorithm based on generative adversarial networks using gradient penalty and perceptual loss to achieve good recovery of blurred images using the GAN’s better texture generation in the generated images. DeblurGan-V2 [11] performed an optimization of the initial DeblurGAN by introducing a lightweight backbone network to improve the speed of the model and introducing a feature pyramid network [18] into the image deblurring model, which can work flexibly with various backbone networks to achieve a balance between performance and efficiency. SAPHNet [19] proposed an efficient deblurring design built on a new convolutional module that uses global attention and adaptive local filters to learn the transformation of features. MAXIM [20] proposed a multi-axis MLP-based architecture with global/local perceptual fields in each module to improve the learning capability of the model. Ref-MFFDN [21] proposed a new reference-based multilevel feature fusion deblurring network for remote sensing image deblurring in which multi-level features are fused in the same location of clear reference images at different moments in time to extract textures and help recover blurred images. However, these deblurring algorithms do not deal well with the problem of feature misalignment in the process of multi-layer feature fusion.

2.2. Feature Alignment

As the convolutional neural network deepens, the feature map gradually shrinks, causing severe loss of detailed information and leading to severe misalignment during feature fusion. Dai et al. [22] first proposed deformable convolution to learn additional offsets to allow the network to obtain information far from its prescribed local neighborhood, thus improving the ability to convolve. TDAN [14] used deformable convolution to align input frames without explicit motion estimation or image warping. EDVR [23] proposed a deformable convolution-based PCD module for feature alignment, which effectively circumvents the need to compute/estimate image optical flow explicitly or implicitly in traditional alignment methods. FaPN [24] proposed a feature alignment module for dense prediction tasks, which learns the transform offset of pixels, aligns upsampled high-level features according to context, and integrates them in a top-down pyramidal architecture. AlignSeg [25] proposed a two-branch bottom-up network that uses two types of alignment modules to solve the feature misalignment problem prior to feature aggregation, including an aligned feature aggregation (AlignFA) module and an aligned context modeling (AlignCM) module. AlignFA uses a learnable interpolation method to learn the transform offset of pixels, and AlignCM enables each pixel to select its contextual information adaptively so that the contextual information can be aligned and fused, which attempts to directly align features of different proportions.

2.3. Feature Selection

Convolutional neural networks are built on convolutional operations that fuse spatial and channel information within the local receptive field to extract information features. Recent research has shown that the performance of the network can be improved by explicitly embedding learning mechanisms that help capture spatial correlations and improve the representational power of the network without additional supervision. Attention mechanisms have been used with great success in many visual tasks, such as channel attention, spatial attention, temporal attention, and branching attention. In the visual system, attentional mechanisms can be considered as a dynamic selection process, achieved by adaptively weighting features according to the importance of the input.
SENet [26] proposed the channel-attention network, which can adaptively predict potential key features and recalibrate channel feature importance by explicitly modeling the interdependencies between channels. GSoP-Net [27] improved the ability to collect global information through SEBlocks by using global second-order pooling (GSoP), GsoPBlock, but it is computationally intensive. ECANet [28] replaced the fully connected layer in SENet with a one-dimensional convolution. GCT [29] proposed a generic and lightweight change module that combines normalization methods and attention mechanisms to analyze the interrelationships between channels. FcaNet [30] proposed an approach with multispectral channel attention. FaPN [24] proposed a feature selection module (FSM) to explicitly model the importance of feature maps, introducing additional jump connections between input and scaled feature maps for necking (i.e., top-down paths) to enhance multiscale feature aggregation.

3. Network Structure of SDD-GAN

As shown in Figure 1, the SDD-GAN in this paper consisted of two main parts: a generator and a discriminator. The main role of the generator is to generate clear images. The main role of the discriminator is to judge the authenticity of the generated images.

3.1. Generator Network Structure of SDD-GAN

An overview of the architecture of the SDD-GAN generator is shown in Figure 2. SDD-GAN restores a single blurred image to a clear image by training the generator. The generator is the core part of the entire generative adversarial network, and the performance of this part determines the quality of the final recovery of the blurred images.
DeblurGAN-v2 [11] proposed a lightweight alternative to multiscale functionality by introducing the feature pyramid network (FPN) [18] structure into the generator. In this paper, we used a feature pyramid similar to DeblurGAN-v2 as the base structure for feature extraction to build a generator for SDD-GAN, including two channels: a bottom-up channel that downsamples the spatial resolution, and a top-down channel that reconstructs a higher-spatial-resolution image, which can generate multiple feature layers, encode different semantics, contain better-quality feature information, and add lateral connections between the bottom-up and top-down paths to complement the high-resolution detail.
The nearest neighbor upsampling in the top-down channel of DeblurGAN-v2 is modified to a null convolution operation, and a feature importance selection module and a feature alignment fusion module are added to enhance the capability of multiscale feature extraction and aggregation.
Separate feature extraction for images of different spatial resolutions followed by upsampling of high-level feature information, feature selection, and feature alignment followed by feature fusion can well extract the detailed information and global information needed to restore the image, so as to generate a feature map with good semantic information in each spatial scale of the image, and reduce the error or loss of details during feature fusion. Five final feature maps at different scales are selected as stage outputs in the top-down channel. After these features are selected by the feature importance selection module, W and H are upsampled to the same size and stitched into a tensor that contains different levels of semantic information. Two additional upsampling and convolution layers are added at the end of the network to recover the original image size and reduce artifacts. Jump connections are introduced from input to output to improve network performance and convergence, with the input image being normalized to [−1, 1] during training, and the output is kept in the same range using the t a n h activation layer.

3.2. Feature Alignment Fusion Module

FaPN [24] proposed a feature alignment fusion module for dense prediction tasks, which learns the transform offset of pixels by deformable convolution. Inspired by it, for the problem of inaccurate correspondence between FPN downsampled and upsampled features, this paper proposes a feature alignment module based on defuzzification model properties and an improved generator structure.
This module learns to align the feature map of the upsampled channel with the corresponding reference feature map of another channel by adjusting each sampled position in the convolution kernel with the learned offset. This paper further proposes a feature selection module that adaptively emphasizes feature maps that contain rich detail. The two modules are then integrated into a top-down pyramidal architecture that progressively aligns features from the coarsest resolution (top) to the finest resolution (bottom), as shown in Figure 3.
The level I network channel of the bottom-up reduced feature map is defined as F i A , and the level II network channel of the top-down enlarged feature map is defined as F i B . Due to the downsampling operation, there is a predictable spatial misalignment between the upsampled feature map F ˙ i B and the corresponding bottom-up feature map F i 1 A . Therefore, feature fusion by direct correspondence summation of elements or channel cascading compromises the accuracy of feature map detail. Prior to feature fusion, aligning the F i B with its reference F i 1 A is an essential part of the process. The F i B is adjusted according to the spatial location information provided by the F i 1 A . Specifically, the feature map F i A of the level I network has a span of 2i pixels relative to the input image.
F i A H 2 i × W 2 i
where H × W is the size of the input image. For brevity, we denote H / 2 i , W / 2 i as ( h i , w i ) .
The output of the fusion of the first feature map F i + 1 B in the top-down path with the F i A of the level I network is defined as F i B . Specifically, the feature of F i + 1 B after deconvolution is defined as F ˙ i + 1 B , F ˙ i + 1 B , and F i A after the first FISM layer to obtain F ¨ i + 1 B and F ¨ i A . The F ¨ i + 1 B feature is aligned to the F i A feature by calculating the offset between F ¨ i + 1 B and F i A , and the feature map is defined as F i + 1 B . F i + 1 B and F i A go through a second FISM layer for feature selection to obtain the feature maps F ˇ i + 1 B and F ˇ i A . Finally, the feature maps F ˇ i A and F ˇ i + 1 B are fused to obtain F i B .
In the feature alignment module, spatial location information is represented by 2D feature maps, where each offset value can be considered as the offset distance in 2D space between each point in F ˙ i + 1 B and the corresponding point in F i A ; the calculation flow for feature alignment is illustrated in Figure 4.
The feature alignment can be expressed mathematically as follows:
F i + 1 B = f a F ¨ i B , Δ i
Δ i = f o λ B F ¨ i + 1 B , λ A F ¨ i A
λ B F ¨ i + 1 B , λ A F ¨ i A is a cascade of F ¨ i + 1 B and F ¨ i A ( λ A , λ B are hyperparameters), which provides the spatial difference between the corresponding top-down feature F i + 1 B and the corresponding bottom-up feature F i A . f a   , denotes the learning offset Δ i according to the spatial difference between the two feature maps, the two feature maps are concatenated to obtain the offset of the XY-axis through convolution, and the parameters are updated through backpropagation during training. f o   , denotes the function that aligns the feature map in the XY-axis direction according to the learned offset Δ i . In this paper, f a   , and f o   , are implemented using deformable convolution [31].

3.3. Feature Importance Selection Module

In this paper, the FISM module is proposed to improve the generator model, with the data flow shown in Figure 5. The feature attention module is constructed in the spatial and channel domains to explicitly model the interdependencies of features, emphasizing lower-level features with rich spatial detail for feature importance selection. As shown in Figure 2 and Figure 4, the FISM module is added to the four jump connections, as well as the feature alignment fusion module in this paper.
First, the information z w , z c ,   and   z h of each input feature Pi mapping in the spatial and channel domains is extracted through a pooling operation, and then feature importance modeling f w , f c ,   and   f h in each of the three domains (i.e., two 1 × 1 Conv layers with a ReLU activation function in between) is performed, using this information to learn to model the importance of each feature mapping in the three domains to ensure that the network can improve its sensitivity to the informative features and output the corresponding three importance vectors S w , S c , and S h , followed by a sigmoid activation function. Next, the original input feature map is scaled using importance vectors, and the scaled feature map is then added to the original feature map. Finally, a feature selection layer F s   .   (using 1 × 1 Conv to weight each layer of features) is introduced over the rescaled feature map to selectively retain important feature maps and suppress useless feature maps to reduce the amount of information.
The FISM process can be simplified using the following formula:
P i = F s ( P i + S w S c S h P i )
S w = f w ( z w ) S c = f c ( z c ) S h = f h ( z h )
where z a = z 1 , z 2 ,   ,     z D   ,   a W , C , H , and z a is calculated as follows:
z d a = 1 H i a × W i a h a = 1 H i a w a = 1 W i a c d a ( h a , w a )

3.4. Loss Function

To train the algorithm in this paper, it is necessary to compare the image in the training phase, the reconstructed image, and the original image under certain metrics. We designed a loss function L consisting of four components and minimized it during the training of the model, expressed as
L = w g L g + w p L p + w x L x + w c L c
This paper continues the partial loss function of DeblurGAN-v2, choosing the mean square error (MSE) loss as L p , and using the RaGAN-LS [11] loss as the global and local discriminator loss function L g , to make the training process smooth and efficient. D denotes discriminator, G denotes generator, I X denotes the blurred image domain, and I Y denotes the clear image domain, where x is the blurred image and y is the corresponding clear image:
L g = E x ~ I X   ( x ) D ( x ) E y ~ I Y ( y ) D ( G ( y ) ) 1 2 + E y ~ I Y ( y ) D ( G ( y ) ) E x ~ I X   ( x ) D ( x ) + 1 2
Perceptual loss [32] is based on the L2 loss of the CNN feature maps of the generated and target images, which measures the high-level perceptual gap and the semantic gap between the two images and is widely used in image transformation tasks. The perceptual loss used in this paper is the difference between a clear image and a recovered blurred image in the conv3.3 feature map of VGG-19.
L X = 1 W i , j H i , j w = 1 W i , j h = 1 H i , j ϕ i , j Y w , h ϕ i , j G X w , h 2
where i, j denotes the feature map ordinal number of the i-th max-pooling layer followed by the j-th convolution (after activation) within the VGG19 network, ϕ i , j denotes the feature map, and W i , j and H i , j are the dimensions of this feature map.
This paper attempts to pre-train VGG19 on ImageNet, the RSDATA dataset, and the dichotomous GoPro dataset, which allows the VGG19 network to better obtain the desired image features while calculating perceptual losses more efficiently.
For the image transformation task, CycleGAN [33] attempts to add a loss between G(Y) and Y, called constant loss. DeepUEP [34] uses color loss to drive the color of the generated image to match the color of the corresponding clear image. On this basis, this paper introduces a color loss function to minimize the error between the clear image and the reconstructed deblurred image, as a way to force the generator to generate images with the same color distribution in detail as the clear image, with the color feature loss function defined as follows:
L c = E x ~ I x   ( x ) , y ~ I Y   ( y ) , p ( ( G ( x ) ) p , ( y ) p ) + E y ~ I Y   ( y ) G ( y ) y 1
Treating the RGB color space of an image as a 3D vector space, where ( , ) is the operator that calculates the angle between two images in vector space,   . p denotes a pixel point of an image, x denotes a blurred image in domain X, and y denotes a clear image in domain Y corresponding to X. G(x) denotes the image after the blurred image has been deblurred, and G(y denotes the image after the clear image has been deblurred. This paper addresses the problem of detailed color distortion during image deblurring by optimizing the sum of errors per pixel point between the deblurred image and the paired clear image. This paper uses the L1 loss function to optimize the error of the clear image and G(y, which can limit the change interval of the generator in the update iteration and make the details of the generated image of the generator more stable and reliable, while having a certain data-widening ability, thus prompting the generator to better restore the color detail features of the image and improve the quality of the generated image.

3.5. Remote Sensing Image Datasets

Motion blur is a common type of image blur in the field of remote sensing. It mainly stems from the irregular movement of the imaging device during exposure. There are no publicly available datasets for remote sensing image deblurring tasks; hence, this paper analyzes scenes that may cause blurring based on the actual process of remote sensing satellite photography, along with detailed calculations of the types of blurring that may occur and the offset range of pixels in the blurred image. The calculation process is described below.
Assume that the rotation speed of the remote sensing satellite around the Earth is V s , the Earth’s rotation speed is V e , ϑ is the correlation coefficient between the rotation of the Earth and the velocity of the satellite at the time of the satellite shot, the exposure time of the satellite photography is T p , the compensation of the satellite motion is Ψ , the radius of the Earth is R, the distance of the satellite from the surface is r, the resolution of the satellite image is Res, and the error of the satellite motion on the photography image during the exposure time is L M , as shown in Equation (12).
L M = V s + ϑ   V e × R × T p R + r × R e s × Ψ
Camera shake can affect satellite photography and blur the image [35]; in this paper, we only study the blurred phase shift caused by this error and define it as L D . The total error in the exposure time of the ground-based static target is calculated using Equation (13).
L s t a t i c = L D + L M .
Assuming that the velocity of the ground-based moving target is V o b j and that ρ is the correlation coefficient between the velocity of the moving target and the velocity of the satellite, the error caused by the moving target during the shot exposure time is L o b j , as expressed in Equation (14):
L o b j = ρ V o b j × T p R e s × Ψ
The blurred images generated by DeepDeblur [17] exhibit real and spatially varying blur caused by moving people and static backgrounds; by the same token, the error in remote sensing satellite photography can be simplified as
L A L L =               L s t a t i c               ,         b a c k g r o u n d L s t a t i c + L o b j   ,         m o v i n g   o b j e c t   a n d   b a c k g r o u n d  
Let the upper left corner of the remote sensing image be the origin, the horizontal axis be the X-axis, the vertical axis be the Y-axis and the angle between the error vector and the X-axis be θ . The phase shift of pixels on the image can be expressed as
Δ x = L A L L × c o s θ Δ y = L A L L × s i n θ
Remote sensing images taken by satellite have both null-variant and non-null-variant types of blur. This paper constructs a remote sensing deblurring task dataset (RSDATA) on this basis, with raw data collected from the remote sensing open datasets UCAS-AOD [36] and DOTA [37]. The UCAS-AOD dataset was collected from Google Earth and the images for the DOTA dataset were collected from Google Earth, GF-2, and JL-1 satellites provided by the China Resource Center Satellite Data and Applications. The dataset is rich in content and includes multiple regions around the world, including airports, vegetation, ports, and high-density urban areas, scaling all images to 0.5 M resolution based on the GSD originally provided with the dataset. The areas were further cropped to a size of 540 × 540 pixels, overlapping by 10 pixels. To simulate a more realistic and complex motion blur process, the random trajectory generation method proposed in DeblurGAN is used in this paper to generate motion blur kernels to generate blurred images; according to Equation (15), the image blur phase shift size varies randomly in the (5, 41) interval on which based 10% of the data are randomly selected to add local null blur, and the phase shift varies randomly in the (1, 5) interval. Eighty percent of the dataset was used as the training set and 20% was used as the test set. The training set contained 1330 pairs and the test set contained 340 pairs, with each pair consisting of a clear image and a corresponding blurred image. An example of our dataset is shown in Figure 6.

4. Results

To verify the generalizability of the training results, comparative experiments were not only conducted on the self-built remote sensing dataset RSDATA but also on the GoPro dataset [17] and RealBlur dataset [38].
To quantitatively assess the performance of different network models, recovered images were evaluated using the peak signal-to-noise ratio (PSNR) and structural similarity metric (SSIM), which are widely used metrics for image quality assessment.
This paper conducts an ablation study in Section 4.3 to verify the effectiveness of the feature alignment fusion module, explore the impact of adding a feature importance selection module with multiple domains on the algorithm and confirm the effectiveness of the color feature loss function.

4.1. Datasets

  • GoPro dataset: The GoPro dataset [17] is currently one of the most commonly used benchmark datasets in the field of image deblurring based on neural network algorithms. The dataset is derived from the results of frame decomposition of HD video images and is rich in content, covering almost all common image features. The entire dataset has a total of 3214 blurred/clear image pairs, each being 720 × 1280 pixels in size. The dataset itself is divided into three parts: Train, Valid, and Test, with a total of 2103 images in the Train group used for training;
  • RealBlur dataset: RealBlur-R [38] is generated from raw camera images and contains 232 different low-light still scenes with a total of 4556 pairs of blurred and sharp images. The images in the dataset are blurred due to camera shake and captured in dimly lit environments (e.g., streets and indoor rooms at night) to cover the most common cases of motion blur, representing blurred images that are more in line with the real world. RealBlur-R randomly selected 182 scenes as the training set, consisting of 3758 image pairs, with the remaining 50 scenes as the test set, consisting of 980 image pairs;
  • Real satellite images: In-orbit real-world data were obtained from the CX-6(02) optical remote sensing satellite, which was launched on 22 December 2016 with an imaging resolution of 2.8 m in a 700 km orbit. In addition to ground-based remote sensing imaging, experimental verification of image super-resolution processing [39] and image deblurring was carried out using CX-6(02) satellite image data. For the deblurring experiments, the satellite attitude was controlled to obtain image shift blurred image data of approximately 25 pixels, with no clear images for pairing, and this set of image data was used to verify the effectiveness of the deblurring algorithm on the ground;
  • RSDATA: In this paper, a remote sensing deblurring mission dataset was constructed with raw data collected from the remote sensing open datasets UCAS-AOD [36] and DOTA [37], details of which were described in Section 3.5.

4.2. Comparative Results

To evaluate the performance of the algorithm in this paper, the method in this paper was compared with other SOTA deblurring methods on three datasets. All comparison procedures and test results were taken from the authors’ official websites.
RSDATA dataset: This paper quantified the results of DeepDeblur [17], DeblurGAN-v2 [11], DBGAN [40], and MAXIM [20] on a self-constructed RSDATA dataset.
The official websites of the comparison algorithms all provide the model weights of the algorithms on the GoPro dataset. Therefore, in this paper, we first chose to train on the GoPro dataset (trained on GoPro) and test on the RSDATA dataset. The average PSNR and average SSIM of different deblurring algorithms are shown in Table 1. The algorithm in this paper outperformed other deblurring methods in terms of average PSNR and average SSIM. The algorithm in this paper achieved a PSNR of 19.96 dB, which was 1.6 dB better than DeblurGAN-v2.
Figure 7 shows the statistical results of comparing the PSNR and SSIM of the algorithms on the RSDATA test set (trained on GoPro). As shown in Figure 7a, the PSNR metric of the recovered images of this paper’s algorithm was higher than that of the other algorithms at the maximum, minimum, and median values. Figure 7b shows that the images recovered by SDD-GAN not only had higher values than the other deblurring algorithms but also had the smallest range of values for the SSIM metric, demonstrating the robustness of SDD-GAN on the RSDATA test set.
A qualitative assessment of the results of the different algorithms is shown in Figure 8. The remote sensing images recovered by DeepDeblur and DeblurGAN-v2 were blurred and lacked texture. The DBGAN approach could recover clear images, but the recovered images also had ghosting, artifacts, and false textures. MAXIM generated images with a good visual appearance, but the results were still poor when recovering some fine textures (e.g., text on airport runways). In contrast, the method proposed in this paper could produce images with fine textures and realistic visual effects, whether for text, vegetation, or buildings.
Trained and tested on the RSDATA dataset, the algorithm in this paper achieved a PSNR of 39.88 dB and an SSIM of 0.92. Compared with training on GoPro, the test results get a substantial improvement, indicating that the non-remote sensing image dataset GoPro was relatively different from the remote sensing fuzzy dataset; thus, a new remote sensing fuzzy dataset is necessary.
GoPro dataset: The results of this paper were compared quantitatively with other methods on the GoPro dataset, including DeblurGAN [5], DeepDeblur [17], SRN [41], STFAN [42], DeblurGAN-v2 [11], and BlurredItV [43].
The average PSNR and average SSIM of different deblurring algorithms trained and tested on the GoPro dataset are shown in Table 2. The algorithm in this paper outperformed other defuzzification methods in terms of average PSNR and average SSIM. The algorithm in this paper achieved a PSNR of 30.95 dB, which was 1.4 dB better compared to DeblurGAN-v2.
RealBlur-R dataset: This paper compared and analyzed the results of DeepDeblur, DeblurGAN [5], DeblurGAN-v2 [11], SRN [41], DMPHN [44], MAXIM [20], MPRNet [45], and DeepRFT [46] on the RealBlur-R dataset.
Two training methods were used; RealBlur-R (GoPro) indicates training on the GoPro dataset and testing on RealBlur-R (to test generalization to real images); RealBlur-R indicates training and testing on RealBlur-R.
The average PSNR and average SSIM of different deblurring algorithms in the RealBlur-R dataset (GoPro) are shown in Table 3. The algorithm in this paper outperformed other defuzzification methods in terms of average PSNR and average SSIM. The algorithm in this paper achieved a PSNR of 36.08 dB, which was an improvement of 0.82 dB compared to DeblurGAN-v2.
The average PSNR and average SSIM of different deblurring algorithms on the RealBlur-R dataset are shown in Table 4. The algorithm in this paper outperformed other defuzzification methods in terms of average PSNR and average SSIM. The PSNR of the algorithm in this paper reached 39.65 dB, which was an improvement of 3.21 dB compared to DeblurGAN-v2.
The recovery results for the RealBlur-R dataset are shown in Figure 9. In the test results of the RealBlur-R dataset, two typical scenes were selected as subjective visual observations to compare the recovery effect of the algorithm in this paper. These two target image scenes were rich in content, and the images contained distinct edges and complex textures. Therefore, such complex scenes were a good test of the recovery ability of this method. As can be seen from the comparative results in the figure, the algorithm in this paper performed well in these samples and basically recovered the original details and colors of the image.
Real satellite images: The results of Restormer [47], DeblurGAN-v2 [11], DBGAN [40], and MAXIM [20] on real satellite remote sensing data were qualitatively analyzed.
The results of the real satellite image data recovery are shown in Figure 10. The images recovered by DBGAN and Restormer lacked many detailed textures, and DBGAN even generated false textures in some areas. The images recovered by DeblurGan-v2 and MAXIM had good visual effects but suffered from ghosting and unclear detail textures. In contrast, the method proposed in this paper produced visually pleasing images with fine and realistic textures for both vegetation and buildings.
Comparing the results of all algorithms on the RSDATA dataset, GOPRO dataset, RealBlur-R dataset, and real satellite blurred images, it can be seen that the algorithm in this paper had good performance metrics and visual results on all four datasets.

4.3. Ablation Study

4.3.1. Effectiveness of the Feature Alignment Fusion Module

The feature alignment fusion module (FAFM) was removed from the generator structure, and the network was retrained on the RSDATA dataset using the same training scheme to verify the effectiveness of the FAFM.
As can be seen from Table 5, the average PSNR performance was improved from 33.13 to 37.20 and the average SSIM performance was improved from 0.81 to 0.91 after the addition of FAFM, verifying the effectiveness of FAFM in the deblurring process.

4.3.2. Impact of the Feature Importance Selection Module on the Algorithm

The FISM of the feature alignment fusion module (FAFM) was removed from the generator network, and the network was retrained on the RSDATA dataset using the same training scheme to verify the effectiveness of the FISM.
In addition, this paper investigated the impact of feature selection on the generator network when different domains were used for FISM (channel domain (C_FISM), channel domain and one spatial domain (C_H_FISM, C_W_FISM), or channel domain and two spatial domains (C_H_W_FISM)).
As can be seen from Table 6, the average PSNR performance was improved from 35.18 to 37.20 and the average SSIM performance was improved from 0.82 to 0.91 with the addition of FISM, verifying the effectiveness of FISM in the deblurring process.
Table 7 shows the results of the quantitative evaluation of multidomain FISM. The SDD-GAN-recovered images using C_FISM scored the lowest in terms of the mean PSNR and mean SSIM metrics, indicating that the use of multidomain feature importance selection in FISM was effective in improving the quality of the recovered images. When C_H_W_FISM was used, the recovered images obtained high scores on the mean PSNR and mean SSIM metrics.

4.3.3. Validity of the Color Feature Loss Function

Without using the color feature loss function, the network was retrained on the RSDATA dataset using the same training scheme to verify its effectiveness.
L = w g L g + w p L p + w x L x
Table 8 shows the results of the quantitative evaluation. The SDD-GAN-recovered images using the color feature loss function scored higher than those without the color feature loss function in terms of both the mean PSNR and the mean SSIM metrics. The effectiveness of the color feature loss function was demonstrated.

5. Conclusions

For the remote sensing image deblurring task, this paper proposed a new end-to-end generative adversarial network (SDD-GAN) for single-image motion deblurring. The proposed feature alignment module (FAFM) and feature importance selection module (FISM) could successfully recover the texture and details of the blurred remote sensing images. The algorithm performed well on both self-constructed remote sensing datasets and publicly deblurred datasets, and the test comparison results on actual remotely sensed blurred images taken by satellite proved the effectiveness of the method proposed in this paper.
Next, we will conduct research on actual remote sensing blurred images, optimize the multiscale fusion architecture, improve the performance of the model in practical applications, and provide theoretical and experimental support for satellite remote sensing image blur recovery techniques.

Author Contributions

Conceptualization, B.Z. and Z.T.; methodology, B.Z.; software, B.Z.; validation, B.Z.; investigation, B.Z., Y.Y., X.S., Y.Z. and Y.T.; writing—original draft preparation, B.Z.; writing—review and editing, B.Z. and Z.T.; project administration, Q.L.; funding acquisition, Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Key Program Project of Science and Technology Innovation of Chinese Academy of Sciences (No. KGFZD-135-20-03-02) and Strategic Priority Research Program of the Chinese Academy of Sciences (Grant No. XDA28050101).

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Himeur, Y.; Rimal, B.; Tiwary, A.; Amira, A. Using Artificial Intelligence and Data Fusion for Environmental Monitoring: A Review and Future Perspectives. Inf. Fusion 2022, 86–87, 44–75. [Google Scholar] [CrossRef]
  2. Liu, S.; Tong, X.; Li, L.; Ye, Z.; Lin, F.; Zhang, H.; Jin, Y.; Xie, H. Geometric Modeling of Attitude Jitter for Three-Line-Array Imaging Satellites. Opt. Express 2021, 29, 20952–20969. [Google Scholar] [CrossRef] [PubMed]
  3. Li, W.; Hu, C.; Zhang, L.; Yan, C. Drift Angle Compensation Method for a High-Resolution and Wide-Range Space Camera. Measurement 2020, 158, 107710. [Google Scholar] [CrossRef]
  4. Preda, V.; Cieslak, J.; Henry, D.; Bennani, S.; Falcoz, A. Robust Microvibration Mitigation and Pointing Performance Analysis for High Stability Spacecraft. Int. J. Robust Nonlinear Control 2018, 28, 5688–5716. [Google Scholar] [CrossRef]
  5. Kupyn, O.; Budzan, V.; Mykhailych, M.; Mishkin, D.; Matas, J. Deblurgan: Blind Motion Deblurring Using Conditional Adversarial Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8183–8192. [Google Scholar]
  6. Levin, A.; Weiss, Y.; Durand, F.; Freeman, W.T. Understanding Blind Deconvolution Algorithms. IEEE Trans. Pattern Anal. Mach. Intell. 2011, 33, 2354–2367. [Google Scholar] [CrossRef] [PubMed]
  7. Lim, H.; Yu, S.; Park, K.; Seo, D.; Paik, J. Texture-Aware Deblurring for Remote Sensing Images Using ℓ0 -Based Deblurring and ℓ2 -Based Fusion. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 3094–3108. [Google Scholar] [CrossRef]
  8. Zhang, Z.; Zheng, L.; Piao, Y.; Tao, S.; Xu, W.; Gao, T.; Wu, X. Blind Remote Sensing Image Deblurring Using Local Binary Pattern Prior. Remote Sens. 2022, 14, 1276. [Google Scholar] [CrossRef]
  9. Burdziakowski, P. A Novel Method for the Deblurring of Photogrammetric Images Using Conditional Generative Adversarial Networks. Remote Sens. 2020, 12, 2586. [Google Scholar] [CrossRef]
  10. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Nets. Adv. Neural Inf. Process. Syst. 2014, 27, 2672–2680. [Google Scholar]
  11. Kupyn, O.; Martyniuk, T.; Wu, J.; Wang, Z. Deblurgan-v2: Deblurring (Orders-of-Magnitude) Faster and Better. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8878–8887. [Google Scholar]
  12. Chan, K.C.; Wang, X.; Yu, K.; Dong, C.; Loy, C.C. Understanding Deformable Alignment in Video Super-Resolution. In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 2–9 February 2021; Volume 35, pp. 973–981. [Google Scholar]
  13. Isobe, T.; Jia, X.; Gu, S.; Li, S.; Wang, S.; Tian, Q. Video Super-Resolution with Recurrent Structure-Detail Network. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany; pp. 645–660. [Google Scholar]
  14. Tian, Y.; Zhang, Y.; Fu, Y.; Xu, C. Tdan: Temporally-Deformable Alignment Network for Video Super-Resolution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3360–3369. [Google Scholar]
  15. Peng, Y.; Tang, Z.; Zhao, G.; Cao, G.; Wu, C. Motion Blur Removal for Uav-Based Wind Turbine Blade Images Using Synthetic Datasets. Remote Sens. 2021, 14, 87. [Google Scholar] [CrossRef]
  16. Sun, J.; Cao, W.; Xu, Z.; Ponce, J. Learning a Convolutional Neural Network for Non-Uniform Motion Blur Removal. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 769–777. [Google Scholar]
  17. Nah, S.; Hyun Kim, T.; Mu Lee, K. Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 3883–3891. [Google Scholar]
  18. Lin, T.-Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 June 2017; pp. 2117–2125. [Google Scholar]
  19. Suin, M.; Purohit, K.; Rajagopalan, A. Spatially-Attentive Patch-Hierarchical Network for Adaptive Motion Deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 3606–3615. [Google Scholar]
  20. Tu, Z.; Talebi, H.; Zhang, H.; Yang, F.; Milanfar, P.; Bovik, A.; Li, Y. Maxim: Multi-Axis Mlp for Image Processing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 5769–5780. [Google Scholar]
  21. Li, Z.; Guo, J.; Zhang, Y.; Li, J.; Wu, Y. Reference-Based Multi-Level Features Fusion Deblurring Network for Optical Remote Sensing Images. Remote Sens. 2022, 14, 2520. [Google Scholar] [CrossRef]
  22. Dai, J.; Qi, H.; Xiong, Y.; Li, Y.; Zhang, G.; Hu, H.; Wei, Y. Deformable Convolutional Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 764–773. [Google Scholar]
  23. Wang, X.; Chan, K.C.; Yu, K.; Dong, C.; Change Loy, C. Edvr: Video Restoration with Enhanced Deformable Convolutional Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  24. Huang, S.; Lu, Z.; Cheng, R.; He, C. FaPN: Feature-Aligned Pyramid Network for Dense Image Prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11 October 2021; pp. 864–873. [Google Scholar]
  25. Huang, Z.; Wei, Y.; Wang, X.; Liu, W.; Huang, T.S.; Shi, H. Alignseg: Feature-Aligned Segmentation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 44, 550–557. [Google Scholar] [CrossRef]
  26. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  27. Gao, Z.; Xie, J.; Wang, Q.; Li, P. Global Second-Order Pooling Convolutional Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 3024–3033. [Google Scholar]
  28. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  29. Yang, Z.; Zhu, L.; Wu, Y.; Yang, Y. Gated Channel Transformation for Visual Recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 11794–11803. [Google Scholar]
  30. Qin, Z.; Zhang, P.; Wu, F.; Li, X. Fcanet: Frequency Channel Attention Networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 11 October 2021; pp. 783–792. [Google Scholar]
  31. Zhu, X.; Hu, H.; Lin, S.; Dai, J. Deformable Convnets v2: More Deformable, Better Results. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 9308–9316. [Google Scholar]
  32. Johnson, J.; Alahi, A.; Li, F.F. Perceptual Losses for Real-Time Style Transfer and Super-Resolution. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Berlin/Heidelberg, Germany; pp. 694–711. [Google Scholar]
  33. Zhu, J.-Y.; Park, T.; Isola, P.; Efros, A.A. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2223–2232. [Google Scholar]
  34. Wang, R.; Zhang, Q.; Fu, C.-W.; Shen, X.; Zheng, W.-S.; Jia, J. Underexposed Photo Enhancement Using Deep Illumination Estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6849–6857. [Google Scholar]
  35. Haghshenas, J. Vibration Effects on Remote Sensing Satellite Images. Adv. Aircr. Spacecr. Sci. 2017, 4, 543. [Google Scholar]
  36. Zhu, H.; Chen, X.; Dai, W.; Fu, K.; Ye, Q.; Jiao, J. Orientation Robust Object Detection in Aerial Images Using Deep Convolutional Neural Network. In Proceedings of the 2015 IEEE International Conference on Image Processing (ICIP), Quebec City, QC, Canada, 27–30 September 2015; pp. 3735–3739. [Google Scholar]
  37. Xia, G.-S.; Bai, X.; Ding, J.; Zhu, Z.; Belongie, S.; Luo, J.; Datcu, M.; Pelillo, M.; Zhang, L. DOTA: A Large-Scale Dataset for Object Detection in Aerial Images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 3974–3983. [Google Scholar]
  38. Rim, J.; Lee, H.; Won, J.; Cho, S. Real-World Blur Dataset for Learning and Benchmarking Deblurring Algorithms. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany; pp. 184–201. [Google Scholar]
  39. Tan, Z.; Lyu, Q.; Sun, J.; Wang, J.; Zhao, N. Super-Resolution Imaging Design for CX6-02 Micro-Satellite. In Proceedings of the Image and Signal Processing for Remote Sensing XXV, SPIE, Strasbourg, France, 9–11 September 2019; Volume 11155, pp. 107–114. [Google Scholar]
  40. Zhang, K.; Luo, W.; Zhong, Y.; Ma, L.; Stenger, B.; Liu, W.; Li, H. Deblurring by Realistic Blurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2737–2746. [Google Scholar]
  41. Tao, X.; Gao, H.; Shen, X.; Wang, J.; Jia, J. Scale-Recurrent Network for Deep Image Deblurring. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8174–8182. [Google Scholar]
  42. Zhou, S.; Zhang, J.; Pan, J.; Xie, H.; Ren, J. Spatio-Temporal Filter Adaptive Network for Video Deblurring. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Korea, 27–28 October 2019. [Google Scholar]
  43. Purohit, K.; Shah, A.; Rajagopalan, A. Bringing Alive Blurred Moments. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 6830–6839. [Google Scholar]
  44. Zhang, H.; Dai, Y.; Li, H.; Koniusz, P. Deep Stacked Hierarchical Multi-Patch Network for Image Deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019; pp. 5978–5986. [Google Scholar]
  45. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H.; Shao, L. Multi-Stage Progressive Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 14821–14831. [Google Scholar]
  46. Mao, X.; Liu, Y.; Shen, W.; Li, Q.; Wang, Y. Deep Residual Fourier Transformation for Single Image Deblurring. arXiv 2021, arXiv:2111.11745. [Google Scholar]
  47. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.-H. Restormer: Efficient Transformer for High-Resolution Image Restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–20 June 2022; pp. 5728–5739. [Google Scholar]
Figure 1. The SDD-GAN framework.
Figure 1. The SDD-GAN framework.
Sensors 22 07894 g001
Figure 2. SDD-GAN generator structure diagram.
Figure 2. SDD-GAN generator structure diagram.
Sensors 22 07894 g002
Figure 3. Diagram of the data flow of FAFM.
Figure 3. Diagram of the data flow of FAFM.
Sensors 22 07894 g003
Figure 4. Schematic of offset calculation in FAFM.
Figure 4. Schematic of offset calculation in FAFM.
Sensors 22 07894 g004
Figure 5. Diagram of the data flow of FISM.
Figure 5. Diagram of the data flow of FISM.
Sensors 22 07894 g005
Figure 6. Example graph of RSDATA dataset.
Figure 6. Example graph of RSDATA dataset.
Sensors 22 07894 g006
Figure 7. Distribution of the mean PSNR and SSIM for different deblurring algorithms on the RSDATA test set (trained on GoPro). The box plots show the maximum, upper quartile, median, lower quartile, and minimum values of the results from top to bottom.
Figure 7. Distribution of the mean PSNR and SSIM for different deblurring algorithms on the RSDATA test set (trained on GoPro). The box plots show the maximum, upper quartile, median, lower quartile, and minimum values of the results from top to bottom.
Sensors 22 07894 g007
Figure 8. Results of different algorithms for image recovery in the RSDATA dataset.
Figure 8. Results of different algorithms for image recovery in the RSDATA dataset.
Sensors 22 07894 g008
Figure 9. Comparison of recovery results for the RealBlur-R dataset.
Figure 9. Comparison of recovery results for the RealBlur-R dataset.
Sensors 22 07894 g009
Figure 10. Test results for real satellite remote sensing blurred image.
Figure 10. Test results for real satellite remote sensing blurred image.
Sensors 22 07894 g010
Table 1. Average PSNR and SSIM for different deblurring methods on the RSDATA dataset (trained on the GoPro dataset). The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
Table 1. Average PSNR and SSIM for different deblurring methods on the RSDATA dataset (trained on the GoPro dataset). The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
TrainTestMethodPSNRSSIM
GoProRSDATADeepDeblur [17]18.850.64
GoProRSDATADeblurGAN-v2 [11]18.360.60
GoProRSDATADBGAN [40]18.670.62
GoProRSDATAMAXIM [20]18.860.64
GoProRSDATASDD-GAN19.960.71
Table 2. Average PSNR and SSIM for different deblurring methods on the GoPro dataset. The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
Table 2. Average PSNR and SSIM for different deblurring methods on the GoPro dataset. The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
MethodPSNRSSIM
DeblurGAN [5]28.700.85
DeepDeblur [17]29.080.91
SRN [41]30.260.93
DeblurGAN-v2 [11]29.550.93
STFAN [42]28.590.86
BlurredItV [43]30.580.94
SDD-GAN30.950.94
Table 3. Average PSNR and SSIM for different deblurring methods on the RealBlur-R dataset (GoPro). The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
Table 3. Average PSNR and SSIM for different deblurring methods on the RealBlur-R dataset (GoPro). The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
MethodPSNRSSIM
DeepDeblur [17]32.510.84
DeblurGAN [5]33.790.90
DeblurGAN-v2 [11]35.260.94
SRN [41]35.660.94
DMPHN [44]35.700.94
MAXIM [20]35.780.94
MPRNet [45]35.990.95
DeepRFT [46]36.060.95
SDD-GAN36.210.96
Table 4. Mean PSNR and SSIM for different deblurring methods on the RealBlur-R dataset. The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
Table 4. Mean PSNR and SSIM for different deblurring methods on the RealBlur-R dataset. The highest scores are highlighted in red, and the second highest scores are highlighted in blue.
MethodPSNRSSIM
SRN [41]38.650.96
DeblurGAN-v2 [11]36.440.93
MAXIM [20]39.450.96
SDD-GAN39.650.97
Table 5. Mean PSNR and SSIM (with FAFM vs. without FAFM) on the RSDATA dataset. The highest scores are highlighted in red.
Table 5. Mean PSNR and SSIM (with FAFM vs. without FAFM) on the RSDATA dataset. The highest scores are highlighted in red.
MethodPSNRSSIM
SDD-GAN without FAFM33.130.81
SDD-GAN37.200.91
Table 6. Mean PSNR and SSIM (with FISM vs. without FISM) on the RSDATA dataset. The highest scores are highlighted in red.
Table 6. Mean PSNR and SSIM (with FISM vs. without FISM) on the RSDATA dataset. The highest scores are highlighted in red.
MethodPSNRSSIM
SDD-GAN without FIFM35.180.82
SDD-GAN37.200.91
Table 7. Average PSNR and SSIM on the RSDATA dataset (using C_FISM, C_W_FISM, C_H_FISM, or C_H_W_FISM). The highest scores are highlighted in red.
Table 7. Average PSNR and SSIM on the RSDATA dataset (using C_FISM, C_W_FISM, C_H_FISM, or C_H_W_FISM). The highest scores are highlighted in red.
MethodPSNRSSIM
SDD-GAN + C_FISM35.510.87
SDD-GAN + C_W_FISM35.690.90
SDD-GAN + C_H_FISM35.740.90
SDD-GAN + C_H_W_FISM37.200.91
Table 8. Average PSNR and SSIM (with LC vs. without LC) on the RSDATA dataset. The highest scores are highlighted in red.
Table 8. Average PSNR and SSIM (with LC vs. without LC) on the RSDATA dataset. The highest scores are highlighted in red.
MethodPSNRSSIM
SDD-GAN without Lc37.200.91
SDD-GAN with Lc39.880.92
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, B.; Lv, Q.; Yang, Y.; Sui, X.; Zhang, Y.; Tang, Y.; Tan, Z. Blind Deblurring of Remote-Sensing Single Images Based on Feature Alignment. Sensors 2022, 22, 7894. https://doi.org/10.3390/s22207894

AMA Style

Zhu B, Lv Q, Yang Y, Sui X, Zhang Y, Tang Y, Tan Z. Blind Deblurring of Remote-Sensing Single Images Based on Feature Alignment. Sensors. 2022; 22(20):7894. https://doi.org/10.3390/s22207894

Chicago/Turabian Style

Zhu, Baoyu, Qunbo Lv, Yuanbo Yang, Xuefu Sui, Yu Zhang, Yinhui Tang, and Zheng Tan. 2022. "Blind Deblurring of Remote-Sensing Single Images Based on Feature Alignment" Sensors 22, no. 20: 7894. https://doi.org/10.3390/s22207894

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop