Next Article in Journal
Model Predictive Control with Variational Autoencoders for Signal Temporal Logic Specifications
Previous Article in Journal
Electrodermal Activity (EDA) Morphologies and Prediction of Engagement with Simple Moving Average Crossover: A Mixed-Method Study
Previous Article in Special Issue
A Novel Adversarial Deep Learning Method for Substation Defect Image Generation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Analysis of the Generalization Ability of Defogging Algorithms on RICE Remote Sensing Images

1
School of Physics and Telecommunication Engineering, Shaanxi University of Technology, Hanzhong 723001, China
2
Trine Engineering Institute, Shaanxi University of Technology, Hanzhong 723001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(14), 4566; https://doi.org/10.3390/s24144566
Submission received: 14 May 2024 / Revised: 1 July 2024 / Accepted: 12 July 2024 / Published: 14 July 2024
(This article belongs to the Special Issue AI-Driven Sensing for Image Processing and Recognition)

Abstract

:
This paper explores the generalization ability of defogging algorithms on RICE (A Remote Sensing Image Dataset for Cloud Removal) remotely sensed images. RICE is a dataset of remotely sensed images used for removing clouds, allowing the researcher to better evaluate the performance of defogging algorithms for cloud removal from remotely sensed images. In this paper, four classical defogging algorithms, including AOD-Net, FFA-Net, dark channel prior, and DehazeFormer, are selected and applied to the task of de-cloud in RICE remote sensing images. The performance of these algorithms on the RICE dataset is analyzed by comparing the experimental results, and their differences, advantages, and disadvantages in dealing with de-clouded remote sensing images are explored. The experimental results show that the four defogging algorithms are capable of performing well on uniform thin cloud images, but there is a color distortion and the performance is weak when it comes to inhomogeneous clouds as well as thick clouds. So, the generalization ability of the algorithms is weak when the defogging algorithms are applied to the problem of cloud removal. Finally, this paper proposes improvement ideas for the de-cloud problem of RICE remote sensing images and looks forward to possible future research directions.

1. Introduction

In today’s digital age, remote sensing technology has become one of the most important means of obtaining information about the Earth’s surface. Through remote sensing images, a variety of information about the Earth’s surface can be obtained, such as topography and geomorphology, vegetation cover, land use, and so on, providing important data support for resource management, environmental monitoring, urban planning, and other fields. However, due to the atmospheric conditions, remote sensing images are often affected by atmospheric disturbances such as haze, rain, fog, clouds, etc., resulting in image quality degradation and blurred information, which reduces their value and usability in practical applications [1].
In recent years, with the rapid development of artificial intelligence technologies such as deep learning, defogging algorithms have received widespread attention as an image enhancement technique [2]. These algorithms effectively remove disturbances such as haze and clouds by modeling and processing factors such as atmospheric scattering and lighting conditions, thus improving the visual quality of remote sensing images. However, most of the current defogging algorithms are designed for specific scenes or specific types of images. For example, the defogging algorithm proposed by Li et al. targets a variety of underwater images [3], the defogging algorithm proposed by Liu et al. targets images containing a large number of sky regions [4], and the algorithm proposed by Li et al. targets images of nighttime haze [5], while in practical applications the scenes and conditions of remote-sensing images are often diverse. Therefore, how to evaluate the generalization ability of defogging algorithms in different scenes has become an important topic in current research.
Among the many remote sensing image-defogging algorithms, algorithms such as AOD-Net (All-in-One Dehazing Network) [6], FFA-Net (Feature Fusion Attention Network for Single Image Dehazing) [7], dark channel a prior (Single Image Haze Removal Using Dark Channel Prior) [1], and DehazeFormer (Vision Transformers for Single Image Dehazing) [8] have achieved some success. These algorithms show different advantages in different scenarios; for example, AOD-Net and FFA-Net have better defogging effects in complex scenarios, the dark channel a prior algorithm has strong robustness, and DehazeFormer has high efficiency in dealing with large-scale data. In addition, some other defogging algorithms, such as GMAN (Single Image Dehazing with a Generic Model-Agnostic Convolutional Neural Network) [9], cycle dehaze (Cycle-Dehaze: Enhanced CycleGAN for Single Image Dehazing) [10], HyLoG-ViT (Complementary Feature Enhanced Network with Vision Transformer for Image Dehazing) [11], etc., have also achieved some results in the field of image defogging. However, there is a lack of in-depth comparisons and analyses of these algorithms for specific scenarios on specific datasets, especially for the de-clouding effect of remotely sensed images.
To fill this research gap, the RICE (A Remote Sensing Image Dataset for Cloud Removal) dataset is selected as the object of study in this paper and the performances of the AOD-Net, FFA-Net, dark channel a prior, and DehazeFormer algorithms on this dataset are comparatively analyzed.
The defogging algorithm based on the dark channel prior is an algorithm in the statistical sense. The dark channel defogging algorithm is an image processing technique based on the a prior theory of dark channels to recover a clear image from a hazy image. The core of the algorithm is to compute the dark channel of an image, which is the smallest value among multiple color channels that is used to represent the darkness of a pixel. The dark channel is then utilized to estimate the transmittance and atmospheric light values to remove the effects of haze. Dr. He summarized a large number of outdoor fog-free images and found that there are some pixels in the local region of the fog-free image. At least one of these pixels has a very low brightness value for at least one of the color channels (the region of the low brightness value excludes the sky region, for example). In the field of computer vision and computer graphics, a widely used formula to describe foggy images [12,13] is:
I x = J x t x + A 1 t x
where I x denotes the foggy image, J x is the fog-free image to be recovered, A is the global atmospheric light composition, and t x is the transmittance. By using this equation, the goal is to compute the original fog-free image, the transmittance, and the estimation of the global atmospheric light composition from the available images.
The AOD-Net defogging algorithm is a deep learning-based approach to defogging that utilizes the powerful feature learning and mapping capabilities of neural networks. Conventional defogging methods typically rely on physical models and prior knowledge to estimate fog concentrations and attempt to remove haze. However, these methods usually require manual parameter tuning and have limited effectiveness in complex situations. Therefore, in recent years, the development of deep learning has provided new ways to solve these problems. Most deep learning methods for image restoration and enhancement fully incorporate end-to-end modeling: training models to regress clear images directly from corrupted ones. Examples include image denoising, deblurring, and super-resolution. In contrast, there is currently no end-to-end deep modeling for defogging that regresses clear images directly from blurred models and current models for defogging lack articulation with advanced vision tasks. By training a large number of fogged and clear image pairs, AOD-Net is able to learn how to convert fogged images into clear ones for efficient defogging. It adopts an end-to-end training approach, outputting clear images directly from the input fogged images, avoiding the tedious process of manually adjusting parameters. In addition, AOD-Net can be seamlessly embedded with other deep models to perform advanced tasks on foggy images [6].
The emergence of the FFA-Net defogging algorithm is mainly based on the development of deep learning techniques, the importance of image defogging, and the limitations of traditional methods. Image defogging is an important and challenging problem in the field of computer vision [14], and in recent years, with the rapid development and application of deep learning techniques, more and more image processing tasks have begun to adopt deep learning methods. These methods are usually able to train on large-scale datasets and achieve good performance on various vision tasks.
Image defogging, as an important image processing task, has received a lot of attention from both academia and industry. Clear images are crucial for many applications, such as unmanned vehicles, surveillance systems, satellite images, etc. Therefore, improving the effectiveness of image defogging has become an important research direction.
Traditional image defogging methods are usually based on physical models or a prior knowledge, which can improve image quality to a certain extent but often do not handle complex scenes well or provide high-quality results. As a result, researchers have begun to explore deep-learning-based approaches to address this problem. In the field of deep learning, feature fusion and attention mechanisms are widely used in a variety of tasks to enhance the network’s representation and improve task performance [15]. The development of these techniques provides new ideas and methods for image-defogging tasks. By combining feature fusion and attention mechanism, FFA-Net can effectively improve the effect of image defogging and provide a new solution to solve the image quality problem in practical applications.
DehazeFormer is an image-defogging method based on the Transformer architecture. The Transformer [16] model was originally proposed as a revolutionary model in the field of natural language processing but has since been shown to have broad applicability in other fields as well, including image processing. In image processing tasks, especially at the pixel level (e.g., image segmentation, denoising, etc.), the Transformer model has been shown to achieve impressive results. Therefore, it is a natural extension for Transformer-based algorithms to be used for tasks like image defogging.
The traditional defogging algorithm and CNN-based defogging algorithm methods have some limitations, such as the dependence on a prior and the inability to deal with high-frequency information. In recent years, with the rise of deep learning in the field of computer vision, image-defogging methods based on deep convolutional neural networks (CNNs) have made significant progress. However, recent studies have shown that the Transformer architecture outperforms most CNN architectures in advanced visual tasks, which has prompted researchers to start exploring the possibility of applying the Transformer in the field of image defogging [17]. DehazeFormer was inspired by the Swin Transformer [18], and through a series of improvements, it significantly outperforms traditional CNN-based methods, especially in processing high-frequency information.
The RICE dataset is a remote sensing image dataset for cloud removal that contains real remotely sensed cloudy images as well as the corresponding remotely sensed cloud-free images, with certain representativeness and practicability [19]. By comparing the experimental results, this paper will evaluate the advantages and disadvantages of each algorithm in terms of de-clouding effect, computational complexity, and applicable scenes and provide an important reference for the research and application of de-clouding algorithms for remote sensing images.

2. Materials and Methods

2.1. Dataset

To verify the effectiveness of the algorithm, two benchmark datasets from RICE were used for the experiments:
  • RICE-I dataset. The RICE-I dataset contains 500 pairs of images, each pair of images has one image with clouds and one without clouds, which were obtained by setting the cloud display parameters on Google Earth. Subsequently, these images were cropped to a size of 512 × 512 pixels and did not contain overlapping regions, as shown in Figure 1
  • RICE-II dataset. The RICE-II dataset is derived from the Landsat 8 OLI_TIRS dataset and utilizes LandsatLook imagery from the US National Geological Survey Earth Explorer database. The dataset uses geo-referenced Landsat-Look imagery, both natural color (landsat8 OLI, bands 6, 5, 4) and high-quality imagery (8-bit files generated from Landsat Level-1 quality bands). The cloud images were produced by acquiring real remote sensing images in the same area at least every 15 days. All images were cropped to a size of 512 × 512 pixels and did not contain overlapping areas. The dataset contains 450 pairs of images, each of which is a reference image without clouds, an image with clouds, and its cloud MASK map. As shown in Figure 2.

2.2. Methods

2.2.1. Dark Channel Prior

  • Theory
In the sky-free region of most fog-free images, there is at least one color channel in the pixels that has a very low luminance value. This lowest luminance value is almost equal to 0. Thus, for an image J , the dark channel is:
J d a r k x = m i n c r , g , b m i n y Ω x J c y
where J c is the color channel of J and Ω x is the local block centred at x. Observations show that the intensity of J d a r k is low and tends to zero except in the sky region if J is a fog-free outdoor image. The dark channel of J is referred to as J d a r k and the above statistical observation or knowledge is referred to as the dark channel prior.
The low brightness in the dark channel comes from several sources:
  • Shadows
  • Colored objects or surfaces
  • Black objects or surfaces
2.
Algorithm design
Assuming first that the atmospheric light component A is known, the defogging model Equation (1) can be reduced to the following equation by dividing both sides of the equation by A at the same time:
I c x A c = t x J c x A c + 1 t x
The superscript c denotes the RGB three-channel, further assuming that the transmittance t x within each filter window is constant, denoted as t ¯ x , and then calculating the dark channel simultaneously for both sides of the equation, i.e., making two minima operations, the following equation can be obtained:
m i n y Ω x m i n c I c y A c = t ¯ x m i n y Ω x m i n c J c y A c + 1 t ¯ x
Since t ¯ x is a constant, it is placed outside the minimum operation, and then, according to the dark-channel prior, J d a r k tends to zero, i.e.:
J d a r k x = m i n y Ω x m i n c J c y = 0
Since A c is always positive, we can obtain:
m i n y Ω x m i n c J c y A c = 0
Substituting Equation (6) into Equation (4) yields an estimate of the transmittance.
t ¯ x = 1 m i n y Ω x m i n c I c y A c
However, in reality, even on a clear day, the atmospheric light component is still present. Especially when it comes to looking at objects in the distance, you will notice that the fog is still there, and in addition, the presence of fog is a fundamental clue to human depth perception [20,21]. If the fog is completely removed, the image may look unnatural and the sense of depth may be lost, therefore, a constant parameter ω ( 0 < ω < 1 ) is introduced to control the degree of defogging. Inside the original paper the authors suggested that ω take 0.95.
t ¯ x = 1 ω m i n y Ω x m i n c I c y A c
Before, it was assumed that A is a known quantity, but the value of A can be approximated by finding the brightest points in the first 0.1% of the dark channels, i.e., the points with the smallest transmittance; for these points, go to the fog map to find the corresponding points and take the largest value of all the channels in them as an approximation to A. At this point, the transmittance t x , the atmospheric light component A, and the fog map I x , are all known and the fog-free map J x can be solved for:
J x = I x A max t x , t 0 + A
where t 0 is a transmittance lower bound. Since the direct recovery when the transmittance t x is close to zero from Equation (1), J x   t x is also zero, which will lose the original image information and makes it easy to introduce noise, so a lower bound is set to retain a certain amount of fog in the place where the density of fog is very high. Inside the original paper, the authors suggested a value of 0.1 for t 0 because the image after defogging will generally look darker, so the exposure can be increased appropriately to obtain better results.

2.2.2. AOD-Net

  • Theory
AOD-Net (All-In-One Dehazing Network) is a deep learning image-defogging algorithm based on a reformulated atmospheric scattering model that is based on the principle of using convolutional neural networks (CNN) and an end-to-end training approach. The core idea of the algorithm is to learn the mapping from foggy images to clear images by training an end-to-end neural network using the feature learning capability of deep learning. During the training of the network, supervised learning is performed using a large number of foggy and clear image pairs, enabling the network to learn the complex mapping relationship between the input image and the output image.
In AOD-Net, a convolutional neural network is first used to learn the features in a foggy image, which include the distribution of haze, the contrast of the image, and so on. Then, the network extracts and abstracts these features through a series of convolution and pooling operations and gradually converts them into a representation of a clear image. Finally, after processing in the output layer of the network, the difference between the generated image and the clear image is minimized, thus achieving the defogging effect.
This end-to-end training approach allows AOD-Net to automatically learn the process of image defogging without the need to manually adjust parameters or rely on prior knowledge. This enables AOD-Net to achieve good defogging results in a variety of different scenarios with strong generalization capabilities.
2.
Algorithm design
Firstly, through the atmospheric scattering model in (1), both sides are simultaneously divided by t x into Equation (10):
J x = 1 t x I x A 1 t x + A
Previous methods such as [2,22] estimated t x and A , respectively, and obtained clear images by (10). Instead of directly minimizing the reconstruction error on J x , they go on to optimize t x . This indirect optimization leads to sub-optimal solutions, so the core idea here is to integrate the two parameters t x and A into a single formula, K x in (11), and to directly minimize the reconstruction error in the pixel domain of the image. Therefore, the formula in (10) is re-expressed as:
J x = K x I x K x + b , w h e r e K x = 1 t x I x A + A b I x 1
In this way, K x integrates t x and A . Here, b is a constant bias with a default value of 1. As can be seen, K x is also dependent on I x and, therefore, an input-adaptive depth model needs to be built and the model needs to be trained by minimizing the error in reconstruction between its output J x and the reconstruction error of the real clear image to train the model.

2.2.3. FFA-Net

  • Theory
FFA-Net is a feature fusion attention network that fuses different types of information through a feature attention module and can process different types of information more flexibly. The structure of this network includes a shallow feature extraction part, a Group Architecture with multiple jump connections [7], a feature attention module, a reconstruction part, and a global residual learning structure.
The feature attention module consists of channel attention and pixel attention, which can provide additional flexibility to process different types of information. Channel attention focuses on the weighted information of different channel features and obtains the weights of different channels through global average pooling and convolutional layers. Feature fusion attention fuses features by splicing the feature maps output from Group Architecture in the channel direction and obtaining adaptive learning weights through the feature attention mechanism.
FFA-Net extracts features from the input haze-affected images using a deep convolutional neural network (CNN). These features usually contain semantic information, edge information, etc., of the image.
FFA-Net uses the technique of feature fusion to fuse features from different layers. The aim is to enhance the representational capabilities of the network so that it can better understand the structure and content of the image.
FFA-Net introduces an attention mechanism for dynamically adjusting how much attention the network pays to different features. This means that the network can automatically focus on feature regions that are more important to the image-defogging task when processing images, thus improving the defogging effect. The attention mechanism allows the network to pay more attention to important pixels or regions when processing different pixels, thus reducing the impact of irrelevant information.
Finally, after the feature fusion and attention mechanism, FFA-Net reconstructs the processed feature maps into a final defogged image. This process is usually achieved by operations such as deconvolution or upsampling.
2.
Algorithm design
The algorithm design of FFA-Net consists of the following key elements: the feature attention module, basic block structure, feature fusion structure, and loss function.
The feature attention module consists of a channel attention module and a pixel attention module (Figure 3). The channel attention mainly focuses on the weighting information of different channel features for image defogging, and the global spatial information related to the channel is put into the channel descriptor by using global average pooling.
g c = H p F c = 1 H × W i = 1 H j = 1 W X c i , j
where X c i , j denotes the value of the cth channel X c at position i , j and H p is the global pooling function. To obtain the weights of the different channels, the features are passed through two convolutional layers and sigmoid [23] and ReLu [24] activation functions.
C A c = σ C o n v δ C o n v g c
where σ is the sigmoid function and δ is the ReLu function. Finally, the input F C is multiplied element-by-element with the weights of the channel C A c .
F C = C A c F c
Pixel attention is similar to channel attention in that we directly feed the input F (the output of the C A ) into two convolutional layers with ReLu and sigmoid activation functions.
The basic block structure consists of local residual learning and feature attention (FA) modules (Figure 4).
The feature fusion architecture first stitches all the feature maps output by multiple Group Architectures in the channel direction and then perform feature fusion by the adaptive learning weights obtained from the Feature Attention (FA) mechanism.
The loss function uses a simple L1 loss function, and the network is trained by optimizing the L1 loss function to improve performance.
L Θ = 1 N i = 1 N I g t i F F A I h a z e i
where Θ denotes the parameters of the FFA-Net network, I g t denotes the ground truth, and I haze denotes the input.

2.2.4. DehazeFormer

  • Theory
The principle of the DehazeFormer algorithm is an image-defogging method based on Transformer architecture. Inspired by Swin Transformer, DehazeFormer surpasses the traditional CNN-based methods through a series of improvements.
DehazeFormer has been designed with some key improvements, including removing the LayerNorm [25], which is commonly used in Vision Transformer [17], and proposing RescaleNorm in place of LayerNorm to better preserve the correlation between features.
For the choice of activation function, the original authors of DehazeFormer proposed SoftReLU, which performs better in the image-defogging task compared to SiLU/Swish [26] and GELU [27]. For window segmentation in Transformer, DehazeFormer proposes an offset window segmentation scheme based on reflection filling and cropping. In terms of spatial information aggregation, DehazeFormer introduces the scheme of W-MHSA combined with parallel convolution to improve the performance of the network by aggregating information within a window while using convolution to aggregate information within a neighborhood. A prior soft reconstruction module, which outperforms global residual learning and SKNet-based multi-scale feature map fusion module [28], is also introduced to replace the cascade fusion.
2.
Algorithm design
The algorithm design of DehazeFormer includes the following key elements: a multi-scale feature map fusion module and a prior soft reconstruction module, normalization substitution, activation function selection, and W-MHSA combined with parallel convolution.
Firstly in the SKNet-based multi-scale feature map fusion module, to first assume that the two features are mapped as x 1 and x 2 , using a linear layer f (·) to project x 1 to x ^ 1 , we use global average pooling GAP , F M L P , the softmax function, and the splitting operation to obtain the fusion weights:
{ a 1 , a 2 } = Split   ( Softmax F M L P GAP x ^ 1 + x 2
Use weights { a 1 , a 2 } to fuse x ^ 1 and x 2 with additional short residuals via y = a 1 x ^ 1 + a 2 x 2 + x 2 .
The a-prior-based soft reconstruction module is a reconstruction of Equation (1):
J x = K x I x + B x + I x
where K ( x ) = 1 / t ( x ) 1 ,   B ( x ) = 1 / t ( x ) 1 A .
RescaleNorm is an improvement on LayerNorm; the normalization process of LayerNorm can be represented as:
x ^ i = x i μ i σ i γ i + β i
Here, μ and σ denote the mean and standard deviation, γ and β are the learned scale factor and bias, and i = ( i b , i n , i c ) denotes the index. Firstly, we first obtain μ , σ R b × 1 × 1 and normalize the input feature x to x ^ by Equation (18); then, we process x ^ to obtain the output y ^ by using the F(·) module. We use two linear layers with weights W γ , W β R 1 × c and bias B γ , B β R ~ 1 × 1 × c , which is passed through γ , β = σ W γ + B γ , μ W β + B β to change the values of µ and σ.
y = F x μ σ γ + β σ W γ + B γ + μ W β + B β
where γ , β R b × 1 × c .
The algorithm design of SoftReLU is an approximation of ReLU [29].
SoftReLU x = x + x 2 + α 2 α 2
where α is a shape parameter and SoftReLU is equivalent to ReLU when α = 0 is set.
W-MHSA combined with parallel convolution is used to introduce parallel convolution based on Multihead Self-Attention Mechanism (MHSA) to improve the network’s ability to extract features at different scales and the effect of defogging. In W-MHSA combined parallel convolution, the MHSA is firstly used for spatial information aggregation and the attention mechanism is applied to the information aggregation within the window to achieve the attention and integration of local information. In contrast to the spatial information aggregation approach of MHSA, W-MHSA introduces an additional convolution operation to aggregate the features but does not consider the window division to achieve the aggregation of neighborhood information. Thus, spatial information aggregation can be expressed as:
  Aggregation   Q , K , V = S o f t m a x Q K T / d + B V + C o n v V ^
where V ^ R b × h × w × c .
Specifically, the computational process of W-MHSA combined with parallel convolution is as follows: firstly, the convolution operation is performed on the input features V before window division to obtain the convolved features; then, the MHSA mechanism is used to perform the attentional weighted aggregation of the information inside the window; finally, the weighted information obtained from MHSA is added with the convolved features to realize the fusion of the in-window and out-of-window information.

2.3. Experimental Setting

2.3.1. Dataset

The RICE remote sensing image dataset was selected as the experimental dataset. The RICE dataset contains a large number of remote sensing images covering different scenarios and conditions. So, a portion of images from this dataset was randomly selected as experimental samples.

2.3.2. Defogging Algorithm

Four different defogging algorithms were selected for comparison, namely:
  • Single Image Haze Removal Using Dark Channel Prior (DCP)
  • All-in-One Dehazing Network (AOD-Net)
  • Feature Fusion Attention Network for Single Image Dehazing (FFA-Net)
  • Vision Transformers for Single Image Dehazing (DehazeFormer)

2.3.3. Experimental Step

Since the dark channel prior algorithm does not use a neural network, it is like a physical filter, so we do not have to perform any training operation on the dark channel prior algorithm model. However, the AOD-Net, FFA-Net, and DehazeFormer algorithms all use neural networks and attention mechanisms and there are original authors and trained pre-training models under each algorithm. So, in order to maintain the consistency and fairness of the experiments, we will first use the three algorithms, AOD-Net, FFA-Net, and DehazeFormer. The original trained pre-trained models are validated on the RICE dataset.
Secondly, we also fine-tuned the parameters on the three algorithmic models of AOD-Net, FFA-Net, and DehazeFormer according to the characteristics of the dataset and then trained using the RICE dataset to obtain the training model under each algorithm, with guided filtering added on top of the dark channel a prior algorithm, which better preserves the edge features after processing.
Finally, we compare the models obtained from the three algorithms (AOD-Net, FFA-Net, and DehazeFormer) trained on the RICE dataset with the pre-trained models provided by the original authors and select the models that are good within the two to be used for testing.

2.3.4. Evaluation Indicators

Two evaluation metrics based on image clarity were chosen to assess the de-cloud effect of different algorithms, PSNR (peak signal-to-noise ratio) [30] and SSIM (Structural Similarity Index) [31], which are used to measure the degree of similarity between the de-clouded image and the original image. PSNR is often used as a measure of the quality of signal reconstruction in areas such as image compression, where a larger PSNR value indicates a better-quality image, which is often defined simply by the Mean Square Error (MSE). For two m × n monochrome images I and K , if one is a noise approximation of the other, then their mean square error is defined as:
M S E = 1 m n i = 0 m 1 j = 0 n 1 [ I ( i , j ) K ( i , j ) ] 2
The peak signal-to-noise ratio is defined as:
P S N R = 10 log 10 M A X I 2 M S E = 20 log 10 M A X I M S E
where MAXI is the maximum value that represents the color of an image point, which is 255 if each sample point is represented by 8 bits.
SSIM is a metric used to measure the degree of similarity between two digital images, the larger the value of SSIM, the higher the structural similarity between the two images. When one of the two images is a distortion-free image and the other is a distorted image, the structural similarity of the two can be seen as an image quality measure of the distorted image. Given two images X and Y, the structural similarity of the two is defined as:
S S I M ( x , y ) = [ l ( x , y ) ] α [ c ( x , y ) ] β [ s ( x , y ) ] γ , l ( x , y ) = 2 μ x μ y + C 1 μ x 2 + μ y 2 + C 1 , c ( x , y ) = 2 σ x σ y + C 2 σ x 2 + σ y 2 + C 2 , s ( x , y ) = σ x y + C 3 σ x σ y + C 3
where l ( x , y ) compares the luminance of X and Y, c ( x , y ) compares the contrast of X and Y, and s ( x , y ) compares the structure of X and Y, with α > 0, β > 0, and γ > 0 as the adjusting l ( x , y ) , c ( x , y ) , s ( x , y ) parameters of relative importance and μ x and μ y and σ x and σ y as the means and standard deviations of X and Y, respectively. σ x σ y are the covariances of X and Y. C 1 , C 2 , and C 3 are constants.
Since remote sensing images have strong structural information and SSIM can better consider the structural similarity of the images, the ratio of PSNR to SSIM is set to be 3:7 and the value of the SSIM index is more valued in the evaluation.

2.3.5. Experimental Procedures

  • Randomly selected a batch of remote sensing images from the RICE dataset as experimental samples.
  • De-cloud selected images using the best model of each of the four algorithms.
  • Evaluate the de-clouded images and record the values of the evaluation metrics.
  • Analyze and compare the experimental results of different algorithms to explore their advantages, disadvantages, and applicable scenarios in terms of cloud removal effects.

3. Results

3.1. Selection of Pre-Trained Models

Among the four algorithms, because the original author of the DehazeFormer algorithm provides pre-training models for multiple scenarios as well as multiple datasets, to ensure the best effect of this algorithm on the RICE dataset, all the pre-training models of this algorithm were validated on this dataset in the preparation stage of the experiments. Because the RICE dataset contains two benchmark datasets and there are only 950 images in total, which constitutes fewer data, and many of the images inside the two benchmark datasets are from the same scene, for validation on the two benchmark datasets, the total number of images validated on each dataset was taken to be 100 images, which were randomly selected inside the respective dataset. Each pre-trained model under the algorithm was then evaluated and the pre-trained model with the best evaluation results under the algorithm was selected.
Each pre-training model was evaluated using the same metrics, i.e., PSNR and SSIM. At the end of the validation of each pre-training model, the PSNR and SSIM were computed for each image processing result and, finally, the PSNR and SSIM were averaged so that the PSNR and SSIM values of each pre-training model were obtained. Then, when evaluating the pre-trained models, the PSNR and SSIM were calculated according to the ratio 3:7. By calculating this ratio, the model with the largest score is selected as the pre-trained model under this algorithm for the corresponding dataset.
Since the DehazeFormer algorithm is trained and tested on the RESIDE [32] dataset and the RS-Haze dataset created by the original author, the original author also divided the RESIDE dataset into two different experimental settings according to different data distributions, namely RESIDE-Full and RESIDE-6K, where RESIDE-Full is further divided into indoor and outdoor. Therefore, there are four categories of pre-trained models under the DehazeFormer algorithm, namely indoor, outdoor, RESIDE-6k, and RS-haze. In the RESIDE-Full experimental setup, the indoor scenario mainly measures the model’s ability to process high-frequency information, while the outdoor scenario mainly measures the model’s convergence speed. In the RESIDE-6K experimental setup, the stability of the model and the ability to extract low-frequency information are mainly measured. In the RS-Haze dataset, the ability of the network to extract semantic features is mainly measured. There are trained pre-trained models under each category. The suffixes of the models, t, s, b, and m, represent tiny, small, basic, and medium, respectively. Under the four categories of the DehazeFormer algorithm, the PSNR and SSIM values of all pre-trained models verified on the RICE-I dataset are shown in Table 1, Table 2, Table 3 and Table 4, respectively.
From the data in the above four tables, it is known that the dehazeformer-m pre-trained model under the outdoor category is the most effective, so the dehazeformer-m pre-trained model under the outdoor category is selected as the pre-trained model under the RICE-I dataset under the DehazeFormer algorithm.
Under the four categories of the DehazeFormer algorithm, the PSNR and SSIM values of all pre-trained models verified on the RICE-II dataset are shown in Table 5, Table 6, Table 7 and Table 8, respectively.
From the data in Table 5, Table 6, Table 7 and Table 8, it is known that the dehazeformer-b pre-trained model under the RS-haze category is the best, so the dehazeformer-b pre-trained model under the RS-haze category is selected as the pre-trained model under the RICE-II dataset under the DehazeFormer algorithm.

3.2. Training Models

In this section, the main focus is on training the three algorithms, AOD-Net, FFA-Net, and DehazeFormer, and we fine-tuned the raw parameters of the three algorithms for better training. During the training process, we recorded the loss function values obtained from the model training, and when the loss function value reached the lowest, we stored the model parameters at that time as the best model under the training of that algorithm. Based on the total number of images, we selected 200 pairs of images for training the model on the RICE-I dataset and 200 pairs of images for training the model on the RICE-II dataset. Because of the small size of the dataset, the 200 pairs of data we selected for training the model in the DehazeFormer algorithm contained data from the 100 pairs that had been used previously in the pre-training model selection for DehazeFormer. This was for the sake of design rationality as well as to avoid overlap between the trained data and the subsequent test data. The models obtained by the three algorithms trained on the RICE-I dataset are shown in Table 9, and the models trained on the RICE-II dataset are shown in Table 10. Here, we name the training model for each algorithm to facilitate subsequent use.

3.3. Comparison of Trained and Pre-Trained Models

On RICE-I and RICE-II, we each selected 100 pairs of data that did not include the data we used to train the models. The results of the comparison between the original pre-trained models of the three algorithms and the models trained on the RICE dataset are shown in Table 11. The model with the highest score for both was selected as the final model to be tested.
Table 11 shows that on the RICE-I dataset, the pre-trained model of AOD-Net performs better than the trained model and the pre-trained model of FFA-Net also performs better than the trained model. Still, the trained model of DehazeFormer performs better than the pre-trained model. On the RICE-II dataset, the trained model of AOD-Net performs better than the pre-trained model, the pre-trained model of FFA-Net still performs better than the trained model, and the trained model of DehazeFormer still performs better than the pre-trained model.
Figure 5, Figure 6 and Figure 7 show the processing results of each algorithm’s pre-trained and trained models on the RICE-I and RICE-II datasets. It can be seen that the results of AOD-Net and FFA-Net processing using trained models are also not good and that the trained model of DehazeFormer performs better. Therefore, the DehazeFormer algorithm is more capable of generalization compared to the AOD-Net and FFA-Net algorithms.

3.4. Model Testing

In this stage of model testing, the pre-training model and the training model under each algorithm will be used as the test model after comparing the best results. To avoid overlap between the data used in testing and the data used in pre-training model selection, the datasets used in testing are excluded by combining the datasets that have been used in the pre-training model selection with those that the training model has used. Because the RICE-I dataset contains a total of 500 pairs of data and the RICE-II dataset includes a total of 450 pairs of data, after removing the 200 pairs of data used in training model selection and the 100 pairs of data used in the model comparison, there are still 200 pairs of data in the RICE-I dataset and 150 pairs of data in the RICE-II dataset. Therefore, for testing, we selected the remaining 200 pairs of images for testing in RICE-I and the remaining 150 pairs of images in RICE-II. Each algorithm is compared on two benchmark datasets, the results of the comparison on the RICE-I dataset are shown in Table 12, and the results of the comparison on the RICE-II dataset are shown in Table 13.
As can be seen from Table 12, the DehazeFormer algorithm has the best performance, with the highest PSNR and SSIM values under the RICE-I dataset. The PSNR value of DCP is the lowest, but as seen in Figure 8, the DCP algorithm is also effective in removing clouds. This is because PSNR mainly measures the overall brightness and noise level of the image. For the processed image, although the visual effect is improved, if the brightness and contrast adjustments are large, it may lead to a decrease in PSNR. In Figure 8, the image processed by the DCP algorithm is compared to the images processed by the other three algorithms when compared to the clear image. The image processed by the DCP algorithm suffers from severe color and contrast distortion and, as a result, the DCP has the lowest PSNR value.
As can be seen from Table 13, the DehazeFormer algorithm still performs best with the RICE-II dataset.

4. Discussion

In the RICE-I dataset, most of the images are fuzzy images shrouded by some uniform thin clouds and there are also some images with relatively thick clouds. Here, three images with different cloud concentrations are selected to evaluate the network’s performance regarding de-clouding clouds. For thin clouds, all four algorithms perform well in cloud removal; however, in restoring a clear image, three of the algorithms (except for the DehazeFormer algorithm) do not handle the color and details very well (see the first and third rows of Figure 8).
The DCP algorithm performs better when the cloud layer is relatively thick and the concentration is relatively high, where it can remove most of the clouds, but when recovering a clear image, the DCP-algorithm-processed image has distortion and poor color reproduction compared to the original clear image. The other three algorithms are not effective in removing most of the clouds when the cloud concentration is relatively high. See the second row of Figure 8.
The reason for the above situation may be because under uniform thin clouds the ground scene is relatively clear, the color and texture changes are small, and the effect of occlusion is relatively small, and these characteristics are conducive to the processing of the image by the defogging algorithm and thus obtaining a better de-clouding effect. The AOD-Net network was originally designed for defogging through a simple end-to-end architecture and may not have been tuned for dealing with a specific type of noise like clouds, which can lead to color distortion. Although FFA-Net introduces feature fusion and attention mechanisms, these mechanisms are mainly used to enhance feature extraction and are not specifically optimized for clouds, a type of noise that can also cause color distortion. DehazeFormer is based on Transformer’s architecture, which excels at capturing global information and reduces color distortion through a self-attention mechanism that allows for more precise adjustment of features.
When clouds are uniformly thicker, thick clouds scatter and absorb light more strongly, making the loss of contrast and detail in the image more severe. This places higher demands on the defogging (cloud removal) algorithm, which requires stronger feature extraction and recovery capabilities. Thick clouds cause many details in the image to be completely obscured, so relying solely on existing feature extraction techniques may not be sufficient to recover these details and hence the three algorithms AOD-Net, FFA-Net, and DehazeFormer do not perform well.
The DCP algorithm in this paper, on the other hand, employs guided filtering using the grayscale map of the image to be processed as the guide map, which better preserves the detailed features of the edges and reduces the blurring and artifacts that can be caused by the cloud removal process. In removing uniform thin clouds, which usually do not completely obscure the background information, the dark channel prior can effectively recognize and remove the interference caused by these thin clouds and restore the contrast of the image. Although thicker clouds are more complex, they still do not completely obscure the background information. By extracting the darkest channel information in the image, the DCP can estimate the atmospheric light and the transmission rate very well, thus effectively removing the effects brought by the clouds and thus removing the clouds with relatively good results. Also, because the luminance variations in clouds are different from haze, this may lead to the failure of the a prior assumptions of the dark channels, especially in thick cloud regions, which may lead to inaccurate color restoration. At the same time, the inconsistent effect of cloud cover on different color channels may lead to poor restoration of certain color channels, which may result in overall image color distortion.
In the RICE-II dataset, there are some fuzzy images shrouded by some irregularly shaped and non-uniform clouds. Again, three different images are selected, and it can be seen from Figure 9 that the three algorithms, DCP, AOD-Net, and FFA-Net, perform poorly on the RICE-II dataset. The DehazeFormer algorithm removes clouds but does not restore a clear image in its entirety.
The AOD-Net, FFA-Net, and DehazeFormer algorithms may produce images like this due to the following reasons: First of all, in the case of uneven cloud shapes, there may be a large cloud mass obscuring the ground scene, resulting in partial or total occlusion of the ground scene. General defogging algorithms find it difficult to deal with this kind of occlusion problem efficiently because they usually assume that the entire image scene is affected by the same degree of haze and ignore the localized occlusion. Secondly, in the case of thicker clouds, the clouds may reflect and scatter light more strongly, resulting in reduced visibility of the ground scene, and the defogging algorithms may not be able to accurately estimate the thickness and optical properties of the clouds to correctly restore the clarity of the ground scene. Finally, due to the uneven cloud shape and thickness inhomogeneity, some areas may be affected by stronger occlusions and reflections, resulting in localized distortions of the color and texture of ground scenes, and the defogging algorithms may also not be able to accurately recover these localized details, leading to poor results.
The reason for the poor performance results of the DCP defogging algorithm on image defogging may be that the principle of the DCP algorithm is based on the dark channel a prior assumption that most of the localized regions in a natural image are relatively dark in at least one color channel. This assumption applies to image blurring due to atmospheric scattering but not to clouds. Clouds may have higher reflectivity and more complex color distributions that do not have a dark channel prior, making it difficult for the defogging algorithm to accurately estimate a clear image.
The performance of these four defogging algorithms on the RICE dataset varies through model testing and analysis. For the RICE-I dataset, all of the algorithms perform better when dealing with uniform thin clouds, but the DCP and DehazeFormer algorithms perform better when dealing with relatively uniform thick clouds. For the RICE-II dataset, which has irregular shapes, unevenly distributed clouds, and thick clouds, the DehazeFormer algorithm has the best de-clouding ability compared to the other three algorithms. The computational complexity of each algorithm is different, so the speed of processing images is different, and each algorithm is tested using Windows with an RTX 4060 graphics card. Table 14 shows the number of parameters and FLOPs of each algorithm test model in this paper by taking the number of parameters and FLOPs index as a measure of the computational complexity because the DCP algorithm in this paper does not use neural networks and attention mechanisms so is not compared in the table; however, the DCP algorithm is designed to have a low complexity, processing an average of 10 images a second with an image size of 512 × 512 using an RTX 4060 graphics card and Windows 11.
As we can see from the data in the table, DehazeFormer has a lot of parameters due to the inclusion of a large number of attention mechanisms, low FLOPs, and a fast processing speed, so it is more suitable for the real-time processing of large datasets and its model performs more powerfully, adapts to more complex tasks, and has a better ability to learn and generalize. However, attention should be paid to preventing overfitting. The AOD-Net and FFA-Net algorithms have high FLOPs because they use convolution a lot, so the processing speed is slower.
In summary, when we are removing thin cloud pictures with a relatively uniform distribution, we can choose the four defogging algorithms in this paper. If the computer’s arithmetic power is relatively low, then we can choose the DCP algorithm. If the arithmetic power is relatively high and you want to process in real time you can choose the DehazeFormer algorithm. The AOD-Net and FFA-Net algorithms can be selected when dealing with a single dataset and real-time processing is not required. When removing thick clouds and images covered by uneven cloud cover, the above four algorithms are no longer applicable and we should improve the defogging algorithm according to the characteristics of the cloud cover of remote sensing images.

5. Conclusions

In this paper, the generalization ability of the current mainstream four defogging algorithms in the direction of remote sensing defogging is compared and analyzed, and after the experiments and the experimental data, it is concluded that the four defogging algorithms in this paper have a weak generalization ability in the direction of remote sensing defogging on the RICE remote sensing image dataset and cannot be applied to remote sensing image defogging.
There are several ideas for improvement in the problem of de-clouding remote sensing images:
  • Design of specialized algorithms for different types of clouds based on the characteristics of different types of clouds; targeted cloud removal algorithms can be designed to improve the algorithm’s effectiveness in dealing with different types of clouds.
  • Combining multiple sensor data for cloud identification and removal, where different types of sensors can provide complementary information to help accurately identify and remove clouds, thus improving the accuracy and robustness of the cloud removal algorithm.
  • End-to-end cloud removal from remote sensing images using deep learning methods that are capable of learning complex image features and patterns to improve the generalization ability and effectiveness of cloud removal algorithms.
  • Addition of ground-truth and synthetic datasets, including various types and densities of cloud cover scenarios. With richer datasets, the de-clouding algorithms can be better trained and evaluated, improving their generalization and applicability.
Future research on this subject will probably be in the following areas:
  • Multimodal data fusion: further research on how to fuse multiple sensor data to improve the accuracy and effectiveness of cloud identification and removal.
  • Optimization for specific application scenarios: optimize the de-clouding algorithm for specific remote sensing application scenarios, such as agricultural monitoring, urban planning, etc., to improve its applicability and effectiveness in specific scenarios.
  • Improvement of real-time capability and efficiency: study how to improve the real-time capability and efficiency of the de-clouding algorithm so that it can be effectively applied in large-scale remote sensing image processing.
  • Processing in conjunction with spatial and temporal information: combining spatial and temporal information to dynamically monitor and process remotely sensed images to better understand and respond to cloud cover changes at different temporal and spatial scales.

Author Contributions

Conceptualization, G.M., Z.Z. and Z.W.; Methodology, G.M.; Software, Z.Z. and G.M.; Validation, G.M., Z.Z. and Z.W.; Formal analysis, G.M.; Resources, G.M., Z.Z. and Z.W.; Data curation, G.M.; Writing—original draft, G.M., Z.Z. and Z.W; Writing—review and editing, G.M., Z.Z. and Z.W.; Visualization, G.M.; Supervision, Z.Z. and Z.W.; Funding acquisition, Z.Z. All authors have read and agreed to the published version of the manuscript.

Funding

Funding was provided by the Key Research and Development Program Projects of Shaanxi Province (Grant: 2023-YBSF-397) and the PhD Startup Fund of Shaanxi University of Technology (Grant: SLGQD2017-06).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used and/or analyzed during the current study, as well as the program code, are available from the corresponding author on reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. In Proceedings of the 2009 IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2009, Miami, FL, USA, 20–25 June 2009; IEEE Computer Society: Washington, DC, USA, 2009; pp. 1956–1963. [Google Scholar]
  2. Ren, W.; Pan, J.; Zhang, H.; Cao, X.; Yang, M.H. Single Image Dehazing via Multi-Scale Convolutional Neural Networks with Holistic Edges. Int. J. Comput. Vis. 2020, 128, 240–259. [Google Scholar] [CrossRef]
  3. Li, C.-Y.; Guo, J.-C.; Cong, R.-M.; Pang, Y.-W.; Wang, B. Underwater Image Enhancement by Dehazing with Minimum Information Loss and Histogram Distribution Prior. IEEE Trans. Image Process. 2016, 25, 5664–5677. [Google Scholar] [CrossRef] [PubMed]
  4. Liu, Y.; Li, H.; Wang, M. Single Image Dehazing via Large Sky Region Segmentation and Multiscale Opening Dark Channel Model. IEEE Access 2017, 5, 8890–8903. [Google Scholar] [CrossRef]
  5. Li, Y.; Tan, R.T.; Brown, M.S. Nighttime Haze Removal with Glow and Multiple Light Colors. In Proceedings of the IEEE International Conference on Computer Vision, Washington, DC, USA, 7–13 December 2015; pp. 226–234. [Google Scholar]
  6. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. AOD-Net: All-in-One Dehazing Network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017. [Google Scholar]
  7. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature Fusion Attention Network for Single Image Dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, Honolulu Hawaii, HI, USA, 27 January–1 February 2019. [Google Scholar]
  8. Song, Y.; He, Z.; Qian, H.; Du, X. Vision Transformers for Single Image Dehazing. IEEE Trans. Image Process. 2022, 32, 1927–1941. [Google Scholar] [CrossRef] [PubMed]
  9. Liu, Z.; Xiao, B.; Alrabeiah, M.; Wang, K.; Chen, J. Single Image Dehazing with a Generic Model-Agnostic Convolutional Neural Network. IEEE Signal Process. Lett. 2019, 26, 833–837. [Google Scholar] [CrossRef]
  10. Engin, D.; Genç, A. Cycle-Dehaze: Enhanced CycleGAN for Single Image Dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  11. Zhao, D.; Li, J.; Li, H.; Xu, L. Complementary Feature Enhanced Network with Vision Transformer for Image Dehazing. arXiv 2021, arXiv:2109.07100. [Google Scholar]
  12. Narasimhan, S.G.; Nayar, S.K. Vision and the Atmosphere. Int. J. Comput. Vis. 2002, 48, 233–254. [Google Scholar] [CrossRef]
  13. Tan, R.T. Visibility in Bad Weather from a Single Image. In Proceedings of the 2008 IEEE Conference on Computer Vision and Pattern Recognition, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8. [Google Scholar]
  14. Guo, F.; Yang, J.; Liu, Z.; Tang, J. Haze Removal for Single Image: A Comprehensive Review. Neurocomputing 2023, 537, 85–109. [Google Scholar] [CrossRef]
  15. Zhang, Y.; Chen, Y.; Huang, C.; Gao, M. Object Detection Network Based on Feature Fusion and Attention Mechanism. Future Internet 2019, 11, 9. [Google Scholar] [CrossRef]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. Adv. Neural Inf. Process. Syst. 2017, 30. [Google Scholar]
  17. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  18. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021. [Google Scholar]
  19. Lin, D.; Xu, G.; Wang, X.; Wang, Y.; Sun, X.; Fu, K. A Remote Sensing Image Dataset for Cloud Removal. arXiv 2019, arXiv:1901.00600. [Google Scholar]
  20. Goldstein, E.B.; Cacciamani, L.; Brazil, A. Sensation and Perception, 11th ed.; Cengage Learning: Singapore, 2022; Available online: https://www.cengageasia.com/title/default/detail?isbn=9780357446478 (accessed on 13 May 2024).
  21. Preetham, A.J.; Shirley, P.; Smits, B. A Practical Analytic Model for Daylight. In Proceedings of the 26th Annual Conference on Computer Graphics and Interactive Techniques, San Antonio, TX, USA, 23–26 July 1999. [Google Scholar]
  22. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. DehazeNet: An End-to-End System for Single Image Haze Removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [PubMed]
  23. Elfwing, S.; Uchibe, E.; Doya, K. Sigmoid-Weighted Linear Units for Neural Network Function Approximation in Reinforcement Learning. Neural Netw. 2017, 107, 3–11. [Google Scholar] [CrossRef] [PubMed]
  24. Agarap, A.F. Deep Learning Using Rectified Linear Units (ReLU). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  25. Ba, J.L.; Kiros, J.R.; Hinton, G.E. Layer Normalization. arXiv 2016, arXiv:1607.06450. [Google Scholar]
  26. Ramachandran, P.; Zoph, B.; Le, Q.V. Searching for Activation Functions. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  27. Hendrycks, D.; Gimpel, K. Gaussian Error Linear Units (GELUs). arXiv 2016, arXiv:1606.08415. [Google Scholar]
  28. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  29. Nair, V.; Hinton, G.E. Rectified Linear Units Improve Restricted Boltzmann Machines. In Proceedings of the 27th International Conference on Machine Learning (ICML-10), Haifa, Israel, 21–24 June 2010. [Google Scholar]
  30. Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar]
  31. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image Quality Assessment: From Error Visibility to Structural Similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef] [PubMed]
  32. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking Single Image Dehazing and Beyond. IEEE Trans. Image Process. 2015, 28, 505. [Google Scholar] [CrossRef] [PubMed]
Figure 1. RICE-I dataset. The cloud column represents a blurred image with clouds, and the label column represents a clear image without clouds.
Figure 1. RICE-I dataset. The cloud column represents a blurred image with clouds, and the label column represents a clear image without clouds.
Sensors 24 04566 g001
Figure 2. RICE-II dataset. The cloud column represents the blurred image with cloud, the label column represents the corresponding clear image without cloud, and the MASK column represents the corresponding mask image with cloud.
Figure 2. RICE-II dataset. The cloud column represents the blurred image with cloud, the label column represents the corresponding clear image without cloud, and the MASK column represents the corresponding mask image with cloud.
Sensors 24 04566 g002
Figure 3. Feature attention module.
Figure 3. Feature attention module.
Sensors 24 04566 g003
Figure 4. Basic block structure.
Figure 4. Basic block structure.
Sensors 24 04566 g004
Figure 5. Comparison of AOD-Net’s pre-trained and trained models on the RICE-I and RICE-II datasets. The first column represents the original input image with fog, the second column represents the image processed using the pre-trained model, the third column represents the image processed using the trained model, and the fourth column represents the original clear image without fog. The first two rows represent the images inside the RICE-I dataset, and the last two rows represent the images inside the RICE-II dataset.
Figure 5. Comparison of AOD-Net’s pre-trained and trained models on the RICE-I and RICE-II datasets. The first column represents the original input image with fog, the second column represents the image processed using the pre-trained model, the third column represents the image processed using the trained model, and the fourth column represents the original clear image without fog. The first two rows represent the images inside the RICE-I dataset, and the last two rows represent the images inside the RICE-II dataset.
Sensors 24 04566 g005
Figure 6. Comparison of FFA-Net’s pre-trained and trained models on the RICE-I and RICE-II datasets.
Figure 6. Comparison of FFA-Net’s pre-trained and trained models on the RICE-I and RICE-II datasets.
Sensors 24 04566 g006
Figure 7. Comparison of DehazeFormer’s pre-trained and trained models on the RICE-I and RICE-II datasets.
Figure 7. Comparison of DehazeFormer’s pre-trained and trained models on the RICE-I and RICE-II datasets.
Sensors 24 04566 g007aSensors 24 04566 g007b
Figure 8. Comparison of RICE-I remote sensing images on four algorithms, DCP, AOD-Net, FFA-Net, and DehazeFormer. The first column represents the original input fogged image, the last column represents the original clear image without fog, and the middle four columns represent the images after being processed by each algorithm.
Figure 8. Comparison of RICE-I remote sensing images on four algorithms, DCP, AOD-Net, FFA-Net, and DehazeFormer. The first column represents the original input fogged image, the last column represents the original clear image without fog, and the middle four columns represent the images after being processed by each algorithm.
Sensors 24 04566 g008
Figure 9. Comparison of RICE-II remote sensing images on four algorithms, DCP, AOD-Net, FFA-Net, and DehazeFormer.
Figure 9. Comparison of RICE-II remote sensing images on four algorithms, DCP, AOD-Net, FFA-Net, and DehazeFormer.
Sensors 24 04566 g009
Table 1. PSNR and SSIM values were obtained from the validation of pre-trained models under the indoor category on the RICE-I dataset.
Table 1. PSNR and SSIM values were obtained from the validation of pre-trained models under the indoor category on the RICE-I dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b18.150.8346.03
dehazeformer-m 17.09 0.773 5.67
dehazeformer-s17.520.8155.83
dehazeformer-t18.300.8486.08
Note: The number highlighted in each column of the table is the maximum value for that column.
Table 2. PSNR and SSIM values were obtained from the validation of pre-trained models under the outdoor category on the RICE-I dataset.
Table 2. PSNR and SSIM values were obtained from the validation of pre-trained models under the outdoor category on the RICE-I dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b20.750.8856.84
dehazeformer-m21.050.8866.93
dehazeformer-s20.810.8826.86
dehazeformer-t20.820.8826.86
Table 3. PSNR and SSIM values were obtained from the validation of pre-trained models under the RESIDE-6K category on the RICE-I dataset.
Table 3. PSNR and SSIM values were obtained from the validation of pre-trained models under the RESIDE-6K category on the RICE-I dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b20.700.8866.83
dehazeformer-m20.670.8736.81
dehazeformer-s20.860.8686.86
dehazeformer-t20.960.8756.90
Table 4. PSNR and SSIM values were obtained from the validation of pre-trained models under the RS-haze category on the RICE-I dataset.
Table 4. PSNR and SSIM values were obtained from the validation of pre-trained models under the RS-haze category on the RICE-I dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b19.130.8006.30
dehazeformer-m19.210.8086.33
dehazeformer-s19.100.8076.30
dehazeformer-t18.610.8036.14
Table 5. PSNR and SSIM values were obtained from the validation of pre-trained models under the indoor category on the RICE-II dataset.
Table 5. PSNR and SSIM values were obtained from the validation of pre-trained models under the indoor category on the RICE-II dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b16.480.6275.38
dehazeformer-m17.160.6445.60
dehazeformer-s17.210.6515.62
dehazeformer-t16.860.6335.50
Table 6. PSNR and SSIM values were obtained from the validation of pre-trained models under the outdoor category on the RICE-II dataset.
Table 6. PSNR and SSIM values were obtained from the validation of pre-trained models under the outdoor category on the RICE-II dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b16.970.6465.54
dehazeformer-m16.590.6335.42
dehazeformer-s16.740.6335.46
dehazeformer-t16.820.6395.49
Table 7. PSNR and SSIM values were obtained from the validation of pre-trained models under the RESIDE-6K category on the RICE-II dataset.
Table 7. PSNR and SSIM values were obtained from the validation of pre-trained models under the RESIDE-6K category on the RICE-II dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b16.250.6155.30
dehazeformer-m16.360.6205.34
dehazeformer-s16.440.6255.37
dehazeformer-t16.330.6215.33
Table 8. PSNR and SSIM values were obtained from the validation of pre-trained models under the RS-haze category on the RICE-II dataset.
Table 8. PSNR and SSIM values were obtained from the validation of pre-trained models under the RS-haze category on the RICE-II dataset.
ModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
dehazeformer-b18.990.6806.17
dehazeformer-m19.000.6586.16
dehazeformer-s18.740.6716.09
dehazeformer-t18.740.6746.09
Table 9. Models obtained by training the three algorithms, AOD-Net, FFA-Net, and DehazeFormer, on the RICE-I dataset.
Table 9. Models obtained by training the three algorithms, AOD-Net, FFA-Net, and DehazeFormer, on the RICE-I dataset.
ArithmeticModel
AOD-NetAOD-train-model-I
FFA-NetFFA-train-model-I
DehazeFormerDehazeFormer-train-model-I
Table 10. Models obtained by training the three algorithms, AOD-Net, FFA-Net, and DehazeFormer, on the RICE-II dataset.
Table 10. Models obtained by training the three algorithms, AOD-Net, FFA-Net, and DehazeFormer, on the RICE-II dataset.
ArithmeticModel
AOD-NetAOD-train-model-II
FFA-NetFFA-train-model-II
DehazeFormerDehazeFormer-train-model-II
Table 11. Comparison between pre-trained and trained models of the three algorithms, AOD-Net, FFA-Net, and DehazeFormer.
Table 11. Comparison between pre-trained and trained models of the three algorithms, AOD-Net, FFA-Net, and DehazeFormer.
RICE-IArithmeticModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
AOD-NetPretrained models17.680.7075.799
AOD-train-model-I15.800.7555.269
FFA-NetPretrained models15.300.6975.078
FFA-train-model-I14.780.4534.751
DehazeFormerPretrained models17.700.8245.887
DehazeFormer-train-model-I18.130.8136.008
RICE-IIArithmeticModelPSNRSSIM p s n r × 0.3 + s s i m × 0.7
AOD-NetPretrained models15.080.4054.808
AOD-train-model-I17.150.7005.635
FFA-NetPretrained models17.920.7275.885
FFA-train-model-I14.570.2914.575
DehazeFormerPretrained models17.300.5335.563
DehazeFormer-train-model-I28.090.8629.030
Note: The number highlighted in the table is the calculated maximum value under the algorithm for that row.
Table 12. PSNR and SSIM values of each algorithm under the RICE-I dataset.
Table 12. PSNR and SSIM values of each algorithm under the RICE-I dataset.
MethodsPSNRSSIM p s n r × 0.3 + s s i m × 0.7
DCP16.520.7985.5146
AOD-Net16.650.7405.513
FFA-Net20.130.7756.5815
DehazeFormer25.310.9028.2244
Table 13. PSNR and SSIM values of each algorithm under the RICE-II dataset.
Table 13. PSNR and SSIM values of each algorithm under the RICE-II dataset.
MethodsPSNRSSIM p s n r × 0.3 + s s i m × 0.7
DCP13.780.5674.5309
AOD-Net19.830.7956.5055
FFA-Net18.450.7566.0642
DehazeFormer28.220.8329.0484
Table 14. Parameters and FLOPs values for each algorithm and its test model.
Table 14. Parameters and FLOPs values for each algorithm and its test model.
Method/ModelParametersFLOPs
DehazeFormer-train-model-I1876.57 Billion4.633504 × 106
DehazeFormer-train-model-II939.92 Billion2.514348 × 106
AOD-Net17612.28851712 × 108
FFA-Net4,455,9132.329659047936 × 1012
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Miao, G.; Zhang, Z.; Wang, Z. Analysis of the Generalization Ability of Defogging Algorithms on RICE Remote Sensing Images. Sensors 2024, 24, 4566. https://doi.org/10.3390/s24144566

AMA Style

Miao G, Zhang Z, Wang Z. Analysis of the Generalization Ability of Defogging Algorithms on RICE Remote Sensing Images. Sensors. 2024; 24(14):4566. https://doi.org/10.3390/s24144566

Chicago/Turabian Style

Miao, Guisheng, Zhongpeng Zhang, and Zhanbei Wang. 2024. "Analysis of the Generalization Ability of Defogging Algorithms on RICE Remote Sensing Images" Sensors 24, no. 14: 4566. https://doi.org/10.3390/s24144566

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop