Next Article in Journal
GammaGAN: Gamma-Scaled Class Embeddings for Conditional Video Generation
Next Article in Special Issue
RCEAU-Net: Cascade Multi-Scale Convolution and Attention-Mechanism-Based Network for Laser Beam Target Image Segmentation with Complex Background in Coal Mine
Previous Article in Journal
Multi-Scale FPGA-Based Infrared Image Enhancement by Using RGF and CLAHE
Previous Article in Special Issue
Wood Veneer Defect Detection Based on Multiscale DETR with Position Encoder Net
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Scale Attention Feature Enhancement Network for Single Image Dehazing

1
School of Opto-Electronic Engineering, Changchun University of Science and Technology, Changchun 130022, China
2
Zhongshan Institute of Changchun University of Science and Technology, Zhongshan 528437, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(19), 8102; https://doi.org/10.3390/s23198102
Submission received: 17 July 2023 / Revised: 18 September 2023 / Accepted: 25 September 2023 / Published: 27 September 2023

Abstract

:
Aiming to solve the problem of color distortion and loss of detail information in most dehazing algorithms, an end-to-end image dehazing network based on multi-scale feature enhancement is proposed. Firstly, the feature extraction enhancement module is used to capture the detailed information of hazy images and expand the receptive field. Secondly, the channel attention mechanism and pixel attention mechanism of the feature fusion enhancement module are used to dynamically adjust the weights of different channels and pixels. Thirdly, the context enhancement module is used to enhance the context semantic information, suppress redundant information, and obtain the haze density image with higher detail. Finally, our method removes haze, preserves image color, and ensures image details. The proposed method achieved a PSNR score of 33.74, SSIM scores of 0.9843 and LPIPS distance of 0.0040 on the SOTS-outdoor dataset. Compared with representative dehazing methods, it demonstrates better dehazing performance and proves the advantages of the proposed method on synthetic hazy images. Combined with dehazing experiments on real hazy images, the results show that our method can effectively improve dehazing performance while preserving more image details and achieving color fidelity.

1. Introduction

With the rapid development of modern industrialization and urbanization, haze has become a common natural phenomenon. Haze is composed of a large number of tiny particles such as water droplets and sulfur dioxide. When optical imaging equipment is collecting images, due to the scattering effect of these tiny particles, the contrast of the collected images is low, and many important details are lost. It is difficult to meet the requirements of object detection [1,2], target recognition [3,4], and other computer vision applications. Therefore, it is of great practical importance and application value to study the excellent performance of the dehazing algorithm to clarify the hazy images.
Single-image dehazing is a challenging problem, and many researchers have started to study single-image dehazing techniques [5,6,7,8]. The purpose of single image dehazing is to restore the image degraded by haze to a clear image. In recent years, most dehazing algorithms are based on a physical scattering model [9,10,11], which can be formulated as
I ( x ) = t ( x ) J ( x ) + A ( 1 t ( x ) )
where I ( x ) is the hazy image captured by the imaging device, J ( x ) is the restored haze-free image, A is the global atmospheric light, and t ( x ) is the transmission map. Unfortunately, in practical applications, both the transmission map and the atmospheric light are unknown. Therefore, most current image dehazing methods first estimate A and t ( x ) , and then recover J ( x ) according to Equation (1).
These studies are based on physical models of atmospheric scattering and aim to restore images with prior knowledge. However, the optimal choice of prior knowledge is still unclear. In addition, the extent to which these priors obey the image statistics and how they influence the dehazing performance also remain unknown [12]. Recently, since the emergence of deep learning, learning-based methods have gradually been applied in the field of image dehazing [13,14]. Although existing learning-based methods have achieved remarkable success, Their performance is still limited by the loss of detailed information such as color [15].
Based on the above analysis, in order to avoid dehazing problems such as color distortion and incomplete dehazing, and to preserve as much image detail information as possible, an end-to-end single image dehazing network with multi-scale attention feature enhancement is proposed. This method can directly restore the input haze image to a clear image without estimating the parameters of the atmospheric scattering model.
The main contributions of this work are summarized as follows:
(1)
We propose an end-to-end Multi-Scale Attention Feature Enhancement Networks for Single Image Dehazing (MAFE). This method has achieved excellent performance in image dehazing. It can adaptively focus on the high-frequency information of the hazy image and retain more detailed information. Since it does not need to rely on the atmospheric scattering model, it is not affected by the estimated atmospheric light value. Additionally, color distortion is avoided.
(2)
We propose an attention feature enhancement module, which can adaptively focus on high-frequency information of hazy images, enhance the relevance of contextual information, suppress redundant information, and compensate for the loss of detailed information.
(3)
We propose a multi-scale attention enhancement module that builds upon the attention feature enhancement module and incorporates a spatial pyramid of dilated convolutions to fully extract and utilize the multi-scale features of the image. This module expands the receptive field and improves the quality of the dehazed image while preserving more detailed information.
(4)
The experimental results on both synthetic and real-world hazy images demonstrate that our proposed method achieves state-of-the-art single image dehazing methods in terms of dehazing performance. It can well preserve details such as color and texture of the image.

2. Related Work

Single image dehazing is an ill-posed problem. At present, the methods for single image dehazing are mainly divided into two categories, one is to artificially summarize the statistical difference between the blurred image and the unblurred image as an empirical prior, and the other is to directly or indirectly learn the mapping function from large-scale datasets of hazy and clear images. Researchers generally refer to the former as prior-based methods and the latter as learning-based methods.
The prior-based methods estimate the transmission rate and global atmospheric light based on some prior knowledge of the clean image. They usually rely on the atmospheric scattering model and handcrafted priors. He et al. [16] found that most local patches in outdoor haze-free images contain some pixels whose intensity is very low in at least one color channel, proposing dark channel priors (DCP). Zhu et al. [17] proposed a color attenuation prior by observing that the presence of haze can lead to image desaturation and brightness increase. Berman et al. [18] found that when haze appears, the pixel clusters of the haze-free image will become haze lines, and proposed a non-local prior to characterize the clean image. Fattal [19] proposes a method for estimating scene reflectance by assuming that the projection of the scene target surface and the propagation of light are partially uncorrelated to achieve dehazing images.
In recent years, with the rapid development of deep learning [20,21,22,23], many researchers have applied deep learning to image dehazing and designed a large number of dehazing neural networks. CAI et al. [24] proposed the Dehaze-net single-image dehazing network, which was the first to introduce convolutional neural networks into image dehazing tasks. It estimates the transmittance through operations such as multi-scale mapping and maximum pooling, and then clears the hazy image according to the atmospheric scattering model. Compared with traditional dehazing methods, Dehaze-net greatly improves the dehazing performance. Ren et al. [25] designed a coarse-scale network and a fine-scale network model to achieve dehazing by extracting and fusing the coarse transmission map and the fine transmission map. Li et al. [26] proposed All-in-One Network (AOD-Net), They unify the two parameters of atmospheric light value and transmittance in the atmospheric scattering model, convert them into a new variable, use lightweight convolution to estimate this variable and generate a dehazed image. Liao et al. [27] proposed HDP-net for night-time dehazing to restore haze-free images by estimating the haze density map. Chen et al. [28] proposed a gated aggregation network that uses dilated convolutions to increase the image receptive field, aggregates the semantic information of the image, and successfully solves the problem of network artifacts. Shao et al. [29] proposed a domain-adaptive dehazing network, which transforms images from one domain to another through a bidirectional transformation network to reduce the error between the synthetic domain and the real domain, and solve the problem of network artifacts. Qin et al. [30] proposed an end-to-end feature fusion attention network (FFA-net). They designed a novel feature attention module, which can effectively focus on the dense haze and high-frequency areas, and achieved excellent results in restoring synthetic hazy images. Liu et al. [31] proposed the GridDehazeNet dehazing network, which can generate learning inputs with better diversity and more relevant features. Through a novel attention-based multi-scale estimation, it effectively alleviates the bottleneck issue often encountered in the conventional multi-scale approach. Wang et al. [32] improved feature extraction and enhanced feature mapping, proposed a multi-scale supervision-guided context aggregation network (MSGCAN) based on two principles, and obtained better dehazing results. Based on the CIELAB color space, Sheng et al. [33] proposed a multi-scale residual attention network for single image dehazing, which improved the color performance of the dehazing method. Zhou et al. [34] proposed an attention-based feature fusion dehazing network, which uses attention-based residual dense blocks to enhance the details of low-light haze images.

3. Proposed Method

In this section, we first describe the haze density image prediction model and propose a haze density image prediction network based on multi-scale attention feature enhancement. Then, we introduce the dehazing fusion block (DFB), multi-scale attention module (MAE), attention feature enhancement module (AFE) and context enhancement module (CEM) in detail. Finally, we provide loss functions for training the network.

3.1. Haze Density Image Prediction Model

Since it is difficult to accurately estimate the atmospheric light value A in the atmospheric scattering model, HDP-net [27] redefines the atmospheric scattering model based on the formation principle of the fog map, which can be expressed as
I ( x ) = J ( x ) + h ( x )
where I ( x ) is the observed hazy image, J ( x ) is the real scene to be recovered, h ( x ) is the haze density image. At the same time, a haze density image prediction model is proposed as
h ( x ) = N ( I ( x ) )
where N ( ) is haze density image prediction network. Thus, the mathematical model of the haze-free image is obtained as
J ( x ) = I ( x ) N ( I ( x ) ) = I ( x ) h ( x )
According to Equation (4), it is only necessary to build a haze density image prediction network N ( ) to obtain the haze density image h ( x ) , and then I ( x ) can restore to J ( x ) . We propose a haze density image prediction network architecture with multi-scale attention feature enhancement. It can predict the haze density image h ( x ) in an end-to-end manner, thus restoring the haze-free image. Below, we will provide a detailed introduction to the multi-scale attention feature enhancement network architecture that we designed.

3.2. Network Architecture

The multi-scale attention feature enhancement dehazing network (MAFE) we propose is shown in Figure 1. MAFE includes feature extraction enhancement module (FEEM) and feature fusion enhancement module (FFEM). FFEM is mainly composed of three dehazing fusion blocks (DFB), which are used to extract and fuse deep and shallow haze information. In DFB, we use the attention mechanism and dilated convolution, so the FEEM module has a larger receptive field and can capture more detailed information.
The FFEM mainly consists of channel attention (CA), pixel attention (PA) [30], context enhancement modules (CEM) [35], and two convolutional layers with a stride of 1.
The input of MAFE is the hazy image of the RGB channel. After it is transmitted to FEEM, the convolution kernel is used to extract the shallow information of the hazy image and passed to DFB. DFB focuses on haze and high-frequency information. This network connects three DFBs in series. As the network gradually deepens, a large amount of detailed information will be lost. Therefore, the features extracted by the three DFBs are connected in the channel direction to realize the fusion of deep and shallow information and preserve more details while retaining rich features. The fusion information obtained from FEEM is transferred to FFEM. First, CA and PA are used to further flexibly adjust the obtained fusion haze density information. Secondly, CEM is used to enhance context information and suppress redundant information. Then, the haze density image h ( x ) is obtained through convolution reconstruction. Finally, Equation (4) is used to obtain a clear haze-free image.

3.3. DFB Mathematical Model

DFB consists of multilevel attention enhancement module (MAE), convolutional layers and skip connections, as shown in Figure 2. MAE is introduced in Section 3.4. We use four consecutive MAE modules in DFB, which not only increases the network depth and expression to improve dehazing performance, but also ensures that the network’s parameter size remains relatively low. The skip connection prevents the gradient from disappearing as the depth increases and speeds up the training. The mathematical model of DFB is expressed as Equation (5).
F D = F D + c o n v ( MAE ( MAE ( MAE ( MAE ( F D ) ) ) ) )
where F D is the input of DFB, F D is the output of DFB.

3.4. MAE Module Mathematical Model

Image features of different scales have different semantic information, so fully extracting and utilizing multi-scale features can effectively improve the quality of dehazing images. Compared with using different ordinary convolution kernels to adjust the size of the receptive field, dilated convolution can achieve the same performance without introducing other calculations and parameters. Mehta et al. [36] proposed the spatial pyramid of dilated convolutions (SPDC), as shown in Figure 3, which can eliminate the grid artifacts well by learning fewer parameters. Although the attention feature enhancement module (AFE) in Section 3.5 adopts the CEM module to improve the receptive field of the network, when the data stream is transmitted in the AFE module, some information will still be lost. Therefore, we propose the MAE module, as shown in Figure 4, to further reduce the loss of detailed information.
The input feature map passes through a local residual block in MAE and is then transmitted in parallel to the AFE module and the SPDC. The feature maps obtained from the two paths are concatenated and output. The multi-scale feature information collected by SPDC is a further supplement to the information obtained by the AFE module, which reduces the loss of haze information and high-frequency information during the transmission of the data stream through the AFE module, and eliminates grid artifacts. The MAE module mathematical model is shown in formula (6).
F M = c o n v ( c a t ( AFE ( F M + c o n v ( F M ) ) , c o n v ( SPDC ( F M + c o n v ( F M ) ) ) ) )
where F M is the input of the MAE module, F M is the output of the MAE module.

3.5. AFE Module Mathematical Model

In order to make up for the loss of some details in the existing attention module during image processing, we propose the AFE module, which is mainly composed of a local residual block, CA, PA and CEM, and the structure is shown in Figure 5. In AFE, local residual learning can avoid haze and low-frequency areas, and focus on effective information. CA uses maximum pooling and average pooling to aggregate channel information and fuse them, and adaptively assigns the weight of fusion information. In CA, inspired by Woo et al. [37], we improve the traditional channel attention mechanism, while using average pooling and max pooling to obtain finer channel attention and retain more detailed information of the image. PA [30] can flexibly focus on dense haze and high-frequency information, and has high sensitivity to images with uneven haze distribution. Then, we use the CEM module to increase the receptive field, improve the relevance of contextual information, and further obtain more detailed information. The AFE module mathematical model is shown by Equation (7).
F A = δ ( c o n v ( F A ) ) + C A ( c o n v ( F A + δ ( c o n v ( F A ) ) ) ) + C E M ( P A ( c o n v ( F A + δ ( c o n v ( F A ) ) ) ) )
where F A is the input of the AFE module, F A is the output of the AFE module, δ is the Relu nonlinear function.
Our improved CA is shown in Figure 1. We combine the average pooling mechanism and the maximum pooling mechanism in parallel to obtain different channel features of the image in two ways and then fuse them to preserve more detailed features. Firstly, we aggregate the spatial information of the feature map using average pooling and max pooling to obtain different channel descriptors. Average pooling and max pooling are represented by Equations (8) and (9), respectively.
g a = H a ( F c ) = 1 H × W i = 1 H j = 1 w X c ( i , j )
g m = H m ( F c ) = M a x ( X c ( i , j ) )
where F c is the input feature map, X c ( i , j ) is the value of the c -th channel X c at the ( i , j ) position, H a is the average function, H m is the maximum function.
Secondly, in order to focus on more image details, two convolutional layers and a Rule activation function are used to fuse the channel features on the two paths. The fusion function is expressed as:
C f = c o n v ( δ ( c o n v ( g a ( F c ) ) ) ) + c o n v ( δ ( c o n v ( g m ( F c ) ) ) )
where C f is the fused channel feature. δ is the Relu function.
Then, we use the sigmoid function to assign the weights of different channels, the expression is as Equation (11).
A c = σ ( C f )
where A c is the weight value of the fusion feature, σ is the sigmoid function.
Finally, the input feature map F c is element-wise multiplied by the weight value A c to obtain the output result F c , as shown by Equation (12).
F c = F c A c
PA module can pay attention to thick-hazed pixels and high-frequency image region. As is shown in Figure 6, It contains two 1 × 1 convolutional layers, Relu and sigmoid activation function. The shape changes from C × H × W to 1 × H × W . PA is shown by Equation (13)
P A = σ ( c o n v ( δ ( c o n v ( F P A ) ) ) )
where F P A is the input feature map of PA. Finally, F P A and PA are multiplied element-wise to obtain the output F P A of PA.
F P A = F P A P A

3.6. Context Enhancement Module Mathematical Model

Although PA can focus on high-frequency information, the receptive field of PA in shallow features is limited, making it difficult to ensure that subsequent feature learning has sufficient context information. In order to capture more context information, we use CEM based on the enhanced attention mechanism, as is shown Figure 1. The CEM module mathematical model is shown by Equations (15)–(17).
f = c o n v ( K )
K = u p ( c o n v ( Re lu ( c o n v ( Re lu ( c o n v ( max ( c o n v ( f ) ) ) ) ) ) ) )
K = = K σ ( c o n v ( c o n v ( c o n v ( K ) ) K ) )
where K and K are the input and output of the CEM module, respectively. f is the output of input K through a 1 × 1 conv, K is the output of six consecutive operations of a 3 × 3 c o n v with a step size of 2, Max Pooling, two consecutive 3 × 3 c o n v with a step size of 1 and the Relu activation function, a 3 × 3 c o n v with a step size of 1 and Upsample. Max Pooling is used to increase the receptive field, enhance contextual semantic information, and reduce redundant information. Its convolution kernel size is 7 and the step size is 3. max is Max Pooling, and u p is Upsample.

3.7. Loss Function

To normalize the learning direction of our proposed network during training, we use the combination of L 1 loss and perceptual loss as the loss function for the whole training process. The L 1 loss is used to correct the difference of pixels between images, and the perceptual loss is used to normalize the human visual perception effect. The total loss function is expressed by Equation (18):
L o s s = λ 1 L 1 + λ p L p
where λ 1 and λ p are the weights of L 1 loss and perceptual loss, respectively. In this paper, we use the L 1 function as the main loss function. The perceptual loss function fine-tunes the dehazed image and the clear image at the depth feature level. According to the training experience, the values of λ 1 and λ p are set to 1 and 0.04, respectively.

3.7.1. L 1 Loss

The L 1 loss function, also known as the mean absolute value error, is a regression model that compares differences pixel by pixel and takes the absolute value. Its mathematical expression is shown by Equation (19)
L 1 = 1 3 N n = 1 N c = 1 3 | I c ( n ) I c g t ( n ) |
where N is the total number of pixels, c is the number of channels, I c ( n ) is the dehazed image restored by the network, and I c g t ( n ) is the original image without haze. The L 1 function will create sparse features, reset the weight of useless features to 0, and have a feature selection effect.

3.7.2. Perceptual Loss

In image restoration tasks, perceptual loss is widely applied to the perceptual difference between two images, which can obtain extremely rich, detailed information. In this paper, VGG-16 [38] is used as the pre-training network for perceptual loss, and the perceptual loss is calculated by the feature map output by VGG-16, which can be expressed as
L p = i = 1 3 1 C j H j W j φ j ( J ) φ j ( J d e h a z e )
where φ j is the feature map of layer j in the VGG-16 network, the size is C j × H j × W j , J is the clear image, and J d e h a z e is the dehazed image.

4. Experimental Results

In this section, first we introduce the dataset and experimental details. The performance of our proposed method is then evaluated on synthetic and real-world datasets and compared with state-of-the-art dehazing methods. Finally, the effectiveness of our proposed module is demonstrated through ablation studies.

4.1. Datasets

We conducted experiments on the publicly available Realistic Single Image Dehazing(RESIDE) dataset [39]. The dataset consists of five subsets, including two training sets and three testing sets. The training sets are the indoor training set (ITS) and the outdoor training set (OTS); the testing sets are the synthetic objective testing set (SOTS), the hybrid subjective testing set (HSTS), and the real-world task-driven testing set (RTTS). For the train, we selected the OTS as the main training set, which contains 8970 clear outdoor images, each of which is combined with 35 hazy images, resulting in a total of 313,950 synthetic hazy images. To further verify the dehazing performance in real scenes, the real-world NHHAZE dataset [40] was also selected for training. For the test, the SOTS-outdoor are used for tests of synthetic hazy images (SOTS includes SOTS-indoor and SOTS-outdoor datasets), while the NHHAZE and RTTS datasets are used for testing real hazy images.

4.2. Implementation Details

We employed the Pytorch framework with NVIDIA RTX8000 GPU on both training and testing stages. The proposed network was trained in RGB channels and augment the training dataset randomly rotated by 90, 180, 270 degrees and horizontal flip. Images were resized to 240 × 240 through preprocessing, and the ADAM optimizer was implemented with a batch size of 4. The whole network is trained for 5 × 10 5 steps on the OTS training set and NHHAZE training set.
The initial learning rate is set to 0.0001, we adopt the cosine annealing strategy [41] to adjust the learning rate from the initial value to 0 by following the cosine function. Expressly, we assume the total step is τ and the initial learning rate is η , the learning rate η s will be updated adaptively by the following strategy.
η s = 1 2 ( 1 + cos ( τ s π τ ) ) η
We train continuously and plot the learning curves by obtaining PSNR and SSIM scores every 5000 steps, as shown in Figure 7. We use skip connections to prevent the gradient from disappearing. When training for 5 × 10 5 steps, PSNR and SSIM level off, and we stop learning. When training for 4.45 × 10 5 steps, the PSNR score and SSIM score reach the maximum, and we select the model at this time as our best model.

4.3. Experimental Results on Synthetic Hazy Images

In this section, we compare our proposed architecture with the following state-of-the-art dehazing methods: DCP [16], CAP [17], AOD-Net [26], Dehaze-Net [24], FFA-Net [30], and GridDehaze-Net [31].
We use Peak Signal-to-Noise Ratio (PSNR), Structural Similarity (SSIM) and Learned Perceptual Image Patch Similarity (LPIPS) among image quality assessment methods to evaluate the dehazing performance of other state-of-the-art methods. The objective results of various dehazing methods on the synthetic dataset are shown in Table 1. It can be observed that the proposed method achieves the best dehazing performance compared to other methods, with PSNR and SSIM scores of 33.74 and 0.9843 on SOTS-outdoor. The LPIPS distance is only 0.0040. This objectively proves the dehazing advantage of the proposed method on synthetic hazy images. Moreover, our proposed method has fewer parameters (params) and floating point operations (Flops). Compared with FFA-Net, our method has 1.39 × 10 6 fewer params and 331.55 × 10 9 fewer Flops, but achieves better dehazing performance.
The subjective results of various dehazing methods on the SOTS-outdoor test set are shown in Figure 8. It can be found that compared with the ground truth image, CAP, DCP and Dehaze-net obviously dehaze excessively, such as the road surface in the first row of Figure 8b–d. AOD-net has the phenomenon of color distortion and dehazing image blur, as shown in the fourth and fifth rows of Figure 8e. GridDehaze-Net and FFA-Net achieve results that are closer to the ground truth images, as shown in Figure 8f,g. However, upon closer inspection, it can be observed that GridDehaze-Net’s and FFA-Net’s dehazed results exhibit subtle haze residue and are generally whiter than the ground truth image, as shown in the ground in the first and second rows of Figure 8f,g. The proposed method captures more detailed information in the images, and the dehazed results are most similar to the ground truth images.

4.4. Experimental Results on Real-World Hazy Images

To verify the dehazing performance of proposed method on real hazy images, we conducted tests on the NHHAZE and RTTS datasets, respectively. The subjective results on the NHHAZE dataset are shown in Figure 9. We can find that the effect of haze removal is not obvious for DCP, CAP, Dehaze-Net, and AOD-Net methods, with a large amount of hazy remaining and color distortion. GridDehaze-Net has obvious color distortion. Both FFA-net and our method have achieved a better dehazing effect. However, compared with the proposed method in this paper, FFA-net has more hazy residue, such as the upper left corner of the trees in the first row of Figure 9g. Our method removes haze obviously and retains detailed information such as color that is closer to the real image on the ground.
We also performed quantitative evaluations on the real hazy images in NHHAZE, as shown in Table 2, to demonstrate the effectiveness of our method further objectively for dehazing. It can be observed that in four randomly selected real hazy images in NHHAZE, our method achieves significantly better PSNR and SSIM scores compared with other dehazing methods. For example, in the first image of Table 2, the PSNR scores are improved by 2.49 and the SSIM scores is improved by 0.0023 compared with the FFA-net, which has the second highest dehazing performance. LPIPS distance is also the shortest among our methods; for example, it is only 0.159 in the fourth row of Table 2.
The subjective results of various dehazing methods on the RTTS dataset are shown in Figure 10 (since the RTTS does not have ground truth images, only qualitative analysis was performed). We can know that DCP, CAP and Dehaze-Net are over-dehazing, the image is overall dark, and there is obvious color distortion, such as the road surface in the third row of Figure 10b–d. AOD-Net exhibits a color cast phenomenon, as shown in the fourth row of Figure 10e. GridDehaze-Net exhibits an uneven dehazing effect and loses the color information of the image, such as the fourth row of Figure 10f. FFA-net leaves a lot of haze, and the dehazing effect is not obvious, as shown in the lawn in the second row of Figure 10g. Our proposed method has a better color fidelity while having a clear dehazing effect.
The subjective results of various dehazing methods in the real world are shown in Figure 11. We can know that DCP and CAP have obvious excessive dehazing phenomenon, such as the reef in the third row of Figure 11b,c. Dehaze-net and AOD-net have obvious color cast phenomena, such as the first rows of Figure 11d,e. GridDehaze-net only performs local dehazing and loses the texture information of the image, such as the upper right corner of the first row and the tree in the second row of Figure 11f. The dehazing effect of FFA-net is not obvious, as shown in the third line of Figure 11g. Our method better preserves the color and texture information of the image.

4.5. Ablation Study

To demonstrate the effectiveness of our proposed method, we conducted an ablation study to analyze the MAE and AFE modules. We cropped the image to 96 × 96 as input with training of 2 × 10 5 steps; other configurations are the same as our implementation details. The quantitative evaluation results of different modules on the SOTS-outdoor dataset are shown in Table 3.
First, we constructed a baseline network without the SPDC and CEM moules, represented as “Baseline” in Table 3. Then, we added the SPDC module and CEM module separately to the baseline network, represented as “Baseline+SPDC” and “Baseline+CEM” in Table 3, respectively. Finally, we added both the SPDC and CEM modules to the baseline network, which is our proposed network model, represented as “Baseline+SPDC+CEM” in Table 3.
According to the quantitative results of the ablation study on the SOTS-outdoor dataset in Table 3, we can draw the following conclusions:
(1)
When our proposed method does not include the SPDC and CEM modules, that is, the baseline network, the dehazing results are the worst.
(2)
When the SPDC module is added to the Baseline network, compared with the dehazing results of the Baseline network, the PSNR scores is increased by 2.26, the SSIM scores is increased by 0.0152, and the LPIPS distance decreased by 0.0028, which proves the effectiveness of the SPDC module in improving the dehazing performance.
(3)
When adding the CEM module on the baseline network, that is, the AFE module we proposed, compared with the dehazing results of the baseline network, the PSNR scores is increased by 2.48, the SSIM scores is increased by 0.0184, and the LPIPS distance decreased by 0.0028, which proves the dehazing performance of the proposed AFE module.
(4)
When both SPDC module and CEM module are added to baseline network, that is, the MAE module proposed in this paper, the network model is our proposed method. The PSNR and SSIM scores are the highest and the LPIPS distance is the shortest in Table 3, indicating that our proposed MAE module and method have the best dehazing performance, demonstrating the superiority of our proposed MAE module and method.

5. Discussion

We developed a new image dehazing network model that demonstrated good dehazing performance. The MAE module and AFE module we proposed use the attention mechanism to adaptively extract haze features, the CEM module enhances contextual information, suppresses redundant information, and dilated convolution expands the receptive field, which can greatly compensate for the loss of detailed information. Therefore, compared with several other state-of-the-art dehazing algorithms such as AOD-net and FFA-net, our network can better capture the color, texture, and other detailed information of the image. Networks such as AOD-net optimize their network models by calculating mean square error (MSE). Our network is the same as FFA-net, using both PSNR and SSIM to optimize the network model. Therefore, our network can directly obtain the best network model corresponding to the maximum PSNR score and SSIM score. In addition, our proposed MAE module and AFE module are universal, which means that they can be easily plugged into network models in other fields. For example, in the field of medical imaging [42], they can retain more detailed information in images, helping doctors to judge the condition more accurately. Our network currently shows good performance in dehazing single images, but has not tested the dehazing performance on hazy video images. In the future, we hope to further expand our method to areas such as video dehazing to achieve real-time recovery of hazy videos.

6. Conclusions

In this paper, we propose an end-to-end multi-scale attention feature enhancement network for single image dehazing, which can well preserve image color, texture, and other detailed information. Around the goal of retaining more detailed information of images, we designed the Attention Feature Enhancement module and the Multi-Scale Attention Enhancement module, which focuses on high-frequency information and haze information. Our proposed network was tested on synthetic and real haze datasets for both qualitative and quantitative evaluation. Experimental results have shown that the proposed method has achieved state-of-the-art results. Through ablation studies, we studied the effectiveness of the different modules proposed.

Author Contributions

Conceptualization, W.D. and C.W.; methodology, W.D. and H.S.; software, experiment, writing—original draft preparation, W.D.; data analysis, investigation, data curation, Y.T.; writing—review and editing, X.X.; validation and supervision, C.W. and X.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the Jilin Province Science and Technology Development Plan Project (no. 20220201074GX), the Jilin Province Industrial Technology Research and Development Project (no. 2023C030-5).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The RESIDE dataset and NH-HAZE dataset are made publicly available for research purposes. For more information, please refer to the websites https://sites.google.com/site/boyilics/website-builder/reside/, and https://data.vision.ee.ethz.ch/cvl/ntire20//nh-haze/ (accessed on 18 December 2022).

Acknowledgments

The authors would like to thank all authors of previous papers for approving the use of their published research results in this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sharma, T.; Debaque, B.; Duclos, N.; Chehri, A.; Kinder, B.; Fortier, P. Deep Learning-Based Object Detection and Scene Perception Under Bad Weather Conditions. Electronics 2022, 11, 563. [Google Scholar] [CrossRef]
  2. Wu, W.; Chang, H.; Zheng, Y.; Li, Z.; Chen, Z.; Zhang, Z. Contrastive Learning-Based Robust Object Detection Under Smoky Conditions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 June 2022; pp. 4295–4302. [Google Scholar]
  3. Amrani, M.; Bey, A.; Amamra, A. New SAR Target Recognition Based on YOLO and Very Deep Multi-Canonical Correlation Analysis. Int. J. Remote Sens. 2022, 43, 5800–5819. [Google Scholar] [CrossRef]
  4. Chen, Z.; Wu, R.; Lin, Y.; Li, C.; Chen, S.; Yuan, Z.; Zou, X. Plant Disease Recognition Model Based on Improved YOLOv5. Agronomy 2022, 12, 365. [Google Scholar] [CrossRef]
  5. Engin, D.; Genç, A.; Kemal Ekenel, H. Cycle-Dehaze: Enhanced Cyclegan for Single Image Dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 825–833. [Google Scholar]
  6. Hodges, C.; Bennamoun, M.; Rahmani, H. Single Image Dehazing Using Deep Neural Networks. Pattern Recogn. Lett. 2019, 128, 70–77. [Google Scholar] [CrossRef]
  7. Song, Y.; He, Z.; Qian, H.; Du, X. Vision Transformers for Single Image Dehazing. IEEE Trans. Image Process. 2023, 32, 1927–1941. [Google Scholar] [CrossRef]
  8. Wang, A.; Wang, W.; Liu, J.; Gu, N. AIPNet: Image-to-Image Single Image Dehazing with Atmospheric Illumination Prior. IEEE Trans. Image Process. 2018, 28, 381–393. [Google Scholar] [CrossRef]
  9. Hide, R. Optics of the atmosphere: Scattering by molecules and particles. Phys. Bull. 1977, 28, 521. [Google Scholar] [CrossRef]
  10. Narasimhan, S.G.; Nayar, S.K. Chromatic Framework for Vision in Bad Weather. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Hilton Head, SC, USA, 15 June 2000; pp. 598–605. [Google Scholar]
  11. Narasimhan, S.G.; Nayar, S.K. Contrast Restoration of Weather Degraded Images. IEEE Trans. Pattern Anal. Mach. Intel. 2003, 25, 713–724. [Google Scholar] [CrossRef]
  12. Zhang, X.; Jiang, R.; Wang, T.; Luo, W. Single image dehazing via dual-path recurrent network. IEEE Trans. Image Process. 2021, 30, 5211–5222. [Google Scholar] [CrossRef]
  13. Chen, Z.; He, Z.; Lu, Z. DEA-Net: Single image dehazing based on detail-enhanced convolution and content-guided attention. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Vancouver, BC, Canada, 18–22 June 2023. [Google Scholar]
  14. Wu, H.; Qu, Y.; Lin, S.; Zhou, J.; Qiao, R.; Zhang, Z.; Ma, L. Contrastive learning for compact single image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Kuala Lumpur, Malaysia, 18–20 December 2021; pp. 10551–10560. [Google Scholar]
  15. Huang, P.; Zhao, L.; Jiang, R.; Wang, T.; Zhang, X. Self-filtering image dehazing with self-supporting module. Neurocomputing 2021, 432, 57–69. [Google Scholar] [CrossRef]
  16. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal. Mach. Intel. 2011, 33, 2341–2353. [Google Scholar]
  17. Zhu, Q.; Mai, J.; Shao, L. A Fast Single Image Haze Removal Algorithm Using Color Attenuation Prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar] [PubMed]
  18. Berman, D.; Avidan, S. Non-local Image Dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1674–1682. [Google Scholar]
  19. Fattal, R. Single Image Dehazing. ACM Trans. Graph. 2008, 27, 1–9. [Google Scholar] [CrossRef]
  20. Dalborgo, V.; Murari, T.B.; Madureira, V.S.; Moraes, J.G.L.; Bezerra, V.M.O.; Santos, F.Q.; Monteiro, R.L. Traffic Sign Recognition with Deep Learning: Vegetation Occlusion Detection in Brazilian Environments. Sensors 2023, 23, 5919. [Google Scholar] [CrossRef]
  21. Esteva, A.; Chou, K.; Yeung, S.; Naik, N.; Madani, A.; Mottaghi, A.; Socher, R. Deep Learning-Enabled Medical Computer Vision. NPJ Digit. Med. 2021, 4, 5. [Google Scholar] [CrossRef]
  22. Ma, W.; Liu, Z.; Kudyshev, Z.A.; Boltasseva, A.; Cai, W.; Liu, Y. Deep Learning for The Design of Photonic Structures. Nat. Photonics 2021, 15, 77–90. [Google Scholar] [CrossRef]
  23. Tran, K.A.; Kondrashova, O.; Bradley, A.; Williams, E.D.; Pearson, J.V.; Waddell, N. Deep Learning in Cancer Diagnosis, Prognosis and Treatment Selection. Genome Med. 2021, 13, 152. [Google Scholar] [CrossRef]
  24. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. Dehazenet: An End-to-End System for Single Image Haze Removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef]
  25. Ren, W.; Liu, S.; Zhang, H.; Pan, J.; Cao, X.; Yang, M.H. Single image dehazing via multi-scale convolutional neural networks. In Proceedings of the 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Springer: Cham, Switzerland, 2016; pp. 154–169. [Google Scholar]
  26. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. AOD-Net: All-in-One Dehazing Network. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 4780–4788. [Google Scholar]
  27. Liao, Y.; Su, Z.; Liang, X.; Qiu, B. Hdp-Net: Haze Density Prediction Network for Nighttime Dehazing. In Proceedings of the Pacific Rim Conference on Multimedia, Hefei, China, 21–22 December 2018; Springer: Cham, Switzerland, 2018; pp. 469–480. [Google Scholar]
  28. Chen, D.; He, M.; Fan, Q.; Liao, J.; Zhang, L.; Hou, D.; Hua, G. Gated context aggregation network for image dehazing and deraining. In Proceedings of the 2019 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 7–11 January 2019; pp. 1375–1383. [Google Scholar]
  29. Shao, Y.; Li, L.; Ren, W.; Gao, C.; Sang, N. Domain Adaptation for Image Dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 2808–2817. [Google Scholar]
  30. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature Fusion Attention Network for Single Image Dehazing. In Proceedings of the 2020 AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 11908–11915. [Google Scholar]
  31. Liu, X.; Ma, Y.; Shi, Z.; Chen, J. Griddehazenet: Attention-based multi-scale network for image dehazing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27–31 October 2019; pp. 7314–7323. [Google Scholar]
  32. Wang, N.; Cui, Z.; Su, Y.; He, C.; Li, A. Multiscale supervision-guided context aggregation network for single image dehazing. IEEE Signal Process. Lett. 2021, 29, 70–74. [Google Scholar] [CrossRef]
  33. Sheng, J.; Lv, G.; Du, G.; Wang, Z.; Feng, Q. Multi-scale residual attention network for single image dehazing. Digit. Signal Process. 2022, 121, 103327. [Google Scholar] [CrossRef]
  34. Zhou, Y.; Chen, Z.; Sheng, B.; Li, P.; Kim, J.; Wu, E. AFF-Dehazing: Attention-based feature fusion network for low-light image Dehazing. Comput. Animat. Virtual Worlds 2021, 32, 3–4. [Google Scholar] [CrossRef]
  35. Wang, S.; Zhou, T.; Lu, Y.; Di, H. Contextual Transformation Network for Lightweight Remote-Sensing Image Super-Resolution. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5615313. [Google Scholar] [CrossRef]
  36. Mehta, S.; Rastegari, M.; Caspi, A.; Shapiro, L.; Hajishirzi, H. Espnet: Efficient Spatial Pyramid of Dilated Convolutions for Semantic Segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 552–568. [Google Scholar]
  37. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  38. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv 2014, arXiv:1409.1556. [Google Scholar]
  39. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking Single-Image Dehazing and Beyond. IEEE Trans. Image Process. 2019, 1, 492–505. [Google Scholar] [CrossRef] [PubMed]
  40. Ancuti, C.O.; Ancuti, C.; Timofte, R. NH-HAZE: An Image Dehazing Benchmark with Non-Homogeneous Hazy and Haze-Free Images. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1798–1805. [Google Scholar]
  41. He, T.; Zhang, Z.; Zhang, H.; Zhang, Z.; Xie, J.; Li, M. Bag of Tricks for Image Classification with Convolutional Neural Networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 558–567. [Google Scholar]
  42. Amyar, A.; Modzelewski, R.; Vera, P.; Morard, V.; Ruan, S. Multi-task multi-scale learning for outcome prediction in 3D PET images. Comput. Biol. Med. 2022, 151, 106208. [Google Scholar] [CrossRef]
Figure 1. The architecture of the MAFE.
Figure 1. The architecture of the MAFE.
Sensors 23 08102 g001
Figure 2. Structure of the DFB module.
Figure 2. Structure of the DFB module.
Sensors 23 08102 g002
Figure 3. Structure of the SPDC.
Figure 3. Structure of the SPDC.
Sensors 23 08102 g003
Figure 4. Structure of the MAE module.
Figure 4. Structure of the MAE module.
Sensors 23 08102 g004
Figure 5. Structure of the AFE module.
Figure 5. Structure of the AFE module.
Sensors 23 08102 g005
Figure 6. Structure of PA.
Figure 6. Structure of PA.
Sensors 23 08102 g006
Figure 7. PSNR and SSIM learning curves. The red line and blue line are the change curves of PSNR and SSIM respectively as steps increase.
Figure 7. PSNR and SSIM learning curves. The red line and blue line are the change curves of PSNR and SSIM respectively as steps increase.
Sensors 23 08102 g007
Figure 8. Visual results comparison on SOTS-outdoor dataset.
Figure 8. Visual results comparison on SOTS-outdoor dataset.
Sensors 23 08102 g008
Figure 9. Visual results comparison on NHHAZE dataset.
Figure 9. Visual results comparison on NHHAZE dataset.
Sensors 23 08102 g009
Figure 10. Visual results comparison on RTTS dataset.
Figure 10. Visual results comparison on RTTS dataset.
Sensors 23 08102 g010
Figure 11. Visual results comparison on real-world images.
Figure 11. Visual results comparison on real-world images.
Sensors 23 08102 g011
Table 1. Quantitative comparisons on synthetic datasets.
Table 1. Quantitative comparisons on synthetic datasets.
MethodSOTS-Outdoor Params / × 10 6 Flops / × 10 9
PSNRSSIMLPIPS
DCP21.810.85830.0527
CAP22.090.88290.0427
AOD-Net20.290.87650.08800.0020.101
Dehaze-Net22.460.85140.03900.0080.450
GridDehaze-Net30.860.98190.00530.956452.01
FFA-Net33.380.98040.00494.4561010.86
Ours33.740.98430.00403.070679.31
Table 2. Quantitative comparisons on NHHAZE dataset (PSNR/SSIM/LPIPS).
Table 2. Quantitative comparisons on NHHAZE dataset (PSNR/SSIM/LPIPS).
Image No.DCPCAPDehaze-NetAOD-NetGridDehaze-NetFFA-NetOur Method
113.36/0.4838/0.38613.52/0.4881/0.39612.48/0.4715/0.41013.38/0.4589/0.43912.77/0.4854/0.41318.65/0.6605/0.21721.14/0.6628/0.213
211.83/0.4087/0.49212.62/0.4060/0.49811.16/0.3726/0.53414.56/0.4253/0.53712.13/0.4005/0.49718.86/0.6109/0.28820.88/0.6261/0.282
311.69/0.4977/0.46012.10/0.4994/0.44810.77/0.4882/0.47711.50/0.4613/0.47510.42/0.5044/0.44617.00/0.6621/0.26418.09/0.6647/0.260
413.84/0.5359/0.34514.15/0.5303/0.34212.94/0.5231/0.35714.27/0.4562/0.43413.65/0.5416/0.38519.85/0.7259/0.16722.34/0.7282/0.159
Table 3. Ablation study on SOTS-outdoor dateset.
Table 3. Ablation study on SOTS-outdoor dateset.
MethodPSNRSSIMLpips
Baseline26.710.95140.0269
Baseline + SPDC28.970.96660.0241
Baseline + CEM29.190.96980.0241
Baseline + SPDC + CEM29.500.97060.0232
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, W.; Wang, C.; Sun, H.; Teng, Y.; Xu, X. Multi-Scale Attention Feature Enhancement Network for Single Image Dehazing. Sensors 2023, 23, 8102. https://doi.org/10.3390/s23198102

AMA Style

Dong W, Wang C, Sun H, Teng Y, Xu X. Multi-Scale Attention Feature Enhancement Network for Single Image Dehazing. Sensors. 2023; 23(19):8102. https://doi.org/10.3390/s23198102

Chicago/Turabian Style

Dong, Weida, Chunyan Wang, Hao Sun, Yunjie Teng, and Xiping Xu. 2023. "Multi-Scale Attention Feature Enhancement Network for Single Image Dehazing" Sensors 23, no. 19: 8102. https://doi.org/10.3390/s23198102

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop