Next Article in Journal
A Novel Decoupled Feature Pyramid Networks for Multi-Target Ship Detection
Next Article in Special Issue
Deep Power Vision Technology and Intelligent Vision Sensors
Previous Article in Journal
Visual-Based Children and Pet Rescue from Suffocation and Incidence of Hyperthermia Death in Enclosed Vehicles
Previous Article in Special Issue
Fittings Detection Method Based on Multi-Scale Geometric Transformation and Attention-Masking Mechanism
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Multi-Patch Hierarchical Transmission Channel Image Dehazing Network Based on Dual Attention Level Feature Fusion

College of Engineering, Sichuan Normal University, Chengdu 610101, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(16), 7026; https://doi.org/10.3390/s23167026
Submission received: 17 July 2023 / Revised: 4 August 2023 / Accepted: 4 August 2023 / Published: 8 August 2023
(This article belongs to the Special Issue Deep Power Vision Technology and Intelligent Vision Sensors)

Abstract

:
Unmanned Aerial Vehicle (UAV) inspection of transmission channels in mountainous areas is susceptible to non-homogeneous fog, such as up-slope fog and advection fog, which causes crucial portions of transmission lines or towers to become fuzzy or even wholly concealed. This paper presents a Dual Attention Level Feature Fusion Multi-Patch Hierarchical Network (DAMPHN) for single image defogging to address the bad quality of cross-level feature fusion in Fast Deep Multi-Patch Hierarchical Networks (FDMPHN). Compared with FDMPHN before improvement, the Peak Signal-to-Noise Ratio (PSNR) and Structural Similarity Index Measure (SSIM) of DAMPHN are increased by 0.3 dB and 0.011 on average, and the Average Processing Time (APT) of a single picture is shortened by 11%. Additionally, compared with the other three excellent defogging methods, the PSNR and SSIM values DAMPHN are increased by 1.75 dB and 0.022 on average. Then, to mimic non-homogeneous fog, we combine the single picture depth information with 3D Berlin noise to create the UAV-HAZE dataset, which is used in the field of UAV power assessment. The experiment demonstrates that DAMPHN offers excellent defogging results and is competitive in no-reference and full-reference assessment indices.

1. Introduction

UAVs have been increasingly employed in power inspection to find safety problems effectively [1]. However, in hilly regions, advection fog, uphill fog, and valley fog are frequently encountered [2,3], causing critical portions of transmission lines or towers to become fuzzy or even wholly concealed and decreasing fault detection accuracy. Image-defogging technology can be used to address the appeal issues. However, the non-homogenous fog is challenging for the current homogenous fog removal method. Additionally, the initial non-homogeneous defogging method FDMPHM exploits residual connections between several levels and ignores the issues with channel redundancy and unequal pixel distribution in cross-level fusion. Based on this, we suggest the Dual Attention Level Feature Fusion Multi-Patch Hierarchical Network (DAMPHN), which aims to enhance the cross-level fusion method of FDMPHN and produce superior defogging effects. Haze non-uniformity is not considered in power inspection image defogging studies due to a lack of non-homogeneous haze datasets. Therefore, to create a dataset that may represent non-homogeneous haze in mountainous places (UAV-HAZE), this paper ingeniously combines image depth measurements with 3D Berlin noise. The suggested DAMPHN performs better in color preservation and haze removal than the other four advanced approaches and can complete the picture preprocessing of transmission channels, according to numerous experiments on three open datasets and UAV-HAZE.

1.1. Related Work

Model-based parameter estimation and model-free picture enhancement methods are currently the main single-image fog removal categories. Additionally, future images for machine vision services will be of higher quality because of advancements in CCD imager technology [4]. Some researchers have used the image defog technique to preprocess photos based on high-quality photographs for transmission channels.

1.1.1. Model-Based Parameter Estimation Method

By predicting the transmission matrix t ( x ) and global atmospheric light A from the haze graph J ( x , λ ) , these approaches, based on the atmospheric scattering model [5], provide images I ( x , λ ) that are devoid of haze. In Equation (1), the atmospheric scattering model is displayed.
I ( x , λ ) = t x J x , λ + A 1 t x
t x = e β λ d x
where d x denotes the depth of the scene and β λ the scattering coefficient. Both the early dark channel prior (DCP) [6] and the color decay prior (CAP) [7] were put out and offered concepts for further study. Convolutional neural networks (CNN) were later developed, and Cai et al. [8] used CNNs with various kernel parameters for the first time to extract the distinctive information of dark channel, color attenuation, maximum contrast, and hue disparity to solve the parameters. Li et al. [9] equalized t x and A as a parameter based on Formula (1) and applied CNN and residual connection to get this parameter. Zhang et al. [10] used the Dense-Net and U-net networks, respectively. A Densely Connected Pyramid Dehazing Network (DCPDN) was subsequently proposed based on the joint discriminator of adversarial networks and the optimization parameter estimate of the edge retention loss function. To achieve adaptive fusion, Li et al. [11] employed a multi-stage deep convolutional network to estimate t x and A and added a memory network and a two-level attention mechanism to determine the weight of findings at each stage. To filter haze residuals step by step and achieve dehazing, Li et al. [12] modified Formula (1) to be task-oriented and assembled recurrent neural networks based on encoder-decoder and space. Bai et al. [13], who combined t x and A into a single parameter and calculated it using the depth pre-defamer. The progressive feature fusion module and the picture recovery module were created to improve parameter estimation.

1.1.2. Model-Free Image Enhancement Method

This technique uses a coding-decoding structure to directly learn the link between the haze/clear image mapping and integrates attention mechanisms, feature fusion, and other techniques to enhance the dehazing performance. Das et al. [14] introduced the Fast Deep Multi-Patch Hierarchical Network (FDMPHN) and Fast Multi-Scale Hierarchical Network (FDMSHN) by improving the loss function, which was inspired by literature [15]. According to Wang et al. [16], a heterogeneous twin network was suggested, U-Net was used to extract haze features, and a detail enhancer network was set up to improve image details. Liu et al. [17] proposed an attention-based multi-scale defogging network (GridDehazeNet), which introduced a channel attention mechanism to improve feature fusion ability among multiple scales. A feature fusion attention network with a channel and pixel focus that prioritizes high-frequency and dense hazy areas was proposed by Qin et al. [18]. To improve the ability to extract edge texture features, Wang et al. [19] created the edge branch module based on the multi-level attention dehazing module and the feature fusion module based on Laplace gradient prior knowledge. Using extended convolution in the multi-scale part, channel attention mechanism in the cross-level fusion part, and frequency domain loss in the loss function part, Yang et al.’s [20] combination of FDMPHN and FDMSHN methods to obtain dense feature maps produced good results. A transfer attention technique was created by Wang et al. [21] to deal with non-uniform noise in images. To focus on the non-uniform hazy region and address the issues of artifacts and excessive smoothing, Zhao et al. [22] developed a dynamic attention module based on the dual attention mechanism. Guo et al. [23] suggested a self-paced half-course learning-driven attention image-generating technique based on the dual attention mechanism to enhance the ability to clear regions with considerable brightness disparities of fog.

1.1.3. Transmission Channel Image Dehazing Method

Recently, researchers have used it in power inspection after taking inspiration from the appeal algorithm. Liu et al. [24] created their own UAV picture collection for transmission line inspection and used the DCPDN approach to achieve dehazing. To address the drawbacks of the DCP method, Zhang et al. [25] divided the sky region by fusing the Canny operator and gradient energy function to obtain a more accurate atmospheric light value, and Zhai et al. [26] optimized the quadtree segmentation method. Both techniques were then applied to the image dehazing of transmission line monitoring systems. To remove haze from photographs of an insulator umbrella disk in transmission lines, Xin et al. [27] coupled a limited-contrast adaptive histogram equalization method with the dark channel, bright channel, and these methods. Gao et al.’s [28] use of DCP to remove haze from fixed-point monitoring photographs of a tower or pole was likewise based on this technique. Yan et al. [29] created their dataset for UAV power inspection and used FDMPHN to achieve dehazing.

1.2. Motivation and Contribution

The model-based parameter estimate methods produces improved outcomes in the area of picture fog removal. However, the overall image that DCP restored is dark, and color distortion can easily happen in areas of bright light. The reduction impact is weak when the depth of field shift in the image is not visible or when there is haze, as CAP is dependent on the color saturation of the image. To maximize the fog removal effect, later researchers used CNN to estimate the parameters t ( x )   and A . However, both the parameter estimation methods based on CNN [8,10,11] and the parameter estimation method after the improved atmospheric scattering model [9,12,13] are subject to artifacts, color distortion, and haze residues because of the shortcomings of the atmospheric scattering model. Although the model-free image enhancement methods are not limited by the model, it depends on the ability of the network to extract and fuse the haze features. Only residual connections are used in the multi-patch network FDMPHN for cross-level feature fusion, disregarding channel differences and pixel distribution non-uniformity. Therefore, when the non-uniform characteristics of haze or the fog area are strong, it is easy for haze residue and detail blur to appear. Later researchers enhanced the network’s capacity for feature extraction by improving the attention mechanism [17,18,19,20,21,22,23], but it was also challenging to address the issue of non-uniform fog.
In the area of fog removal in power inspection images, Refs. [24,25,26,27,28] all use a uniform haze dataset created based on an atmospheric scattering model as the foundation for their analyses, neglecting the non-uniform characteristics of haze distribution in natural settings. As a result, it is only appropriate for processing images with uniform haze distribution. It performs poorly when dealing with powerful light sources and non-uniform haze, and the image quality after recovery is also subpar. Furthermore, power inspection picture fog removal is still in the uniform haze removal stage, and it is challenging to make progress due to the relative paucity of non-uniform haze datasets [30]. Therefore, this paper suggests a Dual Attention Level Feature Fusion Multi-Patch Hierarchical Network (DAMPHN) to enhance the defogging effect of UAV inspection photos of transmission lines in mountainous terrain. This work’s key contributions can be summed up as follows:
  • It is suggested to use a Dual Attention Level Feature Fusion Multi-Patch Hierarchical Network (DAMPHN) that combines an encoder-decoder module with a Dual Attention Level Feature Fusion (DA) module. The experimental results show that the network has low color distortion and a good defogging effect.
  • DA module is proposed. DA makes use of channel attention, pixel attention, and residual connection to enhance the multi-patch layered network’s cross-level feature function strategy. The DA module has strong feature fusion capabilities, as demonstrated by numerous ablation tests.
  • By calculating picture depth information and inserting 3D Berlin noise of various frequencies, 2225 pairs of non-homogeneous haze/clear images datasets are constructed based on the actual situation. The dataset can, as closely as possible, mimic the characteristics of haze dispersal in mountainous regions. Later, it is employed to support DAMPHN training and testing, which can enhance the ability of UAV inspection photos of transmission lines in mountainous locations to remove fog.
Figure 1 illustrates the specifics of our implementation strategy for DAMPHN-based image preprocessing of mountain areas’ transmission channel images. Based on this, Section 2 details the DAMPHN network structure. It also includes the encoder-decoder and DA module’s unique construction and the loss function needed for network training. The datasets required for the ablation and application experiments and the creation of the training parameters are described in Section 3. The usefulness of the suggested DA and DAMPHN is first demonstrated in Section 4 through several ablation experiments, after which many algorithms are trained and tested using real haze photos of mountain power transmission routes and UAV-HAZE datasets. Section 5 discusses and analyzes the experimental results. In Section 6, several conclusions are made.

2. Materials and Methods

In this study, the encoder-decoder and DA module-based DAMPHN are suggested. This section’s first paragraph introduces DAMPHN’s architecture and design principles, as well as those of its submodules. The training and optimization of the DAMPHN loss function are covered in the second section.

2.1. DAMPHN

DAMPHN network is a multi-level structure, and each level comprises corresponding encoders and decoders. The potential of hierarchical feature fusion is further enhanced by a Dual Attention Level Feature Fusion module (DA). Figure 2 displays the structure in its entirety. Figure 2 depicts DAMPHN with i hierarchical structure, where each level processes 4, 2, and 1 picture blocks, respectively, and when i = 1 , 2 , 3 . The j block of level i is represented as I i , j if the input image is I . The first layer then divides I into 4 blocks, identified as I 1 , 1 , I 1 , 2 , I 1 , 3 , and I 1 , 4 , both vertically and horizontally. I is divided vertically into two blocks, designated as I 2 , 1 and I 2 , 2 , by the second stratum. I is directly inputted into the third layer, which is represented as I 3 , 1 .
The pair of encoder decoders that make up each level are denoted as E n c i and D e c i , respectively. The encoding feature Q i , j can be retrieved after the input picture I i , j has sequentially been through the encoder and DA module. In particular, see Equation (3).
Q i , j = C a t [ E n c i ( I i , 2 j 1 ) , E n c i ( I i , 2 j ) ] , i = 1 , j ϵ 1 , 2 E n c i ( D A ( I i , j , J i 1 , j ) ) , i = 2 , j ϵ 1 , 2 E n c i ( D A ( I i , j , J i 1 , j ) ) , i = 3 , j = 1
The local feature output J i , j of all levels can be acquired after the DA module and decoder. J 3 , 1 represents the final dehazing image after DAMPHN feature extraction from the local to the overall concept. The specifics are presented in Equation (4):
J i , j = D e c i ( Q i , j ) , i = 1 , j ϵ 1 , 2 D e c i ( D A ( C a t [ Q i , j , Q i , 2 j ] , C a t [ Q i 1 , j , Q i 1,2 j ] ) ) , i = 2 , j = 1 D e c i ( D A ( Q i , j , C a t [ Q i 1 , j , Q i 1 , 2 j ] ) ) , i = 3 , j = 1

2.1.1. Encoder-Decoder

The encoder is used to extract the feature data from the image, while the decoder reconstructs the image using the feature data. Three convolution layers and three residual modules (Resblock × 3) make up the encoder in this study. The decoder has a similar design to the encoder, with three residual modules, two transposed convolution layers, and one convolution layer. In order to generate a haze-free image and restore the image scale, decoder transposition convolution is utilized. Figure 3 depicts its network structure.

2.1.2. DA Module

After going through the encoder-decoder during the hierarchical fusion process, the local feature J i , j is produced from the foggy picture I input at the first and second levels. The convolution transformation of Q i , j yields each channel of J i , j . As a result, the residual connection in the original FDMPHN network is employed directly in cross-level fusion, and the uneven and redundant channel direction in the fusion feature process is not considered. Additionally, the residual splicing method does not consider the uneven distribution of picture pixels, and the encode-decoder in the original FDMPHN network relies on pixel domain mapping to understand the intricate relationship between the hazy image and the clear image. This led to the development of the DA module provided in this paper, as seen in Figure 4.
The channel domain feature response is first collected by adding the channel attention layer, and subpar or duplicated features are suppressed. Second, by including a pixel attention layer to concentrate on regions of the image with uneven pixel distribution, we may enhance the fusion process’ attention to dense haze or high-frequency regions. After stitching, input the channel attention layer (Ca_layer) and pixel attention layer (Pa_layer), assuming that the feature picture of the current level is F C ϵ R H × W × C and the feature picture of the previous level is F U ϵ R H × W × C . F C A ϵ R H × W × C and F P A ϵ R H × W × C are obtained. Finally, this paper obtains the output F of the final DA module using the convolution joint processing channel and the outcomes of pixel attention processing to make up for the information lost in the extraction process of dual attention layers.
F C A = C a _ l a y e r ( C a t [ F C , F U ] )
F P A = P a _ l a y e r ( F C A )
F = C a t [ c o n v ( F P A ) , F P A , F C A , C a t [ F C , F U ] ]

2.2. Loss of DAMPHN

The total loss function L of DAMPHN is shown in Equation (8), where, respectively, L r , L p , and L t v stand for reconstruction loss, perception loss, and total variational loss.
L = α r L r + α p L p + α t v L t v
  • Reconstruction loss L r ;
Determine the difference between the clear pictures J pixel and the N DAMPHN defogging images J n . MAE and MSE are combined linearly. L r can be written as:
L r = α r 1 1 N i = 1 N J n J + α r 2 1 N i = 1 N J n J 2
  • Perception loss L p ;
The VGG16 network was used to calculate features using the pre-trained model. The network’s convolution layers (Conv1-2, Conv2-2, and Conv3-2) were utilized to calculate differences, designated as φ ( · ) , and extract features. L p is written as:
L p = 1 C K W K H K K = 1 3 φ K ( J n ) φ K ( J )
  • Total variation loss L t v .
L t v is calculated by computing the gradient amplitude of the dehazing image to reduce noise and keep the image smooth. x ( · ) and y ( · ) in Equation (11), respectively, are used to obtain the gradient matrix of the picture in the horizontal and vertical directions.
L t v = x ( J n ) 2 + y ( J ) 2

3. Experiment Setup

3.1. Dataset

3.1.1. Ablation Experimental Dataset

The datasets for the ablation experiment were chosen from three standard datasets from the IEEE CVRP NTIRE Seminar: Dense-HAZE [31], O-HAZE [32], and NH-HAZE [33]. Dense-HAZE includes 55 identical pairs of dense haze/clear images. From the sample, 1–45 pairings were chosen for training, 46–50 pairs for verification, and 51–55 pairs for testing in this study. O-HAZE includes 45 sets of outdoor, non-homogeneous haze/clear images. From that set, 1–35 pairs were chosen for training, 36–40 pairs for verification, and 41–45 pairings for testing in this study. Fifty-five non-homogeneous haze/clear image pairs are included in NH-HAZE. In this study, 1–45 were selected for training, 46–50 for verification, and 51–55 for testing.

3.1.2. Self-Built Transmission Channel Inspection Dataset (UAV-HAZE)

In haze image imaging, because it is often manifested as loss of image visibility, the atmospheric extinction coefficient σ can solve the β ( λ ) in Equation (12).
β ( λ ) = 3.912 σ
Additionally, visibility varies depending on height. Therefore, the depth value of the scene and the vertical field of view of the camera are used to estimate the elevation values of the pixels and their distribution characteristics are calculated to replicate the distribution and color characteristics of genuine haze. To imitate the color features of haze, Formula (1) includes the haze color value I a l as follows:
I ( x , λ ) = t ( x ) J ( x , λ ) + A ( 1 t ( x ) ) × I a l
Taking into account the mountain haze’s irregularly distributed properties. Non-uniform haze is created using 3D Berlin noise, and a haze generator called FOHIS [34] is suggested. They are used to mimic non-uniform haze by making three Berlin noises of varying amplitudes and frequencies, which are then merged with Equation (13) and multiplied by β ( λ ) .
P _ n o i s e = 1 3 i = 1 3 P _ n o i s e i 2 i 1
In light of FOHIS, this work estimated the picture depth value in order to synthesize the mountain transmission into the UAV-HAZE dataset [35]. In the synthesis process, the I a l of the three-color channels of the image RGB is set to [220,220,210], respectively, to simulate the color characteristics of the blue-white mountain fog. Then, to imitate the distribution features of mountain haze, the vertical field of view of the camera is adjusted to 20°. This is combined with the depth value of picture pixels, and the pixel elevation value is calculated. The non-uniform properties of mountain haze were then simulated by creating 3D Berlin noise with three distinct frequency values (f = 130, 60, 10). Finally, the data [700–900], [900–1100], [1100–1300] and [1300,1500] were chosen as the extinction coefficients in Equation (12) using 450 mountain transmission channel photos obtained by UAV inspection as the original dataset. A total of 2225 non-uniform simulated haze/clear images of various concentrations make up UAV-HAZE, which is divided into training sets, verification sets, and test sets in a ratio of 7:2:1. There are 1560 pairs in the training set, 445 pairs in the verification set, and 220 teams in the test set.

3.2. Implementation Details

NVIDIA GeForce RTX3090 (24 GB) was the platform used for the experiment. Data preprocessing involves cropping each training image into 100 non-overlapping image blocks with a size of 120 × 160 pixels and unifying the image resolution of the training set across Dense-HAZE, O-HAZE, NH-HAZY, and UAV-HAZE to 1200 × 1600. The image blocks were simultaneously rotated at random angles of 0, 90, 180, and 270 degrees. The Adam optimizer is initially employed in DAMPHN network training with exponential decay rates γ 1 = 0.9, γ 2 = 0.999, starting learning rates 1 × 10−4, and batch sizes 100. We also adjusted the learning rate using an equally spaced strategy with step size = 10 and gamma = 0.1. Then, the hyperparameters of the loss function are set to α r = 1, α p = 6 × 10−3, α t v = 2 × 10−8, α r 1 = 0.6, α r 2 = 0.4. Finally, when the verification set loss function is stable, the training is stopped and the best model is obtained.

4. Experiment Results

4.1. Ablation Experiment

Two phases of the ablation experiment were conducted. The first and second sections, respectively, confirm the reliability of the DA module and the DAMPHN network.

4.1.1. DA Module

Due to the low cross-level fusion quality of the original multi-patch algorithm FDMPHN, the DA module is proposed in this study. In order to reduce the complexity of the algorithm, the encoder-decoder structure of FDMPHN is diminished. The three sets of experiments listed below are explicitly included in this section:
(I)
The network encoder-decoder has six residual modules (Resblock × 6) using only FDMPHN.
(II)
The approach suggested in this work builds on (I) by adding a DA module (FDMPHN + DA). A DA module plus six residual modules (Resblock × 6) make up the network encoder-decoder.
(III)
To optimize (II) and DAMPHN, the solution presented in this research uses just three residual modules (Resblock × 3).
  • Quantitative evaluation
PSNR [36], SSIM [37], and APT were chosen for quantitative evaluation in this section of the experiment. The visual noise and distortion decrease as the PSNR value rises. The recovery of structural properties such as image brightness and contrast is measured by SSIM. The dehazing is better the higher the value. Table 1 displays the precise outcomes of the three groups of studies. In Table 1, when (I) and (II) are compared, the addition of the DA module raised PSNR and SSIM in the three datasets by an average of 0.35 dB and 0.0073, whereas APT rose by 19% (0.007 s). Comparing (I) and (III), the average PSNR and SSIM in the three datasets are raised by 0.30 dB and 0.011, respectively, and APT is shortened by 11% (0.003 s), respectively, after the encode-decoder structure is streamlined.
  • Convergence analysis
This section assessed the convergence using the dynamic curves for training loss, PSNR, and SSIM. On Dense-HAZE, O-HAZE, and NH-HAZE, Figure 5 displays the training losses, PSNR, and SSIM for the FDMPHN, FDMPHN+DA, and DAMPHN approaches, respectively. Figure 5 shows the training and testing of the three approaches on three separate datasets, with the training losses, PSNR, and SSIM information displayed in the rows and columns, respectively. Figure 5a illustrates how the training loss for the aforementioned approaches steadily lowers as the number of iterations increases and gradually stabilizes at 35–40 rounds. In Figure 5b,c, all three approaches converge after 200 rounds, and the DA module performs better regardless of how complicated or straightforward the encoder-decoder structure is.

4.1.2. DAMPHN Network

To more accurately evaluate DAMPHN, we further conducted quantitative, qualitative, and convergence evaluation on three datasets, Dense-HAZE, O-HAZE, and NH-HAZE, with DCP [6], AOD-Net [9], FDMPHN [14], and GridDehazeNet [17], respectively.
  • Quantitative evaluation
PSNR, SSIM, and APT are also used to gauge how well various techniques remove haze. The outcomes of the quantitative comparison are displayed in Table 2. In Table 2, the blue values represent the optimal values, and the underlined values represent the sub-optimal values. In the three datasets, the PSNR and SSIM values of DAMPHN are 3.72 dB and 0.0666 higher than those of DCP on average, and ART is 94% shorter. The defog quality of AOD-Net in the Dense-HAZE dataset is comparable to that of DAMPHN. However, on the non-uniform haze datasets O-HAZE and NH-HAZE, the PSNR and SSIM values of DAMPHN are increased by 1.72 dB and 0.0446 compared with the average value of AOD-Net. The effect of GridDehazeNet on the fog removal in the three datasets has its own advantages compared with the method in this paper. Specifically, DAMPHN is, on average, 0.38 dB higher than GridDehazeNet’s PSNR value, but the SSIM value is lower than GridDehazeNet’s 0.025. Finally, compared with FDMPHN in the three datasets, the PSNR and SSIM values of DAMPHN are increased by 0.30 dB and 0.011 on average, and ART is shortened by 11%.
  • Qualitative assessment
The experiment’s visual comparison component is the main focus here. Among the images, the haze distribution in the first and second rows is more uniform, and the haze distribution in the third and fourth rows is uneven. The DCP results in Figure 6 reveal color distortion and a significant degree of residual haze. The image’s color changes to dark yellow after AOD-Net fog removal, and a significant quantity of haze residue remains in the non-uniform haze area. GridDehazeNet has a good fog effect when the haze distribution is relatively uniform, but the image’s color after fog removal is darker than that of the clear picture. In addition, in the case of non-uniform haze, GridDehazeNet also shows many haze residues. The image’s overall color after fog removal by FDMPHN is closer to the clear image when the haze distribution is more uniform. Still, the color distortion appears on the ground of the first line of the picture. Regarding non-uniform haze, FDMPHN has a good de-fogging effect, but its de-noising solid ability also causes image smoothing, resulting in blurred details. DAMPHN is visually similar to FDMPHN. However, in the enlarged area of the fourth row of the image, the DAMPHN haze residue is less.
  • Convergence analysis
In this experiment section, the convergence is assessed using the change curves of PSNR and SSIM with the number of training rounds. Figure 7 shows the results of each round of PSNR and SSIM tests for four de-fogging techniques on three datasets. DCP has the fastest convergence rate. AOD-Net uses a relatively lightweight CNN structure in the parameter estimation process, which has poor stability and the slowest convergence rate. When the PSNR value of the current verification set is assumed to be greater than the previous results during GridDehazeNet training, the round model is optimal. Under dynamic control, its convergence rate ranks fourth. The FDMPHN and DAMPHN set the hyperparameters before training, and the validation set is used to optimize the hyperparameter settings. Therefore, both FDMPHN and DAMPHN converge faster. Specifically, in Figure 7a, DAMPHN converges faster than FDMPHN. In Figure 7b,c, FDMPHN and DAMPHN converge at similar speeds. Therefore, DAMPHN in this paper is in second place in terms of convergence speed.

4.2. Transmission Channel Image

4.2.1. Synthetic Dataset UAV-HAZE

DAMPHN can be utilized to clear haze from Sichuan’s mountainous areas’ transmission channel scenery. This section is based on the dataset created in Section 3.1.2, UAV-HAZE. With this collection of data, DCP [6], AOD-Net [9], FDMPHN [14], GridDehazeNet [17], and DAMPHN, the approach in this article, are each examined in turn. This section evaluates both the algorithm’s quantitative and qualitative performance.
  • Quantitative evaluation
PSNR, SSIM, and APT were chosen as evaluation indicators. Table 3 presents the experimental outcomes. In Table 3, the blue font is the optimal value, and the underlined value is the sub-optimal value. The PSNR of DAMPHN is optimal, SSIM and ART are sub-optimal. In this study, DAMPHN’s PSNR and SSIM values are 7.26 dB and 0.0588 greater than DCP’s, respectively. APT barely makes up 4% of DCP techniques. PSNR and SSIM are 9.32 dB and 0.2057 greater in DAMPHN than in AOD-Net, although APT is 14 times higher. The PSNR value of DAMPHN is 0.26 dB higher, and the SSIM value is 0.0007 dB lower than GridDehazeNet. DAMPHN’s SSIM value is the same as FDMPHN’s, but its PSNR is 0.04 dB higher, and its APT is 94% shorter.
  • Qualitative assessment
Figure 8 displays the outcomes of the qualitative comparison between DAMPHN and the techniques mentioned above. DCP has a positive impact in the mist area, according to the analysis of Figure 8. The color of the third row seems distorted when the haze density is excellent, or the randomness of its distribution features is substantial. When dealing with non-uniform haze, AOD-Net’s primary result is that a significant amount of haze is left in the processed image, the details are blurred, and there is evident color distortion. The fog removal quality of GridDehazeNet is superior to that of the first two techniques. However, some fog was still present close to the first row’s wires and the fourth row’s poles and towers. In this study, the FDMPHN and DAMPHN techniques can recover the picture tower’s detailed information with excellent clarity and superb color fidelity. FDMPHN does, however, have a trace amount of haze residue in the first row’s wire area.

4.2.2. Real Image

The actual utility of DAMPHN was confirmed by the refit project from Gangu to Erlang Mountain in Shuzhou and the real hazy photographs of the Sichuan-Tibet network project. The approach was evaluated using both quantitative and qualitative methodologies.
  • Quantitative evaluation
Five non-reference image quality evaluation indexes, including information entropy, standard deviation, clarity, perception-based image quality evaluation method (PIQE) [38], and APT, were chosen for quantitative evaluation because there were insufficient clear reference examples. The more relevant information an image carries, the higher its information entropy. The image’s standard deviation is used to assess its contrast; the lower the standard deviation, the more stable the image is. The greater the value, the higher the sharpness, which is defined as the variance of calculating the absolute value of Laplace. Block effects, blur, and noise distortion are calculated using PIQE, and a lower value corresponds to less distortion. In Table 4, the experimental findings are displayed.
In Table 4, the underlined value and the blue text represent the ideal and sub-optimal values, respectively. This approach performs the best regarding clarity and PIQE, comes in second for ART, and comes in third for information entropy and standard deviation. This approach has reduced standard deviation and higher assessment indices compared to DCP. The proposed method has a clear benefit over AOD-Net regarding image quality, but it takes four times as long to operate. DAMPHN has higher evaluation indexes than GridDehazeNet, except for lower information entropy. DAMPHN is superior to FDMPHN in various assessment indices compared to FDMPHN before improvement, except for the picture information entropy, which is less than 0.17.
  • Qualitative assessment
Figure 9 displays two transmission channel views of the retrofitting project from Gangu to Erlang Mountain in Shuzhou and the haze reduction effect of four groups of the Sichuan-Tibet interconnection project. Uphill fog, uphill fog, advection fog, and radiation fog are all depicted in lines 1 through 4. Intuitive examination reveals that the color of DCP is severely altered and turns blue-purple in the sky area. AOD-Net effectively removes haze. However, it has glaring issues with blurred details and intensified hue. Although GridDehazeNet effectively removes fog, there is still some fog in the third-row valley and second-row tower areas. The image is also slightly lavender once the fog has been eliminated, for instance, the first row’s valley fog area and the fourth row’s pole tower area. In places with high haze density, such as the tower area in the second row and the valley area in the third row, FDMPHN has a competitive dehazing impact but leaves haze residue behind. This technique also results in color distortion, as seen in how the first row of trees on an ascent turned yellow. After adding a DA module, DAMPHN may now pay closer attention to areas with dense fog and a non-uniform haze. As a result, the method suggested in this paper removes fog more thoroughly than GridDehazeNet and FDMPHN in the first-row and third-row valley areas. Additionally, there is no purple or yellowing in terms of color preservation.

5. Discussion

In this paper, the issue of transmission line haze that is unevenly dispersed in mountainous places was studied. A DAMPHN is introduced, an innovative non-uniform haze-defogging network model put forth in this research to facilitate picture preprocessing for UAV transmission channel inspection in mountainous terrain. Similarly, the DAMPHN network model is universal. DAMPHN can be used for preprocessing other images in fog environments, such as unmanned visual perception, surveillance video (road traffic, transmission lines), and tachographs. DCP, AOD-Net, GridDenzeNet, and FDMPHN were utilized in numerous tests using open datasets (Dense-HAZE, O-HAZE, and NH-HAZE) and self-built datasets (UAV-HAZE) to demonstrate the efficacy of DAMPHN.
Notably, because the assumption of uniform distribution of air concentration in the atmospheric scattering model limits both DCP and AOD-Net, the error of estimating parameters is significant in dense fog and non-homogeneous haze. DAMPHN is a multi-level end-to-end fog removal network that seeks to remove fog by discovering the relationship between the haze and clear image mapping. DAMPHN does not, therefore, need to estimate the parameters; instead, it relies on the dataset’s basis, and the higher the base, the higher the quality of fog removal. GridDehazeNet solves the problem of feature fusion between different scales in multi-scale networks by introducing channel attention. DAMPHN solves the problem of feature fusion between different levels in multi-patch networks by introducing channel and pixel attention mechanisms. GridDehazeNet has vital artifact removal, so the SSIM value is stronger than DAMPHN. DAMPHN pays attention to the problem of uneven pixel distribution, pays attention to the removal of non-uniform fog, and has a strong denoising ability and high PSNR value. FDMPHN is identical to a multi-patch defogging network, but the residual connections in hierarchical fusion restrict how well it can fuse features. The pixel attention layer of the DAMPHN’s DA module is designed to pay attention to areas with unequal haze distribution. In contrast, the channel attention layer is designed to appropriately evaluate the channel domain properties. DAMPHN has a better defogging impact as a result than FDMPHN.
Additionally, the frequently used image segmentation algorithms U-Net and GridNet have produced effective outcomes in image segmentation and picture defogging via innovation. DCPDN solves parameter A using the U-Net network. GridDehazeNet proposes a multi-scale attention network based on GridNet. They both have superior defogging effects. With dual U-Net, Amyar et al. [39] created a multi-task and multi-scale network structure that was effectively used for lung tumor segmentation, classification, and prediction. However, DAMPHN accomplishes picture fog removal from the local to the global by helping the feature extraction of the bigger patch image from the top layer with the detailed feature of the lower layer. From the overall to the local picture segmentation, image fog removal, and other tasks, U-Net will employ the more comprehensive information collected from the bottom layer to aid in the development of smaller receptive field information. Consequently, the two networks’ designs have produced successful outcomes in their respective domains.
In conclusion, the DAMPHN approach offers an excellent defogging effect, less color distortion, and quick processing speed. In a location with a lot of fog, it is impossible to eliminate it entirely, and the details are hazy. DAMPHN can improve the defog effect by enhancing the encoder-decoder structure, feature extraction, and reconstruction skills, all of which were influenced by U-Net in the field of image segmentation, or by combining with the conventional image edge previous knowledge to increase the texture information and boost the fog removal effect.

6. Conclusions

This paper proposes that DAMPHN can achieve a good defog effect and restore the color and brightness of the image. The network encoder-decoder module and DA module are composed. The former can learn the mapping relationship between haze and clear pictures and has a strong feature extraction ability. The latter enhances the feature fusion ability by empowering the combination of channel attention and pixel attention. However, in excessive haze density, it cannot be entirely removed, and the details are hazy. Future work will improve the haze removal effect by enhancing texture information through edge prior and enhancing the encoder-decoder structure. Additionally, using 3D Berlin noise and image depth information to simulate haze’s non-uniform distribution characteristics is not only just restricted to UAV mountain transmission channel inspection; it can also be applied to a broader range of situations to enhance generalization performance.

Author Contributions

Conceptualization, methodology, and writing—review and editing, W.Z.; software, validation, data curation, and writing—original draft preparation, L.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was supported by the following project: Innovation and entrepreneurship training program for college students of China: “Vision Engine—Leader of Clear Processing of unmanned image”.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analysis, or interpretation of data; in the writing of the manuscript; or in the decision to publish the results.

Abbreviations

The following abbreviations are used in this manuscript:
UAVUnmanned Aerial Vehicle
FDMPHNFast Deep Multi-Patch Hierarchical Network
DAMPHNDual Attention Level Feature Fusion Multi-Patch Hierarchical Network
PSNRPeak Signal-to-Noise Ratio
SSIMStructural Similarity Index Measure
APTAverage Processing Time
DCPDark Channel Prior
CAPColor Decay Prior
CNNConvolutional neural networks
AOD-NetAll-in-One Dehazing Network
DCPDNDensely Connected Pyramid Dehazing Network
FDMSHNFast Deep Multi-Scale Hierarchical Network
DADual Attention Level Feature Fusion
Ca_layerChannel attention layer
Pa_layerPixel attention layer
PIQEPerception-based Image Quality Evaluation

References

  1. Li, X.; Li, Z.; Wang, H.; Li, W. Unmanned aerial vehicle for transmission line inspection: Status, standardization, and perspectives. Front. Energy Res. 2021, 9, 713634. [Google Scholar] [CrossRef]
  2. Zhang, T.; Tang, Q.; Li, B.; Zhu, X. Genesis and dissipation mechanisms of radiation-advection fogs in Chengdu based on multiple detection data. Meteorol. Sci. Technol. 2019, 47, 70–78. [Google Scholar]
  3. Zhao, L.; Zuo, X.; Zhang, S.; Lu, Y. On restoration of mountain haze image based on non-local prior algorithm. Electron. Opt. Control 2022, 29, 55–58. [Google Scholar]
  4. Imran, A.; Zhu, Q.; Sulaman, M.; Bukhtiar, A.; Xu, M. Electric-Dipole Gated Two Terminal Phototransistor for Charge-Coupled Device. Adv. Opt. Mater. 2023, 2300910. [Google Scholar] [CrossRef]
  5. Swinehart, D.-F. The beer-lambert law. J. Chem. Educ. 1962, 39, 333–335. [Google Scholar] [CrossRef]
  6. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar]
  7. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar]
  8. Cai, B.; Xu, X.; Jia, K.; Qing, C.; Tao, D. DehazeNet: An end-to-end system for single image haze removal. IEEE Trans. Image Process. 2016, 25, 5187–5198. [Google Scholar] [CrossRef] [Green Version]
  9. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. AOD-Net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  10. Zhang, H.; Patel, V.-M. Densely connected pyramid dehazing network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 3194–3203. [Google Scholar]
  11. Li, Y.; Miao, Q.; Quyang, W.; Ma, Z.; Fang, H.; Dong, C.; Quan, Y. LAP-Net: Level-aware progressive network for image dehazing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 3276–3285. [Google Scholar]
  12. Li, R.; Pan, J.; He, M.; Li, Z.; Tang, J. Task-oriented network for image dehazing. IEEE Trans. Image Process. 2020, 29, 6523–6534. [Google Scholar] [CrossRef]
  13. Bai, H.; Pan, J.; Xiang, X.; Tang, J. Self-guided image dehazing using progressive feature fusion. IEEE Trans. Image Process. 2022, 31, 1217–1229. [Google Scholar] [CrossRef]
  14. Das, S.-D.; Dutta, S. Fast deep multi-patch hierarchical network for nonhomogeneous image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 482–483. [Google Scholar]
  15. Zhang, H.; Dai, Y.; Li, H.; Koniusz, P. Deep stacked hierarchical multi-patch network for image deblurring. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 15–21 June 2019; pp. 5978–5986. [Google Scholar]
  16. Wang, K.; Yang, Y.; Li, B.; Cui, L. Uneven image dehazing by heterogeneous twin network. IEEE Access 2020, 8, 118485–118496. [Google Scholar] [CrossRef]
  17. Liu, X.; Ma, Y.; Shi, Z.; Chen, J. Griddehazenet: Attention-based multi-scale network for image dehazing. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 7314–7323. [Google Scholar]
  18. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature fusion attention network for single image dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11908–11915. [Google Scholar]
  19. Wang, C.; Shen, H.-Z.; Fan, F.; Shao, M.-W.; Yang, C.-S.; Luo, J.-C.; Deng, L.-J. EAA-Net: A novel edge assisted attention network for single image dehazing. Knowl.-Based Syst. 2021, 228, 107279. [Google Scholar] [CrossRef]
  20. Yang, K.; Zhang, J.; Fang, Z. Multi-patch and multi-scale hierarchical aggregation network for fast nonhomogeneous image dehazing. Comput. Sci. 2021, 48, 250–257. [Google Scholar]
  21. Wang, K.; Duan, Y.; Yang, Y.; Fei, S. Uneven hazy image dehazing based on transmitted attention mechanism. Pattern Recognit. Artif. Intell. 2022, 35, 575–588. [Google Scholar]
  22. Zhao, D.; Mo, B.; Zhu, X.; Zhao, J.; Zhang, H.; Tao, Y.; Zhao, C. Dynamic Multi-Attention Dehazing Network with Adaptive Feature Fusion. Electronics 2023, 12, 529. [Google Scholar] [CrossRef]
  23. Guo, Y.; Gao, Y.; Liu, W.; Lu, Y.; Qu, J.; He, S.; Ren, W. SCANet: Self-Paced Semi-Curricular Attention Network for Non-Homogeneous Image Dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 1884–1893. [Google Scholar]
  24. Liu, J.; Jia, R.; Li, W.; Ma, F.; Wang, X. Image dehazing method of transmission line for unmanned aerial vehicle inspection based on densely connection pyramid network. Wirel. Commun. Mob. Comput. 2020, 2020, 1–9. [Google Scholar]
  25. Zhang, M.; Song, Z.; Yang, J.; Gao, M.; Hu, Y.; Yuan, C.; Jiang, Z.; Cheng, W. Study on the enhancement method of online monitoring image of dense fog environment with power lines in smart city. Front. Neurorobotics 2022, 16, 299. [Google Scholar] [CrossRef]
  26. Zhai, Y.; Jiang, L.; Long, Y.; Zhao, Z. Dark channel prior dehazing method for transmission channel image based on sky region segmentation. J. North China Electr. Power Univ. 2021, 48, 89–97. [Google Scholar]
  27. Xin, R.; Chen, X.; Wu, J.; Yang, K.; Wang, X.; Zhai, Y. Insulator Umbrella Disc Shedding Detection in Foggy Weather. Sensors 2023, 22, 4871. [Google Scholar] [CrossRef]
  28. Gao, Y.; Yang, J.; Zhang, K.; Peng, H.; Wang, Y.; Xia, N.; Yao, G. A New Method of Conductor Galloping Monitoring Using the Target Detection of Infrared Source. Electronics 2022, 11, 1207. [Google Scholar] [CrossRef]
  29. Yan, L.; Zai, W.; Wang, J.; Yang, D. Image Defogging Method for Transmission Channel Inspection by UAV Based on Deep Multi-patch Layered Network. In Proceedings of the Panda Forum on Power and Energy (PandaFPE), Chengdu, China, 27–30 April 2023; pp. 855–860. [Google Scholar]
  30. Wang, K.; Yang, Y.; Fei, S. Review of hazy image sharpening methods. CAAI Trans. Telligent Syst. 2023, 18, 217–230. [Google Scholar]
  31. Ancuti, C.-O.; Ancuti, C.; Sbert, D.; Timofte, R. Dense-haze: A benchmark for image dehazing with dense-haze and haze-free images. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1014–1018. [Google Scholar]
  32. Ancuti, C.-O.; Ancuti, C.; Timofte, R.; De, C. O-haze: A dehazing benchmark with real hazy and haze-free outdoor images. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 754–762. [Google Scholar]
  33. Ancuti, C.-O.; Ancuti, C.; Timofte, R. NH-HAZE: An image dehazing benchmark with non-homogeneous hazy and haze-free images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 14–19 June 2020; pp. 444–485. [Google Scholar]
  34. Zhang, N.; Zhang, L.; Cheng, Z. Towards simulating foggy and hazy images and evaluating their authenticity. In Proceedings of the Neural Information Processing: 24th International Conference, Guangzhou, China, 14–18 November 2017; pp. 405–415. [Google Scholar]
  35. Harsányi, K.; Kiss, K.; Majdik, A.; Sziranyi, T. A hybrid CNN approach for single image depth estimation: A case study. In Proceedings of the International Conference on Multimedia and Network Information System, Hong Kong, China, 1 June 2019; pp. 372–381. [Google Scholar]
  36. Wang, Z.; Bovik, A.-C. Mean squared error: Love it or leave it? A new look at signal fidelity measures. IEEE Signal Process. Mag. 2009, 26, 98–117. [Google Scholar] [CrossRef]
  37. Wang, Z.; Bovik, A.-C.; Sheikh, H.-R.; Simoncelli, E.-P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar]
  38. Venkatanath, N.; Praneeth, D.; Bh, M.-C.; Channappayya, S.; Medasani, S. Blind image quality evaluation using perception based features. In Proceedings of the 2015 Twenty First National Conference on Communications (NCC), Munbai, India, 27 February–1 March 2015; pp. 1–6. [Google Scholar]
  39. Amyar, A.; Modzelewski, R.; Vera, P.; Morard, V.; Ruan, S. Multi-task multi-scale learning for outcome prediction in 3D PET images. Comput. Biol. Med. 2022, 151, 106208. [Google Scholar] [CrossRef]
Figure 1. Implementation scheme of image preprocessing of mountain transmission channel based on DAMPHN.
Figure 1. Implementation scheme of image preprocessing of mountain transmission channel based on DAMPHN.
Sensors 23 07026 g001
Figure 2. DAMPHN network structure.
Figure 2. DAMPHN network structure.
Sensors 23 07026 g002
Figure 3. Encoder-decoder module structure.
Figure 3. Encoder-decoder module structure.
Sensors 23 07026 g003
Figure 4. DA module structure.
Figure 4. DA module structure.
Sensors 23 07026 g004
Figure 5. Training loss curve and test PSNR and SSIM curve. (a) Training loss. (b) Testing PSNR. (c) Testing SSIM.
Figure 5. Training loss curve and test PSNR and SSIM curve. (a) Training loss. (b) Testing PSNR. (c) Testing SSIM.
Sensors 23 07026 g005
Figure 6. NH-HAZE and O-HAZE dehazing results. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN. (g) Ground truth.
Figure 6. NH-HAZE and O-HAZE dehazing results. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN. (g) Ground truth.
Sensors 23 07026 g006
Figure 7. PSNR and SSIM test curves. (a) Dense-HAZE. (b) NH-HAZE. (c) O-HAZE.
Figure 7. PSNR and SSIM test curves. (a) Dense-HAZE. (b) NH-HAZE. (c) O-HAZE.
Sensors 23 07026 g007
Figure 8. Results of UAV-HAZE dehazing. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN. (g) Ground truth.
Figure 8. Results of UAV-HAZE dehazing. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN. (g) Ground truth.
Sensors 23 07026 g008
Figure 9. Dehazing result of real transmission channel image. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN.
Figure 9. Dehazing result of real transmission channel image. (a) Hazy. (b) DCP. (c) AOD-Net. (d) GridDehazeNet. (e) FDMPHN. (f) DAMPHN.
Sensors 23 07026 g009
Table 1. Results of DA module ablation experiments.
Table 1. Results of DA module ablation experiments.
MethodDense-HAZEO-HAZENH-HAZE
PSNRSSIMAPTPSNRSSIMAPTPSNRSSIMAPT
(I)FDMPHN13.470.43690.03119.930.70450.03016.870.55120.030
(II)FDMPHN + DA14.030.45120.03620.350.69760.03516.940.56560.037
(III)DAMPHN13.890.44970.02720.200.71380.02717.070.56210.027
Table 2. Results of DAMPHN Network quantitative comparison.
Table 2. Results of DAMPHN Network quantitative comparison.
MethodDense-HAZEO-HAZENH-HAZE
PSNRSSIMAPTPSNRSSIMAPTPSNRSSIMAPT
DCP [6]11.600.38540.40615.660.67530.44013.280.46500.416
AOD-Net [9]13.850.47140.02318.190.69500.01015.640.49180.009
GridDehazeNet [17]13.500.47210.02619.820.71080.02616.700.61010.026
FDMPHN [14]13.470.43690.03119.930.70450.03016.870.55120.030
DAMPHN (ours)13.890.44970.02720.200.71380.02717.070.56210.027
Table 3. Quantitative comparison results on UAV-HAZE.
Table 3. Quantitative comparison results on UAV-HAZE.
MethodPSNRSSIMAPT
DCP [6]19.970.88510.352
AOD-Net [9]17.920.73820.001
GridDehazeNet [17]26.980.94760.015
FDMPHN [14]27.200.94390.234
DAMPHN (ours)27.240.94390.014
Table 4. Results of quantitative evaluation of real images.
Table 4. Results of quantitative evaluation of real images.
MethodInformation EntropyStandard
Deviation
ClarityPIQEAPT
DCP [6]17.7832.19459.8627.510.342
AOD-Net [9]16.1045.93452.7928.870.005
GridDehazeNet [17]18.2841.61470.1824.900.021
FDMPHN [14]18.1042.38465.2124.480.270
DAMPHN (ours)17.9341.92536.1123.980.020
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zai, W.; Yan, L. Multi-Patch Hierarchical Transmission Channel Image Dehazing Network Based on Dual Attention Level Feature Fusion. Sensors 2023, 23, 7026. https://doi.org/10.3390/s23167026

AMA Style

Zai W, Yan L. Multi-Patch Hierarchical Transmission Channel Image Dehazing Network Based on Dual Attention Level Feature Fusion. Sensors. 2023; 23(16):7026. https://doi.org/10.3390/s23167026

Chicago/Turabian Style

Zai, Wenjiao, and Lisha Yan. 2023. "Multi-Patch Hierarchical Transmission Channel Image Dehazing Network Based on Dual Attention Level Feature Fusion" Sensors 23, no. 16: 7026. https://doi.org/10.3390/s23167026

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop