Next Article in Journal
Geochemical Characteristics and Depositional Environment from the Permian Qipan Formation Hydrocarbon Source Rocks in the Piedmont of Southwestern Tarim Basin
Previous Article in Journal
Design of a Feed Array Antenna to Obtain a Uniform Near-Field Distribution on a Virtual Surface Placed within a Specified Wavelength
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Remote Sensing Image Dehazing via Dual-View Knowledge Transfer

1
Xi’an Institute of Optics and Precision Mechanics, Chinese Academy of Sciences, Xi’an 710119, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
3
School of Computer Science and Technology, Xi’an University of Posts and Telecommunications, Xi’an 710121, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2024, 14(19), 8633; https://doi.org/10.3390/app14198633
Submission received: 22 August 2024 / Revised: 12 September 2024 / Accepted: 23 September 2024 / Published: 25 September 2024

Abstract

:
Remote-sensing image dehazing (RSID) is crucial for applications such as military surveillance and disaster assessment. However, current methods often rely on complex network architectures, compromising computational efficiency and scalability. Furthermore, the scarcity of annotated remote-sensing-dehazing datasets hinders model development. To address these issues, a Dual-View Knowledge Transfer (DVKT) framework is proposed to generate a lightweight and efficient student network by distilling knowledge from a pre-trained teacher network on natural image dehazing datasets. The DVKT framework includes two novel knowledge-transfer modules: Intra-layer Transfer (Intra-KT) and Inter-layer Knowledge Transfer (Inter-KT) modules. Specifically, the Intra-KT module is designed to correct the learning bias of the student network by distilling and transferring knowledge from a well-trained teacher network. The Inter-KT module is devised to distill and transfer knowledge about cross-layer correlations. This enables the student network to learn hierarchical and cross-layer dehazing knowledge from the teacher network, thereby extracting compact and effective features. Evaluation results on benchmark datasets demonstrate that the proposed DVKT framework achieves superior performance for RSID. In particular, the distilled model achieves a significant speedup with less than 6% of the parameters and computational cost of the original model, while maintaining a state-of-the-art dehazing performance.

1. Introduction

Remote-sensing image dehazing (RSID), as a critical technique in remote-sensing image processing, aims to eliminate the blurring and degradation caused by atmospheric media [1]. With the widespread application of remote-sensing technology in meteorological monitoring, environmental assessment, and other fields, clear remote-sensing images are crucial for the accuracy and timeliness of information extraction. RSID technology not only enhances image quality but also plays a key role in meteorological monitoring [2], disaster response [3], and military surveillance [4].
In recent years, image-dehazing techniques have been extensively studied, and many dehazing methods have been proposed [5,6]. Currently, these dehazing methods can be categorized into two types [7]: prior-based dehazing methods and learning-based dehazing methods. Prior-based dehazing methods relies on manually extracting statistical differences between haze-degraded images and clear images to establish empirical priors [8,9,10]. The most classical prior-based image-dehazing methods recover the clear image based on the atmospheric scattering model (ASM). For instance, He et al. [11] raised the dark channel prior (DCP) method to estimate the transmission map for image dehazing. To reduce the model complexity, He et al. [12] proposed a guided filtering method for image dehazing subsequently. In addition, Zhu et al. [13] proposed the color attenuation prior (CAP) method to recover depth information so as to conduct image dehazing. Fattal [14] noticed that RGB values follow a color line prior (CLP) within each local patch in an image, and estimated the transmission map based on CLP. Although these prior-based methods are easy to implement, they rely on scene-specific prior assumptions, and when these assumptions are not met, the dehazing results may be unsatisfactory [15].
With the advancement of deep learning, learning-based dehazing methods have been widely explored [16,17,18], which mainly include atmospheric scattering model (ASM)-learning methods and end-to-end learning methods. ASM-learning methods achieve image dehazing by designing suitable deep neural networks to estimate the haze parameters in the ASM. For example, Chi et al. [19] incorporated prior knowledge into a gradient-guided Swin Transformer [20] for haze parameter estimation. To avoid explicitly collecting paired training data, Shyam and Yoo [21] proposed an adversarial auto-augmentation approach by combining extended atmospheric scattering model. End-to-end learning methods achieve effective image dehazing by directly learning the underlying mapping relationship between haze-degraded images and clear images [22]. For example, Qin et al. [23] developed a feature fusion attention network, named FFA-Net, for end-to-end image dehazing. Guo et al. [24] proposed a feature modulation module for effectively integrating Transformer and CNN features. Despite the notable achievements of current learning-based approaches, most of them are designed for dehazing natural scenes [25,26]. Compared to images in natural scenes, remote-sensing images are typically larger, more complex, and exhibit variable and irregular haze intensity distributions. Therefore, the application of existing end-to-end dehazing methods for haze-degraded remote-sensing images may not be ideal [27].
Recently, researchers have explored numerous methods specifically tailored for RSID [28], achieving significant results. For example, Li and Chen [29] designed a two-stage coarse-to-fine dehazing network FCTF-Net for gradually refining dehazing result. Song et al. [30] introduced a detail-compensated split attention to capture global and local regional dependencies for RSID. Zhang and Wang [31] proposed a dynamic collaborative reasoning learning framework for RSID. Zhang et al. [32] developed a frequency-oriented FOTformer for recovering haze-free images. Zheng et al. [33] achieved unsupervised RSID by introducing an enhanced attention-guided GAN network. Yu et al. [34] designed a hierarchical sliced interactive multi-layer collaborative decoding network (HSMD-Net) to achieve image restoration by realizing cross-layer global and local information interaction. In general, most state-of-the-art methods employ heavy backbone networks (e.g., Transformer models [28]) to obtain better performance. However, due to the demanding computational cost and slow running speed, these models are highly inefficient. Therefore, a fundamental question is how to obtain efficient RSID models rather than relying on heavy backbone networks. Furthermore, in comparison to dehazing tasks within natural scenes, there exists a significant scarcity of RSID datasets that incorporate unambiguously labeled images, resulting in suboptimal performance. Although a large number of lightweight techniques, such as quantization [35] and pruning [36], have been explored, there is still no good way to solve the problem of scarcity of RSID datasets. Fortunately, the emergence of knowledge distillation technology provides an alternative for addressing these issues.
As a result, this study proposes a Dual-View Knowledge Transfer (DVKT) framework for simultaneously distilling hierarchical features and cross-layer correlation knowledge. The framework encompasses two innovative modules: Intra-layer Knowledge Transfer (Intra-KT) and Inter-layer Knowledge Transfer (Inter-KT). Specifically, the Intra-KT module leverages representative features extracted from a well-trained teacher network to sequentially guide the corresponding features of the student network. This mechanism progressively corrects the learning bias of the student, enabling its features to exhibit similar visual and semantic patterns to those of the teacher. In contrast, the Inter-KT module calculates the relationships between features of the teacher network layer-by-layer and utilizes this knowledge to aid the student network in modulating its hierarchical features, thereby enabling the student to learn the solution process of the teacher. Based on the proposed DVKT framework, the distilled student network can effectively acquire compact and knowledge-rich features, resulting in high-quality haze-free remote-sensing images. With less than 6% of the parameters of the teacher model, the distilled model achieves nearly an 8-fold speedup while maintaining competitive dehazing performance.
In summary, the main contributions and novelty of this work are three-fold:
  • A Dual-View Knowledge Transfer (DVKT) framework is proposed for effective RSID: to the extent of our knowledge, the proposed DVKT framework is the first work to concentrate on transferring shared knowledge from natural scene dehazing models to RSID models and learning remote-sensing scene-specific knowledge for lightweight models.
  • Two complementary knowledge transfer modules, Intra-KT and Inter-KT, are integrated to comprehensively transfer the structured knowledge from well-trained natural scene-dehazing models to lightweight RSID models.
  • Extensive experiments on benchmark datasets demonstrate the effectiveness of the proposed DVKT framework. In particular, the proposed DVKT framework achieves satisfactory dehazing results while significantly improving processing speed.

2. Materials and Methods

This study proposes a Dual-View Knowledge Transfer (DVKT) framework to address the efficiency and limited labeled data issues in RSID. As depicted in Figure 1, the framework aims to effectively transfer hierarchical and cross-layer knowledge from a well-trained teacher network to a student network based on multiple Intra-KT and Inter-KT modules. The teacher network in the proposed DVKT framework adopts our earlier SCSNet [37]. Differently, the SCSNet used in this study is pre-trained on the natural-scene-dehazing dataset RESIDE [38], instead of limited labeled remote-sensing images. The student network is 1/n-SCSNet, where the number of channels in its convolutional layers (except the last layer) is 1/n of the teacher network. Compared to the complex SCSNet, the 1/n-SCSNet model has only 1/n2 of the parameters and computational cost of SCSNet, but its performance degrades slightly.

2.1. Feature Extraction

It is typical for teacher networks to be pre-trained on standard benchmark datasets, and the knowledge they have acquired can be represented explicitly as parameters or implicitly embedded in features. Building on previous research [39], knowledge transfer is conducted at the feature level, using the features of a pre-trained teacher network to guide the learning of a student network’s features hierarchically.
As illustrated in Figure 1, any given hazy image I is simultaneously fed into SCSNet and 1/n-SCSNet for feature learning. For the sake of simplicity, the features produced by the i-th block of SCSNet are uniformly represented as t i , while the features of the corresponding layer in 1/n-SCSNet are denoted as s i . Note that t i and s i have the same resolution, but the number of channels of s i is 1/n of that of t i . To reduce the computational cost during training, only some representative features are distilled:
T = t 1 ,   t 2 , , t 8 ,
S = s 1 ,   s 2 , , s 8 ,
where T and S represent the feature sets to be distilled from SCSNet and 1/n-SCSNet, respectively.

2.2. Intra-Layer Knowledge Transfer

As discussed in the previous section, the learned features t i implicitly inherently encompass the knowledge acquired by SCSNet. To enhance the efficiency of the student network 1/n-SCSNet, a simple yet powerful Intra-KT module is designed, which enables the sequential transfer of hierarchical knowledge from SCSNet to 1/n-SCSNet. Specifically, s i is responsible for learning the semantic patterns of t i and optimizing the parameters of 1/n-SCSNet by enhancing their distributional similarity.
Concretely, the Intra-KT module consists of two steps: channel reduction and knowledge transfer. Since the number of channels of features s i and t i is different, it is not suitable to directly calculate their similarity. To address this issue, a set of embedded features E = e 1 ,   e 2 , , e 8 is generated by inputting each feature s i into a 1 × 1 convolutional layer for channel reduction, which can be expressed as follows:
e i = s i k 1 × 1 ,
where k 1 × 1 denotes the to-be-learned parameters in the 1 × 1 convolutional layer. The output e i is the embedded feature of s i , and it has the same channel number with t i .
The second step entails a similarity calculation and knowledge transfer. Given that the Euclidean distance is excessively strict and may result in mere rote memorization, a relatively relaxed metric, cosine similarity, is employed to quantify the similarity between two features. In particular, the similarity between t i and e i at position x , y is calculated as follows:
s i m i x , y = c o s t i x , y , e i x , y = c = 1 C t i x , y , c · e x , y , c t i x , y · e i x , y ,
where C denotes the channel number t i , and t i x , y , c indicates the response value at position x , y of the c-th channel in t i . · denotes the length of a vector. Therefore,
L i n t r a = x = 1 H i y = 1 W i 1 s i m i x , y ,
where H i and W i denote the spatial size of t i . By minimizing this loss, the parameters in 1/n-SCSNet can be effectively optimized. In comparison to computationally expensive pair-wise distillation that can only be performed on low-resolution high-level features, our Intra-KT is more efficient to transfer implicit knowledge embedded in each layer.

2.3. Inter-Layer Knowledge Transfer

To regularize the student network learning flow of the solution process (FSP) of the teacher network, an Inter-KT module is designed. This module calculates the pair-wise feature relationship within the teacher network layer by layer to regulate features from the student network. A detailed description of the Inter-KT module is as follows:
The Inter-KT module evaluates the feature interdependencies between pairs within the teacher network on a layer-by-layer basis, aiming to govern both the prolonged and immediate development of feature attributes in the student network. Below is an elaborative explanation of the Inter-KT module’s functionality.
Firstly, the FSP matrices need to be generated layer by layer. For any two general features X 1 and X 2 , their FSP matrix F X 1 , X 2 R m × n can be calculated through channel-wise inner product:
F X 1 , X 2 = x = 1 h y = 1 w X 1 x , y · X 2 x , y h · w ,
It is worth noting that the FSP matrix is computed on features with the same resolution. However, the features in T have different resolutions. To address this issue, the resolution of all features in T is adjusted to that of t 4 through max pooling so as to simultaneously reduce the computational cost. The adjusted features of t i are denoted as R t i . Similarly, corresponding resolution adjustment is applied to features in feature set E , so as to align them with the resolution of t i .
In this study, the FSP matrices of adjacent features are calculated in order to grasp the evolution of features. In particular, an FSP matrix F R t i , R t i + 1 is generated for t i , t i + 1 in the teacher network. Similarly, the FSP matrix F R e i , R e i + 1 is also calculated for e i , e i + 1 of in the student network. Finally, the computation of the loss function for Inter-RT can be carried out using the following:
L i n t e r = i = 1 7 F R t i , R t i + 1 F R e i , R e i + 1 2 ,
By reducing the gap among these FSP matrices, the knowledge in the teacher network SCSNet can be effectively transferred into the student network 1/n-SCSNet.

2.4. Loss Function

At the output level, the smooth L 1 loss is employed to train the proposed DVKT framework. The smooth L 1 loss guarantees proximity between the reconstructed haze-free image and the ground truth, demonstrating superior performance over L 2 loss and the traditional L 1 loss in numerous low-level visual tasks. Let I denote the input haze-degraded image, O t denote the reconstructed haze-free image obtained by the teacher network, O s denote the reconstructed haze-free image obtained by the student network, and I g t denote the ground truth clear image. The formula for the smooth L 1 loss can be articulated as follows:
L S L 1 = 1 K k = 1 K l = 1 3 s m o o t h L 1 O s l k I g t l k + 1 K k = 1 K l = 1 3 s m o o t h L 1 O s l k O t l k ,
where
s m o o t h L 1 r = 0.5 r 2 , i f   r < 1 r 0.5 , o t h e r w i s e ,
where K is the total number of pixels within the image, and l denotes the l-th color channel.
Ultimately, the optimization of parameters within the 1/n-SCSNet is achieved through the minimization of the overall loss function:
L t o t a l = λ 1 · L i n t r a + λ 2 · L i n t e r + λ 3 · L S L 1 ,
where λ 1 , λ 2 , and λ 3 represent the respective weights assigned to the various loss components.

3. Results and Discussion

To illustrate the efficacy of the proposed DVKT framework, four public datasets are applied for evaluation. Furthermore, the experimental setup, comparison with the state of the art, and some ablation studies are given successively.

3.1. Datasets

Extensive experiments were conducted on four public remote-sensing-image-dehazing datasets: SateHaze1k [40], RS-Haze [28], RSID [19], RRSD300 [41].
  • SateHaze1k: The SateHaze1k dataset [40] consists of 1200 image pairs, divided into three sub-datasets with different haze densities. The SateHaze1k-thin sub-dataset contains images with thin haze, the SateHaze1k-moderate sub-dataset contains images with moderate haze, and the SateHaze1k-thick sub-dataset contains images with thick haze. The images in each sub-dataset are split among three categories: 320 for training, 35 for validation, and 45 for testing.
  • RS-Haze: The RS-Haze dataset [28] is a comprehensive and challenging dataset designed for image dehazing tasks, featuring 51,300 image pairs. Among these, 51,030 pairs are designated for training purposes, while the remaining 270 pairs are reserved for testing. It encompasses various scenarios and levels of haze density, such as urban landscapes, forests, beaches, and mountain regions.
  • RSID: The RSID dataset [19] comprises 1000 image pairs, with each pair consisting of a hazy image and a corresponding haze-free image. In line with prior research [41], 900 of these pairs are randomly chosen for the training set, while the rest are utilized for testing.
  • RRSD300: The RRSD300 dataset [41] comprises 300 real-world hazy images collected from DIOR dataset [42] and Microsoft Bing. This dataset is used to verify the generalization of various models.

3.2. Experimental Details and Evaluation Criteria

(1) Experimental Details: The proposed DVKT framework is implemented with PyTorch 1.13, and the experiments are conducted on a PC with NVIDIA RTX 3090 GPU (NVIDIA, Santa Clara, CA, USA). To augment the training dataset, random rotations of 90°, 180°, and 270°, as well as horizontal flipping are applied. The inputs to the DVKT framework are with the spatial size of 256 × 256. For the training process, a batch size of 32 was used, and the number of epochs was set to 20. The DVKT framework was optimized using the Adam optimizer. The initial learning rate configured at 1.0 × 10−4, with a cosine annealing strategy employed to progressively reduce it to 0. To address the high cost associated with tuning weight coefficients, a multi-task learning approach based on equal variance and uncertainty [43] was implemented to determine the optimal weight coefficients λ 1 , λ 2 , and λ 3 . Empirically, the hyperparameter n in 1/n-SCSNet was set to 5.
(2) Evaluation Criteria: To assess the performance of the proposed DVKT framework, the Peak Signal-to-Noise Ratio (PSNR), the Structural Similarity Index Measure (SSIM), and the Learned Perceptual Image Patch Similarity (LPIPS) are used as quantitative metrics. Higher PSNR and SSIM or lower LPIPS values signify superior quality in the restored images. They can be expressed as follows:
P S N R = 20 · l o g M A X I M S E ,
S S I M I g t , O s = 2 μ g t μ s + c 1 2 σ g t s + c 2 μ g t 2 + μ s 2 + c 1 σ g t 2 + σ s 2 + c 2 ,
L P I P S I g t , O s = i 1 H i W i h , w w i · I h w i O h w i 2 2 ,
where M A X I denotes the maximum pixel value of I g t , and MSE represents the mean squared error between the corresponding pixels of I g t and the reconstructed haze-free image O s . μ g t and μ s are the means of I g t and O s , respectively. σ g t and σ s mean the standard deviations of I g t and O s , respectively. σ g t s is the covariance between I g t and O s . w i means the weight for each image channel. c 1 and c 2 are constants introduced to maintain stability in the computation.
Furthermore, to evaluate the model size and computational cost, Params and floating-point operations (FLOPs) are adopted:
P a r a m s = K h · K w · C i n + 1 · C o u t ,
F L O P S = 2 H W K h · K w · C i n + 1 · C o u t ,
where C i n represents the channel number for the input feature, K h and K w denote the kernel size, and C o u t signifies the channel number for the output feature. H and W correspond to the height and width of the feature map. In this study, the units of Params and FLOPs are millions (M) and billions (G), respectively. Furthermore, the model’s inference time, measured in seconds, is determined by averaging the processing times of 100 haze-degraded images. During the experiments, all comparative methods utilize input images with a spatial dimension of 256 × 256.

3.3. Comparison with Other Image Dehazing Methods

This section reports comprehensive quantitative and qualitative comparisons against state-of-the-art methods on the adopted datasets. The compared methods are categorized into two primary groups: (1) those developed for natural image dehazing, encompassing techniques like DCP [11], CEP [44], SLP [45], AOD-Net [46], MOF [47], FFA-Net [23], Light-DehazeNet [48], Restormer [49], DehazeFormer [28], and Trinity-Net [19]; and (2) methods tailored specifically for dehazing remote-sensing images, such as FCTF-Net [29], DCRD-Net [50], and SCSNet [37]. To guarantee a fair assessment, the officially released code for each method was utilized, and a consistent data division strategy was applied during training.

3.3.1. Quantitative Evaluations

(1) Performance Comparison: The quantitative comparison results on the SateHaze1k dataset are summarized in Table 1. As indicated, traditional prior-driven dehazing methods such as DCP, CEP, MOF, and SLP exhibit lower PSNR and SSIM values, suggesting they struggle with the complexities of remote-sensing image dehazing, thus delivering subpar performance. Similarly, AOD-Net and Light-DehazeNet show rather mediocre results when applied to remote-sensing imagery, possibly due to the limitations of early lightweight neural networks in capturing the intricate mappings between hazy and clear images. Conversely, data-driven approaches like FFA-Net, Restormer, Trinity-Net, and DehazeFormer demonstrate higher PSNR and SSIM scores, highlighting their potential in enhancing remote-sensing image dehazing. Models such as DCRD-Net, FCTF-Net, and SCSNet, which are specifically crafted for remote-sensing dehazing tasks, are also included in the comparison. It can be observed that both DCRD-Net and SCSNet achieve very good dehazing performance, and SCSNet performs best among all comparison methods. In contrast, FCTF-Net performs generally, which may be due to its relatively complex network structure, while the data volume in the SateHaze1k dataset is limited, leading to model overfitting. It is worth noting that, except for the teacher network SCSNet used in this study, the proposed DVKT framework delivers a state-of-the-art dehazing performance. Additionally, our DVKT framework matches the dehazing performance of the teacher network, SCSNet, but with a reduced number of parameters and faster inference times.
Table 2 illustrates the quantitative comparison results across the RS-Haze and RSID datasets. As indicated, the proposed method demonstrates notable advantages over other compared approaches. Specifically, aside from the teacher network SCSNet utilized in this study, the DVKT framework shows superior dehazing performance. On the RS-Haze dataset, it achieves PSNR and SSIM values of 31.4227 dB and 0.9135, respectively. For the RSID dataset, the PSNR and SSIM values are 24.1082 dB and 0.9426, respectively.
(2) Efficiency Comparison: One of the core objectives of this work is to enhance the efficiency of RSID models. To substantiate the effectiveness of the proposed DVKT framework in dehazing tasks, an exhaustive comparison of inference efficiency is conducted with the existing RSID models. Table 3 details the model sizes and inference efficiencies of various models. Concretely, it lists the model parameters (Params), FLOPs, and the average time required to process 100 haze-degraded images (inference time). As observed, while FFA-Net, Restormer, Trinity-Net, and DehazeFormer deliver superior dehazing outcomes, they come with larger Params and FLOPs, which can constrain their practical deployment. On the other hand, AOD-Net and Light-DehazeNet boast smaller parameter sizes, albeit at the expense of considerable performance loss. When compared to FCTF-Net and DCRD-Net, both designed specifically for RSID, the proposed DVKT framework exhibits the lowest computational cost and inference time, underscoring its advantage. Notably, the proposed DVKT framework achieves almost identical PSNR and SSIM values to the teacher model SCSNet but with reduced computational requirements and faster inference times.

3.3.2. Qualitative Evaluations

Figure 2 showcases the qualitative results of various methods applied to the SateHaze1k-thin dataset. The dehazed images from DCP and FCTF-Net display noticeable color distortions. A significant amount of haze persists in the images processed by AOD-Net and Light-DehazeNet. Methods such as FCTF-Net and Light-DehazeNet achieve some level of dehazing; however, compared to the ground truth images, there is still a minor presence of haze in the restored images. Similarly, DehazeFormer, SLP, Trinity-Net, and DCRD-Net manage to remove some haze, but the overall restoration still leaves a slight haze residue compared to the ground truth. The dehazed images produced by the teacher model SCSNet and our DVKT framework are more closely aligned with the ground truth images, achieving a higher level of detail recovery. It is important to note that our DVKT framework boasts a smaller model size and higher inference efficiency.
Figure 3 displays the visual results of multiple methods on the SateHaze1k-moderate test set. Under moderate haze conditions, crucial information in remote sensing images can become obscured. Apart from SCSNet and the proposed DVKT framework, other dehazing methods result in images with some residual haze. DCP, AOD-Net, SLP, and Light-DehazeNet show significant color distortion. While DCRD-Net and Trinity-Net achieve a certain level of dehazing, there remain noticeable differences when compared to the ground truth image. In contrast, both the teacher model SCSNet and the proposed DVKT framework produce visually more pleasing results, with enhanced color fidelity and better preservation of texture details.
Figure 4 illustrates the qualitative results of different methods when applied to the SateHaze1k-thick test set. This particular set includes remote-sensing images heavily affected by dense haze, causing a significant loss of texture detail information. Methods such as DCP, AOD-Net, Light-DehazeNet, SLP, and FCTF-Net show evident color distortions. Although DehazeFormer, Trinity-Net, and DCRD-Net achieve good dehazing effects, the texture details in the images produced by these methods do not match the ground truth images as closely. In contrast, both the teacher model SCSNet and the proposed DVKT framework effectively preserve the overall structure, local details, and color accuracy of the ground truth images, resulting in visually pleasing and realistic dehazed images.
Figure 5 showcases the visual results of various methods on the RS-Haze test set. While several methods demonstrate varying degrees of success, some exhibit notable shortcomings. For instance, SLP- and DCP-restored images often suffer from color distortion and residual haze. AOD-Net, DehazeFormer, Trinity-Net, DCRD-Net, and FCTF-Net frequently over-dehaze, leading to a loss of fine details. Light-DehazeNet, on the other hand, struggles with color distortion and detail preservation. In contrast, both SCSNet and the proposed DVKT framework effectively restore texture details and enhance image quality, aligning more closely with the ground truth.
Figure 6 presents the visual comparison of various methods on the RSID test set. Obviously, SLP and DCP-restored images suffer from severe color distortion. The images restored by AOD-Net and Light-DehazeNet suffer from a significant loss of texture details and exhibit notable color distortions. DehazeFormer, Trinity-Net, DCRD-Net, and FCTF-Net can avoid color distortions to some extent, but the generated images still have some haze that cannot be removed. In contrast, both SCSNet and the proposed DVKT framework effectively preserve color fidelity and texture details, producing visually superior results.
To further demonstrate the generalization capability of the proposed method, we conducted performance verification on the RRSD300 dataset. Since the RRSD300 dataset only contains real-world hazy images without corresponding ground truth images, all models are trained on SateHaze1k dataset and trained on the RRSD300 dataset. The visualization results are shown in Figure 7. It can be observed that the dehazing results obtained by DCP and SLP exhibit significant color distortion. AOD-Net and DCRD-Net introduce some artifacts, resulting in poorer dehazing outcomes. Light-DehazeNet’s dehazing results lose considerable detail information. Although FCTF-Net is a dehazing method tailored for remote-sensing images, it also suffers from color distortion. In contrast, the adopted teacher network SCSNet and the proposed DVKT framework achieve better dehazing results, indicating the good generalization and practicality of the proposed DVKT framework method.

3.4. Ablation Study

This study focuses on RSID and optimizes the dehazing model by reducing the number of channels in each layer. Typically, for models with the same network architecture, their size and computational cost are positively correlated with dehazing performance. To investigate this further, a study was conducted to evaluate the impact of different Channel Preservation Rates (CPRs) on the dehazing performance of the model. As shown in Table 4, a detailed performance comparison of dehazing models trained with different CPRs was presented. The original teacher model has 1.53 M Params. For further optimization, the experiments were conducted by compressing the model using CPRs of 1/2, 1/3, 1/4, and 1/5. Significant degradation in dehazing performance is not observed when the parameter count is decreased to 0.06 M. Specifically, when CPR was set to 1/5, the PSNR and SSIM reached 23.9271 and 0.9105, respectively, which were nearly comparable to those of the original model (25.1759 and 0.9223). At this time, the Params and FLOPs of the student model are 0.06 M and 0.44 G, respectively, which is a significant optimization compared to the 1.53 M parameters and 10.79 GFLOPs of the teacher model. Overall, our method demonstrates a strong knowledge distillation capability in the field of dehazing and can generate an efficient model that is easy to deploy and apply on edge computing devices and smart terminals.
Additionally, ablation experiments were conducted to evaluate the effects of different knowledge transfer settings, and the results are presented in Table 5. When no transfer strategy was employed and 1/5-SCSNet was trained directly, the performance was moderate. The incorporation of the Intra-KT strategy led to a significant enhancement in model performance, with PSNR and SSIM metrics increasing to 22.3718 dB and 0.9051, respectively. This indicates that intra-layer knowledge transfer can effectively transfer knowledge from the teacher network to the student network. When only the Inter-KT transfer strategy was used, the model’s PSNR, SSIM, and LPIPS metrics were 21.5281 dB, 0.8937, and 0.0768, respectively. Compared to the results without any transfer strategy, this also showed substantial improvement, suggesting that the proposed Inter-KT transfer module can effectively transfer cross-layer knowledge from the teacher network to the student network. The optimal dehazing performance was achieved when both Intra-KT and Inter-KT were combined. This is because the two proposed transfer modules facilitate knowledge transfer from complementary perspectives, demonstrating the superiority of the proposed approach.

3.5. Discussion

In this study, we have presented a comprehensive evaluation of the proposed Dual-View Knowledge Transfer (DVKT) framework for RSID. Our results demonstrate the efficacy of the DVKT approach in improving both the performance and efficiency of RSID models.
Firstly, the quantitative and qualitative evaluations on benchmark datasets (SateHaze1k, RS-Haze, and RSID) confirm that the DVKT framework almost achieves a state-of-the-art dehazing performance, comparable to heavy backbone models like SCSNet, while requiring significantly fewer parameters and computational resources. This highlights the effectiveness of our knowledge distillation approach in transferring structured knowledge from a well-trained teacher network to a lightweight student network.
Secondly, the ablation study investigating the impact of different channel preservation rates further validates the robustness of our framework. By progressively reducing the number of channels, we show that the DVKT framework can maintain high dehazing quality even with extremely compressed models. This finding underscores the practicality of our method for deployment on edge devices and mobile platforms with limited computational resources.
Finally, the proposed DVKT framework addresses two critical issues in remote-sensing image dehazing: lack of labeled data and high computational costs. By leveraging the abundance of natural-image-dehazing datasets and transferring knowledge from a pre-trained teacher model, we overcome the data scarcity challenge. Additionally, the lightweight student model generated through DVKT achieves significant speedups compared to heavy backbone networks, demonstrating its potential to enable real-time dehazing for various remote-sensing applications.

4. Conclusions

In summary, this study proposes an innovative Dual-View Knowledge Transfer (DVKT) framework to tackle the challenges of efficiency and dataset scarcity in RSID. The framework includes two innovative modules: Intra-layer Knowledge Transfer (Intra-KT) and Inter-layer Knowledge Transfer (Inter-KT), which can successfully distill hierarchical features and cross-layer correlations from a pre-trained teacher network to a lightweight student network. Specifically, the Intra-KT module can correct the learning bias of the student network by the guidance of representative features from a well-trained teacher network. The Inter-KT module can extract compact and effective features for RSID by distilling and transferring knowledge about cross-layer correlations. Comprehensive evaluations confirm that the proposed DVKT framework successfully compresses RSID models, yielding dual enhancements in both dehazing performance and efficiency.
While the proposed DVKT framework exhibits promising results, it currently relies on a pre-trained teacher model trained on natural-image-dehazing datasets. This may introduce a domain gap between the teacher and student networks, limiting the optimal performance for remote-sensing images. Future work could explore domain adaptation techniques to better bridge this gap.

Author Contributions

Conceptualization, L.Y. and J.C.; methodology, L.Y.; software, H.B.; validation, R.Q.; formal analysis, H.G.; investigation, L.Y.; resources, H.B.; data curation, R.Q.; writing—original draft preparation, L.Y.; writing—review and editing, L.Y.; visualization, H.G.; supervision, J.C. and H.N.; project administration, J.C.; funding acquisition, H.N. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China under Grant 62201452.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study.

Acknowledgments

We would also like to express our gratitude to the anonymous reviewers and the editors for their valuable advice and assistance.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Huang, Y.; Xiong, S. Remote sensing image dehazing using adaptive region-based diffusion models. IEEE Geosci. Remote Sens. Lett. 2023, 20, 8001805. [Google Scholar] [CrossRef]
  2. Su, H.; Liu, L.; Jeon, G.; Wang, Z.; Guo, T.; Gao, M. Remote Sensing Image Dehazing Based on Dual Attention Parallelism and Frequency Domain Selection Network. IEEE Trans. Consum. Electron. 2024; early access. [Google Scholar]
  3. Zheng, Y.; Su, J.; Zhang, S.; Tao, M.; Wang, L. Dehaze-TGGAN: Transformer-Guide Generative Adversarial Networks with Spatial-Spectrum Attention for Unpaired Remote Sensing Dehazing. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5634320. [Google Scholar] [CrossRef]
  4. Kulkarni, A.; Murala, S. Aerial image dehazing with attentive deformable transformers. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 2–7 January 2023; pp. 6305–6314. [Google Scholar]
  5. Wang, Y.; Yan, X.; Wang, F.L.; Xie, H.; Yang, W.; Zhang, X.P.; Qin, J.; Wei, M. Ucl-dehaze: Towards real-world image dehazing via unsupervised contrastive learning. IEEE Trans. Image Process. 2024, 33, 1361–1374. [Google Scholar] [CrossRef]
  6. Zheng, Y.; Zhan, J.; He, S.; Dong, J.; Du, Y. Curricular contrastive regularization for physics-aware single image dehazing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023; pp. 5785–5794. [Google Scholar]
  7. Guo, F.; Yang, J.; Liu, Z.; Tang, J. Haze removal for single image: A comprehensive review. Neurocomputing 2023, 537, 85–109. [Google Scholar] [CrossRef]
  8. Berman, D.; Avidan, S. Non-local image dehazing. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 1674–1682. [Google Scholar]
  9. Xu, L.; Zhao, D.; Yan, Y.; Kwong, S.; Chen, J.; Duan, L.Y. IDeRs: Iterative dehazing method for single remote sensing image. Inf. Sci. 2019, 489, 50–62. [Google Scholar] [CrossRef]
  10. Liang, S.; Gao, T.; Chen, T.; Cheng, P. A Remote Sensing Image Dehazing Method Based on Heterogeneous Priors. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5619513. [Google Scholar] [CrossRef]
  11. He, K.; Sun, J.; Tang, X. Single image haze removal using dark channel prior. IEEE Trans. Pattern Anal. Mach. Intell. 2010, 33, 2341–2353. [Google Scholar] [PubMed]
  12. He, K.; Sun, J.; Tang, X. Guided image filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 1397–1409. [Google Scholar] [CrossRef]
  13. Zhu, Q.; Mai, J.; Shao, L. A fast single image haze removal algorithm using color attenuation prior. IEEE Trans. Image Process. 2015, 24, 3522–3533. [Google Scholar]
  14. Fattal, R. Dehazing using color-lines. ACM Trans. Graph. (TOG) 2014, 34, 1–14. [Google Scholar] [CrossRef]
  15. Zheng, Z.; Ren, W.; Cao, X.; Hu, X.; Wang, T.; Song, F.; Jia, X. Ultra-high-definition image dehazing via multi-guided bilateral learning. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 16180–16189. [Google Scholar]
  16. Wang, Y.; Xiong, J.; Yan, X.; Wei, M. Uscformer: Unified transformer with semantically contrastive learning for image dehazing. IEEE Trans. Intell. Transp. Syst. 2023, 24, 11321–11333. [Google Scholar] [CrossRef]
  17. Li, C.; Yu, H.; Zhou, S.; Liu, Z.; Guo, Y.; Yin, X.; Zhang, W. Efficient dehazing method for outdoor and remote sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2023, 16, 4516–4528. [Google Scholar] [CrossRef]
  18. Tong, L.; Liu, Y.; Li, W.; Chen, L.; Chen, E. Haze-Aware Attention Network for Single-Image Dehazing. Appl. Sci. 2024, 14, 5391. [Google Scholar] [CrossRef]
  19. Chi, K.; Yuan, Y.; Wang, Q. Trinity-Net: Gradient-guided Swin transformer-based remote sensing image dehazing and beyond. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4702914. [Google Scholar] [CrossRef]
  20. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  21. Shyam, P.; Yoo, H. Data Efficient Single Image Dehazing via Adversarial Auto-Augmentation and Extended Atmospheric Scattering Model. In Proceedings of the IEEE/CVF International Conference on Computer Vision Workshops, Paris, France, 2–6 October 2023; pp. 227–237. [Google Scholar]
  22. Leng, C.; Liu, G. IFE-Net: An Integrated Feature Extraction Network for Single-Image Dehazing. Appl. Sci. 2023, 13, 12236. [Google Scholar] [CrossRef]
  23. Qin, X.; Wang, Z.; Bai, Y.; Xie, X.; Jia, H. FFA-Net: Feature fusion attention network for single image dehazing. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 11908–11915. [Google Scholar]
  24. Guo, C.; Yan, Q.; Anwar, S.; Cong, R.; Ren, W.; Li, C. Image dehazing transformer with transmission-aware 3d position embedding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5812–5820. [Google Scholar]
  25. Huang, P.; Zhao, L.; Jiang, R.; Wang, T.; Zhang, X. Self-filtering image dehazing with self-supporting module. Neurocomputing 2021, 432, 57–69. [Google Scholar] [CrossRef]
  26. Wang, C.; Shen, H.Z.; Fan, F.; Shao, M.W.; Yang, C.S.; Luo, J.C.; Deng, L.J. EAA-Net: A novel edge assisted attention network for single image dehazing. Knowl.-Based Syst. 2021, 228, 107279. [Google Scholar] [CrossRef]
  27. Sun, H.; Luo, Z.; Ren, D.; Hu, W.; Du, B.; Yang, W.; Wan, J.; Zhang, L. Partial siamese with multiscale bi-codec networks for remote sensing image haze removal. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4106516. [Google Scholar] [CrossRef]
  28. Song, Y.; He, Z.; Qian, H.; Du, X. Vision transformers for single image dehazing. IEEE Trans. Image Process. 2023, 32, 1927–1941. [Google Scholar] [CrossRef]
  29. Li, Y.; Chen, X. A coarse-to-fine two-stage attentive network for haze removal of remote sensing images. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1751–1755. [Google Scholar] [CrossRef]
  30. Song, T.; Fan, S.; Li, P.; Jin, J.; Jin, G.; Fan, L. Learning an effective transformer for remote sensing satellite image dehazing. IEEE Geosci. Remote Sens. Lett. 2023, 20, 8002305. [Google Scholar] [CrossRef]
  31. Zhang, L.; Wang, S. Dense haze removal based on dynamic collaborative inference learning for remote sensing images. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5631016. [Google Scholar] [CrossRef]
  32. Zhang, Y.; He, X.; Zhan, C.; Li, J. Frequency-Oriented Transformer for Remote Sensing Image Dehazing. Sensors 2024, 24, 3972. [Google Scholar] [CrossRef] [PubMed]
  33. Zheng, Y.; Su, J.; Zhang, S.; Tao, M.; Wang, L. Dehaze-AGGAN: Unpaired remote sensing image dehazing using enhanced attention-guide generative adversarial networks. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5630413. [Google Scholar] [CrossRef]
  34. Yu, M.; Xu, S.; Sun, H.; Zheng, Y.; Yang, W. Hierarchical slice interaction and multi-layer cooperative decoding networks for remote sensing image dehazing. Image Vis. Comput. 2024, 148, 105129. [Google Scholar] [CrossRef]
  35. Ayazoglu, M. Extremely lightweight quantization robust real-time single-image super resolution for mobile devices. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 2472–2479. [Google Scholar]
  36. Xue, L.; Wang, X.; Wang, Z.; Li, G.; Song, H.; Song, Z. A Lightweight Patch-Level Change Detection Network Based on Multi-layer Feature Compression and Sensitivity-Guided Network Pruning. IEEE Trans. Geosci. Remote Sens. 2024, 62, 5624219. [Google Scholar] [CrossRef]
  37. Yang, L.; Cao, J.; Wang, H.; Dong, S.; Ning, H. Hierarchical Semantic-Guided Contextual Structure-Aware Network for Spectral Satellite Image Dehazing. Remote Sens. 2024, 16, 1525. [Google Scholar] [CrossRef]
  38. Li, B.; Ren, W.; Fu, D.; Tao, D.; Feng, D.; Zeng, W.; Wang, Z. Benchmarking single-image dehazing and beyond. IEEE Trans. Image Process. 2018, 28, 492–505. [Google Scholar] [CrossRef]
  39. Beyer, L.; Zhai, X.; Royer, A.; Markeeva, L.; Anil, R.; Kolesnikov, A. Knowledge distillation: A good teacher is patient and consistent. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 10925–10934. [Google Scholar]
  40. Huang, B.; Zhi, L.; Yang, C.; Sun, F.; Song, Y. Single satellite optical imagery dehazing using SAR image prior based on conditional generative adversarial networks. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Snowmass Village, CO, USA, 1–5 March 2020; pp. 1806–1813. [Google Scholar]
  41. Wen, Y.; Gao, T.; Zhang, J.; Li, Z.; Chen, T. Encoder-free multi-axis physics-aware fusion network for remote sensing image dehazing. IEEE Trans. Geosci. Remote Sens. 2023, 61, 4705915. [Google Scholar] [CrossRef]
  42. Li, K.; Wan, G.; Cheng, G.; Meng, L.; Han, J. Object detection in optical remote sensing images: A survey and a new benchmark. ISPRS J. Photogramm. Remote Sens. 2020, 159, 296–307. [Google Scholar] [CrossRef]
  43. Kendall, A.; Gal, Y.; Cipolla, R. Multi-task learning using uncertainty to weigh losses for scene geometry and semantics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7482–7491. [Google Scholar]
  44. Bui, T.M.; Kim, W. Single image dehazing using color ellipsoid prior. IEEE Trans. Image Process. 2017, 27, 999–1009. [Google Scholar] [CrossRef] [PubMed]
  45. Ling, P.; Chen, H.; Tan, X.; Jin, Y.; Chen, E. Single Image Dehazing Using Saturation Line Prior. IEEE Trans. Image Process. 2023, 32, 3238–3253. [Google Scholar] [CrossRef] [PubMed]
  46. Li, B.; Peng, X.; Wang, Z.; Xu, J.; Feng, D. Aod-net: All-in-one dehazing network. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 4770–4778. [Google Scholar]
  47. Zhao, D.; Xu, L.; Yan, Y.; Chen, J.; Duan, L.Y. Multi-scale optimal fusion model for single image dehazing. Signal Process. Image Commun. 2019, 74, 253–265. [Google Scholar] [CrossRef]
  48. Ullah, H.; Muhammad, K.; Irfan, M.; Anwar, S.; Sajjad, M.; Imran, A.S.; de Albuquerque, V.H.C. Light-DehazeNet: A novel lightweight CNN architecture for single image dehazing. IEEE Trans. Image Process. 2021, 30, 8968–8982. [Google Scholar] [CrossRef]
  49. Zamir, S.W.; Arora, A.; Khan, S.; Hayat, M.; Khan, F.S.; Yang, M.H. Restormer: Efficient transformer for high-resolution image restoration. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 18–24 June 2022; pp. 5728–5739. [Google Scholar]
  50. Huang, Y.; Chen, X. Single remote sensing image dehazing using a dual-step cascaded residual dense network. In Proceedings of the 2021 IEEE International Conference on Image Processing (ICIP), Anchorage, AK, USA, 19–22 September 2021; pp. 3852–3856. [Google Scholar]
Figure 1. The framework of the proposed Dual-View Knowledge Transfer (DVKT) for remote-sensing image dehazing.
Figure 1. The framework of the proposed Dual-View Knowledge Transfer (DVKT) for remote-sensing image dehazing.
Applsci 14 08633 g001
Figure 2. Visual comparisons of dehazing performance across various methods on the SateHaze1k-thin dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Figure 2. Visual comparisons of dehazing performance across various methods on the SateHaze1k-thin dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Applsci 14 08633 g002
Figure 3. QVisual comparisons of dehazing performance across various methods on the SateHaze1k-moderate dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Figure 3. QVisual comparisons of dehazing performance across various methods on the SateHaze1k-moderate dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Applsci 14 08633 g003
Figure 4. Visual comparisons of dehazing performance across various methods on the SateHaze1k-thick dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Figure 4. Visual comparisons of dehazing performance across various methods on the SateHaze1k-thick dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Applsci 14 08633 g004
Figure 5. Visual comparisons of dehazing performance across various methods on the RS-Haze dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Figure 5. Visual comparisons of dehazing performance across various methods on the RS-Haze dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Applsci 14 08633 g005
Figure 6. Visual comparisons of dehazing performance across various methods on the RSID dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Figure 6. Visual comparisons of dehazing performance across various methods on the RSID dataset. The red box emphasizes key areas for comparison. (a) Hazy Image; (b) GroundTruth; (c) DCP [11]; (d) AOD-Net [46]; (e) Light-DehazeNet [48]; (f) DehazeFormer [28]; (g) SLP [45]; (h) Trinity-Net [19]; (i) DCRD-Net [50]; (j) FCTF-Net [29]; (k) SCSNet (Teacher Network) [37]; (l) DVKT (ours).
Applsci 14 08633 g006
Figure 7. Visual comparisons of dehazing performance across various methods on the RRSD300 dataset. (a) Hazy Image; (b) DCP [11]; (c) AOD-Net [46]; (d) Light-DehazeNet [48]; (e) DehazeFormer [28]; (f) SLP [45]; (g) DCRD-Net [50]; (h) FCTF-Net [29]; (i) SCSNet (Teacher Network) [37]; (j) DVKT (ours).
Figure 7. Visual comparisons of dehazing performance across various methods on the RRSD300 dataset. (a) Hazy Image; (b) DCP [11]; (c) AOD-Net [46]; (d) Light-DehazeNet [48]; (e) DehazeFormer [28]; (f) SLP [45]; (g) DCRD-Net [50]; (h) FCTF-Net [29]; (i) SCSNet (Teacher Network) [37]; (j) DVKT (ours).
Applsci 14 08633 g007
Table 1. Qualitative comparison results on the SateHaze1k dataset.
Table 1. Qualitative comparison results on the SateHaze1k dataset.
MethodsThin HazeModerate HazeThick HazeMixed Haze
PSNRSSIMLPIPSPSNRSSIMLPIPSPSNRSSIMLPIPSPSNRSSIMLPIPS
DCP [11]19.11830.85180.145119.83840.88120.119816.79300.77010.210318.58330.83440.1602
CEP [44]13.59970.72220.320114.21220.72700.229416.08240.77620.209714.69500.75120.2703
MOF [47]15.38910.72910.300214.74180.62560.330116.24950.67670.330315.51460.68590.3108
AOD-Net [46]19.05480.77770.190319.42110.70150.240116.46720.71230.270217.48590.63320.3382
Light-DehazeNet [48]18.48680.86580.129918.39180.88250.100216.76620.76970.225317.81320.83520.1499
FFA-Net [23]20.14100.85820.114922.55860.91320.065719.12550.79760.199221.28730.86630.0998
Restormer [49]20.98290.86860.104923.15740.90360.059819.69840.77390.213720.78920.83790.1479
DehazeFormer [28]21.92740.88430.091824.44070.92680.057920.21330.80490.163522.00660.86590.1099
SLP [45]16.41620.74280.276218.61820.78300.183716.50710.76590.225017.24830.77060.1852
Trinity-Net [19]19.37080.78160.176119.69180.70620.225217.12810.77230.211617.65080.78640.1940
DCRD-Net [50]20.84730.87670.097823.31190.92250.058919.72500.81210.137821.74680.88120.0849
FCTF-Net [29]18.32620.83690.136220.90570.85530.129517.25510.69220.289719.58830.84340.1498
SCSNet (Teacher Network) [37]26.14600.94150.058328.35010.95660.039824.65420.90150.064925.17590.92230.0519
DVKT24.72840.91830.061827.04620.94170.057923.31360.88620.091923.92710.91050.0609
The bold font highlights our proposed method.
Table 2. Qualitative results on the RS-Haze and RSID datasets.
Table 2. Qualitative results on the RS-Haze and RSID datasets.
MethodsRS-HazeRSID
PSNRSSIMLPIPSPSNRSSIMLPIPS
DCP [11]18.10030.67040.280117.32560.79270.1699
CEP [44]15.90970.57720.350214.23750.70340.2192
MOF [47]16.16080.56280.360312.90520.64850.2901
AOD-Net [46]23.96770.72070.190218.70370.74240.1781
Light-DehazeNet [48]25.59650.82090.120317.92790.84140.1099
FFA-Net [23]29.19320.88460.089821.28760.90420.0749
Restormer [49]25.67000.75630.200111.72400.59710.3203
DehazeFormer [28]29.34190.87300.079822.68590.91180.0649
SLP [45]17.38050.65270.288216.72600.77510.1735
Trinity-Net [19]29.70820.89730.071423.52730.92070.0562
DCRD-Net [50]29.67800.88780.074922.16430.89260.0898
FCTF-Net [29]29.62400.89580.069820.25560.83970.1203
SCSNet (Teacher Network) [37]32.25040.92710.049925.38210.95850.0398
DVKT31.42270.91350.054124.10820.94260.0449
The bold font highlights our proposed method.
Table 3. Comparison of inference efficiency on the Satehaze1k dataset.
Table 3. Comparison of inference efficiency on the Satehaze1k dataset.
MethodsParamsFLOPSTimePSNRSSIMLPIPS
DCP [11]0.260418.58330.83440.1602
AOD-Net [46]0.010.150.045217.48590.63320.2498
Light-DehazeNet [48]0.031.970.057117.81320.83520.1199
FFA-Net [23]4.45287.530.446521.28730.86630.0898
Restormer [49]26.13141.170.315720.78920.83790.0749
DehazeFormer [28]9.6889.850.243822.00660.86590.0599
Trinity-Net [19]9.8290.170.261517.65080.78640.1940
DCRD-Net [50]0.0414.920.182721.74680.88120.0549
FCTF-Net [29]0.1610.040.157419.58830.84340.1098
SCSNet (Teacher Network) [37]1.5310.790.171325.17590.92230.0419
DVKT0.060.440.022523.92710.91050.0509
The bold font highlights our proposed method.
Table 4. Ablation study about channel preservation rates (CPR) on SateHaze1k dataset.
Table 4. Ablation study about channel preservation rates (CPR) on SateHaze1k dataset.
CPRParamsFLOPSPSNRSSIMLPIPS
11.5310.7925.17590.92230.0386
1/20.392.7124.67170.91840.0407
1/30.181.2024.38040.91620.0424
1/40.100.6824.11630.91400.0441
1/50.060.4423.92710.91050.0509
Table 5. Ablation study about different transfer settings on SateHaze1k dataset.
Table 5. Ablation study about different transfer settings on SateHaze1k dataset.
Transfer SettingsPSNRSSIMLPIPS
W/T Transfer19.82060.86250.1357
Only Intra-KT22.37180.90510.0521
Only Inter-KT21.52810.89370.0768
Intra-KT + Inter-KT23.92710.91050.0509
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Yang, L.; Cao, J.; Bian, H.; Qu, R.; Guo, H.; Ning, H. Remote Sensing Image Dehazing via Dual-View Knowledge Transfer. Appl. Sci. 2024, 14, 8633. https://doi.org/10.3390/app14198633

AMA Style

Yang L, Cao J, Bian H, Qu R, Guo H, Ning H. Remote Sensing Image Dehazing via Dual-View Knowledge Transfer. Applied Sciences. 2024; 14(19):8633. https://doi.org/10.3390/app14198633

Chicago/Turabian Style

Yang, Lei, Jianzhong Cao, He Bian, Rui Qu, Huinan Guo, and Hailong Ning. 2024. "Remote Sensing Image Dehazing via Dual-View Knowledge Transfer" Applied Sciences 14, no. 19: 8633. https://doi.org/10.3390/app14198633

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop