Next Article in Journal
Exploring Stand Parameters Using Terrestrial Laser Scanning in Pinus tabuliformis Plantation Forests
Previous Article in Journal
Spatiotemporal Trends and Zoning Geospatial Assessment in China’s Offshore Mariculture (2018–2022)
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MFAFNet: Multi-Scale Feature Adaptive Fusion Network Based on DeepLab V3+ for Cloud and Cloud Shadow Segmentation

1
School of Future Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
2
School of Automation, Nanjing University of Information Science and Technology, Nanjing 210044, China
3
Collaborative Innovation Center on Atmospheric Environment and Equipment Technology, Nanjing University of Information Science and Technology, Nanjing 210044, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2025, 17(7), 1229; https://doi.org/10.3390/rs17071229
Submission received: 23 February 2025 / Revised: 26 March 2025 / Accepted: 28 March 2025 / Published: 30 March 2025

Abstract

:
The accurate segmentation of clouds and cloud shadows is crucial in meteorological monitoring, climate change research, and environmental management. However, existing segmentation models often suffer from issues such as losing fine details, blurred boundaries, and false positives or negatives. To address these challenges, this paper proposes an improved model based on DeepLab v3+. First, to enhance the model’s ability to extract fine-grained features, a Hybrid Strip Pooling Module (HSPM) is introduced in the encoding stage, effectively preserving local details and reducing information loss. Second, a Global Context Attention Module (GCAM) is incorporated into the Atrous Spatial Pyramid Pooling (ASPP) module to establish pixel-wise long-range dependencies, thereby effectively integrating global semantic information. In the decoding stage, a Three-Branch Adaptive Feature Fusion Module (TB-AFFM) is designed to merge multi-scale features from the backbone network and ASPP. Finally, an innovative loss function is employed in the experiments, significantly improving the accuracy of cloud and cloud shadow segmentation. Experimental results demonstrate that the proposed model outperforms existing methods in cloud and cloud shadow segmentation tasks, achieving more precise segmentation performance.

1. Introduction

In the field of remote sensing image processing, cloud and cloud shadow detection is a critical research area that plays a key role in enhancing the quality of remote sensing images. In satellite imagery, cloud and cloud shadow often obscure ground targets, leading to reduced illumination uniformity and color accuracy. This obstruction prevents the image from faithfully reflecting surface information, impacting the accuracy and reliability of downstream tasks such as land cover classification and object detection. Therefore, in applications that require precise surface information, such as land and water resource monitoring, the accurate segmentation of cloud and cloud shadow is crucial. Segmentation refers to dividing an image into multiple regions with distinct characteristics. In this study, it specifically involves categorizing remote sensing images into cloud, cloud shadow, and surface regions, thus improving image quality and minimizing analytical errors. Furthermore, the segmented cloud data can be utilized to study cloud distribution, movement patterns, and their impact on climate forecasting and disaster assessment. Thus, enhancing the accuracy of cloud and cloud shadow detection is of paramount importance.
Traditional cloud and cloud shadow detection methods can be categorized into three main types: threshold-based methods, spatial feature-based methods, and machine learning-based methods. Threshold-based methods utilize the spectral brightness characteristics of clouds and cloud shadows to achieve segmentation through fixed or adaptive thresholds. For example, Rossow et al. [1] proposed the ISCCP algorithm which detects cloud regions using fixed thresholds; however, this method suffers from reduced accuracy in complex scenes and is susceptible to misclassification. To address the limitations of threshold-based methods, Sun et al. [2] proposed the UDTCDA algorithm, which dynamically adjusts thresholds to enhance its adaptability to various cloud types and complex surface backgrounds. Spatial feature-based methods enhance detection performance by analyzing the texture or geometric properties of clouds and cloud shadows. For instance, Vásquez and Manian [3] proposed a cloud detection method based on texture features in remote sensing images, achieving better results than pixel-level threshold classification. Zhu et al. [4] developed the Fmask algorithm, which integrates spectral, shadow, and thermal features to generate highly accurate cloud and cloud shadow masks. Additionally, Danda et al. [5] introduced a morphology-based detection method that improves segmentation by analyzing the shape and spatial distribution of clouds. However, these methods often struggle to distinguish thin clouds and irregular cloud shadows, as their texture and geometric properties closely resemble those of background objects. With the rapid advancement of machine learning, models such as random forests and support vector machines have also been applied to cloud and cloud shadow detection. For example, Le Hégarat-Mascle and André [6] combined the Markov Random Field (MRF) model for the automatic detection of clouds and cloud shadows in high-resolution optical imagery. Cheng and Lin [7] combined multi-scale neighborhood features with multiple classifiers to enhance segmentation accuracy. Wei et al. [8] proposed the RFmask algorithm, which integrates energy-driven superpixel segmentation features to optimize cloud detection in Landsat imagery. Although machine learning methods significantly improve detection accuracy, they rely heavily on feature engineering. The handcrafted features may fail to capture the diverse characteristics of cloud and cloud shadow in complex scenes, limiting both detection accuracy and generalization capability.
Deep learning-based cloud and cloud shadow segmentation methods have gained increasing attention in the field of remote sensing image processing due to their end-to-end network structures, which enable high-precision segmentation in complex scenes. Since 2015, numerous classical network architectures have been proposed and applied to cloud and cloud shadow segmentation tasks. Long et al. [9] first introduced the Fully Convolutional Network (FCN), which is an end-to-end pixel-wise classification network that established the foundation for semantic segmentation. In the same year, Ronneberger et al. [10] developed the U-Net network for biomedical image segmentation, effectively recovering information lost during downsampling through skip connections, thereby significantly improving segmentation accuracy. The application of these network structures in cloud and cloud shadow segmentation has been progressively validated. In 2018, Miller et al. [11] applied Convolutional Neural Networks (CNNs) to meteorological image recognition, demonstrating superior segmentation accuracy compared to traditional random forest classifiers. In the same year, Mohajerani et al. [12] employed FCN for the pixel-level annotation of cloud regions in Landsat 8 imagery, achieving substantial improvements in the Jaccard index and recall rate. In 2019, Gonzales and Sakla [13] utilized a pre-trained deep U-Net for the semantic segmentation of clouds in satellite images, outperforming state-of-the-art networks on benchmark datasets based on several segmentation metrics. To further enhance segmentation accuracy and network efficiency, encoder–decoder architectures have become a key research focus. In 2017, Badrinarayanan et al. [14] proposed SegNet, which minimizes information loss during upsampling by employing pooling indices in the decoding stage. Subsequently, Lu et al. [15] developed a cloud segmentation method based on SegNet, investigating the impact of symmetric network structures on segmentation accuracy. In the same year, Zhao et al. [16] introduced the Pyramid Scene Parsing Network (PSPNet), which aggregates global contextual information to improve segmentation precision. Later, Chen et al. [17] proposed the DeepLab v3+ model in 2018, incorporating depthwise separable convolutions in the spatial pyramid pooling and decoder modules to enhance segmentation efficiency and accuracy. Additionally, in 2019, Sun et al. [18] introduced the High-Resolution Network (HRNet), which maintained continuous bidirectional information flow between high- and low-resolution features, achieving more precise cloud boundary detection. Recent studies have increasingly focused on the integration of attention mechanisms and multi-scale feature fusion to enhance segmentation performance. In 2021, Qu et al. [19] proposed the Strip Pooling Channel Spatial Attention Network (SP_CSANet), which enhances cloud and cloud shadow edge segmentation accuracy through strip pooling while integrating shallow and deep contextual information to improve model performance. However, its reliance on strip pooling may limit its ability to capture fine-grained boundary details, particularly in areas with thin clouds or light cloud shadows. In 2022, Lu et al. [20] introduced a Dual-Branch Network (DBNet) combining Transformer and convolutional networks, leveraging an interactive guidance module to extract deep features and refining coarse segmentation boundary representations in the decoding stage. Compared with SP_CSANet, DBNet provides more refined segmentation details; however, due to its complex structure, it incurs higher computational costs. In 2023, Dai et al. [21] designed the Location Pooling Multi-Scale Network (LPMSNet), which enhances target region attention by aggregating multi-scale information and incorporating attention mechanisms. Compared with the previous two models, LPMSNet achieves a balance between segmentation accuracy and computational efficiency. Nevertheless, in scenarios with severe interference, LPMSNet may still encounter limitations. The most recent research by Hu et al. [22] in 2024 introduced the Multibranch Hybrid Segmentation Network (HyCloudX), which integrates spectral band information in a multibranch architecture, demonstrating significant improvements in segmentation performance and generalization capability under complex noise interference.
Although existing deep learning models have made significant progress in cloud and cloud shadow segmentation tasks, several challenges still remain. First, the multi-scale feature fusion is insufficient and inefficient, leading to the loss of fine-grained details and the weakening of semantic information. Second, the models lack sufficient attention to fine-grained features and boundary information, resulting in blurred segmentation in transition areas. Finally, current methods exhibit limited global context awareness and struggle to capture long-range dependencies within scenes, making them less adaptable to diverse scenarios. In recent years, attention mechanisms have emerged as an important approach to enhance the model’s adaptability to complex scenarios due to its advantage in modeling global context information [23,24,25]. Simultaneously, multi-scale feature fusion techniques have proven effective in improving segmentation accuracy for diverse cloud types and complex boundaries while preserving local details, further enhancing the performance of deep learning models in remote sensing image semantic segmentation [26,27,28,29]. Therefore, to address the aforementioned issues, this paper proposes a Multi-scale Feature Adaptive Fusion Network (MFAFNet) based on the DeepLab v3+ model. DeepLab v3+ is selected as the foundational model due to its strong capability to capture multi-scale contextual information, which aligns well with the significant scale variations in cloud and cloud shadow segmentation. Furthermore, as a widely recognized model in the field of semantic segmentation, DeepLab v3+ offers a clear and modular architecture that facilitates modifications and enhancements. This provides a solid foundation for evaluating the effectiveness of our proposed improvement modules fairly and reliably. In the encoder, we replace the global pooling in the Atrous Spatial Pyramid Pooling (ASPP) with a Hybrid Strip Pooling Module (HSPM), avoiding the spatial information loss that may occur in traditional global pooling. Additionally, a Global Context Attention Module (GCAM) is added to each branch of the ASPP to optimize feature extraction. In the decoding stage, we use a Three-Branch Adaptive Feature Fusion Module (TB-AFFM) to adaptively fuse mid-level and high-level features from the backbone network with deep high-level features extracted by the ASPP, effectively improving feature fusion. Furthermore, we propose a hybrid loss function that combines focal loss and Dice loss, balancing pixel-level and region-level segmentation accuracy. The main contributions of this work are as follows:
  • A Multi-scale Feature Adaptive Fusion Network (MFAFNet) based on DeepLab v3+ is proposed to enhance feature extraction capabilities, optimize multi-scale feature fusion, and improve the accuracy of semantic segmentation for clouds and cloud shadows.
  • To avoid the potential spatial information loss caused by global pooling, we replace the global pooling in the ASPP structure with a Hybrid Strip Pooling Module (HSPM), which can extract global distribution information while focusing on features of locally salient regions, thus enhancing adaptability to different cloud types and complex cloud shadow boundaries.
  • Considering the spatial correlation between clouds and cloud shadows, we introduce a Global Context Attention Module (GCAM) into each branch of the ASPP, enabling the branches to better handle high-level semantic information and optimize feature extraction.
  • A Three-Branch Adaptive Feature Fusion Module (TB-AFFM) is employed to fuse mid-level and high-level features extracted from the backbone network with deep high-level features from the ASPP. This module adaptively adjusts weights in both the channel and spatial dimensions, improving the semantic understanding of complex cloud shadow scenes while preserving key detail information.
  • A weighted hybrid loss function combining focal loss and Dice loss is adopted to enhance the overall segmentation accuracy and improve boundary segmentation performance.

2. Methodology

2.1. Network Structure

Based on these observations, this paper proposes an improved DeepLab v3+ network model, as illustrated in Figure 1. In the encoding phase, the global pooling module in the ASPP structure is replaced with a Hybrid Strip Pooling Module (HSPM) to mitigate information loss during pooling. Additionally, a Global Context Attention Module (GCAM) is incorporated into each branch to enhance feature extraction by integrating global contextual information. In the decoding phase, the original DeepLab v3+ model utilizes only low-level features from Block2 of the backbone network. In contrast, the proposed improved model introduces a Three-Branch Adaptive Feature Fusion Module (TB-AFFM) in the decoding stage, adaptively fusing mid-level and high-level features from Block 3 and Block 4 of the backbone network with deep high-level features extracted by ASPP. This fusion effectively balances global semantics and local detail representation, significantly improving cloud and cloud shadow segmentation performance.

2.2. Hybrid Strip Pooling Module (HSPM)

The global pooling layer generates global contextual information through simple average pooling; however, it compresses all spatial information of the entire feature map into a single scalar, often leading to the loss of local details and spatial information. In cloud and cloud shadow segmentation tasks, local information such as boundaries and fine details is crucial for precise segmentation. Therefore, this paper proposes a Hybrid Strip Pooling Module (HSPM), which builds upon traditional strip pooling [30] by integrating both horizontal and vertical average and max pooling operations. The structure of HSPM is illustrated in Figure 2.
Given an input feature map X with dimensions H × W , average and max pooling operations are applied along the height direction, producing an output of size H × 1 . Similarly, average and max pooling operations are performed along the width direction, resulting in an output of size 1 × W . The specific formulations are as follows:
y a v g h = 1 W 0 j < W x i , j
y a v g v = 1 H 0 i < H x i , j
y m a x h = max 0 j < W ( x i , j )
y m a x v = max 0 i < H ( x i , j )
The results of each pooling operation are then fused across channels through convolution and expanded to match the original feature dimensions H × W . These outputs are combined through additive fusion to generate a feature map enriched with contextual information:
y = y a v g h + y a v g v + y m a x h + y m a x v
The fused feature map is further processed using a 1 × 1 convolution for dimensionality reduction and subsequently activated by the Sigmoid function to produce an attention-weight map. This weight map is applied to the input features via element-wise multiplication, yielding the final output Z as follows:
z = S i g m o i d ( x , σ ( C o n v ( y ) ) )

2.3. Three-Branch Adaptive Feature Fusion Module (TB-AFFM)

The scale variation in clouds and cloud shadows is often significant, and fusing features from different scales is crucial for improving segmentation accuracy. The original DeepLab v3+ model only fuses low-level features (block2) with deep high-level features from the ASPP module, neglecting the important information carried by intermediate layer features. To utilize more comprehensive multi-scale feature information, this paper introduces mid-level and high-level features from the backbone network (Block 3 and Block 4). After upsampling to a unified spatial resolution, these features are fused with the deep high-level features output by the ASPP module. Unlike traditional concatenation or simple addition, this paper employs a three-branch adaptive feature fusion module to dynamically allocate fusion weights to the three input features, as shown in Figure 3.
For the input feature map X R C × H × W , we design three parallel branches for feature processing. In the global branch, global average pooling is used to extract global distribution information M R C × 1 × 1 from the channel domain. A 1 × 1 point convolution layer with a scaling factor of r is applied to reduce and increase the channel dimension, which is followed by batch normalization and a ReLU activation function to obtain the global channel attention features. The formulation is as follows:
G ( X ) = B ( W 2 ( δ ( B ( W 1 ( g ( X ) ) ) ) ) )
where g ( ) represents global average pooling, B ( ) denotes batch normalization, δ ( ) is the ReLU activation function, and W 1 and W 2 are point convolutions with kernel sizes 1 × 1 × C × C r and 1 × 1 × C r × C , respectively.
In the local branch, global average pooling is omitted, and local feature information is directly extracted using a sequence of point convolutions while preserving the original spatial resolution. The formulation is as follows:
L ( X ) = B ( W 4 ( δ ( B ( W 3 ( X ) ) ) ) )
where W 3 and W 4 are point convolutions with kernel sizes 1 × 1 × C × C r and 1 × 1 × C r × C , respectively.
Additionally, a hybrid strip pooling branch is introduced, combining max and average pooling operations to capture strip-like feature patterns. This branch also undergoes a sequence of point convolutions, batch normalization, and ReLU transformations, as expressed in the following formula:
H ( X ) = B ( W 6 ( δ ( B ( W 5 ( h ( X ) ) ) ) ) )
where h ( ) denotes hybrid strip pooling, and W 5 and W 6 are point convolutions with kernel sizes 1 × 1 × C × C r and 1 × 1 × C r × C , respectively.
The outputs of the three branches are processed through a softmax operation to generate normalized weights ω , ensuring that the sum of the weights equals 1. Finally, the features from the three branches are fused by weighted summation:
ω = s o f t m a x ( G ( X ) + L ( X ) + H ( X ) )
o u t p u t = ω 1 X + ω 2 Y + ω 3 Z
where ω 1 , ω 2 , and ω 3 represent the weights generated by the softmax operation, and denotes element-wise multiplication. This adaptive weighting mechanism dynamically adjusts the contribution of each branch based on the importance of the input features, effectively integrating multi-scale feature information.

2.4. Global Context Attention Module (GCAM)

A significant characteristic of cloud and cloud shadow segmentation is the notable spatial dependency between clouds and cloud shadows, where the morphological features and spatial distribution of cloud shadows directly correspond to the cloud body. To effectively model this spatial dependency, we introduce the Global Context Attention Module (GCAM) into each branch of the Atrous Spatial Pyramid Pooling (ASPP) module. This module establishes long-range dependencies to model global contextual information, enabling the model to not only extract local features but also to effectively integrate the global semantic details of the image, thereby enhancing the accuracy of cloud shadow segmentation. The module is simplified from the non-local module [31], and its specific structure is shown in Figure 4.
Studies [32] have found that for different query positions in the image, the global context modeled by the non-local module is almost identical. Therefore, to simplify computation, the global context modeling of one position can be used to replace other positions with the linear transformation matrix W v placed outside the multiplication operation. The simplified process is described by Equation (12):
z i = x i S i g m o i d ( W v S o f t m a x ( W k x i ) x i )
where x R C × H × W is the input feature, z R C × H × W is the output feature, W k is the global information modeling matrix, and W v represents the linear transformation matrix.
In this module, the input features are first linearly mapped to obtain W k , and then a weight matrix is generated by softmax normalization. The matrix W v is weighted, and different weighting coefficients are assigned to the input features so that the model can focus on the important feature regions. Then, a gating value is generated through the sigmoid function to select important features for enhancement. Finally, the activated results are multiplied by the original feature map, allowing the model to dynamically adjust the feature map based on the global context.

2.5. Loss Function

In cloud and cloud shadow segmentation tasks, the cloud shadow regions are typically small and difficult to segment accurately, while the background region occupies most of the pixels, causing the model to be biased toward the background class. To address the class imbalance issue and the challenge of segmenting small targets in cloud and cloud shadow segmentation, this paper adopts a weighted loss function that combines focal loss [33] and Dice loss [34].
Focal loss is an improved version of cross-entropy loss, which assigns higher weights to hard-to-classify samples, thus allowing the model to focus more on the difficult-to-classify cloud shadow regions. It is suitable for handling class imbalance tasks. The formula is as follows:
L f o c a l ( p t n ) = n = 1 N ( 1 p t n ) γ log ( p t n )
where p t n represents the predicted probability of the model for the true class of the n-th sample. γ is a modulation factor that controls the influence of easily classified samples and increases the weight of hard-to-classify samples, and N is the total number of samples.
Dice loss is a commonly used metric for measuring the overlap between the predicted region and the true region. It is defined as the size of the intersection between the predicted segmentation mask and the true segmentation mask divided by the size of their union. Since cloud shadow regions are relatively small and have complex shapes, Dice loss can effectively enhance the model’s attention to the cloud shadow regions, ensuring more accurate segmentation results. The formula is as follows:
L d i c e = 2 Y T Y + T
where Y is the predicted segmentation mask and T is the true segmentation mask.
This paper combines these two loss functions with weighted coefficients, which helps address the class imbalance between cloud shadow regions and background regions. It also optimizes the overlap of the segmentation results, enhancing the model’s learning of cloud shadow boundaries and details. The form of the hybrid loss function is as follows:
L = α L f o c a l + β L d i c e
where α and β are hyperparameters. In the experiments, multiple combinations of hyperparameters were tested, as shown in Table 1. It is evident that when α = 0.6 and β = 0.4 , the model achieves the highest segmentation accuracy. Therefore, this combination of hyperparameters was chosen for the subsequent experiments.

3. Experimental Analysis

3.1. Experimental Datasets

The publicly available dataset used in this study is the GF-1 WHU cloud and cloud shadow detection dataset, which is derived from images captured by the Chinese Gaofen-1 (GF-1) satellite. It was created by researchers from Wuhan University [35]. The images obtained by the GF-1 satellite’s Wide Field of View (WFV) sensor have a spatial resolution of 16 m and span four multispectral bands across the visible to near-infrared spectral range. The dataset includes 108 Level 2A scenes collected under various cloud conditions, covering multiple land cover types such as forests, barren land, water bodies, wetlands, and urban areas. The annotations in the dataset were manually completed by experts with cloud, cloud shadow, clear sky, and background pixels labeled as 255, 128, 1, and 0, respectively. Some sample images and labels from the dataset are shown in Figure 5.
The original images were divided into smaller 256 × 256 pixel sub-images with black borders and unclear images removed to avoid negative impacts on model performance. The dataset was then split into training and validation sets in an 8:2 ratio, providing adequate sufficient data for training and evaluation. To increase the model’s diversity and robustness, data augmentation techniques, such as horizontal/vertical flipping and random rotations, were applied to generate additional variations of the images. Ultimately, 5428 images were used for training, while 1360 images were used for validation and testing.
To evaluate the generalization capability of the model, we also conducted experiments on the HRC_WHU and Cloud and Cloud Shadow datasets. The HRC_WHU dataset, also provided by the laboratory of Wuhan University [36], contains 150 high-resolution images from various regions around the world. This dataset only labels cloud and non-cloud pixels and includes images with five different background types (water, vegetation, urban, snow/ice, and barren land). Due to GPU memory limitations, we cropped the images into 3200 small 256 × 256 patches and divided them into training and testing sets with an 8:2 ratio, resulting in 2560 images for training and 640 images for testing. The Cloud and Cloud Shadow dataset, derived from high-resolution Google Earth images, covers complex backgrounds like plains, farmlands, towns, and water bodies. For training convenience, we cropped the high-resolution images into 224 × 224 patches and applied data augmentation techniques such as rotation and flipping. A total of 2630 images were obtained, which were then split into training and validation sets with an 8:2 ratio. This dataset labels clouds, cloud shadows, and background information.

3.2. Experimental Details

To ensure a fair comparison, all models were tested under the same experimental conditions and with identical hyperparameter settings. The experiments were conducted on an NVIDIA RTX 3090 GPU, manufactured by NVIDIA Corporation, headquartered in Santa Clara, California, USA. The framework used was PyTorch 1.7.0 with CUDA version 11.0. Experimental results show that most models converge within 200–250 epochs, so we set the number of training epochs to 300. Considering the GPU memory limitations, the batch size was set to 16. Due to the good convergence and stability of the Adam optimizer, we selected Adam for the experiment. Except for our model, which uses the fusion loss function proposed in this paper, all other comparative models used the commonly applied cross-entropy loss function. The initial learning rate (LR) was set to 0.001, and a cosine annealing learning rate decay strategy was employed to ensure gradual convergence during training. During training, the learning rate was adjusted according to the following formula:
L R = 0.001 × ( 1 e p o c h 300 ) 2
To evaluate the algorithm’s performance in cloud and cloud shadow segmentation tasks, we selected several quantitative metrics: precision (P), recall (R), F1 score, pixel accuracy (PA), average pixel accuracy (MPA), mean intersection over union (MIoU), and frequency-weighted intersection over union (FWIoU). The formulas for each of these metrics are as follows:
P = T P T P + F P
R = T P T P + F N
F 1 = 2 × P × R P + R
P A = i = 0 k p i , i i = 0 k j = 0 k p i , j
M P A = 1 k i = 0 k p i , i j = 0 k p i , j
M I o U = 1 k + 1 i = 0 k p i , i j = 0 k p i , j + j = 0 k p j , i p i , i
F W I o U = 1 i = 0 k j = 0 k p i , j i = 0 k j = 0 k p i j p i i j = 0 k p i j + j = 0 k p j i p i i
where T P represents the number of true positive samples, F P represents the number of false positive samples, and F N represents the number of false negative positive samples. p i , i denotes the number of correctly classified pixels of the i-th class, p i , j denotes the number of pixels of the i-th class incorrectly predicted as the j-th class, and p j , i denotes the number of pixels of the j-th class incorrectly predicted as the i-th class. k represents the number of classes (including background).
In the experiments on the GF-1 WHU dataset and the Cloud and Cloud Shadow Dataset, we consider three classes (cloud, cloud shadow, and background) for metric computation. For the HRC_WHU dataset experiments, we use two classes (cloud and background) for evaluation. In addition, we also calculated the number of floating-point operations (FLOPs) and the time taken to train a single image (Time).

3.3. Ablation Experiment

To evaluate the effectiveness of the proposed modules in improving segmentation results, we conducted ablation experiments. The original DeepLab v3+ was used as the baseline model, and we progressively introduced the HSPM, TB-AFFM, and GCAM modules to verify the effectiveness of each module. Finally, we incorporated our proposed innovative loss function to validate the feasibility of the entire model. The metrics used in the ablation experiments were MPA and MIoU with the specific results shown in Table 2. It can be observed that the model that includes all the modules and uses the innovative loss function achieved the best results, improving the MPA and MIoU metrics by 1.47% and 1.88%, respectively, compared to the baseline model.
  • Ablation for HSPM: Global pooling leads to a significant loss of detailed information, while replacing global pooling with HSPM helps retain both global information and important local details. This enables the model to better recognize the complex shapes and boundaries of clouds and cloud shadows. The experiment shows that the HSPM improved the MPA and MIoU by 84.87% and 75.79%, respectively, validating the module’s effectiveness in improving model accuracy.
  • Ablation for TB-AFFM: The proper use of multi-scale features is a crucial way to enhance model performance. In the decoding stage, the TB-AFFM module adaptively fuses mid-level and high-level features from the backbone network with deep high-level features extracted by ASPP. This helps the model automatically learn the importance of different feature levels and better recover and strengthen the details of clouds and cloud shadows during decoding. The experiment shows that TB-AFFM improved the model’s MPA and MIoU by 0.45% and 0.62%, respectively.
  • Ablation for GCAM: ASPP, using different dilated convolutions, can only capture local information at different scales, but it fails to capture global contextual relationships effectively. Considering that the morphology of clouds and the distribution of cloud shadows are correlated, we introduced GCAM into each branch of ASPP to enhance the model’s ability to perceive long-range pixel relationships and improve its resistance to complex background interference. The experiment shows that GCAM increased the MPA from 85.32% to 85.75% and the MIoU from 76.41% to 76.93%.

3.4. Comparison Experiments

To demonstrate the superiority of the proposed model in cloud and cloud shadow segmentation, we conducted comparative experiments with several outstanding semantic segmentation models. In addition to the models mentioned in the introduction, the comparative experiments also included CSDNet, based on CNN architecture, SegFormer, based on Transformer architecture, and DASUNet and MAINet, which are both based on hybrid architectures. The experimental results are shown in Table 3. From the perspective of mIoU, FCN-8s exhibited the worst segmentation performance, followed by SegNet, PSPNet, HRNet, UNet, DeepLab v3+, SP_CSANet, CSDNet, SegFormer, DASUNet, and MAINet, with the latter four models achieving better segmentation results. Our model achieved PA, MPA, MIoU, and FWIoU scores of 90.57%, 86.08%, 77.28%, and 83.05%, respectively, outperforming all other models.
To further analyze the advantages of our model, we computed the segmentation metrics for cloud and cloud shadow. The results are shown in Table 4. For cloud detection, our model achieved precision, recall, and F1 score values of 93.78%, 92.36%, and 92.63%, respectively, with precision and F1 score outperforming other models, while recall was slightly lower than DASUNet. For cloud shadow detection, our model achieved precision, recall, and F1 score values of 77.13%, 73.64%, and 75.34%, respectively, outperforming all other models in every metric. In conclusion, our model significantly improves both cloud and cloud shadow detection accuracy.
Figure 6 visually compares the segmentation results of different models for cloud and cloud shadow detection. In the figure, black represents the background, white represents the cloud, and gray represents the cloud shadow. It can be seen that the segmentation results of SegNet are relatively coarse especially in terms of cloud shadow extraction. Due to its simple network structure, SegNet struggles to capture detailed features, leading to inaccuracies in edge detection and frequent false negatives and false positives, particularly for small cloud and cloud shadow areas, as seen in the first and third sets of images. HRNet produces better segmentation results with smoother boundaries and more precise edge detection, though it still exhibits noticeable false negatives and false positives (third and fourth sets of images). From the first and fifth sets of images, it is evident that the encoder–decoder architecture of UNet helps preserve structural details in the image, enabling more accurate cloud shadow boundary detection. However, it may still result in false positives when handling complex backgrounds or overlapping cloud shadows, as seen in the fourth set of images. DeepLab v3+, by incorporating dilated convolutions, enhances the model’s ability to capture features at different scales, yielding fine segmentation results and strong edge extraction. However, in cases like the fourth set of images, where the cloud shadow boundaries are unclear and there is significant background interference, it may mistakenly classify some non-cloud shadow pixels as cloud shadows. Our proposed method, in the feature extraction phase, first uses HSPM to enhance the extraction of fine details. GCAM is then introduced to help the model better understand the relationships between clouds and the background, as well as between clouds and cloud shadows, reducing false positives and false negatives. Finally, in the decoding stage, TB-AFFM adaptively fuses mid-level and high-level features from the backbone network with deep features from the ASPP, further improving segmentation accuracy for detailed and complex regions. Particularly, as seen in the second and third sets of images, the boundaries of the segmentation results from our model are the most accurate.
To better demonstrate the generalization capability of our model in different scenarios, we compared the segmentation results of various models on seven different scenes: grassland, barren land, water bodies, desert, wasteland, urban areas, and snow/ice. As shown in Figure 7, in the first set of images, SegNet, HRNet, and UNet fail to detect small gaps in the clouds, and their edge extraction is not sufficiently accurate. Although DeepLab v3+ successfully detects the small gaps in the clouds, its segmentation of shapes is not accurate enough, and its detailed handling needs improvement. Our model, however, accurately identifies the gaps in the cloud shapes, with richer details. In the second set of images, HRNet’s edge extraction for irregular cloud boundaries is coarse, while other models perform better in edge segmentation, but their detailed rendering is still inferior to our model. In the third set of images, where the boundaries are unclear and difficult to segment, our model can still precisely locate the cloud boundaries, whereas other models struggle to extract accurate boundaries in this scenario. In the fourth set, the dark background increases the difficulty of cloud shadow detection. Except for our model, other models fail to detect the small cloud shadow area at the bottom of the image. In the fifth set of images, the dark regions in the background are more likely to be incorrectly detected as cloud shadows. The results show that UNet and DeepLab v3+ almost identify the entire dark area as cloud shadow, while SegNet identifies part of the dark area as cloud shadow. HRNet, although successfully distinguishing the dark background from the cloud shadow, does not segment the shapes of other clouds and cloud shadows as accurately as our model. In the sixth set of images, influenced by the urban background, SegNet, HRNet, and UNet show a relatively coarse segmentation of cloud and cloud shadow boundaries, whereas our model demonstrates the highest accuracy in boundary segmentation. In the seventh set of images, SegNet almost fails to distinguish between the snow on the mountaintop and the clouds, and other models also mistakenly classify some mountaintop snow as clouds. In contrast, our model exhibits the lowest false positive rate, demonstrating its superior performance in complex backgrounds.
Figure 8 demonstrates the segmentation results of clouds with varying thickness and shapes. The central portion of thick clouds, due to their lower translucency, appears as bright white regions, which are typically well recognized by the model. However, the edges of thick clouds, where light reflection is weaker and if the background is relatively dark, there is a low contrast between the cloud edges and the background, making it difficult for the model to distinguish the cloud boundaries, leading to false negatives or false positives. Our model can more accurately identify thick clouds and extract clearer boundaries. Thin or light clouds typically have blurred edges and are more susceptible to background interference, often resulting in discontinuous segmentation and false positives. For these types of clouds, our model also exhibits better segmentation performance. In conclusion, our model improves segmentation accuracy for clouds of different scales and shapes by aggregating global contextual information and adaptively assigning attention.

3.5. Generalization Performance Analysis

To further evaluate the generalization ability of our model on different datasets, we conducted comparative experiments on the HRC_WHU dataset (Dataset 1) and the Cloud and Cloud Shadow dataset (Dataset 2). The experiments were evaluated using three metrics: PA, MPA, and MIoU. The experimental results are shown in Table 5.
The experimental results indicate that on Dataset 1, our model demonstrates outstanding performance, leading all comparison models in the PA (94.44%) and MIoU (87.55%) metrics. In the MPA metric, our model ranks second only to DASUNet. On Dataset 2, our model also shows the best overall performance, with the PA metric reaching 97.62%, the MPA metric at 96.56%, and the MIoU metric at 93.02%. Overall, although other models such as CSDNet and DASUNet achieve competitive results in certain metrics, our model outperforms them across both datasets in terms of overall performance. This strongly demonstrates the efficiency and strong generalization capability of our model for cloud and cloud shadow segmentation tasks.

4. Discussion

In this study, we propose an improved model based on DeepLab v3+ aimed at enhancing the segmentation accuracy of clouds and cloud shadows. Our improvement strategy includes the Global Context Attention Module (GCAM), Hybrid Strip Pooling Module (HSPM), and Three-Branch Adaptive Feature Fusion Module (TB-AFFM). Experimental results show that our improved model achieved PA, MPA, MIoU, and FWIoU scores of 90.57%, 86.08%, 77.28%, and 83.05%, respectively, on the GF1_WHU dataset. Our model also outperforms other comparative models in segmentation accuracy on other datasets.
Although our research is primarily based on the DeepLab v3+ architecture, we believe that these methods are equally applicable to other image segmentation models. For instance, multi-scale feature fusion techniques can enhance the integration of multi-scale features, thereby improving the handling of images with varying resolutions and complex backgrounds. Similar approaches have been widely applied in other segmentation networks with good results. For example, Surya et al. [41] improved the fusion of multi-scale features in the U-Net architecture, successfully enhancing cloud shadow segmentation performance. Qu et al. [19] combined strip pooling with residual networks, significantly improving edge segmentation accuracy.
Beyond cloud and cloud shadow segmentation, our strategies may also yield similar effects in other task models, particularly in areas such as change detection [42,43,44]. For example, Jiang et al. [45] improved urban change detection by utilizing cross-scale fusion. Ren et al. [46] integrated self-attention and cross-attention mechanisms to enhance the model’s focus on real change regions.
In summary, we believe that our improvement strategies not only enhance the performance of DeepLab v3+ but also have potential applications in other image segmentation models, providing support for a broader range of remote sensing image tasks.

5. Conclusions

The accurate segmentation of clouds and cloud shadows is crucial in meteorological monitoring and remote sensing image analysis. This paper proposes an improved model based on DeepLab v3+, significantly enhancing cloud and cloud shadow segmentation accuracy. The model achieves this by adding a Global Context Attention Module to the ASPP branches, replacing global pooling with hybrid strip pooling, and incorporating an adaptive weighting mechanism to fuse mid- and high-level features from the backbone network with deep features from the ASPP during decoding. Additionally, we adopt an innovative loss function in the experiments. Specifically, the HSPM reduces spatial information loss and enhances the ability to extract detailed features. The GCAM models global contextual information and enhances the model’s perception of the global background through attention allocation, effectively reducing background noise interference. The TB-AFFM uses three parallel branches to adaptively generate weights, effectively integrating multi-scale feature information. Experimental results show that our model outperforms traditional segmentation models and some recent segmentation models across various metrics. In terms of segmentation performance, our model effectively handles different types of clouds and demonstrates strong generalization ability in complex scenarios, particularly excelling in edge extraction and preserving detailed features. However, there are still some limitations: (1) The model still experiences some false negatives and false positives in complex cloud shadow scenes, such as images with strong noise interference. (2) The segmentation performance for large areas of thin clouds needs further improvement. (3) The introduction of complex modules increases the number of parameters, leading to slower inference speed and higher computational overhead. In the future, we will address these limitations through further research to provide more accurate and real-time cloud and cloud shadow segmentation models for related fields.

Author Contributions

Methodology, Y.F. and Z.F.; Software, Y.F., Z.F. and Y.Y.; Validation, Y.F., Y.Y., Z.J. and S.Z.; Formal analysis, Y.F., Z.F. and Y.Y.; Investigation, Y.F. and Z.F.; Data curation, Y.F., Y.Y. and S.Z.; Writing—original draft, Y.F. and Z.F.; Writing—review & editing, Y.F. and Z.F.; Visualization, Y.F. and Z.J. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in this study are included in the article. Further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Rossow, W.B.; Schiffer, R.A. Advances in Understanding Clouds from ISCCP. Bull. Am. Meteorol. Soc. 1999, 80, 2261–2288. [Google Scholar]
  2. Sun, L.; Wei, J.; Wang, J.; Mi, X.; Guo, Y.; Lv, Y.; Yang, Y.; Gan, P.; Zhou, X.; Jia, C.; et al. A Universal Dynamic Threshold Cloud Detection Algorithm (UDTCDA) Supported by a Prior Surface Reflectance Database. J. Geophys. Res. Atmos. 2016, 121, 7172–7196. [Google Scholar] [CrossRef]
  3. Vásquez, R.E.; Manian, V.B. Texture-Based Cloud Detection in MODIS Images. In Proceedings of the SPIE Remote Sensing; SPIE: Bellingham, WA, USA, 2003. [Google Scholar]
  4. Zhu, Z.; Woodcock, C.E. Object-Based Cloud and Cloud Shadow Detection in Landsat Imagery. Remote Sens. Environ. 2012, 118, 83–94. [Google Scholar]
  5. Danda, S.; Challa, A.; Sagar, B.S.D. A Morphology-Based Approach for Cloud Detection. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 80–83. [Google Scholar]
  6. Le Hégarat-Mascle, S.; André, C. Use of Markov Random Fields for Automatic Cloud/Shadow Detection on High Resolution Optical Images. ISPRS J. Photogramm. Remote Sens. 2009, 64, 351–366. [Google Scholar]
  7. Cheng, H.-Y.; Lin, C.-L. Cloud Detection in All-Sky Images via Multi-Scale Neighborhood Features and Multiple Supervised Learning Techniques. Atmos. Meas. Tech. 2017, 10, 199–208. [Google Scholar] [CrossRef]
  8. Wei, J.; Huang, W.; Li, Z.; Sun, L.; Zhu, X.; Yuan, Q.; Liu, L.; Cribb, M. Cloud Detection for Landsat Imagery by Combining the Random Forest and Superpixels Extracted via Energy-Driven Sampling Segmentation Approaches. Remote Sens. Environ. 2020, 248, 112005. [Google Scholar]
  9. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  10. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015; Navab, N., Hornegger, J., Wells, W.M., Frangi, A.F., Eds.; Springer International Publishing: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  11. Miller, J.; Nair, U.; Ramachandran, R.; Maskey, M. Detection of Transverse Cirrus Bands in Satellite Imagery Using Deep Learning. Comput. Geosci. 2018, 118, 79–85. [Google Scholar]
  12. Mohajerani, S.; Krammer, T.A.; Saeedi, P. A Cloud Detection Algorithm for Remote Sensing Images Using Fully Convolutional Neural Networks. In Proceedings of the 2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP), Vancouver, BC, Canada, 29–31 August 2018; pp. 1–5. [Google Scholar]
  13. Gonzales, C.; Sakla, W.A. Semantic Segmentation of Clouds in Satellite Imagery Using Deep Pre-Trained U-Nets. In Proceedings of the 2019 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), Washington, DC, USA, 15–17 October 2019; pp. 1–7. [Google Scholar]
  14. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar]
  15. Lu, J.; Wang, Y.; Zhu, Y.; Ji, X.; Xing, T.; Li, W.; Zomaya, A.Y. P_Segnet and NP_Segnet: New Neural Network Architectures for Cloud Recognition of Remote Sensing Images. IEEE Access 2019, 7, 87323–87333. [Google Scholar]
  16. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6230–6239. [Google Scholar]
  17. Chen, L.-C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. In Proceedings of the European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
  18. Sun, K.; Zhao, Y.; Jiang, B.; Cheng, T.; Xiao, B.; Liu, D.; Mu, Y.; Wang, X.; Liu, W.; Wang, J. High-Resolution Representations for Labeling Pixels and Regions. arXiv 2019, arXiv:1904.04514. [Google Scholar]
  19. Qu, Y.; Xia, M.; Zhang, Y. Strip Pooling Channel Spatial Attention Network for the Segmentation of Cloud and Cloud Shadow. Comput. Geosci. 2021, 157, 104940. [Google Scholar] [CrossRef]
  20. Lu, C.; Xia, M.; Qian, M.; Chen, B. Dual-Branch Network for Cloud and Cloud Shadow Segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5410012. [Google Scholar] [CrossRef]
  21. Dai, X.; Chen, K.; Xia, M.; Weng, L.; Lin, H. LPMSNet: Location Pooling Multi-Scale Network for Cloud and Cloud Shadow Segmentation. Remote Sens. 2023, 15, 4005. [Google Scholar] [CrossRef]
  22. Hu, Z.; Weng, L.; Xia, M.; Hu, K.; Lin, H. HyCloudX: A Multibranch Hybrid Segmentation Network With Band Fusion for Cloud/Shadow. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 6762–6778. [Google Scholar] [CrossRef]
  23. Kang, J.; Liu, L.; Zhang, F.; Shen, C.; Wang, N.; Shao, L. Semantic Segmentation Model of Cotton Roots In-Situ Image Based on Attention Mechanism. Comput. Electron. Agric. 2021, 189, 106370. [Google Scholar] [CrossRef]
  24. Li, D.; Yang, Y.; Zhao, S.; Ding, J. Segmentation of Underwater Fish in Complex Aquaculture Environments Using Enhanced Soft Attention Mechanism. Environ. Model. Softw. 2024, 181, 106170. [Google Scholar] [CrossRef]
  25. Jiang, J.; Feng, X.; Ye, Q.; Hu, Z.; Gu, Z.; Huang, H. Semantic Segmentation of Remote Sensing Images Combined with Attention Mechanism and Feature Enhancement U-Net. Int. J. Remote Sens. 2023, 44, 6219–6232. [Google Scholar] [CrossRef]
  26. Ding, Z.; Wang, T.; Sun, Q.; Wang, H. Adaptive Fusion with Multi-Scale Features for Interactive Image Segmentation. Appl. Intell. 2021, 51, 5610–5621. [Google Scholar] [CrossRef]
  27. Wei, D.; Wang, H. MFFLNet: Lightweight Semantic Segmentation Network Based on Multi-Scale Feature Fusion. Multim. Tools Appl. 2023, 83, 30073–30093. [Google Scholar]
  28. Li, Y.; Huang, M.; Zhang, Y.; Bai, Z. Attention Guided Multi Scale Feature Fusion Network for Automatic Prostate Segmentation. Comput. Mater. Contin. 2024, 78, 1649–1668. [Google Scholar] [CrossRef]
  29. Ji, H.; Xia, M.; Zhang, D.; Lin, H. Multi-Supervised Feature Fusion Attention Network for Clouds and Shadows Detection. ISPRS Int. J. Geo-Inf. 2023, 12, 247. [Google Scholar] [CrossRef]
  30. Hou, Q.; Zhang, L.; Cheng, M.-M.; Feng, J. Strip Pooling: Rethinking Spatial Pooling for Scene Parsing. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 4002–4011. [Google Scholar]
  31. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-Local Neural Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7794–7803. [Google Scholar]
  32. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. GCNet: Non-Local Networks Meet Squeeze-Excitation Networks and Beyond. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 1971–1980. [Google Scholar]
  33. Lin, T.-Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2999–3007. [Google Scholar]
  34. Milletari, F.; Navab, N.; Ahmadi, S.-A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–26 October 2016; IEEE Computer Society: Los Alamitos, CA, USA, 2016; pp. 565–571. [Google Scholar]
  35. Li, Z.; Shen, H.; Li, H.; Xia, G.; Gamba, P.; Zhang, L. Multi-Feature Combined Cloud and Cloud Shadow Detection in GaoFen-1 Wide Field of View Imagery. Remote Sens. Environ. 2017, 191, 342–358. [Google Scholar] [CrossRef]
  36. Li, Z.; Shen, H.; Cheng, Q.; Liu, Y.; You, S.; He, Z. Deep Learning Based Cloud Detection for Medium and High Resolution Remote Sensing Images of Different Sensors. ISPRS J. Photogramm. Remote Sens. 2019, 150, 197–212. [Google Scholar] [CrossRef]
  37. Zhang, G.; Gao, X.; Yang, Y.; Wang, M.; Ran, S. Controllably Deep Supervision and Multi-Scale Feature Fusion Network for Cloud and Snow Detection Based on Medium- and High-Resolution Imagery Dataset. Remote Sens. 2021, 13, 4805. [Google Scholar] [CrossRef]
  38. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  39. Wei, F.; Wang, S.; Sun, Y.; Yin, B. A Dual Attentional Skip Connection Based Swin-UNet for Real-Time Cloud Segmentation. IET Image Process. 2024, 18, 3460–3479. [Google Scholar]
  40. Ding, L.; Xia, M.; Lin, H.; Hu, K. Multi-Level Attention Interactive Network for Cloud and Snow Detection Segmentation. Remote Sens. 2024, 16, 112. [Google Scholar] [CrossRef]
  41. Surya, S.R.; Abdul Rahiman, M. CSDUNet: Automatic Cloud and Shadow Detection from Satellite Images Based on Modified U-Net. J. Indian Soc. Remote Sens. 2024, 52, 1699–1715. [Google Scholar] [CrossRef]
  42. Zhan, Z.; Ren, H.; Xia, M.; Lin, H.; Wang, X.; Li, X. AMFNet: Attention-Guided Multi-Scale Fusion Network for Bi-Temporal Change Detection in Remote Sensing Images. Remote Sens. 2024, 16, 1765. [Google Scholar] [CrossRef]
  43. Wang, Z.; Gu, G.; Xia, M.; Weng, L.; Hu, K. Bitemporal Attention Sharing Network for Remote Sensing Image Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 10368–10379. [Google Scholar] [CrossRef]
  44. Zhu, T.; Zhao, Z.; Xia, M.; Huang, J.; Weng, L.; Hu, K.; Lin, H.; Zhao, W. FTA-Net: Frequency-Temporal-Aware Network for Remote Sensing Change Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2025, 18, 3448–3460. [Google Scholar] [CrossRef]
  45. Jiang, S.; Lin, H.; Ren, H.; Hu, Z.; Weng, L.; Xia, M. MDANet: A High-Resolution City Change Detection Network Based on Difference and Attention Mechanisms under Multi-Scale Feature Fusion. Remote Sens. 2024, 16, 1387. [Google Scholar] [CrossRef]
  46. Ren, H.; Xia, M.; Weng, L.; Lin, H.; Huang, J.; Hu, K. Interactive and Supervised Dual-Mode Attention Network for Remote Sensing Image Change Detection. IEEE Trans. Geosci. Remote Sens. 2025, 63, 5612818. [Google Scholar] [CrossRef]
Figure 1. Multi-scale feature adaptive fusion network structure based on DeepLab v3+.
Figure 1. Multi-scale feature adaptive fusion network structure based on DeepLab v3+.
Remotesensing 17 01229 g001
Figure 2. The structure of the Hybrid Strip Pooling Module. (The intensity of the color represents the magnitude of the feature values, with darker colors indicating larger values).
Figure 2. The structure of the Hybrid Strip Pooling Module. (The intensity of the color represents the magnitude of the feature values, with darker colors indicating larger values).
Remotesensing 17 01229 g002
Figure 3. The structure of the three-branch adaptive feature fusion module.
Figure 3. The structure of the three-branch adaptive feature fusion module.
Remotesensing 17 01229 g003
Figure 4. The structure of the Global Context Attention Module.
Figure 4. The structure of the Global Context Attention Module.
Remotesensing 17 01229 g004
Figure 5. Some data from the GF-1 WHU dataset. The first row shows the original images, and the second row corresponds to their labels: (a) forest; (b) Gobi desert; (c) barren land; (d) wetland; and (e) urban areas.
Figure 5. Some data from the GF-1 WHU dataset. The first row shows the original images, and the second row corresponds to their labels: (a) forest; (b) Gobi desert; (c) barren land; (d) wetland; and (e) urban areas.
Remotesensing 17 01229 g005
Figure 6. Comparison of segmentation results of clouds and cloud shadows by different methods: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes highlight the areas where the segmentation results differ between models).
Figure 6. Comparison of segmentation results of clouds and cloud shadows by different methods: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes highlight the areas where the segmentation results differ between models).
Remotesensing 17 01229 g006
Figure 7. Comparison of segmentation results of clouds and cloud shadows by different methods in different scenarios: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes and red circles highlight the areas where the segmentation results differ between models).
Figure 7. Comparison of segmentation results of clouds and cloud shadows by different methods in different scenarios: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes and red circles highlight the areas where the segmentation results differ between models).
Remotesensing 17 01229 g007
Figure 8. Comparison of segmentation results of different cloud types and cloud shadows by different methods: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes highlight the areas where the segmentation results differ between models).
Figure 8. Comparison of segmentation results of different cloud types and cloud shadows by different methods: (a) test image; (b) labels; (c) segmentation result of our method; (d) segmentation result of SegNet; (e) segmentation result of HRNet; (f) segmentation result of UNet; (g) segmentation result of DeepLab v3+. (The red boxes highlight the areas where the segmentation results differ between models).
Remotesensing 17 01229 g008
Table 1. The segmentation results of the model under different combinations of hyperparameters.
Table 1. The segmentation results of the model under different combinations of hyperparameters.
α β MPAMIoU
0.20.885.8077.16
0.30.785.9877.20
0.40.685.4876.84
0.50.586.0877.06
0.60.486.5077.43
0.70.385.5276.87
0.80.285.2776.80
Note: Bold indicates the best result, and the underline indicates the second-best result.
Table 2. The ablation experiment results of different module combinations.
Table 2. The ablation experiment results of different module combinations.
MethodMPA (%)MIoU (%)
DeepLab v3 plus84.6175.40
DeepLab v3 plus + HSPM84.8775.79 (0.39 ↑)
DeepLab v3 plus + HSPM + TB-AFFM85.3276.41 (0.62 ↑)
DeepLab v3 plus + HSPM + TB-AFFM + GCAM85.7576.93 (0.52 ↑)
DeepLab v3 plus + HSPM + TB-AFFM + GCAM + loss (Ours)86.0877.28 (0.35 ↑)
Note: Bold indicates the best result, and the arrow represents the improvement in model performance.
Table 3. Comparison of segmentation results of different models on the GF1_WHU dataset.
Table 3. Comparison of segmentation results of different models on the GF1_WHU dataset.
MethodPA (%)MPA (%)MIoU (%)FWIoU (%)FLOPs (G)Time (ms)
FCN-8s [9]86.1580.2069.0574.5215.073.73
SegNet [14]88.4882.8572.8379.7318.154.26
PSPNet [16]88.5382.6472.9879.6924.826.97
HRNet [18]89.1484.5074.7780.7234.6430.93
Unet [10]89.3285.0275.0881.0717.362.90
DeepLab v3+ [17]89.7284.6575.4481.6620.7813.53
SP_CSANet [19]89.6985.3575.7882.7225.0423.21
CSDNet [37]89.8385.5176.3482.5516.2118.48
SegFormer [38]90.2485.4576.4582.6320.3215.23
DASUNet [39]90.1185.8376.5782.8119.248.54
MAINet [40]90.3285.6876.7382.7225.3215.43
MFAFNet (Ours)90.5786.0877.2883.0523.0116.93
Note: Bold indicates the best result, and the underline indicates the second-best result.
Table 4. Comparison of segmentation results of different models in each category.
Table 4. Comparison of segmentation results of different models in each category.
MethodCloudCloud Shadow
P (%)R (%)F1 (%)P (%)R (%)F1 (%)
FCN-8s87.8486.3488.6368.5765.0167.78
SegNet90.3392.0191.1671.8966.2868.97
PSPNet89.7390.4190.0774.7866.0670.15
HRNet91.4990.7291.1074.8271.3773.05
UNet91.7291.2391.4773.6072.6873.14
DeepLab v3+92.2491.4691.8575.7470.3172.92
SP_CSANet92.4591.7392.1274.9471.4873.03
CSDNet92.7292.0292.2375.8373.0774.53
SegFormer92.8592.0792.4576.2572.6374.33
DASUNet93.2192.4192.3176.4773.1173.67
MAINet93.4992.2592.4276.3272.8374.24
MFAFNet (Ours)93.7892.3692.6377.1373.6475.34
Note: Bold indicates the best result, and the underline indicates the second-best result.
Table 5. Comparison of segmentation results of different models on other datasets.
Table 5. Comparison of segmentation results of different models on other datasets.
MethodDataset 1Dataset 2
PA (%)MPA (%)MIoU (%)PA (%)MPA (%)MIoU (%)
FCN-8s92.9391.8386.0894.5892.6387.34
SegNet93.0692.0286.1194.8692.7487.55
PSPNet93.3392.2586.5995.2393.6189.86
HRNet93.2192.0686.2796.7395.0791.35
UNet93.5692.8786.7897.0295.7792.10
DeepLab v3+93.3592.1486.5497.2695.8192.25
SP_CSANet93.7293.0686.8597.7296.3392.76
CSDNet94.3593.5087.2797.5896.2792.64
SegFormer93.8793.3586.9297.3296.2292.43
DASUNet94.3293.6387.3297.4596.4392.81
MAINet93.9593.3887.0197.5296.1892.68
MFAFNet (Ours)94.4493.5887.5597.6296.5693.02
Note: Bold indicates the best result, and the underline indicates the second-best result.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Feng, Y.; Fan, Z.; Yan, Y.; Jiang, Z.; Zhang, S. MFAFNet: Multi-Scale Feature Adaptive Fusion Network Based on DeepLab V3+ for Cloud and Cloud Shadow Segmentation. Remote Sens. 2025, 17, 1229. https://doi.org/10.3390/rs17071229

AMA Style

Feng Y, Fan Z, Yan Y, Jiang Z, Zhang S. MFAFNet: Multi-Scale Feature Adaptive Fusion Network Based on DeepLab V3+ for Cloud and Cloud Shadow Segmentation. Remote Sensing. 2025; 17(7):1229. https://doi.org/10.3390/rs17071229

Chicago/Turabian Style

Feng, Yijia, Zhiyong Fan, Ying Yan, Zhengdong Jiang, and Shuai Zhang. 2025. "MFAFNet: Multi-Scale Feature Adaptive Fusion Network Based on DeepLab V3+ for Cloud and Cloud Shadow Segmentation" Remote Sensing 17, no. 7: 1229. https://doi.org/10.3390/rs17071229

APA Style

Feng, Y., Fan, Z., Yan, Y., Jiang, Z., & Zhang, S. (2025). MFAFNet: Multi-Scale Feature Adaptive Fusion Network Based on DeepLab V3+ for Cloud and Cloud Shadow Segmentation. Remote Sensing, 17(7), 1229. https://doi.org/10.3390/rs17071229

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop