Next Article in Journal
Particle Scattering Induced Orbital Angular Momentum Spectrum Change of Vector Bessel–Gaussian Vortex Beam
Next Article in Special Issue
An Empirical Study on Retinex Methods for Low-Light Image Enhancement
Previous Article in Journal
On the Accuracy and Consistency of Quintuple Collocation Analysis of In Situ, Scatterometer, and NWP Winds
Previous Article in Special Issue
Enhanced Multi-Stream Remote Sensing Spatiotemporal Fusion Network Based on Transformer and Dilated Convolution
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Towards Robust Semantic Segmentation of Land Covers in Foggy Conditions

1
School of Instrument Science and Engineering, Southeast University, Nanjing 210096, China
2
Health Sciences Center, University of Oklahoma, Oklahoma City, OK 73106, USA
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(18), 4551; https://doi.org/10.3390/rs14184551
Submission received: 5 August 2022 / Revised: 6 September 2022 / Accepted: 8 September 2022 / Published: 12 September 2022

Abstract

:
When conducting land cover classification, it is inevitable to encounter foggy conditions, which degrades the performance by a large margin. Robustness may be reduced by a number of factors, such as aerial images of low quality and ineffective fusion of multimodal representations. Hence, it is crucial to establish a reliable framework that can robustly understand remote sensing image scenes. Based on multimodal fusion and attention mechanisms, we leverage HRNet to extract underlying features, followed by the Spectral and Spatial Representation Learning Module to extract spectral-spatial representations. A Multimodal Representation Fusion Module is proposed to bridge the gap between heterogeneous modalities which can be fused in a complementary manner. A comprehensive evaluation study of the fog-corrupted Potsdam and Vaihingen test sets demonstrates that the proposed method achieves a mean F 1 s c o r e exceeding 73%, indicating a promising performance compared to State-Of-The-Art methods in terms of robustness.

1. Introduction

Computer vision has emerged as a powerful and labor-saving tool for automatic scene parsing of remote sensing images (RSIs). Land cover classification (LCC) of aerial imagery, which is also known as semantic labeling or segmentation, assigns a class label to each pixel in RSIs. As an integral part of computer vision, semantic segmentation plays a pivotal role in remote sensing for rapid and accurate detection. A considerable amount of literature has been published concerning applications like landslide extraction [1], road extraction [2], collapsed building detection [3], and so on. Semantic segmentation models can be summarized into three categories [4], namely FCNets [5] which yield a coarse feature map directly from the low-resolution representation, UNets [6] which perform high-resolution recovery from the downsampled representation as well as HRNets [4] which retain the high-resolution representation during all procedures. Despite plenty of research exploring semantic segmentation based on natural scene images (NSIs), there is still a lack of scientific literature specifically focusing on robust remote sensing in foggy conditions.
RSIs with fog are distinguished from NSIs by a number of challenging characteristics that may result in decreased classification robustness. It is believed that there exist two major challenges when it comes to LCC with fog. First, model robustness is susceptible to fog corrupted RSIs, which refer to a series of issues [7], including intra-class heterogeneity, inter-class homogeneity, geometric size diversity, and so on. In terms of intra-class heterogeneity, the models tend to classify objects with distinctive appearances as disparate species; they may belong to the identical yet [8]. For instance, various materials and structures may lead to different appearances and textures. Whereas, fog-covered objects affiliated with diverse species frequently exhibit close characteristics when they are made up of the same material. The concrete building and the impervious surface in Figure 1 appear similar. Models assume they belong to the same class by mistake accidentally, which is inter-class homogeneity. Additionally, objects under fog exhibit geometric size diversity and only a robust model can capture multi-scale attributes in RSIs. Overall, RSIs with low quality can exert negative impacts on classification robustness. Second, when dealing with dense fog, a single sensor is not always effective. It is difficult for a single optical camera to classify objects robustly when they are partially obscured. Figure 1 illustrates the failing cases, such as the shadow of buildings, fog coverage, and cars parked under the trees. The optical input is rich in semantic details, while Digital Surface Model (DSM) provides discriminative height information. It is imperative to excavate informative cues from multimodal inputs.
To address the listed challenges of LCC in foggy conditions, we propose a framework with superior robustness based on attention mechanisms and multimodal fusion. We adopt HRNet as the backbone. Through compiling representations from all the high-to-low resolution streams in parallel, HRNet is robust to intra-class heterogeneity and geometric size diversity. The proposed Spectral and Spatial Representation Learning (SSRL) module probes into the relationship between spectral channels and spatial locations to improve robustness to intra-class heterogeneity. Thus, the output representation is gifted with semantic information and spatial accuracy. The introduced Multimodal Representation Fusion Module (MRFM) investigates the fusion of multimodal remote sensing data to learn the boundary connectivity and contour closure in RSIs to cope well with object occlusion and inter-class homogeneity issues. In summary, the main contributions are as follows:
  • Based on multimodal fusion and attention mechanisms, we propose a robust end-to-end model that can fuse the optical and DSM input for LCC.
  • Adopting HRNet as the backbone, we propose and incorporate SSRL and MRFM into the framework. To enhance the semantic information, a lightweight SSRL is inserted to capture the long-range dependencies and explore the interactions between various spectral channels. MRFM is employed for the effective fusion of multimodal remote sensing data. All the components cooperate and contribute to the classification robustness.
  • We conduct an ablation study to evaluate the effectiveness of the proposed framework, including functions of different modules and modal inputs.
  • We compare our model with SOTA methods to demonstrate the robustness against natural noise.

2. Related Work

There is a large volume of published studies describing how to conduct LCC. Most publications concentrate on accuracy instead of robustness, which is also vital in daily application. Thus far, several studies investigating robustness are predominantly associated with NSIs, such as scenes in ImageNet, Cityscape, BDD100k, and so on. Nonetheless, there is still a lack of relevant research focusing on LCC robustness. RSIs captured in foggy conditions are characterized by low quality, which poses challenges to robustness. This paper aims to design a robust model which can improve the classification performance in harsh environments. Our work refers to LCC, model robustness, and multimodal fusion. This section discusses the related work from these three perspectives.

2.1. Land Cover Classification

LCC actually refers to semantic segmentation of land covers using computer vision. There has been a great deal of research into semantic segmentation focusing on classification accuracy. Conventional segmentation algorithms are normally put forward on the premise of basic image attributes, e.g., grey-scale mutations are utilized to detect edges. Comparable grey scale values are partitioned into several regions according to the predefined criteria. However, it is extremely complex to detect boundaries when there exist substantial grey-scale changes. Considerable evidence has accumulated to show that deep learning-based models are more suitable for the semantic segmentation of NSIs. These models can usually be divided into three groups [4], namely, FCNet [5] type, UNet [6,9] type, and HRNet [4] type. FCNets learn the representation from high to low resolution in series to extract coarse feature maps. This group includes models like Deeplab [10], DenseASPP [11], PSPNet [12], and so on. UNets learn the encoded low-resolution representation and then recover to the high-resolution representation. Analogous models are DeepLabV3+ [13], SegNet [14], and so on. Moreover, ref. [9] inserts a cascaded dilated convolution in UNet to capture objects of diverse shapes, which is an effective approach to enhance robustness to multi-scale issues. Different from [9], we alleviate the influence of diverse shapes by adopting HRNet as the backbone because it retains a high-resolution representation throughout the process [15].

2.2. Model Robustness

Although neural networks are highly accurate for classification, they are not always as robust as human beings while actually applied [16]. Ref. [16] suggests that building multimodal and multitasking systems based on multi-sensor fusion is indispensable for robust decisions. Noises are categorized into three main groups [17], mainly adversarial noises, systematic noises, and natural noises. By comparing the robustness of three types of models, ref. [17] shows that CNN is more robust under natural noise and systematic noise, while Transformer is more robust against adversarial noise. Adversarial noise is the result of ambiguous decision-making at boundaries due to the limited training dataset and the inability to cover the whole sampling space. A small perturbation will always lead to completely distinct results. Ref. [18] constructed ImageNet-P and ImageNet-C datasets on top of ImageNet to facilitate researchers to evaluate and test the corruption and perturbation robustness. Based on this work, ref. [19] investigates the robustness of semantic segmentation. Researchers find that the Atrous Spatial Pyramid Pooling module significantly improves robustness, while the generalization performance depends heavily on the corruption degrees. Furthermore, there are some recent works on the adversarial noise robustness of Visual Transformers (ViTs). Ref. [20] found that shallow features in ViTs enable it to possess a better generalization than CNN, thus, better coping with adversarial noise. Meanwhile, the ensemble operation of CNN and ViTs can also improve the model robustness [21].

2.3. Attention Mechanism

A large and growing body of literature has investigated the role of attention mechanism [22] in deep visual models. Ref. [23] proposes an efficient channel module to explore the cross-channel interactions without dimension reduction. Ref. [24] designs a global context module to model the long-range dependencies with significantly less computation. There is a consensus among researchers that multi-head attention in ViTs[25] has acquired SOTA due to the uniform representation. Swin transformer [26] is capable of modeling input of different scales flexibly and the complexity is linear with input sizes. SETR [27] presents a multi-level feature fusion module to classify each pixel at a fine-grained level. Segformer [28] removes the complex position encoding binding a lightweight multi-layer perceptron to output feature maps of various sizes. Volo [29] introduces a novel outlook attention to grasp both coarse and fine-grained representations.

2.4. Multimodal Fusion

Data collected by a single sensor are often flawed and multi-sensor fusion is fundamental for accurate and robust decisions. Ref. [30] put forward a top-down pyramid fusion architecture for multimodal fusion. It is lightweight and can extract complementary features from multi-sources. Ref. [31] proposes a new and lightweight depth fusion transformer network for LCC, with different backbones extracting features from various inputs. Ref. [32] explores the pros and cons of early fusion along with late fusion to show that they all can utilize the complementarity of multimodal inputs. To calibrate features of the current modality from spatial and channel dimensions, ref. [33] has developed a Cross-Modal Feature Rectification Module for feature extraction. A multimodal fusion module was proposed in [15] to explore complementary features of heterogeneous inputs. In contrast, our method exploits the discriminate representation of each input from the perspective of the channel and spatial location before multimodal fusion, which greatly enhances the robustness of low-quality RSIs.
Overall, most studies remain narrow in focusing only on NSIs instead of RSIs. RSIs are characteristic of challenges of high resolution, multi-scales, class imbalance, occlusion, and so on. The attention mechanism is capable of capturing long-range dependencies. We extend ideas from deep learning and RGB-D semantic segmentation as well as an attention mechanism to establish a practical framework that can robustly classify land covers in foggy conditions.

3. Core of the Framework

3.1. Overview

The overall framework is illustrated in Figure 2. HRNetV2-W48 is adopted as the backbone for feature extraction. UperHead from [34] serves as the decoder. Each batch combines an Optical image ( X RGB R C × H × W ) with the corresponding DSM ( X DSM R 1 × H × W ), which contains the height information of land covers. H and W denote the height and width of RSIs, respectively. There is a considerable amount of noise in low-quality RSIs. We intend to design SSRL in such a manner that it would extract useful and discriminate representation efficiently without incurring excessive computational costs. Conventional methods simply aggregate two modalities without obtaining complementary features effectively. By contrast, MRFM, which is dedicated to multimodal fusion, exploits the complementarity between heterogeneous data. The backbone and decoder can be replaced by the other models. We will dig into the detailed design of HRNet, SSRL, and MRFM in Section 3.2, Section 3.3, and Section 3.4, respectively.

3.2. Backbone

Low-quality RSIs collected under fog suffer from the issues of intra-class heterogeneity and diverse geometric shapes, which impairs classification robustness. LCC is actually a dense pixel prediction task that requires a strong backbone with powerful modeling capability. Semantic segmentation networks are often constructed based on encoder-decoder architecture, where ResNet [35] is usually applied as the encoder. It is characterized by the accurate prediction of the spatial location at the low-level stage but is limited by a small receptive field that lacks consistent semantic information. This could lead to blurry classification. At the high-level stage, the network possesses a larger receptive field to make fine semantic predictions, but is deficient in the global representation. Consequently, conventional CNNs are susceptible to the above issues which are attributed to a loss of spatial details with the degradation of resolution.
HRNet is composed of four parallel branches pertaining to different resolutions. They constantly exchange information across multiple scales. High-resolution representations contain more spatial details, while the low-resolution representations are more capable of fine-grained classification. By maintaining a high-resolution representation, HRNet can generate spatially accurate feature maps that contain semantic information abundantly. Accordingly, HRNet-W48 is selected as our backbone for robust semantic segmentation. It can learn discriminative and distinct representations efficiently. Furthermore, HRNet is capable of integrating both local and global features with multiple scales, thereby increasing robustness in the presence of fog corruptions.
The hierarchical structure of HRNet is illustrated in Figure 3, which consists of 4 multi-resolution branches, each with resolutions of 1 / 4 , 1 / 8 , 1 / 16 , 1 / 32 . Each branch can be partitioned into 4 stages, and the output channel number of each branch is C , 2 C , 4 C , 8 C . Between each stage, there are blocks of multi-resolution fusion which consist of a 3 × 3 stride convolution integrating with a 1 × 1 upsampling layer, represented by the crossed lines. The fusion module serves as a mechanism for transferring feature information between branches of different resolutions. HRNet can be applied to semantic segmentation by accessing a 1 × 1 convolution for mixing and merging the representations from four branch outputs to align channel numbers. Detailed convolution parameters are shown in Table 1 where parameters for each stage are in the form [ a × a , n C ] × b × c . ‘[]’ represents the residual connection unit. Parameters a,b represent kernel size and duplication times of the residual unit separately. c means to repeat entire modularized part c times. Four basic blocks like 3 × 3 , n C 3 × 3 , n C accompanying with fusion modules constitute each branch of HRNet.

3.3. Spectral and Spatial Representation Learning

Generally, RSIs captured in foggy conditions are of low quality, with a significant amount of noise existing in heterogeneous inputs. Merely incorporating encoded representations into the decoder will bring excess redundancy to the network, which reduces the classification robustness. There have been some proposals to enhance semantic features through an attention mechanism. The non-local operation proposed by [22] can obtain the attention map corresponding to a specific query tensor for modeling the global context. After rigorous experiments in [24], researchers argue that the gap between attention activation maps corresponding to different query locations is narrow, illustrating the non-necessity of query weights. In addition, ViT [25] can also enhance the semantic information through multi-head attention. It is featured with numerous parameters, high complexity as well as overfitting. This computationally intensive approach, however, ignores the correlation between various spectral channels. Additionally, the channel attention mechanism [23,36] has also been proposed to explore the interaction between different channels. Nevertheless, simply regarding 2D images as 1D disrupts the dependencies between different positions, which reduces the robustness in capturing long-range relationships. Inspired by CBAM [37], we propose SSRL which is composed of Spectral Attention and Spatial Attention. Spectral attention is to explore interdependencies between different spectral channels, thereby improving semantic representations. Spatial attention is to capture long-range dependencies. Thus, SSRL can generate a global context and acquire correlations between various pixels and spectral channels to improve the robustness of intra-class heterogeneity.
Spectral Attention is illustrated in the upper half of Figure 2b. It is constructed for acquiring the spectral-level representation weight s p e ( s p e R C × 1 × 1 ) . SSRL firstly transforms the dimension of X R G B or X D S M ( X R G B R C × H × W or X D S M R C × H × W ) to 1 × 1 × C ( C is the channel number) using a global average pooling layer (GAP). The compressed tensor is fed into MLP ( W mlp ) to compute the interaction between k adjacent channels. The MLP consists of two one-dimensional convolution layers with kernel size 1 and k, denoted as C o n v 1 D _ 1 ( W 1 D _ 1 ) and C o n v 1 D _ k ( W 1 D _ k ) , respectively. C o n v 1 D _ 1 is applied for the dimension reduction, converting the channel number from C to C . The same learnable weight values are shared among channels, where the efficiency is significantly improved because only k values are noted. Detailed formulas about Spectral Attention are as the following:
s p e = σ W m l p 1 W H i = 1 , j = 1 W , H X i j
W m l p ( x ) = W 1 D _ k [ F M i s h ( W 1 D _ 1 ( x ) ) ]
F M i s h ( y ) = x · tanh ln 1 + e y
In Equation (1), σ is the sigmoid activation function. Local cross-channel attention interaction coverage is adaptively and dynamically adjusted according to the input channel numbers. There is a nonlinear mapping between the total number of channels C and the kernel size k of C o n v 1 D . k increases with the number of channels. odd means to select the nearest odd number. F s p e means the feature map acquired through the Spectral Attention. The specific correspondence is as follows:
k = ψ ( C ) = log 2 ( C ) 2 + 1 2 · 1 1 + e C + 1 2 o d d
F s p e = s p e X
Spatial Attention is illustrated in the lower panel of Figure 2b. This part is constructed for acquiring the spatial-level weight s p e ( s p e R H W × H W ). Through the linear transformation of s p a , we can get three weight matrixes W q , W k , W v . By calculating the interaction between positions (softmax operation, Equation (6)), the long range dependencies s p a can be captured. Finally, the feature map F s p a after Spectral Attention and Spatial Attention is obtained through matrix multiplication.
s p a = exp W q F s p e , W k F s p e m exp W q F s p e , W k F s p e
F s p a = W c [ s p a ( W v F s p e ) ]

3.4. Multimodal Representation Fusion Module

RSIs are collected by satellites or drones far from the ground, which will inevitably cause pixel loss owing to the atmosphere and clouds. Meanwhile, the terrestrial environment is a three-dimensional space and there exist complicated interactions between land covers. The quality of RSIs with fog is low as some of the land covers will be obscured by fog, resulting in a single optical sensor failing. This phenomenon reduced the classification robustness by a large margin. DSM built with lidar will not suffer from this. We fuse multiple inputs by designing an effective MRFM to explore the respective characteristics of each modality. Different from the early and late fusion strategy in [32], we exploit semantic representation of different modalities through the cross attention mechanism. This can improve robustness to inter-class homogeneity and object occlusion issues.
The structure of MRFM is illustrated in Figure 2c. Red and green dashed boxes in Figure 2c function as context modeling and transform respectively. We extract the coarse representation D S M from DSM since it contains the height information of each land cover. Firstly, utilizing 1 × 1 convolution ( W v ) with softmax in Equation (8) to obtain the global semantic key weight from DSM in the batch, this step is to obtain the coarse correlation feature maps between different locations. N p signifies the number of all pixels. j is for pixel indexing. Then, the computational cost is reduced by bottleneck. r is the reduction coefficient, which is set to 16 by default. Layer Normalization and GELU activation functions are integrated which enable a faster convergence as well as a stable training process. This step plays a role of transform in exploring channel-wise features while Channel Embedding ( W C E ) can enhance the nonlinearity as well as reduce the dimension. Finally, the linked adaptive average pooling layer ( AAP ) is utilized for the late fusion. γ , β in F L N are trainable vectors, which is for affine transformation and ϵ is for numerical stability. E and Var mean expectation and standard deviation separately. We extract the fine-grained feature map from RGB input as a result of the abundant semantic features in this modality. Like non-local operation [22], we obtain the RGB feature weight R G B through linear transformation and softmax. To exploit the complementary representation, we fuse both modalities using matrix multiplication accompanied by the residual connection. We likewise incorporate the residual structure into MRFM to make it easier for information to flow between layers, including providing feature reuse during forwarding propagation and mitigating the gradient vanishing phenomenon during backward propagation. Adding the original DSM input, dependencies between different positions obtained from DSM are fused with the exhaustive global information obtained from the optical so that each coarse position in DSM has an element-wise corresponding response generated.
α ( j ) = e W v X D j m = 1 N p e W v X D m
F G E L U ( x ) = 0.5 x 1 + tanh 2 / π x + 0.044715 x 3
F L N ( x ) = x E [ x ] Var [ x ] + ϵ γ + δ
D S M = A A P F L N { F G e l u [ W C E j = 1 N p α ( j ) X D j ] }
R G B = exp W q X RGB , W k X RGB m exp W q X RGB , W k X RGB
F M R F M = R G B D S M + X D S M

3.5. Loss Function

From the analysis in Section 4.1, we can observe that the class imbalance issue exists in the ISPRS dataset. The proportion of impervious surface is sixteen times higher than the vehicle which is the tail class. Many deep models are heavily biased towards the dominant class during the training process and fail to classify the tail classes instead. Drawing on an extensive range of sources, we propose a unified loss function for our framework following a series of studies like [38,39]. Three elements constitute the unified loss. It is generalized for RSIs with long-tail class imbalance. FC, FT, and CE in Equation (14) stand for the modified focal loss [40], focal Tversky loss [41], and cross-entropy loss [42], respectively. f refers to the final loss originating from [38], whose input is the prediction result of MRFM. aux is the auxiliary loss employed to supervise the coarse object area estimation of SSRL output. x stands for the input data and y gt is the ground truth. y pred is the prediction output values. back denotes the background class. gt , pred , and coar are ground truth, prediction, and coarse feature map, respectively.
L u n i f i e d = α L F C ˜ f + ( 1 α ) L F T ˜ f + λ x { r g b , d } L C E a u x y p r e d c o a r , y g t
L F C ˜ f = 1 N c + 1 [ j = 1 N c δ L C E + ( 1 δ ) ( 1 y p r e d b a c k ) γ 1 L C E ]
L F T ˜ f = 1 N c + 1 [ j = 1 N c ( 1 D S C ( x j ) ) 1 γ 2 + ( 1 D S C ( x b a c k ) ) ]
D S C ( x ) = T P + ε T P + δ F N + ( 1 δ ) F P + ε
L C E = j = 1 N c [ y g t l o g ( y p r e d ) + ( 1 y g t ) l o g ( 1 y p r e d ) ]
α (e.g., 0.5) is designed to balance the relative weights of the final loss while λ is the weight for auxiliary bootstrap loss. N C is the number of classes. δ is the threshold parameter (e.g., 0.7) related to the proportion of positive and negative samples. γ controls the degree of down-weighting of easy samples while enhancing the rare. γ 1 | γ 2 are 2 and 0.75 by default. ε is the small number for numerical stability in DSC , which acts similarly to the Tversky index to control the optimizing for output imbalance.

4. Experiment

4.1. Dataset Overview

ISPRS provided true orthophotos (TOP) of Potsdam and Vaihingen for model training and validation, whose resolutions are 6000 × 6000 and average 2500 × 2500 , respectively. Ground sampling distance (GSD) is 5 cm and 9 cm. We need to eliminate the Image Index 7–10 in Potsdam dataset due to the error in GT. Each TOP has six classes: car, low vegetation, tree, building, impervious surface, and cluttered background. We split both datasets into the training, test, and validation sets according to the ratios of 0.8, 0.1, and 0.1. DSM, which is acquired through lidars, contains the height information for land covers in 32-bit floating type. In addition to DSM, we use RGB and IRRG optical bands of Potsdam and Vaihingen for training and inference. nDSM is the normalized DSM which supplies the height of each pixel accompanied by ground elevation subtracted.
In order to intuitively observe the composition of different classes, we pull out all pixels from GT for an exploratory data analysis. Class proportions are shown in Figure 4. In both datasets, the ratios of each class are similar, with the proportion of cars being the smallest, less than 2 % . The volume of buildings along with impervious surfaces, which are difficult to distinguish, is between 26 % and 28 % . Potsdam has a proportion of low vegetation that is 8.92% higher than the number of trees, whereas the difference between the two classes in Vaihingen is only 2.35%.

4.2. Implementation Details

ISPRS datasets consist of high-resolution RSIs. Limited by GPU memory, we crop the optical image into 512 × 512 for model training and inference. In view of reducing overfitting, we augment the training set. Each pair of the image and the corresponding GT are rotated in arbitrary directions. Basic attributes such as contrast, brightness, saturation, and so on are randomly set for the augmentation. Reflection padding is conducted at the edges after cropping, this adjustment is particularly effective for urban complexes like buildings. Details are inevitably lost when cropping randomly, hence the symmetry of the buildings is well preserved by adopting reflection padding. Figure 5 illustrates samples for two datasets.
The hardware and software environment is listed in Table A1. We choose HRNet-W48 as the backbone, whose four branches yield feature maps ( R 1 , D 1 R 4 , D 4 ) with 1/4, 1/8, 1/16, 1/32 of the original size. UPerHead [34] is selected as the decoder. The learning rate is set to 0.00006 with the AdamW optimizer. The weight decay is 2 × 10 2 and the power of poly optimization strategy is 1. A mixed precision scheme is employed to reduce memory usage. The model is trained for 40k iterations loaded with weight pretrained on the ImageNet.

4.3. Test Set Transformation

The original dataset was captured in normal weather conditions. It is necessary to augment the test set for robustness evaluation. Inspired by the generation of corrupted ImageNet data sets in [18], we also render the corrupted RSIs test set with different degrees of fog and average the classification results for judging the robustness performance. We model the fog corruption through a diamond square algorithm which is to create a weighted heat map blended with the clean image. Thus, we can acquire corrupted test sets of Potsdam and Vaihingen, which are employed to measure the classification robustness in clean and foggy conditions. Corrupted samples are displayed in Figure 6, with fog generated corresponding to five levels of severity. This facilitates the robustness evaluation in various foggy conditions. Evaluation values are averaged over all five severity levels. As can be observed from Figure 6, the fog-covered region is indistinguishable from the actual scenario.

4.4. Metrics

Metrics like overall accuracy (OA), F 1 s c o r e ( F 1 s c o r e ) are selected to evaluate the classification accuracy. Following the robustness evaluation in [18,19], we take Corruption degradation ( C D ) and relative corruption degradation ( r C D ) into consideration for measuring LCC robustness. Specifically, metrics for accuracy evaluation are defined in Equation (19) to Equation (21). TP, FP, TN, FN represent true positive, false positive, true negative, and false negative classifications, respectively. Higher F 1 s c o r e and O A indicate a better classification accuracy.
P r e c i s i o n = T P T P + F P R e c a l l = T P T P + F N
O A = T P + T N T P + F P + T N + F N
F 1 s c o r e = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
From Equation (22) to Equation (24), r e f stands for the baseline which is regarded as a reference model in the ablation and comparison experiments. D refers to the degree of degradation. S signifies obtaining the mean value across different corruption degrees. f is the selected model. C l e a n and F represents clean and corrupted datasets. D s , f F ˜ means that we acquire the average degradation value using model f on RSIs across different degrees of fog corruption.
D = 1 F 1 s c o r e
C D F f = D s , F f ˜ D s , F r e f ˜ × 100 %
r C D F f = D s , F f ˜ D c l e a n f D s , F r e f ˜ D c l e a n r e f × 100 %
C D is a measure of absolute robustness. C D greater than 100% indicates a decline in robustness compared to the reference. The part over 100% represents the degradation in performance. To evaluate the relative robustness, r C D takes the performance on the clean dataset into account. Based on the reference, it is a proportional measure of the degradation in robustness relative to the clean data. When r C D < 100%, it indicates that the performance degradation in foggy conditions is less than that of the corresponding reference value compared with the clean. When C D or r C D > 100%, it means that model is not as robust as the reference. Robustness is better when both values are lower.

5. Result

5.1. Impact of Different Modules

To verify the effectiveness of each component, we conduct an ablation study on Vaihingen by removing or replacing the original part. From the qualitative visualization in Figure 7, it can be noted that when SSRL is added alone, the model can enhance capturing the global context and semantic features for classifying land cover edges. However, owing to the lack of height information, it is tough to grasp the correlation in the vertical space precisely (e.g., car in the box in Figure 7d). When MRFM is incorporated, the model can obtain the complementary information of multiple modalities, yet boundary labeling is coarse due to the absence of semantic representation details. The integration of both increases the model robustness and allows it to classify various land covers in foggy conditions more accurately.
In Table 2, Table 3, Table 4, Table 5, Table 6 and Table 7, ✓ indicates the element that we incorporate on top of the baseline. Values in the fog column refer to the average value across five severity levels of fog corruption. We first illustrate the effectiveness of each constituent. To demonstrate the robustness and accuracy of HRNet-W48, ResNet50 is selected for comparison in view of the comparable size. H and R in Table 2, Table 3 and Table 4 correspond to HRNet and ResNet respectively. When only SSRL is integrated, we directly transfer the Optical and DSM feature maps following convolutional layers into the decoder. When only MRFM is available, we exclude SSRL and transfer both modalities into MRFM. U and C in loss column represent the Unified loss and Cross-entropy loss. I m p S u r f and L o w V e g signify the impervious surface and low vegetation.
Combining Table 2, Table 3 and Table 4 with Figure 8, we observe that the incorporation of SSRL alone could improve the accuracy, but for impervious surface ( C D = 145.23%, r C D = 105.12%) and low vegetation ( C D = 119.53%, r C D = 135.26%), values of both exceed 100% and robustness is still inferior to the others. In the absence of effective multimodal fusion, SSRL is more biased towards the regular shape and small objects in Vaihingen. Moreover, the addition of MRFM is conducive to the improvement of robustness. Misclassification result of tail-end distributed cars ( F 1 s c o r e = 64.08%, C D = 131.45%) manifests if just cross-entropy loss function is utilized. When compared with the Unified loss, F 1 s c o r e is reduced by 8.59%, while C D is increased by 31.45%, indicating that UFL improves the robustness of imbalanced distributed objects ( F 1 s c o r e = 72.67%).

5.2. Impact of Multimodal Fusion

We also perform an ablation study with different inputs to investigate the improvement of robustness under multimodal fusion. ✓ represents the input modality. In the case of single modal input, the original multimodality is replaced by the identical modal input. From Table 5, Table 6 and Table 7, we can conclude that utilizing a single modality alone is less effective than multimodal fusion. When using DSM alone, the model performs poorly because DSM contains fewer semantic features compared with the optical. There is an 8.35% and 35.95% performance loss compared to the corresponding result of the optical input. Fusing multimodalities improves accuracy and robustness in foggy environments. Compared to the single optical modal input, the accuracy is 5.9% higher and r C D is 36.18% lower. From Figure 9, we can observe that our model is capable of classifying edges of the cars robustly in dense fog.

6. Discussion

To further elucidate the model robustness, we select some existing SOTA methods to conduct a comparison experiment on the Potsdam dataset. Models in the experiment can be grouped into CNNs and Transformers. Specifically, CNN-based models contain FCN [5], UNet [6], PSPNet [12], DeepLabV3+[13], CCNet [43], OCRNet [44], TRM [15], and Transformer-based models include SETR [27], Segmenter [45], and Segformer [28]. To ensure that models in the experiment have comparable parameters, we adopt ResNet101 [35] as the backbone for FCN, UNet, PSPNet, DeepLabV3+, and CCNet. TRM, OCRNet, and our framework are built on top of HRNet-W48 [4]. Encoder backbones selected for SETR, Segmenter, and Segformer are DeiT-B [46], DeiT-B, and MiT-B5 [28], respectively. Thus, the parameter of each model is around 70–90M in size. To utilize multimodal data, we stack the optical and DSM inputs in the channel dimension for all models excluding ours.
Results obtained from Figure 10 and Figure 11 and Table 8 show that Transformers perform better compared to CNNs on the clean test set with comparable sizes. However, the robustness of most ViTs is significantly reduced on the fog corrupted test set, with the exception of Segformer. There is a coarse classification of cars and edges in the box regions. Multimodal fusion allows our model to precisely learn the hierarchical features of inter-modal and the relationship between neighboring objects and the global. In this way, edges and interiors can be accurately classified. Compared to our previously proposed algorithm TRM, the accuracy and robustness LCC have been improved as a result of SSRL and MRFM, which enhance the ability to capture semantic information in low-quality images with a more effective data fusion approach. Specifically speaking, the F 1 s c o r e on the corrupted dataset improved by 1.3%, while there is a reduction of over 3% on both C D and r C D .
It can be concluded from Table 8, Table 9 and Table 10 that performance on the clean and fog corrupted test set varies significantly with regard to different models, whereas C D and r C D results are generally stable. ViTs perform better than CNNs on the clean test set (Segformer is 0.31% better on the Mean F 1 s c o r e and 0.71% better on OA), which can be attributed to the capability of capturing global information. However, in terms of robustness, CNNs represented by OCRNet are stronger than the best-performing ViTs (OCRNet’s C D decreased by 0.36% and r C D decreased by 2.23% in the mean F 1 s c o r e compared to Segformer), which is attributable to the fact that ViTs require more sophisticated training strategies with data augmentation. We are able to achieve a balance between accuracy and robustness compared to several SOTAs, as shown in Figure 12, where our model encircles a relatively smaller area ( C D decreases by 3.96% and r C D decreases by 2.87% on OA compared to OCRNet). The balanced classification result of each class also reflects the robustness. This illustrates the effectiveness of the proposed framework in generalizing and handling with class imbalance.

7. Conclusions

This study set out to design a robust model for LCC. The framework utilizes multimodal fusion and attention mechanisms to achieve a robust segmentation of RSIs in foggy conditions. We transfer heterogeneous data into HRNet, which serves as the backbone to maintain the high-resolution representation. Incorporating MRFM into the framework can exploit cross-modal complementary fusion. SSRL is deployed for exploring the correlations between different channels and positions. Unified loss helps to mitigate class imbalance issues. Multiple experiment analyses reveal that the proposed model has superior robustness on the fog-corrupted Potsdam and Vaihingen test sets. In addition, this study has also confirmed that in terms of robustness, ViTs are often inferior to CNNs in the presence of natural noises. Overall, this study highlights the importance of multimodal fusion and attention mechanisms for enhancing segmentation robustness. Our future research plans to investigate this topic further by combining ViTs with more fundamental attributes of RSIs.

Author Contributions

Conceptualization, W.S.; methodology, W.S.; validation, W.S.; writing—review and editing, W.S. and A.C.; project administration, W.Q. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Key R&D Program of Jiangsu Province under Grant BE2019311, Jiangsu Modern Agricultural Industry Key Technology Innovation Project under Grant CX(20)2013 and National Key Research and Development Program under Grant 2020YFB160070301.

Data Availability Statement

No new data were created or analyzed in this study.

Acknowledgments

A publicly available GitHub repository fastai was adapted for the experiment.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
RSIsRemote Sensing Images
NSIsNatural Scene Images
LCCLand Cover Classification
SSRLSpectral and Spatial Representation Learning
MRFMMultimodal Representation Fusion Module
DSMDigital Surface Model
SOTAState Of The Art
CNNConvolution Neural Network
ViTVisual Transformer
TOPTrue Ortho Photos
GSDGround Sampling Distance
OAOverall Accuracy
C D Corruption Degradation
r C D Relative Corruption Degradation
GTGround Truth
ISPRSInternational Society for Photogrammetry and Remote Sensing

Appendix A

Table A1. Experiment Environment.
Table A1. Experiment Environment.
SoftwareSoftware VersionHardwareHardware Version
CUDA10.2CPUi7-5930K CPU @ 3.50 GHz
cuDNN7.6GPU2 × Titan XP(12G)
Pytorch1.7RAM64 GB
Fast.ai2.2.2HARD DISKToshiba SSD 2T
Wandb0.1.20SYSTEMUbuntu 18.0.4

References

  1. He, H.; Li, C.; Yang, R.; Zeng, H.; Li, L.; Zhu, Y. Multisource Data Fusion and Adversarial Nets for Landslide Extraction from UAV-Photogrammetry-Derived Data. Remote Sens. 2022, 14, 3059. [Google Scholar] [CrossRef]
  2. Shao, S.; Xiao, L.; Lin, L.; Ren, C.; Tian, J. Road Extraction Convolutional Neural Network with Embedded Attention Mechanism for Remote Sensing Imagery. Remote Sens. 2022, 14, 2061. [Google Scholar] [CrossRef]
  3. Ding, J.; Zhang, J.; Zhan, Z.; Tang, X.; Wang, X. A Precision Efficient Method for Collapsed Building Detection in Post-Earthquake UAV Images Based on the Improved NMS Algorithm and Faster R-CNN. Remote Sens. 2022, 14, 663. [Google Scholar] [CrossRef]
  4. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. arXiv 2020, arXiv:1908.07919. [Google Scholar] [CrossRef]
  5. Shelhamer, E.; Long, J.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  6. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. arXiv 2015, arXiv:1505.04597. [Google Scholar]
  7. Xu, Q.; Yuan, X.; jun Ouyang, C.; Zeng, Y. Attention-Based Pyramid Network for Segmentation and Classification of High-Resolution and Hyperspectral Remote Sensing Images. Remote Sens. 2020, 12, 3501. [Google Scholar] [CrossRef]
  8. Zhang, G.; Lei, T.; Cui, Y.; Jiang, P. A dual-path and lightweight convolutional neural network for high-resolution aerial image segmentation. ISPRS Int. J. Geo-Inf. 2019, 8, 582. [Google Scholar] [CrossRef]
  9. Li, X.; Jiang, Y.; Peng, H.; Yin, S. An aerial image segmentation approach based on enhanced multi-scale convolutional neural network. In Proceedings of the 2019 IEEE International Conference on Industrial Cyber Physical Systems (ICPS), Taipei, Taiwan, 6–9 May 2019; pp. 47–52. [Google Scholar]
  10. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. arXiv 2017, arXiv:1606.00915. [Google Scholar] [CrossRef]
  11. Yang, M.; Yu, K.; Zhang, C.; Li, Z.; Yang, K. DenseASPP for Semantic Segmentation in Street Scenes. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; IEEE: Salt Lake City, UT, USA, 2018; pp. 3684–3692. [Google Scholar] [CrossRef]
  12. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. arXiv 2017, arXiv:1612.01105. [Google Scholar]
  13. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-Decoder with Atrous Separable Convolution for Semantic Image Segmentation. arXiv 2018, arXiv:1802.02611. [Google Scholar]
  14. Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. arXiv 2016, arXiv:1511.00561. [Google Scholar] [CrossRef] [PubMed]
  15. Shi, W.; Qin, W.; Yun, Z.; Chen, A.; Huang, K.Y.; Zhao, T. Land Cover Classification in Foggy Conditions: Toward Robust Models. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  16. Zhang, X.Y.; Liu, C.L.; Suen, C.Y. Towards Robust Pattern Recognition: A Review. Proc. IEEE 2020, 108, 894–922. [Google Scholar] [CrossRef]
  17. Tang, S.; Gong, R.; Wang, Y.; Liu, A.; Wang, J.; Chen, X.; Yu, F.; Liu, X.; Song, D.; Yuille, A.; et al. RobustART: Benchmarking Robustness on Architecture Design and Training Techniques. arXiv 2021, arXiv:2109.05211. [Google Scholar]
  18. Hendrycks, D.; Dietterich, T. Benchmarking Neural Network Robustness to Common Corruptions and Perturbations. arXiv 2019, arXiv:1903.12261. [Google Scholar]
  19. Kamann, C.; Rother, C. Benchmarking the Robustness of Semantic Segmentation Models. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 8825–8835. [Google Scholar] [CrossRef]
  20. Shao, R.; Shi, Z.; Yi, J.; Chen, P.Y.; Hsieh, C.J. On the Adversarial Robustness of Visual Transformers. arXiv 2021, arXiv:2103.15670. [Google Scholar]
  21. Mahmood, K.; Mahmood, R.; van Dijk, M. On the Robustness of Vision Transformers to Adversarial Examples. arXiv 2021, arXiv:2104.02610. [Google Scholar]
  22. Wang, X.; Girshick, R.; Gupta, A.; He, K. Non-Local Neural Networks. arXiv 2018, arXiv:1711.07971. [Google Scholar]
  23. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Zuo, W.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. arXiv 2020, arXiv:1910.03151. [Google Scholar]
  24. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. GCNet: Non-local Networks Meet Squeeze-Excitation Networks and Beyond. arXiv 2019, arXiv:1904.11492. [Google Scholar]
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image Is Worth 16x16 Words: Transformers for Image Recognition at Scale. arXiv 2021, arXiv:2010.11929. [Google Scholar]
  26. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer Using Shifted Windows. arXiv 2021, arXiv:2103.14030. [Google Scholar]
  27. Zheng, S.; Lu, J.; Zhao, H.; Zhu, X.; Luo, Z.; Wang, Y.; Fu, Y.; Feng, J.; Xiang, T.; Torr, P.H.S.; et al. Rethinking Semantic Segmentation from a Sequence-to-Sequence Perspective with Transformers. arXiv 2021, arXiv:2012.15840. [Google Scholar]
  28. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and Efficient Design for Semantic Segmentation with Transformers. arXiv 2021, arXiv:2105.15203. [Google Scholar]
  29. Yuan, L.; Hou, Q.; Jiang, Z.; Feng, J.; Yan, S. VOLO: Vision Outlooker for Visual Recognition. arXiv 2021, arXiv:2106.13112. [Google Scholar]
  30. Gu, Y.; Hao, J.; Chen, B.; Deng, H. Top-Down Pyramid Fusion Network for High-Resolution Remote Sensing Semantic Segmentation. Remote Sens. 2021, 13, 4159. [Google Scholar] [CrossRef]
  31. Yan, L.; Huang, J.; Xie, H.; Wei, P.; Gao, Z. Efficient Depth Fusion Transformer for Aerial Image Semantic Segmentation. Remote Sens. 2022, 14, 1294. [Google Scholar] [CrossRef]
  32. Audebert, N.; Le Saux, B.; Lefèvre, S. Beyond RGB: Very High Resolution Urban Remote Sensing with Multimodal Deep Networks. ISPRS J. Photogramm. Remote Sens. 2018, 140, 20–32. [Google Scholar] [CrossRef]
  33. Liu, H.; Zhang, J.; Yang, K.; Hu, X.; Stiefelhagen, R. CMX: Cross-Modal Fusion for RGB-X Semantic Segmentation with Transformers. arXiv 2022, arXiv:2203.04838. [Google Scholar]
  34. Xiao, T.; Liu, Y.; Zhou, B.; Jiang, Y.; Sun, J. Unified Perceptual Parsing for Scene Understanding. arXiv 2018, arXiv:1807.10221. [Google Scholar]
  35. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. arXiv 2015, arXiv:1512.03385. [Google Scholar]
  36. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. arXiv 2019, arXiv:1709.01507. [Google Scholar]
  37. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. arXiv 2018, arXiv:1807.06521. [Google Scholar]
  38. Yeung, M.; Sala, E.; Schönlieb, C.B.; Rundo, L. Unified Focal Loss: Generalising Dice and Cross Entropy-Based Losses to Handle Class Imbalanced Medical Image Segmentation. Comput. Med Imaging Graph. 2022, 95, 102026. [Google Scholar] [CrossRef] [PubMed]
  39. Ma, J.; Chen, J.; Ng, M.; Huang, R.; Li, Y.; Li, C.; Yang, X.; Martel, A.L. Loss Odyssey in Medical Image Segmentation. Med. Image Anal. 2021, 71, 102035. [Google Scholar] [CrossRef]
  40. Lin, T.Y.; Goyal, P.; Girshick, R.B.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 318–327. [Google Scholar] [CrossRef] [PubMed]
  41. Abraham, N.; Khan, N.M. A Novel Focal Tversky Loss Function With Improved Attention U-Net for Lesion Segmentation. In Proceedings of the 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019), Venice, Italy, 8–11 April 2019; pp. 683–687. [Google Scholar]
  42. Zhang, Z.; Sabuncu, M.R. Generalized Cross Entropy Loss for Training Deep Neural Networks with Noisy Labels. In Proceedings of the NeurIPS, Montreal, QC, Canada, 3–8 December 2018. [Google Scholar]
  43. Huang, Z.; Wang, X.; Wei, Y.; Huang, L.; Shi, H.; Liu, W.; Huang, T.S. CCNet: Criss-Cross Attention for Semantic Segmentation. arXiv 2020, arXiv:1811.11721. [Google Scholar] [CrossRef]
  44. Yuan, Y.; Chen, X.; Chen, X.; Wang, J. Segmentation Transformer: Object-Contextual Representations for Semantic Segmentation. arXiv 2021, arXiv:1909.11065. [Google Scholar]
  45. Strudel, R.; Garcia, R.; Laptev, I.; Schmid, C. Segmenter: Transformer for Semantic Segmentation. arXiv 2021, arXiv:2105.05633. [Google Scholar]
  46. Touvron, H.; Cord, M.; Douze, M.; Massa, F.; Sablayrolles, A.; Jégou, H. Training Data-Efficient Image Transformers & Distillation through Attention. arXiv 2021, arXiv:2012.12877. [Google Scholar]
Figure 1. Challenges associated with robust LCC in the areas covered with fog. The first row is the fog corrupted image and the second row is the corresponding ground truth (GT).
Figure 1. Challenges associated with robust LCC in the areas covered with fog. The first row is the fog corrupted image and the second row is the corresponding ground truth (GT).
Remotesensing 14 04551 g001
Figure 2. (a) Illustration of the proposed framework for LCC in foggy conditions. The input consists of a pair of the optical image and DSM. UperHead is selected as the decoder. (b) Detailed design of SSRL, which is composed of Spectral Attention and Spatial Attention modules. C and C are different channel numbers. (c) The framework of MRFM is to explore the complementarity of heterogeneous inputs.
Figure 2. (a) Illustration of the proposed framework for LCC in foggy conditions. The input consists of a pair of the optical image and DSM. UperHead is selected as the decoder. (b) Detailed design of SSRL, which is composed of Spectral Attention and Spatial Attention modules. C and C are different channel numbers. (c) The framework of MRFM is to explore the complementarity of heterogeneous inputs.
Remotesensing 14 04551 g002
Figure 3. Overview of the backbone HRNet. HRNet maintains high-resolution representations and exchanges information throughout branches by means of 1 × 1 and 3 × 3 convolution. It can cope well with intra-class heterogeneity and multi-scale issues in RSIs with fog.
Figure 3. Overview of the backbone HRNet. HRNet maintains high-resolution representations and exchanges information throughout branches by means of 1 × 1 and 3 × 3 convolution. It can cope well with intra-class heterogeneity and multi-scale issues in RSIs with fog.
Remotesensing 14 04551 g003
Figure 4. Illustrations about the composition of each class in ISPRS datasets.
Figure 4. Illustrations about the composition of each class in ISPRS datasets.
Remotesensing 14 04551 g004
Figure 5. Illustrations about samples of Potsdam and Vaihingen dataset. Augmented means the image after the augmentation operation.
Figure 5. Illustrations about samples of Potsdam and Vaihingen dataset. Augmented means the image after the augmentation operation.
Remotesensing 14 04551 g005
Figure 6. Illustrations of five severity levels of fog rendered the ISPRS dataset. First row: Potsdam. Second row: Vaihingen.
Figure 6. Illustrations of five severity levels of fog rendered the ISPRS dataset. First row: Potsdam. Second row: Vaihingen.
Remotesensing 14 04551 g006
Figure 7. Illustrations about some ablation results of each component in the framework. In this case, optical images are corrupted by fog, which belongs to severity level 2.
Figure 7. Illustrations about some ablation results of each component in the framework. In this case, optical images are corrupted by fog, which belongs to severity level 2.
Remotesensing 14 04551 g007
Figure 8. Radar plot visualization of the ablation study results. The first in top two is the classification result when different modules are integrated into the backbone. The second is about different inputs. Based on the radar plots regarding C D and r C D , the smaller envelope range is indicative of a model that is more robust on a fog-corrupted test set.
Figure 8. Radar plot visualization of the ablation study results. The first in top two is the classification result when different modules are integrated into the backbone. The second is about different inputs. Based on the radar plots regarding C D and r C D , the smaller envelope range is indicative of a model that is more robust on a fog-corrupted test set.
Remotesensing 14 04551 g008
Figure 9. Comparing the LCC results across different corruption levels. 1st column displays the clean image and ground truth. Others are images with fog of various severity levels, accompanied by the corresponding semantic labeling result. From left to right, the fog intensity increases gradually.
Figure 9. Comparing the LCC results across different corruption levels. 1st column displays the clean image and ground truth. Others are images with fog of various severity levels, accompanied by the corresponding semantic labeling result. From left to right, the fog intensity increases gradually.
Remotesensing 14 04551 g009
Figure 10. Qualitative comparisons between different methods applied to semantic segmentation of RSIs. The optical image is corrupted by the third severity level of fog.
Figure 10. Qualitative comparisons between different methods applied to semantic segmentation of RSIs. The optical image is corrupted by the third severity level of fog.
Remotesensing 14 04551 g010
Figure 11. Qualitative comparisons between different methods applied to semantic segmentation of RSIs. The optical image is corrupted by the fourth severity level of fog.
Figure 11. Qualitative comparisons between different methods applied to semantic segmentation of RSIs. The optical image is corrupted by the fourth severity level of fog.
Remotesensing 14 04551 g011
Figure 12. Radar plot for the robust performance of several SOTAs on the Potsdam test set. The envelope area of a robust model should be small and balanced. Although ViTs can boost the performance, CNNs manifest stronger robustness compared with ViTs.
Figure 12. Radar plot for the robust performance of several SOTAs on the Potsdam test set. The envelope area of a robust model should be small and balanced. Although ViTs can boost the performance, CNNs manifest stronger robustness compared with ViTs.
Remotesensing 14 04551 g012
Table 1. Detailed HRNET specifications about every stage and channels.
Table 1. Detailed HRNET specifications about every stage and channels.
Downsp. RateStage1Stage2Stage3Stage4
1 × 1 , 64 3 × 3 , 64 1 × 1 , 256 × 4 × 1 3 × 3 , C 3 × 3 , C × 4 × 1 3 × 3 , C 3 × 3 , C × 4 × 1 3 × 3 , C 3 × 3 , C × 4 × 3
3 × 3 , 2 C 3 × 3 , 2 C × 4 × 1 3 × 3 , 2 C 3 × 3 , 2 C × 4 × 4 3 × 3 , 2 C 3 × 3 , 2 C × 4 × 3
16 × 3 × 3 , 4 C 3 × 3 , 4 C × 4 × 4 3 × 3 , 4 C 3 × 3 , 4 C × 4 × 3
32 × 3 × 3 , 8 C 3 × 3 , 8 C × 4 × 3
Table 2. Ablation study results with different components integrated based on the clean and fog corrupted variants of the Vaihingen test set. The F 1 s c o r e is averaged across five severity levels. The best results are marked in bold.
Table 2. Ablation study results with different components integrated based on the clean and fog corrupted variants of the Vaihingen test set. The F 1 s c o r e is averaged across five severity levels. The best results are marked in bold.
MethodPer-Class F 1 s c o r e  (%)Mean F1 (%) OA (%)
Imp Surf*BuildingLow Veg*TreeCar
BaselineSSRLMRFMLossCleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
H U84.9372.1888.7176.7181.7068.3282.5170.9383.4663.8284.2670.3985.9273.12
H U88.1375.1090.9780.2182.6670.3884.3671.3285.7965.9286.3872.5886.9175.16
RU85.9475.8688.8577.0482.9665.7183.8369.0184.0963.6585.1370.2586.6372.53
HC91.7480.7092.0580.0881.8671.4786.9773.7484.2164.0887.3674.0188.5576.97
HU92.9880.8593.1981.5483.3973.5087.9175.2892.1272.6789.9276.7790.4979.15
Table 3. C D in ablation study with different components integrated based on clean and fog corrupted variants of the Vaihingen test set. Our framework is regarded as the reference and values above 100% denote a decrease in robustness compared with the reference. The highest C D is bold.
Table 3. C D in ablation study with different components integrated based on clean and fog corrupted variants of the Vaihingen test set. Our framework is regarded as the reference and values above 100% denote a decrease in robustness compared with the reference. The highest C D is bold.
Method  C D for Per-Class F 1 s c o r e  (%)Mean C D (%)  C D for OA (%)
Imp Surf*BuildingLow Veg*TreeCar
BaselineSSRLMRFMLossCleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
H U214.50145.23165.73126.15110.16119.53144.69117.61209.78132.41156.06127.44148.08128.94
H U169.01130.01132.58107.19104.39111.78129.40116.04180.21124.74135.05118.01137.69119.14
RU200.23126.05163.72124.36102.58129.40133.78125.37201.80133.03147.44128.04140.65131.75
HC117.66100.77116.64107.90109.22107.66107.81106.24200.27131.45125.30111.86120.36110.48
HU100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00
Table 4. The relative C D in ablation study with different components integrated based on clean and fog corrupted variants of the Vaihingen test set. Lower r C D suggests an improvement of robustness in the presence of fog corruption. The highest r C D is bold.
Table 4. The relative C D in ablation study with different components integrated based on clean and fog corrupted variants of the Vaihingen test set. Lower r C D suggests an improvement of robustness in the presence of fog corruption. The highest r C D is bold.
Method r C D for Per-Class F 1 s c o r e  (%)Mean r C D  (%) r C D for OA (%)
BaselineSSRLMRFMLossImp Surf*BuildingLow Veg*TreeCar
H U105.12103.01135.2691.70101.03105.49112.89
H U107.4292.35124.19103.25102.24104.94103.59
RU83.10101.36174.43117.31105.13113.17124.28
HC90.99102.78105.05104.74103.54101.55102.19
HU100.00100.00100.00100.00100.00100.00100.00
Table 5. Quantitative evaluation for clean and fog corrupted variants of the Vaihingen test set about using different input data. Every F 1 s c o r e is averaged over all five severity levels. The best results are marked in bold.
Table 5. Quantitative evaluation for clean and fog corrupted variants of the Vaihingen test set about using different input data. Every F 1 s c o r e is averaged over all five severity levels. The best results are marked in bold.
MethodPer-Class F 1 s c o r e  (%)Mean F1 (%)OA (%)
Optical DSM Imp Surf*BuildingLow Veg*TreeCar
CleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
90.7077.0591.5883.4982.3270.9185.3273.5690.6365.8788.1174.1788.6973.25
85.5069.2888.6679.1776.2159.7083.5965.1283.8955.8583.5765.8280.1865.21
92.9880.8593.1981.5483.3973.5087.9175.2892.1272.6789.9276.7790.4979.15
Table 6. Quantitative evaluation of C D on clean and fog corrupted variants of the Vaihingen test set about using different input data. Our framework is regarded as the reference and values above 100% denote a decrease in robustness compared with the reference. The highest C D is bold.
Table 6. Quantitative evaluation of C D on clean and fog corrupted variants of the Vaihingen test set about using different input data. Our framework is regarded as the reference and values above 100% denote a decrease in robustness compared with the reference. The highest C D is bold.
Method C D for Per-Class F 1 s c o r e  (%)Mean C D (%) C D for OA (%)
OpticalDSMImp Surf*BuildingLow Veg*TreeCar
CleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
132.42119.84123.5989.42106.41109.76121.47106.96118.83124.91117.91111.16118.93128.32
206.37160.39166.46112.84143.21152.07135.73141.11204.35161.56162.92147.11208.43166.88
100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00
Table 7. Quantitative evaluation of r C D on clean and fog corrupted variants of the Vaihingen test set in ablation study about using different input data evaluated. Our framework is regarded as the reference and lower r C D indicates an improvement of robustness in the presence of fog corruption. The highest r C D is bold.
Table 7. Quantitative evaluation of r C D on clean and fog corrupted variants of the Vaihingen test set in ablation study about using different input data evaluated. Our framework is regarded as the reference and lower r C D indicates an improvement of robustness in the presence of fog corruption. The highest r C D is bold.
Method r C D for Per-Class F 1 s c o r e  (%)Mean r C D  (%) r C D for OA (%)
OpticalDSMImp Surf*BuildingLow Veg*TreeCar
112.5669.44115.4093.08127.37105.98136.18
133.7681.49166.94146.25144.21134.98132.04
100.00100.00100.00100.00100.00100.00100.00
Table 8. Quantitative comparison with SOTA methods on the clean and fog corrupted variants of the Potsdam test set. Each F 1 s c o r e is averaged over all severity levels. The best results are marked in bold.
Table 8. Quantitative comparison with SOTA methods on the clean and fog corrupted variants of the Potsdam test set. Each F 1 s c o r e is averaged over all severity levels. The best results are marked in bold.
MethodPer-Class F 1 s c o r e (%)Mean F1 (%)OA (%)
Imp Surf*BuildingLow Veg*TreeCar
CleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
FCN [5]85.1563.2286.0867.1279.0263.1581.7263.5878.2455.6182.0462.5485.8658.33
UNet [6]85.6166.9388.1772.7979.3163.9979.1664.7780.6162.8982.5766.2786.2561.85
PSPNet [12]86.3466.5289.7972.5577.3462.4679.8864.7380.9561.7482.8665.6088.3462.31
DeepLabV3+[13]87.7371.8390.0675.0880.2769.5983.8970.5681.2363.1284.6470.0488.7670.86
CCNet [43]89.4970.5590.3773.2483.9169.7184.4770.5980.6360.9885.7769.0188.2764.91
OCRNet [44]88.7774.8690.6978.1984.0272.9584.1171.0285.3370.7986.5873.5689.3174.92
SETR [27]86.0973.3889.5373.0484.5768.4484.6170.9885.9364.7186.1570.1189.2670.99
Segmenter [45]89.9472.3090.2275.6184.5869.2184.7771.9084.8564.8386.8770.7789.8873.42
SegFormer [28]89.5375.5292.4080.0284.0769.9985.3872.1383.0569.4886.8973.4390.0174.91
TRM [15]89.1175.5791.4580.6183.8369.2985.5672.3783.5869.6286.7173.4989.9275.27
Ours89.3175.1692.4980.7084.2370.1585.8372.4582.0869.8386.7973.6690.1776.57
Table 9. Quantitative evaluation of C D on clean and fog corrupted variants of Potsdam test set about different SOTA methods. FCN is regarded as the reference and values lower than 100% represent an improvement in the robust performance compared with the reference. The minimum in each Fog column is bold.
Table 9. Quantitative evaluation of C D on clean and fog corrupted variants of Potsdam test set about different SOTA methods. FCN is regarded as the reference and values lower than 100% represent an improvement in the robust performance compared with the reference. The minimum in each Fog column is bold.
Method  C D for Per-Class F 1 s c o r e (%) C D for Mean F1 (%) C D for OA (%)
Imp Surf*BuildingLow Veg*TreeCar
CleanFogCleanFogCleanFogCleanFogCleanFogCleanFogCleanFog
FCN [5]100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00100.00
UNet [6]96.9089.9184.9982.7698.6297.72114.0096.7389.1183.6097.0590.0297.2491.55
PSPNet [12]91.9991.0373.3583.49108.01101.87110.0796.8487.5586.1995.4491.8282.4690.45
DeepLabV3+[13]82.6376.5971.4175.7994.0482.5288.1380.8386.2683.0885.5679.9879.4969.93
CCNet [43]70.7780.0769.1881.3976.6982.2084.9680.7589.0287.9079.2282.7182.9684.21
OCRNet [44]75.6268.3566.8866.3376.1773.4186.9379.5767.4265.8074.7170.5775.6060.19
SETR [27]93.6772.3875.2282.0073.5585.6484.1979.6864.6679.5077.1579.7875.9569.62
Segmenter [45]67.7475.3170.2674.1873.5083.5583.3277.1669.6279.2373.1078.0271.5763.79
SegFormer [28]70.5166.5654.6060.7775.9381.4479.9876.5277.9068.7573.0370.9370.6560.21
TRM [15]73.3366.4261.4258.9777.0783.3478.9975.8675.4668.4474.0370.7671.2959.35
Ours71.9967.5453.9558.7075.1781.0077.5275.6582.3567.9773.5770.3169.5256.23
Table 10. Quantitative evaluation of r C D on clean and fog corrupted variants of Potsdam test set about different SOTA methods. FCN is regarded as the reference and values lower than 100% represent an improvement in the robust performance compared to the reference. The minimum in each column is in bold.
Table 10. Quantitative evaluation of r C D on clean and fog corrupted variants of Potsdam test set about different SOTA methods. FCN is regarded as the reference and values lower than 100% represent an improvement in the robust performance compared to the reference. The minimum in each column is in bold.
Method r C D for Per-Class F 1 s c o r e  (%)Mean r C D  (%) r C D for OA (%)
Imp Surf*BuildingLow Veg*TreeCar
FCN [5]100.00100.00100.00100.00100.00100.00100.00
UNet [6]85.1881.1296.5379.3378.3083.5588.63
PSPNet [12]90.3890.9393.7683.5284.8988.4994.55
DeepLabV3+ [13]72.5079.0167.3073.4880.0374.8565.02
CCNet [43]86.3790.3589.4876.5286.8385.9284.85
OCRNet [44]63.4365.9369.7572.1664.2566.7652.27
SETR [27]57.9686.97101.6475.1493.7782.2166.36
Segmenter [45]80.4477.0696.8570.9588.4782.5559.79
SegFormer [28]63.8965.3088.7273.0459.9668.9954.85
TRM [15]61.7457.1791.6272.7161.6967.7453.21
Ours64.5262.1888.7273.7654.1367.3149.40
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Shi, W.; Qin, W.; Chen, A. Towards Robust Semantic Segmentation of Land Covers in Foggy Conditions. Remote Sens. 2022, 14, 4551. https://doi.org/10.3390/rs14184551

AMA Style

Shi W, Qin W, Chen A. Towards Robust Semantic Segmentation of Land Covers in Foggy Conditions. Remote Sensing. 2022; 14(18):4551. https://doi.org/10.3390/rs14184551

Chicago/Turabian Style

Shi, Weipeng, Wenhu Qin, and Allshine Chen. 2022. "Towards Robust Semantic Segmentation of Land Covers in Foggy Conditions" Remote Sensing 14, no. 18: 4551. https://doi.org/10.3390/rs14184551

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop