Next Article in Journal
Local Distributed Node for Power Quality Event Detection Based on Multi-Sine Fitting Algorithm
Previous Article in Journal
Analysis of Post-Processed Pseudorange-Based Point Positioning with Different Data Sources for the Current Galileo Constellations
Previous Article in Special Issue
Building Flexible, Scalable, and Machine Learning-Ready Multimodal Oncology Datasets
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Guidance Fusion Network for Cross-Modal Semantic Segmentation

College of Information, Shanghai Ocean University, No. 999 Hucheng Ring Road, Shanghai 201306, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(8), 2473; https://doi.org/10.3390/s24082473
Submission received: 24 February 2024 / Revised: 31 March 2024 / Accepted: 1 April 2024 / Published: 12 April 2024

Abstract

:
Leveraging data from various modalities to enhance multimodal segmentation tasks is a well-regarded approach. Recently, efforts have been made to incorporate an array of modalities, including depth and thermal imaging. Nevertheless, the effective amalgamation of cross-modal interactions remains a challenge, given the unique traits each modality presents. In our current research, we introduce the semantic guidance fusion network (SGFN), which is an innovative cross-modal fusion network adept at integrating a diverse set of modalities. Particularly, the SGFN features a semantic guidance module (SGM) engineered to boost bi-modal feature extraction. It encompasses a learnable semantic guidance convolution (SGC) designed to merge intensity and gradient data from disparate modalities. Comprehensive experiments carried out on the NYU Depth V2, SUN-RGBD, Cityscapes, MFNet, and ZJU datasets underscore both the superior performance and generalization ability of the SGFN compared to the current leading models. Moreover, when tested on the DELIVER dataset, the efficiency of our bi-modal SGFN displayed a mIoU that is comparable to the hitherto leading model, CMNEXT.

1. Introduction

Semantic segmentation presents a formidable challenge in several sectors, including autonomous driving, robotics, and virtual reality [1,2,3,4]. While neural networks have brought about a significant improvement in the accuracy and speed of RGB-based semantic segmentation in recent years, these models fall short when tasked with comprehensive 3D scene understanding. This is largely due to their inability to perceive crucial three-dimensional geometric information [5], thereby narrowing their practical applicability. Researchers have sought to mitigate these deficiencies by incorporating depth information [6] or thermal maps [7] into the semantic segmentation process. With advancements in sensor technology, an increasing number of sensors can now provide complementary detail to RGB images. However, the integration of multiple modalities is more challenging than a purely RGB-focused approach, as it requires effectually incorporating and leveraging the disparate characteristics of each modality. Therein lies the need for a flexible and adaptable network architecture specifically suited to multimodal segmentation.
The majority of current techniques for multimodal fusion can be categorized into two main strategies: early fusion and intermediate fusion. Early fusion involves merging RGB data with another modality at the channel level during the input stage [8,9,10]. Its downside is that it treats different modalities as uniform, thereby limiting its ability to fully exploit complementary information and pose challenges when applied to various modality combinations. On the flip side, intermediate fusion approaches [11,12,13] typically consist of parallel branches, where each branch is tasked with processing data from a specific modality. The features extracted at each layer from different modalities are then combined and passed through subsequent layers for semantic prediction. This layout is easily extendable to different RGB-based combinations. However, most existing models are tailor-made for specific combinations such as ACNet [14] and SA-Gate [11] for RGB-Depth data, resulting in underperformance when used for RGB-T segmentation (Figure 1). Therein lies the need for a robust and versatile multimodal network, one capable of adapting to diverse sensors without the necessity for dedicated architectures per modality. Unfortunately, the pursuit of this type of multimodal segmentation remains insufficient.
Moreover, many existing models [6,18,19] operate by assuming that each modality in multimodal data is invariably accurate. However, real-world measurement systems are prone to sensor faults, resulting in partial inaccuracies. For instance, issues such as LiDAR Jitter can introduce misalignments to sensing data [17]. Crucially, attempts at fusing such misaligned information could negatively impact overall segmentation performance, as depicted in Figure 1. Furthermore, multimodal data often display substantial levels of noise across the various sensing modalities. This noise can arise from various factors, including limited depth detection ranges [11], resulting in low-quality distance estimations and uncertainties revolving around dynamic cases [20]. It is critical to address these noise factors when dealing with multimodal data to ensure the reliability and accuracy of segmentation.
In light of these challenges, we present the semantic guidance fusion network (SGFN), a versatile cross-modal fusion network for multimodal semantic segmentation. Our model comprises two parallel transformers [21] paired with a novel Semantic Guidance Feature Fusion Module (SG-FFM) for pixel prediction. During the fusion, we introduce a semantic guidance module (SGM) for the extraction of supplementary multimodal information, which encompasses a learnable semantic guidance convolution (SGC). The SGC calculates the semantic distance of the supplementary modality to rectify the RGB data. This adjustment allows the fused feature to concentrate more on the complementary information while minimizing the negative impact of misalignments and noise from diverse modalities effectively.
In order to evaluate the efficacy of our proposed SGFN, we conducted exhaustive evaluations on six datasets, covering five distinct combinations of multiple modalities: RGB-Depth, RGB-Thermal, RGB-Polarization, RGB-Event, and RGB-LiDAR. Our results are noteworthy, with the highest mIoU of 57.6% on the NYU Depth V2 (RGB-D) dataset [15], 59.9% on the MFNet dataset (RGB-T) [7], and 93.1% on the ZJU dataset (RGB-P) [16]. Moreover, when tested on the DELIVER dataset [17], the performance of the SGFN is found to be on par with the previous best method. It is important to underline the fact that the SGFN surpasses specific architectures and outperforms existing multimodal methods, thereby affirming its effectiveness and superiority in delivering precise and robust multimodal semantic segmentation.
In summary, our research makes the following contributions:
  • We introduce a novel semantic guidance convolution (SGC) operation that calculates the similarity of adjacent pixels under the guidance of another modality to enhance the complementary cues and reduce the noise;
  • We further propose a new general multimodal segmentation network named SGFN, which is built on the SDC. This network is adept at effectively integrating and fusing bi-modal features from any amalgamation of modalities;
  • With comprehensive experiments on six datasets, our SGFN achieves state-of-the-art performance, covering RGB-D, RGB-T, RGB-P, RGB-E, and RGB-L tasks.

2. Related Works

2.1. Semantic Segmentation

Semantic segmentation is an intensive prediction task that needs to make predictions at the pixel level. Fully convolutional networks (FCNs) [22] first use a fully convolutional architecture to make pixel-wise prediction, which opens a new era of semantic segmentation. However, one upsampling step is insufficient to fully exploit the rich semantic information contained within the feature map. U-Net [23] uses an encoder-decoder structure to restore the image to its original resolution by multi-step upsampling while combining high-level features with low-level features through skip connections to improve accuracy. After that, a slew of advancements have been made to increase the receptive field for superior contextual comprehension. For example, PSPNet [24] introduced a pyramid pooling module to obtain contextual information at different scales. In parallel, the deepLab series [25,26,27,28] proposed atrous convolution, which broadened the receptive field without computational overhead. Some methods focus on improving semantic segmentation by enhancing boundaries [29,30,31] or applying attention blocks [32,33,34].
Recently, vision transformers [35] have been adopted as the backbone in dense prediction tasks [36,37] as well as in semantic segmentation tasks [21,38,39,40], demonstrating the effectiveness of global receptive fields. Although previous approaches have achieved impressive performance, they primarily rely on RGB images and, consequently, may suffer in challenging real-world scenarios. For instance, situations such as low-light conditions or fast-changing areas may expose the limitations of RGB images in accurately capturing minute scene details.

2.2. Multimodal Semantic Segmentation

Multimodal semantic segmentation is considerably enriched by the incorporation of disparate modalities, affording a more comprehensive understanding of scenes and bolstering overall performance. Numerous pieces of research have been undertaken on the fusion of RGB data with depth [6,9,11,19] and thermal [41,42,43,44] data, as they provide complementary information for scene analysis. Additional cues, such as polarization [45,46] and events [47,48], have also proven valuable in refining scene understanding. In the realms of autonomous driving, the integration of LiDAR data [49,50] and optical flow data [51] has drawn significant attention due to their essential contribution to perception tasks.
Despite these strides, many current approaches within this field tend to be tailored to specific modalities. This lack of adaptability to different combinations of sensing data impedes their broad application across various scenarios. CMX [52] offers a step towards handling this by putting forth a unified cross-modal fusion architecture for RGB-X segmentation, integrating cross-modal feature rectification and cross-attention feature fusion. Nonetheless, these methods often struggle when confronted with misaligned data resulting from sensor discrepancies.
By considering these constraints, we present a versatile framework that integrates cross-modal feature guidance to achieve robust segmentation. Our framework not only handles diverse combinations of modalities but also successfully mitigates the issues associated with misaligned data caused by sensor faults.

2.3. Central Difference Convolution

In order to improve edge performance, researchers have integrated gradient operators into vanilla convolutions, as the original convolution operation tends to smooth local features, resulting in decreased edge sharpness. By utilizing the fixed binary values, which are treated as filters, in convolution instead of learnable kernel weights, local binary convolution (LBC) [53,54] has been explored as an efficient alternative to traditional convolutions in various computer vision tasks. In the context of central difference convolution (CDC) [55,56,57], learnable kernels are employed to capture edge and texture details from the central difference map effectively; that is, y c = 0 n 1 w i · x i x c e n t e r , where w indicates the kernel weights, and x i represents the surrounding pixel of the center entry in the local patch. It calculates the difference between pixel values in the horizontal and vertical directions to estimate the gradient information. By aggregating the gradient-level details within the local patch, CDC demonstrated impressive performance in anti-spoofing tasks. Furthermore, pixel difference convolution (PDC) [58] offers a more versatile approach to encoding local differences by modifying sampling strategies within the local region. This flexibility allows PDC to explore microstructures with greater adaptability. Semantic difference convolution (SDC) [59], on the other hand, draws inspiration from the diffusion process [60] and amplifies semantic boundary awareness by incorporating a similarity map, which is generated by calculating semantic similarity. However, all these aforementioned operators entirely concentrate on the extraction of features from the current modality to enhance edge representation. In contrast to previous work [11,52,61], we are dedicated to developing an innovative and effective operator-level solution that incorporates a guidance map derived from another modality to extract complementary information.

3. Proposed Method

In this section, we will initially provide a detailed elaboration of the SGFN framework designed for multimodal semantic segmentation (Section 3.1), then the Semantic Guidance Feature Fusion Module (SG-FFM) is covered in Section 3.2, the Semantic Guidance Convolution (SGC) is covered in Section 3.2.1, and the corresponding semantic guidance module (SGM) is covered in Section 3.2.2.

3.1. Framework Overview

In Figure 2a, We apply an encoder-decoder structure to our SGFN. The encoder consists of two parallel backbones designed to extract features from RGB images and other modalities, including Depth, Thermal, Polarization, Event, LiDAR, and more. By following most of the previous works [24,32,62], we used a four-stage structure in the backbone to extract pyramidal features from each modality. At the end of each stage, the features from different modalities are rectified by the cross-modal feature rectification module (FRM) [52], which is crucial to promote interactions and reduce noise. Thus, the calibrated features are sent back to the backbone to continue the extraction of deeper characteristics. Moreover, as shown in Figure 2b, we designed a Semantic Guidance Feature Fusion Module (SG-FFM) to fuse the rectified features at each stage of the encoder, termed F. Within SGM, we introduce a semantic guidance convolution operator that takes the feature maps of another modality as a guide, thereby formulating a more enriched feature map. Ultimately, the features of four stages, F F 1 , F 2 , F 3 , F 4 , are passed to the decoder to predict the semantic map.

3.2. Semantic Guidance Feature Fusion Module

It is worth highlighting that the pixels share the same semantic label, demonstrating a higher degree of similarity. Extracting valuable features is essential when dealing with different modalities, as they often possess complementary information [11,14]. In this subsection, shown in Figure 2, we propose a new learnable approach known as the Semantic Guidance Feature Fusion Module (SG-FFM) to interact with two distinct modalities effectively. The SG-FFM consists of a semantic guidance module, comprising a parametric semantic guidance convolution operator, followed by a straightforward feature fusion process that generates enhanced features. Specifically, as indicated in Figure 2b, the channel dimensions of both modalities are initially compressed by a factor of one-eighth for reduction. Next, the modalities are sent into the SGM, which generates enhanced features. Subsequently, the outputs of SGM are convolved to achieve channel-wise alignment. At last, the generated outputs are integrated with “X” features and RGB features by performing an element-wise summation.

3.2.1. Semantic Guidance Convolution

By taking inspiration from SDC, which effectively applies semantic difference convolution to mimic the diffusion process and has shown substantial improvements in boundary performance, our approach also incorporates central difference into our SGC, emulating the diffusion process. The SGC encompasses two primary stages. In our method, the sampling step over the input feature follows a similar pattern to vanilla convolution. However, an amendment is introduced in the succeeding amalgamation step. The semantic guidance convolution, as depicted in Figure 3, focuses on aggregating the center-oriented gradient of local patches from complementary modalities. SGC generates the output value Y by taking the feature map V and another modality’s guidance map, U, as input. Consequently, we express the formula as follows:
Y c = i = 0 n 1 W i · S U i U c e n t e r · V i V c e n t e r
where i enumerates the pixels in the current patch. The first term, W, represents the learnable kernel weights, which have the same size as U and V. The second term, S U i U c e n t e r , known as the semantic guidance term, quantifies the semantic distance between the central pixel and its surrounding counterparts at the same location in another modality. Specifically, S U i U c e n t e r = U i U c e n t e r 2 . The last term V i V c e n t e r , known as the central difference term, evaluates the disparity between adjacent pixels at the pixel level. The semantic guidance term (S) and central difference term (D) have a mutual influence on each other, with S capable of acting as a guiding factor to enhance or suppress D, and conversely, D can also have a similar effect on S. This dynamic interaction enables our network to successfully extract complementary information and effectively handle challenges such as noise or misalignments.
In semantic segmentation tasks, assimilating information from both the intensity and gradient levels holds critical value. Therefore, by combining vanilla convolution with semantic guidance convolution, we augment the capability to capture diverse and informative features, leading to improved robustness and accuracy in semantic segmentation. As a result, the semantic guidance convolution can be represented as
Y c = θ · i = 0 n 1 W i · S U i U c e n t e r · V i V c e n t e r semantic guidance term + 1 θ · i = 0 n 1 W i · V i vanilla term
In this context, θ 0 , 1 serves as a hyperparameter to govern the trade-off between the gradient term and intensity term. The ablation of θ will be demonstrated in Section 5.5.

3.2.2. Semantic Guidance Module

As previously discussed, SGC concentrates on drawing out complementary information from different modalities. In order to further augment the cross-modal interactions, we introduce an efficient yet simple module called the semantic guidance module (SGM), which builds upon our proposed SGC approach.
As shown in Figure 2c, our SGM utilizes a branching framework that accommodates two inputs,
F s g m = S G M F r g b , F x
where F r g b R C × H × W is the RGB features, and F x R C × H × W is the guidance feature from another modality. The process can be formulated as
F ^ x = C o n v 1 × 1 F x
F s g c = BN-Relu S G C F r g b , F ^ x
w = S i g m o i d C o n v 1 × 1 F s g c
F s g m = w · F s g c
In Equation (5), by utilizing a 1 × 1 convolution, the feature F x is reduced to one-eighth (by default) of its original size along the channel dimension. Afterward, the bi-modal inputs are fed into the proposed SGC, which is then followed by batch normalization (BN) [63] and the ReLU [64] activation function. Finally, a Sigmoid function is applied to compute the attention weight after recovering the channel with a 1 × 1 convolution.

4. Experiments

4.1. Datasets

In order to validate our proposed SGFN (semantic guidance fusion network), we conducted experiments on three datasets relating to RGB-Depth semantic segmentation, as well as datasets involving combinations of the RGB-Thermal, RGB-Polarization, RGB-Event, and RGB-LiDAR modalities.
NYU Depth V2 [15] is an indoor RGB-D dataset with a total of 1449 images categorized into 40 classes, displayed at a resolution of 640 × 480 pixels. The dataset is divided into a training set of 795 images and a testing set of 654 images.
SUN-RGBD [65] is an indoor RGB-D dataset containing 10,335 images classified into 37 categories. It’s split into 5285/5050 for training/testing. We cropped and resized the image to 480 × 480 .
Cityscapes [3] is a benchmark for outdoor datasets featuring urban street scenes; it is divided into training/validation/testing sets of 2975/500/1525 samples, respectively. It comprises fine annotations for 19 different classes. We took an input at the resolution of 1024 × 512 .
RGB-T MFNet [7]. The MFNet dataset comprises 1569 RGB-thermal pairs captured from urban street scenes, with a resolution of 640 × 480 and eight classes. Among these, 820 pairs were captured during the daytime, and the remaining were captured at night.
RGB-P ZJU [16] is an RGB-P dataset collected from college street scenes, which comprises a total of 394 images, with 344/50 pairs for training/evaluation. Each image pair in the dataset is annotated for eight distinct semantic classes. Each image was cropped and resized to 612 × 512 .
DELIVER [17] consists of Depth, LiDAR, Event, and RGB data, each with a size of 1024 × 1024 . It comprises a total of 7885 samples, including sensor faults such as LiDAR-Jitter and Event Low-resolution. It is divided into training (3983), validation (2005), and testing (1897) sets, with 25 classes. Our study focuses on exploring the combinations of the RGB-Event and RGB-LiDAR modalities.

4.2. Implementation Details

We used the parallel Mix-Transformer-B2 (MiT-B2) [21] architecture pre-trained on the ImageNet [66] dataset as the backbone and UPernet-decoder [67] for our model. We trained our models on Pytorch 1.8.1 with four 3090 GPUs. We chose the AdamW optimizer [68] with epsilon 10 8 , and weight decay 10 2 . The initial learning rate (LR) was set as 6 × 10 5 with a poly strategy. We warmed up the first 10 epochs with 0.1× the original LR. We used the cross-entropy loss function. Throughout the training, we performed data augmentation by random flipping and random scaling. For NYU Depth V2 and SUN RGB-D, we used multiscale flip testing for a fair comparison. Like most of the previous works [15,65,69], we applied two common metrics, pixel accuracy (Pixel Acc.) and mean intersection over union (mIoU), to evaluate our model.

5. Experiment Results and Analyses

In this section, we present extensive experiment results on six multimodal benchmarks to verify the efficacy of our proposed SGFN for multimodal semantic segmentation. The results are compared with state-of-the-art methods, as shown in the following subsections.

5.1. Results of the RGB-Depth Datasets

NYU Depth V2: The results of the NYU Depth V2 dataset with 40 categories are shown in Table 1. It is evident that the exceptional performance of our approach surpasses previous methods in terms of the scores. Specifically, our proposed method (utilizing MiT-B2) already achieves remarkable results with a mIoU of 53.4%. In building upon this success, our SGFN models based on MiT-B4 and B5 demonstrate significant enhancements, significantly elevating the mIoU to 56.9% and 57.6%, respectively. These outstanding improvements highlight the effectiveness and superiority of our approach in cross-modal semantic segmentation.
SUN-RGBD: As shown in Table 2, our method achieves the best scores in two evaluation metric when compared with previous works. Precisely, our models leveraging MiT-B2 and B5 achieve impressive mIoU scores of 50.4% and 52.8%, respectively. These results serve as strong evidence of the remarkable effectiveness of our proposed method and underscore the superiority of our approach in multimodal semantic segmentation.
Cityscapes: In order to examine its applicability to outdoor environments, we evaluated the performance of SGFN on the Cityscapes dataset. The findings, displayed in detail in Table 3, provide a comparative study between our models and the cutting-edge RGB and RGB-D methods. Remarkably, our RGB-D approach demonstrates a notable enhancement of 0.9% in mIoU over the MiT-B2 (RGB) model. Furthermore, our method leveraging MiT-B4 achieves a leading score of 83.1%, outperforming all other RGB-D techniques available in the field.

5.2. Results of the RGB-Thermal Dataset

In Table 4, the results of a set of experiments on the MFNet dataset to evaluate the generalization capability of our model can be seen. Our SGFN model, utilizing MiT-B4, achieves state-of-the-art performance, surpassing CMX [52] by 0.2% in mIoU. Our methods based on MiT-B2 already surpass RGB-only models as well as RGB-T methods such as FEANet [61], ABMDRNet [12], and GMNet [43], achieving a mIoU of over 59%. Models designed for RGB-D segmentation, such as ACNet [14] and SA-Gate [11], are not applicable to RGB-T scenarios, as they only interact in the channel dimension while neglecting the crucial pixel-wise information. It is evident that our SGFN achieves remarkable success in extracting multimodal information effectively.
As shown in Table 5, we conducted separate experiments for both daytime and nighttime conditions. In the daytime scenario, our method achieves a comparable mIoU of 52.5% compared to CMX [52]. However, in nighttime conditions, despite the noise caused by poor light, our model surpasses all other RGB-T methods, achieving an impressive mIoU that exceeds 60.0%.

5.3. Results of the RGB-Polarization Dataset

In order to demonstrate the generality of our method, we conducted further evaluations on the ZJU-RGB-P dataset [16]. By leveraging trichromatic representations, which have proven to be more informative than monochromatic representations [16,84], we introduced the trichromatic degree of linear polarization (DoLP) as supplementary data. In Table 6, our approach utilizing MiT-B2 already surpasses all other RGB-P models, achieving a mIoU of 92.8% and outperforming the previous leading model CMX by 0.2%. Furthermore, our top-performing model with MiT-B4 attains a leading score of 93.1%, validating the generalizability of our SGFN approach on RGB-P data.

5.4. Results of the DELIVER Dataset

In Table 7, we compare our SGFN with recent models on the DELIVER dataset to study the generalizability of our approach in RGB-Event and RGB-LiDAR. Overall, SGFN, with MiT-B2, obtains state-of-the-art status for fusing RGB data and Event data, reaching the same mIoU (57.48) as CMNeXt. Additionally, our model performs comparably to CMNeXt in the fusion of RGB and LiDAR data. These results show that our SGM plays a positive role in fusing dense-sparse data.

5.5. Ablation Study

In order to gain a deeper understanding of how the various components of our architecture contribute to the segmentation task, we conducted a comprehensive set of ablation studies. For a fair comparison, we took MiT-B2 as the backbone to evaluate the performance of our model on the NYU Depth V2 dataset.
Ablation of the SGFN architecture. As shown in Table 8, we performed ablation experiments on our SGFN architecture. If we remove the SGM module, the complementary modalities are simply combined with average fusion. The results show a decrease in mIoU by 1.5% compared to the baseline. This decline highlights the critical role of our SGM in facilitating effective cross-modal fusion. Similarly, when removing the FRM module, the modalities are extracted independently, without any interaction or influence on each other during the process. This leads to a decline in the value by 2.2%, indicating the importance of the FRM module in enabling robust cross-modal interactions between the RGB feature and the supplementary feature.
Comparing SGC with other Convolutions: In order to evaluate the effectiveness of semantic guidance convolution (SGC) in cross-modal fusion, we conducted experiments by replacing the SGC in SGM with two alternative convolutions: vanilla convolution [86] and CDC. The results, as shown in Table 8, demonstrate that SGC outperforms vanilla convolution by a margin of 1.0%. Interestingly, the CDC achieves a much lower score, which may be attributed to its design for edge detection rather than semantic segmentation tasks. These findings emphasize the irreplaceable role of our SGC in achieving accurate multimodal semantic segmentation.
Impact of  θ  in SGC: As mentioned in Section 3.2.1, the parameter θ controls the influence of the gradient-level details and intensity-level details in SGC. We systematically varied θ as a hyperparameter in the range of 0 to 1 to investigate the impact of the semantic guidance term on the overall performance of SGC. Notably, in Figure 4, when θ exceeds 0.3, SGC consistently outperforms vanilla convolution ( θ = 0 ), indicating the superior performance of SGC in capturing both gradient-level and intensity-level details. Since the highest performance is observed when θ is set to 0.5, we set θ = 0.5 as the default configuration for all experiments.
Impact of Kercel Size in SGC: In order to evaluate the impact of kernel size and dilation rate, we conducted an ablation study. As shown in Table 9, increasing the size of the kernel in SGC did not result in any improvement. This suggests that enlarging the receptive fields may not be necessary for our SGC, as it primarily focuses on capturing bi-modal interactions in the local region. Additionally, we observe negative effects when increasing the dilation rate, which potentially results in the loss of local details.

5.6. Qualitative Analysis

Visualization of Prediction Results: Figure 5 displays the visual outcomes of our cross-modal segmentation, indicating the proficiency of our method across various modalities. Specifically, for RGB-D results, the SGM successfully discerns depth details and rightly segments the bed. Conversely, the baseline, which relies exclusively on RGB images, misclassifies the bed as a sofa, thereby demonstrating the enhanced accuracy of our approach. In the context of RGB-T segmentation, the baseline model struggles under low illumination conditions, leading to erroneous segmentations. In contrast, our network overcomes this challenge by effectively correcting the errors and achieving clearer distinctions between objects and persons. Moreover, the integration of polarization cues in RGB-P segmentation enables the more precise segmentation of car and glass areas, further enhancing the overall performance of our method. Notably, our SGM exhibits robust generalization capabilities in dynamic scenes, effectively accommodating moving objects and improving segmentation accuracy. By leveraging the advantages of LiDAR points, our network successfully segments the complete structure of the bridge, whereas the baseline misclassifies it as part of the sky. These qualitative analyses collectively demonstrate the strong generalization ability of our approach across multiple modalities.
Visualization of Feature Maps: In order to demonstrate the efficacy of our semantic guidance module (SGM), we visualized the feature maps extracted from the first layer of the backbone. By comparing the feature maps before and after applying SGM, as depicted in Figure 6, we can clearly see the remarkable ability of our semantic guidance fusion network (SGFN) to accentuate edge details and effectively suppress noise for both RGB-D and RGB-T tasks. In RGB-D tasks, depth features provide crucial geometric information that can complement RGB features, enhancing edge performance. For RGB-T tasks, thermal images become more important, especially in nighttime scenarios where RGB images may struggle due to inadequate lighting conditions. The effective utilization of these complementary modalities demonstrates the efficacy of our SGM.

6. Conclusions

In this study, we introduce a novel approach for universal cross-modal semantic segmentation called semantic guidance fusion network (SGFN). Our method leverages a vision transformer architecture to extract meaningful features from diverse modalities. We put forward a cross-modal Semantic Guidance Feature Fusion Module (SG-FFM) for comprehensive interactions between diverse modalities. SG-FFM utilizes a novel operator-level operation to augment the RGB feature in boundary awareness so as to enhance the performance of multimodal fusion. Specifically, we calculate the semantic distance of the supplementary information to guide the pixel-wise relevance derived from RGB information. The extensive experiments conducted on six benchmark datasets, including RGB-Depth, RGB-Thermal, RGB-Polarization, RGB-Event, and RGB-LiDAR combinations, demonstrate the superior performance of our proposed SGFN compared to existing state-of-the-art methods for cross-modal semantic segmentation.
In the future, our goal is to tackle the challenge of cross-modal fusion beyond the current scope by adapting the SGFN framework to accommodate the integration of three or more distinct sensor data types.

Author Contributions

Conceptualization, P.Z. and M.C.; methodology, P.Z.; software, P.Z.; validation, P.Z. and M.G.; formal analysis, P.Z. and M.G.; investigation, P.Z. and M.G.; resources, P.Z. and M.G.; data curation, P.Z. and M.G.; writing—original draft preparation, P.Z.; writing—review and editing, P.Z. and M.C.; visualization, P.Z.; supervision, M.C.; project administration, M.C.; funding acquisition, M.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Research and Development Planning in Key Areas of Guang dong Province (No. 2021B0202070001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Weng, X.; Yan, Y.; Chen, S.; Xue, J.H.; Wang, H. Stage-Aware Feature Alignment Network for Real-Time Semantic Segmentation of Street Scenes. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 4444–4459. [Google Scholar] [CrossRef]
  2. Sheng, H.; Cong, R.; Yang, D.; Chen, R.; Wang, S.; Cui, Z. UrbanLF: A Comprehensive Light Field Dataset for Semantic Segmentation of Urban Scenes. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 7880–7893. [Google Scholar] [CrossRef]
  3. Cordts, M.; Omran, M.; Ramos, S.; Rehfeld, T.; Enzweiler, M.; Benenson, R.; Franke, U.; Roth, S.; Schiele, B. The Cityscapes Dataset for Semantic Urban Scene Understanding. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 3213–3223. [Google Scholar] [CrossRef]
  4. Brooks, F. What’s Real About Virtual Reality? In Proceedings of the IEEE Virtual Reality (Cat. No. 99CB36316), Virtual, 13–17 March 1999; pp. 2–3. [Google Scholar] [CrossRef]
  5. Liu, J.; Zhou, W.; Cui, Y.; Yu, L.; Luo, T. GCNet: Grid-like context-aware network for RGB-thermal semantic segmentation. Neurocomputing 2022, 506, 60–67. [Google Scholar] [CrossRef]
  6. Yang, J.; Bai, L.; Sun, Y.; Tian, C.; Mao, M.; Wang, G. Pixel Difference Convolutional Network for RGB-D Semantic Segmentation. IEEE Trans. Circuits Syst. Video Technol. 2023, 34, 1481–1492. [Google Scholar] [CrossRef]
  7. Ha, Q.; Watanabe, K.; Karasawa, T.; Ushiku, Y.; Harada, T. MFNet: Towards real-time semantic segmentation for autonomous vehicles with multi-spectral scenes. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5108–5115. [Google Scholar] [CrossRef]
  8. Gupta, S.; Girshick, R.; Arbeláez, P.; Malik, J. Learning rich features from RGB-D images for object detection and segmentation. In Proceedings of the Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, 6–12 September 2014; Part VII. Springer: Berlin/Heidelberg, Germany, 2014; pp. 345–360. [Google Scholar]
  9. Cao, J.; Leng, H.; Lischinski, D.; Cohen-Or, D.; Tu, C.; Li, Y. ShapeConv: Shape-aware Convolutional Layer for Indoor RGB-D Semantic Segmentation. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 7068–7077. [Google Scholar] [CrossRef]
  10. Wang, J.; Wang, Z.; Tao, D.; See, S.; Wang, G. Learning common and specific features for RGB-D semantic segmentation with deconvolutional networks. In Proceedings of the Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, 11–14 October 2016; Part V. Springer: Berlin/Heidelberg, Germany, 2016; pp. 664–679. [Google Scholar]
  11. Chen, X.; Lin, K.Y.; Wang, J.; Wu, W.; Qian, C.; Li, H.; Zeng, G. Bi-directional cross-modality feature propagation with separation-and-aggregation gate for RGB-D semantic segmentation. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 561–577. [Google Scholar]
  12. Zhang, Q.; Zhao, S.; Luo, Y.; Zhang, D.; Huang, N.; Han, J. ABMDRNet: Adaptive-weighted Bi-directional Modality Difference Reduction Network for RGB-T Semantic Segmentation. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 2633–2642. [Google Scholar] [CrossRef]
  13. Hazirbas, C.; Ma, L.; Domokos, C.; Cremers, D. Fusenet: Incorporating depth into semantic segmentation via fusion-based cnn architecture. In Proceedings of the Computer Vision–ACCV 2016: 13th Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; Revised Selected Papers, Part I. Springer: Berlin/Heidelberg, Germany, 2017; pp. 213–228. [Google Scholar]
  14. Hu, X.; Yang, K.; Fei, L.; Wang, K. ACNET: Attention Based Network to Exploit Complementary Features for RGBD Semantic Segmentation. In Proceedings of the 2019 IEEE International Conference on Image Processing (ICIP), Taipei, Taiwan, 22–25 September 2019; pp. 1440–1444. [Google Scholar] [CrossRef]
  15. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from rgbd images. In Proceedings of the 12th European Conference on Computer Vision—ECCV 2012, Florence, Italy, 7–13 October 2012; Part V. Springer: Berlin/Heidelberg, Germany, 2012; pp. 746–760. [Google Scholar]
  16. Xiang, K.; Yang, K.; Wang, K. Polarization-driven semantic segmentation via efficient attention-bridged fusion. Opt. Express 2021, 29, 4802–4820. [Google Scholar] [CrossRef]
  17. Zhang, J.; Liu, R.; Shi, H.; Yang, K.; Reiß, S.; Peng, K.; Fu, H.; Wang, K.; Stiefelhagen, R. Delivering Arbitrary-Modal Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 1136–1147. [Google Scholar]
  18. Valada, A.; Mohan, R.; Burgard, W. Self-supervised model adaptation for multimodal semantic segmentation. Int. J. Comput. Vis. 2020, 128, 1239–1285. [Google Scholar] [CrossRef]
  19. Seichter, D.; Köhler, M.; Lewandowski, B.; Wengefeld, T.; Gross, H.M. Efficient RGB-D Semantic Segmentation for Indoor Scene Analysis. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13525–13531. [Google Scholar] [CrossRef]
  20. Zhang, J.; Yang, K.; Stiefelhagen, R. ISSAFE: Improving Semantic Segmentation in Accidents by Fusing Event-based Data. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 1132–1139. [Google Scholar] [CrossRef]
  21. Xie, E.; Wang, W.; Yu, Z.; Anandkumar, A.; Alvarez, J.M.; Luo, P. SegFormer: Simple and efficient design for semantic segmentation with transformers. Adv. Neural Inf. Process. Syst. 2021, 34, 12077–12090. [Google Scholar]
  22. Long, J.; Shelhamer, E.; Darrell, T. Fully Convolutional Networks for Semantic Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015. [Google Scholar]
  23. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Part III. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  24. Zhao, H.; Shi, J.; Qi, X.; Wang, X.; Jia, J. Pyramid Scene Parsing Network. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  25. Chen, L.C.; Papandreou, G.; Schroff, F.; Adam, H. Rethinking atrous convolution for semantic image segmentation. arXiv 2017, arXiv:1706.05587. [Google Scholar]
  26. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Semantic image segmentation with deep convolutional nets and fully connected crfs. arXiv 2014, arXiv:1412.7062. [Google Scholar]
  27. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. DeepLab: Semantic Image Segmentation with Deep Convolutional Nets, Atrous Convolution, and Fully Connected CRFs. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 834–848. [Google Scholar] [CrossRef]
  28. Chen, L.C.; Zhu, Y.; Papandreou, G.; Schroff, F.; Adam, H. Encoder-decoder with atrous separable convolution for semantic image segmentation. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 801–818. [Google Scholar]
  29. Takikawa, T.; Acuna, D.; Jampani, V.; Fidler, S. Gated-SCNN: Gated Shape CNNs for Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 5228–5237. [Google Scholar] [CrossRef]
  30. Ding, H.; Jiang, X.; Liu, A.Q.; Thalmann, N.M.; Wang, G. Boundary-Aware Feature Propagation for Scene Segmentation. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6818–6828. [Google Scholar] [CrossRef]
  31. Yuan, Y.; Xie, J.; Chen, X.; Wang, J. Segfix: Model-agnostic boundary refinement for segmentation. In Proceedings of the 16th European Conference Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; Part XII. Springer: Berlin/Heidelberg, Germany, 2020; pp. 489–506. [Google Scholar]
  32. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual Attention Network for Scene Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 3141–3149. [Google Scholar] [CrossRef]
  33. Zhao, H.; Zhang, Y.; Liu, S.; Shi, J.; Loy, C.C.; Lin, D.; Jia, J. PSANet: Point-wise Spatial Attention Network for Scene Parsing. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  34. Huang, Z.; Wang, X.; Huang, L.; Huang, C.; Wei, Y.; Liu, W. CCNet: Criss-Cross Attention for Semantic Segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27–28 October 2019. [Google Scholar]
  35. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16 × 16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations, Vienna, Austria, 3–7 May 2021. [Google Scholar]
  36. Ranftl, R.; Bochkovskiy, A.; Koltun, V. Vision Transformers for Dense Prediction. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 12179–12188. [Google Scholar]
  37. Yuan, Y.; Fu, R.; Huang, L.; Lin, W.; Zhang, C.; Chen, X.; Wang, J. HRFormer: High-Resolution Vision Transformer for Dense Predict. In Advances in Neural Information Processing Systems; Ranzato, M., Beygelzimer, A., Dauphin, Y., Liang, P., Vaughan, J.W., Eds.; Curran Associates, Inc.: Red Hook, NY, USA, 2021; Volume 34, pp. 7281–7293. [Google Scholar]
  38. Zhang, Y.; Pang, B.; Lu, C. Semantic Segmentation by Early Region Proxy. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1248–1258. [Google Scholar] [CrossRef]
  39. He, H.; Cai, J.; Pan, Z.; Liu, J.; Zhang, J.; Tao, D.; Zhuang, B. Dynamic Focus-aware Positional Queries for Semantic Segmentation. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 11299–11308. [Google Scholar] [CrossRef]
  40. Gu, J.; Kwon, H.; Wang, D.; Ye, W.; Li, M.; Chen, Y.H.; Lai, L.; Chandra, V.; Pan, D.Z. Multi-Scale High-Resolution Vision Transformer for Semantic Segmentation. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12084–12093. [Google Scholar] [CrossRef]
  41. Shivakumar, S.S.; Rodrigues, N.; Zhou, A.; Miller, I.D.; Kumar, V.; Taylor, C.J. PST900: RGB-Thermal Calibration, Dataset and Segmentation Network. In Proceedings of the 2020 IEEE International Conference on Robotics and Automation (ICRA), Paris, France, 31 May–31 August 2020; pp. 9441–9447. [Google Scholar] [CrossRef]
  42. Wu, W.; Chu, T.; Liu, Q. Complementarity-aware cross-modal feature fusion network for RGB-T semantic segmentation. Pattern Recognit. 2022, 131, 108881. [Google Scholar] [CrossRef]
  43. Zhou, W.; Liu, J.; Lei, J.; Yu, L.; Hwang, J.N. GMNet: Graded-Feature Multilabel-Learning Network for RGB-Thermal Urban Scene Semantic Segmentation. IEEE Trans. Image Process. 2021, 30, 7790–7802. [Google Scholar] [CrossRef]
  44. Sun, Y.; Zuo, W.; Liu, M. RTFNet: RGB-Thermal Fusion Network for Semantic Segmentation of Urban Scenes. IEEE Robot. Autom. Lett. 2019, 4, 2576–2583. [Google Scholar] [CrossRef]
  45. Kalra, A.; Taamazyan, V.; Rao, S.K.; Venkataraman, K.; Raskar, R.; Kadambi, A. Deep Polarization Cues for Transparent Object Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  46. Mei, H.; Dong, B.; Dong, W.; Yang, J.; Baek, S.H.; Heide, F.; Peers, P.; Wei, X.; Yang, X. Glass Segmentation Using Intensity and Spectral Polarization Cues. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–20 June 2022; pp. 12622–12631. [Google Scholar]
  47. Alonso, I.; Murillo, A.C. EV-SegNet: Semantic segmentation for event-based cameras. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Long Beach, CA, USA, 16–17 June 2019. [Google Scholar]
  48. Zhang, J.; Yang, K.; Stiefelhagen, R. Exploring Event-Driven Dynamic Context for Accident Scene Segmentation. IEEE Trans. Intell. Transp. Syst. 2022, 23, 2606–2622. [Google Scholar] [CrossRef]
  49. Chang, X.; Pan, H.; Sun, W.; Gao, H. A Multi-Phase Camera-LiDAR Fusion Network for 3D Semantic Segmentation With Weak Supervision. IEEE Trans. Circuits Syst. Video Technol. 2023, 33, 3737–3746. [Google Scholar] [CrossRef]
  50. Yuan, Z.; Song, X.; Bai, L.; Wang, Z.; Ouyang, W. Temporal-Channel Transformer for 3D Lidar-Based Video Object Detection for Autonomous Driving. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2068–2078. [Google Scholar] [CrossRef]
  51. Rashed, H.; Yogamani, S.; El-Sallab, A.; Krizek, P.; El-Helw, M. Optical flow augmented semantic segmentation networks for automated driving. arXiv 2019, arXiv:1901.07355. [Google Scholar]
  52. Zhang, J.; Liu, H.; Yang, K.; Hu, X.; Liu, R.; Stiefelhagen, R. CMX: Cross-Modal Fusion for RGB-X Semantic Segmentation with Transformers. IEEE Trans. Intell. Transp. Syst. 2023, 24, 14679–14694. [Google Scholar] [CrossRef]
  53. Juefei-Xu, F.; Naresh Boddeti, V.; Savvides, M. Local binary convolutional neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 19–28. [Google Scholar]
  54. Zhang, X.; Liu, L.; Xie, Y.; Chen, J.; Wu, L.; Pietikainen, M. Rotation invariant local binary convolution neural networks. In Proceedings of the IEEE International Conference on Computer Vision Workshops, Venice, Italy, 22–29 October 2017; pp. 1210–1219. [Google Scholar]
  55. Yu, Z.; Qin, Y.; Zhao, H.; Li, X.; Zhao, G. Dual-cross central difference network for face anti-spoofing. arXiv 2021, arXiv:2105.01290. [Google Scholar]
  56. Yu, Z.; Wan, J.; Qin, Y.; Li, X.; Li, S.Z.; Zhao, G. NAS-FAS: Static-Dynamic Central Difference Network Search for Face Anti-Spoofing. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3005–3023. [Google Scholar] [CrossRef]
  57. Yu, Z.; Zhao, C.; Wang, Z.; Qin, Y.; Su, Z.; Li, X.; Zhou, F.; Zhao, G. Searching central difference convolutional networks for face anti-spoofing. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 5295–5305. [Google Scholar]
  58. Su, Z.; Liu, W.; Yu, Z.; Hu, D.; Liao, Q.; Tian, Q.; Pietikäinen, M.; Liu, L. Pixel Difference Networks for Efficient Edge Detection. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 5097–5107. [Google Scholar] [CrossRef]
  59. Tan, H.; Wu, S.; Pi, J. Semantic diffusion network for semantic segmentation. Adv. Neural Inf. Process. Syst. 2022, 35, 8702–8716. [Google Scholar]
  60. Sapiro, G. Geometric partial differential equations in image analysis: Past, present, and future. In Proceedings of the International Conference on Image Processing, Washington, DC, USA, 23–26 October 1995; Volume 3, pp. 1–4. [Google Scholar] [CrossRef]
  61. Deng, F.; Feng, H.; Liang, M.; Wang, H.; Yang, Y.; Gao, Y.; Chen, J.; Hu, J.; Guo, X.; Lam, T.L. FEANet: Feature-Enhanced Attention Network for RGB-Thermal Real-time Semantic Segmentation. In Proceedings of the 2021 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Prague, Czech Republic, 27 September–1 October 2021; pp. 4467–4473. [Google Scholar] [CrossRef]
  62. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid Vision Transformer: A Versatile Backbone for Dense Prediction Without Convolutions. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 568–578. [Google Scholar]
  63. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning, Lille, France, 7–9 July 2015; pp. 448–456. [Google Scholar]
  64. Glorot, X.; Bordes, A.; Bengio, Y. Deep sparse rectifier neural networks. In Proceedings of the Fourteenth International Conference on Artificial Intelligence and Statistics—JMLR Workshop and Conference Proceedings, Fort Lauderdale, FL, USA, 11–13 April 2011; pp. 315–323. [Google Scholar]
  65. Song, S.; Lichtenberg, S.P.; Xiao, J. SUN RGB-D: A RGB-D scene understanding benchmark suite. In Proceedings of the 2015 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Boston, MA, USA, 7–12 June 2015; pp. 567–576. [Google Scholar] [CrossRef]
  66. Russakovsky, O.; Deng, J.; Su, H.; Krause, J.; Satheesh, S.; Ma, S.; Huang, Z.; Karpathy, A.; Khosla, A.; Bernstein, M.; et al. Imagenet large scale visual recognition challenge. Int. J. Comput. Vis. 2015, 115, 211–252. [Google Scholar] [CrossRef]
  67. Xiao, T.; Liu, Y.; Zhou, B.; Jiang, Y.; Sun, J. Unified perceptual parsing for scene understanding. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 418–434. [Google Scholar]
  68. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  69. Gupta, S.; Arbeláez, P.; Malik, J. Perceptual Organization and Recognition of Indoor Scenes from RGB-D Images. In Proceedings of the 2013 IEEE Conference on Computer Vision and Pattern Recognition, Portland, OR, USA, 23–28 June 2013; pp. 564–571. [Google Scholar] [CrossRef]
  70. Qi, X.; Liao, R.; Jia, J.; Fidler, S.; Urtasun, R. 3D Graph Neural Networks for RGBD Semantic Segmentation. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 5209–5218. [Google Scholar] [CrossRef]
  71. Xu, D.; Ouyang, W.; Wang, X.; Sebe, N. Pad-net: Multi-tasks guided predictionand-distillation network for simultaneous depth estimation and scene parsing, in 2018 IEEE. In Proceedings of the CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 675–684. [Google Scholar]
  72. Zhang, Z.; Cui, Z.; Xu, C.; Yan, Y.; Sebe, N.; Yang, J. Pattern-Affinitive Propagation Across Depth, Surface Normal and Semantic Segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4101–4110. [Google Scholar] [CrossRef]
  73. Yang, Y.; Xu, Y.; Zhang, C.; Xu, Z.; Huang, J. Hierarchical Vision Transformer with Channel Attention for RGB-D Image Segmentation. In Proceedings of the 4th International Symposium on Signal Processing Systems, Xi’an, China, 25–27 March 2022; pp. 68–73. [Google Scholar]
  74. Wu, Z.; Zhou, Z.; Allibert, G.; Stolz, C.; Demonceaux, C.; Ma, C. Transformer Fusion for Indoor rgb-d Semantic Segmentation. SSRN. 2022. Available online: https://ssrn.com/abstract=4251286 (accessed on 18 October 2022).
  75. Chen, L.Z.; Lin, Z.; Wang, Z.; Yang, Y.L.; Cheng, M.M. Spatial Information Guided Convolution for Real-Time RGBD Semantic Segmentation. IEEE Trans. Image Process. 2021, 30, 2313–2324. [Google Scholar] [CrossRef] [PubMed]
  76. Zhang, G.; Xue, J.H.; Xie, P.; Yang, S.; Wang, G. Non-Local Aggregation for RGB-D Semantic Segmentation. IEEE Signal Process. Lett. 2021, 28, 658–662. [Google Scholar] [CrossRef]
  77. Zhou, H.; Qi, L.; Huang, H.; Yang, X.; Wan, Z.; Wen, X. CANet: Co-attention network for RGB-D semantic segmentation. Pattern Recognit. 2022, 124, 108468. [Google Scholar] [CrossRef]
  78. Oršic, M.; Krešo, I.; Bevandic, P.; Šegvic, S. In Defense of Pre-Trained ImageNet Architectures for Real-Time Semantic Segmentation of Road-Driving Images. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 12599–12608. [Google Scholar] [CrossRef]
  79. Sun, L.; Yang, K.; Hu, X.; Hu, W.; Wang, K. Real-time fusion network for RGB-D semantic segmentation incorporating unexpected obstacle detection for road-driving images. IEEE Robot. Autom. Lett. 2020, 5, 5558–5565. [Google Scholar] [CrossRef]
  80. Romera, E.; Álvarez, J.M.; Bergasa, L.M.; Arroyo, R. ERFNet: Efficient Residual Factorized ConvNet for Real-Time Semantic Segmentation. IEEE Trans. Intell. Transp. Syst. 2018, 19, 263–272. [Google Scholar] [CrossRef]
  81. Wang, P.; Chen, P.; Yuan, Y.; Liu, D.; Huang, Z.; Hou, X.; Cottrell, G. Understanding convolution for semantic segmentation. In Proceedings of the 2018 IEEE Winter Conference on Applications of Computer Vision (WACV), Lake Tahoe, NV, USA, 12–15 March 2018; pp. 1451–1460. [Google Scholar]
  82. Wang, J.; Sun, K.; Cheng, T.; Jiang, B.; Deng, C.; Zhao, Y.; Liu, D.; Mu, Y.; Tan, M.; Wang, X.; et al. Deep High-Resolution Representation Learning for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2021, 43, 3349–3364. [Google Scholar] [CrossRef]
  83. Xu, J.; Lu, K.; Wang, H. Attention fusion network for multi-spectral semantic segmentation. Pattern Recognit. Lett. 2021, 146, 179–184. [Google Scholar] [CrossRef]
  84. Yan, R.; Yang, K.; Wang, K. NLFNet: Non-Local Fusion Towards Generalized Multimodal Semantic Segmentation across RGB-Depth, Polarization, and Thermal Images. In Proceedings of the 2021 IEEE International Conference on Robotics and Biomimetics (ROBIO), Jinghong, China, 27–31 December 2021; pp. 1129–1135. [Google Scholar] [CrossRef]
  85. Broedermann, T.; Sakaridis, C.; Dai, D.; Van Gool, L. HRFuser: A multi-resolution sensor fusion architecture for 2D object detection. arXiv 2022, arXiv:2206.15157. [Google Scholar]
  86. Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
Figure 1. Comparative results of our SGFN on different benchmarks. (a) RGB and Depth on NYU Depth V2 [15] dataset; (b) RGB and Thermal on the MFNet [7] dataset; (c) RGB and degree of linear Polarization (DoLP) on ZJU [16] dataset; (d,e) RGB-Event and RGB-LiDAR on DELIVER [17] dataset.
Figure 1. Comparative results of our SGFN on different benchmarks. (a) RGB and Depth on NYU Depth V2 [15] dataset; (b) RGB and Thermal on the MFNet [7] dataset; (c) RGB and degree of linear Polarization (DoLP) on ZJU [16] dataset; (d,e) RGB-Event and RGB-LiDAR on DELIVER [17] dataset.
Sensors 24 02473 g001
Figure 2. (a) The overall architecture of our SGFN. It is composed of a parallel encoder to fuse RGB data and other modality data. (b) Details of the Semantic Guidance Feature Fusion Module (SG-FFM). (c) Detailed architecture of semantic guidance module (SGM).
Figure 2. (a) The overall architecture of our SGFN. It is composed of a parallel encoder to fuse RGB data and other modality data. (b) Details of the Semantic Guidance Feature Fusion Module (SG-FFM). (c) Detailed architecture of semantic guidance module (SGM).
Sensors 24 02473 g002
Figure 3. Semantic guidance convolution.
Figure 3. Semantic guidance convolution.
Sensors 24 02473 g003
Figure 4. Impact of θ on SGC.
Figure 4. Impact of θ on SGC.
Sensors 24 02473 g004
Figure 5. Visualization of qualitative comparison of RGB-only and our SGFN, arranged from top to bottom, showcasing the results for the RGB-Depth, RGB-Thermal, RGB-Polarization (AoLP), RGB-Event, and RGB-LiDAR semantic segmentation tasks.
Figure 5. Visualization of qualitative comparison of RGB-only and our SGFN, arranged from top to bottom, showcasing the results for the RGB-Depth, RGB-Thermal, RGB-Polarization (AoLP), RGB-Event, and RGB-LiDAR semantic segmentation tasks.
Sensors 24 02473 g005
Figure 6. Visualization of the feature maps before and after the application of the SGM.
Figure 6. Visualization of the feature maps before and after the application of the SGM.
Sensors 24 02473 g006
Table 1. Results of NYU Depth V2. “*” denotes the multi-scale test.
Table 1. Results of NYU Depth V2. “*” denotes the multi-scale test.
MethodmIoU (%)Acc (%)
3DGNN [70]43.1-
ACNet [14]48.3-
PADNet [71]50.262.3
PAP [72]50.462.5
Swin-RGBD [73]50.964.2
TransD [74]55.569.4
SGNet [75]51.176.8
ShapeConv [9]51.376.4
SA-Gate [11]52.477.9
CMX (MiT-B2) * [52]54.479.9
CMX (MiT-B5) * [52]56.980.1
SGFN (Mit-B2) *53.478.5
SGFN (Mit-B5) *57.680.5
Table 2. Results of SUN RGB-D. “*” denotes the multi-scale test.
Table 2. Results of SUN RGB-D. “*” denotes the multi-scale test.
MethodmIoU (%)Acc (%)
3DGNN [70]45.9-
ACNet [14]48.1-
SGNet [75]48.682.0
ShapeConv [9]48.682.2
NANet [76]48.882.3
PDCNet [6]49.2-
CANet [77]49.360.5
TransD [74]51.964.1
SA-Gate [11]49.482.5
CMX (MiT-B2) * [52]49.782.8
CMX (MiT-B5) * [52]52.483.8
SGFN (Mit-B2) *50.883.0
SGFN (Mit-B5) *53.184.1
Table 3. Results of Cityscapes.
Table 3. Results of Cityscapes.
MethodModalBackbonemIoU (%)
SwiftNet [78]RGBResNet-1870.4
ESANet [19]RGBResNet-5079.2
DANet [32]RGBResNet-10181.5
SegFormer [21]RGBMiT-B281.0
SegFormer [21]RGBMiT-B482.3
RFNet [79]RGB-DResNet-1872.5
PADNet [71]RGB-DResNet-5076.1
ESANet [19]RGB-DResNet-5080.0
SA-Gate [11]RGB-DResNet-5080.7
SA-Gate [11]RGB-DResNet-10181.7
CMX [52]RGB-DMiT-B281.6
CMX [52]RGB-DMiT-B482.6
SGFNRGB-DMiT-B281.6
SGFNRGB-DMiT-B483.0
Table 4. Per-class comparison on the MFNet dataset for RGB-Thermal semantic segmentation.
Table 4. Per-class comparison on the MFNet dataset for RGB-Thermal semantic segmentation.
MethodModalUnlabeledCarPersonBikeCurveCar StopGuardrailColor ConeBumpmIoU
DANet [32]RGB96.371.348.151.830.218.20.730.318.841.3
SegNet [75]RGB96.765.355.751.138.410.00.012.051.542.3
UNet [23]RGB96.966.260.546.241.617.91.830.644.245.1
PSPNet [24]RGB96.874.861.350.238.415.80.033.244.446.1
ERFNet [80]RGB96.767.156.234.330.69.40.00.130.536.1
DUC [81]RGB97.782.569.458.940.120.93.442.140.950.7
HRNet [82]RGB98.086.967.359.235.323.11.746.647.351.7
SegFormer-B2 [21]RGB97.987.462.863.231.725.69.850.949.653.2
SegFormer-B4 [21]RGB98.088.964.062.838.125.96.950.857.754.8
MFNet [7]RGB-T96.965.958.942.929.99.90.025.227.739.7
SA-Gate [11]RGB-T96.873.859.251.338.419.30.024.548.845.8
ACNet [14]RGB-T96.779.464.752.732.928.40.816.944.446.3
RTFNet [44]RGB-T98.587.470.362.745.329.80.029.155.753.2
AFNet [83]RGB-T98.086.067.462.043.028.94.644.956.654.6
ABMDRNet [12]RGB-T98.684.869.660.345.133.15.147.450.054.8
FEANet [61]RGB-T98.387.871.161.146.522.16.655.348.955.3
GMNet [43]RGB-T97.586.573.161.744.042.314.548.747.457.3
CMX (MiT-B2) [52]RGB-T98.389.474.864.747.330.18.152.459.458.2
CMX (MiT-B4) [52]RGB-T98.390.175.264.550.235.38.554.260.659.7
SGFN (MiT-B2)RGB-T98.389.476.066.149.332.710.952.456.159.0
SGFN (MiT-B4)RGB-T98.490.976.766.149.235.77.555.159.159.9
Table 5. Experiments on the nighttime and daytime images of the MFNet dataset.
Table 5. Experiments on the nighttime and daytime images of the MFNet dataset.
MethodModalDaytime mIoU (%)Nighttime mIoU (%)
SegFormer-B2 [21]RGB48.649.2
SegFormer-B4 [21]RGB49.452.4
GMNet [43]RGB-T49.057.7
MFNet [7]RGB-T36.136.8
RTFNet [44]RGB-T45.854.8
ABMDRNet [12]RGB-T46.755.5
CMX (MiT-B2) [52]RGB-T51.357.8
CMX (MiT-B4) [52]RGB-T52.559.4
SGFN (MiT-B2)RGB-T52.058.7
SGFN (MiT-B4)RGB-T52.560.0
Table 6. Results of the ZJU dataset for RGB-Polarization segmentation.
Table 6. Results of the ZJU dataset for RGB-Polarization segmentation.
MethodModalmIoU (%)
SwiftNet [78]RGB80.3
SegFormer-B2 [21]RGB89.6
NLFNet [84]RGB-P84.4
EAFNet [16]RGB-P85.7
CMX (SegFormer-B2) [52]RGB-AoLP92.0
CMX (SegFormer-B4) [52]RGB-AoLP92.6
CMX (SegFormer-B2) [52]RGB-DoLP92.2
CMX (SegFormer-B4) [52]RGB-DoLP92.5
SGFN (SegFormer-B2)RGB-DoLP92.8
SGFN (SegFormer-B4)RGB-DoLP93.2
Table 7. Results of the DELIVER datasets for RGB-Event and RGB-LiDAR semantic segmentation.
Table 7. Results of the DELIVER datasets for RGB-Event and RGB-LiDAR semantic segmentation.
MethodModalBackbonemIoU (%)
HRFuser [85]RGBHRFormer-T47.95
SegFormer [21]RGBMiT-B257.20
HRFuser [85]RGB-EventHRFormer-T42.22
CMX [52]RGB-EventMiT-B256.52
CMNeXt [17]RGB-EventMiT-B257.48
SGFNRGB-EventMiT-B257.48
HRFuser [85]RGB-LiDARHRFormer-T43.13
CMX [52]RGB-LiDARMiT-B256.37
CMNeXt [17]RGB-LiDARMiT-B258.04
SGFNRGB-LiDARMiT-B257.70
Table 8. Albation on the SGFN architecture.
Table 8. Albation on the SGFN architecture.
StructuremIoU (%)
SGFN (MiT-B2)53.4
   -without SGM51.6 (−1.8)
   -without FRM50.6 (−2.8)
   -with CDC instead of SGC51.4 (−2.0)
   -with vanilla instead of SGC52.3 (−1.1)
   -with SDC instead of SGC52.6 (−0.8)
Table 9. Impact of the kernel size of the SGC operator.
Table 9. Impact of the kernel size of the SGC operator.
Kernel SizeDilation RatePixel Acc. (%)mIoU (%)
3 × 3178.553.4
3 × 3378.253.1
3 × 3577.952.7
5 × 5178.453.3
7 × 7178.252.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, P.; Chen, M.; Gao, M. Semantic Guidance Fusion Network for Cross-Modal Semantic Segmentation. Sensors 2024, 24, 2473. https://doi.org/10.3390/s24082473

AMA Style

Zhang P, Chen M, Gao M. Semantic Guidance Fusion Network for Cross-Modal Semantic Segmentation. Sensors. 2024; 24(8):2473. https://doi.org/10.3390/s24082473

Chicago/Turabian Style

Zhang, Pan, Ming Chen, and Meng Gao. 2024. "Semantic Guidance Fusion Network for Cross-Modal Semantic Segmentation" Sensors 24, no. 8: 2473. https://doi.org/10.3390/s24082473

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop