Next Article in Journal
Three-Dimensional Defect Characterization of Ultrasonic Detection Based on GCNet Improved Contrast Learning Optimization
Next Article in Special Issue
A Neural Multi-Objective Capacitated Vehicle Routing Optimization Algorithm Based on Preference Adjustment
Previous Article in Journal
Temporal-Aware Neural Network for Dense Non-Rigid Structure from Motion
Previous Article in Special Issue
Anomaly Detection Model of Network Dataflow Based on an Improved Grey Wolf Algorithm and CNN
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interactive Efficient Multi-Task Network for RGB-D Semantic Segmentation

1
Key Laboratory of Machine Perception, Shenzhen Graduate School, Peking University, Shenzhen 518055, China
2
School of Intelligent Systems Engineering, Sun Yat-sen University, Shenzhen 518107, China
*
Author to whom correspondence should be addressed.
Electronics 2023, 12(18), 3943; https://doi.org/10.3390/electronics12183943
Submission received: 8 August 2023 / Revised: 29 August 2023 / Accepted: 6 September 2023 / Published: 19 September 2023
(This article belongs to the Special Issue Deep Perception in Autonomous Driving)

Abstract

:
Semantic segmentation is significant for robotic indoor activities. However, relying solely on RGB modality often leads to poor results due to limited information. Introducing other modalities can improve performance but also increases complexity and cost, making it unsuitable for real-time robotic applications. To address the balance issue of performance and speed in robotic indoor scenarios, we propose an interactive efficient multitask RGB-D semantic segmentation network (IEMNet) that utilizes both RGB and depth modalities. On the premise of ensuring rapid inference speed, we introduce a cross-modal feature rectification module, which calibrates the noise of RGB and depth modalities and achieves comprehensive cross-modal feature interaction. Furthermore, we propose a coordinate attention fusion module to achieve more effective feature fusion. Finally, an instance segmentation task is added to the decoder to assist in enhancing the performance of semantic segmentation. Experiments on two indoor scene datasets, NYUv2 and SUNRGB-D, demonstrate the superior performance of the proposed method, especially on the NYUv2, achieving 54.5% mIoU and striking an excellent balance between performance and inference speed at 42 frames per second.

1. Introduction

In recent years, improving robot perception capabilities [1,2] has become a hotspot in various applications. Recent research employs sensors such as lasers or LIDAR [3,4], but the absence of visual information limits optimal performance. Consequently, semantic segmentation [5], which effectively utilizes visual data, has gained prominence. It aims to classify images at the pixel level, assigning each pixel to specific categories.
With the emergence of deep learning, an increasing number of high-performance models [6,7,8,9] have shown remarkable results, gradually replacing traditional semantic segmentation methods. However, in complex indoor scenes, relying solely on the RGB modality is insufficient for refined segmentation due to the lack of object boundary and geometric information. However, depth data, which contains rich position, contour, and geometric information, is easily collected using RGB-D sensors such as Microsoft Kinect, making it an excellent modality. Thus, various approaches for RGB-D semantic segmentation [10,11,12] have been studied. Nonetheless, these methods possess limited interaction within cross-modality features and often ignore the impact of noise of both modalities.
Moreover, most RGB-D semantic segmentation works focus on feature fusion [13,14,15] between RGB and depth modality. Although these methods provide instructive fusion modules to integrate the two kinds of information, fully utilizing both modalities for detailed semantic segmentation remains a challenge.
On the other hand, the increased complexity introduced by depth information substantially raises inference time, making it challenging for real-time applications. In semantic segmentation, numerous studies [16,17,18] employ different techniques to minimize computational costs. However, for RGB-D semantic segmentation, most methods do not focus on the balance issue of performance and speed, causing slow inference speed.
To solve the above problems, we propose an interactive efficient multi-task RGB-D semantic segmentation network, IEMNet. By introducing a cross-modality feature rectification module [15], feature interaction and rectification are increased by leveraging channel-wise and spatial-wise correlations. In addition, we also propose a coordinate attention fusion module for capturing long-range interactions and achieving comprehensive feature fusion. Meanwhile, lightweight and efficient architecture [19] is utilized to ensure efficient inference speed. Without affecting inference speed, we also use a multi-task decoder [19], training the model with semantic segmentation and instance segmentation tasks simultaneously, further enhancing semantic segmentation performance. Evaluations of two indoor scene datasets NYUv2 [20] and SUNRGB-D [21] demonstrate that IEMNet achieves an optimal balance between performance and speed compared to other advanced methods.

2. Related Work

RGB-D semantic segmentation. Incorporating depth modality in RGB modality can increase segmentation performance. CNN-based RGB-D semantic segmentation has achieved great advances in recent years. Various methods have been proposed for higher accuracy in different aspects. CANet [10] employed a three-branch encoder comprising RGB, deep, and mixed branches, effectively complementing information via a co-attention module. PGDENet [11] utilized depth data with a dual-branch encoder and depth enhancement module. FRNet [12] proposed cross-level enriching and cross-modality awareness modules to boost representative information and obtain rich contextual information. FSFNet [13] introduced a symmetic cross-modality residual fusion module. CEN [14] proposed a novel feature fusion mechanism that replaces the current channel feature with a feature on the same channel from the other modality according to the scale factor of the batch normalization layer. CMX [15] performed a feature fusion module with a cross attention mechanism. Shapeconv [22] introduced a shape-aware convolutional layer to utilize the shape information from depth modality. SGNet [23] proposed spatial information-guided convolution to help integrate the RGB feature and 3D spatial information. The SA-Gate [24] employed a separation-and-aggregation gating operation for jointly filtering and recalibrating both modalities. Different from the previous works, we improve feature interaction and rectification with spatial and channel correlation and propose a coordinate attention fusion module to integrate features from RGB and depth modality.
Efficient semantic segmentation. The objective of efficient semantic segmentation is to achieve high-quality segmentation accuracy with low computational cost. Several efficient segmentation networks have been reported to satisfy this requirement. MiniNet [16] introduced a multi-dilated depthwise convolution with fewer parameters and better performance. WFDCNet [17] proposed a depthwise factorized convolution to separate dimension and improve computational efficiency. FANet [18] presented a fast attention module and manually added a downsampling layer in ResNet to reduce the computational cost. FRNet [25] used an asymmetric encoder–decoder architecture with factorized and regular blocks, making a trade-off between accuracy and speed. DSANet [26] employed a channel split and shuffle to reduce the computation and maintain higher segmentation accuracy. In this work, we utilize a shallow backbone and a simple architecture to ensure efficient inference speed.

3. Proposed Method

3.1. Overview

We propose an Interactive Efficient Multi-Task RGB-D Semantic Segmentation Network (IEMNet) following the EMSANet [19], as shown in Figure 1. The network extracts RGB and depth information through two encoder branches with ResNet as backbone, and replaces each 3 × 3 convolution of the original ResNet with a Non-Bottleneck-1D-Block (NBt1D) to improve performance. The NBt1D consists of a 3 × 1 convolution layer, a ReLU activation function, and a 1 × 3 convolution layer.
Moreover, the Cross-Modal Feature Rectification Module (CM-FRM) [15] is introduced and placed between two encoder branches to integrate and rectify information from both the modalities. In this way, the interacted and rectified features will be fed into the next stage to further promote feature extraction. Meanwhile, we also propose the coordinate attention fusion module (CAFM) to replace the RGB-D fusion module in EMSANet, enabling more comprehensive feature fusion through the coordinate attention mechanism, merging two feature maps into one, and sending it to the decoder via the skip connection.
Our decoder consists of two branches [19]: one for semantic segmentation and the other for instance segmentation. The semantic branch includes three decoder modules, a 3 × 3 convolution layer, and two learned upsampling blocks. It incorporates shallow structure with skip connections and applies 4× upsampling to reduce computational cost. Each decoder module contains a 3 × 3 convolution layer, three NBt1D blocks, and a learned upsampling block. Additionally, we calculate loss on a multi-scale to improve performance and training effectiveness. The instance branch is essentially the same as the semantic branch, except for the task heads. Instance segmentation results are represented by 2D gaussian heatmaps encoded from their centers and offset vectors pointing from each pixel to the corresponding instance center in the x and y directions.

3.2. Cross-Modal Feature Rectification Module

To address noise rectification and facilitate feature interaction, we introduces the Cross-Modality Feature Rectification Module (CM-FRM) [15], which calibrates features from both modalities at different stages in the encoder. The detailed structure of the CM-FRM is depicted in Figure 2, where it performs feature rectification in both channel and spatial dimensions by leveraging information from the other modality.
Channel-wise feature rectification. Given the features of two modalities RGB in R H × W × C and Depth in R H × W × C , maximum and average pooling are first applied simultaneously across their channel dimensions, obtaining four output vectors. These vectors are concatenated and processed by a multi-layer perceptron (MLP) with a sigmoid activation function to obtain the channel weights, which are then split into W RGB C and W Depth C . Then, they are multiplied with the input separately on channel dimension, obtaining the channel-wise rectificated features as follows:
RGB rec C = W Depth C Depth in , Depth rec C = W RGB C RGB in ,
where ⊗ denotes channel-wise multiplication.
Spatial-wise feature rectification. Given the same features of two modalities RGB in and Depth in , they are concatenated on channel dimension and processed by two convolutional layers and a ReLU function. A sigmoid activation function is then applied to obtain the feature map F R H × W × 2 . Afterwards, F is split into two spatial weight maps W RGB S and W Depth S . Then, they are multiplied with the input separately on spatial dimension, obtaining the spatial-wise rectificated features as follows:
RGB rec S = W Depth S Depth in , Depth rec S = W RGB S RGB in ,
where ∗ denotes spatial-wise multiplication.
The final rectificated features of CM-FRM are obtained as follows:
RGB out = RGB in + λ C RGB rec C + λ S RGB rec S , Depth out = Depth in + λ C Depth rec C + λ S Depth rec S ,
where λ C and λ S are two hyperparameters, and both are set to 0.5 by default. Note that RGB rec C and RGB rec S are obtained by Depth in multiplying the corresponding weights. Utilizing the information of the other modality, the CM-FRM can achieve comprehensive feature rectification and interaction.

3.3. Coordinate Attention Fusion Module

Inspired by [27], we propose a coordinate attention fusion module (CAFM) that receives the rectificated features and applies coordinate attention for feature fusion. As shown in Figure 3, CAFM consists of two stages, the coordinate attention stage and the fusion stage. Through the coordinate attention mechanism, our network is able to capture long-range interactions and know which features to pay more attention to and which features to suppress.
Coordinate attention stage. Given the rectificated features RGB out R H × W × C and Depth out R H × W × C , spatially factorized adaptive pooling is first applied to both features to aggregate features along the two spatial directions, respectively, yielding four direction-aware feature maps. Then, we concatenate the feature maps on spatial dimension after the reshape operation and process the result with a 1 × 1 convolution and a non-linear activation function, yielding
Z = δ Conv 1 × 1 ( [ RGB h , RGB w , Depth h , Depth w ] ) ,
where [ . , . ] denotes the concatenation operation along the spatial dimension, δ is a non-linear activation function, and Z is the intermediate feature map.
Then, Z is split into two separate tensors G h and G w . Another two 1 × 1 convolutions are utilized to separately transform G h and G w to tensors with the same channel number to the input. Then, we split G h and G w as attention weights for two modalities. They are multiplied with the input separately. The process is expressed as follows:
G RGB h , G Depth h = F split σ Conv 1 × 1 ( G h ) , G RGB w , G Depth w = F split σ Conv 1 × 1 ( G w ) ,
RGB out CA = RGB out G RGB w G RGB h , Depth out CA = Depth out G Depth w G Depth h ,
where ⊗ denotes matrix multiplication.
Finally, RGB out CA and Depth out CA are concatenated on the channel dimension as the output of coordinate attention stage.
Fusion stage. In this stage, we use a simple channel embedding [15] to merge the features. Two 1 × 1 convolutions, a 3 × 3 depthwise convolution, and an intermediate ReLU activation function constitute the main path of the channel embedding. A 1 × 1 convolution forms the skip connection path. The concatenated features pass through these two paths separately and are then added to obtain the final output, which serves as the input for each stage of the decoder.

3.4. Loss Function

For semantic segmentation, we utilize the cross-entropy loss on a multi-scale, which is formulated as follows:
L sem = 1 4 i = 1 4 L CE ( S out i , S gt i ) ,
where S out i and S gt i denote semantic predictions and labels in a different scale.
For instance segmentation, Mean Squared Error (MSE) loss and L1 loss are used for center and offset supervision, respectively. The instance loss is computed as follows:
L ins = 1 4 λ MSE i = 1 4 L MSE ( C out i , C gt i ) + 1 4 λ L 1 i = 1 4 L L 1 ( O out i , O gt i ) ,
where C out i and C gt i denote center predictions and labels in different scales, O out i and O gt i denote offset predictions and labels in different scales, λ MSE and λ L 1 are the loss weight of the MSE loss and L1 loss, and we set them as 2 and 1 as default.
The total loss is organized as:
L = λ sem L sem + λ ins L ins ,
where λ sem and λ ins are hyperparameters, denoting task weights of semantic and instance segmentation.

4. Experiment

4.1. Datasets and Evaluation Measures

We evaluate the proposed method on two indoor datasets, NYUv2 and SUNRGB-D. NYUv2 [20] contains 1449 indoor RGB-D images with detailed annotations for semantic and instance segmentation. The standard dataset split is adopted, 795 images for training and the remaining 654 images for testing. SUNRGB-D [21] integrates multiple indoor RGB-D datasets, including NYUv2, with a total of 10,335 indoor images, annotated in 37 classes. Of these, 5285 images are used for training, and the remaining 5050 images are used for testing. The instance annotations of SUNRGB-D are extracted using 3D bounding boxes and semantic annotations [19]. Moreover, we use the Hypersim [28] dataset for pretraining. Hypersim is a realistic synthetic dataset, including 77,400 samples with masks for semantic and instance segmentation. The high-quality and large number of samples make it highly suitable for pretraining. We evaluated the proposed IEMNet and existing SOTA methods in terms of the mean intersection over union (mIoU), pixel accuracy (PA) [29], and frames per second (FPS).

4.2. Implementation Details

Our model was implemented based on Pytorch, and all training and testing processes were completed on the RTX A5000 GPU. During the experiments, the pre-trained weights from [19] on ImageNet were used to initialize the encoders of the two branches, and the model was trained for 500 epochs with a batch size of 8. We used the AdamW [30] optimizer with a weight decay of 0.01 for training, while the initial learning rate was set to 6 × 10 5 and 4 × 10 5 for NYUv2 and SUNRGB-D, and the poly learning rate schedule was applied to further adjust the learning rate. For data augmentation, random scaling, random cropping, and random flipping were employed, and for RGB images, slight color jitter augmentation was added in the HSV color space.

4.3. Quantitative Results on NYUv2 and SUNRGB-D

Table 1, Figure 4 and Figure 5 show the quantitative results of the proposed method and other methods on NYUv2 and SUNRGB-D, including the results after pretraining on the Hypersim and fine-tuning on the corresponding datasets. Task weights λ sem and λ ins were set as 1:3 for NYUv2 and 1:2 for SUNRGB-D, determined through extensive experiments. The pretraining process uses only 20% of Hypersim data in each epoch to accelerate training, with an initial learning rate of 6 × 10 5 and task weights of 1:2 for semantic and instance segmentation. Fine-tuning was performed on NYUv2 and SUNRGB-D, with the same task weights and learning rate. FPS is calculated on NYUv2.
It can be observed that introducing the instance segmentation task significantly improves performance on both datasets compared to using only semantic segmentation. Combining multiple tasks does not increase the computational cost during testing, as only semantic results are predicted. Meanwhile, our method gets further improvement after pretraining on Hypersim without increasing the inference cost. Using data several times larger than the original training set for pretraining achieves a huge improvement for our method, as the impact of the data on the model performance is significant. However, our method still achieves enhancement on this basis, which can be seen from the comparison with EMSANet and subsequent ablation experiments. In terms of mIoU comparison, our method ranks just behind CMX, but our inference speed is much faster, reaching 42.97 FPS. In terms of FPS comparison, our method is only slower than ESANet and EMSANet, but the mIoU is higher, achieving 54.5% on NYUv2 and 49.1% on SUNRGB-D, surpassing most methods. Thus, compared to other methods, our IEMNet achieves the optimal balance between performance and inference speed.

4.4. Qualitative Results on NYUv2

We visualize five semantic segmentation results of ESANet, EMSANet, and our method on the NYUv2 dataset, as shown in Figure 6. Visually, the method proposed in this paper performs better in segmentation accuracy than the ESANet and EMSANet. For example, our method IEMNet correctly identifies the TV cabinet and the washbasin while the ESANet and EMSANet wrongly classify them with other classes. It demonstrates that our proposed method enhances the network’s ability to utilize depth information through more comprehensive feature interaction and fusion, resulting in better performance.

4.5. Ablation Study on NYUv2

To investigate the impact of different components on segmentation performance and inference speed, we conducted ablation studies using only semantic decoder on NYUv2 and SUNRGB-D. The same training strategy and various model settings, including the CM-FRM, CAFM, and different backbones, are employed as shown in Table 2, where the RGB-D Fusion module is the original fusion module in the EMSANet.
Firstly, it can be seen that our CAFM outperforms RGB-D Fusion in fusing cross-modal features and achieving better results on both datasets. The significant improvement after introducing the CM-FRM demonstrates the effectiveness of enhanced feature interaction and calibration. Although using the CM-FRM and CAFM increases the inference time, the FPS remains at 42.97, achieving efficient speed. Note that the FPS is very slow when using Res50. This is because the block and number of layers are different with Res18 and Res34, causing a significant increase in parameter and computational complexity. Moreover, using NBt1D, PA, and mIoU is generally better, while the inference speed decreases. This discrepancy with [19,31,32] is due to their FPS calculations being performed on a robotic platform, while ours are on the Pytorch platform. Thus, NBt1D is more suitable for robotic platforms compared to normal convolution. Considering the performance and inference speed of various backbones, using Res34NBt1D as the backbone achieves better balance between performance and speed.

5. Conclusions

In order to achieve good balance between performance and speed for RGB-D semantic segmentation, we propose an interactive efficient multitask RGB-D semantic segmentation network. Based on the EMSANet, the cross-modal feature rectification module and coordinate attention fusion module are introduced into the encoder branches to accomplish feature rectification and fusion at different scales, constructing an interactive encoder structure that enables comprehensive feature interaction. Furthermore, without compromising the inference speed of the semantic segmentation task, our method incorporates an instance segmentation task, thereby enhancing the semantic segmentation performance. Results of the NYUv2 and SUNRGB-D datasets indicate that the proposed method exhibits optimal performance in balancing performance and speed under indoor scenes. In future work, we will port our method to robotic platforms to verify its performance. In addition, integrating more tasks into the current framework is expected to further improve performance, such as depth estimation.

Author Contributions

Conceptualization, X.X. and J.L.; methodology, X.X. and J.L.; software, X.X.; validation, X.X.; formal analysis, X.X.; investigation, X.X.; resources, X.X.; data curation, X.X.; writing—original draft preparation, X.X.; writing—review and editing, J.L. and H.L.; visualization, X.X.; supervision, J.L.; project administration, J.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded partly by the National Natural Science Foundation of China, grant number 62073004, partly by the Science and Technology Plan of Shenzhen, grant number JCYJ20200109140410340, and partly by the Shenzhen Fundamental Research Program, grant number GXWD20201231165807007-20200807164903001.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Liu, H.; Jin, Y.; Zhao, C. Real-time trust region ground plane segmentation for monocular mobile robots. In Proceedings of the 2017 IEEE International Conference on Robotics and Biomimetics (ROBIO), Macau, China, 5–8 December 2017; pp. 952–958. [Google Scholar]
  2. Li, Y.; Liu, H.; Tang, H. Multi-modal perception attention network with self-supervised learning for audio-visual speaker tracking. In Proceedings of the AAAI Conference on Artificial Intelligence, Online, 22 February–1 March 2022; Volume 36, pp. 1456–1463. [Google Scholar]
  3. Hu, P.; Held, D.; Ramanan, D. Learning to optimally segment point clouds. IEEE Robot. Autom. Lett. 2020, 5, 875–882. [Google Scholar] [CrossRef]
  4. Nowicki, M.R. Spatiotemporal calibration of camera and 3D laser scanner. IEEE Robot. Autom. Lett. 2020, 5, 6451–6458. [Google Scholar] [CrossRef]
  5. Li, X.; Zhong, Z.; Wu, J.; Yang, Y.; Lin, Z.; Liu, H. Expectation-maximization attention networks for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9167–9176. [Google Scholar]
  6. Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 3431–3440. [Google Scholar]
  7. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention—MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015, Proceedings, Part III 18; Springer: Cham, Switzerland, 2015; pp. 234–241. [Google Scholar]
  8. Xiao, X.; Lian, S.; Luo, Z.; Li, S. Weighted Res-UNet for high-quality retina vessel segmentation. In Proceedings of the 2018 9th International Conference on Information Technology in Medicine and Education (ITME), Hangzhou, China, 19–21 October 2018; pp. 327–331. [Google Scholar]
  9. Jha, D.; Smedsrud, P.H.; Riegler, M.A.; Johansen, D.; De Lange, T.; Halvorsen, P.; Johansen, H.D. Resunet++: An advanced architecture for medical image segmentation. In Proceedings of the 2019 IEEE International Symposium on Multimedia (ISM), San Diego, CA, USA, 9–11 December 2019; pp. 225–2255. [Google Scholar]
  10. Zhou, H.; Qi, L.; Huang, H.; Yang, X.; Wan, Z.; Wen, X. Canet: Co-attention network for RGB-D semantic segmentation. Pattern Recognit. 2022, 124, 108468. [Google Scholar]
  11. Zhou, W.; Yang, E.; Lei, J.; Wan, J.; Yu, L. PGDENet: Progressive guided fusion and depth enhancement network for RGB-D indoor scene parsing. IEEE Trans. Multimed. 2022, 25, 3483–3494. [Google Scholar] [CrossRef]
  12. Zhou, W.; Yang, E.; Lei, J.; Yu, L. FRNet: Feature reconstruction network for RGB-D indoor scene parsing. IEEE J. Sel. Top. Signal Process. 2022, 16, 677–687. [Google Scholar] [CrossRef]
  13. Su, Y.; Yuan, Y.; Jiang, Z. Deep feature selection-and-fusion for RGB-D semantic segmentation. In Proceedings of the 2021 IEEE International Conference on Multimedia and Expo (ICME), Shenzhen, China, 5–9 July 2021; pp. 1–6. [Google Scholar]
  14. Wang, Y.; Sun, F.; Huang, W.; He, F.; Tao, D. Channel exchanging networks for multimodal and multitask dense image prediction. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 45, 5481–5496. [Google Scholar] [CrossRef] [PubMed]
  15. Liu, H.; Zhang, J.; Yang, K.; Hu, X.; Stiefelhagen, R. CMX: Cross-modal fusion for RGB-X semantic segmentation with transformers. arXiv 2022, arXiv:2203.04838. [Google Scholar]
  16. Alonso, I.; Riazuelo, L.; Murillo, A.C. Mininet: An efficient semantic segmentation convnet for real-time robotic applications. IEEE Trans. Robot. 2020, 36, 1340–1347. [Google Scholar] [CrossRef]
  17. Hao, X.; Hao, X.; Zhang, Y.; Li, Y.; Wu, C. Real-time semantic segmentation with weighted factorized-depthwise convolution. Image Vis. Comput. 2021, 114, 104269. [Google Scholar] [CrossRef]
  18. Hu, P.; Perazzi, F.; Heilbron, F.C.; Wang, O.; Lin, Z.; Saenko, K.; Sclaroff, S. Real-time semantic segmentation with fast attention. IEEE Robot. Autom. Lett. 2020, 6, 263–270. [Google Scholar] [CrossRef]
  19. Seichter, D.; Fischedick, S.B.; Köhler, M.; Groß, H.M. Efficient multi-task RGB-D scene analysis for indoor environments. In Proceedings of the 2022 International Joint Conference on Neural Networks (IJCNN), Padua, Italy, 18–23 July 2022; pp. 1–10. [Google Scholar]
  20. Silberman, N.; Hoiem, D.; Kohli, P.; Fergus, R. Indoor segmentation and support inference from RGBD images. ECCV 2012, 7576, 746–760. [Google Scholar]
  21. Song, S.; Lichtenberg, S.P.; Xiao, J. Sun RGB-D: A RGB-D scene understanding benchmark suite. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 567–576. [Google Scholar]
  22. Cao, J.; Leng, H.; Lischinski, D.; Cohen-Or, D.; Tu, C.; Li, Y. ShapeConv: Shape-aware convolutional layer for indoor RGB-D semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 7088–7097. [Google Scholar]
  23. Chen, L.Z.; Lin, Z.; Wang, Z.; Yang, Y.L.; Cheng, M.M. Spatial information guided convolution for real-time RGBD semantic segmentation. IEEE Trans. Image Process. 2021, 30, 2313–2324. [Google Scholar] [CrossRef]
  24. Chen, X.; Lin, K.Y.; Wang, J.; Wu, W.; Qian, C.; Li, H.; Zeng, G. Bi-directional cross-modality feature propagation with separation-and-aggregation gate for RGB-D semantic segmentation. In Computer Vision—ECCV 2020: 16th European Conference, Glasgow, UK, 23–28 August 2020, Proceedings, Part XI; Springer: Cham, Switzerland, 2020; pp. 561–577. [Google Scholar]
  25. Lu, M.; Chen, Z.; Wu, Q.; Wang, N.; Rong, X.; Yan, X. FRNet: Factorized and regular blocks network for semantic segmentation in road scene. IEEE Trans. Intell. Transp. Syst. 2020, 23, 3522–3530. [Google Scholar] [CrossRef]
  26. Elhassan, M.; Huang, C.; Yang, C.; Munea, T. DSANet: Dilated spatial attention for real-time semantic segmentation in urban street scenes. Expert Syst. Appl. 2021, 183, 115090. [Google Scholar] [CrossRef]
  27. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  28. Roberts, M.; Ramapuram, J.; Ranjan, A.; Kumar, A.; Bautista, M.A.; Paczan, N.; Webb, R.; Susskind, J.M. Hypersim: A photorealistic synthetic dataset for holistic indoor scene understanding. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10912–10922. [Google Scholar]
  29. Zhou, W.; Lin, X.; Lei, J.; Yu, L.; Hwang, J.N. MFFENet: Multiscale feature fusion and enhancement network for RGB–thermal urban road scene parsing. IEEE Trans. Multimed. 2021, 24, 2526–2538. [Google Scholar] [CrossRef]
  30. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  31. Seichter, D.; Köhler, M.; Lewandowski, B.; Wengefeld, T.; Gross, H.M. Efficient RGB-D semantic segmentation for indoor scene analysis. In Proceedings of the 2021 IEEE International Conference on Robotics and Automation (ICRA), Xi’an, China, 30 May–5 June 2021; pp. 13525–13531. [Google Scholar]
  32. Romera, E.; Alvarez, J.M.; Bergasa, L.M.; Arroyo, R. Erfnet: Efficient residual factorized convnet for real-time semantic segmentation. IEEE Trans. Intell. Transp. Syst. 2017, 19, 263–272. [Google Scholar] [CrossRef]
Figure 1. Interactive Efficient Multi-Task RGB-D Semantic Segmentation Network.
Figure 1. Interactive Efficient Multi-Task RGB-D Semantic Segmentation Network.
Electronics 12 03943 g001
Figure 2. Cross-Modal Feature Rectification Module.
Figure 2. Cross-Modal Feature Rectification Module.
Electronics 12 03943 g002
Figure 3. Coordiante Attention Fusion Module.
Figure 3. Coordiante Attention Fusion Module.
Electronics 12 03943 g003
Figure 4. Quantitative results on NYUv2.
Figure 4. Quantitative results on NYUv2.
Electronics 12 03943 g004
Figure 5. Quantitative results on SUNRGB-D.
Figure 5. Quantitative results on SUNRGB-D.
Electronics 12 03943 g005
Figure 6. Qualitative results on NYUv2.
Figure 6. Qualitative results on NYUv2.
Electronics 12 03943 g006
Table 1. Quantitative results on NYUv2 and SUNRGB-D.
Table 1. Quantitative results on NYUv2 and SUNRGB-D.
MethodBackboneNYUv2SUNRGB-DFPS
PA (%)mIoU (%)PA (%)mIoU (%)
ShapeConv * [22]ResNext101 32x8d76.451.3--25.97
ShapeConv * [22]Res101--82.248.625.97
SGNet * [23]Res10176.851.182.048.6N/A
SA-Gate [24]Res5077.952.482.549.424.71
CMX * [15]SegFormer-B580.156.983.852.411.13
ESANet (pre. SceneNet) [31]Res34NBt1D-51.6-48.543.66
EMSANet (pre. Hypersim) [19]Res34NBt1D78.153.381.948.546.23
Ours (Semantic only)Res34NBt1D76.750.382.048.142.97
OursRes34NBt1D76.851.381.948.342.97
Ours (pre. Hypersim)Res34NBt1D78.854.582.349.142.97
*: additional test-time augmentation, N/A: no implementation available, pre. SceneNet: additional pretraining on SceneNet, pre. Hypersim: additional pretraining on Hypersim.
Table 2. Ablation for CM-FRM, CAFM and Backbone.
Table 2. Ablation for CM-FRM, CAFM and Backbone.
BackboneCM-FRMFusionNYUv2SUNRGB-DFPS
mIoU (%)mIoU (%)
Res34NBt1D×RGB-D Fusion49.2046.8146.18
Res34NBt1DRGB-D Fusion50.2747.8843.26
Res34NBt1D×CAFM49.4147.1145.28
Res34NBt1DCAFM50.3848.1042.97
Res18CAFM46.0445.1763.37
Res34CAFM47.0646.3654.57
Res50CAFM49.8647.0528.03
Res18NBt1DCAFM46.7446.0955.21
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, X.; Liu, J.; Liu, H. Interactive Efficient Multi-Task Network for RGB-D Semantic Segmentation. Electronics 2023, 12, 3943. https://doi.org/10.3390/electronics12183943

AMA Style

Xu X, Liu J, Liu H. Interactive Efficient Multi-Task Network for RGB-D Semantic Segmentation. Electronics. 2023; 12(18):3943. https://doi.org/10.3390/electronics12183943

Chicago/Turabian Style

Xu, Xinhua, Jinfu Liu, and Hong Liu. 2023. "Interactive Efficient Multi-Task Network for RGB-D Semantic Segmentation" Electronics 12, no. 18: 3943. https://doi.org/10.3390/electronics12183943

APA Style

Xu, X., Liu, J., & Liu, H. (2023). Interactive Efficient Multi-Task Network for RGB-D Semantic Segmentation. Electronics, 12(18), 3943. https://doi.org/10.3390/electronics12183943

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop