Next Article in Journal
CNN-Based Kidney Segmentation Using a Modified CLAHE Algorithm
Previous Article in Journal
Plug-and-Play PRNU Enhancement Algorithm with Guided Filtering
Previous Article in Special Issue
Research on Visual Perception of Speed Bumps for Intelligent Connected Vehicles Based on Lightweight FPNet
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach

1
Guangdong Provincial Engineering Research Center for Optoelectronic Instrument, School of Electronic Science and Engineering (School of Microelectronics), South China Normal University, Foshan 528225, China
2
School of Computer Science, Nanjing University of Information Science and Technology, Nanjing 210044, China
3
Inspur Group, Ji’nan 250000, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2024, 24(23), 7702; https://doi.org/10.3390/s24237702
Submission received: 29 October 2024 / Revised: 23 November 2024 / Accepted: 1 December 2024 / Published: 2 December 2024
(This article belongs to the Special Issue Intelligent Sensing and Computing for Smart and Autonomous Vehicles)

Abstract

:
Vehicle-to-vehicle communication enables capturing sensor information from diverse perspectives, greatly aiding in semantic scene completion in autonomous driving. However, the misalignment of features between ego vehicle and cooperative vehicles leads to ambiguity problems, affecting accuracy and semantic information. In this paper, we propose a Two-Stream Multi-Vehicle collaboration approach (TSMV), which divides the features of collaborative vehicles into two streams and regresses interactively. To overcome the problems caused by feature misalignment, the Neighborhood Self-Cross Attention Transformer (NSCAT) module is designed to enable the ego vehicle to query the most similar local features from collaborative vehicles through cross-attention, rather than assuming spatial-temporal synchronization. A 3D occupancy map is finally generated from the features of collaborative vehicle aggregation. Experimental results on both V2VSSC and SemanticOPV2V datasets demonstrate TSMV outpace state-of-the-art collaborative semantic scene completion techniques.

1. Introduction

Autonomous vehicles (AVs) necessitate precise ambient perception and comprehensive scene understanding to achieve secure and robust autonomous driving [1,2]. Many researchers depict the surrounding environment by using bird’s eye view (BEV) [3,4]. However, the BEV map compresses the height information and limits the fine granularity in the height dimension. Similarly, 3D object detection [5,6] assigns fixed-size bounding boxes to objects, which is difficult to deal with general and extra-lexical objects. Semantic scene completion (SSC) [7,8,9], which can estimate the occupancy state and semantics of the scene, has become an emerging task to overcome these limitations. SSC can provide occupancy for each small unit, rather than assigning bounding boxes, which is conducive to identifying irregular obstacles and less limiting the shape and category of objects. 3D SSC can represent not only the spatial information on the ground plane but also the height dimension, which enhances its capability on rugged roads. Semantic comprehension has a direct impact on subsequent tasks, such as object detection and segmentation [10], which are essential to autonomous driving.
A basic challenge of SSC is to obtain comprehensive information from incomplete observations, often necessitating the model to address occlusion and thoroughly investigate the obscured portion of the scene. Due to the complexity of the occlusion problem, relying solely on single-view sensors may not be the ultimate solution. The existing works [11,12] usually use LiDAR points as input. They aggregate Lidar points from multiple frames to provide rich point cloud information and compensate for long-distance sparsity and partial occlusion problems by adding time-continuous information. However, the fusion of Lidar frames also introduces motion blur, which makes the edges of environmental objects blurred and cannot be accurately identified, suppressing the advantages of a single frame in accurate object positioning. DynStatF [13] proposed a two-stream architecture for the first time to extract features from single-frame and multi-frame LiDAR inputs and interactively fuse them. However, when facing scenes such as intersections, these methods will fail due to the lack of observation from others’ perspectives.
The Internet of Vehicles (IoV) [14] has the potential to ensure driving safety and optimize driving behavior [15]. With the advancement of communication technology, vehicle-to-vehicle (V2V) communication technology [16,17] has gradually matured. Semantic scene completion with V2V provides observation from different perspectives in the same scene which makes it possible to overcome the limitations of a single-vehicle SSC. Recent research for V2V collaboration [18,19,20] exploit the advantages of multiple perspectives. Previous multi-vehicle cooperation works can be divided into three strategies: early, intermediate, and late fusion. Early fusion [21] transmits the original sensor data while late fusion [22] transmits the detection results. The former requires a large bandwidth, and the latter needs additional manual operators to fuse results, such as non-maximum suppression due to the problems of overlap and ambiguity. The fundamental reason for overlap and ambiguity lies in the misalignment of features among cooperative vehicles, leading to boundary ambiguity such as bloated vehicles. This problem is directly reflected in the late fusion and inevitably appears in the other strategies. The last strategy is intermediate fusion, which transmits intermediate features [19,23], performing feature fusion through self-attention. These methods usually have high computational complexity. To solve this problem, self-attention will be restricted to the corresponding position within the feature map [24,25], or the window [26] will be divided to reduce the secondary complexity of attention. However, the features of the corresponding positions of different vehicle feature maps do not strictly belong to the same position.
Therefore, to address the aforementioned issue of the limitation of self-attention and the problem of feature misalignment, we propose a Two-Stream Multi-Vehicle collaboration approach (TSMV), which is the first framework to generate 3D semantic occupancy maps using the collaboration of ego stream and fuse stream. The ego stream pertains to the local environment of the ego vehicle, whereas the fuse stream relates to the comprehensive, interconnected environment of all collaborative vehicles. The two-stream method allows for retaining the relatively accurate environmental boundary information from the ego stream while learning global scene information in the fuse stream, thus alleviating the issue of boundary ambiguity. We believe that the relevant features of the same object should be in the vicinity of the neighborhood. The proposed NSCAT module computes self attention and cross attention by constraining the neighborhood, effectively limiting computational costs while enhancing the features of the two-stream. Our experiments are carried out using the V2VSSC and SemanticOPV2V datasets, and TSMV achieves top performance on collaborative SSC tasks. In conclusion, our work has three primary contributions:
  • A novel collaboration approach called TSMV was proposed for collaborative semantic scene completion, which includes a two-stream architecture to effectively fuse features from collaborative vehicles. TSMV alleviates the boundary ambiguity problem caused by feature misalignment in multi-vehicle cooperative perception.
  • A NSCAT module was proposed by combining self-attention and cross-attention transformer. NSCAT recurrently aggregates features of two streams through local and global interaction.
  • Experiments on both V2VSSC and SemanticOPV2V datasets show that TSMV achieves improved performance compared to the prior arts in both LiDAR-based and Camera-based methods.

2. Related Work

2.1. Semantic Scene Completion

Semantic occupancy perception originated from SUNCG [27]. The algorithm indicates that learning semantic segmentation and scene completion can be mutually beneficial. Subsequently, more and more attention has been paid to the work of indoor small-scale fixed scenes, such as NYUCAD [28], ScanNet [29] and PALNet [30]. With the success of using dense point clouds or depth maps in small indoor scenes on SSC tasks, researchers have begun to explore dynamic large-scale outdoor scenes. SFSU [31] introduces semantic foggy scene understanding into foggy driving scenes. Ref. [32] proposes an anisotropic network, which decomposes 3D convolution into three consecutive dimensional convolutions, and further enhances modeling capability by stacking multiple kernel-size selection convolutions and kernel modulation convolutions. SemanticKITT [33] extends occupancy perception to driving scenarios, but it only evaluates front-view occupancy prediction. Towards a comprehensive study of surrounding perception algorithms, Openccupancy [34] proposes the first surrounding SSC benchmark. Furthermore, A cascade occupancy network was proposed to refine the coarse prediction dealing with the complexity of high-resolution 3D predictions. SSCBench [35] integrates well-known datasets (e.g., KITTI-360, nuScenes, and Waymo) in autonomous driving to generate SSC benchmarks. V2VSSC [25] is a large-scale V2V 3D semantic occupancy dataset collected by CARLA and collaborative driving automation tool OpenCDA [36]. It can provide observations of multiple vehicle perspectives in the same scene from the simulation server, making up for the limitations of single-view semantic occupancy perception. SemanticOPV2V [37] equips vehicles with semantic LiDAR to obtain 12 different 3D semantic occupancy labels.

2.2. Multi-Vehicle Perception

Sharing collective perception messages [38] including GPS and sensor data (lidar, camera, etc.) between vehicles helps to make up for the limitations of current visual perception technologies, such as limited resolution, weather effects, and occlusions. AVs can exchange traffic information with surrounding cooperative vehicles to achieve cooperative perception and improve the safety and robustness of autonomous driving systems. Ref. [39] proposed a novel neural network V2VNet using a spatial perception map for multi-vehicle feature aggregation. FCooper [40] uses a simple max-out operation to fuse features. OPV2V [24] fuses all the features by point-to-point attention at the same location in the feature map. DiscoNet [41] uses knowledge distillation to limit the corresponding features to enhance training. Ref. [42] applies an adaptive feature-level fusion model to fuse multi-vehicle data in simultaneous localization and mapping tasks. V2VFomer++ [43] construct the first multi-model multi-vehicles cooperative perception framework, where individual camera-LiDAR representation is incorporated with dynamic channel fusion module at BEV space. EdgeCooper [44] combines the individual views of the vehicle to form an overall view with higher resolution and proposes a two-dimensional graph coloring algorithm to eliminate conflicts. CoBEVT [45] proposed fused axial attention (FAX) to capture local and global information between different vehicles. V2XViT [26] proposed a novel heterogeneous multi-agent self-attention module (HMSA), which can adaptively integrate heterogeneous agents, such as vehicles and vehicles or vehicles and infrastructure. DAIR-V2X [46] proposed a Time Compensation Late Fusion (TCLF) framework to alleviate the problem of time asynchrony. Lu et al. [47] proposed a scalable heterogeneous cooperative sensing framework Heterogeneous Alliance (HEAL), which uses multi-scale foreground perception to establish a unified feature space.

2.3. Transformer in Feature Fusion

Transformer was first proposed for machine translation [48] to capture remote interactions between words by stacking multi-headed self-attention and feed-forward layers. MorphText [49] creates reliable connections between text segments. Dos. A. et al. [50] proposed ViT for image recognition tasks, which directly applies self-attention by cutting image blocks and adding position embedding. The Transformer architecture’s ability to fuse features from different sources or patterns has been verified in previous work. Emerging work series [51,52] integrates multi-modal fusion of images and texts. TransFusion [53] proposed a two-stage Transformer framework. The first stage uses LiDAR features to obtain a set of sparse object queries, and the second stage uses another transformer structure to obtain cross-attention of camera and lidar data. The deformable DETR [54] extracts the cross-attention of the current frame and the previous historical frame. Li et al. [4] used deformable DETR to obtain spatial-temporal attention between multiple images. Although it has global interaction ability, the computational complexity is linearly related to image resolution, and high-resolution images will lead to a quadratic increase in model complexity and memory usage. To solve this problem, many methods introduce locality into self-attention. In SASA [55], each pixel only focuses on the self-attention of a window around it. Swim [56] proposed W-MSA and SW-MSA modules to reduce the secondary complexity and achieve global information interaction. The W-MSA module reduces the amount of calculation by dividing the window, and the SW-MSA solves the problem that information cannot be exchanged between different windows by shifting the window. CSwin [57] proposed a cross-shaped window attention to achieve parallel computing of horizontal and vertical self-attention. Ali et al. [58] proposed a flexible and efficient explicit sliding window attention mechanism, called neighborhood attention. It locates the attention range of each pixel to its nearest field. DynStatF [13] uses a neighborhood attention mechanism to capture the cross-attention of multi-frame and single-frame LiDAR. Different from the previous work, to preserve the spatial correlation, our approach uses the transformers of neighborhood attention to obtain the cross-attention between the LiDAR input features of the cooperative vehicles and the ego vehicle. CoHFF [37] achieves advanced camera-based collaborative semantic occupancy prediction technology through feature sharing in V2X communication networks.

3. Methodology

3.1. Overall Architecture

The overall system framework of TSMV is illustrated in Figure 1. First, a spatial graph will be constructed on the ego vehicle using V2V communication technology with surrounding vehicles. Following the voxel-based LiDAR detector, we extract the features of the vehicles in the spatial graph and transform them to voxel spatial features F v . Then the voxel spatial features are compressed using 3D convolution. The height dimension is compressed to the channel dimension and voxel space features F v is converted into intermediate features F m . Subsequently, the nearby AVs broadcast intermediate features F m using the designated method for sharing messages. When the ego vehicle receives the broadcast information, features of all vehicles in the spatial graph will be divided into ego stream and fuse stream. Feature fusion between two streams occurs regressively using the neighborhood self-cross attention module (NSCAT) in our TSMV. Finally, the detection head performs a series of lightweight convolutional operations and applies a reshape operation along the channel dimension to restore the height dimension from the channel dimension. The 3D semantic occupancy map is generated by the detection head.

3.1.1. Spatial Graph Construction

The role of the space graph is to transmit extrinsic and relative positions to each cooperative vehicle. Since the vehicle is moving, the spatial graph will be updated dynamically in real-time. For an ego-vehicle, it searches N [ 1 , 7 ] neighboring vehicles within the communication range 70 m as the center around and constructs a spatial graph. Each vehicle within the communication range is represented by a node in the spatial graph, with each communication channel connecting the ego and collaborative nodes being represented by an edge. The node will be deleted from the spatial graph when it exceeds the communication range. On the contrary, when the node enters the communication range, it will be added to the spatial graph to realize the dynamic perception of nearby vehicles by ego vehicle.

3.1.2. Feature Extraction and Compression

The point clouds of neighboring vehicles are projected into the LiDAR coordinate system of the ego vehicle using shared location information. Following the voxel-based LiDAR detection setting, we use VoxelNet [59] to extract features. The projection point clouds will be converted to voxel space features, represented as F v R C × D × H × W , in which C represents the dimension of voxel feature, D / H / W represents the size of voxel grid map along z / y / x direction. The transmission bandwidth of hardware imposes limitations on V2V communication. Due to the high bandwidth requirements for transmitting 3D high-dimensional feature maps, it is essential to compress these feature maps. To further reduce computational burden and the pressure of communication bandwidth, inspired by [60], we use 3D convolution to compress F v and concat the height dimension to the channel dimension, which F m R C × H × W . The compression of height information into channel dimensions can be viewed as a form of BEV which reduces the communication delay and calculation. The compressed intermediate features F m of all vehicles will be fed into the two-stream multi-vehicle collaboration module for fusion, which outputs the final aggregated features F o u t . The two-stream multi-vehicle collaboration module is introduced in Section 3.2.

3.1.3. Head and Losses

The aggregated features F o u t will be fed to the detection head to generate occupancy probability and semantic category scores. Similar to the Unet structure, we perform a series of lightweight convolutional operations on F o u t , which include several downsampling convolution and upsampling deconvolution layers. Inspired by FlashOcc [60], we perform a simple reshaping operation in the channel dimension to transform the feature map from shape B , C , H , W to B , C , Z , H , W , where B represents batch size, C represents channel number in F o u t , Z , H , W represent the resolutions in the z / y / x dimensions of 3D space, C is set to 1 in occupancy prediction, while to the number of semantic category in semantic prediction and C = C × Z . Finally, we restore the height dimension from the channel dimension and output the final 3D semantic occupancy map.
To train the proposed approach, we leverage geometric loss L g and semantic loss L s to optimize the network. The geometric loss is computed using binary cross-entropy (BCE) loss, indicating whether the voxel position is occupied, while the semantic loss is calculated using cross-entropy (CE) loss, representing whether each unit of voxel is correctly classified semantic category. The general loss function can be derived as:
L = λ g L g + λ s L s
where λ g and λ s represent the weight of geometric loss and semantic loss respectively.

3.2. Two-Stream Multi-Vehicle Collaboration

The fusion of intermediate features from multi-vehicle requires local and global interactions across the spatial location. We divide the intermediate feature of the ego vehicle and the received vehicles’ intermediate features into two streams, called ego stream and fuse stream. The ego stream contains only the intermediate feature of the ego vehicle, denoted as F e R C × H × W . Given the input intermediate features { F m i | i 1 N } in the spatial graph, N represents the total number of vehicles, i represents the index of vehicles in the spatial graph, especially F m 1 represents F e . Formally, let F g R N × C × H × W be the stacked intermediate features from N vehicles. We map F g to fuse stream features F f R C × H × W by N-dimensional addition:
F f = a d d ( F g , d i m = N )
where add represents features addition. Then, TSMV utilizes the Neighborhood Self-Cross Attention Transformer module (NSCAT) to regress features from the ego local stream to learn features from the fuse global stream. This module is inspired by [13,45,58] and combines self-attention and cross-attention within local windows to enhance features. In contrast to V2VSSC, which applies self-attention to regions at the same position in space to capture representative features. However, simply weighting features of the same spatial position will greatly limit spatial correlation. When faced with issues such as sensor noise, communication delays, localization errors, etc., leading to performance decline. To preserve spatial correlations, we believe that relevant features of the same environmental object should be located in similar positions and neighboring regions in a two-stream features map. Therefore, we establish cross-attention in the intermediate feature for local information nearing a specific query. Furthermore, the intermediate feature map for segmentation tasks is relatively large, limiting the field size helps save computational costs. We apply neighborhood attention(NA), which is a restricted self-attention method used in image classification. Compared to attention methods like Swin, NA achieved efficient GPU acceleration and ran faster while using less memory. Moreover, we extend NA from self-attention to cross-attention to gather information in the local perception field.
Neighborhood Self-Attention (NSA) and Neighborhood Cross-Attention (NCA) can be represented as:
Q f = X f W q , K f = X f W k , V f = X f W v N S A = s ( Q f K f T v ) V f
Q e = X e W q , K f = X f W k , V f = X f W v N C A = s ( Q e K f T v ) V f
where X represents neighborhood token features and subscripts represents different streams, W represntes trainable parameter matrix, v represents feature dimension scale, s represents softmax. Specifically, taking the extended neighborhood cross-attention as an example, the convolutional layer is first used to mark the intermediate feature map, which is expressed as a token of m-dimensional feature vectors. For example, the token feature for the ego stream can be expressed as X e t o k e n R n × m . Token Feature X e t o k e n in ego stream are linearly projected to the query Q e R n × q , while the token features in the fuse stream are also linearly projected to the key K f R n × q and value V f R n × v . For the i-th token, the attention weight with neighborhood size of k ( A i k ) is:
A i k = Q e K f ρ 1 ( i ) T + B ( i , ρ 1 ( i ) ) Q e K f ρ 2 ( i ) T + B ( i , ρ 2 ( i ) ) Q e K f ρ k ( i ) T + B ( i , ρ k ( i ) )
where ρ j ( i ) represents the j-th nearest neighbor of the i-th token, B ( i , ρ j ( i ) ) represents relative positional biases between j-th nearest neighbor and i-th token. The attention value can be expressed as:
V i k = V f ρ 1 ( i ) T V f ρ 2 ( i ) T V f ρ k ( i ) T
so the l-th token cross attention is defined as:
N C A k ( i ) = s o f t m a x ( A i k ) v V i k
where v represents the feature dimension scale. We compute the cross attention for each pixel according to the method outlined previously. Similarly, the pixel-by-pixel fuse feature can be obtained by calculating the self-attention of Q K V from the fuse stream. The illustration of the self and cross attention are shown in Figure 2.
We combine two kinds of attention with the classic design of Transformers, including Layer Normalization (LN), Multilayer Perceptrons (MLPs), Feedforward Neural Networks (FFN), and skip-connections as exemplified in Figure 3. In the l-th block ( l 1 , 2 , . . . , L ), given the intermediate feature from ego stream F e R C × H × W and feature from fuse stream F f R C × H × W , the output of the NSCAT module can be expressed as:
F ^ f l = α l ( L N ( F f l 1 ) ) + F f l 1
F ^ e l = β l ( L N ( F e l 1 ) , L N ( F f l 1 ) ) + F e l 1
F f l = M L P ( L N ( F ^ f l ) ) + F ^ f l
F e l = M L P ( L N ( F ^ e l ) ) + F ^ e l
where α l ( · ) represents NSA module and β l ( · ) represents NCA module. After all L numbers of NSCAT blocks, we add the intermediate features of the two streams together to sufficiently consolidate features.
Considering that the intermediate features from the ego stream provide accurate environmental perception, while the fuse stream provides rich long-range perception and environmental perception of partially occluded regions from the ego stream, the NSCAT module utilizes the features in ego stream F e l as queries, generating keys and values from fuse stream feature F f l to achieve cross attention and enhance the fuse stream features F f l using the neighborhood self-attention to optimize of the feature map through collaboration. Finally, the outputs of the NSCAT module are added to obtain the final features F o u t , and F o u t is fed into the detection head.

4. Experiments

4.1. Datasets and Evaluation Metrics

We conduct experiments on two challenging collaborative autonomous driving datasets, namely V2VSSC dataset and SemanticOPV2V dataset. Both datasets are the extension of the groundbreaking OPV2V dataset [24] which contains over 70 different scenarios. In each scenario, different numbers (2 to 7) of vehicles appear at the same time. Therefore, both datasets are the multi-vehicle collaborative SSC dataset with 6764 training sequences, 1980 validation sequences and 2170 test sequences.
The V2VSSC dataset [25] is a LiDAR-based multi-vehicle collaborative SSC dataset. Each sample consists of a LiDAR sensor and a GNSS/IMU sensor. For the SSC task, the V2VSSC dataset procures precise occupancy and semantic details directly from the server. There are 10,914 annotated semantic occupancy maps with 6 different semantic labels. For the evaluation metrics, The V2VSSC dataset utilizes Intersection of Union (IoU) [33] as the geometric metric, the mean IoU (mIoU) overall categories are utilized as the semantic metric. The IoU and mIoU can be derived as:
I o U = T P T P + F P + F N
m I o U = 1 C c = 1 C T P c T P c + F P c + F N c
where T P , F P , F N correspond to the number of true positive, false positive, and false negative predictions, the subscript ∘ is denoted for occupied voxels while c for categories respectively. In addition, the critical objects of autonomous driving encompass road and vehicles. Therefore, cIoU is defined as the mean of these two categories.
The SemanticOPV2V dataset [37] is a Camera-based multi-vehicle collaborative SSC dataset. Each sample consists of RGB images from 4 cameras with 360 horizontal FOV and a GNSS/IMU sensor. SemanticOPV2V equips each CAV with 4 semantic LiDARs and associates the ground truth data of all CAVs to capture an accurately collaborative semantic ground truth, thus integrating 12 different 3D semantic occupancy labels. We utilize the metric Intersection Over Union for the evaluation of each category and the mean Intersection Over Union (mean IoU) on the SemanticOPV2V dataset.

4.2. Implementation Details

Following previous methods [25,37], we adopt Voxelnet [59] as LiDAR-based backbone and ResNet101 [61] as Camera-based backbone. For experiments on V2VSSC, the perception ranges are [−50 m, 50 m] for the x, y axis and the size of resolution of the occupancy map is 0.78 m while the perception range of height is [−3 m, 5 m] with the resolution of 0.4 m, so the occupancy map resolution is 20 × 128 × 128 . We adopt Adam optimizer [62] with an initial learning rate of 2 × 10 3 and employ cosine anneal warm learning rate scheduler with an initial learning rate of 2 × 10 4 , η m i n = 2 × 10 5 . We conduct training for all models with 30 epochs. Besides, we set up two evaluation models, (1) Perfect setting, in which the positioning is accurate, and all transmission data is synchronized and real-time. (2) Noisy setting, considering the positioning inaccuracies and time delay. Specifically, following the real-world noise levels [63,64], the positional and heading noise are taken from the Gaussian distribution with a default standard deviation of 0.2 m and 0 . 4 respectively. In addition, the time delay of all evaluation models is set to 100 ms which reflects the usual transmission delay in V2V communication system [65].
For experiments on SemanticOPV2V, we utilize a 40 × 40 × 3.2 m detection area with a grid size of 100 × 100 × 8 , resulting in a voxel size of 0.4 m3. We add the NSCAT module to the V2X feature fusion module in CoHFF [37] to implement a camera-based TSMV method and we set the share features with a length of 64 for Feature Fusion. Following the settings in CoHFF [37], we pretrained Occupancy Network, Semantic Segmentation Network and finally trained Hybrid Feature Fusion Network with pretrained network.

4.3. Comparison

4.3.1. LiDAR-Based Results on V2VSSC

For multi-vehicle perception tasks, we compare our proposed model with no-fusion baseline, early and late fusion, as well as various state-of-the-art intermediate fusion LiDAR-based models including FCooper [40], V2VSSC [25], and V2XViT [26].
In Table 1, we report the results on the V2VSSC test set in perfect settings. Under perfect setting, all collaboration methods surpass the no-fusion method, which proves the effectiveness of multi-vehicle collaboration in SSC tasks. Among all fusion models, TSMV outperforms all other methods in IoU, mIoU, and cIoU metrics, while achieving the best in most categories, which demonstrates the effectiveness of TSMV in occupancy and semantic prediction. Specifically, for the comprehensive metric mIoU, it surpasses the second-best by 2%. Moreover, it outperforms the second-best by 2.2% in the car category metric that is most relevant to autonomous driving, with a 2.9% higher accuracy in the cIoU metric including both road and car metrics.
We also conduct experiments under noisy settings on V2VSSC, as shown in Table 2. We fine-tuned the models from the perfect setting. Considering position errors and time delays, the effectiveness of all fusion models shows a decline. The performance of the late fusion method shows a sharp decline, even worse than the no-fusion method. Specifically, for the IoU and mIoU metrics, except for the late fusion method, all other methods are ahead of no-fusion, with TSMV remaining the best performance. Due to AVs being in highly dynamic environments, there was a significant performance decrease across all models in the cIoU metric. Except for TSMV maintaining a lead of 4.4%, the performance of the other methods was either comparable to or worse than the no-fusion baseline method, which proves that our proposed method has certain anti-interference capability and is applicable in normal noisy environments.

4.3.2. Camera-Based Results on SemanticOPV2V

We conduct experiments on SemanticOPV2V val set, as shown in Table 3. Following [37], we calculate IoU for each individual category and the mean IoU across all categories for occupancy and semantic occupancy prediction.
When focusing solely on binary occupancy prediction(as shown at Occ. Pred. in Table 3), we compare the method with Raw LiDAR baseline and CoHFF [37]. It is observed that TSMV has achieved the highest mean IoU metric and has led in many categories. In the final evaluation of semantic occupancy prediction(as shown at Sem. Occ. Pred. in Table 3), we further demonstrate the benefits brought by two-stream architecture. Our method achieves state-of-the-art results on the camera-based SSC task with an overall mean IoU increase of approximately 2.32%. As a result, the TSMV method can strengthen collaborative semantic scene completion tasks.

4.4. Run Time Efficiency

The real-time performance is displayed in Table 4. Notably, the time we calculate here only includes inference time on the ego vehicle. We tested the real-time performance on NVIDIA RTX 3060Ti. The results demonstrate that the proposed approach achieves competitive running efficiency and state-of-the-art model performance.

4.5. Detection Visualization

Figure 4 shows the ego vehicle’s front view (a), semantic occupancy ground truth (b), and the detection visualization of No Fusion (c), V2VSSC (d), TSMV (e) method. Our model can predict the occupancy of 3D semantic scenes more accurately and align with the ground truth better. Compared with the method of no-fusion, the multi-vehicle cooperative method can better deal with some occluded areas. For example, where the red box is marked in the picture, from the front view, we can not observe these AVs, but the different perspectives brought by other vehicles can solve the problem of single-view occlusion. TSMV can more accurately describe the boundaries of each category, such as the blue box marked out, no-fusion, and V2VSSC method will show adhesion to adjacent vehicles.
Figure 5 presents visual results of the CoHFF and TSMV model in SemanticOPV2V, which are compared with the ground truth data from multiple perspectives. All models can accurately predict various categories of voxels, such as roads, sidewalks, and vehicles. Our TSMV is better than the CoHFF model in terms of overall building prediction(red voxel in the picture) and is closer to the ground truth.

4.6. Ablation Studies

4.6.1. Contribution of Major Components

We investigate the effectiveness of each component of TSMV on the V2VSSC dataset. Ablation experiments are performed on the two-stream architecture and NSCAT module proposed in this study. Our base model is a single vehicle baseline that only uses the ego stream and we compare the method that only uses the fuse stream. Then, we assess the influence of each component by incrementally incorporating a two-stream structure and an NSCAT module. As shown in Table 5, the proposed two-stream architecture and NSCAT module are conducive to performance improvement which proves the effectiveness of the two components. The two-stream architecture and NSCAT module increase mIoU by 2.1% and 2.7% respectively.

4.6.2. Impact of the Kernel Size in the Neighborhood

Here we examine the impact of the kernel size of the neighborhood on TSMV in the V2VSSC dataset. As outlined in Table 6, enlarging the kernel size can effectively expand the global receptive field, particularly benefiting larger object categories (e.g., Road, Terrian, and Vegetable). Conversely, reducing the kernel size enhances the precision in localizing smaller object categories (e.g., Car and Pole), leading to improved performance. Thus, the choice of the appropriate kernel size of the neighborhood necessitates consideration of both global and local receptive fields. Finally, we choose nine as the kernel size to better balance the accuracy of each category.

4.6.3. Performance of Positional Error

We quantify the level of noise via positional errors separately. The Gaussian distribution is utilized for sampling noise with the standard deviation σ x y z [ 0 , 0.4 ] m. As illustrated in Figure 6, TSMV achieves a significant advantage over other models when the positional discrepancy remains within the normal range ( σ x y z 0.2 m [63,64]). Even with substantial noise ( σ x y z = 0.4 m, twice higher than the maximum normal noise value [63,64]), TSMV outperforms the no-fusion baseline model, while the remaining models fall below the performance of the no-fusion baseline model, indicating our model possesses certain anti-interference capabilities.

5. Discussion

In the previous section, this study analyzed five experiments conducted to evaluate the performance of our proposed collaboration approach, TSMV. The following section is delved into these experimental results’ theoretical and practical significance. It will also discuss this study’s limitations and future research directions.
The results of Experiment One demonstrates that TSMV performs well across all metrics in both perfect and normal noisy environments, highlighting the potential of the two-stream architecture in feature fusion and the reliability of TSMV in normal noisy settings. Experiment Two demonstrates that TSMV can enhance collaborative semantic scene completion tasks in camera-based approaches. Experiment Three demonstrates that our proposed method balances performance and efficiency. Experiment Four, through visualization, illustrates the superiority of the multi-vehicle collaborative method in semantic scene completion, with TSMV providing more precise object localization and semantic segmentation within the collaborative approaches. Experiment Five demonstrates the effectiveness of components in TSMV. The two-stream architecture improves performance, with further enhancements from the NSCAT module. Experiment Six demonstrates the neighborhood size also impacts the performance. Larger neighborhoods improve performance for larger object categories, while smaller neighborhoods benefit smaller object categories by concentrating receptive fields within a smaller spatial range. Thus, selecting an appropriate neighborhood size can better balance performance across different categories. Experiment seven, through the quantification of noise, illustrates our proposed method presents a certain anti-interference capacity within a certain range of noise. These results not only validate the feasibility of our proposed method but also strongly support the applicability of collaborative perception in semantic scene completion in autonomous driving scenarios. The experimental results of TSMV not only enhance our understanding of the performance of collaborative perception in semantic scene completion but also guide the collaborative perception feature fusion method.
Despite significant advancements, TSMV has some limitations. The framework relies heavily on robust V2V communication, which may face reliability challenges with poor connectivity or network failures. To validate the influence of poor connectivity, we randomly delay signals obtained from other vehicles by different time ranges. The results are shown in Table 7. Network failure dramatically degrades the model performance for all V2V methods, but is still better than the no-fusion baseline. In addition, the extreme noise results in positioning errors, which leads to feature misalignment and restricts its application in highly adverse circumstances. Despite a considerable performance decline in extreme noise environments, the performance of TSMV can still yield the best result (and above the no-fusion setting) within a certain noise range. We will carry out further optimization in the future.

6. Conclusions

In this paper, we propose a novel two-stream multi-vehicle collaboration approach (TSMV) to address the ambiguity issue caused by feature misalignment between cooperative vehicles. The key component is a Neighborhood Self-Cross Attention Transformer module (NSCAT), which enables interactive learning and fusion of two-stream features at a reduced time cost. Features learn global environmental information from the fuse stream while retaining the accurate environmental boundary information from the ego stream which alleviates the boundary ambiguity issue and achieves collaborative perception. Comprehensive experiments in collaborative V2VSSC and SemanticOPV2V datasets showed that TSMV improves the performance of collaborative SSC tasks and achieves state-of-the-art results. In the future, we plan to test and validate our method in real-world scenarios to further evaluate its stability and robustness.

Author Contributions

Conceptualization, J.L. and Y.Z.; methodology, J.L.; software, J.H.; validation, J.L., Y.Z. and J.H.; formal analysis, J.L.; investigation, Y.Z.; resources, K.L.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, J.H., K.L. and P.H.; visualization, J.L.; funding acquisition, Y.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Startup Foundation for Introducing Talent of NUIST (2024r026), and the Shandong Province Science Foundation for Youths (ZR2024QF128).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data available on request due to restrictions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Fayyad, J.; Jaradat, M.A.; Gruyer, D.; Najjaran, H. Deep Learning Sensor Fusion for Autonomous Vehicle Perception and Localization: A Review. Sensors 2020, 20, 4220. [Google Scholar] [CrossRef] [PubMed]
  2. Wang, R.; Luo, X.; Ye, Q.; Jiang, Y.; Liu, W. Research on Visual Perception of Speed Bumps for Intelligent Connected Vehicles Based on Lightweight FPNet. Sensors 2024, 24, 2130. [Google Scholar] [CrossRef]
  3. Wang, J.; Li, F.; An, Y.; Zhang, X.; Sun, H. Toward Robust LiDAR-Camera Fusion in BEV Space via Mutual Deformable Attention and Temporal Aggregation. IEEE Trans. Circuits Syst. Video Technol. 2024, 34, 5753–5764. [Google Scholar] [CrossRef]
  4. Li, Z.; Wang, W.; Li, H.; Xie, E.; Sima, C.; Lu, T.; Qiao, Y.; Dai, J. Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022; pp. 1–18. [Google Scholar]
  5. Liang, T.; Xie, H.; Yu, K.; Xia, Z.; Lin, Z.; Wang, Y.; Tang, T.; Wang, B.; Tang, Z. Bevfusion: A simple and robust lidar-camera fusion framework. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), New Orleans, LA, USA, 28 November–9 December 2022; Volume 35, pp. 10421–10434. [Google Scholar]
  6. Yan, Y.; Mao, Y.; Li, B. SECOND: Sparsely Embedded Convolutional Detection. Sensors 2018, 18, 3337. [Google Scholar] [CrossRef]
  7. Rist, C.; Emmerichs, D.; Enzweiler, M.; Gavrila, D. Semantic Scene Completion Using Local Deep Implicit Functions on LiDAR Data. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 7205–7218. [Google Scholar] [CrossRef]
  8. Zhang, Z.; Han, X.; Dong, B.; Li, T.; Yin, B.; Yang, X. Point Cloud Scene Completion with Joint Color and Semantic Estimation from Single RGB-D Image. IEEE Trans. Pattern Anal. Mach. Intell. 2023, 45, 11079–11095. [Google Scholar] [CrossRef]
  9. Wilson, J.; Song, J.; Fu, Y.; Zhang, A.; Capodieci, A.; Jayakumar, P.; Barton, K.; Ghaffari, M. MotionSC: Data Set and Network for Real-Time Semantic Mapping in Dynamic Environments. IEEE Robot. Autom. Lett. 2022, 7, 8439–8446. [Google Scholar] [CrossRef]
  10. Park, J.; Yoo, H.; Wang, Y. Drivable Dirt Road Region Identification Using Image and Point Cloud Semantic Segmentation Fusion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 23, 13203–13216. [Google Scholar] [CrossRef]
  11. Meyer, G.P.; Charland, J.; Pandey, S.; Laddha, A.; Gautam, S.; Vallespi-Gonzalez, C.; Wellington, C.K. LaserFlow: Efficient and Probabilistic Object Detection and Motion Forecasting. IEEE Robot. Autom. Lett. 2021, 6, 526–533. [Google Scholar] [CrossRef]
  12. Yuan, Z.; Song, X.; Bai, L.; Wang, Z.; Ouyang, W. Temporal-Channel Transformer for 3D Lidar-Based Video Object Detection for Autonomous Driving. IEEE Trans. Circuits Syst. Video Technol. 2022, 32, 2068–2078. [Google Scholar] [CrossRef]
  13. Rong, Y.; Wei, X.; Lin, T.; Wang, Y.; Kasneci, E. DynStatF: An Efficient Feature Fusion Strategy for LiDAR 3D Object Detection. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Vancouver, BC, Canada, 17–24 June 2023; pp. 3238–3247. [Google Scholar] [CrossRef]
  14. Ullah, I.; Ali, F.; Khan, H.; Khan, F.; Bai, X. Ubiquitous computation in internet of vehicles for human-centric transport systems. Comput. Hum. Behav. 2024, 161, 108394. [Google Scholar] [CrossRef]
  15. Naeem, H.M.Y.; Bhatti, A.I.; Butt, Y.A.; Ahmed, Q.; Bai, X. Energy Efficient Solution for Connected Electric Vehicle and Battery Health Management Using Eco-Driving Under Uncertain Environmental Conditions. IEEE Trans. Intell. Veh. 2024, 9, 4621–4631. [Google Scholar] [CrossRef]
  16. Nardini, G.; Virdis, A.; Campolo, C.; Molinaro, A.; Stea, G. Cellular-V2X Communications for Platooning: Design and Evaluation. Sensors 2018, 18, 1527. [Google Scholar] [CrossRef] [PubMed]
  17. Ku, Y.J.; Baidya, S.; Dey, S. Uncertainty-Aware Task Offloading for Multi-Vehicle Perception Fusion Over Vehicular Edge Computing. IEEE Trans. Veh. Technol. 2023, 72, 14906–14923. [Google Scholar] [CrossRef]
  18. Cui, G.; Zhang, W.; Xiao, Y.; Yao, L.; Fang, Z. Cooperative Perception Technology of Autonomous Driving in the Internet of Vehicles Environment: A Review. Sensors 2022, 22, 5535. [Google Scholar] [CrossRef]
  19. Li, Y.; Ma, D.; An, Z.; Wang, Z.; Zhong, Y.; Chen, S.; Feng, C. V2X-Sim: Multi-Agent Collaborative Perception Dataset and Benchmark for Autonomous Driving. IEEE Robot. Autom. Lett. 2022, 7, 10914–10921. [Google Scholar] [CrossRef]
  20. Malik, S.; Khan, M.A.; El-Sayed, H. Collaborative Autonomous Driving—A Survey of Solution Approaches and Future Challenges. Sensors 2021, 21, 3783. [Google Scholar] [CrossRef]
  21. Chen, Q.; Tang, S.; Yang, Q.; Fu, S. Cooper: Cooperative Perception for Connected Autonomous Vehicles Based on 3D Point Clouds. In Proceedings of the IEEE 39th International Conference on Distributed Computing Systems (ICDCS), Dallas, TX, USA, 7–10 July 2019; pp. 514–524. [Google Scholar] [CrossRef]
  22. Rawashdeh, Z.Y.; Wang, Z. Collaborative Automated Driving: A Machine Learning-based Method to Enhance the Accuracy of Shared Information. In Proceedings of the International Conference on Intelligent Transportation Systems (ITSC), Maui, HI, USA, 4–7 November 2018; pp. 3961–3966. [Google Scholar] [CrossRef]
  23. Hu, Y.; Fang, S.; Lei, Z.; Zhong, Y.; Chen, S. Where2comm: Communication-efficient collaborative perception via spatial confidence maps. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 9–14 December 2024. [Google Scholar]
  24. Xu, R.; Xiang, H.; Xia, X.; Han, X.; Li, J.; Ma, J. Opv2v: An open benchmark dataset and fusion pipeline for perception with vehicle-to-vehicle communication. In Proceedings of the International Conference on Robotics and Automation (ICRA), Philadelphia, PA, USA, 23–27 May 2022; pp. 2583–2589. [Google Scholar]
  25. Zhang, Y.; Li, J.; Luo, K.; Yang, Y.; Han, J.; Liu, N.; Qin, D.; Han, P.; Xu, C. V2VSSC: A 3D Semantic Scene Completion Benchmark for Perception with Vehicle to Vehicle Communication. arXiv 2024, arXiv:2402.04671. [Google Scholar]
  26. Xu, R.; Xiang, H.; Tu, Z.; Xia, X.; Yang, M.H.; Ma, J. V2X-ViT: Vehicle-to-Everything Cooperative Perception with Vision Transformer. In Proceedings of the European Conference on Computer Vision (ECCV), Tel Aviv, Israel, 23–27 October 2022. [Google Scholar]
  27. Song, S.; Yu, F.; Zeng, A.; Chang, A.X.; Savva, M.; Funkhouser, T. Semantic Scene Completion from a Single Depth Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 190–198. [Google Scholar] [CrossRef]
  28. Firman, M.; Aodha, O.M.; Julier, S.; Brostow, G.J. Structured Prediction of Unobserved Voxels from a Single Depth Image. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 5431–5440. [Google Scholar] [CrossRef]
  29. Dai, A.; Chang, A.X.; Savva, M.; Halber, M.; Funkhouser, T.; Niessner, M. ScanNet: Richly-Annotated 3D Reconstructions of Indoor Scenes. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2432–2443. [Google Scholar] [CrossRef]
  30. Li, J.; Liu, Y.; Yuan, X.; Zhao, C.; Siegwart, R.; Reid, I.; Cadena, C. Depth Based Semantic Scene Completion With Position Importance Aware Loss. IEEE Robot. Autom. Lett. 2020, 5, 219–226. [Google Scholar] [CrossRef]
  31. Sakaridis, C.; Dai, D.; Van Gool, L. Semantic Foggy Scene Understanding with Synthetic Data. Int. J. Comput. Vis. 2018, 126, 973–992. [Google Scholar] [CrossRef]
  32. Li, J.; Wang, P.; Han, K.; Liu, Y. Anisotropic Convolutional Neural Networks for RGB-D Based Semantic Scene Completion. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 44, 8125–8138. [Google Scholar] [CrossRef] [PubMed]
  33. Behley, J.; Garbade, M.; Milioto, A.; Quenzel, J.; Behnke, S.; Stachniss, C.; Gall, J. SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 9296–9306. [Google Scholar] [CrossRef]
  34. Wang, X.; Zhu, Z.; Xu, W.; Zhang, Y.; Wei, Y.; Chi, X.; Ye, Y.; Du, D.; Lu, J.; Wang, X. OpenOccupancy: A Large Scale Benchmark for Surrounding Semantic Occupancy Perception. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Paris, France, 1–6 October 2023; pp. 17804–17813. [Google Scholar] [CrossRef]
  35. Li, Y.; Li, S.; Liu, X.; Gong, M.; Li, K.; Chen, N.; Wang, Z.; Li, Z.; Jiang, T.; Yu, F.; et al. Sscbench: A large-scale 3d semantic scene completion benchmark for autonomous driving. arXiv 2023, arXiv:2306.09001. [Google Scholar]
  36. Xu, R.; Guo, Y.; Han, X.; Xia, X.; Xiang, H.; Ma, J. OpenCDA: An Open Cooperative Driving Automation Framework Integrated with Co-Simulation. In Proceedings of the International Conference on Intelligent Transportation Systems (ITSC), Indianapolis, IN, USA, 19–22 September 2021; pp. 1155–1162. [Google Scholar] [CrossRef]
  37. Song, R.; Liang, C.; Cao, H.; Yan, Z.; Zimmer, W.; Gross, M.; Festag, A.; Knoll, A. Collaborative Semantic Occupancy Prediction with Hybrid Feature Fusion in Connected Automated Vehicles. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 17–21 June 2024. [Google Scholar]
  38. Yuan, Y.; Cheng, H.; Sester, M. Keypoints-Based Deep Feature Fusion for Cooperative Vehicle Detection of Autonomous Driving. IEEE Robot. Autom. Lett. 2022, 7, 3054–3061. [Google Scholar] [CrossRef]
  39. Wang, T.H.; Manivasagam, S.; Liang, M.; Yang, B.; Zeng, W.; Urtasun, R. V2VNet: Vehicle-to-Vehicle Communication for Joint Perception and Prediction. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 605–621. [Google Scholar] [CrossRef]
  40. Chen, Q.; Ma, X.; Tang, S.; Guo, J.; Yang, Q.; Fu, S. F-cooper: Feature based cooperative perception for autonomous vehicle edge computing system using 3D point clouds. In Proceedings of the ACM/IEEE Symposium on Edge Computing, Arlington, VA, USA, 7–9 November 2019; pp. 88–100. [Google Scholar] [CrossRef]
  41. Mehr, E.; Jourdan, A.; Thome, N.; Cord, M.; Guitteny, V. DiscoNet: Shapes Learning on Disconnected Manifolds for 3D Editing. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Repulic of Korea, 27 October–2 November 2019; pp. 3473–3482. [Google Scholar] [CrossRef]
  42. Fang, S.; Li, H. Multi-Vehicle Cooperative Simultaneous LiDAR SLAM and Object Tracking in Dynamic Environments. IEEE Trans. Intell. Transp. Syst. 2024, 25, 11411–11421. [Google Scholar] [CrossRef]
  43. Yin, H.; Tian, D.; Lin, C.; Duan, X.; Zhou, J.; Zhao, D.; Cao, D. V2VFormer++: Multi-Modal Vehicle-to-Vehicle Cooperative Perception via Global-Local Transformer. IEEE Trans. Intell. Transp. Syst. 2024, 25, 2153–2166. [Google Scholar] [CrossRef]
  44. Luo, G.; Shao, C.; Cheng, N.; Zhou, H.; Zhang, H.; Yuan, Q.; Li, J. EdgeCooper: Network-Aware Cooperative LiDAR Perception for Enhanced Vehicular Awareness. IEEE J. Sel. Areas Commun. 2024, 42, 207–222. [Google Scholar] [CrossRef]
  45. Xu, R.; Tu, Z.; Xiang, H.; Shao, W.; Zhou, B.; Ma, J. CoBEVT: Cooperative Bird’s Eye View Semantic Segmentation with Sparse Transformers. In Proceedings of the Conference on Robot Learning, Atlanta, GA, USA, 14–18 December 2023; Volume 205, pp. 989–1000. [Google Scholar]
  46. Yu, H.; Luo, Y.; Shu, M.; Huo, Y.; Yang, Z.; Shi, Y.; Guo, Z.; Li, H.; Hu, X.; Yuan, J.; et al. DAIR-V2X: A Large-Scale Dataset for Vehicle-Infrastructure Cooperative 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 21329–21338. [Google Scholar] [CrossRef]
  47. Lu, Y.; Hu, Y.; Zhong, Y.; Wang, D.; Chen, S.; Wang, Y. An Extensible Framework for Open Heterogeneous Collaborative Perception. arXiv 2024, arXiv:2401.13964. [Google Scholar]
  48. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is all you need. In Proceedings of the Advances in Neural Information Processing Systems (NIPS’17), Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
  49. Xu, C.; Jia, W.; Wang, R.; Luo, X.; He, X. MorphText: Deep Morphology Regularized Accurate Arbitrary-shape Scene Text Detection. IEEE Trans. Multimed. 2022, 25, 4199–4212. [Google Scholar] [CrossRef]
  50. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 4 May 2021. [Google Scholar]
  51. Xu, C.; Jia, W.; Wang, R.; He, X.; Zhao, B.; Zhang, Y. Semantic Navigation of PowerPoint-Based Lecture Video for AutoNote Generation. IEEE Trans. Learn. Technol. 2022, 16, 1–17. [Google Scholar] [CrossRef]
  52. Xu, C.; Jia, W.; Cui, T.; Wang, R.; Zhang, Y.F.; He, X. Arbitrary-shape scene text detection via visual-relational rectification and contour approximation. IEEE Trans. Multimed. 2022, 25, 4052–4066. [Google Scholar] [CrossRef]
  53. Bai, X.; Hu, Z.; Zhu, X.; Huang, Q.; Chen, Y.; Fu, H.; Tai, C.L. TransFusion: Robust LiDAR-Camera Fusion for 3D Object Detection with Transformers. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 1080–1089. [Google Scholar] [CrossRef]
  54. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable Transformers for End-to-End Object Detection. In Proceedings of the International Conference on Learning Representations (ICLR), Vienna, Austria, 4 May 2021. [Google Scholar]
  55. Ramachandran, P.; Parmar, N.; Vaswani, A.; Bello, I.; Levskaya, A.; Shlens, J. Stand-alone self-attention in vision models. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
  56. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin Transformer: Hierarchical Vision Transformer using Shifted Windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 9992–10002. [Google Scholar] [CrossRef]
  57. Dong, X.; Bao, J.; Chen, D.; Zhang, W.; Yu, N.; Yuan, L.; Chen, D.; Guo, B. CSWin Transformer: A General Vision Transformer Backbone with Cross-Shaped Windows. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 12114–12124. [Google Scholar] [CrossRef]
  58. Hassani, A.; Walton, S.; Li, J.; Li, S.; Shi, H. Neighborhood Attention Transformer. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 17–24 June 2023; pp. 6185–6194. [Google Scholar] [CrossRef]
  59. Zhou, Y.; Tuzel, O. VoxelNet: End-to-End Learning for Point Cloud Based 3D Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4490–4499. [Google Scholar] [CrossRef]
  60. Yu, Z.; Shu, C.; Deng, J.; Lu, K.; Liu, Z.; Yu, J.; Yang, D.; Li, H.; Chen, Y. FlashOcc: Fast and Memory-Efficient Occupancy Prediction via Channel-to-Height Plugin. arXiv 2023, arXiv:2311.12058. [Google Scholar]
  61. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar] [CrossRef]
  62. Kingma, D.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diega, CA, USA, 7–9 May 2015. [Google Scholar]
  63. Xia, X.; Hang, P.; Xu, N.; Huang, Y.; Xiong, L.; Yu, Z. Advancing Estimation Accuracy of Sideslip Angle by Fusing Vehicle Kinematics and Dynamics Information With Fuzzy Logic. IEEE Trans. Veh. Technol. 2021, 70, 6577–6590. [Google Scholar] [CrossRef]
  64. Li, Y.; Zhuang, Y.; Hu, X.; Gao, Z.; Hu, J.; Chen, L.; He, Z.; Pei, L.; Chen, K.; Wang, M.; et al. Toward Location-Enabled IoT (LE-IoT): IoT Positioning Techniques, Error Sources, and Error Mitigation. IEEE Internet Things J. 2021, 8, 4035–4062. [Google Scholar] [CrossRef]
  65. Tsukada, M.; Oi, T.; Ito, A.; Hirata, M.; Esaki, H. AutoC2X: Open-source software to realize V2X cooperative perception among autonomous vehicles. In Proceedings of the 2020 IEEE 92nd Vehicular Technology Conference (VTC2020-Fall), Victoria, BC, Canada, 18 November–16 December 2020; pp. 1–6. [Google Scholar] [CrossRef]
Figure 1. The system framework of TSMV. It consists of four steps: spatial graph construction, feature extraction and compression, two-stream multi-vehicle collaboration, and the detection head. The details of each component are illustrated in Section 3.
Figure 1. The system framework of TSMV. It consists of four steps: spatial graph construction, feature extraction and compression, two-stream multi-vehicle collaboration, and the detection head. The details of each component are illustrated in Section 3.
Sensors 24 07702 g001
Figure 2. Neighborhood Self Attention (NSA) module (a) and Neighborhood Cross Attention (NCA) module (b). We use the dilation parameters (control the spacing between the kernel points), where the dilation is 2 in both the NSA and the NCA module.
Figure 2. Neighborhood Self Attention (NSA) module (a) and Neighborhood Cross Attention (NCA) module (b). We use the dilation parameters (control the spacing between the kernel points), where the dilation is 2 in both the NSA and the NCA module.
Sensors 24 07702 g002
Figure 3. Architectures of Neighborhood Self-Cross Attention Transformer (NSCAT).
Figure 3. Architectures of Neighborhood Self-Cross Attention Transformer (NSCAT).
Sensors 24 07702 g003
Figure 4. Visualization result. The red area in the figure represents the location of the ego vehicle.
Figure 4. Visualization result. The red area in the figure represents the location of the ego vehicle.
Sensors 24 07702 g004
Figure 5. Illustration of collaborative semantic occupancy prediction from multiple perspectives, compared to the ground truth.
Figure 5. Illustration of collaborative semantic occupancy prediction from multiple perspectives, compared to the ground truth.
Sensors 24 07702 g005
Figure 6. Positional error performance. The abscissa represents the standard deviation of the positioning sampling noise of the Gaussian distribution, and the ordinate represents the mIoU metric.
Figure 6. Positional error performance. The abscissa represents the standard deviation of the positioning sampling noise of the Gaussian distribution, and the ordinate represents the mIoU metric.
Sensors 24 07702 g006
Table 1. Semantic scene completion performance comparison on V2VSSC test set in perfect settings in detail. Bold font indicates the best on the column.
Table 1. Semantic scene completion performance comparison on V2VSSC test set in perfect settings in detail. Bold font indicates the best on the column.
MethodIoUmIoUcIoURoadCarTerrianBuildingVeg. 1Pole
No Fusion0.5120.4180.6040.6430.5660.4640.3240.3830.126
Late Fusion0.540.4330.6080.670.5460.4780.3680.4190.115
Early Fusion0.570.4610.6490.6820.6160.5080.340.4440.174
Fcooper [40]0.5560.4550.640.6760.6040.5130.3230.440.172
V2XViT [26]0.580.440.580.6580.5010.5580.3590.4330.131
V2VSSC [25]0.5660.4620.6580.6720.6440.4990.3360.4460.177
TSMV(Ours)0.5860.4920.6870.7080.6660.5770.3380.4810.18
1 Veg. is the abbreviation of Vegetation.
Table 2. Semantic scene completion performance comparison on V2VSSC test set in noisy settings. Bold font indicates the best on the column in noisy settings.
Table 2. Semantic scene completion performance comparison on V2VSSC test set in noisy settings. Bold font indicates the best on the column in noisy settings.
MethodPerfectNoisy
IoU mIoU cIoU IoU mIoU cIoU
No Fusion0.5120.4180.6040.5120.4180.604
Late Fusion0.540.4330.6080.5260.4030.544
Early Fusion0.570.4610.6490.5590.4410.613
FCooper [40]0.5560.4550.640.5520.4470.626
V2XViT [26]0.580.440.580.570.4310.565
V2VSSC [25]0.5660.4620.6580.5590.4430.618
TSMV(Ours)0.5860.4920.6870.5720.4680.648
Table 3. Semantic occupancy prediction on SemanticOPV2V val set. Bold font indicates the best on the line of different prediction tasks.
Table 3. Semantic occupancy prediction on SemanticOPV2V val set. Bold font indicates the best on the line of different prediction tasks.
Task TypeOcc Pred.Sem. Occ. Pred.
MethodRL 1CoHFF [37]TSMVCoHFF [37]TSMV
mean IoU57.1265.2465.7631.7234.04
Building (5.40%)67.544.0843.6324.0031.49
Fence (0.85%)59.4063.6364.1522.8627.71
Terrain (4.80%)43.6076.9977.3710.2112.89
Pole (0.39%)66.361.5761.2946.3349.93
Road (40.53%)51.4787.988.6864.8065.80
Side walk (35.64%)45.4686.2786.4957.4463.11
Vegetation (1.11%)43.6133.2633.9410.4710.86
Vehicles (9.14%)41.470.1572.6282.2184.11
Wall (2.01%)71.5180.4881.1338.2337.74
Guard rail (0.04%)49.6739.9139.7623.8117.97
Traffic signs (0.05%)68.9857.0658.10.246.93
Bridge (0.04%)76.5381.5981.970.000.00
1 RL(Raw LiDAR) is used as a baseline for the evaluation on the task of occupancy prediction.
Table 4. Inference time measured on GPU NVIDIA RTX 3060Ti.
Table 4. Inference time measured on GPU NVIDIA RTX 3060Ti.
MethodParameter (M)Time (ms)mIoU (prf)
No Fusion10.03500.418
Late Fusion10.03500.433
Early Fusion10.03540.461
V2XViT [26]15.07880.44
V2VSSC [25]10.03310.462
TSMV(Ours)11.23480.492
Table 5. Component ablation study. Two-stream represents two-stream architecture, and NSCAT represents the Neighborhood Self-Cross Attention Transformer module.
Table 5. Component ablation study. Two-stream represents two-stream architecture, and NSCAT represents the Neighborhood Self-Cross Attention Transformer module.
Two-StreamNSCATmIoU (prf)
Ego Stream Fuse Stream
××0.418
××0.444
×0.465
0.492
Table 6. Kernel size of neighborhood ablation study. Bold font indicates the best on the column.
Table 6. Kernel size of neighborhood ablation study. Bold font indicates the best on the column.
SizemIoURoadCarTerrianBuidingVegetablePole
30.4750.70.6730.5280.3030.4620.183
50.4770.7090.6620.5320.3340.4610.17
70.480.7110.6580.5630.3470.4660.132
90.4920.7080.6660.5770.3380.4810.18
110.4860.7150.6530.5860.3330.4870.143
Table 7. Performance of different transmission delays.
Table 7. Performance of different transmission delays.
Method00–5000–10000–20000–3000
No Fusion0.4180.4180.4180.4180.418
Late Fusion0.4330.4140.4030.3980.397
Early Fusion0.4610.4160.4040.3970.395
V2VSSC [25]0.4620.4270.4150.4090.406
TSMV0.4920.4520.440.4350.433
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Zhang, Y.; Han, J.; Han, P.; Luo, K. Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach. Sensors 2024, 24, 7702. https://doi.org/10.3390/s24237702

AMA Style

Li J, Zhang Y, Han J, Han P, Luo K. Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach. Sensors. 2024; 24(23):7702. https://doi.org/10.3390/s24237702

Chicago/Turabian Style

Li, Junxuan, Yuanfang Zhang, Jiayi Han, Peng Han, and Kaiqing Luo. 2024. "Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach" Sensors 24, no. 23: 7702. https://doi.org/10.3390/s24237702

APA Style

Li, J., Zhang, Y., Han, J., Han, P., & Luo, K. (2024). Semantic Scene Completion in Autonomous Driving: A Two-Stream Multi-Vehicle Collaboration Approach. Sensors, 24(23), 7702. https://doi.org/10.3390/s24237702

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop