Next Article in Journal
Comparison of the Accuracy of Ground Reaction Force Component Estimation between Supervised Machine Learning and Deep Learning Methods Using Pressure Insoles
Previous Article in Journal
Multi-Channel Multi-Scale Convolution Attention Variational Autoencoder (MCA-VAE): An Interpretable Anomaly Detection Algorithm Based on Variational Autoencoder
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Camera-Radar Fusion with Radar Channel Extension and Dual-CBAM-FPN for Object Detection

1
Guangxi Key Laboratory of Precision Navigation Technology and Application, Guilin University of Electronic Technology, Guilin 541004, China
2
Information and Communication School, Guilin University of Electronic Technology, Guilin 541004, China
3
National & Local Joint Engineering Research Center of Satellite Navigation Positioning and Location Service, Guilin 541004, China
4
GUET-Nanning E-Tech Research Institute Co., Ltd., Nanning 530031, China
*
Authors to whom correspondence should be addressed.
Sensors 2024, 24(16), 5317; https://doi.org/10.3390/s24165317
Submission received: 2 June 2024 / Revised: 5 August 2024 / Accepted: 15 August 2024 / Published: 16 August 2024
(This article belongs to the Section Radar Sensors)

Abstract

:
When it comes to road environment perception, millimeter-wave radar with a camera facilitates more reliable detection than a single sensor. However, the limited utilization of radar features and insufficient extraction of important features remain pertinent issues, especially with regard to the detection of small and occluded objects. To address these concerns, we propose a camera-radar fusion with radar channel extension and a dual-CBAM-FPN (CRFRD), which incorporates a radar channel extension (RCE) module and a dual-CBAM-FPN (DCF) module into the camera-radar fusion net (CRF-Net). In the RCE module, we design an azimuth-weighted RCS parameter and extend three radar channels, which leverage the secondary redundant information to achieve richer feature representation. In the DCF module, we present the dual-CBAM-FPN, which enables the model to focus on important features by inserting CBAM at the input and the fusion process of FPN simultaneously. Comparative experiments conducted on the NuScenes dataset and real data demonstrate the superior performance of the CRFRD compared to CRF-Net, as its weighted mean average precision (wmAP) increases from 43.89% to 45.03%. Furthermore, ablation studies verify the indispensability of the RCE and DCF modules and the effectiveness of azimuth-weighted RCS.

1. Introduction

Road traffic safety and efficiency are the key challenges in modern transportation [1]. An intelligent roadside unit (RSU), which is capable of achieving over-the-horizon perception, is used to monitor traffic flow and alert road users to dangerous situations [2,3]. To achieve this, intelligent RSUs must possess a stable and reliable perception system [1].
However, due to the complexities presented by the road environment, such as object occlusion, rain, and lighting changes, a single sensor cannot meet the demand for reliable environment perception; this has prompted the emergence of multi-sensor fusion as a research hotspot [4]. Among the variety of sensors utilized in intelligent transportation, millimeter-wave radar and cameras exhibit complementary characteristics. Specifically, millimeter-wave radars are robust to severe weather conditions, providing accurate velocity and depth information of objects. However, their azimuth resolution is limited, and they fail to recognize the appearance and shape of objects. Cameras, on the other hand, are capable of achieving a high azimuth resolution, accurately capturing information about objects’ physical characteristics; however, they fall short when it comes to depth information and are vulnerable to weather conditions and lighting conditions. Therefore, the fusion of the millimeter-wave radar and a camera has attracted significant attention from researchers [5].
Compared to single sensors, which have issues with misidentification or, in some cases, failure to detect objects altogether, multi-sensor fusion is an advanced object detection method. Several researchers have used it to great effect; J. Dudczyk et al. [6] realized object detection through a multi-sensory data fusion method, while X. Liu et al. [7] iteratively updated the radar and camera through an interactive module and aggregated the radar and image features with a set of sparse 3D object queries, all while preserving the integrity of the original radar features to prevent information loss to facilitate more accurate object detection. In another experiment, H. Sun et al. [8] used height extension and azimuth uncertainty to achieve more significant feature extraction, generating denser radar input and achieving better alignment between the radar data and the objects on the image. M. Zong et al. [9] designed a spatial attention concatenate fusion (SAC) module that focuses on the feature layer, combining the positional information of shallow feature maps with deep semantic information, to improve the detection of both large and small targets. Meanwhile, S. He et al. [10] created the multi-scale fusion obstacle detection algorithm, which combines different scales of radar and visual feature maps at three different levels, making it easier for the model to detect small targets in the environment.
Radar and camera fusion methods can be classified as data-level, feature-level, and decision-level fusion based on the fusion stage. Most studies fusing the mmWave radar data and camera data of intelligent RSUs take place at the decision level. J. Lin et al. [3] proposed a deep-learning-based early fusion between a mmWave radar and an RGB camera sensor, but the feature extraction prior to fusion was too simple. Since radar data and camera data have distinct representations, feature-level fusion emerged as the more suitable approach. The feature-level fusion of radar data and camera data currently encompasses two approaches: fusion after feature extraction and fusion during feature extraction. For fusion that takes place after feature extraction, radar features and camera features extracted from the last layer of each extraction network are employed [11,12,13]. This method ignores the low-level features, and as such, its capacity to detect smaller objects is limited. When the fusion takes place during feature extraction however, radar features and camera features extracted from each layer of each extraction network are all used to good advantage [14,15,16]. By utilizing the high-level and low-level features simultaneously, fusion that takes place during feature extraction considers both small and larger objects. Camera-radar fusion net (CRF-Net) is a representative model of fusion during feature extraction, which builds on RetinaNet with a Visual Geometry Group-16 (VGG16) backbone and a feature pyramid network (FPN) [17]. Radar and camera data are concatenated and fed into the VGG16 network; then, the radar data are concatenated to the output of the previous fused network block. Subsequently, the FPN is introduced to enhance the fusion of the radar features. CRF-Net automatically learns the level at which the fusion of the sensor data is most beneficial for the detection result, making it a highly promising method [18].
However, CRF-Net has several issues that have yet to be resolved. First, it uses limited radar channels, which are generated by distance and radar cross-section (RCS). Second, within the FPN structure, there is a lack of emphasis on important features. This inevitably affects the feature extraction of objects whose features are less prominent, such as small and occluded objects, and may have a negative effect on the detection performance. Therefore, it is crucial to minimize information loss during feature representation [19] and enhance the representation ability of important features during feature fusion [20] to improve the performance of CRF-Net.
To solve the abovementioned problems, we propose camera-radar fusion with radar channel extension and dual-CBAM-FPN (CRFRD), a novel camera-radar fusion model based on CRF-Net for object detection. By extending three radar channels and inserting CBAM at the input and the fusion process of FPN simultaneously, the CRFRD effectively enhances object detection accuracy.
The main contributions of this paper are as follows:
  • We propose the CRFRD model, which incorporates radar channel extension (RCE) and dual-CBAM-FPN (DCF) into the CRF-Net model. The CRFRD enriches the representation of radar features and pays more attention to the important information of multi-scale features that are fused by radar and camera features.
  • We introduce a new parameter azimuth-weighted RCS to construct a radar channel, making use of the azimuth and RCS parameters to achieve richer feature representation. Along with velocity, azimuth and azimuth-weighted RCS are chosen to construct additional radar channels and undergo experimental evaluation.
  • We present a dual-CBAM-FPN strategy to direct the model’s focus toward pivotal features along the channel and spatial dimensions. CBAM is inserted into both the input and the fusion process of FPN, which significantly enhances the feature representation, particularly with regard to smaller objects.
  • Numerous experiments verify the effectiveness of the CRFRD model in improving the detection accuracy of CRF-Net. The weighted mean average precision (wmAP) increases from 43.89% to 45.03%, and more small and occluded objects are detected by CRFRD.
The remainder of the paper is structured as follows. In Section 2, we introduce a series of relevant studies and provide a concise overview of our background knowledge in this area. In Section 3, we describe the structure of the proposed model, as well as that of RCE and DCF. In Section 4, we present the results and analysis of comparative experiments and ablation studies. Finally, in Section 5, we list the conclusions of our work.

2. Related Work and Background Knowledge

2.1. Related Work

The limited utilization of radar features and insufficient extraction of important features are not unique to CRF-Net, and numerous researchers have endeavored to address or mitigate these problems.
Two common approaches are used to fuse the different representations of radar data and camera data. One involves projecting the radar point clouds onto the image plane to construct additional channels with parameters of the radar point clouds [19], from which 2D bounding boxes are obtained. Another option is to project the multi-view camera features to the 3D ego-car coordinate to generate bird’s eye view (BEV) features [21], which provide 3D bounding boxes. In this paper, we focused on the fusion of RSUs, making multi-view camera features more difficult to achieve; therefore, we chose to project the radar point clouds onto the image plane to generate 2D bounding boxes for object detection. According to NuScenes, a comprehensive dataset for radar-camera fusion, there are 18 parameters for each radar point, including the position coordinate in three dimensions ( x , y , z ) , the radial velocity in the x and y directions ( v x , v y ) , dynamic properties p D indicating whether the point is stationary or in motion, point identification, RCS σ , radial velocity compensated by ego-motion in the x and y directions ( v x _ c o m p , v y _ c o m p ) , clustering validity states, Doppler ambiguity solution states, the false-alarm probability of the cluster, quality validity states, and the uncertainty of x and y position and velocity ( x r m s , y r m s , v x r m s , v y r m s ) [22]. In most cases, distance d and azimuth θ are also calculated to assist with the construction of radar channels. The parameters chosen for individual models may vary. For instance, S. Chang et al. [23] created the radar channels using three parameters, including distance and velocity ( d , v x , v y ) . These three parameters were also implemented by several other researchers [11,24]. L. Li et al. [22] selected five parameters to construct the radar channels, including RCS, distance, velocity, and azimuth ( σ , d , v x , v y , θ ) . In addition to several of the previously mentioned parameters, L. Stacker et al. [19] proposed two novel parameters which they referred to as azimuth uncertainty and azimuth-uncertainty-weighted RCS (UwRCS), respectively, and constructed the radar channels using one of the proposed parameters together with the distance d and RCS σ . They then compared the model’s performance with different proposed parameters, demonstrating that a model using UwRCS outperformed the model that implemented azimuth uncertainty. Constructing more comprehensive radar channels, either by leveraging more radar output parameters or through the design of new representation parameters, to minimize the loss of radar feature information is an area that merits further investigation.
FPN is a network structure that extracts features from different layers and makes full use of multi-scale features. However, it is susceptible to loss and underutilization of information during the multi-scale fusion process. Introducing the attention mechanism has proven to be an effective solution to the problem [25]. The convolutional block attention module (CBAM), an effective mechanism for enhancing convolutional neural networks (CNNs), has been inserted into the FPN model to improve its performance. The combination of CBAM and FPN takes three distinctive forms. First, CBAM is incorporated in the backbone network [26], where it takes the information refined by the spatial and channel attention mechanism as the input of FPN C i . This method emphasizes the local efficient information of the feature maps, enhancing the detection capability for occluded or truncated objects. Second, CBAM is inserted into the input or output of FPN [27,28]. Q. Guo et al. [27] added CBAM before the FPN output of the lowest-level feature map to emphasize the object region of interest, improving the model’s ability to detect small objects. J.C.Á. Cerón et al. [28] employed CBAM before each feature map C i input into FPN and before FPN output each feature map P i , respectively. This aggregation of local information with its corresponding global dependencies extracts richer context, refining the feature representation of objects. Third, CBAM is introduced into the fusion process of each scale of the feature maps [29,30]. In their work, Z. Gui et al. [29] began by upsampling the high-level features of FPN P i and inputting them into CBAM to obtain the attention-weighted features A i 1 , which were further fused with C i 1 via concatenation to receive the refined features P i 1 . These refined feature maps highlight salient features at specific scales. Y. Han et al. [30] used the upsampled high-level features of FPN and C i 1 as the input of CBAM to obtain the refined feature maps P i 1 . By improving the feature representation, this model extracts important features to enhance detection accuracy. All the above-described combinations of CBAM with FPN emphasize the local information of the feature map. When inserting CBAM at the input or output of FPN, the feature map in each scale only focuses on its own local information. When CBAM is inserted during the fusion of feature maps at each scale, low-level feature maps focus on the local information of feature maps fused by higher-level feature maps. It is worth fully leveraging the various forms of CBAM-FPN to enhance the multi-scale feature fusion performance of FPN.

2.2. Radar Data Preprocessing

To generate radar images, it is necessary to solve the problems of projection to the image plane, height loss, and data sparsity. The radar data are processed via spatial calibration, line expansion, and aggregation of multiple radar sweeps.
Radar and camera data are composed of two different structures, with different coordinate systems. In order to fuse radar data with camera data, it is essential to unify the coordinates of the two sensors. This is known as spatial calibration—a coordinate transformation between the radar coordinate system, the global coordinate system, the camera coordinate system, and the image coordinate system, which allows the radar point clouds to be projected onto the image plane [15]. The coordinate transformation process is depicted in Figure 1. First, the radar points in the radar coordinate system are rotated and translated to the ego coordinate system under the radar timestamp using the radar’s external parameters. Then, they are converted to the global coordinate system. After that, the global coordinates are transformed to the ego coordinates under the camera timestamp. Next, the ego coordinates under the camera timestamp are transformed to camera coordinates using the camera’s external parameters. Finally, the camera coordinates are transformed to the image coordinates using the camera’s intrinsic parameters. When the platform installing the radar and camera is moving, the ego coordinate system under the radar timestamp may differ from the ego coordinates under the camera timestamp; when the platform is stationary, the ego coordinate system under the radar timestamp, the global coordinate system and the ego coordinates under the camera timestamp are identical. To generate the radar image, the parameters of the radar point clouds are stored as pixel values for positions in the image plane; these should be converted to a value between 0 and 255. At the location of image pixels where no radar returns are present, the pixel values are filled by 0. The radar image plane generated by each radar parameter serves as a channel for the radar image [22].
As radar detection does not provide information about objects’ height, the radar point projection is extended along the vertical direction to three meters [19]. The radar point is mapped into the image plane with a pixel width of one. Figure 2 shows the projection results of radar point clouds onto the image plane, where Figure 2a is the projection result prior to the line expansion and Figure 2b is a projection result after the line expansion. Figure 2 depicts the image coordinate system, which is denoted as ( y I , z I ) .
Radar sensors can only provide limited point clouds for object detection and are therefore incapable of performing classification. To address the sparsity of radar point clouds, we accumulate the point clouds data from several radar scans as a single frame. This significantly enhances the density of the data [17].

2.3. CRF-Net

CRF-Net takes the camera image I c a m e r a R m × n × l c and the generated radar image I r a d a r R m × n × l r as inputs, where m and n represent the dimensions of the images; l c represents the number of channels of the camera image, l c = 3 ; and l r represents the number of channels of the radar image, l r = 2 . The camera image and the radar image are fused by concatenation and, thereafter, are considered as the input of the VGG16 network, namely, C 0 = I r a d a r I c a m e r a , where represents concatenation. Subsequently, the output of each block of the VGG16 network undergoes further concatenation with the max-pooled radar images F i r a d a r to obtain the fused features C i .
{ F i r a d a r = M a x P o o l ( F i 1 r a d a r ) , C i = B l o c k i ( C i 1 ) F i r a d a r , i = 1 , 2 , , 5 ,
where F 0 r a d a r = I r a d a r . M a x P o o l ( ) represents maximum pooling and B l o c k i ( ) represents the output after the i -th block of VGG16.
During feature extraction of VGG16 and multi-scale feature fusion of FPN, the radar channels undergo additional concatenation at each level. As shown in Equations (2) and (3), using C 3 C 5 and F 3 r a d a r F 7 r a d a r as the input, FPN outputs five feature maps N 3 N 7 , which are in different scales [31].
P i = { Conv 3 × 3 ( Conv 1 × 1 ( C i ) + Upsample ( Conv 1 × 1 ( C i + 1 ) ) ) ,   Conv 3 × 3 ( Conv 1 × 1 ( C i ) ) ,   Conv 3 × 3 ( C i 1 ) , Conv 3 × 3 ( Relu ( P i 1 ) ) , i = 3 , 4 i = 5 i = 6 i = 7 ,
N i = F i r a d a r P i , i = 3 , 4 , , 7
where P i represents the intermediate feature maps of FPN; Conv 1 × 1 ( ) represents 2D convolution with 1 × 1 convolution kernel and stride 1, which is used to convert the number of channels of the feature map; Conv 3 × 3 ( ) represents 2D convolution with a 3 × 3 convolution kernel; and Upsample ( ) represents upsampling. The purpose of the upsampling operation is to obtain a dimension that can be added to the previous layer; Relu ( x ) = max ( 0 , x ) .
The loss function of CRF-Net consists of a classification loss L c l s and a regression loss L r e g . The total training loss L t o t a l for the whole task is constructed as Equation (4).
L t o t a l ( x i , y i ) = L c l s ( x i , x i ) + β L r e g ( y i , y i )
L c l s ( x i , x i ) = { α ( 1 x i ) γ log ( x i ) ,   x i = 1 ( 1 α ) ( x i ) γ log ( 1 x i ) , x i = 0
L r e g ( y i , y i ) = { 0.5 δ 2 ( y i y i ) 2 ,   i f | y i y i | 1 δ 2 | y i y i | 0.5 δ 2 , otherwise
where i is the anchor index; x i is the ground truth score, which is 1 if the anchor is positive and 0 if it is negative; x i is the model’s estimated probability for the class with the label x i = 1 ; y i is the ground truth of the bounding box; y i is the predicted output of the regression subnetwork; and β denotes the weight of L r e g . The classification loss L c l s is computed as described by the authors of [32], where α = 0.25 , γ = 2.0 . The regression loss L r e g is computed as stated by the authors of [33], where δ = 2.0 , β = 1.0 .

2.4. CBAM

CBAM is a lightweight and general attention module that combines the channel attention mechanism and spatial attention mechanism for feed-forward convolutional neural networks. It can conduct end-to-end training with basic CNN and achieves a guaranteed performance improvement with less overhead [34]. The channel attention module enhances the feature representation of different channels, while spatial attention facilitates the extraction of important information regarding various locations in the space [35]. The structure of CBAM is illustrated in Figure 3. The input feature map F R H × W × C is processed by the channel attention weights T c R 1 × 1 × C and the spatial attention weights T s R H × W × 1 in turn, which allows the refined feature map F R H × W × C to be obtained.
F = T c F F = T s F

2.4.1. Channel Attention Module

The channel attention module focuses on the specific meaningful contents of input feature maps [36]. The spatial information of a feature map is aggregated using both average pooling and max-pooling operations, and two different spatial context descriptors are generated. These descriptors are then fed into a shared network composed of a multi-layer perceptron (MLP) to learn the importance of each channel adaptively. The output feature vectors are summed and then activated by a sigmoid function to generate the channel attention weights T c R 1 × 1 × C , which are calculated as
T c ( F ) = S i g m o i d ( M L P ( A v g P o o l ( F ) ) + M L P ( M a x P o o l ( F ) ) ) .

2.4.2. Spatial Attention Module

The spatial attention module places greater emphasis on the specific locations of input feature maps [37]. Initially, the max-pooling and average-pooling operations are applied along the channel axis to generate features with different context scales. Then, the outcomes of both operations are concatenated along the channel axis to create a feature map with different scales of contextual information. Finally, a 2D convolution with 7 × 7 convolution kernel is utilized to generate a spatial attention weighted map, which is normalized with the sigmoid function T s R H × W × 1 . The calculation formula is as follows:
T s ( F ) = S i g m o i d ( Conv 7 × 7 ( [ A v g P o o l ( F ) ;   M a x P o o l ( F ) ] ) ) .

3. Camera-Radar Fusion with RCE and DCF

3.1. The Overall Structure of CRFRD

The block diagram of the CRFRD, shown in Figure 4, is divided into four main components: the input channel generation module, the feature extraction and fusion module, the dual-CBAM-FPN module, and the classification and regression module. The input channel generation module includes camera channel and radar channel generation. Camera images are composed of R, B, and G channels, while radar images comprise five channels: two that are the same as the baseline and three additional channels. The feature extraction and fusion module fuses the camera and radar features during the feature extraction process, since it is based on VGG16, and this means that the features extracted from each block are all fused features. For the dual-CBAM-FPN module, CBAM is inserted into the input and the fusion process of FPN simultaneously. Meanwhile, for the classification and regression module, the fused features are used for both classification and regression tasks.

3.2. Radar Channel Extension

In this section, we describe the radar channel extension module of CRFRD, adding three channels to enrich the representations of radar features.
The radial velocity in the x and y directions ( v x , v y ) indicate the motion state of point clouds. Point clouds with the same radial velocity in the x and y directions are likely to be part of the same motion entity. This facilitates the clustering of point clouds and helps with the identification of different objects, especially those that are occluded. Therefore, the two parameters ( v x , v y ) are utilized for image plane projection, and two additional radar images are constructed; there are then considered to be two extra radar channels.
Acknowledging the inherent redundancy of azimuth with respect to the position and distance of point clouds, we introduce a new parameter named azimuth-weighted RCS, rather than directly projecting it to the image plane to construct radar image. In the Experimental Section, we evaluate how introducing either the azimuth or the azimuth-weighted RCS as a channel affects the detection accuracy.
According to the description of spatial calibration in Section 2.2, the i -th radar point is projected from the 3D coordinates in the radar coordinate system P i R = ( x i R , y i R , z i R ) to the 2D coordinates in the image coordinate system P i I = ( y i I , z i I ) through a series of matrix transformations.
( y i I , z i I ) = T C I ( x i C , y i C , z i C ) = T C I T E ( t ) C T G E ( t ) T E ( t k ) G T R E ( t k ) ( x i R , y i R , z i R ) ,
where R , C , and I represent the radar, camera, and image coordinate system, respectively; T Y X represents the transformation from a Y coordinate system to an X coordinate system; E represents the vehicle coordinate system; t represents time; k represents the delay between the radar timestamp and the camera timestamp; and G represents the global coordinate system. The conversion from ( x i R , y i R , z i R ) to ( x i C , y i C , z i C ) is achieved by a rotation matrix and translation vector from 3D to 3D without any loss of information. However, the projection to the image plane loses one dimension of information when converted from the camera coordinate system to the image coordinate system.
{ y i I = f y i C x i C z i I = f z i C x i C ,
where f is the distance between the origin of the image coordinate system and the origin of the camera coordinate system. The information loss can be compensated by d i C = ( x i C ) 2 + ( y i C ) 2 + ( z i C ) 2 or θ i C = arctan ( y i C / x i C ) or φ i C = arctan ( ( x i C ) 2 + ( y i C ) 2 / z i C ) . Assuming that d i C is used for compensation, both θ i C and φ i C are functions of d i C , y i I , and z i I . Since ( d i C , θ i C , φ i C ) and ( d i R , θ i R , φ i R ) can be converted from each other, θ i R is also a function of d i R , y i I , and z i I , which can be denoted as
θ i R = H ( d i R , y i I , z i I ) .
Here, only the relationship between θ i R and d i R is discussed. The relationship between φ i R and d i R is neglected because in the radar coordinate system, z i R = 0 and φ i R = 90 .
According to Equation (12), if parameter θ i R is utilized to construct a radar image, it can be considered as a redundancy of a radar image constructed from distance, and it only increases the weight of the distance channel. Therefore, in this paper, we utilize θ i R in a novel way to construct a new parameter that will enhance the RCS of the targets whose azimuth is close to zero radian.
Azimuth-weighted RCS σ θ : We design σ θ by multiplying the azimuth and RCS as follows:
σ θ = θ i R σ .
For two points with different RCS but similar azimuths close to zero radian, σ θ for these two points are both close to zero. After normalization, the difference of RCS between these two points is reduced. In other words, azimuth helps to enhance the RCS of points close to zero radian. Since the azimuths between π / 3 and π / 3 are all approximately less than 1 radian, the effect of shrinkage has a large scope. A radar image constructed from RCS and a radar image constructed from azimuth-weighted RCS are compared in Figure 5, where the green dashed rectangles are indicative of a point with enhanced RCS and the blue dashed rectangles represent a point with a similar azimuth.
In summation, the extended radar channel is constructed as follows:
E C R F R D r a d a r = { E C R F N e t r a d a r , v x , v y , σ θ } ,
where E C R F N e t r a d a r = { d , σ } .

3.3. Dual-CBAM-FPN

The structure of the dual-CBAM-FPN module is depicted in Figure 6, which comprises the input module, the multi-scale feature fusion module, and the output module.
The input module consists of CBAM, as described in Section 2.4, and 2D convolution with a 1 × 1 convolution kernel and stride 1. C 3 , C 4 , and C 5 are the three high-level features extracted by the feature extraction and fusion module of the CRFRD. C 4 and C 5 have 517 channels, 512 of which are extracted by VGG16, while the remaining 5 are obtained from a radar image. Meanwhile, C 3 possesses 261 channels, with 256 channels extracted by VGG16 and 5 that are obtained from the radar image. The 2D convolution reduces the channel dimensions of C i to match the number of subsequent channels. The weighted features A 3 , A 4 , and A 5 are obtained after the processing of the input module,
A i = Conv 1 × 1 ( M ( C i ) ) , i = 3 , 4 , 5 ,
where M ( ) represents the CBAM processing and Conv 1 × 1 ( ) represents 2D convolution with 1 × 1 convolution kernel.
The multi-scale feature fusion module is a bottom–up network that begins with A 5 . As described by Equation (2), the feature is upsampled and processed via CBAM to obtain B i , which is then added to A i 1 to obtain D i 1 .
B i = M ( Upsample ( A i ) ) , i = 4 , 5 ,
D i 1 = A i 1 + B i i = 4 , 5 .
For the output module, D i is processed by 2D convolution with 3 × 3 convolution kernel and stride 1 to obtain P i . Furthermore, P 5 is obtained from A 5 , P 6 is obtained from C 5 , and P 7 is obtained from P 6 . Since it is necessary to carry out further concatenation with the five channels of the radar image, the number of channels of P i must be set to 251 to maintain the number of channels after fusion at 256. The final output of the dual-CBAM-FPN module is represented by the multi-scale features P 3 , P 4 , P 5 , P 6 , P 7 .
P i = { Conv 3 × 3 ( D i ) ,   Conv 3 × 3 ( A i ) ,   Conv 3 × 3 ( C i 1 ) , Conv 3 × 3 ( Relu ( P i 1 ) ) , i = 3 , 4 i = 5 i = 6 i = 7 ,
where Conv 3 × 3 ( ) represents 2D convolution with 3 × 3 convolution kernel used to alleviate the aliasing effect of upsampling.
The CBAM in the input module serves to refine feature representations of the input, emphasizing the interdependencies between channels and spatial regions. The CBAM in the multi-scale feature fusion module highlights features for specific scales. By inserting CBAM into both the input and the multi-scale feature fusion of FPN, the high-level and low-level features can be more effectively fused, which is beneficial for the detection of small or occluded objects.

4. Experiment and Analysis of Results

In this section, we evaluate our proposed CRFRD on the NuScenes dataset and real data. We begin by introducing the dataset, along with a detailed description of the experimental settings and evaluation criteria. After that, the numerical results and the visualization results of comparative experiments are presented. Finally, to validate the effectiveness of our proposed modules, several ablation studies are conducted.

4.1. Dataset and Experimental Settings

As we are unable to access open datasets for RSUs of mmWave radar sensors and camera sensors, in our experiments, we train and evaluate our model on the simplified NuScenes dataset to mimic the scenario of RSUs and then test it on the real data collected in the RSU scenario.
NuScenes is a comprehensive sensor dataset comprising data from six cameras and five mmWave radars [38]. However, we only employ the front camera and the front radar, as was the case in a previous study with a similar objective [19]. The original resolution of the camera images in the NuScenes dataset was 1600 × 900; however, we reduce this to 640 × 360, which allows the training to be completed more quickly. The front radar samples and the front camera samples in the NuScenes dataset are processed by the NuScenes generator module, providing a total of 20,480 training samples, 6839 validation samples, and 6830 test samples. We then classify the obtained samples based on what they depict; the seven chosen categories are human, bicycle, bus, car, motorcycle, trailer, or truck.
To verify the proposed model’s performance in RSUs scenarios, an experiment is conducted on the roadside. We use a ARS408-21 mmWave radar manufactured by German Continental to obtain the radar point clouds and a USB camera to capture video footage. The device used in the experiment is depicted in Figure 7a. The parameters of ARS408-21 are as follows: the cycle time is approximately 72 ms; the maximum distance is 70 m for near range and 250 m for far range; the distance resolution is 0.39 m for near range and 1.79 m for far range; the velocity range is −400 km/h to 200 km/h; and the velocity resolution is 0.1 km/h. The camera is configured with an output image revolution of 640 × 480 and a frame rate of 20 fps. The experimental device is raised up to 1.0 m high and the data are transferred to a laptop for storage. The experimental setup is shown in Figure 7b. Data are collected from three scenes: scene 1, which is recorded during the daytime; scene 2, which depicts cloudy conditions; and scene 3, which takes place at night. To address the sparsity of the radar point cloud, we accumulate the point cloud data from 3 radar scans as a single frame, which is approximately 0.25 s. The data collection duration is approximately 9.5 s per scene, and in this time, 39 radar point clouds data are obtained. The corresponding 39 images are selected via time calibration, where the radar data are matched to the nearest camera data. The final test set comprises 117 samples.
We train, evaluate, and test the model on a device with 128 GB RAM and GPU configured as an RTX 3090 and CUDA 11.3. The most appropriate learning rate contributes to the model’s strong convergence. The initial learning rate is set to 2 × 10−5. Similar to the baseline CRF-Net [17], we conduct training for 25 epochs with a batch size of 1, which helps to prevent model overfitting.

4.2. Evaluation Criteria

In this paper, the weighted mean average precision (wmAP) [17] is utilized to evaluate the performance of the object detection model. The wmAP is calculated as a weighted mean of the average precision (AP) for each category.
wmAP = 1 N i = 1 C ( N i AP i ) ,
where N represents the total number of objects in all categories; C represents the number of object categories (and C = 7 in this paper); N i represents the number of the i -th category of object; and A P i represents the average accuracy of the i -th category of object.
AP = 0 1 p ( r ) d r ,
where p ( r ) represents the precision-recall curve; p represents the precision; and r represents the recall. When calculating the precision and recall, the intersection over union (IoU) is set to 0.5.

4.3. Comparative Experiments

To verify the effectiveness of the CRFRD model, we compare its detection results with one image-only network RetinaNet [32] and five millimeter-wave radar and camera fusion models that generate 2D bounding boxes: CRF-Net, the research conducted by Nabati and Qi [39], REF-Net [40], RSA + CA2 [41], and the research conducted by H. Sun et al. [42]. We do not compare our results with those described in reference [3], in which mmWave radar data and camera data for intelligent RSUs are fused at the feature level, because the authors of that study trained and evaluated the model’s performance based on their own dataset. The comparison results are summarized in Table 1, describing the AP for each category as well as the wmAP. The best results are written in bold text, while the second-best results are underlined. The results of CRF-Net are reproduced in the table below, and the wmAP is 43.89%, which is consistent with the reported result of 43.95%. The results of other models are directly cited from the referenced papers. To further compare the computational complexity of our method and the CRF-Net, we also illustrate two metrics, parameters [10] and frames per second (FPS) [43], which are usually used to evaluate the computational burden of deep learning models. The number of parameters indicates the memory occupied, while frames per second (FPS) is indicative of the inference time.
As illustrated in Table 1, according to wmAP, the CRFRD outperforms all the other models, especially for the image-only model RetinaNet. The wmAP of CRFRD surpasses that of RetinaNet, CRF-Net, Nabati and Qi, REF-Net, RSA + CA2, and H. Sun et al. by 1.45%, 1.14%, 0.54%, 0.27%, 1.11%, and 0.15%, respectively. The CRFRD achieves the best AP performance in two categories, human and car, and achieves the second-best performance in three categories, bicycle, motorcycle, and trailer; as such, it can be considered the best model in Table 1. These outcomes demonstrate the effectiveness of the proposed CRFRD model in improving detection performance.
Moreover, Table 1 shows that the computational complexity of the CRFRD has a slight depravation compared to the CRF-Net, where the parameters of CRFRD are 2.6% greater than those of CRF-Net, and the FPS of CRFRD is 3.6% less than that of CRF-Net. The performance of the proposed CRFRD is improved by sacrificing the computational complexity. Furthermore, as the radar data preprocessing also takes time, we have included the preprocessing times of radar data for one image of CRF-Net and CRFRD here; these were 0.15 milliseconds and 0.18 milliseconds, respectively.
Providing further evidence of how the CRFRD enhances the detection performance, several visualizations provided by CRF-Net and CRFRD are depicted in Figure 8. Each row in Figure 8 exemplifies a detection result from a distinct scene, during the daytime, in cloudy conditions, in rainy conditions, and at night, respectively. The cloudy, rainy, and night-time scenes pose difficulties for camera detection, which supports the advantages of millimeter-wave radar and camera fusion. The images in the first column show the ground truth of the object in each scene. The second column contains the detection result of CRF-Net, while the third column depicts the detection result of CRFRD. The yellow dashed ellipses and the red dashed ellipses indicate the small and the occluded objects that are overlooked by CRF-Net but successfully detected by CRFRD.
For each scene, several small objects and occluded objects are overlooked by CRF-Net but successfully detected by CRFRD. Specifically, in the daytime scenario, there is a car at a distance of 88.8 m that goes undetected by CRF-Net but is correctly identified by CRFRD. In the cloudy scenario, there is a car at a distance of 70.6 m that is overlooked by CRF-Net but recognized by CRFRD, and a car at a distance of 30.0 m that is occluded by another car is missed by CRF-Net but is successfully identified by CRFRD. In the rainy scenario, there is a car at a distance of 28.2 m that is occluded by another car and is not detected by CRF-Net but is identified by CRFRD. In the night-time scenario, there is a truck at a distance of 30.4 m that is not observed by CRF-Net but is recognized by CRFRD, and there is a bus with a distance of 48.0 m that is not detected by CRF-Net but is successfully detected by CRFRD. Moreover, there is a car with a distance of 48.7 m that is occluded by another car that is overlooked by CRF-Net but is identified by CRFRD. These results demonstrate CRFRD’s effective detection of both small and occluded objects, which is essential for improving the amount of time intelligent RSUs take to identify a hazard.
To test the performance of CRFRD in the RSU scenario, we evaluate CRF-Net and CRFRD using real data and illustrate some examples of the visualization results. Since the number of samples in the test set is limited, it is difficult to evaluate the AP for each category. Therefore, in the test experiment involving real data, we count the number of targets that are detected correctly for each frame to evaluate the performance of the models. As we annotate the ground truth of the targets, which include the bounding boxes and categories, a correctly detected target indicates that the IoU of the predicted bounding box and the true bounding box is larger than 0.5 and the predicted category is correct. Examples of the visualization results of CRF-Net and CRFRD using real data are shown in Figure 9. Each row depicts the detection result for a specific scene recorded during the daytime, in cloudy conditions, and at night, respectively. The images in the first column show the ground truth of the object in each scene. The second column showcases the detection results of CRF-Net, while the third column depicts the detection results of CRFRD. The numbers in the ground-truth images represent the distance between the target and the detection platform. The numbers in the detection results of the CRF-Net and CRFRD represent the probability that the target is in a specific category. The yellow dashed ellipses highlight the small objects that are overlooked by CRF-Net but successfully detected by CRFRD and the red dashed ellipses indicate the occluded objects that are missed by CRF-Net but successfully identified by CRFRD.
Figure 9 shows that in the image of the daytime scenario, there are four targets, of which two are correctly detected by CRF-Net, while all four are successfully identified by the proposed CRFRD. CRF-Net misidentifies a car at a distance of 62.3 m and an occluded human, but CRFRD does not exhibit the same limitations. In the image of the cloudy scenario, there are six targets, four of which are correctly detected by CRF-Net, while all six targets are correctly identified by the proposed CRFRD. CRF-Net misidentifies a car at a distance of 34.3 m and an occluded human, while CRFRD successfully identifies them. In the image of the night-time scenario, there are four targets, of which one is correctly detected by CRF-Net, while all four are successfully identified by the proposed CRFRD. CRF-Net fails to detect a car at a distance of 144.3 m, a car at a distance of 51.2 m, and an occluded motorcycle, while CRFRD successfully identifies them. The superiority of the model in detecting small targets and occluded targets is verified by the data collected from real roadside scenes.
By gathering the statistics of all 39 images, the results of each scene can be evaluated. In the daytime scenes, there are 142 targets; of these, 50 targets are correctly identified by CRF-Net, while 74 targets are correctly identified by the proposed CRFRD. In the scene of cloudy conditions, there are 216 targets, of which 150 are successfully identified by CRF-Net, while 201 targets are correctly identified by the proposed CRFRD. In the scenes recorded during the night, there are 128 targets, of which 36 targets are detected correctly by CRF-Net and 67 targets are correctly identified by the proposed CRFRD. The evaluation results verify the superior performance of the CRFRD in various scenes.

4.4. Ablation Study

To verify the effectiveness of the RCE and DCF modules in CRFRD and the influence of θ and σ θ in the RCE module on enhancing the detection performance, several ablation studies are performed.
Table 2 depicts the results of the ablation study of two main modules: RCE and DCF. As shown, the wmAP decreases when any module is removed. When both modules are utilized, the model achieves the highest wmAP of 45.03%, performs best on AP for three categories, and performs the second best on AP for two categories, demonstrating a superior performance compared to any other case in which one or more modules are removed. When RCE module is removed, the radar input is reduced to two channels where the features are not rich enough. When DCF module is removed, during the process of feature extraction and fusion, the feature map has no attention mechanism with which to extract key information and is unable to make full use of the information between channels and spaces, which affects its ability to detect small and occluded objects. Consequently, both the RCE and the DCF modules are indispensable in CRFRD.
Table 3 shows the results of the ablation study about azimuth θ and azimuth-weighted RCS σ θ in the RCE module as additional radar-extended channels along with velocity. The observations reveal that using azimuth-weighted RCS σ θ as a parameter to construct additional radar channels offers the highest wmAP of 44.74%. Specifically, it performs best on AP for three categories and performs second best for two categories, surpassing any other scenarios. These results verify that the utilization of richer redundant information σ θ improves the detection performance more effectively compared to redundant information θ .

5. Conclusions

This paper proposes a camera-radar fusion model named CRFRD, which incorporates an RCE module and a DCF module into the CRF-Net model to enhance object detection performance. By adding the RCE and DCF modules simultaneously, we enrich the radar feature representation and direct the model’s focus toward the important information. In the RCE module, we design an azimuth-weighted RCS parameter that leverages the secondary redundant information to generate richer feature representation and extend three radar channels, namely, the velocity in the x-direction, the velocity in the y-direction, and azimuth-weighted RCS. In the DCF module, the dual-CBAM-FPN structure is proposed, enabling the model to focus on important feature information by inserting CBAM at the input and the fusion process of FPN simultaneously. This is beneficial for the detection of small or occluded objects, since the high-level and low-level features are more effectively fused. Comparative experiments conducted on the NuScenes dataset and real data demonstrate the superior performance of the CRFRD model compared to CRF-Net, particularly with regard to the detection of small and occluded objects, and the wmAP increases from 43.89% to 45.03%. Additionally, our ablation studies verify the indispensability of the RCE and DCF modules and the effectiveness of azimuth-weighted RCS.
Several of the obstacles observed in this study warrant further exploration in the future; for example, the proposed CRFRD method is more likely to overlook or misidentify motorcycles, buses, and trailers. The reason for this may be that azimuth-weighted RCS only enhances the RCS of the targets whose azimuth is close to zero radian; for targets whose azimuth is far beyond zero radian, azimuth-weighted RCS plays a limited role. Therefore, in future work, we will seek more suitable feature representation for targets that have a large azimuth.

Author Contributions

Conceptualization, H.Q. and J.L.; methodology, H.Q. and X.S.; software, Y.J. (Yaoyu Jiang) and H.Q.; validation, Y.J. (Yaoyu Jiang), H.Q. and X.S.; investigation, Y.J. (Yaoyu Jiang) and Y.J. (Yuanfa Ji); data curation, Y.J. (Yaoyu Jiang) and X.S.; Writing—original draft preparation, H.Q. and J.L.; Writing—review and editing, H.Q., J.L. and Y.J. (Yuanfa Ji). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant numbers 62061010, 62161007, 62172350, and U23A20280; Guangxi Science and Technology Department Project, grant numbers AB23026120, AA23062038, and AB24010085; Nanning Scientific Research and Technology Development Project, grant number 20231029; Guangxi Key Laboratory of Precision Navigation Technology and Application, Guilin University of Electronic Technology, grant number DH202230; Key Laboratory of Cognitive Radio and Information Processing, Ministry of Education, Guilin University of Electronic Technology, grant number CRKL220208.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

Author Xiyan Sun was employed by the company GUET-Nanning E-Tech Research Institute Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Bai, J.; Li, S.; Zhang, H.; Huang, L.; Wang, P. Robust Target Detection and Tracking Algorithm Based on Roadside Radar and Camera. Sensors 2021, 21, 1116. [Google Scholar] [CrossRef] [PubMed]
  2. Liu, P.; Yu, G.; Wang, Z.; Zhou, B.; Chen, P. Object Classification Based on Enhanced Evidence Theory: Radar–Vision Fusion Approach for Roadside Application. IEEE Trans. Instrum. Meas. 2022, 71, 1–12. [Google Scholar] [CrossRef]
  3. Lin, J.-J.; Guo, J.-I.; Shivanna, V.M.; Chang, S.-Y. Deep Learning Derived Object Detection and Tracking Technology Based on Sensor Fusion of Millimeter-Wave Radar/Video and Its Application on Embedded Systems. Sensors 2023, 23, 2746. [Google Scholar] [CrossRef]
  4. Ounoughi, C.; Ben Yahia, S. Data Fusion for ITS: A Systematic Literature Review. Inf. Fusion 2023, 89, 267–291. [Google Scholar] [CrossRef]
  5. Kim, Y.; Kim, S.; Choi, J.W.; Kum, D. CRAFT: Camera-Radar 3D Object Detection with Spatio-Contextual Fusion Transformer. Procee. AAAI Conf. Artif. Intell. 2023, 37, 1160–1168. [Google Scholar] [CrossRef]
  6. Dudczyk, J.; Czyba, R.; Skrzypczyk, K. Multi-Sensory Data Fusion in Terms of UAV Detection in 3D Space. Sensors 2022, 22, 4323. [Google Scholar] [CrossRef] [PubMed]
  7. Liu, X.; Li, Z.; Zhou, Y.; Peng, Y.; Luo, J. Camera–Radar Fusion with Modality Interaction and Radar Gaussian Expansion for 3D Object Detection. Cyborg Bionic Syst. 2024, 5, 0079. [Google Scholar] [CrossRef]
  8. Sun, H.; Feng, H.; Stettinger, G.; Servadei, L.; Wille, R. Multi-Task Cross-Modality Attention-Fusion for 2D Object Detection. In Proceedings of the 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC), Bilbao, Spain, 24–28 September 2023; pp. 3619–3626. [Google Scholar] [CrossRef]
  9. Zong, M.; Wu, J.; Zhu, Z.; Ni, J. A Method for Target Detection Based on Mmw Radar and Vision Fusion. arXiv 2024, arXiv:2403.16476. [Google Scholar]
  10. He, S.; Lin, C.; Hu, Z. A Multi-Scale Fusion Obstacle Detection Algorithm for Autonomous Driving Based on Camera and Radar. SAE Intl. J CAV 2023, 6, 333–343. [Google Scholar] [CrossRef]
  11. Nabati, R.; Qi, H. CenterFusion: Center-Based Radar and Camera Fusion for 3D Object Detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 1527–1536. [Google Scholar] [CrossRef]
  12. Chadwick, S.; Maddern, W.; Newman, P. Distant Vehicle Detection Using Radar and Vision. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 8311–8317. [Google Scholar] [CrossRef]
  13. Wang, Z.; Miao, X.; Huang, Z.; Luo, H. Research of Target Detection and Classification Techniques Using Millimeter-Wave Radar and Vision Sensors. Remote Sens. 2021, 13, 1064. [Google Scholar] [CrossRef]
  14. Ni, J.; Liu, J.; Li, X.; Chen, Z. SFA-Net: Scale and Feature Aggregate Network for Retinal Vessel Segmentation. J. Healthc. Eng. 2022, 2022, e4695136. [Google Scholar] [CrossRef] [PubMed]
  15. Wang, J.; Kong, L.; Chang, D.; Kong, Z.; Zhao, Y. Interactive Guidance Network for Object Detection Based on Radar-Camera Fusion. Multimedia Tools Appl. 2024, 83, 28057–28075. [Google Scholar] [CrossRef]
  16. Lo, C.-C.; Vandewalle, P. RCDPT: Radar-Camera Fusion Dense Prediction Transformer. In Proceedings of the IEEE International Conference on Acoustics, Speech, and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar] [CrossRef]
  17. Nobis, F.; Geisslinger, M.; Weber, M.; Betz, J.; Lienkamp, M. A Deep Learning-Based Radar and Camera Sensor Fusion Architecture for Object Detection. In Proceedings of the 2019 Sensor Data Fusion: Trends, Solutions, Applications (SDF), Bonn, Germany, 15–17 October 2019; pp. 1–7. [Google Scholar] [CrossRef]
  18. Cui, C.; Ma, Y.; Lu, J.; Wang, Z. REDFormer: Radar Enlightens the Darkness of Camera Perception with Transformers. IEEE Trans. Intell. Veh. 2024, 9, 1358–1368. [Google Scholar] [CrossRef]
  19. Stacker, L.; Heidenreich, P.; Rambach, J.; Stricker, D. Fusion Point Pruning for Optimized 2D Object Detection with Radar-Camera Fusion. In Proceedings of the 2022 IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 5–7 January 2022; pp. 1275–1282. [Google Scholar]
  20. Xiao, M.; Yang, B.; Wang, S.; Zhang, Z.; Tang, X.; Kang, L. A Feature Fusion Enhanced Multiscale CNN with Attention Mechanism for Spot-Welding Surface Appearance Recognition. Comput. Ind. 2022, 135, 103583. [Google Scholar] [CrossRef]
  21. Stäcker, L.; Mishra, S.; Heidenreich, P.; Rambach, J.; Stricker, D. RC-BEVFusion: A Plug-In Module for Radar-Camera Bird’s Eye View Feature Fusion. In Proceedings of the DAGM German Conference on Pattern Recognition, Heidelberg, Germany, 19–22 September 2023; pp. 178–194. [Google Scholar] [CrossRef]
  22. Li, L.; Xie, Y. A Feature Pyramid Fusion Detection Algorithm Based on Radar and Camera Sensor. In Proceedings of the 2020 15th IEEE International Conference on Signal Processing (ICSP), Beijing, China, 6–9 December 2020; pp. 366–370. [Google Scholar] [CrossRef]
  23. Chang, S.; Zhang, Y.; Zhang, F.; Zhao, X.; Huang, S.; Feng, Z.; Wei, Z. Spatial Attention Fusion for Obstacle Detection Using MmWave Radar and Vision Sensor. Sensors 2020, 20, 956. [Google Scholar] [CrossRef] [PubMed]
  24. Liu, Y.; Chang, S.; Wei, Z.; Zhang, K.; Feng, Z. Fusing mmWave Radar With Camera for 3-D Detection in Autonomous Driving. IEEE Internet Things J. 2022, 9, 20408–20421. [Google Scholar] [CrossRef]
  25. Dang, J.; Tang, X.; Li, S. HA-FPN: Hierarchical Attention Feature Pyramid Network for Object Detection. Sensors 2023, 23, 4508. [Google Scholar] [CrossRef]
  26. Sheng, W.; Yu, X.; Lin, J.; Chen, X. Faster RCNN Target Detection Algorithm Integrating CBAM and FPN. Appl. Sci. 2023, 13, 6913. [Google Scholar] [CrossRef]
  27. Guo, Q.; Liu, J.; Kaliuzhnyi, M. YOLOX-SAR: High-Precision Object Detection System Based on Visible and Infrared Sensors for SAR Remote Sensing. IEEE Sens. J. 2022, 22, 17243–17253. [Google Scholar] [CrossRef]
  28. Cerón, J.C.Á.; Ruiz, G.O.; Chang, L.; Ali, S. Real-Time Instance Segmentation of Surgical Instruments Using Attention and Multi-Scale Feature Fusion. Med. Image Anal. 2022, 81, 102569. [Google Scholar] [CrossRef]
  29. Cui, Z.; Li, Q.; Cao, Z.; Liu, N. Dense Attention Pyramid Networks for Multi-Scale Ship Detection in SAR Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8983–8997. [Google Scholar] [CrossRef]
  30. Han, Y.; Ding, T.; Li, T.; Li, M. An Improved Anchor-Free Object Detection Method. In Proceedings of the 2022 International Conference on Machine Learning and Intelligent Systems Engineering (MLISE), Guangzhou, China, 5–7 August 2022; pp. 6–9. [Google Scholar] [CrossRef]
  31. Lin, T.L.; Piotr, D.; Ross, G.; He, K.M.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR 2017), Honolulu, HI, USA, 22–25 July 2017; pp. 936–944. [Google Scholar]
  32. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.M.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision (ICCV 2017), Venice, Italy, 22–29 October 2017; pp. 770–778. [Google Scholar]
  33. Dong, C.; Duoqian, M. Control Distance IoU and Control Distance IoU Loss for Better Bounding Box Regression. Pattern Recognit. 2023, 137, 109256. [Google Scholar] [CrossRef]
  34. Ganguly, A.; Ruby, A.U.; Chandran J, G.C. Evaluating CNN Architectures Using Attention Mechanisms: Convolutional Block Attention Module, Squeeze, and Excitation for Image Classification on CIFAR10 Dataset. Res. Sq. 2023, 1–13. [Google Scholar] [CrossRef]
  35. Woo, S.; Park, J.; Lee, J.-Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  36. Zhang, J.; Wu, J.; Wang, H.; Wang, Y.; Li, Y. Cloud Detection Method Using CNN Based on Cascaded Feature Attention and Channel Attention. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–17. [Google Scholar] [CrossRef]
  37. Wang, B.; Zhang, S.; Zhang, J.; Cai, Z. Architectural Style Classification Based on CNN and Channel–Spatial Attention. Signal, Image Video Process. 2023, 17, 99–107. [Google Scholar] [CrossRef]
  38. Caesar, H.; Bankiti, V.; Lang, A.H.; Vora, S.; Liong, V.E.; Xu, Q.; Krishnan, A.; Pan, Y.; Baldan, G.; Beijbom, O. nuScenes: A Multimodal Dataset for Autonomous Driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 14–19 June 2020; pp. 11621–11631. [Google Scholar]
  39. Nabati, R.; Qi, H. Radar-Camera Sensor Fusion for Joint Object Detection and Distance Estimation in Autonomous Vehicles. arXiv 2020, arXiv:2009.08428. [Google Scholar]
  40. Gu, Y.; Meng, S.; Shi, K. Radar-Enhanced Image Fusion-Based Object Detection for Autonomous Driving. In Proceedings of the 2022 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xi’an, China, 25–27 October 2022; pp. 1–6. [Google Scholar] [CrossRef]
  41. Feng, B.; Li, B.; Wang, S.; Ouyang, N.; Dai, W. RSA-Fusion: Radar Spatial Attention Fusion for Object Detection and Classification. Multimed. Tools Appl. 2024, 1–20. [Google Scholar] [CrossRef]
  42. Sun, H.; Feng, H.; Mauro, G.; Ott, J.; Stettinger, G.; Servadei, L.; Wille, R. Enhanced Radar Perception via Multi-Task Learning: Towards Refined Data for Sensor Fusion Applications. arXiv 2024, arXiv:2404.06165. [Google Scholar]
  43. Kim, Y.; Shin, J.; Kim, S.; Lee, I.-J.; Choi, J.W.; Kum, D. CRN: Camera Radar Net for Accurate, Robust, Efficient 3D Perception. In Proceedings of the IEEE International Conference on Computer Vision (ICCV 2023), Paris, France, 2–6 October 2023; pp. 17615–17626. [Google Scholar]
Figure 1. Flow diagram of coordinate transformation from radar coordinate system to image coordinate system. The radar points in the radar coordinate system are rotated and translated to the ego coordinate system under the radar timestamp using the radar’s external parameters then converted to the global coordinate system. After that, the global coordinates are transformed to the ego coordinates under the camera timestamp then transformed into camera coordinates using the camera’s external parameters. Finally, the camera coordinates are transformed into image coordinates using the camera’s intrinsic parameters.
Figure 1. Flow diagram of coordinate transformation from radar coordinate system to image coordinate system. The radar points in the radar coordinate system are rotated and translated to the ego coordinate system under the radar timestamp using the radar’s external parameters then converted to the global coordinate system. After that, the global coordinates are transformed to the ego coordinates under the camera timestamp then transformed into camera coordinates using the camera’s external parameters. Finally, the camera coordinates are transformed into image coordinates using the camera’s intrinsic parameters.
Sensors 24 05317 g001
Figure 2. Projection results of radar point clouds onto the image plane: (a) is the projection result prior to the line expansion, whereas (b) is the projection result after the line expansion. The image coordinate system was used here and is denoted as ( y I , z I ) .
Figure 2. Projection results of radar point clouds onto the image plane: (a) is the projection result prior to the line expansion, whereas (b) is the projection result after the line expansion. The image coordinate system was used here and is denoted as ( y I , z I ) .
Sensors 24 05317 g002
Figure 3. CBAM. The input feature is processed by the channel attention module and the spatial attention module in turn to obtain a refined feature. The channel attention module utilizes both max-pooling outputs and average-pooling outputs with a shared network; the spatial attention module utilizes similar two outputs that are pooled along the channel axis and forwarded to a convolution layer.
Figure 3. CBAM. The input feature is processed by the channel attention module and the spatial attention module in turn to obtain a refined feature. The channel attention module utilizes both max-pooling outputs and average-pooling outputs with a shared network; the spatial attention module utilizes similar two outputs that are pooled along the channel axis and forwarded to a convolution layer.
Sensors 24 05317 g003
Figure 4. The block diagram of CRFRD is composed of four modules: input generation, feature extraction and fusion, dual-CBAM-FPN, and classification and regression. Input generation is a generator module for camera and radar channels. The feature extraction and fusion module consists of five feature extraction and fusion blocks, with max-pooling and VGG16 being the most important. Dual-CBAM-FPN combines CBAM and FPN. The classification and regression module fulfills the task of object detection.
Figure 4. The block diagram of CRFRD is composed of four modules: input generation, feature extraction and fusion, dual-CBAM-FPN, and classification and regression. Input generation is a generator module for camera and radar channels. The feature extraction and fusion module consists of five feature extraction and fusion blocks, with max-pooling and VGG16 being the most important. Dual-CBAM-FPN combines CBAM and FPN. The classification and regression module fulfills the task of object detection.
Sensors 24 05317 g004
Figure 5. Comparison of radar image constructed from RCS and azimuth-weighted RCS. The blue dashed and green dashed rectangles indicate two different radar points: (a) shows two points with similar azimuths but different RCS; (b) shows that the points with similar azimuths also have similar azimuth-weighted RCS, indicating that the RCS of the point represented by the green dashed rectangle is enhanced after azimuth weighting.
Figure 5. Comparison of radar image constructed from RCS and azimuth-weighted RCS. The blue dashed and green dashed rectangles indicate two different radar points: (a) shows two points with similar azimuths but different RCS; (b) shows that the points with similar azimuths also have similar azimuth-weighted RCS, indicating that the RCS of the point represented by the green dashed rectangle is enhanced after azimuth weighting.
Sensors 24 05317 g005
Figure 6. DCF is composed of the input module, the multi-scale feature fusion module, and the output module. The input module consists of CBAM and 2D convolution with 1 × 1 convolution kernel and stride 1. The multi-scale feature fusion module is a bottom–up network in which the feature is upsampled, processed by CBAM, and added to the output of the input module. The output module achieves the multi-scale features from the multi-scale feature fusion module as well as from the input module using 2D convolution and relu operation.
Figure 6. DCF is composed of the input module, the multi-scale feature fusion module, and the output module. The input module consists of CBAM and 2D convolution with 1 × 1 convolution kernel and stride 1. The multi-scale feature fusion module is a bottom–up network in which the feature is upsampled, processed by CBAM, and added to the output of the input module. The output module achieves the multi-scale features from the multi-scale feature fusion module as well as from the input module using 2D convolution and relu operation.
Sensors 24 05317 g006
Figure 7. Experimental device and experimental setup: (a) is the experimental device composed of the mmWave radar and the camera; (b) is the experimental setup. The experimental device is raised up to 1.0 m high, and all the data are transferred to a laptop for storage.
Figure 7. Experimental device and experimental setup: (a) is the experimental device composed of the mmWave radar and the camera; (b) is the experimental setup. The experimental device is raised up to 1.0 m high, and all the data are transferred to a laptop for storage.
Sensors 24 05317 g007
Figure 8. Visualization results of CRF-Net and CRFRD. Each row shows the examples of detection result in a distinct scene during the daytime, cloudy conditions, rainy conditions, and night, respectively. The images in the first column show the ground truth of the object in each scene. The second column contains the detection results of CRF-Net, and the third column contains the detection results of CRFRD. The numbers in the ground-truth images represent the distance between the target and the detection platform. The numbers in the detection results of CRF-Net and CRFRD represent the probability that the target will fit into this category. The yellow dashed ellipses indicate the small objects that are missed by CRF-Net but successfully detected by CRFRD, and the red dashed ellipses point out the occluded objects that are missed by CRF-Net but successfully detected by CRFRD.
Figure 8. Visualization results of CRF-Net and CRFRD. Each row shows the examples of detection result in a distinct scene during the daytime, cloudy conditions, rainy conditions, and night, respectively. The images in the first column show the ground truth of the object in each scene. The second column contains the detection results of CRF-Net, and the third column contains the detection results of CRFRD. The numbers in the ground-truth images represent the distance between the target and the detection platform. The numbers in the detection results of CRF-Net and CRFRD represent the probability that the target will fit into this category. The yellow dashed ellipses indicate the small objects that are missed by CRF-Net but successfully detected by CRFRD, and the red dashed ellipses point out the occluded objects that are missed by CRF-Net but successfully detected by CRFRD.
Sensors 24 05317 g008
Figure 9. Examples of the visualization results of CRF-Net and CRFRD using real data. Each row depicts the detection results for a distinct scene recorded during the daytime, in cloudy conditions, or at night, respectively. The images in the first column show the ground truth of the object in each scene. The second column depicts the detection results of CRF-Net, while the third column showcases the detection results of CRFRD. The numbers in the ground-truth images represent the distance between the target and the detection platform. The numbers in the detection results of CRF-Net and CRFRD represent the probability that the target falls within a specific category. The yellow dashed ellipses highlight the small objects that are overlooked by CRF-Net but successfully detected by CRFRD, and the red dashed ellipses indicate the occluded objects that are missed by CRF-Net but correctly identified by CRFRD.
Figure 9. Examples of the visualization results of CRF-Net and CRFRD using real data. Each row depicts the detection results for a distinct scene recorded during the daytime, in cloudy conditions, or at night, respectively. The images in the first column show the ground truth of the object in each scene. The second column depicts the detection results of CRF-Net, while the third column showcases the detection results of CRFRD. The numbers in the ground-truth images represent the distance between the target and the detection platform. The numbers in the detection results of CRF-Net and CRFRD represent the probability that the target falls within a specific category. The yellow dashed ellipses highlight the small objects that are overlooked by CRF-Net but successfully detected by CRFRD, and the red dashed ellipses indicate the occluded objects that are missed by CRF-Net but correctly identified by CRFRD.
Sensors 24 05317 g009
Table 1. The result of comparative experiments. The best results are written in bold text, while the second-best results are underlined.
Table 1. The result of comparative experiments. The best results are written in bold text, while the second-best results are underlined.
ClassHumanBicycleBusCarMotorcycleTrailerTruckwmAPParametersFPS
Model
RetinaNet [32]40.257.1417.1953.3319.365.9525.0043.58
CRF-Net38.4114.9933.6752.9626.6018.3725.8543.89 (43.95 )22.45 M4.12
Nabati and Qi [39] *27.5925.0048.3052.3125.97 34.4544.49
REF-Net [40] 44.76
RSA + CA2 [41]38.8715.1935.3952.4224.0915.0228.5343.92
H. Sun et al. [42] 44.88
CRFRD40.3517.6933.0854.2726.2217.7327.4645.0323.12 M3.97
* The results in this row were reported by Nabati and Qi [39]. The results were reported by CRF-Net [17].
Table 2. Ablation study of two main modules. The √ means that this module is used in the model. The best results are written in bold text, while the second-best results are underlined.
Table 2. Ablation study of two main modules. The √ means that this module is used in the model. The best results are written in bold text, while the second-best results are underlined.
RCEDCFHumanBicycleBusCarMotorcycleTrailerTruckwmAP
38.4114.9933.6752.9626.6018.3725.8543.89
39.6615.2334.7653.7927.6916.2926.4344.74
40.2614.7827.2553.1426.3517.3727.4844.41
40.3517.6933.0854.2726.2217.7327.4645.03
Table 3. Ablation study of θ and σ θ as additional channel/channels in the RCE module. The √ means that this parameter is used in the model. The best results are written in bold text, while the second-best results are underlined.
Table 3. Ablation study of θ and σ θ as additional channel/channels in the RCE module. The √ means that this parameter is used in the model. The best results are written in bold text, while the second-best results are underlined.
θ σ θ HumanBicycleBusCarMotorcycleTrailerTruckwmAP
39.0713.2028.6353.4022.9118.2126.8244.20
39.6414.4431.4152.9328.5017.0529.1844.49
39.6615.2334.7653.7927.6916.2926.4344.74
39.9013.2830.2053.6820.4016.8626.3444.52
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Sun, X.; Jiang, Y.; Qin, H.; Li, J.; Ji, Y. Camera-Radar Fusion with Radar Channel Extension and Dual-CBAM-FPN for Object Detection. Sensors 2024, 24, 5317. https://doi.org/10.3390/s24165317

AMA Style

Sun X, Jiang Y, Qin H, Li J, Ji Y. Camera-Radar Fusion with Radar Channel Extension and Dual-CBAM-FPN for Object Detection. Sensors. 2024; 24(16):5317. https://doi.org/10.3390/s24165317

Chicago/Turabian Style

Sun, Xiyan, Yaoyu Jiang, Hongmei Qin, Jingjing Li, and Yuanfa Ji. 2024. "Camera-Radar Fusion with Radar Channel Extension and Dual-CBAM-FPN for Object Detection" Sensors 24, no. 16: 5317. https://doi.org/10.3390/s24165317

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop