Next Article in Journal
Development of a Rehabilitation Chair Design Based on a Functional Technology Matrix and Multilevel Evaluation Methods
Previous Article in Journal
Traffic-Driven Proactive Slot Orchestration in an Optical Network with Fine-Granularity Flexible Ethernet Interfaces
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

YOLOv7-3D: A Monocular 3D Traffic Object Detection Method from a Roadside Perspective

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
Institute of Applied Electronics, CAEP, Mianyang 621900, China
*
Authors to whom correspondence should be addressed.
Appl. Sci. 2023, 13(20), 11402; https://doi.org/10.3390/app132011402
Submission received: 28 August 2023 / Revised: 10 October 2023 / Accepted: 13 October 2023 / Published: 17 October 2023

Abstract

:
Current autonomous driving systems predominantly focus on 3D object perception from the vehicle’s perspective. However, the single-camera 3D object detection algorithm in the roadside monitoring scenario provides stereo perception of traffic objects, offering more accurate collection and analysis of traffic information to ensure reliable support for urban traffic safety. In this paper, we propose the YOLOv7-3D algorithm specifically designed for single-camera 3D object detection from a roadside viewpoint. Our approach utilizes various information, including 2D bounding boxes, projected corner keypoints, and offset vectors relative to the center of the 2D bounding boxes, to enhance the accuracy of 3D object bounding box detection. Additionally, we introduce a 5-layer feature pyramid network (FPN) structure and a multi-scale spatial attention mechanism to improve feature saliency for objects of different scales, thereby enhancing the detection accuracy of the network. Experimental results demonstrate that our YOLOv7-3D network achieved significantly higher detection accuracy on the Rope3D dataset while reducing computational complexity by 60%.

1. Introduction

With the continuous development and intelligence of urban transportation, the safety and efficiency of autonomous driving have become a focal point for researchers [1,2,3,4]. Visual traffic monitoring systems, as an essential component of traffic management and safety, have broad application prospects. These systems [5,6,7,8] utilize cameras and other sensor devices to capture real-time traffic scenes and analyze and process images or videos using computer vision algorithms to extract traffic information and key data. Visual traffic monitoring systems consist of functions such as vehicle detection, vehicle tracking, license plate recognition, and behavior analysis, which are used to extract traffic information and features. Furthermore, they enable functions such as vehicle count, traffic flow monitoring, congestion detection, and traffic accident warning through the utilization of traffic information and features [9,10]. Visual traffic monitoring systems have extensive applications in traffic management, urban planning, and traffic safety. They can assist traffic management departments in better monitoring and managing traffic flow, improving road utilization efficiency, reducing traffic congestion and the probability of accidents, and providing a safer and more convenient traffic environment for urban residents.
Traditional traffic monitoring systems primarily rely on 2D image processing [11,12,13], which can only provide limited information, thereby limiting the accurate recognition and stereo perception capabilities of traffic targets. To address this issue, in recent years, researchers have started to incorporate 3D object detection techniques from the computer vision field into roadside monitoring viewpoints to provide more comprehensive and accurate traffic object information [14,15]. Single-camera 3D traffic object detection, as one of the methods, estimates the three-dimensional positions and orientations of traffic objects based on the texture and contextual information in the images, in the case of a single camera.
There are several advantages to using single-camera 3D object detection in roadside traffic monitoring systems:
(1)
Spatial perception: By using 3D detection, the system can accurately perceive the precise position, size, and orientation information of vehicles, enabling accurate perception of vehicles in three-dimensional space. This helps to better understand vehicle motion behavior and spatial layout.
(2)
Distance estimation: 3D detection provides distance estimation between vehicles and monitoring cameras. This is crucial for evaluating key parameters such as distance, speed, and acceleration between vehicles and cameras. Such information is vital for traffic flow statistics, behavior analysis, and accident warning applications.
(3)
Enhanced safety: Through 3D detection, the monitoring system can more accurately identify the position and motion status of vehicles, providing more reliable traffic information. This helps in the timely detection of traffic violations, abnormal driving behaviors, and accident risks, enabling appropriate warnings and measures to be taken and improving road safety.
(4)
Cost-effectiveness: Compared to multi-sensor fusion solutions, monocular vision systems have lower costs. They only require a single camera to capture image information without the need for additional sensor devices, thereby reducing installation and maintenance costs.
However, the use of a monocular 3D object detection method requires precise camera intrinsic and extrinsic parameters during the model training and deployment process, and the dataset annotation process is more cumbersome. However, these limitations cannot overshadow the advantages of this algorithm.
Moreover, single-camera 3D object detection faces several challenges and difficulties. Firstly, cameras in roadside monitoring systems have different installation heights, pitch angles, and complex road environments. Due to the limitations of a single-camera viewpoint, the camera can only provide a limited field of view, leading to situations where targets are occluded, or there are variations in the viewpoint, thereby increasing the complexity of object detection. Secondly, due to the lack of accurate depth information in images, relying solely on 2D images for object detection makes it difficult to obtain the true size and shape of objects, affecting precise localization and pose estimation. Additionally, environmental factors such as lighting changes and weather conditions can also affect the accuracy of single-camera 3D object detection. Furthermore, real-time performance is also highly demanded for traffic object detection algorithms. For example, Adib Hosseiny et al. [16] achieved accelerated deployment of the YOLOv7-tiny algorithm on FPGAs using high-level synthesis (HLS) tools. This method significantly reduced the usage of digital signal processing (DSP) units and flip-flops, achieving remarkable real-time application latency.
To address the challenges of single-camera 3D object detection, this paper thoroughly investigates and explores intelligent transportation systems and object detection. We propose a novel algorithm based on single-camera vision to address the issue of poor accuracy and precision in object detection. The algorithm utilizes information such as the 2D bounding box of the traffic object, projected corner keypoints, and offset vectors relative to the center of the 2D bounding box to assist in detecting 3D object bounding boxes. Additionally, we introduce a 4-layer feature pyramid network (FPN) structure and a multi-scale spatial attention mechanism to enhance the saliency of features for objects at different scales and improve detection accuracy.
Through experimental evaluations on roadside monitoring datasets [14], our method achieves significant performance results in single-camera 3D traffic object detection. In conclusion, single-camera 3D traffic object detection from a roadside monitoring viewpoint has significant research and application value. By overcoming the limitations and challenges of a single-camera viewpoint, our algorithm can achieve more accurate and reliable traffic object detection, providing a more reliable guarantee for urban traffic safety and efficiency. In the future, we will further optimize algorithm performance and explore integration with other traffic monitoring systems to achieve more intelligent and efficient traffic management and control.

2. Related Work

There are two categories of single-camera 3D object detection methods: pseudo-LIDAR/depth-based methods and methods that do not introduce additional information.
Pseudo-LIDAR/Depth-based methods: Due to the lack of depth information in monocular vision, the accuracy of these methods is significantly lower compared to methods that incorporate LIDAR and image fusion. Therefore, some pseudo-LIDAR/depth-based methods introduce additional depth estimation modules and/or point cloud assistance [17,18,19,20]. These methods convert image pixels into pseudo-LIDAR point clouds using existing depth estimation techniques, allowing them to leverage LIDAR-based methods for object detection. Ref. [21] introduces a method to generate pseudo-LIDAR data from visual depth estimation, effectively improving the accuracy of 3D object detection in autonomous driving scenarios requiring simulated LIDAR sensors. Mono3D-PLiDAR [22] proposes a novel method that utilizes urban maps to enhance long-range 3D object detection performance. In order to improve the performance of monocular 3D target detection, refs. [23,24,25] use the method of converting 2D images to pseudo-LIDAR point clouds. The other method is DA-3Ddet [26], which adjusts features from the unreliable image-based pseudo-LIDAR domain to the reliable LIDAR domain to improve monocular 3D object detection performance.
Methods without introducing additional information: M3D-RPN uses an anchor-based approach, utilizing a series of predefined 3D bounding box positions called “anchors” and estimating offsets relative to these anchors. One of these methods is M3D-RPN [27], which employs a 3D region proposal network and geometric constraints from both 2D and 3D perspectives to directly regress the 3D position and size of the objects. MonoDLE [28] proposes the use of 3D projected center coordinates to aid in estimating coarse center-aware 3D geometric information. MonoCon [29], an extension of MonoDLE, incorporates a 2D auxiliary learning module for projected objects and introduces Attentive Normalization to all heads. MonoFlex [30] combines the idea of predicting object depth using height ratios with uncertainty theory to construct an ensemble learning approach for estimating the object center position. Arsalan Mousavian et al. [31,32] propose a hybrid approach combining deep learning and geometry to accurately estimate 3D bounding boxes for objects in images. Refs. [33,34,35,36] focus on improving object detection in the context of monocular images. They aim to enhance the accuracy and efficiency of detecting 3D objects using deep learning techniques. RTM-3D [37] and MonoGRNet [14] adopt keypoint-based methods aiming to directly regress keypoints and optimize the estimation of 3D bounding box size and position from the keypoints’ image locations.
However, it is worth noting that existing single-camera 3D detection methods are primarily designed for autonomous driving scenarios and have not been optimized for roadside monitoring scenes. Therefore, this paper proposes the YOLOv7-3D algorithm, which can simultaneously predict the 2D and 3D bounding boxes of traffic objects.

3. Proposed Method

First, we provide a task description of monocular 3D traffic object detection from a roadside viewpoint. Then, we present detailed descriptions of the model architecture and the design of the loss function in this paper.

3.1. Task Description

In this work, we aim to detect 3D bounding boxes of vehicle objects in RGB images using the YOLOv7-3D model. Given an input image I R H × W × 3 , camera intrinsic parameters K R 3 × 3 , and camera extrinsic parameters P R 3 × 4 , the output of the YOLOv7-3D model provides 3D bounding boxes for the traffic objects in the image, which can be represented as B = { B 1 , B 2 , , B n } , where each bounding box B i consists of seven degrees of freedom:
B i = x i , y i , z i , l i , w i , h i , θ i
Here, ( x , y , z ) represents the position of the center point of each 3D bounding box in meters. ( l , w , h ) denote the length, width, and height of the cuboid, respectively, in meters. θ represents the global orientation of each traffic object in space, which is the angle between the object’s heading direction and the x-axis of the camera coordinate system, in the range of [ π , π ] .
The camera intrinsic parameters K R 3 × 3 include parameters such as focal lengths, optical centers, and pixel sizes, which describe the optical characteristics and imaging process inside the camera. f u and f v are the focal lengths, while c u and c v represent the pixel coordinates of the optical center.
K = f u 0 c u 0 f v c v 0 0 1
The camera extrinsic parameters describe the camera’s position and orientation in the world coordinate system. It consists of the translation vector T R 3 × 1 and the rotation matrix R R 3 × 3 , which map points in the camera coordinate system to the world coordinate system.
For a 3D point X = [ x , y , z ] T , it can be projected onto the image plane as Y = [ u , v , 1 ] T using the camera projection matrix:

3.2. Model Structure

In terms of the model structure, YOLOv7-3D can be divided into three parts: the backbone feature extraction network, the multi-scale feature pyramid, and the multi-branch decoupled detection head. Due to the significant scale differences among various traffic objects in the Rope3D dataset, this paper proposes a 5-layer FPN structure and a multi-scale spatial attention mechanism (MSSA) to improve feature saliency and detection accuracy for different-scale objects. The overall network architecture of the model is shown in Figure 1.
In this work, the model does not directly predict the seven degrees of freedom of the targets ( x , y , z , l , w , h , θ ) . Instead, it predicts nine degrees of freedom for the targets ( x c , y c , φ z , σ z , l , w , h , sin ( θ ) , cos ( θ ) ) . Here, x c and y c represent the projected points of the center of the 3D bounding box on the 2D image. φ z and σ z represent the depth value and the logarithmic variance of the depth value, respectively. The introduction of these parameters aims to quantify the uncertainty in the predictions by considering the variance. l, w, and h represent the length, width, and height of the 3D bounding box, respectively, while sin ( θ ) and cos ( θ ) are values required to compute the heading angle. The schematic diagram of degrees of freedom is shown in Figure 2.

3.2.1. Multi-Scale Feature Pyramids

The backbone network is primarily responsible for extracting features from the image and is structured as shown in the diagram. It consists of CBS convolutional layers, E-ELAN convolutional modules, MP-Conv convolutional modules, and SPPCSPC modules. The CBS convolutional layer comprises a convolutional layer, batch normalization layer (BN), and SiLU activation function, which are used to extract features at different scales. The E-ELAN convolutional module is an efficient aggregation network that modifies the calculation block while preserving the original transition layer structure. It enhances the network’s learning capacity without disrupting the original gradient paths through arithmetic techniques and allows the calculation blocks of different feature groups to learn more diverse features. The MP-Conv convolutional module combines the results of downsampling using max pooling and convolutional blocks, reducing computational complexity while increasing the receptive field and effectively propagating global information backward. The SPPCSPC module includes a spatial pyramid pooling layer that adapts to different resolutions using max pooling at four different scale sizes, distinguishing objects of different sizes.
In this paper, an additional layer, P 6 R 11 × 19 × 1024 , is added to the original backbone feature network to detect large objects such as buses. Additionally, a larger-scale feature P 2 R 176 × 304 × 64 is utilized to detect smaller vehicle targets at a greater distance from the camera. Moreover, the channel numbers in the backbone feature extraction network are halved to avoid an excessive increase in parameters and computational complexity caused by increasing the depth of the model.

3.2.2. Multi-Scale Spatial Attention Mechanism

Y = K · [ R | T ] · X
FPN (Feature Pyramid Network) is a feature pyramid network used for multi-scale object detection. It constructs multi-scale feature maps by adding extra lateral connections in the Backbone network. These lateral connections extract features from different levels of the Backbone feature maps and merge them into the feature maps of the previous layer. This way, FPN can obtain rich semantic information at different scales and provide feature maps with different resolutions. This is crucial for detecting objects of different scales, as objects may have varying sizes and proportions. In this paper, the original 3-layer FPN structure in the YOLOv7 algorithm is modified to a 5-layer structure to accommodate more scale variations of the targets. As shown in the diagram, the upsampling structure of the FPN feature pyramid consists of CBS, UpSample, Concat, CatConv, and the multi-scale spatial attention mechanism.
Attention mechanisms are widely used to improve the performance of deep learning models by selectively focusing on relevant information and suppressing irrelevant or noisy information [38,39,40]. However, these attention mechanisms overlook the scale information of feature maps. This paper proposes a Multi-Scale Spatial Attention mechanism (MSSA), which adaptively adjusts the weights of features from P2 to P6. It increases the weights of scale features that are more beneficial for the recognition task while suppressing the weights of other scale features. Spatially, the model can focus on image textures and contextual information that are advantageous for the recognition task. The structure of MSSA is shown in the Figure 3. MSSA performs average pooling on the input 2D feature map G R C × ( H × W ) , with pooling kernel sizes of K h and K w , and strides of S h and S w along the horizontal and vertical directions, respectively. This process generates condensed features W avg c and H avg c . Subsequently, the features from these two condensed layers are aggregated:
(1) First, concatenate the compacted features. Since the dimensions of features W avg c and H avg c do not match, the width and height dimensions of feature H avg c need to be transposed before concatenating it with H avg c to obtain the feature map M.
(2) Second, set a hyperparameter r such that M is passed through a 2D convolution to obtain the feature map M 1 , where the number of channels changes from c to c r In this paper, r is set, and the number of channels in M 1 should not be less than 8. Then, insert a BN layer and a SiLU activation function to obtain the feature map M 2 . At this stage, M 2 incorporates both the feature compaction of the input feature G along the x-axis and y-axis, allowing spatial information of the input feature G to interact.
The mixed spatial information in M 2 is then divided, transposed, and passed through another 2D convolution to restore the channel number to c, resulting in W and H . These two feature maps represent the spatial weights. Finally, element-wise multiplication is performed between W , H , and the corresponding elements of matrix G to obtain G . This process combines the spatial weights in the input feature map, where the spatial weights beneficial for the recognition task are increased. In the diagram, ⊙ represents element-wise multiplication between matrices. Thus, the spatial information S and channel information C are adaptively adjusted. This process can be represented as:
G = σ ( f ( [ A v g P o o l W ( G ) ; A v g P o o l H ( G ) ] ) ) = σ f W avg c ; H a v g c
where σ represents the sigmoid function, ⊗ represents the convolution process, and f is a convolution kernel.
Next, the feature map G is adaptively average-pooled at scale L to obtain a compacted feature of size 2 × 1 × 1w. Then, an MLP is used to adjust the compacted feature, and a sigmoid function is applied to obtain the activated weights. Finally, the weights are added to the original feature map G to obtain the feature map G . This process adapts the weights at scale L. It can be represented as:
G = σ ( f ( [ A v g P o o l L ( G ) ] ) ) = σ ( f ( [ W avg l ] ) )

3.2.3. Multi-Branch Decoupling Detection Head

In object detection, the conflict between classification and regression tasks is a common problem. Therefore, the decoupling of classification and localization heads has been widely applied in both one-stage and two-stage detectors. However, despite the evolution of the YOLO series in terms of backbone and feature pyramid, its detection head remains coupled, as shown in Figure 4. In this paper, a multi-branch decoupled detection head is used in the detection stage. The model’s output includes predictions for 2D bounding box regression, object classification, projection of the 3D box’s center point ( x c , y c ) onto the 2D image, depth value and logarithmic variance ( φ z , σ z ) of the depth value, scale of the 3D box in terms of length, width, and height ( l , w , h ) , and the values required to compute the heading angle sin ( θ ) and cos ( θ ) . The decoupled detection head consists of the following branches:
(1)
Reg2D(h, w, 4): Used to determine the regression parameters for each feature point, allowing adjustment of the predicted bounding boxes.
(2)
Obj(h, w, 1): Used to determine whether each feature point contains an object.
(3)
Cls(h, w, num classes): Used to determine the object class for each feature point.
(4)
CenterReg3D(h, w, 2): Used to determine the regression parameters for the 2D projection of the 3D object’s center point on the image.
(5)
Depth(h, w, 2): Used to determine the depth value and logarithmic variance for the 3D object.
(6)
Dim(h, w, 3): Used to determine the 3D dimensions (length, width, height) of the object.
(7)
Theta(h, w, 2): Used to determine the viewing angle of the 3D object.
Figure 4. Schematic diagram of Decoupled Head.
Figure 4. Schematic diagram of Decoupled Head.
Applsci 13 11402 g004
Finally, the seven prediction results are stacked together. The results obtained for each feature layer are of size Out(h, w, 13 + 1 + num classes). The first four parameters are used to determine the 2D regression parameters for each feature point, allowing adjustment of the predicted bounding boxes. The fifth parameter is used to determine whether each feature point contains an object. Parameters six to thirteen are used to obtain the 9 degrees of freedom for the object in 3D space. Finally, the last num class parameters are used to determine the object class for each feature point.

3.3. Loss Function

In this work, the model does not directly predict the seven degrees of freedom of the object ( x , y , z , l , w , h , θ ) but predicts nine degrees of freedom ( x c , y c , φ z , σ z , l , w , h , sin ( θ ) , cos ( θ ) ) . Here, ( x c , y c ) represent the projected 2D coordinates of the center point of the 3D bounding box on the image. ( φ z , σ z ) denote the depth compensation value and the logarithmic variance of the depth value. The introduction of σ z aims to quantify the uncertainty of the predictions by considering the variance. ( l , w , h ) represent the length, width, and height of the 3D bounding box. ( sin ( θ ) , cos ( θ ) ) are used to compute the heading angle.
Classification loss function: The paper introduces Focal Loss to address the issue of class imbalance in the Rope3D dataset. Focal Loss reduces the weight assigned to easily classified samples and increases the weight assigned to hard samples in order to focus more on challenging samples. This helps improve the performance of the model in scenarios with class imbalance:
L c l s = α t ( 1 P t ) γ l o g ( P t )
where P t is the predicted probability by the model, indicating the probability of the sample belonging to the target class. α t is the balance factor used to adjust the weights of positive and negative samples. In general, it is defined as the ratio of the number of positive samples to the total number of samples. γ is the focusing parameter used to adjust the weight difference between easily classified samples and hard samples. A larger value of γ will make the model pay more attention to hard samples.
2D box regression loss function: For the prediction of the 2D bounding box, this paper uses the CIoU (Complete Intersection over Union) loss.
L I O U = 1 I O U + ρ 2 b , b * c 2 + α v
where IOU represents the traditional Intersection over Union, ρ 2 ( b , b * ) denotes the Euclidean distance between the center points of the predicted box and the ground truth box. α is a balance factor used to adjust the weights of the position and shape losses, and v represents the position and shape loss. The specific formula is as follows:
α = v 1 I O U + v v = 4 π 2 arctan w * h * arctan w h 2
Projection center point loss function: For the projection of the 3D object center point onto the 2D image, this paper uses L1 loss for direct regression. Unlike previous methods, this paper modifies the prediction of the center point to be an offset relative to the center point of the 2D bounding box rather than an offset relative to the entire image. Here, P represents the ground truth projected point coordinates, and P * represents the predicted projected point coordinates.
L c e n t e r = P P * 1
Dimension-aware loss function: For the prediction of the 3D bounding box dimensions, this paper employs a dimension-aware L1 loss function that is suitable for size estimation. The loss function is defined as follows:
L d i m = s s * s 1
Here, · 1 denotes the L1 norm. Previous algorithms used a baseline L1 loss, i.e., L d i m = s s * 1 . It is worth noting that, compared to the baseline L1 loss L d i m , the dimension-aware loss used in this paper incorporates dynamic compensation generation. By re-computing L d i m , compensation weights w d i m = L d i m / L d i m are generated to make the average value of the final loss function L d i m equal to the standard loss. In this way, the proposed loss can be seen as a redistribution of the standard L1 loss.
Viewpoint loss function: Instead of directly predicting the object’s yaw angle (the global orientation angle of the object in the camera coordinate system, ranging from π to π ), this paper chooses to regress the observation angle θ for each object. The observation angle θ represents the angle between the object’s direction and the camera axis, with the camera origin as the center and the line connecting the camera origin to the object’s center as the radius, after rotating the object around the camera’s y-axis to align with the camera’s z-axis. The transformation relationship between these two angles is shown in Figure 5. In addition, each is encoded as a vector. Using the target position, the yaw angle can be obtained: Furthermore, each θ is encoded as a vector sin ( θ ) , cos ( θ ) T . By using θ and the object’s position, the yaw angle r y can be obtained:
r y + π 2 θ = α + π 2 α = r y θ
The final viewpoint loss function computes the loss using SmoothL1Loss on the normalized sin ( θ ) , cos ( θ ) T .
Depth loss function: In 3D object detection, depth estimation is a task with uncertainty since the depth of an object in real-world scenarios can be influenced by various factors such as viewpoint, occlusion, and image noise. This paper utilizes the Laplacian Aleatoric Uncertainty Loss to predict the depth of 3D objects, quantifying the uncertainty of the predictions by introducing a variance term. This loss reduces the weight of depth predictions with higher variance (i.e., greater uncertainty), thereby increasing the accuracy and reliability of depth predictions with lower uncertainty.
L d e p t h = 1 | P | ( x c , y c ) P 2 σ z z z * + log σ z
Here, P represents the set of real 2D bounding box centers, ( φ z , σ z ) denotes the depth offset and logarithmic variance, z is the true depth value of the object, and z * is the predicted depth value of the object. The predicted depth value z * is obtained by reversing the sigmoid transformation of the model’s predicted depth offset φ z back to the original linear range:
z * = 1 s i g m o i d ( φ z ) + ϵ 1
By combining the projected points of the center on the feature map ( x c , y c ) , the inverse matrix of the camera intrinsic parameters K 1 R 3 × 3 , the 3D center position of the object in meters can be calculated as:
x y z 3 D = K 3 × 3 1 z x c z y c z 2 D
Once the seven degrees of freedom for the 3D bounding box are obtained, the coordinates of the eight corners of the 3D box on the 2D image can be calculated as:
B = R θ ± h / 2 ± w / 2 ± l / 2 + x y z
Here, R θ is the rotation matrix for the yaw angle, [ h , w , l ] T represents the object’s 3D dimensions, and [ x , y , z ] T represents the object’s center coordinates.The overall loss is the sum of all the individual loss terms, with each loss term having a balancing weight parameter.

4. Experiments

In this section, the paper describes the details of model training and the effectiveness of data augmentation, conducts ablation experiments, and analyzes the experimental results. Additionally, in the last subsection, a quantitative comparison is made between YOLOv7-3D and some state-of-the-art monocular 3D object detection algorithms. The paper employed 4 NVIDIA GeForce RTX 3090 GPUs for training the model with a batch size of 4 and a total of 300 epochs. The SGD optimizer with betas ( 0.9 , 0.009 ) , an initial learning rate of 0.01, and a weight decay of 0.0005 was used.

4.1. Dataset

This paper uses the Rope3D dataset, which is based on roadside surveillance perspectives. This dataset has collected a total of 50,009 image frames under different times (day, night, dusk/dawn), different weather conditions (sunny, cloudy, rainy), different densities (crowded, normal, sparse), and different distributions of traffic elements, etc. It has the characteristics of being large-scale and multi-view. Some of its samples are shown in Figure 6. The dataset includes 13 object categories with their corresponding labels, 2D attributes (occlusion, truncation), and 7 degrees of freedom for 3D bounding boxes: position (x, y, z), dimensions (width W, length L, height H), and orientation (heading angle θ ). Additionally, the Rope3D dataset presents greater detection challenges due to significant occlusion, as shown in Figure 6. Over half of the objects are partially or fully occluded, whereas the occlusion percentage in the KITTI dataset [41] ranges from 5% to 30%. This highlights the increased difficulty of detection in the Rope3D dataset.
Additionally, this paper conducted a statistical analysis of the distribution of different object categories in the training and validation sets of the Rope3D dataset, as shown in Figure 7. It was found that there is a significant class imbalance in the dataset. To address this issue and improve the classification accuracy, the study introduced the Focal Loss technique.
Furthermore, this study employed the K-means++ algorithm to cluster the sizes of the target objects in the Rope3D dataset into several clusters, with each cluster’s centroid size serving as the prior box size. As shown in Figure 8, there is a significant variation in size within each cluster and distinct scale differences and boundaries between clusters. This observation reflects the substantial diversity in object scales within the dataset. Therefore, this study designed a 5-layer Feature Pyramid Network (FPN) architecture and a multi-scale spatial attention mechanism.

4.2. Real-Time Data Augmentation

Data augmentation is a common technique in 2D object detection to increase training data diversity and enrich the model’s robustness and generalization capabilities. In this paper, the commonly used 2D data augmentation methods were adapted to 3D object detection. Random scaling, translation, horizontal flipping, color transformations, and brightness adjustments were implemented for 3D objects. The scaling range was set from 0.75 to 2. The results, as shown in Figure 9, present the 2D bounding boxes in green, the 3D bounding boxes in white, black dots representing the center of the 2D boxes, red dots representing the center of the 3D boxes, and green dots representing the center of the bottom face of the 3D boxes. The gaps resulting from scaling were filled with gray pixels.

4.3. Experiments and Analysis

In the experiments, this paper primarily compared its approach with the following monocular visual 3D object detection algorithms: (1) KM3D, which employed ResNet34 as the backbone network. (2) Kinematic3D, a video-based monocular 3D object detector utilizing DenseNet121 as the backbone network. (3) MonoDLE, a center-point-based single-stage detector based on CenterNet, employing DLA34 as the backbone network.
In the experimental setup, “Ours Base” represents the improved baseline model of YOLOv7 capable of simultaneously detecting 2D and 3D objects. “Ours +5Head” indicates the model with five detection heads built upon the base model. “Ours +MSSA” signifies the YOLOv7-3D model incorporating the multi-scale spatial attention mechanism on top of the previous model.
The comparison of the experimental and ablation study results is shown in Table 1. The analysis of experimental results reveals that KM3D achieved lower average precision across all categories. It performed relatively well in the Cyclist and Motorcyclist categories but poorly in the Bus and Truck categories. Kinematic3D, on the other hand, obtained higher average precision compared to KM3D across all categories. It performed well in the Cyclist, Motorcyclist, and Pedestrian categories but poorly in the Van and Bus categories. MonoDLE achieved higher average precision than KM3D and Kinematic3D across all categories. It performed well in the Cyclist, Motorcyclist, and Pedestrian categories. Ours Base showed poor performance in the Car, Van, and Bus categories but relatively good performance in the Cyclist, Motorcyclist, and Pedestrian categories. Adding five detection heads to the Base model resulted in higher average precision across all categories compared to Ours Base. It performed well in the Cyclist, Motorcyclist, and Pedestrian categories. Finally, incorporating MSSA resulted in the highest average precision across all categories. It performed best in the Cyclist, Motorcyclist, and Pedestrian categories. Overall, based on the given experimental data, the Ours + MSSA method achieved the best average precision across multiple categories. Additionally, categories such as Bus achieved lower precision compared to the aforementioned methods due to the imbalance in the dataset, where the Bus and Truck categories have a very small number of instances, resulting in less training exposure for the model to learn other categories effectively.
The model’s parameter count and computational load are crucial metrics for assessing algorithm detection speed. The results indicate that Ours Base has a relatively large parameter count compared to other models. However, by modifying it to incorporate five detection heads and halving the main channel count, the model’s parameters increased from 37M to 42M. Notably, the computational load decreased significantly from 222G to 87G, accompanied by improvements in the supervised accuracy for various categories. In contrast, Ours + MSSA showed only a slight increase in parameter count yet achieved enhanced detection accuracy for different target categories. This suggests that Ours + MSSA improves model performance without introducing a substantial number of redundant parameters and computational burden. Considering accuracy, model parameter count, and computational load comprehensively, Ours + MSSA demonstrates outstanding performance in traffic object detection tasks. The method enhances detection accuracy while maintaining a relatively modest model complexity, showcasing excellent practicality.

4.4. Detection Performance in Different Ranges

To further analyze the performance of the models in different distance ranges, this study employed the Score metric, which encompasses the following components: ACS (Average Center Score): This represents the average score of the object’s center point. It is calculated based on the distance or similarity between the estimated center point and the ground truth center point. ACS is used to evaluate the accuracy of the object detection algorithm in estimating the object’s center point.
A C S = 1 | D | s D 1 min 1 , Δ s c C s
D represents the set of true positive samples, C s is the norm of the ground truth center, and Δ s c is the Euclidean distance between the predicted ground center and the ground truth center for sample s. | D | is the total number of true positive objects. AOS (Average Overlap Score): It represents the average overlap score of the object bounding boxes. It is computed by measuring the overlap between the detected bounding boxes and the ground truth bounding boxes. AOS is used to evaluate the accuracy of object localization.
A A S = 1 | D | s D 1 min 1 , Δ s A A s
Δ s θ represents the angular difference for sample s, and cos 2 Δ s θ accounts for the fact that during evaluation, no distinction is made between whether the head or tail of an object is facing the camera. ASS (Average Size Score): It represents the average score for object size. It is calculated by measuring the size difference between the detected bounding boxes and the ground truth bounding boxes. Δ s A is the absolute area difference, and A s is the ground truth area. ASS evaluates the accuracy of object size estimation.
A O S = 1 | D | s D 1 + cos 2 Δ s θ 2
AGS (Average Four Ground Points Distance and Similarity): This represents the average geometric distance score (relative value) for object shape. It is calculated by measuring the geometric distance between the detected shape and the ground truth shape of the object. s g ^ and s g represent the g-th predicted point and the g-th ground truth point for sample s, respectively. K = 4 denotes the total number of ground truth points.
A G S = 1 | D | s D 1 min 1 , 1 K g = 0 K 1 s g s g ^ | c ^ |
(ACS + AOS + ASS + AGS)/4.0: It represents the overall score, which is the average of the four mentioned scores. This overall score can be used to comprehensively evaluate the performance of the object detection algorithm.
As shown in Table 2, the model achieves high detection accuracy in the mid-range distances, particularly within the range of 30–60 m. However, the detection accuracy is lower in the range of 90–120 m. This can be attributed to two reasons: the region with strong features for learning in the image is too small, and there is a significant reduction in 3D annotations due to occlusion in distant regions. MonoDLE performs well in pedestrian detection, while the proposed algorithm in this paper shows significant improvement in Car, Big Vehicle, and Cyclist categories.

5. Conclusions

The paper proposes a method called YOLOv7-3D for single-camera 3D traffic object detection in roadside monitoring scenarios. Through experiments and evaluations on the Rope3D dataset, our YOLOv7-3D algorithm achieves significant performance results in single-camera 3D traffic object detection. Compared to traditional methods based on 2D image processing, our algorithm accurately captures the position, size, and orientation information of traffic objects, as well as estimates the distance between vehicles and the camera. By overcoming the limitations and challenges of a single-camera perspective, our algorithm enables more accurate and reliable traffic object detection, providing a more reliable guarantee for urban traffic safety and efficiency. Future research directions include further optimizing algorithm performance and exploring integration with other traffic monitoring systems to achieve more intelligent and efficient traffic management and control.

Author Contributions

Conceptualization, software, formal analysis, investigation, resources, data curation, writing—original draft preparation, Z.Y.; validation, Z.Y., H.Z., J.G., and X.L.; writing—review and editing, supervision, project administration, funding acquisition, H.Z. and J.G.; visualization, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the Innovation and Development Fund Project of the China Academy of Engineering Physics(CX2020033).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Cui, J.; Qiu, H.; Chen, D.; Stone, P.; Zhu, Y. Coopernaut: End-to-end driving with cooperative perception for networked vehicles. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 July 2022; pp. 17252–17262. [Google Scholar]
  2. Huang, J.; Huang, G.; Zhu, Z.; Ye, Y.; Du, D. Bevdet: High-performance multi-camera 3d object detection in bird-eye-view. arXiv 2021, arXiv:2112.11790. [Google Scholar]
  3. Yu, H.; Luo, Y.; Shu, M.; Huo, Y.; Yang, Z.; Shi, Y.; Guo, Z.; Li, H.; Hu, X.; Yuan, J.; et al. Dair-v2x: A large-scale dataset for vehicle-infrastructure cooperative 3d object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 July 2022; pp. 21361–21370. [Google Scholar]
  4. Xu, R.; Xiang, H.; Tu, Z.; Xia, X.; Yang, M.H.; Ma, J. V2x-vit: Vehicle-to-everything cooperative perception with vision transformer. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2022; pp. 107–124. [Google Scholar]
  5. Adaimi, G.; Kreiss, S.; Alahi, A. Deep Visual Re-identification with Confidence. Transp. Res. Part C Emerg. Technol. 2021, 126, 103067. [Google Scholar] [CrossRef]
  6. Ghahremannezhad, H.; Shi, H.; Liu, C. Real-Time Accident Detection in Traffic Surveillance Using Deep Learning. In Proceedings of the 2022 IEEE International Conference on Imaging Systems and Techniques (IST), Kaohsiung, Taiwan, 21–23 June 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1–6. [Google Scholar]
  7. Hu, Z.; Lam, W.H.; Wong, S.C.; Chow, A.H.; Ma, W. Turning traffic surveillance cameras into intelligent sensors for traffic density estimation. Complex Intell. Syst. 2023, 1–25. [Google Scholar] [CrossRef]
  8. Naphade, M.; Wang, S.; Anastasiu, D.C.; Tang, Z.; Chang, M.C.; Yao, Y.; Zheng, L.; Rahman, M.S.; Arya, M.S.; Sharma, A.; et al. The 7th AI City Challenge. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; IEEE: Piscataway, NJ, USA, 2023; pp. 5537–5547. [Google Scholar]
  9. Fernandez-Sanjurjo, M.; Bosquet, B.; Mucientes, M.; Brea, V.M. Real-Time Visual Detection and Tracking System for Traffic Monitoring; Elsevier: Amsterdam, The Netherlands, 2019; Volume 85, pp. 410–420. [Google Scholar]
  10. Zhang, C.; Ren, K. LRATD: A Lightweight Real-Time Abnormal Trajectory Detection Approach for Road Traffic Surveillance; Springer: Berlin/Heidelberg, Germany, 2022; Volume 34, pp. 22417–22434. [Google Scholar]
  11. Ghahremannezhad, H.; Shi, H.; Liu, C. Object Detection in Traffic Videos: A Survey. IEEE Trans. Intell. Transp. Syst. 2023, 24, 6780–6799. [Google Scholar] [CrossRef]
  12. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2112.11790. [Google Scholar]
  13. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  14. Ye, X.; Shu, M.; Li, H.; Shi, Y.; Li, Y.; Wang, G.; Tan, X.; Ding, E. Rope3d: The roadside perception dataset for autonomous driving and monocular 3d object detection task. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, New Orleans, LA, USA, 19–24 July 2022; pp. 21341–21350. [Google Scholar]
  15. Yang, L.; Yu, K.; Tang, T.; Li, J.; Yuan, K.; Wang, L.; Zhang, X.; Chen, P. BEVHeight: A Robust Framework for Vision-based Roadside 3D Object Detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 21611–21620. [Google Scholar]
  16. Hosseiny, A.; Jahanirad, H. Hardware acceleration of YOLOv7-tiny using high-level synthesis tools. Real-Time Image Proc. 2023, 20, 75. [Google Scholar] [CrossRef]
  17. Chen, H.; Huang, Y.; Tian, W.; Gao, Z.; Xiong, L. Monorun: Monocular 3d object detection by reconstruction and uncertainty propagation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Nashville, TN, USA, 20–25 June 2021; pp. 10379–10388. [Google Scholar]
  18. Ding, M.; Huo, Y.; Yi, H.; Wang, Z.; Shi, J.; Lu, Z.; Luo, P. Learning Depth-Guided Convolutions for Monocular 3D Object Detection. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020. [Google Scholar]
  19. Reading, C.; Harakeh, A.; Chae, J.; Waslander, S.L. Categorical Depth Distribution Network for Monocular 3D Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  20. Wang, L.; Du, L.; Ye, X.; Fu, Y.; Guo, G.; Xue, X.; Feng, J.; Zhang, L. Depth-conditioned Dynamic Message Propagation for Monocular 3D Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  21. Wang, Y.; Chao, W.L.; Garg, D.; Hariharan, B.; Campbell, M.; Weinberger, K.Q. Pseudo-lidar from visual depth estimation: Bridging the gap in 3D object detection for autonomous driving. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 8445–8453. [Google Scholar]
  22. Carrillo, J.; Waslander, S. Urbannet: Leveraging urban maps for long range 3D object detection. In Proceedings of the 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), Indianapolis, IN, USA, 19–22 September 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 3799–3806. [Google Scholar]
  23. Weng, X.; Kitani, K. Monocular 3D Object Detection with Pseudo-LiDAR Point Cloud. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 857–866. [Google Scholar]
  24. Mousavian, A.; Anguelov, D.; Flynn, J.; Kosecka, J. 3D bounding box estimation using deep learning and geometry. In Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 7074–7082. [Google Scholar]
  25. Ma, X.; Liu, S.; Xia, Z.; Zhang, H.; Zeng, X.; Ouyang, W. Rethinking pseudo-lidar representation. In Computer Vision–ECCV 2020, Proceedings of the 16th European Conference, Glasgow, UK, 23–28 August 2020; Springer: Berlin/Heidelberg, Germany, 2020; pp. 311–327. [Google Scholar]
  26. Ye, X.; Du, L.; Shi, Y.; Li, Y.; Tan, X.; Feng, J.; Ding, E.; Wen, S. Monocular 3D Object Detection via Feature Domain Adaptation. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; pp. 17–34. [Google Scholar]
  27. Brazil, G.; Liu, X. M3D-RPN: Monocular 3D Region Proposal Network for Object Detection. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November.
  28. Ma, X.; Zhang, Y.; Xu, D.; Zhou, D.; Yi, S.; Li, H.; Ouyang, W. Delving into localization errors for monocular 3D object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 4721–4730. [Google Scholar]
  29. Liu, X.; Xue, N.; Wu, T. Learning Auxiliary Monocular Contexts Helps Monocular 3D Object Detection. AAAI Proc. Aaai Conf. Artif. Intell. 2022, 36, 1810–1818. [Google Scholar] [CrossRef]
  30. Zhang, Y.; Lu, J.; Zhou, J. Objects are Different: Flexible Monocular 3D Object Detection. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Virtual, 19–25 June 2021. [Google Scholar]
  31. Zhou, X.; Koltun, V.; Krähenbühl, P. Tracking objects as points. In European Conference on Computer Vision; Springer: Berlin/Heidelberg, Germany, 2020; pp. 474–490. [Google Scholar]
  32. Simonelli, A.; Bulo, S.R.; Porzi, L.; López-Antequera, M.; Kontschieder, P. Disentangling monocular 3D object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 1991–1999. [Google Scholar]
  33. Liu, Z.; Wu, Z.; Tóth, R. Smoke: Single-stage monocular 3D object detection via keypoint estimation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops, Seattle, WA, USA, 13–19 June 2020; IEEE: Piscataway, NJ, USA, 2020; pp. 996–997. [Google Scholar]
  34. Zhou, X.; Karpur, A.; Gan, C.; Luo, L.; Huang, Q. Unsupervised domain adaptation for 3D keypoint estimation via view consistency. In Proceedings of the European Conference on Computer Vision (ECCV); Springer: Berlin/Heidelberg, Germany, 2018; pp. 137–153. [Google Scholar]
  35. Li, Z.; Chen, Z.; Li, A.; Fang, L.; Jiang, Q.; Liu, X.; Jiang, J. Unsupervised domain adaptation for monocular 3D object detection via self-training. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2022; pp. 245–262. [Google Scholar]
  36. Adam, M.G.; Piccolrovazzi, M.; Eger, S.; Steinbach, E. Bounding box disparity: 3D metrics for object detection with full degree of freedom. In Proceedings of the 2022 IEEE International Conference on Image Processing (ICIP), Bordeaux, France, 16–19 October 2022; IEEE: Piscataway, NJ, USA, 2022; pp. 1491–1495. [Google Scholar]
  37. Li, P.; Zhao, H.; Liu, P.; Cao, F. RTM3D: Real-time Monocular 3D Detection from Object Keypoints for Autonomous Driving. In Proceedings of the Computer Vision—ECCV 2020, Glasgow, UK, 23–28 August 2020; pp. 644–660. [Google Scholar]
  38. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  39. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Computer Vision—ECCV 2018, Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  40. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021. [Google Scholar]
  41. Geiger, A.; Lenz, P.; Stiller, C.; Urtasun, R. Vision Meets Robotics: The Kitti Dataset; Sage Publications: London, UK, 2013; Volume 32, pp. 1231–1237. [Google Scholar]
Figure 1. The network architecture diagram depicts, from left to right, the specific structures of the main feature extraction network, the Feature Pyramid Network (FPN), and the five multi-branch decoupled detection heads.
Figure 1. The network architecture diagram depicts, from left to right, the specific structures of the main feature extraction network, the Feature Pyramid Network (FPN), and the five multi-branch decoupled detection heads.
Applsci 13 11402 g001
Figure 2. Schematic diagram of the seven degrees of freedom of the target.
Figure 2. Schematic diagram of the seven degrees of freedom of the target.
Applsci 13 11402 g002
Figure 3. The overall structure diagram of the Multi-Scale Spatial Attention mechanism.
Figure 3. The overall structure diagram of the Multi-Scale Spatial Attention mechanism.
Applsci 13 11402 g003
Figure 5. The diagram illustrating the viewing angle.
Figure 5. The diagram illustrating the viewing angle.
Applsci 13 11402 g005
Figure 6. Visualization of Partial Images and Labels from the Rope3D Dataset.
Figure 6. Visualization of Partial Images and Labels from the Rope3D Dataset.
Applsci 13 11402 g006
Figure 7. Statistical distribution of various target categories in the Rope3D training and validation sets.
Figure 7. Statistical distribution of various target categories in the Rope3D training and validation sets.
Applsci 13 11402 g007
Figure 8. The result of K-means for anchors.
Figure 8. The result of K-means for anchors.
Applsci 13 11402 g008
Figure 9. Presentation of data augmentation results. We implemented random scaling, cropping, changes in color space, and brightness adjustments in the data augmentation algorithm to enhance the robustness of the model.
Figure 9. Presentation of data augmentation results. We implemented random scaling, cropping, changes in color space, and brightness adjustments in the data augmentation algorithm to enhance the robustness of the model.
Applsci 13 11402 g009
Table 1. The comparison of the experimental and ablation study.
Table 1. The comparison of the experimental and ablation study.
MethodBackboneAP3D/IoU = 0.5Params (M)GFlops
CarVanBusTruckCyclistMotorcyclistTricyclistPedestrian
KM3DResNet348.977.773.974.9411.8111.3510.3912.6126.881422.276
Kinematic3DDenseNet12111.4210.654.024.8311.3515.0811.2414.4325.913296.441
MonoDLEDLA3412.8411.214.195.2513.2316.5811.7214.7520.310264.898
Ours BaseCSPDarknet11.7510.981.864.5014.2417.4611.0313.1937.626222.563
Ours+5HeadCSPDarknet12.2510.731.783.6815.4518.7311.6614.8342.23587.625
Ours + MSSACSPDarknet13.0211.332.444.5315.6919.1311.7714.8142.95788.453
Table 2. The comparison results of detection performance in different distance ranges.
Table 2. The comparison results of detection performance in different distance ranges.
MethodRange (m)Score
CarBig VehicleCyclistPedestrian
MonoDLEall92.886.388.592.2
0–3089.884.087.192.0
30–6091.785.388.792.7
60–9092.790.988.991.9
90–12092.8989.386.291.3
Oursall94.489.789.191.9
0–3092.886.988.591.8
30–6094.792.089.392.5
60–9094.591.988.692.4
90–12093.790.088.890.9
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Ye, Z.; Zhang, H.; Gu, J.; Li, X. YOLOv7-3D: A Monocular 3D Traffic Object Detection Method from a Roadside Perspective. Appl. Sci. 2023, 13, 11402. https://doi.org/10.3390/app132011402

AMA Style

Ye Z, Zhang H, Gu J, Li X. YOLOv7-3D: A Monocular 3D Traffic Object Detection Method from a Roadside Perspective. Applied Sciences. 2023; 13(20):11402. https://doi.org/10.3390/app132011402

Chicago/Turabian Style

Ye, Zixun, Hongying Zhang, Jingliang Gu, and Xue Li. 2023. "YOLOv7-3D: A Monocular 3D Traffic Object Detection Method from a Roadside Perspective" Applied Sciences 13, no. 20: 11402. https://doi.org/10.3390/app132011402

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop