Next Article in Journal
ϕ-Informational Measures: Some Results and Interrelations
Next Article in Special Issue
Pairwise Elastic Net Representation-Based Classification for Hyperspectral Image Classification
Previous Article in Journal
Tsallis q-Stat and the Evidence of Long-Range Interactions in Soil Temperature Dynamics
Previous Article in Special Issue
An Improved Stereo Matching Algorithm for Vehicle Speed Measurement System Based on Spatial and Temporal Image Fusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Robust Vehicle Speed Measurement Based on Feature Information Fusion for Vehicle Multi-Characteristic Detection

1
School of Electronic and Information, Zhongyuan University of Technology, Zhengzhou 450007, China
2
Dongjing Avenue Campus, Kaifeng University, Kaifeng 475004, China
3
School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China
4
Department of Electrical and Computer Engineering, Texas A&M University, College Station, TX 77843, USA
*
Authors to whom correspondence should be addressed.
Entropy 2021, 23(7), 910; https://doi.org/10.3390/e23070910
Submission received: 17 June 2021 / Revised: 13 July 2021 / Accepted: 14 July 2021 / Published: 17 July 2021
(This article belongs to the Special Issue Advances in Image Fusion)

Abstract

:
A robust vehicle speed measurement system based on feature information fusion for vehicle multi-characteristic detection is proposed in this paper. A vehicle multi-characteristic dataset is constructed. With this dataset, seven CNN-based modern object detection algorithms are trained for vehicle multi-characteristic detection. The FPN-based YOLOv4 is selected as the best vehicle multi-characteristic detection algorithm, which applies feature information fusion of different scales with both rich high-level semantic information and detailed low-level location information. The YOLOv4 algorithm is improved by combing with the attention mechanism, in which the residual module in YOLOv4 is replaced by the ECA channel attention module with cross channel interaction. An improved ECA-YOLOv4 object detection algorithm based on both feature information fusion and cross channel interaction is proposed, which improves the performance of YOLOv4 for vehicle multi-characteristic detection and reduces the model parameter size and FLOPs as well. A multi-characteristic fused speed measurement system based on license plate, logo, and light is designed accordingly. The system performance is verified by experiments. The experimental results show that the speed measurement error rate of the proposed system meets the requirement of the China national standard GB/T 21555-2007 in which the speed measurement error rate should be less than 6%. The proposed system can efficiently enhance the vehicle speed measurement accuracy and effectively improve the vehicle speed measurement robustness.

1. Introduction

Vehicle speed measurement is one of the most important tasks of intelligent traffic monitoring system [1,2,3,4]. It helps to monitor speeding behavior and improve road safety. Common vehicle speed measurement methods can be divided into intrusive and nonintrusive. The intrusive method uses inductive loop detectors (ILDs) embedded in the road to measure the average vehicle speed, which is difficult to install and maintain due to the damage to road [5]. The nonintrusive speed measurement methods include Radar [6], Lidar [7], and video-based speed measurement systems [8]. The Radar method uses Doppler effect produced by the relative motion between vehicle and radar equipment (fixed or mobile) to measure the vehicle speed [6]. The Lidar method uses a laser to measure the distance from a fixed lidar equipment to vehicle twice within a specific time interval, and calculates the average vehicle speed [7]. Both methods transmit signals, which are easy to be detected and hard to make secret measurements. Video-based speed measurement methods can be further divided into two categories: monocular 2D video-based method and binocular 3D video-based method [9]. The monocular 2D video-based methods [8,10,11] use the perspective projection relationship of 2D imaging to estimate the distance traveled by the vehicle within a fixed frame interval, thereby calculate the vehicle speed. Due to the perspective projection relationship, they can only accurately measure the speed of vehicle in straight-line motion, but not in curved motion. The binocular 3D video-based methods [12,13,14] use the stereo imaging principle of the binocular camera, which can directly calculate the distance traveled by the vehicle within a fixed frame interval, thereby calculate the vehicle speed. However, the existing methods still have disadvantages such as low object detection efficiency and lack of intelligence, which need to be further improved.
To overcome the shortcomings of the existing methods, a binocular stereovision-based vehicle speed measurement system is proposed in [14]. The system consists of three parts: vehicle characteristic detection, vehicle tracking and stereo matching, and vehicle speed and trajectory measurement. The system uses a SSD (Single-Shot Multibox Detector) network [15] optimized for license plate detection to efficiently detect the license plate in binocular stereo video, performs fast stereo matching [16,17] in the detected left-view and right-view license plate areas, calculates the 3D coordinates of the matching point pairs, eliminates the abnormal points and selects the point closest to the license plate center as the speed measurement point of the current frame, and calculates the vehicle speed by dividing the distance between the speed measurement points in consecutive frames by the frame interval. This system is of low cost and high accuracy. It can realize nonintrusive and secret vehicle speed measurement, and can simultaneously measure the speed of multiple vehicles in different motions and on multiple lanes. The system is based on vehicle license plate detection, and the detection network is optimized for license plate detection. Moreover, the system designed is also based on the premise that the license plate characteristic of the vehicle can be accurately detected. However, vehicles license plate violation often exists, including no license plate, deliberately shielding or blurring license plate, faking license plate, etc. [18]. According to traffic statistical reports [19,20,21,22], license plate violation accounts for a relatively high proportion of traffic violation, which poses a challenge to the binocular stereovision-based vehicle speed measurement system with license plate in [14]. Once license plate violation occurs, the system will be invalid.
A robust binocular stereovision-based vehicle speed measurement system with vehicle multi-characteristic detection is proposed in this paper. The system model is shown in Figure 1, wherein, A, B, C and D respectively represent the multiple characteristics of vehicle, i.e., mirror, light, logo, and license plate. Δ S 1 , Δ S 2 , Δ S 3 , and Δ S 4 , respectively, represent the displacements of multiple characteristics in the speed measurement frame interval Δ t . As the multiple characteristics of the vehicle are of multiple scales at multiple distances, You Only Look Once (YOLO) v4 network is selected, and a YOLOv4-based vehicle multi-characteristic detection network is proposed, which can efficiently detect multiple characteristics of vehicle. Combined with attention mechanism, Efficient Channel Attention (ECA) module is chosen, and an ECA-YOLOv4-based vehicle multi-characteristic detection network is proposed, which can further improve the detection efficiency and accuracy of the network for multiple characteristics of the vehicle. Accordingly, a multi-characteristic combined speed measurement system based on binocular stereovision with license plate, logo and light detection is proposed. In normal case, this system can efficiently enhance the speed measurement accuracy. In the case of license plate violation, this system can effectively improve the speed measurement robustness, which can solve the problem that the system in [14] cannot measure the vehicle speed in the case of license plate violation.
The rest of the paper is organized as follows. In Section 2, related work is discussed. In Section 3, a robust binocular stereovision-based vehicle speed measurement system with vehicle multi-characteristic detection is proposed. The main work focuses on the optimization of object detection algorithm for multiple characteristics of vehicle and the design of the multi-characteristic combined vehicle speed measurement system. In Section 4, the experimental setup and results are reported. In Section 5, the conclusion is drawn.

2. Related Works

2.1. Object Detection Algorithm

Accurate vehicle characteristic detection, namely, accurate object detection is an important premise of the vehicle speed measurement system proposed in this paper, and is also the main research content of this paper. Object detection includes two broad categories: traditional object detection and deep learning-based object detection [23]. Traditional object detection extracts informative feature sets by handcrafted features, while deep learning based object detection extracts informative feature sets by various end-to-end deep learning networks [24]. In recent years, thanks to the rapid development of artificial intelligence, object detection based on deep learning has become the mainstream, with better adaptability and more intelligence. Deep learning networks can be divided into Convolutional Neural Networks (CNNs) for Euclidean data and the currently state-of-the-art Graph Neural Networks (GNNs) for non-Euclidean data [25]. CNNs extract multi-scale local spatial information and fuse them to construct feature representation [26]. GNNs capture the dependency in graph by information transfer between nodes in graph, among which Graph Convolutional Networks (GCNs) have become a hot topic. An improved version of GCNs, known as miniGCNs, has been proposed, which allows to train in a minibatch fashion and is capable of inferring out-of-sample data without retraining [27]. However, as the object detection in this paper is performed on the captured stereo images which belong to Euclidean data, CNNs based object detection is enough for this application scene.
CNN-based object detection algorithm aims to detect the object of interest from the image, determine the category of each object, and locate the bounding box of each object by self-learning the high-level features of the images. Feature extraction backbone network includes AlexNet [28], VGG [29], Inception [30], ResNet [31], DenseNe [32], DarkNet [33], CSPDarkNet [34], and so on. According to the different ways of using the extracted feature maps, the object detection algorithms can be divided into three categories: object detection algorithm based on single feature map, object detection algorithm based on pyramid feature hierarchy, and object detection algorithm based on Feature Pyramid Network (FPN) [35].
R-CNN [36], Fast R-CNN [37], Faster R-CNN [38], and YOLOv1 [33] are all based on the single feature map. The advantage is the fast detection speed and little memory requirement. The disadvantage is that only the features of the last high-level are used; the resolution is low and is not conductive to small object detection.
SSD [15], RFB [39], and YOLOv2 [40] are all based on the pyramid feature hierarchy. The advantage is the simultaneous use of multi-layer feature maps. The disadvantage is that the calculated low-level high-resolution feature maps are not reused, the spatial information in the low-level feature maps is not fully utilized, and the small object detection task is still not well completed.
The FPN-based object detection algorithm takes the feature pyramid hierarchy as the foundation structure and makes joint prediction by the fused feature map of each layer. Commonly used feature fusion schemes include additive fusion, elementwise multiplicative fusion, and concatenation fusion [27]. Several state-of-the-art fusion strategies include concatenation-based fusion: early fusion, middle fusion, and late fusion, and compactness-based fusion: encoder–decoder fusion and cross-fusion [41]. The purpose of feature fusion is to combine the resulting features using different fusion strategies before the final classification. By feature fusion from the top layer to the bottom layer, the fused feature map contains both rich high-level semantic information and detailed low-level location information [42,43]. Then, the predicted joint feature pyramid hierarchy is used for object detection. For distinguishing multiple objects of multiple scales in the image, multiple feature maps in the predicted joint feature pyramid hierarchy can be used. YOLOv3 [44], M2Det [45], Retinanet [46], YOLOv4 [34], and EfficientDet [47] are all based on FPN. The advantage is that it can transfer the semantic information from the high-level to the low-level by information fusion and effectively supplement the semantic information to the low-level, therefore obtain features with high-resolution and rich semantic information, which can successfully complete the small object detection task.

2.2. Attention Mechanism

The essence of the attention mechanism is to locate the information of interest and suppress the useless information. In object detection network, the convolution operation module in the feature extraction backbone network can be replaced by the attention module to assign more weight to the part containing the information of interest and reduce the weight of the part containing the interference information. Thereby, the object detection accuracy can be improved.
The attention module can be divided into spatial attention module, channel attention module, and mixed attention module according to the attention principle [48,49,50]. The task of spatial attention module is to find the 2D spatial position containing the information of interest in a single feature map. It performs average pooling and maximum pooling on the single feature map of the input features, connects the average pooling, and maximum pooling results by convolution, and generates a corresponding 2D spatial attention features map.
The task of channel attention module is to find the channel position containing the information of interest among different channels. Squeeze-and-Excitation (SE) [51] and ECA [52] are typical representatives of channel attention module. SE performs global average pooling on the feature map of each channel, obtains the correlation between the two fully connected layers by sharing the multi-layer perception (MLP), and generates a corresponding channel attention map. For the two convolutions between the two fully connected layers in SE by which the channel dimension is first reduced and then increased, ECA replace them with a one-dimensional convolution that utilizes the local cross channel interaction information which fuses the information of each channel and its k neighbors. Dimension reduction can be avoided by cross-channel interaction which is a kind of information fusion as well. ECA can significantly reduce the model complexity while improving the network performance. It is an extremely lightweight channel attention module.
The task of mixed attention module is to simultaneously find the channel position containing the information of interest among different channels and the 2D spatial position containing the information of interest in each channel. Convolutional Block Attention Module (CBAM) [53] is a typical representative of mixed attention module. It divides the attention process into two independent cascaded modules: channel attention module and spatial attention module. First, channel attention map is generated by the channel attention module for input features, and the refined features of channel attention optimization are obtained by multiplying the input features with the generated channel attention map. Then, spatial attention map is generated by the spatial attention module for the refined feature, and the features of channel attention and spatial attention joint optimization are obtained by multiplying the refined features of channel attention optimization with the generated spatial attention map. Mixed attention model can refine features from both spatial and channel dimensions at the same time, which can better obtain the feature information of the region of interest and provide more effective information for prediction, thus improving the performance of object detection algorithm.

2.3. Cross-Entropy Loss

Cross-entropy is an important concept in Shannon’s information theory. Cross-entropy loss is an important index to measure the performance of an object detection classification model in deep learning training. It measures the similarity between the prediction and the actual target, which can effectively avoid the problem of learning rate decline in gradient descent. Cross-entropy loss increases as the predicted probability diverges from the actual label. For multi-class classification, the cross-entropy loss function is as shown in Equation (1)  [54].
L = 1 N i = 1 N j = 1 C y i j log p i j
wherein, C is the number of classes; N is the number of samples; y i j is the indicator variable, which is 1 if the predicted class of sample i is the same as the actual class j, and 0 otherwise; p i j is the predicted probability that the sample i belongs to the actual class j. The smaller the cross-entropy loss is, the better the model prediction will be.

3. The Proposed Method

The system proposed is built by two industrial cameras and a laptop. The Hikvision MV-CA050-11UC industrial camera has a resolution of 2448 × 2048, with a Wallis WL1608-5MP fixed-focus lens of 8 mm. The laptop is equipped with an Intel Core i7-10750H CPU, 16 GB RAM, and a Nvidia RTX2060 6G graphics card. The stereo camera is calibrated by Zhengyou Zhang’s calibration method, and the cell size of the calibration board is 30 mm × 30 mm. The system configuration is shown in Figure 2.
The whole procedure of the proposed system is shown in Figure 3, which includes three parts: vehicle multi-characteristic detection and decision, stereo matching, and speed measurement. In the vehicle multi-characteristic detection and decision part, the proposed ECA-YOLOv4 object detection network is trained by the vehicle multi-characteristic dataset to get a model. The model is used to detect the left-view and right-view images separately to obtain multiple characteristics of vehicle, i.e., vehicle, license plate, logo, and light. The bounding box of the detected vehicle is used to constrain the detected multiple characteristics into the same vehicle area. Assuming the light characteristic always exists and can be detected, the speed measurement scheme is decided according to whether license plate and logo exist in the detection results, which can be divided into four cases: (1) three vehicle characteristics detected, i.e., license plate, logo and light detected; (2) two vehicle characteristics detected, i.e., license plate and light detected; (3) two vehicle characteristics detected, i.e., logo and light detected; (4) one vehicle characteristic detected, i.e., only light detected. In the stereo matching part, the stereo matching algorithm in [14] is reused, while the speed measurement point selection is slightly adjusted for the multiple characteristics with irregular shapes. In the speed measurement part, the binocular stereovision calibration algorithm in [14] is reused to perform speed measurement of single characteristic respectively, then the vehicle speed is calculated according to the speed measurement scheme decided in the detection and decision part.

3.1. Vehicle Multi-Characteristic Detection Based on YOLOv4

First, the vehicle multi-characteristic dataset is constructed. We randomly select 6103 images with a resolution of 1600 × 1200 from the Open ITS dataset [55], 1921 images with a resolution of 1920 × 1080 from the BIT Vehicle dataset [56], and 280 images with a resolution of 720 × 1160 from the CCPD dataset [57]. In addition, we capture 3351 images with a resolution of 6000 × 4000 by Nikon d3200 SLR camera and 480 images with a resolution of 2448 × 2048 by the Hikvision MV-CA050-11UC camera. The dataset has a total of 12,135 vehicle images with multiple characteristics. Figure 4 shows some image examples of our vehicle multi-characteristic dataset.
According to regulation on the relative size of small objects by SPIE, an object with a pixel ratio less than 0.12% can be regarded as small object, otherwise it can be regarded as regular object [58]. Statistical analysis is performed on the pixel ratio of four common vehicle characteristics (license plate, logo, light, and mirror) in the image captured at different distance within the speed measurement range. As shown in Figure 5, the distance threshold of small object for the license plate is about 9 m; the distance threshold of small object for the logo is about 4 m; the distance threshold of small object for the light is about 9 m; and the distance threshold of small object for the mirror is about 7 m. Therefore, the detection of multiple characteristics of vehicle is a multi-scale object detection problem.
As mentioned above, object detection algorithms based on single feature map and pyramidal feature hierarchy are not suitable for this multi-scale varying object detection problem, while object detection algorithm based on FPN is suitable for this multi-scale varying object detection problem. Therefore, the object detection algorithm based on FPN is selected to detect the multiple characteristics of vehicle. Among them, YOLOv4 based on FPN is chosen due to its faster speed and higher accuracy [34]. Figure 6 is the schematic block diagram of the YOLOv4-based vehicle multi-characteristic object detection algorithm. Features are extracted from the input vehicle images by the feature extraction backbone network CSPDarknet53. The extracted features of different scales are fused by SPP and PANet. Finally, prediction is performed on the three different scale feature maps output by PANet, so as to obtain the bounding box, category, and confidence of the multiple characteristics of vehicle.
There are 23 Cross-Stage Partial (CSP) modules in CSPDarknet53. The CSP module enhances the learning ability of CNN and reduces the memory usage, whose structure is shown in Figure 7. The CBM module is composed of convolution, batch normalization and mish activation function. The structure of the residual unit is shown in Figure 8. The input feature map is added to the original feature map after two convolution operations.
After three convolution operations on the last feature map of CSPDarknet53, the Spatial Pyramid Pooling (SPP) module executes maximum pooling operations of four different scales, i.e., 1 × 1 , 5 × 5 , 9 × 9 , and 13 × 13 , and then the maximum pooling results of multiple scales are cascaded. After that, the feature maps are fused from the bottom to the top by PANet, and then the fused feature maps are fused again from the top to the bottom. Finally, the bounding box, category, and confidence of the object are predicted by the fused feature maps of three different scales.
The performance of the YOLOv4-based vehicle multi-characteristic detection algorithm is experimentally verified. Six representative object detection algorithms, i.e., Faster R-CNN, SSD, RFB, Retinanet, M2Det, and YOLOv3 are also selected for performance comparison. During the experiment, the model training parameters are the same, and the same model measurement indexes AP and mAP are selected. The experimental results are shown in Table 1. YOLOv4 has an AP of 96.47% for car, 92.13% for license plate, 87.72% for logo, 94.17% for light, and 91.2% for mirror, whose detection accuracy for vehicle single-characteristic detection is better than the other six algorithms. The average accuracy mAP of YOLOv4 is 92.34%, which is also better than that of the other six algorithms. Experimental results show that the proposed algorithm based on YOLOv4 is suitable for vehicle multi-characteristic detection and performs well.

3.2. Vehicle Multi-Characteristic Detection Based on ECA-YOLOv4

The vehicle characteristics to be detected are local information, and the attention mechanism that conforms to the human perception mechanism helps to focus on the local information. As mentioned above, the multiple vehicle characteristics to be detected are multi-scale varying, and a large proportion of them are small objects. Thus, the channel attention mechanism is suited to solve such problem. However, the position of the vehicle characteristics in the imaging plane constantly changes with the moving of the vehicle, and the spatial attention mechanism is not applicable. Therefore, the channel attention module is selected to optimize the YOLOv4-based vehicle multi-characteristic detection algorithm. In the two commonly used channel attention modules, the ECA module is an upgrade version of the SE module. Therefore, the ECA module is chosen to optimize the network structure of the YOLOv4 and improve the detection performance.
In the ECA module [52], cross-channel interaction and channel weight sharing are adopted. A weight matrix W k is defined for the cross-channel interaction, as shown in Equation (2).
W k = ω 1 1 ω 1 k 0 0 0 0 ω 2 2 ω 2 k + 1 0 0 0 0 0 ω C C k + 1 ω C C
wherein W k has k × C parameters. For the ith channel y i , as shown in Equation (3), the interactive relationship of k adjacent channels y i j (including y i ) to y i should be considered to calculate its weight ω i , j = 1 , , k .
ω i = σ j = 1 k ω i j y i j , y i j Ω i k
wherein Ω i k represents the set is composed of the k adjacent channels of y i
After cross-channel interaction, channel weight sharing is carried out, so that all the channels share the same k parameters, as shown in Equation (4).
ω i = σ j = 1 k ω j y i j , y i j Ω i k
In this way, the parameter number of the weight matrix W k are reduced from k × C to k. Weight parameter learning can be achieved by 1D fast convolution with a convolution kernel size of k.
ω = σ C 1 D k ( y )
wherein C 1 D represents 1D convolution operation.
As shown in Figure 9, the two CBM modules in each residual unit in each CSP module of YOLOv4 are replaced by one ECA module, and the YOLOv4 is improved to ECA-YOLOv4. Moreover, SE channel attention module and CBAM mixed attention module are also, respectively, used to improve YOLOv4 to SE-YOLOv4 and CBAM-YOLOv4 in the same way for performance comparison. The model parameter sizes and floating point operations (FLOPs) of the object detection networks with different attention modules are shown in Table 2. For the same input size of 416 × 416 , the model parameter size of YOLOv4 is 244.29 M, the model parameter size of CBAM-YOLOv4 is 180.51 M with a compression rate of 26.10%, the model parameter size of SE-YOLOv4 is 179.66 M with a compression rate of 26.45%, and the model parameter size of ECA-YOLOv4 is 178.82 M with a compression rate of 26.8%. ECA-YOLOv4 has the smallest model parameter size. For the same input size of 416 × 416 , the FLOPs of YOLOv4 is 29.95 G, the FLOPs of CBAM-YOLOv4 is 19.28 G, the FLOPs of SE-YOLOv4 is 19.28 G, and the FLOPs of ECA-YOLOv4 is 19.28 G. ECA-YOLOv4, SE-YOLOv4, and CBAM-YOLOv4 have the same FLOPs, which is smaller than that of YOLOv4.
During the experiment, the model training parameters are the same, and the same model measurement indexes AP, mAP, and FPS are selected. The experimental results are shown in Table 3. From the AP of vehicle, license plate, logo, and mirror, the detection performances of CBAM-YOLOv4, SE-YOLOv4, and ECA-YOLOv4 are all superior to that of YOLOv4. From the AP of light, the detection performances of SE-YOLOv4 and ECA-YOLOv4 are superior to that of YOLOv4. From the mAP, the detection performances of CBAM-YOLOv4, SE-YOLOv4, and ECA-YOLOv4 are all superior to that of YOLOv4. From the FPS, only the detection speed of ECA-YOLOv4 is faster than that of YOLOv4. Note that for all four networks, the AP of logo is slightly lower than that of other vehicle characteristics. This is because the shapes of logos are more diverse than those of license plates, lights and mirrors, and a few special logos may not be accurately detected, which will reduce the precision corresponding to some recall and thus reduce the AP of logo.
Considering the four indexes of model parameter size, AP, mAP, and FPS. ECA-YOLOv4 has the smallest model parameter size and the fastest FPS, with AP for vehicle single-characteristic detection and mAP for vehicle multi-characteristic superior to that of YOLOv4 and equivalent to that of CBAM-YOLOv4, and SE-YOLOv4. In summary, the performance of ECA-YOLOv4 is better than that of the other three algorithms.

3.3. Optimal Design of Binocular Stereovision-Based Vehicle Speed Measurement System with Vehicle Multi-Characteristic Detection

After the multiple characteristics of the vehicle are detected, as shown in Figure 10, stereo matching is performed on each detected vehicle characteristic by the stereo matching algorithm in [14]. Thus, the stereo matching point pairs in the left-view and right-view images of the vehicle characteristic are obtained. Then, the matching point pair with the smallest square sum of the Euclidean distance to the bounding box center is selected as the speed measurement point. Zhengyou Zhang’s calibration method is used to calculate the 3D coordinates of the selected speed measurement point, the displacement of adjacent speed measurement points are calculated, and the vehicle speed can be obtained by dividing the displacement by the time interval.
The video capture frame rate is 30FPS, the vehicle speed is set to 43 km/h, the vehicle drives to the camera in a straight line from far and near at a constant speed, and the speed measurement is performed ten times per second. The speed measurement data by a professional satellite speed meter P-Gear P-510 are taken as the ground truth for comparison. The speed measurement results of single characteristic are shown in Table 4, and the error rate curve is shown in Figure 11. It can be seen from Table 4 that the speeds measured by the license plate, logo, and light are all consistent with the satellite ground truth. The error rate range is [−2.49%, 4.55%], [−3.71%, 3.20%], and [−4.16%, 5.66%], respectively, all of which do not exceed the 6% error rate limit specified by the China national standard GB/T21255-2007 [59]. However, the speeds measured by the mirror differs greatly from the satellite ground truth, with an error rate range of [−82.27%, 184.69%], which is far more than the 6% error rate limit.
By analyzing the speed measurement procedure of the mirror, the large error is mainly due to the smooth surface which is lack of detailed information such as texture. Thus, large error occurs in the stereo matching, which leads to the incorrect speed measurement. Therefore, in the following design of the multi-characteristic speed measurement system, only three vehicle characteristics (license plate, logo, and light) are selected, and the mirror characteristic is abandoned.
Assuming that the vehicle light characteristic always exists, the vehicle speed measurement algorithm of the proposed binocular stereovision vehicle speed measurement system with vehicle multi-characteristic detection is designed as Algorithm 1.
Algorithm 1 Vehicle Speed Measurement Algorithm
Input:  Binocular Stereovision Video Sequence
Output:  Vehicle Speed
  1:  function Vehicle speed measurement
  2:        if License plate detected then
  3:            if Logo detected then
  4:                 v = v p l a t e + v l o g o + v l i g h t 3
  5:            else Logo undetected
  6:                 v = v p l a t e + v l i g h t 2
  7:            end if
  8:        else License plate undetected
  9:             if Logo detected then
10:                 v= v l o g o + v l i g h t 2
11:             else Logo undetected
12:                 v = v l i g h t
13:             end if
14:        end if
15:        return v
16:  end function

4. Experiments

In the actual speed measurement test, the video capture frame rate is 30 FPS, the vehicle drives to the camera in a straight line from far and near at a constant speed, and the speed measurement is performed ten times per second. The speed measurement data by the professional satellite speed meter P-Gear P-510 is taken as the ground truth for comparison. To verify the algorithm, four speed measurement scenarios are set up, as shown in Figure 12. Figure 12a,b shows the vehicle speed measurement experiments with license plate detected, while Figure 12c,d shows the vehicle speed measurement experiments with license plate undetected.

4.1. Vehicle Speed Measurement Experiments with License Plate Detected

First, experiments are carried out in the speed measurement scenario where license plate, logo, and light all exist, as shown in Figure 12a. Table 5 shows the speed measurement results of a single experiment in this scenario, in which the vehicle drives in a straight line at a constant speed of 46 km/h.
As shown in Table 5, the speed measurement error rate range is [−3.49%, 1.50%] for the license plate, [−4.07%, 5.01%] for the logo, [−5.97%, 2.28%] for the light, and [−3.53%, 1.52%] for the average of all the three characteristics of vehicle. The average speed measurement of all the three vehicle characteristics can effectively reduce the error rate fluctuation range of the individual speed measurement by individual vehicle characteristic. Multiple experiments are also carried out in this scenario, and the speed measurement results are shown in Table 6. The vehicle speeds set for test 2–8 are 44 km/h, 36 km/h, 32 km/h, 38 km/h, 40 km/h, 43 km/h, and 30 km/h, respectively. The error rate ranges of test 2–8 are [−3.08%, 2.92%], [−1.36%, 2.62%], [−2.23%, 3.61%], [−1.39%, 2.87%], [−2.04%, 3.97%], [−2.86%, 4.12%], and [−0.84%, 3.20%], respectively. The maximum absolute error rate is 4.12%, which meets the 6% error rate limit specified by the China national standard GB/T21255-2007 [59].
Next, as shown in Figure 12b, the logo is artificially covered to simulate the speed measurement scenario where license plate and light exist but logo does not exist, and the experiments are carried out therein. Table 7 shows the speed measurement results of a single experiment in this scenario, in which the vehicle drives in a straight line at a constant speed of 30 km/h.
As shown in Table 7, the speed measurement error rate range is [−4.76%, 5.22%] for the license plate, [−4.35%, 5.90%] for the light, and [−1.97%, 4.12%] for the average of the two characteristics of vehicle. The average speed measurement of the two vehicle characteristics can reduce the error rate fluctuation range of the individual speed measurement by individual vehicle characteristic. Multiple experiments are also carried out in this scenario, and the speed measurement results are shown in Table 8. The vehicle speeds set for test 2–8 are 45 km/h, 36 km/h, 43 km/h, 38 km/h, 46 km/h, 33 km/h, and 31 km/h, respectively. The error rate ranges of test 2–8 are [−4.16%, 4.50%], [−3.36%, 3.33%], [−2.76%, 4.58%], [−2.08%, 4.24%], [−4.02%, 1.84%], [−2.14%, 3.37%], and [−1.57%, 4.22%], respectively. The maximum absolute error rate is 4.58%, which also meets the 6% error rate limit specified by the China national standard GB/T21255-2007 [59].

4.2. Vehicle Speed Measurement Experiments with License Plate Undetected

Then, as shown in Figure 12c, the license plate is artificially covered to simulate the speed measurement scenario where logo and light exist but license plate does not exist, and the experiments are carried out therein. Table 9 shows the speed measurement results of a single experiment in this scenario, in which the vehicle drives in a straight line at a constant speed of 45 km/h.
As shown in Table 9, the speed measurement error rate range is [−3.20%, 4.38%] for the logo, [−4.72%, 4.84%] for the light, and [−2.79%, 2.30%] for the average of the two characteristics of vehicle. The average speed measurement of the two vehicle characteristics can reduce the error rate fluctuation range of the individual speed measurement by individual vehicle characteristic. Multiple experiments are also carried out in this scenario, and the speed measurement results are shown in Table 10. The vehicle speeds set for test 2–8 are 43 km/h, 36 km/h, 32 km/h, 38 km/h, 46 km/h, 33 km/h, and 31 km/h, respectively. The error rate ranges of test 2–8 are [−5.61%, 4.33%], [−2.99%, 3.93%], [−1.39%, 3.45%], [−3.85%, 3.84%], [−3.63%, 1.77%], [−2.50%, 4.99%], and [−1.64%, 4.62%], respectively. The maximum absolute error rate is 5.61%, which also meets the 6% error rate limit specified by the China national standard GB/T21255-2007 [59].
Finally, as shown in Figure 12d, the license plate and the logo are artificially covered to simulate the speed measurement scenario where only light exists but license plate and logo do not exist, and the experiments are carried out therein. Table 11 shows the speed measurement results of a single experiment in this scenario, in which the vehicle drives in a straight line at a constant speed of 36 km/h.
As shown in Table 11, the speed measurement error rate range is [−4.68%, 5.30%] for the light. Multiple experiments are also carried out in this scenario, and the speed measurement results are shown in Table 12. The vehicle speeds set for test 2–8 are 31 km/h, 32 km/h, 38 km/h, 46 km/h, 33 km/h, 30 km/h, and 45 km/h, respectively. The error rate ranges of test 2–8 are [−4.35%, 5.90%], [−3.87%, 4.57%], [−4.86%, 4.16%], [−5.97%, 2.28%], [−2.13%, 5.21%], [−3.90%, 4.67%], and [−4.72%, 4.84%], respectively. The maximum absolute error rate is 5.97%, which still meets the 6% error rate limit specified by the China national standard GB/T21255-2007 [59].

4.3. Contrast Experiments

The speed measurement results of the proposed system are compared with that of the original system in [14] in the same speed measurement scenario, including the speed measurement scenario with license plate not covered and the speed measurement scenario with license plate covered. The former is to verify the speed measurement accuracy of the proposed system, while the latter is to verify the speed measurement robustness of the proposed system.
Table 13 shows the speed measurement results of two contrast experiments in the scenario with license plate not covered, in which the vehicle drives in a straight line at a constant speed of 32 km/h. Figure 13 is the corresponding speed measurement curve. As shown in Table 13, in the scenario with license plate not covered, the speed measurement error rate range of the proposed system is [−2.23%, 3.61%] and the maximum absolute error rate is 3.61%, while the speed measurement error rate range of the system in [14] is [−3.91%, 4.53%] and the maximum absolute error rate is 4.53%. Both meet the 6% error rate limit specified by the China national standard GB/T21255-2007 [59]. However, both the error rate range and the maximum absolute error rate of the proposed system are smaller than those of the system in [14]. As can be seen from the speed measurement curve in Figure 13, the proposed system has a smaller error fluctuation range, which is closer to the satellite ground truth, and has higher speed measurement accuracy.
Meanwhile, the speed measurement performances are compared between the proposed system and the various existing speed measurement methods. Table 14 shows a comparison of the speed measurement error between the proposed system and the other four methods in the scenario with license plate not covered. It can be seen that the root mean square error (RMSE) of the proposed system is smaller than that of the other four methods, and the maximum error is also smaller than that of the other four methods. Therefore, in the scenario with license plate not covered, the speed measurement accuracy of the proposed method is superior to that of the other four methods, that is, the speed measurement accuracy of the system is improved.
Table 15 shows the speed measurement results of two contrast experiments in the scenario with license plate covered, in which the vehicle drives in a straight line at a constant speed of 33 km/h. As shown in Table 15, in the scenario with license plate covered, the speed measurement error rate range of the proposed system is [−2.50%, 4.99%] and the maximum absolute error rate is 4.99%, which meets the 6% error rate limit specified by the China national standard GB/T21255-2007 [59], while the system in [14] fails in the speed measurement. The proposed system can still measure the vehicle speed accurately in the scenario with license plate covered.
Meanwhile, the speed measurement performances are compared between the proposed system and the various existing speed measurement methods. Table 16 shows a comparison of the speed measurement error between the proposed system and the other four methods in the scenario with license plate covered. It can be seen that the methods in [8,14] fail, and the RMSE and maximum error of the proposed system are smaller than that of the other two methods. Therefore, in the scenario with license plate covered, the speed measurement robustness of the proposed method is superior to that of the other two invalid methods, and the speed measurement accuracy of the proposed method is superior to that of the other two valid methods, that is, the speed measurement robustness of the system is improved.

5. Conclusions

In this study, we solved the problem of effectively measuring the vehicle speed based on binocular stereovision system in case of license plate violation. We proposed a robust binocular stereovision-based vehicle speed measurement system with vehicle multi-characteristic detection. We optimized the object detection algorithm for vehicle multi-characteristic detection, and thus optimized the whole vehicle speed measurement system. The FPN-based YOLOv4 object detection algorithm was selected for the vehicle multi-characteristic detection, and the ECA channel attention mechanism was combined to improve YOLOv4. An improved ECA-YOLOv4 object detection algorithm was proposed for the vehicle multi-characteristic detection, which was trained and verified by our constructed vehicle multi-characteristic dataset. The experimental results showed that the proposed ECA-YOLOv4 object detection algorithm can efficiently improve the detection accuracy of multiple characteristics of the vehicle and minimize the network model parameter size. Three vehicle characteristics, that is, license plate, logo, and light, were chosen to design the corresponding speed measurement system for vehicle multi-characteristic detection. Many experiments were carried out in different speed measurement scenarios. Experimental results show that the proposed speed measurement system improves the speed measurement accuracy efficiently and effectively solve the problems in robustness of the system in [14].

Author Contributions

Conceptualization, L.Y. and X.S.; methodology, L.Y. and J.L.; formal analysis, L.Y.; data curation, J.L.; writing—original draft preparation, J.L.; writing—review and editing, L.Y., X.S., M.L. and P.W.; supervision, L.Y., X.S. and Z.X.; funding acquisition, X.S. All authors read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the ZhongYuan Science and Technology Innovation Leading Talent Program under Grant 214200510013, in part by the Key Research Project of Colleges and Universities in Henan Province under Grant 19A510005, Grant 21A510016, and Grant 21A520052, in part by the Scientific Research Grants and Start-up Projects for Overseas Student under Grant HRSS2021-36, and in part by the Major Project Achievement Cultivation Plan of Zhongyuan University of Technology under Grant K2020ZDPY02.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pramanik, A.; Sarkar, S.; Maiti, J. A real-time video surveillance system for traffic pre-events detection. Accid. Anal. Prev. 2021, 154, 106019. [Google Scholar] [CrossRef]
  2. Wang, Z.; Huang, J.; Xiong, N.N.; Zhou, X.; Lin, X.; Ward, T.L. A Robust Vehicle Detection Scheme for Intelligent Traffic Surveillance Systems in Smart Cities. IEEE Access 2020, 8, 139299–139312. [Google Scholar] [CrossRef]
  3. Phan, H.N.; Pham, L.H.; Chung, N.M.; Ha, S.V.U. Improved Shadow Removal Algorithm for Vehicle Classification in Traffic Surveillance System. In Proceedings of the 2020 RIVF International Conference on Computing and Communication Technologies (RIVF), Ho Chi Minh City, Vietnam, 14–15 October 2020; pp. 647–655. [Google Scholar]
  4. Tang, Y.; Zhang, C.; Gu, R.; Li, P.; Yang, B. Vehicle detection and recognition for intelligent traffic surveillance system. Multimed. Tools Appl. 2017, 76, 5817–5832. [Google Scholar] [CrossRef]
  5. Jiang, H.; Polak, J.W.; Barria, J.; Krishnan, R. On the estimation of space-mean-speed from inductive loop detector data. Transp. Plan. Technol. 2010, 33, 91–104. [Google Scholar]
  6. Jeng, S.L.; Chieng, W.H.; Lu, H.P. Estimating Speed Using a Side-Looking Single-Radar Vehicle Detector. IEEE Trans. Intell. Transp. Syst. 2014, 15, 607–614. [Google Scholar] [CrossRef]
  7. Mao, X.; Hou, J. Object-based forest gaps classification using airborne LiDAR data. J. For. Res. 2018, 30, 617–627. [Google Scholar] [CrossRef]
  8. Luvizon, D.C.; Nassu, B.T.; Minetto, R. A Video-Based System for Vehicle Speed Measurement in Urban Roadways. IEEE Trans. Intell. Transp. Syst. 2017, 18, 1393–1404. [Google Scholar] [CrossRef]
  9. Llorca, D.F.; Martínez, A.H.; Daza, I.G. Vision-based Vehicle Speed Estimation for ITS: A Survey. arXiv 2021, arXiv:2101.06159. [Google Scholar]
  10. Vakili, E.; Shoaran, M.; Sarmadi, M.R. Single–camera vehicle speed measurement using the geometry of the imaging system. Multimed. Tools Appl. 2020, 79, 19307–19327. [Google Scholar] [CrossRef]
  11. Sonth, A.; Settibhaktini, H.; Jahagirdar, A. Vehicle speed determination and license plate localization from monocular video streams. In Proceedings of 3rd International Conference on Computer Vision and Image Processing; Springer: Singapore, 2020; pp. 267–277. [Google Scholar]
  12. Jalalat, M.; Nejati, M.; Majidi, A. Vehicle detection and speed estimation using cascade classifier and sub-pixel stereo matching. In Proceedings of the 2016 2nd International Conference of Signal Processing and Intelligent Systems (ICSPIS), Tehran, Iran, 14–15 December 2016; pp. 1–5. [Google Scholar]
  13. Zhu, J.; Yuan, L.; Zheng, Y.F.; Ewing, R.L. Stereo Visual Tracking Within Structured Environments for Measuring Vehicle Speed. IEEE Trans. Circuits Syst. Video Technol. 2012, 22, 1471–1484. [Google Scholar] [CrossRef] [Green Version]
  14. Yang, L.; Li, M.; Song, X.; Xiong, Z.; Hou, C.; Qu, B. Vehicle Speed Measurement Based on Binocular Stereovision System. IEEE Access 2019, 7, 106628–106641. [Google Scholar] [CrossRef]
  15. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.E.; Fu, C.Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. In 14th European Conference on Computer Vision, ECCV 2016; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  16. Jiang, X.; Ma, J.; Xiao, G.; Shao, Z.; Guo, X. A review of multimodal image matching: Methods and applications. Inf. Fusion 2021, 73, 22–71. [Google Scholar] [CrossRef]
  17. Ma, J.; Jiang, X.; Fan, A.; Jiang, J.; Yan, J. Image Matching from Handcrafted to Deep Features: A Survey. Int. J. Comput. Vis. 2021, 129, 23–79. [Google Scholar] [CrossRef]
  18. Tan, Y.; Zhang, Y. Beijing: Forging “Sharp Swords” to Severely Punish license plate violation. Road Traffic Manag. 2019, 4, 44–45. [Google Scholar]
  19. Beijing, Tianjin and Hebei jointly investigated and dealt with illegal acts of false license plates for motor vehicles. Auto Saf. 2018, 1, 27.
  20. Police in Quanzhou, Fujian Province, Have Dealt with 8328 Cases of Drunk Driving and Four of Drag Racing This Year. Available online: http://www.rmjtxw.com/news/rxw/59216.html (accessed on 16 July 2021).
  21. Big Data Power Public Security Traffic Police, Set of Cars No Place to Hide. Available online: http://www.rmjtxw.com/news/dsj/36721.html (accessed on 16 July 2021).
  22. Three Months to Investigate and Deal with Fake Card Set more than 3500 Illegal Cases. Available online: https://www.chinacourt.org/article/detail/2017/06/id/2897523.shtml (accessed on 16 July 2021).
  23. Zheng, Z.; Zhen, S.; Yu, G.; Jie, Y. Object Detection in 20 Years: A Survey. arXiv 2019, arXiv:1905.05055. [Google Scholar]
  24. Zong, W.; Shi, P.; Feng, C.; Guo, L.; Cheng, Z.; Philip, S.Y. A Comprehensive Survey on Graph Neural Networks. IEEE Trans. Neural Netw. 2021, 32, 4–24. [Google Scholar]
  25. Jie, Z.; Gan, C.; Zheng, Z.; Cheng, Y.; Zhi, L.; Li, W.; Chang, L.; Mao, S. Graph neural networks: A review of methods and applications. AI Open 2020, 1, 57–81. [Google Scholar]
  26. Yang, J.; Guo, J.; Yue, H.; Liu, Z.; Hu, H.; Li, K. CDnet: CNN-Based Cloud Detection for Remote Sensing Imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6195–6211. [Google Scholar] [CrossRef]
  27. Hong, D.; Gao, L.; Yao, J.; Zhang, B.; Plazas, A.; Chanussot, J. Graph Convolutional Networks for Hyperspectral Image Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 5966–5978. [Google Scholar] [CrossRef]
  28. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  29. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the ICLR 2015: International Conference on Learning Representations 2015, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  30. Lin, M.; Chen, Q.; Yan, S. Network In Network. In Proceedings of the ICLR 2014: International Conference on Learning Representations (ICLR) 2014, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Iandola, F.; Moskewicz, M.; Karayev, S.; Girshick, R.; Keutzer, K. DenseNet: Implementing Efficient ConvNet Descriptor Pyramids. arXiv 2014, arXiv:1404.1869. [Google Scholar]
  33. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  34. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  35. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  36. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  37. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision (ICCV), Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  38. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  39. Liu, S.; Huang, D.; Wang, Y. Receptive Field Block Net for Accurate and Fast Object Detection. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 404–419. [Google Scholar]
  40. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 6517–6525. [Google Scholar]
  41. Dan, H.; Lian, G.; Naoto, Y.; Jing, Y.; Jocelyn, C.; Qian, D.; Bing, Z. More Diverse Means Better: Multimodal Deep Learning Meets Remote-Sensing Imagery Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 4340–4354. [Google Scholar]
  42. Zhu, X.; Xu, X.; Mu, N. Saliency Detection Based on the Combination of High-Level Knowledge and Low-Level Cues in Foggy Images. Entropy 2019, 21, 374. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Gu, Y.; Wang, X.; Zhang, C.; Li, B. Advanced Driving Assistance Based on the Fusion of Infrared and Visible Images. Entropy 2021, 23, 239. [Google Scholar] [CrossRef]
  44. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  45. Zhao, Q.; Sheng, T.; Wang, Y.; Tang, Z.; Ling, H. M2Det: A Single-Shot Object Detector Based on Multi-Level Feature Pyramid Network. Proc. AAAI Conf. Artif. Intell. 2019, 33, 9259–9266. [Google Scholar] [CrossRef]
  46. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 1, 2999–3007. [Google Scholar]
  47. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  48. Li, X.; Wang, W.; Hu, X.; Yang, J. Selective Kernel Networks. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 510–519. [Google Scholar]
  49. Pan, Z.; Yu, W.; Lei, J.; Ling, N.; Kwong, S. TSAN: Synthesized View Quality Enhancement via Two-Stream Attention Network for 3D-HEVC. IEEE Trans. Circuits Syst. Video Technol. 2021, 1–14. [Google Scholar] [CrossRef]
  50. Gupta, A.K.; Seal, A.; Prasad, M.; Khanna, P. Salient Object Detection Techniques in Computer Vision-A Survey. Entropy 2020, 22, 1174. [Google Scholar] [CrossRef] [PubMed]
  51. Hu, J.; Shen, L.; Albanie, S.; Sun, G.; Wu, E. Squeeze-and-Excitation Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 42, 2011–2023. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Wang, Q.; Wu, B.; Zhu, P.; Li, P.; Hu, Q. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  53. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  54. Zhang, Z.; Sabuncu, M.R. Generalized cross entropy loss for training deep neural networks with noisy labels. In Proceedings of the 32nd International Conference on Neural Information Processing Systems, NIPS’18, Montreal, QC, Canada, 3–8 December 2018; Volume 31, pp. 8792–8802. [Google Scholar]
  55. Guangdong Key Laboratory of Intelligent Transportation System. OpenIts Research Project. Available online: http://www.openits.cn/ (accessed on 16 July 2021).
  56. Dong, Z.; Wu, Y.; Pei, M.; Jia, Y. Vehicle Type Classification Using a Semisupervised Convolutional Neural Network. IEEE Trans. Intell. Transp. Syst. 2015, 16, 2247–2256. [Google Scholar] [CrossRef]
  57. Xu, Z.; Yang, W.; Meng, A.; Lu, N.; Huang, H.; Ying, C.; Huang, L. Towards End-to-End License Plate Detection and Recognition: A Large Dataset and Baseline. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 261–277. [Google Scholar]
  58. Xie, W.; Zhu, D.; Tong, X. A small object detection method based on visual attention. Comput. Eng. Appl. 2013, 49, 125–128. [Google Scholar]
  59. Zou, Y. Motor Vehicle Speed Detector. In China National Standard GB/T 21255-2007; State Administration for Market Regulation, Standardization Administration: Beijing, China, 2007. [Google Scholar]
  60. Tang, Z.; Wang, G.; Xiao, H.; Zheng, A.; Hwang, J.N. Single-Camera and Inter-Camera Vehicle Tracking and 3D Speed Estimation Based on Fusion of Visual and Semantic Features. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Salt Lake City, UT, USA, 18–22 June 2018; pp. 108–1087. [Google Scholar]
  61. El Bouziady, A.; Thami, R.O.H.; Ghogho, M.; Bourja, O.; El Fkihi, S. Vehicle speed estimation using extracted SURF features from stereo images. In Proceedings of the 2018 International Conference on Intelligent Systems and Computer Vision (ISCV), Fez, Morocco, 2–4 April 2018; pp. 1–6. [Google Scholar]
Figure 1. The proposed system model.
Figure 1. The proposed system model.
Entropy 23 00910 g001
Figure 2. System configuration.
Figure 2. System configuration.
Entropy 23 00910 g002
Figure 3. The whole procedure of the proposed vehicle speed measurement system with vehicle multi-characteristic detection.
Figure 3. The whole procedure of the proposed vehicle speed measurement system with vehicle multi-characteristic detection.
Entropy 23 00910 g003
Figure 4. Image examples of our vehicle multi-characteristic dataset. (a) Image from CCPD. (b) Image from BIT Vehicle. (c) Image from Open ITS. (d) Self-captured image.
Figure 4. Image examples of our vehicle multi-characteristic dataset. (a) Image from CCPD. (b) Image from BIT Vehicle. (c) Image from Open ITS. (d) Self-captured image.
Entropy 23 00910 g004
Figure 5. Pixel ratio graph of multiple characteristics in the speed measurement range of 0 to 15 m.
Figure 5. Pixel ratio graph of multiple characteristics in the speed measurement range of 0 to 15 m.
Entropy 23 00910 g005
Figure 6. Schematic block diagram of the YOLOv4-based vehicle multi-characteristic object detection algorithm.
Figure 6. Schematic block diagram of the YOLOv4-based vehicle multi-characteristic object detection algorithm.
Entropy 23 00910 g006
Figure 7. CSP structure in YOLOv4.
Figure 7. CSP structure in YOLOv4.
Entropy 23 00910 g007
Figure 8. Residual unit structure in YOLOv4.
Figure 8. Residual unit structure in YOLOv4.
Entropy 23 00910 g008
Figure 9. Residual unit structure in ECA-YOLOv4.
Figure 9. Residual unit structure in ECA-YOLOv4.
Entropy 23 00910 g009
Figure 10. Vehicle multi-characteristic detection results: (a) left-view and (b) right-view.
Figure 10. Vehicle multi-characteristic detection results: (a) left-view and (b) right-view.
Entropy 23 00910 g010
Figure 11. Speed measurement error rate curve for single characteristic of vehicle.
Figure 11. Speed measurement error rate curve for single characteristic of vehicle.
Entropy 23 00910 g011
Figure 12. Four speed measurement scenarios. (a) Speed measurement scenario where license plate, logo and light all exist; (b) Speed measurement scenario where license plate and light exist; (c) Speed measurement scenario where logo and light exist; (d) Speed measurement scenario where only light exists.
Figure 12. Four speed measurement scenarios. (a) Speed measurement scenario where license plate, logo and light all exist; (b) Speed measurement scenario where license plate and light exist; (c) Speed measurement scenario where logo and light exist; (d) Speed measurement scenario where only light exists.
Entropy 23 00910 g012
Figure 13. Speed measurement curve of two contrast experiments in the scenario with license plate not covered.
Figure 13. Speed measurement curve of two contrast experiments in the scenario with license plate not covered.
Entropy 23 00910 g013
Table 1. Performance comparison of multiple object detection algorithms for vehicle multi-characteristic detection.
Table 1. Performance comparison of multiple object detection algorithms for vehicle multi-characteristic detection.
AlgorithmAP car (%)AP plate (%)AP logo (%)AP light (%)AP mirror (%)mAP (%)
Faster R-CNN87.7545.8253.9179.3268.8467.13
SSD94.9077.1671.6887.7980.8282.47
RFB94.6256.9360.2182.6780.0874.90
Retinanet93.9733.1451.6173.5956.8061.82
M2Det95.8768.7571.1587.2784.8481.58
YOLOv393.9188.0681.9490.8986.9988.35
YOLOv496.4792.1387.7294.1791.2092.34
Table 2. The model parameter size and FLOPs of the object detection networks with different attention modules.
Table 2. The model parameter size and FLOPs of the object detection networks with different attention modules.
AlgorithmInput SizeModel Parameters SizeModel Compression RateFLOPs
YOLOv4416 × 416244.29 MB/29.95 G
CBAM-YOLOv4416 × 416180.51 MB26.10%19.28 G
SE-YOLOv4416 × 416179.66 MB26.45%19.28 G
ECA-YOLOv4416 × 416178.82 MB26.80%19.28 G
Table 3. Performance comparison of the object detection networks with different attention modules.
Table 3. Performance comparison of the object detection networks with different attention modules.
AlgorithmAP car (%)AP plate (%)AP logo (%)AP light (%)AP mirror (%)mAP (%)FPS
YOLOv496.4792.1387.7294.1791.2092.3422
ECA-YOLOv497.0193.1987.9894.5391.9992.9423
SE-YOLOv497.0293.2488.3094.4992.2993.0721
CBAM-YOLOv497.2793.4788.4893.9191.7692.9816
Table 4. Speed measurement results of single characteristics of the vehicle.
Table 4. Speed measurement results of single characteristics of the vehicle.
NumberPlateLogoLightMirrorSatellite Speed (km/h)
Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)
142.29−2.4943.700.7642.53−1.9479.6583.6543.37
243.550.4143.410.0843.500.3022.43−48.2843.37
344.522.2844.712.7144.472.1639.10−10.1843.53
445.323.5645.163.2046.215.6058.2433.0943.76
543.28−1.1045.163.2043.09−1.5313.04−70.2043.76
643.22−1.0642.06−3.7143.940.60124.35184.6843.68
744.632.5343.640.2641.72−4.169.57−78.0243.53
842.39−2.2542.05−3.0441.95−3.277.69−82.2743.37
945.344.5544.041.5643.560.4475.3473.7143.37
1043.090.3441.62−3.0745.375.6642.48−1.0742.94
Error rate range[−2.49%, 4.55%][−3.71%, 3.20%][−4.16%, 5.66%][−82.27%, 184.69%]
Table 5. The speed measurement results in the scenario where license plate, logo, and light all exist.
Table 5. The speed measurement results in the scenario where license plate, logo, and light all exist.
NumberPlateLogoLightAverageSatellite Speed (km/h)
Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)
144.81−2.4845.30−1.4144.64−2.8544.92−2.2545.95
245.79−0.7446.500.8046.570.9546.290.3446.13
347.081.5048.274.0845.34−2.2446.901.1146.38
447.021.0148.885.0145.87−1.4647.261.5246.55
547.421.5046.71−0.0243.93−5.9746.02−1.5046.72
647.211.4144.67−4.0447.612.2846.50−0.1246.55
746.870.4845.64−2.1446.17−1.0146.23−0.8946.64
845.22−3.3946.36−0.9646.54−0.5846.04−1.6446.81
945.26−3.4946.25−1.3944.77−4.5445.43−3.1446.9
1045.16−3.3344.82−4.0745.23−3.1945.07−3.5346.72
Error rate range[−3.49%, 1.50%][−4.07%, 5.01%][−5.97%, 2.28%][−3.53%, 1.52%]
Table 6. The error rates of multiple experiments in the average speed measurement of the three vehicle characteristics (license plate, logo, and light).
Table 6. The error rates of multiple experiments in the average speed measurement of the three vehicle characteristics (license plate, logo, and light).
NumberError Rate of Test 2 (%)Error Rate of Test 3 (%)Error Rate of Test 4 (%)Error Rate of Test 5 (%)Error Rate of Test 6 (%)Error Rate of Test 7 (%)Error Rate of Test 8 (%)
1−3.080.25−0.011.55−2.04−1.22−0.77
2−1.141.931.842.82−1.390.271.90
32.921.202.771.800.862.38−0.84
40.47−1.280.871.96−1.074.12−0.28
5−1.910.843.61−0.881.240.193.20
61.021.892.202.060.26−1.391.65
71.61−0.88−2.232.87−1.55−0.461.06
8−2.771.880.33−1.39−0.51−2.860.78
9−2.23−1.361.250.36−0.042.181.07
101.652.62−0.241.543.970.981.65
Error rate range[−3.08%, 2.92%][−1.36%, 2.62%][−2.23%, 3.61%][−1.39%, 2.87%][−2.04%, 3.97%][−2.86%, 4.12%][−0.84%, 3.20%]
Table 7. The speed measurement results in the scenario where license plate and light exist.
Table 7. The speed measurement results in the scenario where license plate and light exist.
NumberPlateLightAverageSatellite Speed (km/h)
Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)Speed (km/h)Error Rate (%)
128.84−3.62%30.923.35%29.88−0.14%29.92
229.972.62%30.574.68%30.273.65%29.2
330.581.83%31.364.42%30.973.12%30.03
431.805.22%30.962.44%31.383.83%30.22
531.041.44%29.62−3.20%30.33−0.88%30.6
631.121.69%29.72−2.87%30.42−0.59%30.6
730.910.41%29.45−4.35%30.18−1.97%30.79
829.05−4.76%32.305.90%30.670.57%30.5
929.78−2.68%30.35−0.81%30.07−1.74%30.6
1032.014.51%31.773.74%31.894.12%30.63
Error rate range[−4.76%, 5.22%][−4.35%, 5.90%][−1.97%, 4.12%]
Table 8. The error rates of multiple experiments in the average speed measurement of the two vehicle characteristics (license plate and light).
Table 8. The error rates of multiple experiments in the average speed measurement of the two vehicle characteristics (license plate and light).
NumberError Rate of Test 2 (%)Error Rate of Test 3 (%)Error Rate of Test 4 (%)Error Rate of Test 5 (%)Error Rate of Test 6 (%)Error Rate of Test 7 (%)Error Rate of Test 8 (%)
1−3.021.53−2.211.96−2.67−0.570.14
2−1.552.110.362.670.11−1.951.48
34.501.322.220.76−0.371.41−1.57
4−1.48−3.364.583.53−0.23−0.57−0.02
5−2.240.92−1.31−2.08−2.241.952.58
60.421.63−0.234.241.841.030.19
72.890.33−0.822.54−0.26−0.874.22
8−2.991.54−2.76−0.68−1.991.531.07
9−4.16−0.262.49−1.44−4.02−2.143.25
101.023.333.002.67−3.263.371.95
Error rate range[−4.16%, 4.50%][−3.36%, 3.33%][−2.76%, 4.58%][−2.08%, 4.24%][−4.02%, 1.84%][−2.14%, 3.37%][−1.57%, 4.22%]
Table 9. The speed measurement results in the scenario where logo and light exist.
Table 9. The speed measurement results in the scenario where logo and light exist.
NumberLogoLightAverageSatellite Speed (km/h)
Result (km/h)Error Rate (%)Result (km/h)Error Rate (%)Result (km/h)Error Rate (%)
142.96−3.2043.59−1.7843.28−2.4944.38
244.61−0.3444.47−0.6544.54−0.4944.76
344.73−0.2447.014.8445.872.3044.84
446.984.3844.75−0.5945.871.9045.01
544.52−1.2643.14−4.3243.83−2.7945.09
646.092.2244.40−1.5445.250.3445.09
744.68−0.9446.442.9745.561.0245.1
844.04−2.3244.28−1.8044.16−2.0645.09
945.831.6542.96−4.7244.40−1.5445.09
1046.442.9045.280.3345.861.6245.13
Error rate range[−3.20%, 4.38%][−4.72%, 4.84%][−2.79%, 2.30%]
Table 10. The error rates of multiple experiments in the average speed measurement of the two vehicle characteristics (logo and light).
Table 10. The error rates of multiple experiments in the average speed measurement of the two vehicle characteristics (logo and light).
NumberError Rate of Test 2 (%)Error Rate of Test 3 (%)Error Rate of Test 4 (%)Error Rate of Test 5 (%)Error Rate of Test 6 (%)Error Rate of Test 7 (%)Error Rate of Test 8 (%)
1−1.380.510.451.85−2.13−2.50−0.02
2−0.030.973.262.360.880.673.57
32.271.401.902.480.922.38−1.64
41.29−0.910.520.011.77−1.11−0.11
50.212.293.45−1.16−3.002.524.31
60.011.451.990.93−0.88−1.264.62
74.33−2.99−1.393.84−1.58−0.61−0.82
8−5.613.930.42−3.85−0.770.221.24
9−0.81−1.081.980.76−2.961.02−0.02
100.362.031.25−0.11−3.634.990.62
Error rate range[−5.61%, 4.33%][−2.99%, 3.93%][−1.39%, 3.45%][−3.85%, 3.84%][−3.63%, 1.77%][−2.50%, 4.99%][−1.64%, 4.62%]
Table 11. The speed measurement results in the scenario where only light exists.
Table 11. The speed measurement results in the scenario where only light exists.
NumberLight (km/h)Satellite Speed (km/h)Error Rate (%)
137.3136.123.30
236.2436.110.37
336.8636.191.84
434.5636.26−4.68
537.7536.333.90
636.6736.490.49
735.2236.19−2.68
838.1736.255.30
936.7236.211.41
1037.2636.222.87
Error rate range[−4.68%, 5.30%]
Table 12. The error rates of multiple experiments in the speed measurement of the one vehicle characteristic (light).
Table 12. The error rates of multiple experiments in the speed measurement of the one vehicle characteristic (light).
NumberError Rate of Test 2 (%)Error Rate of Test 3 (%)Error Rate of Test 4 (%)Error Rate of Test 5 (%)Error Rate of Test 6 (%)Error Rate of Test 7 (%)Error Rate of Test 8 (%)
13.351.822.96−2.85−0.032.56−1.78
24.684.571.610.951.614.39−0.65
34.42−1.231.07−2.245.01−3.904.84
42.440.741.21−1.46−0.150.59−0.59
5−3.204.41−3.82−5.975.214.18−4.32
6−2.872.104.162.28−1.254.67−1.54
7−4.35−0.034.16−1.011.673.642.97
85.90−3.87−4.86−0.585.032.28−1.80
9−0.814.14−2.46−4.54−2.133.25−4.72
103.744.330.51−3.194.810.190.33
Error rate range[−4.35%, 5.90%][−3.87%, 4.57%][−4.86%, 4.16%][−5.97%, 2.28%][−2.13%, 5.21%][−3.90%, 4.67%][−4.72%, 4.84%]
Table 13. Speed measurement results of two contrast experiments in the scenario with license plate not covered.
Table 13. Speed measurement results of two contrast experiments in the scenario with license plate not covered.
NumberProposed System (km/h)System in [14] (km/h)Satellite Speed (km/h)
Speed (km/h)Error (km/h)Error Rate (%)Speed (km/h)Error (km/h)Error Rate (%)
132.390.01−0.0332.10−0.30−0.9432.4
233.000.601.8432.07−0.33−1.0232.4
333.340.902.7733.911.474.5332.44
432.750.280.8732.980.511.5632.47
533.611.173.6133.721.283.9432.44
633.110.712.2033.240.842.6132.4
731.59−0.72−2.2331.05−1.26−3.9132.31
832.420.110.3332.360.050.1532.31
932.740.401.2532.27−0.07−0.2132.34
1032.23−0.08−0.2431.27−1.04−3.2132.31
Error rate range[−2.23%, 3.61%][−3.91%, 4.53%]
Table 14. Error comparison among different speed measurement methods in the scenario with license plate not covered.
Table 14. Error comparison among different speed measurement methods in the scenario with license plate not covered.
RMSE (km/h)Max Error (km/h)
Luvizon et al. [8]1.36[−4.68, +6.00]
Tang et al. [60]6.59NA
VSS-SURF [61]1.29[−2.0, +2.0]
System in [14]0.65[−1.6, +1.1]
Proposed System0.43[−0.72, +1.17]
Table 15. Speed measurement results of two contrast experiments in the scenario with license plate covered.
Table 15. Speed measurement results of two contrast experiments in the scenario with license plate covered.
NumberProposed System (km/h)System in [14] (km/h)Satellite Speed (km/h)
Speed (km/h)Error (km/h)Error Rate (%)Speed (km/h)Error (km/h)Error Rate (%)
132.96−0.84−2.50InvalidInvalidInvalid33.8
233.790.220.67InvalidInvalidInvalid33.56
334.570.812.38InvalidInvalidInvalid33.76
433.39−0.37−1.11InvalidInvalidInvalid33.76
534.610.852.52InvalidInvalidInvalid33.76
633.24−0.42−1.26InvalidInvalidInvalid33.66
733.36−0.20−0.61InvalidInvalidInvalid33.56
833.480.080.22InvalidInvalidInvalid33.4
933.700.341.02InvalidInvalidInvalid33.36
1034.911.664.99InvalidInvalidInvalid33.25
Error rate range[−2.50%, 4.99%]NA
Table 16. Error comparison among different speed measurement methods in the scenario with license plate covered.
Table 16. Error comparison among different speed measurement methods in the scenario with license plate covered.
RMSE (km/h)Max Error (km/h)
Luvizon et al. [8]InvalidInvalid
Tang et al. [60]6.59NA
VSS-SURF [61]1.29[−2.0, +2.0]
System in [14]InvalidInvalid
Proposed System0.48[−0.84, +1.66]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, L.; Luo, J.; Song, X.; Li, M.; Wen, P.; Xiong, Z. Robust Vehicle Speed Measurement Based on Feature Information Fusion for Vehicle Multi-Characteristic Detection. Entropy 2021, 23, 910. https://doi.org/10.3390/e23070910

AMA Style

Yang L, Luo J, Song X, Li M, Wen P, Xiong Z. Robust Vehicle Speed Measurement Based on Feature Information Fusion for Vehicle Multi-Characteristic Detection. Entropy. 2021; 23(7):910. https://doi.org/10.3390/e23070910

Chicago/Turabian Style

Yang, Lei, Jianchen Luo, Xiaowei Song, Menglong Li, Pengwei Wen, and Zixiang Xiong. 2021. "Robust Vehicle Speed Measurement Based on Feature Information Fusion for Vehicle Multi-Characteristic Detection" Entropy 23, no. 7: 910. https://doi.org/10.3390/e23070910

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop