Next Article in Journal
Depth-Sensing-Based Algorithm for Chest Morphology Assessment in Children with Cerebral Palsy
Previous Article in Journal
Electrochemical Sensors for Antibiotic Detection: A Focused Review with a Brief Overview of Commercial Technologies
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Detection of Road Risk Sources Based on Multi-Scale Lightweight Networks

by
Rong Pang
1,2,3,
Jiacheng Ning
4,
Yan Yang
1,*,
Peng Zhang
2,3,
Jilong Wang
2,3 and
Jingxiao Liu
2,3
1
School of Computing and Artificial Intelligence, Southwest Jiaotong University, Chengdu 611756, China
2
China Merchants Chongqing Road Engineering Inspection Center Co., Ltd., Chongqing 400067, China
3
National Mountain Highway Engineering Technology Research Center, Chongqing 400067, China
4
Chongqing Transportation Planning and Technology Development Center, Chongqing 400067, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5577; https://doi.org/10.3390/s24175577
Submission received: 3 July 2024 / Revised: 14 August 2024 / Accepted: 27 August 2024 / Published: 28 August 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Timely discovery and disposal of road risk sources constitute the cornerstone of road operation safety. Presently, the detection of road risk sources frequently relies on manual inspections via inspection vehicles, a process that is both inefficient and time-consuming. To tackle this challenge, this paper introduces a novel automated approach for detecting road risk sources, termed the multi-scale lightweight network (MSLN). This method primarily focuses on identifying road surfaces, potholes, and scattered objects. To mitigate the influence of real-world factors such as noise and uneven brightness on test results, pavement images were carefully collected. Initially, the collected images underwent grayscale processing. Subsequently, the median filtering algorithm was employed to filter out noise interference. Furthermore, adaptive histogram equalization techniques were utilized to enhance the visibility of cracks and the road background. Following these preprocessing steps, the MSLN model was deployed for the detection of road risk sources. Addressing the challenges associated with two-stage network models, such as prolonged training and testing times, as well as deployment difficulties, this study adopted the lightweight feature extraction network MobileNetV2. Additionally, transfer learning was incorporated to elevate the model’s training efficiency. Moreover, this paper established a mapping relationship model that transitions from the world coordinate system to the pixel coordinate system. This model enables the calculation of risk source dimensions based on detection outcomes. Experimental results reveal that the MSLN model exhibits a notably faster convergence rate. This enhanced convergence not only boosts training speed but also elevates the precision of risk source detection. Furthermore, the proposed mapping relationship coordinate transformation model proves highly effective in determining the scale of risk sources.

1. Introduction

At present, China is in the critical period of the development of intelligent transportation, and the construction of intelligent transportation has an important guiding role in the maintenance and management of basic road facilities. China’s road network stretched to 5.198 million kilometers, marking a growth of 185,600 km over the prior year’s end [1]. Among them, the total mileage of road maintenance is 5.144 million km, accounting for 99.0% of the total mileage of roads [2]. Cracks are one of the main factors affecting the performance of roads.
For a long time, road risk source detection has been an important research topic in academic and engineering circles at home and abroad. From the traditional image processing technology to the current hot deep learning technology, road risk source detection technology is constantly updated and has achieved some excellent research results. Crack detection based on threshold segmentation is the earliest algorithm studied [3]. Oh et al. proposed an iterative threshold segmentation algorithm in 1998 [4], but the threshold needs to be set according to engineering experience. Then, Li et al. proposed an improved OSTU algorithm based on the Hough transform [5]. Next, Sun et al. proposed a detection algorithm based on the adaptive threshold Canny algorithm, which no longer required manual threshold setting [6]. But when the light or noise in the image is not uniform, it still cannot produce a stable effect [7,8,9,10,11,12,13]. Subirats et al. proposed a crack detection method based on continuous wavelet transform, which could effectively detect and segment crack images but still had some noise [14]. Hou et al. proposed research and comparison of road surface detection methods based on data depth enhancement, which mainly adopted a data depth enhancement method integrating WGAN-GP and Poisson migration algorithms to supplement and balance training sample data by generating road pit pictures under different occlusion and different light conditions [15]. Huang et al. described a height mutation recognition method based on 3D data. Compared with the gray threshold method and artificial intelligence method, the three-dimensional detection of pavement potholes by linear structured light scanning was proposed [16]. Cha Y J et al. first applied convolutional neural networks (CNN) to detect cracks in concrete, abandoning the traditional practice of manual feature extraction. CNN can accurately identify cracks in the original images that contain variations in lighting and shadows, significantly improving the level of automation and accuracy of the detection [17]. Cha Y J et al. were also the first to adopt the Faster R-CNN visual detection framework, achieving simultaneous recognition of various structural damages. This framework can accurately detect a variety of damage types, including concrete cracks, different degrees of steel corrosion, bolt corrosion, and steel delamination [18]. Xu et al. used ResNet50, the residual network of cavity convolution optimization, as the backbone neural network to extract features and carry out multi-scale processing and proposed a road scatter detection algorithm based on center mask optimization of the case segmentation model [19]. Tien et al. proposed a crack target recognition method FFA that uses both luminance and morphological features to segment fracture targets, which can effectively improve the accuracy of fracture detection [20]. In recent years, due to the rapid development of deep learning, an increasing number of lightweight road risk source detection methods have emerged. Cha et al. proposed a road surface detection method based on a convolutional neural network, which classifies whether each sub-area contains a crack region by the output value of the model. Gopalakrishnan et al. [21] proposed a road crack detection method based on transfer learning, which used the pre-trained convolutional neural network based on the ImageNet dataset and then fine-tuned it to obtain the final detection model. Chen et al. realized the use of the NB-CNN network for crack detection but could not achieve accurate extraction of crack shape [22]. Carr et al. proposed a methodology utilizing a single stage of the target detection network [23]. RetinaNet for road crack detection uses a neural network to extract feature residual and uses the feature extraction of a multi-scale pyramid network diagram, with a slide on each scale using anchor boxes to generate multiple candidate areas and classifying the candidate region and border position return for detecting cracks [24,25]. Liu et al. proposed PDT-YOLO, an efficient roadside object detection algorithm addressing challenges in multi-scale, occlusion, and model deployment for intelligent roadside perception systems [26]. Yuan et al. proposed a super-resolution reconstruction method for pavement crack images based on an improved generative adversarial network (GAN), addressing the issue of significant disparities in image quality due to variable equipment and illumination conditions during the image-collecting stage. As demonstrated in experiments, the method effectively improves the accuracy of pavement crack detection [27]. Rehana K C et al. proposed the YOLOv5-PD model specifically for detecting common defects in asphalt pavements, such as cracks, patches, and potholes, demonstrating the effectiveness of this model in handling defects of varying scales [28]. Kang D et al. proposed an integrated Faster R-CNN algorithm for detecting crack area, and combined an improved tubularity flow field (TuFF) algorithm to achieve pixel-level crack segmentation. In addition, this method also employed a modified distance transform method (DTM) to accurately measure the thickness and length of cracks [29].
In this paper, a two-stage lightweight network is applied to road hazard source detection. First, a road hazard source dataset based on forward computer vision is constructed, and MobileNetV2 is introduced to modify the categories of the computational model. The bounding box loss functions, Alpha IOU and PolyLoss, are improved, thereby enhancing the main feature extraction network of the model. The MSLN network model is proposed. Experiments show that it can effectively improve the efficiency of model training and detection and increase detection accuracy. Additionally, by utilizing visual coordinate transformation to estimate the scale of hazard sources, the model can estimate the scale of hazard sources, meeting the requirements of road inspection.

2. Construction and Augmentation of Road Risk Source Datasets

To obtain the image data of road potholes and sprinkles that can be used for model training, we first obtained the road image of actual road maintenance work. Due to the constraints of acquisition equipment and environment, the obtained image has uneven illumination and many noise points, which leads to low image quality. Therefore, we need to preprocess the image first. The road image is preprocessed by the traditional digital image processing method. A noise removal algorithm is applied to the grayscale image, followed by histogram equalization to boost the contrast of the pits, sprinkling objects, and background, and the first batch of images is obtained. Finally, for the first batch of images, random clipping, rotation, flipping, and other common methods in deep learning are used to augment the images to obtain the final experimental dataset.

2.1. Grayscale Images of Risk Sources

The images collected in road maintenance are RGB three-color channel images, which occupy a large memory. However, in the actual pit and sprinkling recognition and detection work, the pixel difference between the target and the background is large, so the image gray processing does not affect the recognition, and detection work can reduce memory usage and is conducive to the subsequent image noise reduction and enhancement. Therefore, in this paper, the color image is gray processed to obtain a single-channel gray image, mainly taking the pit as an example, as shown in Figure 1.

2.2. Crack Image Denoising

In the way of daily maintenance, collected road pavement images are often affected by environmental and human factors, such as a lot of noise, so reducing the noise influence on the crack detection requires a filter algorithm of image noise reduction, which will preserve image details and remove noise as much as possible. This paper mainly uses a Gaussian filtering algorithm to process the image.
Gaussian filtering is a kind of classic filtering method, which realizes the filter in the same calculation template size of neighborhood pixel gray value of the mean as a filter within the collection after the goal of pixel values is reached; however, in the Gaussian filter, the farther the distance, the smaller the target pixel distance position influence coefficient, thus reducing the after filter in the neighborhood pixel difference caused by the blur effect. Therefore, compared with other filters, Gaussian filter can better retain the detailed information of crack targets [29]. Gaussian noise is a common noise that conforms to the normal distribution model and the probability density function formula such as Equation (1):
P ( z ) = 1 2 π σ e ( z u ) 2 / 2 σ 2
where z represents the gray value of the image, u represents the mean value of gray, and σ represents the standard value of gray. The image processed by Gaussian filtering is shown in Figure 2.

2.3. Enhancement of the Fracture Image

After image processing with a denoising filtering algorithm, the pavement crack image is relatively smooth. To enhance the difference between the target and background of road cracks and improve the detection and recognition effect, this paper mainly enhances the image by histogram equalization and expands the sample number to enhance the generalization of the model by using common data augmentation methods in deep learning, such as random clipping, rotation, and gamet transformation.
Histogram equalization is a method to enhance image contrast. Its main idea is to change the histogram distribution of an image into an approximate uniform distribution to enhance the image contrast [30]. A typical histogram equalization formula is presented in Equation (2):
S i = T ( r i ) = j = 0 k n i N
where S i represents the gray value of the output image, r i corresponds to the gray value obtained from the input image, n i is the number of pixels in the gray value r i , N is the total number of pixels in the image, and k is the gray level. The sum term indicates the percentage of pixels with a value of i relative to the overall pixel count. The image after equalization is shown in Figure 3.
Then, for the images after equalization processing, we used the method of data augmentation to construct the final sample set of road cracks, with a total of 1200 images. The image after data augmentation is shown in Figure 4.

3. Recognition Algorithm Based on Lightweight Network

3.1. Improve the Trunk Feature Extraction Network

As one of the two-stage target detection networks, Faster RCNN has a long training detection time and high hardware requirements, which is one of its major shortcomings. In the daily road maintenance and inspection scene, the hardware facilities of the terminal equipment often cannot meet the requirements, so it is urgent to develop a lightweight network model to improve this disadvantage and improve the training and detection efficiency of the network model.
MobileNetV2 focuses on lightweight convolutional neural networks in mobile terminals or embedded devices and is proposed to solve the shortcomings of traditional convolutional neural networks, such as large memory requirements and large computation. Depth-detachable convolution in the network can greatly reduce the number of parameters in the operation of our network and improve the operation speed. Traditional convolution and depth-detachable convolution are shown in Figure 5 and Figure 6.
It can be seen from Figure 5 and Figure 6 that in traditional convolution, the convolution kernel channel is equivalent to the input eigenmatrix channel depth. For depth-separable convolution, the DW convolution, with the depth of convolution kernel equal to 1 and the number of convolution kernels as the input eigenmatrix, is first used to obtain the output of the middle layer. The final output is derived by convolving with a 1 × 1 convolution kernel. Assuming that D F denotes the dimensions of the input eigenmatrix, D K denotes the size of the convolution kernel, M denotes the depth of the input eigenmatrix, and N denotes the depth of the output eigenmatrix, the calculation formula of the depthable convolution parameter and ordinary convolution is given Equation (3):
D K × D K × M × D F × D F + M × N × D F × D F D K × D K × M × N × D F × D F = 1 N + 1 D K 2 = 1 N + 1 9
When our convolution kernel is 3 times 3, the ordinary convolution is 8 or 9 times bigger than our deep-separable convolution.
The backward residual structure in the MobileNetV2 network can increase the feature information extraction of the input image, as shown in Figure 7.
Departing from the conventional residual structure, the traditional approach initially uses a 1 × 1 convolution kernel to compress our input eigenmatrix, that is, to reduce the channel of our input eigenmatrix. And then, it is processed by a 3 × 3 convolution kernel; finally, we use a 1 × 1 convolution kernel to extend its channel. This creates a bottleneck structure that is wide at both ends and narrow in the middle. In contrast, for the inverted residual structure features, the 1 × 1 convolution kernel is first used for dimension raising operation, and the number of channels becomes deeper. Then, the DW convolution operation of the 3 × 3 convolution kernel is used for convolution, and more convolution is extracted for dimension reduction through the 1 × 1 convolution. Therefore, the trunk feature extraction network in Faster R-CNN is replaced with MobileNetV2 to shorten the model training time and improve training efficiency.

3.2. Improve Loss Function

The loss function is of great significance in the training of neural networks. In the two-stage target detection network model Faster RCNN, loss in the network is mainly divided into two parts: regional suggestion network loss and Fast RCNN network loss. In the calculation area, the category of the network is suggested, that is, to distinguish the loss between the foreground and the background where binary cross-entropy loss is mainly used, and Smooth L1 loss is mainly used for the positioning loss. In the calculation of the category loss of the Fast RCNN network, cross-entropy loss is mainly used, and Smooth L1 loss is still used for the positioning loss of candidate box coordinates.
To use a more flexible form of the loss function, the PolyLoss function [31] was introduced to replace the cross-entropy loss function in the Fast RCNN network. Designing the loss function as a linear combination of polynomial functions facilitates the adjustment of the importance of different polynomials according to various target tasks and datasets. Its calculation formula is presented in Equation (4):
L P o l y = j = 1 α j ( 1 P t ) j
where α j represents the polynomial coefficient, and P t represents the probability of predicting categories.
Meanwhile, taking into account Smooth L1 losses, its calculation formula is shown in Equation (5):
Smooth   L 1 = 0.5 x 2 , | x | < 1 | x | 0.5 , x < 1   or   x > 1
When calculating the loss of the boundary box, the loss function first independently calculates the loss of four points—A, B, C, and D—and then adds them together, as depicted in Figure 8. In the figure, red represents the ground truth and green represents the prediction box.
This approach presumes the mutual independence of the four points; however, in reality, there exists a correlation between the candidate box’s four coordinates because they all depend on the same target. Therefore, alpha-iOU loss [30] is introduced in this paper to replace Smooth L1 loss of the region suggestion network and Fast RCNN network, respectively, for accurate candidate box regression and target detection. The formula of alpha-iOU loss function is given in Equation (6). The IOU calculation of the candidate box and the ground truth is shown in Figure 9.
L α IOU = 1 IOU α
where α represents power operation.

4. Dimension Estimation of Disease

In the field of road surface detection, the identification and assessment of the severity of pavement distress are crucial for maintaining road safety and extending the service life of the road. When pavement distress is detected, merely identifying its type and location is insufficient; it is also necessary to calculate its dimensions to more comprehensively evaluate the extent of pavement damage. Dimensional information can assist engineers in determining the severity of the distress, devising appropriate repair schemes, and predicting future maintenance needs. To compute the dimension of pavement distress, it is essential to convert the three-dimensional position information of the distress in the world coordinate system to two-dimensional pixel coordinates in the image coordinate system. This involves a series of complex coordinate system transformation processes requiring the use of camera calibration and pose estimation techniques.
In the domains of image processing and stereo vision, accurately depicting the position and orientation of objects relies on four levels of coordinate systems: the world coordinate system, the camera coordinate system, the image coordinate system, and the pixel coordinate system. These coordinate systems form an interconnected network that translates the position of objects in the real world into pixel values that computers can interpret. The world coordinate system provides a global reference framework, defining the absolute positions of objects within a scene, akin to a coordinate system on a map. The camera coordinate system uses the camera as the reference center, describing the relative position and orientation of objects for the camera. The image coordinate system resides on the imaging plane and is responsible for projecting three-dimensional world coordinates onto a two-dimensional image plane. The pixel coordinate system further converts image coordinates into pixel values that computers can directly process, and it is the most directly relevant coordinate system for computer vision processing. Through this series of coordinate transformations, precise mapping from the real world to digital images can be achieved, providing robust technical support for road distress analysis and object localization.

4.1. Camera Parameter Acquisition

According to the Chandraker M calibration method [32], the mapping relationship between the world coordinate system and the camera coordinate system is established. First, Figure 10 shows the chessboard calibration pattern. After obtaining an image of the calibration pattern, an image detection algorithm is applied to detect the pixel coordinates of each corner point. Since the world coordinate system of the calibration pattern is predefined by humans, and the dimension of each square on the calibration board is known, the physical coordinates of each corner point in the world coordinate system can be calculated. Using this information, the camera calibration is performed, yielding the mapping relationship formula Equation (7) between the world coordinate system and the camera coordinate system.
[ x c y c z c 1 ] = [ R T 0 1 ] [ x w y w z w 1 ]
where ( x w , y w , z w ) represents the world coordinate system, ( x c , y c , z c ) represents the camera coordinate system, R ( 3 × 3 ) represents the rotation vectors along the x , y , and z axes, and T ( 3 × 1 ) represents the translation vector.
Then, through the conversion from the camera coordinate system to the image coordinate system and from the image coordinate system to the pixel coordinate system, the mapping relationship from the world coordinate system to the pixel coordinate system is obtained, with the formula given as (8).
[ u v 1 ] = [ f x 0 u 0 0 f y v 0 0 0 1 ] [ R T 0 1 ] [ X w Y w Z w 1 ]
where Z c represents the vertical coordinate in the camera coordinate system, ( X w , Y w , Z w ) denotes the coordinates in the world coordinate system, ( u , v ) denotes the coordinates in the pixel coordinate system, ( u 0 , v 0 ) represents the coordinates of the camera sensor center in the pixel coordinate system, ( f x , f y ) represents the camera focal length, R denotes the rotation matrix, and T represents the translation vector.

4.2. Camera Pose Estimation

In practical applications, the direct measurement of the rotation matrix R and the translation vector T in Equation (8) presents operational inconvenience and challenges in precision. Therefore, this paper employs camera pose estimation techniques to obtain these parameters.
Specifically, based on the intrinsic and distortion parameters obtained from camera calibration in Section 4.1, we can estimate the camera pose by solving the Perspective-n-Point (PnP) problem. The PnP problem involves calculating the camera’s projection relationship given N feature points in the world coordinate system and their corresponding image points in the image coordinate system, thereby obtaining the pose of the camera or object. This paper employs the P3P algorithm, which is a PnP algorithm that requires only three pairs of matching points to complete the camera pose estimation (Figure 11). Through the P3P algorithm, we can solve for the rotation angles of the camera around the three coordinate axes, i.e., Roll (roll angle), Pitch (pitch angle), and Yaw (yaw angle). These parameters comprehensively describe the camera pose, laying the foundation for subsequent tasks of calculating the disease dimensions.
In the P3P problem, we consider three points—A, B, and C—in space, along with their known corresponding projections a , b , and c on the image plane. The coordinates of A, B, and C in the world coordinate system are known, but we need to solve for their coordinates in the camera coordinate system. According to the derivation, Formula (9) was obtained.
( 1 u ) y 2 y x 2 cos ( b , c ) y + 2 u x y cos ( a , b ) + 1 = ( 1 w ) x 2 w y 2 cos ( a , c ) x + 2 w x y cos ( a , b ) + 1
Equation (9) is a binary quadratic equation about x and y . By solving it using the elimination method, a maximum of four solutions may be obtained. Therefore, in addition to the three points, a set of matching points is needed for verification. With this, the coordinates of A, B, and C in the camera coordinate system can be determined using x and y . Hence, the 3D-2D problem transforms into a 3D-3D pose estimation problem.
3D-3D pose estimation pertains to the situation where the three-dimensional coordinates of a point in space are known in two camera coordinate systems. According to the iterative closest point (ICP) algorithm [24], if a series of points in space have their three-dimensional coordinates as C = c 1 , c n in the first camera coordinate system and as C = c 1 , c n in the second camera coordinate system, Equation (10) is derived.
i , c i = R c i + T
where R , T represent the rotation matrix and the translation vector, respectively.
Then, the linear algebraic solution is applied to Equation (10) to construct the least squares error term as in Equation (11), resulting in the final rotation matrix R and translation matrix T .
min R , t J = 1 2 i = 1 n c i ( R c i + T ) 2 2

5. Experiment

5.1. Datasets

In this paper, a highway crack disease dataset was constructed to meet the training requirements of industrial recognition algorithms. The dataset collected part of the pavement disease data from March to September 2022 through the vehicle-mounted front camera and used LabeImg to mark the diseases in each image as scattered objects and potholes, respectively.
The number of dataset divisions and image sizes are shown in Table 1:

5.2. Experimental Environment

The experimental setup consists of a Windows operating system running on an Intel® Core™ i5-8265u CPU (Santa Clara, CA, USA), equipped with 16 GB of random-access memory (RAM) and an NVIDIA GeForce RTX 3090 graphics card featuring 24 GB of display memory. The deep learning framework utilized is PyTorch 3.3. For this experiment, a dataset comprising a total of 6000 images was employed, with 5000 images designated as the training set and 1000 images allocated as the test set. To facilitate the differentiation between positive and negative samples during the training phase, this study employs a method to compute the degree of overlap between candidate regions and the ground truth.
The specific screening rules are shown in Table 2:

5.3. Analysis of Experimental Results

This paper introduces modifications to the original Faster R-CNN model, enhancing its backbone feature extractor and loss function. Initial training parameters for the enhanced algorithm are listed in Table 3. Table 4 shows the performance of various two-stage object detection models and the proposed model on the road damage dataset.
The experiment results are shown in Table 3. The first group is the baseline Faster RCNN algorithm; the second group modifies the backbone network of Faster RCNN to MobileNetV2 (MF RCNN). The third group modifies the cross-entropy loss of the Fast RCNN part to PolyLoss based on Faster RCNN (PF RCNN), and the fourth group replaces the region of Fast RCNN with smooth L1, namely Alpha IOU (AF1 RCNN). The fifth group replaces the Smooth L1 of the Fast RCNN network part with Alpha IOU (AF2 RCNN). The sixth group is the fusion of the backbone network MobileNetV2 and the loss functions Alpha IOU and PolyLoss, which is the MSLN algorithm proposed in this article.
As can be seen from Table 3, although the MAP of MSLN from RseNet50 to MobileNetV2 was reduced by 0.9%, the training time was reduced by 31.7%, and the model training parameters were reduced by 34.6%, effectively saving training time and improving training efficiency. Based on Faster RCNN, the cross-entropy loss of the Fast RCNN part was modified to PolyLoss, the MAP of PF RCNN reached 61.8%, and the model parameters only increased slightly. After replacing the Smooth L1 loss function of the Faster RCNN with the Alpha IOU loss function, the mAP trained by the model is improved by 0.8% and 0.7%, respectively. The Alpha IOU loss function enhanced model detection accuracy by incorporating inter-boundary box distance metrics. As can be seen from the last line, the detection accuracy of the improved MSLN proposed in this paper reaches 63.1%, and the training time is also greatly shortened compared with the Faster RCNN, which can effectively improve the training accuracy of the model and reduce the trainable parameters of the model.

5.4. Experimental Results of Disease Dimension Estimation

By substituting the coordinates of the upper-left and lower-right corners of the MSLN object detection box into Equation (8), we reverse-calculate the specific position of the disease in the world coordinate system. Based on this, the length and width of the disease are represented by Equation (12):
w = x w 2 x w 1 h = y w 2 y w 1
By capturing images of scattered objects and real pothole diseases in multiple different placement positions, this paper calculates their lengths and widths. To verify the accuracy of the measurements, the calculated dimensions were compared with the actual dimensions of the diseases, and the errors and error ratios in length and width were calculated and are presented in Table 5.
Table 4 illustrates the size errors of scattered objects and potholes in two different placement positions. In the experiment, we tested 10 different scattered objects and potholes. The average length error ratio of scattered objects was 24.2%, and the average width error ratio of potholes was 32.2%. The overall results showed that the errors of scattered objects were generally smaller. Furthermore, as can be roughly seen from the pattern in Table 4, larger objects or diseases generally have smaller errors, such as Disease 3, Disease 5, and Disease 6. Since road inspection mainly aims to identify significant risk sources, this work meets the actual needs of highway inspection and is significant for the detection and identification of road risk sources.

6. Results and Discussion

Three images in the test set were randomly selected for testing of different models, and the results are shown in Figure 12.
As can be seen from Figure 12a, both basic Faster RCNN and AF RCNN, which individually modified the loss function or the trunk feature extraction network, identify the white paper in the upper-left corner of the picture as a pit. However, the MSLN proposed in this paper, on the one hand, does not have false recognition and reduces the false positive rate. On the other hand, it also improves the recognition rate of pits. It can be seen in Figure 12b that the basic Faster RCNN model can identify overlapping pit and groove diseases, but the detection accuracy of small targets in the overlapping part is low. After the Smooth L1 loss function in the Faster RCNN is modified, the detection accuracy of small targets in the overlapping part can be significantly improved. Finally, the small target detection accuracy of the overlapping part of MSLN proposed in this paper reaches 89%. Figure 12c illustrates that the Faster R-CNN model tailored for scattered object detection achieves notable recognition performance; however, by further enhancing the feature extraction network and the loss function, our proposed MSLN managed to elevate the detection accuracy for scattered objects even further. To sum up, the MSLN proposed in this paper can reduce the false positive rate of detected objects on the one hand and improve the detection accuracy of objects on the other hand, thus improving the detection efficiency of the algorithm.
Figure 13 presents the dimension measurement results of two different-volume scattered objects. Figure 13a reflects the measurement of scattered object disease 1 listed in Table 4. Figure 13b depicts another scattered object. In Figure 13b, it can be observed that for scattered objects with a certain height, the dimension calculation error is relatively large. This is because the target detection box typically only provides two-dimensional position information of the object in the image, which limits the accuracy of the conversion from the image to the three-dimensional space. When a scattered object has a certain height, the two-dimensional information is converted into a three-dimensional position in the world coordinate system. The calculation error in the Z-axis is particularly significant, which directly affects the accuracy of the dimension measurement.

7. Conclusions

This paper presents a novel MSLN approach for the automated detection of road risk sources, including road surfaces, potholes, and scattered objects. The proposed method addresses the inefficiencies of traditional manual inspection methods by leveraging advancements in deep learning and computer vision techniques. The MSLN model integrates several key innovations to improve the detection efficiency and accuracy. First, MobileNetV2, a lightweight feature extraction network, was adopted to replace the conventional backbone network, significantly reducing the training time and hardware requirements. Second, the loss functions were optimized by introducing PolyLoss and Alpha IoU loss, improving the model’s ability to detect objects accurately. Furthermore, a mapping relationship model was established to convert the detected objects from the pixel coordinate system to the world coordinate system, enabling the estimation of the dimensions of road risk sources. Experimental results demonstrate the effectiveness of the proposed MSLN model. Compared with the baseline Faster R-CNN model, the MSLN achieved a notable improvement in both training speed and detection accuracy, with a mean average precision (MAP) of 63.1% and a reduced training time of 31.7%. The mapping relationship model also showed promising results in estimating the dimensions of scattered objects and potholes.
The method proposed in this paper has demonstrated its value over an application journey of more than 20,000 km, but at the same time, we have also discovered its limitations in the actual environment. Especially under conditions of insufficient ambient light and backlight, the recognition was significantly affected. Future research will focus on enhancing the MSLN model in several directions. First, backbone fusion networks that combine lightweight and powerful networks will be explored to improve detection accuracy without sacrificing efficiency. Second, we will develop recognition algorithms for road conditions under low-light and backlight conditions to enhance their identification accuracy. Third, depth information from depth cameras or stereo vision will be incorporated to enhance dimension estimation, particularly for objects with height. Real-time detection performance and ease of deployment will be optimized through wider testing and efficient data processing pipelines. Multimodal data fusion, combining image data with LiDAR and radar, will be investigated to improve robustness in complex environments. Finally, the model’s generalizability to other road infrastructures, such as bridges and tunnels, will be assessed to broaden its application scope. These efforts aim to refine and extend the MSLN model for widespread adoption in intelligent transportation systems.

Author Contributions

R.P. and J.N.: conceptualization; R.P. and P.Z.: methodology; R.P. and P.Z.: software; J.W. and J.L.: validation; R.P. and Y.Y.: formal analysis; R.P.: investigation; J.N. and P.Z.: resources; R.P.: data curation; R.P. and Y.Y.: writing—original draft preparation; Y.Y. and J.L.: writing—review and editing; J.N.: visualization; Y.Y. and P.Z.: supervision; J.N. and J.W.: funding acquisition. All authors have read and agreed to the published version of the manuscript.

Funding

Supported by the National Natural Science Foundation of China (No. 61976247); Chongqing Technology Innovation and Application Development Special (No. CSTB2022TIAD-GPX0029) and Chongqing Meritocratic Plan (cstc2021ycjh-bgzxm0264).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Due to privacy restrictions, some data are not publicly available but can be made available from the authors upon reasonable request and with permission from the China Merchants Chongqing Road Engineering Inspection Center Co., Ltd.

Acknowledgments

The authors would like to thank the Chongqing Transportation Planning and Technology Development Center for funding this research. We also express our gratitude to the China Merchants Chongqing Road Engineering Inspection Center Co., Ltd., and the National Mountainous Highway Engineering Technology Research Center for providing the necessary facilities and technical support.

Conflicts of Interest

Author Rong Pang was employed by the company China Merchants Chongqing Road Engineering Inspection Center Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Zhou, B.; Pei, J.; Zhang, J.; Wang, C. Joint design and load transfer capacity analysis of photovoltaic/thermal integrated pavement unit. J. Clean. Prod. 2022, 380, 135029. [Google Scholar] [CrossRef]
  2. The Comprehensive Department of the Ministry of Transport of the People’s Republic of China. 2023 Statistical Bulletin of Transportation Industry Development; Ministry of Transport of the People’s Republic of China: Beijing, China, 2024.
  3. Hsieh, Y.A.; Tsai, Y.J. Machine learning for crack detection: Review and model performance comparison. J. Comput. Civ. Eng. 2020, 34, 04020038. [Google Scholar] [CrossRef]
  4. Oh, H.; Garrick, N.W.; Achenie, L. Segmentation Algorithm Using Iterative Clipping for Processing Noisy Pavement Images. In Imaging Technologies; American Society of Civil Engineers: Reston, VA, USA, 1998. [Google Scholar]
  5. Li, Y.; Huang, Q.; Hou, Z. Research on Improved OSTU Algorithm Based on Hough Transform for Pavement Crack Detection. Electron. Des. Eng. 2016, 24, 4. [Google Scholar]
  6. Sun, L.; Xing, J.; Xie, L.; Wang, J. Research on Crack Detection Method based on Adaptive Threshold Canny Algorithm. Microcomput. Appl. 2017, 36, 4. [Google Scholar]
  7. Zhang, J.; Kang, X.; Ni, H.; Ren, F. Surface defect detection of steel strips based on classification priority YOLOv3-dense network. Ironmak. Steelmak. 2021, 48, 547–558. [Google Scholar] [CrossRef]
  8. Li, S.; Gu, X.; Xu, X.; Xu, D.; Zhang, T.; Liu, Z.; Dong, Q. Detection of concealed cracks from ground penetrating radar images based on deep learning algorithm. Constr. Build. Mater. 2021, 273, 121949. [Google Scholar] [CrossRef]
  9. Fei, Y.; Wang, K.C.P.; Zhang, A.; Chen, C.; Li, J.Q.; Liu, Y.; Yang, G.; Li, B. Pixel-level cracking detection on 3D asphalt pavement images through deep-learning-based CrackNet-V. IEEE Trans. Intell. Transp. Syst. 2019, 21, 273–284. [Google Scholar] [CrossRef]
  10. Liu, Y.; Yao, J.; Lu, X.; Xie, R.; Li, L. DeepCrack: A deep hierarchical feature learning architecture for crack segmentation. Neurocomputing 2019, 338, 139–153. [Google Scholar] [CrossRef]
  11. Meng, D.; Cao, G.; Duan, Y.; Zhu, M.; Tu, L.; Xu, D.; Xu, J. Tongue images classification based on constrained high dispersal network. Evid.-Based Complement. Altern. Med. 2017, 2017, 7452427. [Google Scholar] [CrossRef]
  12. Shelhamer, E.; Long, J.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  13. Zou, Q.; Zhang, Z.; Li, Q.; Qi, X.; Wang, Q.; Wang, S. Deepcrack: Learning hierarchical convolutional features for crack detection. IEEE Trans. Image Process. 2018, 28, 1498–1512. [Google Scholar] [CrossRef] [PubMed]
  14. Subirats, P.; Dumoulin, J.; Legeay, V.; Barba, D. Automation of Pavement Surface Crack Detection using the Continuous Wavelet Transform. In Proceedings of the 2006 IEEE International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006. [Google Scholar]
  15. Hou, Y.; Zhang, H.; Gao, Z.; Wang, D.; Liu, P.; Oeser, M.; Wang, L.; Chen, N. Research and comparison of intelligent pavement disease detection methods based on data depth enhancement. J. Beijing Univ. Technol. 2022, 48, 13. [Google Scholar]
  16. Huang, F.H.; Cao, Y.S.; Jiang, Y.S. Progress of research application of machine learning in bridge disease detection and identification. Road Traffic Sci. Technol. Appl. Technol. Ed. 2019, 114–116. [Google Scholar]
  17. Cha, Y.J.; Choi, W.; Suh, G.; Mahmoudkhani, S.; Büyüköztürk, O. Autonomous structural visual inspection using region-based deep learning for detecting multiple damage types. Comput.-Aided Civ. Infrastruct. Eng. 2018, 33, 731–747. [Google Scholar] [CrossRef]
  18. Cha, Y.J.; Choi, W.; Büyüköztürk, O. Deep Learning-Based Crack Damage Detection Using Convolutional Neural Networks. Comput.-Aided Civ. Infrastruct. Eng. 2017, 32, 361–378. [Google Scholar] [CrossRef]
  19. Xu, Z.; Guan, H.; Kang, J.; Lei, X.; Ma, L.; Yu, Y.; Chen, Y.; Li, J. Pavement crack detection from CCD images with a locally enhanced transformer network. Int. J. Appl. Earth Obs. Geoinf. 2022, 110, 102825. [Google Scholar] [CrossRef]
  20. Nguyen, T.S.; Begot, S.; Duculty, F.; Avila, M. Free-form anisotropy: A new method for crack detection on pavement surface images. In Proceedings of the 2011 18th IEEE International Conference on Image Processing, Brussels, Belgium, 11–14 September 2011. [Google Scholar]
  21. Gopalakrishnan, K.; Khaitan, S.K.; Choudhary, A.; Agrawal, A. Deep Convolutional Neural Networks with transfer learning for computer vision-based data-driven pavement distress detection. Constr. Build. Mater. 2017, 157, 322–330. [Google Scholar] [CrossRef]
  22. Chen, F.C.; Jahanshahi, R. NB-CNN: Convolutional Neural Network and Naive Bayes Data Fusion Using Deep Learn-based Crack Detection. IEEE Trans. Ind. Electron. 2018, 65, 4392–4400. [Google Scholar] [CrossRef]
  23. Carr, T.A.; Jenkins, M.D.; Iglesias, M.I.; Buggy, T.; Morison, G. Road crack detection using a single stage detector based deep neural network. In Proceedings of the 2018 IEEE Workshop on Environmental, Energy, and Structural Monitoring Systems (EESMS), Salerno, Italy, 21–22 June 2020. [Google Scholar]
  24. Pang, R.; Yang, Y.; Huang, A.; Liu, Y.; Zhang, P.; Tang, G. Multi-scale feature fusion model for bridge appearance defects detection. Big Data Min. Anal. 2024, 7, 1–11. [Google Scholar] [CrossRef]
  25. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 42, 2999–3007. [Google Scholar]
  26. Liu, R.; Huang, M.; Wang, L.; Bi, C.; Tao, Y. PDT-YOLO: A Roadside Object-Detection Algorithm for Multiscale and Occluded Targets. Sensors 2024, 24, 2302. [Google Scholar] [CrossRef] [PubMed]
  27. Yuan, B.; Sun, Z.; Pei, L.; Li, W.; Ding, M.; Hao, X. Super-Resolution Reconstruction Method of Pavement Crack Images Based on an Improved Generative Adversarial Network. Sensors 2022, 22, 9092. [Google Scholar] [CrossRef] [PubMed]
  28. Rehana, K.C.; Remya, G. Road Damage Detection and Classification Using YOLOv5. In Proceedings of the 2022 Third International Conference on Intelligent Computing Instrumentation and Control Technologies (ICICICT), Kannur, India, 11–12 August 2022; pp. 489–494. [Google Scholar]
  29. Kang, D.; Benipal, S.S.; Gopal, D.L.; Cha, Y.J. Hybrid pixel-level concrete crack segmentation and quantification across complex backgrounds using deep learning. Autom. Constr. 2020, 118, 103291. [Google Scholar] [CrossRef]
  30. He, J.; Erfani, S.; Ma, X.; Bailey, J.; Chi, Y.; Hua, X.S. Alpha-IoU: A Family of Power Intersection over Union Losses for Bounding Box Regression. Adv. Neural Inf. Process. Syst. 2021, 34, 20230–20242. [Google Scholar] [CrossRef]
  31. Leng, Z.; Tan, M.; Liu, C.; Cubuk, E.D.; Shi, X.; Cheng, S.; Anguelov, D. PolyLoss: A Polynomial Expansion Perspective of Classification Loss Functions. arXiv arXiv:2204.12511, 2022. [CrossRef]
  32. Chandraker, M.; Song, S.; Lin, Y.; Wang, X. Moving Object Localization in 3D Using a Single Camera. US2014270484A1, 30 June 2015. [Google Scholar]
Figure 1. Image grayscale processing: (a) the original image; (b) road grayscale diagram.
Figure 1. Image grayscale processing: (a) the original image; (b) road grayscale diagram.
Sensors 24 05577 g001
Figure 2. Image processed by Gaussian filtering; (a) road grayscale diagram; (b) filter processing diagram.
Figure 2. Image processed by Gaussian filtering; (a) road grayscale diagram; (b) filter processing diagram.
Sensors 24 05577 g002
Figure 3. Image processed after equalization: (a) filter processing diagram; (b) equalization processing diagram.
Figure 3. Image processed after equalization: (a) filter processing diagram; (b) equalization processing diagram.
Sensors 24 05577 g003
Figure 4. Data augmentation: (a) the original image; (b) random cutting; (c) translation; (d) random rotation; (e) horizontal reversal; (f) vertical flip.
Figure 4. Data augmentation: (a) the original image; (b) random cutting; (c) translation; (d) random rotation; (e) horizontal reversal; (f) vertical flip.
Sensors 24 05577 g004
Figure 5. Traditional convolution, * denotes matrix multiplication.
Figure 5. Traditional convolution, * denotes matrix multiplication.
Sensors 24 05577 g005
Figure 6. Depth-separable convolution, * denotes matrix multiplication.
Figure 6. Depth-separable convolution, * denotes matrix multiplication.
Sensors 24 05577 g006
Figure 7. Inverted residual structure, * denotes matrix multiplication.
Figure 7. Inverted residual structure, * denotes matrix multiplication.
Sensors 24 05577 g007
Figure 8. Smooth L1 calculation method.
Figure 8. Smooth L1 calculation method.
Sensors 24 05577 g008
Figure 9. Alpha-IOU calculation method.
Figure 9. Alpha-IOU calculation method.
Sensors 24 05577 g009
Figure 10. Checkerboard calibration pattern.
Figure 10. Checkerboard calibration pattern.
Sensors 24 05577 g010
Figure 11. Camera pose illustration.
Figure 11. Camera pose illustration.
Sensors 24 05577 g011
Figure 12. Comparison of experimental results.
Figure 12. Comparison of experimental results.
Sensors 24 05577 g012
Figure 13. Scattered object dimension test.
Figure 13. Scattered object dimension test.
Sensors 24 05577 g013
Table 1. Dataset division and image size.
Table 1. Dataset division and image size.
DatasetNumberImage Size
Training set50001080 × 720
Validation set10001080 × 720
Test set10001080 × 720
Table 2. Filtering rules.
Table 2. Filtering rules.
IOUSample ClassificationTag
>0.7Positive samples1
[0.3, 0.7]Negative samples−1
<0.3Abandon the sample0
Table 3. Model initialization parameters.
Table 3. Model initialization parameters.
The Size of the BatchBasis
Vector
Vector
Attenuation Way
The Weight
Damping Item
Learning Rate ParameterTraining
Number of Rounds
120.01Step0.00050.1200
Table 4. Details of MSLN and Faster RCNN experimental results.
Table 4. Details of MSLN and Faster RCNN experimental results.
AlgorithmBackboneLoss FunctionTraining Time/minMAPModel ParametersFPS
Faster RCNNRseNet50L1 Loss + cross-entropy loss148.261.2%8238500624
MF RCNNMobileNetV2L1 Loss + cross-entropy loss105.361.1%6135228127
PF RCNNRseNet50PolyLoss135.161.8%6274653427
AF1 RCNNRseNet50Alpha IOU13562.0%6789465726
AF2 RCNNRseNet50Alpha IOU134.961.9%6845653626
MSLN(ours)MobileNetV2PolyLoss + Alpha IOU101.263.1%5384596832
Table 5. Errors in disease dimension calculation.
Table 5. Errors in disease dimension calculation.
TypeNumberTrue Length (cm)True Width (cm)Estimated Length (cm)Estimated Width (cm)Length ErrorWidth ErrorAverage Error
Scattered objectsDisease 1 (pos1)111116135234.1%
Disease 1 (pos2)111171544
Disease 2 (pos1)18720132539.1%
Disease 2 (pos2)187211234
Disease 3 (pos1)41153015909.6%
Disease 3 (pos2)4115371641
Disease 4 (pos1)71810213325.4%
Disease 4 (pos2)71882315
Disease 5 (pos1)533060357512.7%
Disease 5 (pos2)5330592644
PotholeDisease 6 (pos1)4141573716420.7%
Disease 6 (pos2)41412714140
Disease 7 (pos1)1962596341.0%
Disease 7 (pos2)196221034
Disease 8 (pos1)18713105325.4%
Disease 8 (pos2)18715631
Disease 9 (pos1)1261092339.6%
Disease 9 (pos2)126111115
Disease 10 (pos1)841153134.3%
Disease 10 (pos2)8412341
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Pang, R.; Ning, J.; Yang, Y.; Zhang, P.; Wang, J.; Liu, J. Detection of Road Risk Sources Based on Multi-Scale Lightweight Networks. Sensors 2024, 24, 5577. https://doi.org/10.3390/s24175577

AMA Style

Pang R, Ning J, Yang Y, Zhang P, Wang J, Liu J. Detection of Road Risk Sources Based on Multi-Scale Lightweight Networks. Sensors. 2024; 24(17):5577. https://doi.org/10.3390/s24175577

Chicago/Turabian Style

Pang, Rong, Jiacheng Ning, Yan Yang, Peng Zhang, Jilong Wang, and Jingxiao Liu. 2024. "Detection of Road Risk Sources Based on Multi-Scale Lightweight Networks" Sensors 24, no. 17: 5577. https://doi.org/10.3390/s24175577

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop