Next Article in Journal
User Evaluation of a Shared Robot Control System Combining BCI and Eye Tracking in a Portable Augmented Reality User Interface
Previous Article in Journal
A Fast and High-Accuracy Foreign Object Detection Method for Belt Conveyor Coal Flow Images with Target Occlusion
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8

1
Key Laboratory of Opto-Electronic Technology and Intelligent Control, Ministry of Education, Lanzhou Jiaotong University, Lanzhou 730070, China
2
National and Provincial Joint Engineering Laboratory of Road & Bridge Disaster Prevention and Control, Lanzhou Jiaotong University, Lanzhou 730070, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(16), 5252; https://doi.org/10.3390/s24165252
Submission received: 12 July 2024 / Revised: 9 August 2024 / Accepted: 12 August 2024 / Published: 14 August 2024
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Concrete surface crack detection is a critical research area for ensuring the safety of infrastructure, such as bridges, tunnels and nuclear power plants, and facilitating timely structural damage repair. Addressing issues in existing methods, such as high cost, lengthy processing times, low efficiency, poor effectiveness and difficulty in application on mobile terminals, this paper proposes an improved lightweight concrete surface crack detection algorithm, YOLOv8-Crack Detection (YOLOv8-CD), based on an improved YOLOv8. The algorithm integrates the strengths of visual attention networks (VANs) and Large Convolutional Attention (LCA) modules, introducing a Large Separable Kernel Attention (LSKA) module for extracting concrete surface crack and local feature information, adapted for features such as fracture susceptibility, large spans and slender shapes, thereby effectively emphasizing crack shapes. The Ghost module in the YOLOv8 backbone efficiently extracts essential information from original features at a minimal cost, enhancing feature extraction capability. Moreover, replacing the original convolution structure with GSConv in the neck network and employing the VoV-GSCSP module adapted for the YOLOv8 framework reduces floating-point operations during feature channel fusion, thereby lowering computational complexity whilst maintaining model accuracy. Experimental results on the RDD2022 and Wall Crack datasets demonstrate the improved algorithm increases in mAP50 by 15.2% and 12.3%, respectively, and in mAP50-95 by 22.7% and 17.2%, respectively, whilst achieving a reduced model computational load of only 7.9 × 109, a decrease of 3.6%. The algorithm achieves a detection speed of 88 FPS, enabling real-time and accurate detection of concrete surface crack targets. Comparison with other mainstream object detection algorithms validates the effectiveness and superiority of the proposed approach.

1. Introduction

Since the 1990s, a large number of social and civil infrastructure, such as bridges, tunnels and nuclear power plants, have been constructed throughout China. Owing to its low cost, versatility and extensive use, concrete has become one of the most widely used materials in various types of infrastructure. However, concrete structures inevitably face crack problems caused by creep, shrinkage and loads, which may compromise their safety [1]. Moreover, as buildings age, they are increasingly affected by factors such as material aging, long-term loads, frequent and sudden natural disasters and human-induced damage, leading to varying degrees of crack damage in concrete structures. If not repaired promptly, these cracks may pose marked safety risks to concrete infrastructure and potentially threaten human life. Therefore, timely detection of concrete surface cracks has profound significance for preventing infrastructure damage and maintaining its safety.
Concrete crack detection methods primarily include traditional and deep learning methods. Traditional methods are based on crack detection using traditional digital image processing. The detection procedure based on image processing generally includes three stages: firstly, pre-processing crack images through denoising and filtering; secondly, binarization and morphological processing by setting thresholds; and thirdly, classification detection using classifiers. Abdel-Qader et al. [2] analyzed crack edge information using the Fourier and Hough transform algorithms and segmented cracks using the Canny operator to extract crack edge information. Salman et al. [3] designed an automatic crack digital image classification method using Gabor filters. Zhou et al. [4] proposed an automatic method for detecting cracks using frequency domain filtering and contour analysis of 3D laser range data. Vivekananthan et al. [5] combined grayscale discrimination and the Otsu method to successfully detect target cracks in different images. Zhu et al. [6] proposed a crack detection framework based on two-dimensional digital image correlation and a displacement-based robust crack detection method for concrete cracking phenomena, which evaluated their fracture performance.
In recent years, with the rapid development of deep learning technology, crack detection algorithms based on object detection techniques have become mainstream. Object detection techniques are classified into two- and single-stage algorithms. The current mainstream two-stage algorithms include R-CNN, Fast R-CNN and U-Net, to name a few. Rosso et al. [7] developed an artificial intelligence (AI)-based automatic road tunnel defects hierarchical classification framework to improve the efficiency of the indirect surveying method. Shahin et al. [8] constructed a new concrete crack detection model by combining ViT models with multiple image enhancement detectors. Chun et al. [9] proposed a method combining fully convolutional neural networks and semi-supervised learning to segment pavement crack images, but this method produced rough results that were prone to false detections. Ghosh et al. [10] improved the convolutional layer structure of U-Net using ResNet residual blocks, validating their performance on public datasets and achieving excellent results. Kang et al. [11] used an ensemble approach, incorporating the Faster R-CNN algorithm for faster regional convolutional neural networks to detect crack areas. Meng et al. [12] classified cracks, coarse segmentation and fine segmentation and extracted maximum widths using lightweight classification algorithms, lightweight segmentation algorithms, high-precision segmentation algorithms and crack width measurement algorithms to achieve an automatic real-time crack detection method based on drones. Another category is single-stage detection methods, which directly separate specific categories and regress boundaries. The single-stage object detection algorithms primarily include the You Only Look Once (YOLO) series and the Single Shot MultiBox Detector (SSD). Compared with single-stage algorithms, two-stage algorithms have certain advantages in detection accuracy, but their detection speed is much slower than single-stage algorithms. Therefore, single-stage algorithms are extra concerned with concrete surface crack detection. Chen et al. [13] introduced a deep learning framework, combining convolutional neural networks (CNNs) with naive Bayes‘ decision algorithms to detect boundary box cracks on underwater nuclear power station surfaces. Deng et al. [14] applied YOLOv2 for automatic crack detection in concrete, evaluating the robustness of the trained detector against interference by handwritten notes and finding that YOLOv2 can automatically locate a cracked bounding box from original images, even under handwriting interference. Liu et al. [15] proposed a bridge crack detection algorithm, R-YOLO v5, largely improving crack detection accuracy by incorporating attention mechanisms and optimizing loss functions. Wu et al. [16] proposed an improved YOLOv4 network using pruning techniques and EvoNorm-S0 structures, which better identified concrete cracks with several misleading targets. They found that whilst maintaining high accuracy, the model could correctly classify cracks at a faster computational speed. Ye et al. [17] proposed an improved YOLOv7 network and enhanced three different self-developed modules to better identify concrete cracks from several misleading targets, demonstrating effective detection of cracks of different sizes and robustness in validating images contaminated with various types and intensities of noise. Jiang et al. [18] optimized the YOLO-v3 and SSD algorithms using depth-separable convolution, inverse residual networks and linear bottleneck structures, finding that optimized detection accuracy increased by 3.25% and 4.04%, respectively.
As one of the representatives of single-stage object detection algorithms, the YOLO algorithm has received widespread attention from scholars since its inception. In recent years, with continuous optimization updates, the YOLO algorithm has achieved increasingly excellent results in object detection. In 2023, the Ultralytics team proposed YOLOv8, which not only boasts of high-precision recognition but also excellent real-time performance. The team further lightweighted the entire framework, rendering it extra suitable for concrete surface crack detection. Therefore, based on YOLOv8n, this paper optimizes the model to improve the accuracy of concrete surface crack detection, proposing a YOLOv8-CD algorithm for concrete surface crack detection. Firstly, by introducing the Large Separable Kernel Attention module [19], other abundant and fine characteristic information about cracks on concrete surfaces is captured, providing strong support for accurate crack identification. Secondly, to further enhance the feature extraction capabilities, the Ghost module [20] is integrated into the backbone network of YOLOv8, improving detection accuracy and optimizing overall model performance. Finally, in the neck network of the model, the GSConv is used to replace traditional convolutional structures and integrate the VoV-GSCSP module [21]. This method not only ensures model accuracy but also effectively reduces floating-point operations during feature channel fusion, further reducing model parameters and enhancing computational efficiency. Thus, the proposed algorithm in this paper can further strengthen the application of deep learning algorithms in concrete crack detection technology, offering ideas for deploying related detection algorithms to embedded rapid detection devices, improving the accuracy and efficiency of crack detection and providing technical support for the timely elimination of safety risks in concrete infrastructure.

2. Algorithms

2.1. The YOLOv8 Algorithm

The You Only Look Once (YOLO) algorithm was initially proposed by Joseph Redmon et al. in 2016. Unlike traditional object detection methods, YOLO adopts a novel approach by transforming the object detection task into a regression problem of a single neural network. It divides the image into multiple grids, predicts bounding boxes and object categories within each grid and utilizes the Non-Maximum Suppression (NMS) algorithm to eliminate overlapping bounding boxes, thus achieving real-time object detection.
The YOLO series algorithms represent notable advancements in the field of object detection. YOLOv1 [22] was the first single-stage object detection algorithm, achieving end-to-end prediction. YOLOv2 [23] introduced the Darknet-19 network, profoundly enhancing detection accuracy and speed. YOLOv3 [24] employed the deeper Darknet-53 network. YOLOv4 [25], proposed by Alexey Bochkovskiy et al., used the CSPDarknet53 network to improve performance. YOLOv5, introduced by Glenn Jocher, achieved remarkable results in many object detection tasks. YOLOv6 [26], developed by Meituan, focused on industrial applications. YOLOv7 [27], proposed by the YOLOv4 team in 2022, transformed object detection into a regression problem through convolutional neural networks and fully connected layers for prediction.
The YOLOv8 algorithm used in this paper was introduced by Ultralytics in 2023. It adopts a lightweight network structure whilst maintaining the efficiency and ease of use typical of the YOLO series, further enhancing algorithm performance and flexibility for tasks such as image classification, object detection and instance segmentation, supporting both CPU and GPU platforms. YOLOv8 introduces several innovations, including a new backbone network, Anchor-Free detection head and loss functions and deformable convolution DCNv3, significantly boosting performance. Additionally, the algorithm employs dual-path prediction and densely connected convolution networks, decomposing object detection into classification and localization subtasks. Then, it applies cascading and pyramid concepts to detect objects of different sizes.
Figure 1 is the network structure of YOLOv8. The backbone consists of C2f and SPPF modules with the CSP concept, whilst the neck network of the backbone adopts the ELAN structure design principles from YOLOv7, replacing the C3 module of YOLOv5 with the more gradient flow of the C2f structure. As shown in Figure 2, the C2f structure enhances the feature fusion capability of convolutional neural networks, improves inference speed and achieves further lightweighting. Moreover, YOLOv8 uses the SPPF module from YOLOv5 architecture, as shown in Figure 3, which is a spatial pyramid pooling layer that expands the receptive field, facilitates local and global feature fusion and enriches feature information. The head section undergoes remarkable changes compared with YOLOv5, adopting the mainstream decoupled head structure and separating classification and detection heads. YOLOv8 departs from Anchor-Base methods to embrace Anchor-Free principles; on loss functions, it employs the BCE Loss for classification and the DFL Loss + CIOU Loss for regression. In terms of sample matching, YOLOv8 uses Task-Aligned Assigner instead of IOU matching or single-sided proportional allocation.

2.2. Problems with YOLOv8

When deploying the YOLO algorithm to mobile devices, the limited performance of mobile devices needs consideration. At the same time, the requirement for real-time detection of concrete surface cracks is high. Accordingly, this paper selects YOLOv8n, the minimum weight model algorithm of YOLOv8. However, in the actual detection process, YOLOv8n also has certain defects. Firstly, the YOLOv8n algorithm uses numerous standard convolutions and C2f modules, which improves the accuracy of the algorithm but reduces the running speed and increases the parameters of the model. Secondly, the detection scene changes quickly in mobile detection, and ensuring sufficient detection accuracy is difficult. Therefore, the YOLOv8n algorithm is not ideal for the detection and processing of concrete surface cracks, as it is prone to problems such as false detection and missing detection.

3. Improvements to YOLOv8

To improve the accuracy of concrete crack detection, this paper proposes an improved detection network model, called YOLOv8-CD, based on YOLOv8n, as shown in Figure 4. Prior to the SPPF module in the YOLOv8n backbone network, the LSKA module, incorporating visual attention networks and a large convolutional attention, was embedded to accommodate features such as brittle cracking, large span and the elongation of crack objects. Additionally, the Ghost module was employed within the backbone network to extract necessary information from original features at a lower cost, thereby enhancing feature extraction capability. In the neck network, GSConv replaced the original convolution structure, and the VoV-GSCSP module was utilized to reduce floating-point operations during feature channel fusion and decrease model parameters whilst ensuring accuracy.
Additionally, this paper conducts detailed hyperparameter tuning for the YOLOv8-CD algorithm. Specific adjustments include determining an appropriate learning rate to ensure stable model convergence; selecting a suitable batch size to effectively utilize GPU memory and maintain training stability; adjusting the number of network layers and structure, based on the characteristics of the dataset, to fully extract and represent crack features; optimizing the size and ratio of anchor boxes to improve detection accuracy and recall rate; and adjusting L2 regularization parameters to prevent overfitting. These hyperparameter adjustments significantly improved the model’s detection accuracy and generalization capability on the dataset.

3.1. The LSKA Module

Owing to the complex and variable environments where concrete surface cracks occur, enhancing the model’s ability to represent crack damage features is crucial. In the YOLOv8 backbone network, an attention mechanism was integrated to locally enhance features, allowing the network to ignore irrelevant information interference and incorporate other valuable information into the fused feature maps. Traditional attention mechanisms, such as Squeeze and Excitation (SE) and the Convolutional Block Attention Module (CBAM), have notable limitations. SE attention focuses solely on inter-channel dependencies, consequently neglecting spatial features. Meanwhile, the CBAM introduces large-scale convolution kernels to extract spatial features but overlooks long-range dependency issues.
Large Separable Kernel Attention (LSKA) is an improved module for the application of the traditional Large Kernel Attention (LKA) module in visual attention networks. Traditional LKA modules have the disadvantages of high computational complexity and memory requirements when handling large convolution kernels. To mitigate these problems, LSKA employs oversized convolution kernels in the attention modules of VAN, as shown in Figure 5. It decomposes the two-dimensional convolution kernels of deep convolution layers into stacked horizontal and vertical one-dimensional kernels, achieving the goal of reducing computational complexity and memory usage. This improvement allows the model to maintain high performance whilst alleviating computational and memory burdens. In this paper, introducing LSKA further enhanced the detection accuracy and efficiency of the model. By optimizing network structure and attention mechanisms, the model better captured critical features in images, thereby improving the accuracy of object detection. Figure 6 shows the block design structure of LSKA within the visual attention network.

3.2. Ghost Module

The conventional method for feature extraction involves convolving all channels of input feature maps with multiple convolution kernels. However, in deep networks, stacking numerous convolutional layers requires a substantial number of parameters and computations, resulting in many redundant feature maps. Therefore, several studies have proposed model compression methods, such as pruning, quantization and knowledge distillation [28,29,30]. Some problems, such as complicated model design and difficult training, exist despite their effectiveness in reducing parameters. Other approaches focus on optimizing network structures, such as MobileNet and ShuffleNet [31,32]. However, a 1 × 1 convolutional layer still consumes considerable memory and Giga floating-point operations per second (GFLOPs). Hence, acquiring essential feature maps is needed without complete convolutional operations. The Ghost module (in Figure 7) uses standard convolutions to obtain partial feature maps, linear operations to generate additional feature maps and, finally, concatenates these two sets of feature maps to a specified dimension, achieving other feature mappings with fewer parameters and computations.
The operation of generating n feature maps in any convolutional layer can be defined as follows:
Y = X × f + b
where X R c × h × w is the input, c is the number of input channels and h and w are the height and width of the input feature map. Y R h × w × n is the output feature map with n channels, h and w are the height and width of the output feature map, f R c × k × k × n is the convolutional filter of this layer with a kernel size of k · k and b is the bias term. The computation result of this convolution is as follows:
n · h · w · c · k · k
For the Ghost module, the first step is to obtain the intrinsic feature maps ( Y ) using conventional convolution, with a computational complexity of the following:
Y = X × f + b
Then, each channel of feature maps Y is processed using operation φ i j to generate the Ghost feature maps Y i j :
y i j = φ i , j ( y i ) ,   i = 1 , , m ,   j = 1 , , s
Finally, the intrinsic feature maps obtained in the first step are concatenated (identity concatenate) with the Ghost feature maps obtained in the second step to obtain the final result.
If the input tensor is c · h · w (input channels, feature map height and width), after one convolution, the output tensor is n · h · w (output channels, feature map height and width). If the conventional convolution kernel size is k and the linear transformation convolution kernel size is d , after s times transformations, the computational complexity comparison between regular convolution and Ghost convolution operations is as follows:
r s = n · h · w · c · k · k n s · h · w · c · k · k + ( s 1 ) · n s · h · w · d · d = c · k · k 1 s · c · k · k + s 1 s · d · d s
where n / s is the number of output channels after the first transformation; owing to the identity mapping, s 1 does not need to be calculated, but it is considered part of the second transformation. Therefore, the computational complexity of the Ghost module is approximately 1 / s times that of regular convolution.
Taking advantage of the Ghost module, a Ghost Bottleneck is achieved specifically for CNNs. As shown in Figure 8, the Ghost Bottleneck resembles the basic residual blocks in ResNet [33], incorporating multiple convolutional layers.
This study enhances feature extraction capability by leveraging the Ghost module within the backbone network, extracting the necessary information from raw features at minimal cost. It simultaneously reduces the overall model size, thereby ensuring lightweight design without compromising accuracy.

3.3. GSConv and VoV-GSCSP Modules

Real-time performance is crucial for object detection, as it requires accurate identification of objects in images or videos within extremely short timeframes, providing relevant information such as their position and size. This scenario demands algorithms not only with efficient processing capabilities but also with rational optimization and configuration of hardware devices. However, lightweight models constructed from numerous depthwise separable convolution layers often fail to achieve sufficient accuracy. Therefore, this paper introduces a novel approach, GSConv, which maintains adequate accuracy whilst reducing model complexity.
The GSConv is introduced to solve the speed problem of predictive computing in Convolutional Neural Networks (CNNs). In the backbone network of CNNs, input images typically undergo a gradual spatial-to-channel transformation process. Each spatial compression and channel expansion of feature maps leads to a partial loss of semantic information. Channel-dense convolution (SC) maximally preserves implicit connections between each channel, whereas depthwise separable convolution (DSC) completely severs these connections. GSConv aims to maintain these connections as much as possible whilst reducing time complexity. Additionally, GSConv is used to design GS Bottleneck, as shown in Figure 9, which enhances feature non-linear expression and information reuse based on GSConv. As shown in Figure 10, the VoV-GSCSP module is a cross-stage partial network module designed by using a one-time aggregation method, which is employed to carry out effective information fusion between feature graphs of different stages.
As seen in Figure 11, GSConv begins by performing downsampling with regular convolution on its input. Subsequently, it applies DWConv (depthwise convolution), concatenates the results of both convolutions and, finally, performs a shuffle operation to align corresponding channel numbers from the previous two convolutions together.
As shown in Figure 12, the new YOLOv8 neck and head networks replace Conv and C2f modules with GSConv and VoV-GSCSP modules, granting the model sufficient feature information to understand input data whilst reducing the model’s parameter number, achieving the purpose of lightweighting.

4. Dataset

This study used two datasets to train the model, including the RDD2022 dataset and the Wall Crack dataset.
RDD2022: Released by the University of Tokyo, it includes 47,420 road surface images collected from six countries: Japan, India, the Czech Republic, Norway, the United States and China. The dataset captures four types of concrete pavement crack damages, namely, D00 (longitudinal cracks), D10 (transverse cracks), D20 (mesh cracks) and D40 (potholes). Owing to the excessively large image sizes from Norway in comparison to other countries, they are deemed unsuitable for training. Therefore, in this paper, 16,648 photos from countries other than Norway are selected as the training set in the RDD2022 dataset.
Wall Crack: This dataset consists of 5882 concrete wall images collected from the Internet and literature, from which 6 different concrete surface damages can be captured, namely, D0 (exposed steel bars), D1 (weathering), D2 (cracks), D3 (delamination), D4 (spalling) and D5 (rust).
The RDD2022 dataset and the Wall Crack dataset are both used for detecting damage in concrete structures through image recognition. The RDD2022 dataset includes a large number of road crack images from different countries. The selection of the training set ensures a relatively balanced quantity of each type of damage and images from different countries. This allows the trained model to adequately represent different scenarios and types of damage, demonstrating excellent balance. The Wall Crack dataset consists of concrete wall images collected from the Internet and literature, capturing six different types of concrete surface damage. Although the number of images is relatively small, each type of damage is well-represented in the dataset, with a fairly balanced sample distribution. This ensures that the model can learn and recognize each type of damage fairly during training. Overall, these two datasets perform well in providing data support for concrete damage detection and also show excellent balance in data distribution.

5. Experiments and Results

5.1. Experimental Equipment and Evaluation Index

The primary development tool for this model is Python, with the open-source deep learning framework PyTorch as the network framework, accelerated by CUDA 11.8 for training. The hardware testing environment for this model includes an Intel® Xeon® Platinum 8375C CPU @ 2.90 GHz for the CPU and an NVIDIA RTX 4090 GPU with 24 GB of VRAM.
During training, the input images were set to 640 × 640, and SGD was used as the optimization function to train the model. The training consisted of 300 epochs with a batch size of 16 and an initial learning rate of 0.01. The experiment used the data augmentation algorithms identical to those used in the original YOLOv8 algorithm.
The evaluation indices used in this study included F1 score, mean Average Precision (mAP), GFLOPs and Frames Per Second (FPS). Precision and recall were fundamental indices, with F1 score and mAP, calculated on the bases of precision and recall, serving as the primary evaluation metrics to assess the model’s recognition accuracy. GFLOPs measured the complexity of the model or algorithm, and smaller GFLOPs indicated lower computational requirements and easier construction on low-end devices with reduced hardware performance demands. FPS refers to the number of frames detected per second, which depends not only on algorithm weights but also on the hardware configuration of the experimental equipment.
The precision ratio is the percentage of all results predicted correctly for a positive sample:
P r e c i s i o n = T P T P + F P
The recall rate is calculated on the basis of the proportion of all targets correctly predicted:
R e c a l l = T P T P + F N
where T P is the number of correctly detected targets, F P is the number of falsely detected targets and F N is the number of missed targets among the correct targets.
The formula for calculating the average accuracy across n classes is as follows:
m A P = 1 n i = 1 n 0 1 P r e c i s i o n ( R e c a l l ) d ( R e c a l l )
The F1 score comprehensively considers both precision and recall, providing a holistic reflection of the overall performance of the network. It is calculated as the harmonic mean of the two parameters as shown in the following:
F 1 = 2 P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l

5.2. Experimental Results and Analysis

5.2.1. Comparison of Ablation Experiments

To validate the accuracy of the proposed algorithm improvements, experiments were conducted with several other models: YOLOv8n, YOLOv8n-L, YOLOv8n-G, YOLOv8n-GV, YOLOv8n-LG, YOLOv8n-LGV, YOLOv8n-GGV and YOLOv8-CD. Specifically, YOLOv8n-L incorporated the LSKA module before the SPPF module in the YOLOv8n backbone network, and YOLOv8n-G used the Ghost module within the YOLOv8n backbone network. YOLOv8n-GV introduced GSConv and VoV-GSCSP structures into the neck network of YOLOv8n. YOLOv8-CD is the proposed algorithm in this paper.
As shown in Table 1, the improved algorithm adopts an efficient network structure to improve the network structure of YOLOv8n, which improved the accuracy and reduced the calculation amount of the model. It also demonstrates that the Ghost module did not compromise the algorithm’s accuracy. Furthermore, the new head network increased accuracy whilst reducing model computational requirements. The introduced GSConv module only marginally increased the computational load. By integrating these improvements with the YOLOv8n model, it effectively reduced the deployment difficulty and cost on mobile terminals whilst remarkably enhancing accuracy in real-time scenarios. Figure 13 presents the detection performance on the RDD2022 dataset before and after improvements, showing the significant improvements in the detection accuracy of cracks in (a), (b), (c) and (d).

5.2.2. Evaluation of Practical Application Detection

Application in the RDD2022 Dataset

Figure 14 displays the detection results of the YOLOv8-CD algorithm in the RDD2022 dataset. The improved algorithm identified more crack targets than the original algorithm, with varying degrees of detection accuracy enhancement shown in the figures. Both (a) and (d) detected cracks that were missed by other algorithms. The results show that the improved YOLOv8-CD algorithm effectively detected concrete surface cracks, accurately identifying their positions and categories and demonstrating strong robustness and accuracy.
To further validate the detection performance of the model for different targets, Table 2 shows the performance of YOLOv8n and the improved model, YOLOv8-CD, under various conditions. The data show that YOLOv8-CD achieved higher detection accuracy in each category than YOLOv8n. Specifically, mAP50 and mAP50-90 improved by 15.2% and 22.7%, respectively.
To further validate the algorithm’s performance, this study compared YOLOv8-CD with other object detection algorithms using the RDD2022 dataset. Table 3 shows the results.

Application in the Wall Crack Dataset

The Wall Crack dataset was also tested in this study. Figure 15 shows the detection results of the YOLOv8-CD algorithm using the Wall Crack dataset. Remarkable improvements in crack detection accuracy were observed in Figure 15a,b, whereas Figure 15c,d demonstrated cracks detected by YOLOv8-CD that were missed by other algorithms. The results indicate that the optimized YOLOv8-CD algorithm performed excellently in detecting concrete surface crack objects, accurately locating and identifying various types of cracks and exhibiting high stability and reliability. Additionally, the algorithm showed outstanding performance in detecting other types of damage on concrete surfaces.
To further validate the model’s detection performance with different targets, Table 4 presents the performance of YOLOv8n and the improved model YOLOv8-CD under various conditions. The table shows that YOLOv8-CD achieved higher detection accuracy in each category than YOLOv8n. Specifically, mAP50 and mAP50-90 improved by 12.3% and 17.2%, respectively.
To further validate the algorithm’s performance, this study also compared YOLOv8-CD with other object detection algorithms using the Wall Crack dataset. Table 5 shows the results.

5.2.3. Cross-Validation

To ensure the generalization ability and robustness of the model, this paper uses the Wall Crack dataset as an example, and a five-fold cross-validation method was employed in the experiments. Specifically, the entire dataset was randomly divided into five equal subsets. In each iteration, four subsets were used for model training, and the last subset for model validation. The process was repeated five times, each time selecting a different subset as the validation set, while the remaining four subsets were used for training. The final model performance was evaluated by averaging the results of the five validations.
The advantage of five-fold cross-validation is that it fully utilizes every sample in the dataset for both training and validation, thereby reducing the randomness caused by data partitioning and obtaining more reliable model performance evaluation metrics. In experiments, five-fold cross-validation not only effectively evaluated the model’s accuracy and recall rate, but also helped adjust the hyperparameters of the model and prevent the occurrence of overfitting. Table 6 shows the results of the five-fold cross-validation.
After cross-validation, it was found that the model’s performance was consistent across all folds, with an average precision of 91.5%, an average recall of 85.9%, an average mAP50 of 93.3%, and an average mAP50-95 of 77.6%. These results indicate that the model had high precision and recall, and it also demonstrated high average precision across different overlap thresholds. So, the model has good generalization ability and robustness.

6. Conclusions

Concrete surface crack detection is an important research area in structural safety, particularly in the structural assessment of roads, buildings and bridges. However, traditional crack detection methods typically rely on manual inspection, which is not only time-consuming and labor-intensive but also inefficient for meeting the demands of large-scale detection. Therefore, the automation of crack detection with deep learning algorithms has become a hotspot in related fields. Research have shown that traditional deep learning algorithms have been applied in crack detection. For example, Zhang et al. [34] proposed a deep learning model that was able to automatically detect and identify cracks from road images, which was validated in actual road crack detection. Zou et al. [35] proposed a DeepCrack model, which was also tested in various actual scenarios. Although these methods have been used in practice, their generalization ability and detection accuracy may be limited when dealing with the complexity and diversity of concrete surfaces.
As an advanced object detection algorithm, the YOLOv8 algorithm has shown a high potential in crack detection owing to its speed and high accuracy. However, there are numerous limitations when dealing with cracks on concrete surfaces. For instance, YOLOv8 may be disrupted by other textures, stains or shadows on concrete surfaces, leading to decreased accuracy in crack recognition. Therefore, this paper proposes an improved concrete surface crack detection algorithm based on YOLOv8n. The improved YOLOv8-CD algorithm of this paper is primarily applied to the crack detection on concrete surfaces. It can handle various types of cracks on concrete surfaces, including small and large cracks, while maintaining high detection accuracy and detection efficiency. This makes it suitable for large-scale crack detection tasks. This method is of great significance for the maintenance and repair of concrete structures. The main conclusions drawn are as follows:
(1)
In the crack detection process of concrete surfaces, to better capture the unique characteristics of concrete surface cracks, this study innovatively combines the advantages of visual attention networks and large-scale convolutional attention, introducing a novel LSKA module. This module not only effectively extracts the entire feature information of concrete surface cracks but also focuses on the local details of cracks, particularly adapting to the characteristics of cracks that are prone to breakage, have large spans and are thin and long. Through the LSKA module, the model pays extra attention to the shape of cracks, ensuring accurate identification and localization of cracks in complex concrete textures and backgrounds, thereby improving the accuracy and efficiency of crack detection.
(2)
The Ghost module is integrated into the backbone network of YOLOv8, aiming to efficiently extract and refine crucial information from raw features at minimal computational cost, largely enhancing the model’s feature extraction capability. The Ghost module reduces redundant computations by introducing a novel feature generation method. Accordingly, it not only effectively reduces the number of parameters and computation amount of the model, but also further improves its feature extraction capability and detection performance in complex scenarios whilst maintaining efficiency.
(3)
In the neck network of YOLOv8, to enhance the efficiency and performance of the model, the GSConv structure is introduced to replace the traditional convolutional structure. Simultaneously, a novel VoV-GSCSP module is incorporated on the basis of the characteristics of the YOLOv8 framework. This module integrates the GSConv structure whilst maintaining the advantages of the original network, allowing the effective utilization of computational resources during feature channel fusion and reducing floating-point operations. By introducing the GSConv and VoV-GSCSP modules, significant optimization is achieved in the neck network of YOLOv8 for feature extraction and fusion. This optimization enables the model to operate efficiently and accurately when dealing with complex images and detection tasks.
(4)
This paper achieves efficient and accurate crack detection by improving the YOLOv8 algorithm. The improved algorithm has achieved notable performance improvement with multiple datasets, thus proving its effectiveness and practicability in the field of crack detection. Compared with the existing models, this method has higher detection accuracy whilst reducing demands on platform computing and storage capabilities, and deploying on resource-limited devices becomes easy. In the future, we will continue exploring other advanced deep learning algorithms and technologies to further enhance the accuracy and efficiency of crack detection. Additionally, we aim to deploy the improved model on resource-limited embedded detection devices to refine the proposed algorithm in practical applications.

Author Contributions

Conceptualization, X.D. and J.D.; methodology, Y.L.; software, Y.L.; validation, X.D and J.D.; formal analysis, Y.L.; investigation, Y.L. and X.D.; resources, X.D. and J.D.; data curation, Y.L. and X.D.; writing—original draft preparation, Y.L.; writing—review and editing, Y.L.; visualization, Y.L.; supervision, X.D.; project administration, X.D. and J.D. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Natural Science Foundation of China (Grant No. 52368032, 51808272), the China Postdoctoral Science Foundation (Grant No. 2023M741455), the Tianyou Youth Talent Lift Program of Lanzhou Jiaotong University, the Gansu Province Youth Talent Support Project (Grant No. GXH20210611-10) and in part by the Natural Science Foundation of Gansu Province (Grant No. 23JRRA889), the Innovation Fund Project of Colleges and Universities in Gansu Province (Grant No. 2024B-057).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request. The data are not publicly available due to privacy.

Acknowledgments

The authors would like to thank the technical team of the Key Laboratory of Opto-Electronic Technology and Intelligent Control of the Ministry of Education, Lanzhou Jiaotong University and the National and Provincial Joint Engineering Laboratory of Road & Bridge Disaster Prevention and Control, Lanzhou Jiaotong University for their technical support.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Pan, Y.; Zhang, X.; Jin, X.; Yu, H.; Rao, J.; Tian, S.; Luo, L.; Li, C. Road pavement condition mapping and assessment using remote sensing data based on MESMA. In Proceedings of the 9th Symposium of the International Society for Digital Earth, Halifax, NS, Canada, 5–9 October 2015. [Google Scholar]
  2. Abdel-Qader, I.; Pashaie-Rad, S.; Abudayyeh, O.; Yehia, S. PCA-Based algorithm for unsupervised bridge crack detection. Adv. Eng. Softw. 2006, 37, 771–778. [Google Scholar] [CrossRef]
  3. Salman, M.; Mathavan, S.; Kamal, M.; Rahman, M. Pavement Crack Detection Using the Gabor Filter. In Proceedings of the 16th International IEEE Annual Conference on Intelligent Transportation Systems, Hague, The Netherlands, 6–9 October 2013. [Google Scholar]
  4. Zhou, S.; Song, W. Robust Image-Based Surface Crack Detection Using Range Data. J. Comput. Civ. Eng. 2019, 34, 04019054. [Google Scholar] [CrossRef]
  5. Vivekananthan, V.; Vignesh, R.; Vasanthaseelan, S.; Joel, E.; Kumar, K.S. Concrete bridge crack detection by image processing technique by using the improved OTSU method. Mater. Today 2023, 74, 1002–1007. [Google Scholar] [CrossRef]
  6. Zhu, Z.; Al-Qadi, I.L. Crack Detection of Asphalt Concrete Using Combined Fracture Mechanics and Digital Image Correlation. J. Transp. Eng. Part B Pavements 2023, 149, 04023012. [Google Scholar] [CrossRef]
  7. Rosso, M.M.; Marasco, G.; Aiello, S.; Aloisio, A.; Chiaia, B.; Marano, G.C. Convolutional networks and transformers for intelligent road tunnel investigations. Comput. Struct. 2023, 275, 106918. [Google Scholar] [CrossRef]
  8. Shahin, M.; Chen, F.F.; Maghanaki, M.; Hosseinzadeh, A.; Zand, N.; Khodadadi Koodiani, H. Improving the Concrete Crack Detection Process via a Hybrid Visual Transformer Algorithm. Sensors 2024, 24, 3247. [Google Scholar] [CrossRef] [PubMed]
  9. Chun, C.; Ryu, S.-K. Road Surface Damage Detection Using Fully Convolutional Neural Networks and Semi-Supervised Learning. Sensors 2019, 19, 5501. [Google Scholar] [CrossRef] [PubMed]
  10. Ghosh, S.; Singh, S.; Maity, A.; Maity, H.K. CrackWeb: A modified U-Net based segmentation architecture for crack detection. In Proceedings of the 3rd International Conference on Advances in Mechanical Engineering and its Interdisciplinary Areas, Kolaghat, India, 5–7 January 2021. [Google Scholar]
  11. Kang, D.; Benipal, S.S.; Gopal, D.L.; Cha, Y.J. Hybrid pixel-level concrete crack segmentation and quantification across complex backgrounds using deep learning. Autom. Constr. 2020, 118, 103291. [Google Scholar] [CrossRef]
  12. Meng, S.; Gao, Z.; Zhou, Y.; He, B.; Djerrad, A. Real-time automatic crack detection method based on drone. Comput.-Aided Civ. Infrastruct. Eng. 2022, 38, 849–872. [Google Scholar] [CrossRef]
  13. Chen, F.C.; Jahanshahi, M.R. NB-CNN: Deep Learning-Based Crack Detection Using Convolutional Neural Network and Naive Bayes Data Fusion. IEEE Trans. Ind. Electron. 2018, 65, 4392–4400. [Google Scholar] [CrossRef]
  14. Deng, J.; Lu, Y.; Lee, V.C.S. Imaging-based crack detection on concrete surfaces using You Only Look Once network. Struct. Health Monit. 2020, 20, 484–499. [Google Scholar] [CrossRef]
  15. Liu, Y.; Zhou, T.; Xu, J.; Hong, Y.; Pu, Q.; Wen, X. Rotating Target Detection Method of Concrete Bridge Crack Based on YOLO v5. Appl. Sci. 2023, 13, 11118. [Google Scholar] [CrossRef]
  16. Wu, P.; Liu, A.; Fu, J.; Ye, X.; Zhao, Y. Autonomous surface crack identification of concrete structures based on an improved one-stage object detection algorithm. Eng. Struct. 2022, 272, 114962. [Google Scholar] [CrossRef]
  17. Ye, G.; Qu, J.; Tao, J.; Dai, W.; Mao, Y.; Jin, Q. Autonomous surface crack identification of concrete structures based on the YOLOv7 algorithm. J. Build. Eng. 2023, 73, 106688. [Google Scholar] [CrossRef]
  18. Jiang, Y.; Pang, D.; Li, C. A deep learning approach for fast detection and classification of concrete damage. Autom. Constr. 2021, 128, 103785. [Google Scholar] [CrossRef]
  19. Lau, K.W.; Po, L.M.; Rehman, Y.A.U. Large Separable Kernel Attention: Rethinking the Large Kernel Attention design in CNN. Expert Syst. Appl. 2024, 236, 121352. [Google Scholar] [CrossRef]
  20. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C.; Xu, C. GhostNet: More Features from Cheap Operations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
  21. Li, H.; Li, J.; Wei, H.; Liu, Z.; Zhan, Z.; Ren, Q. Slim-neck by GSConv: A lightweight-design for real-time detector architectures. J. Real-Time Image PR 2024, 21, 62. [Google Scholar] [CrossRef]
  22. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  23. Redmon, J.; Farhadi, A. YOLO9000: Better, Faster, Stronger. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  24. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. Available online: https://arxiv.org/abs/1804.02767 (accessed on 22 May 2024).
  25. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. Available online: https://arxiv.org/abs/2004.10934 (accessed on 20 May 2024).
  26. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. Available online: https://arxiv.org/abs/2209.02976 (accessed on 25 May 2024).
  27. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 17–24 June 2023. [Google Scholar]
  28. Han, S.; Pool, J.; Tran, J.; Dally, W.J. Learning Both Weights and Connections for Efficient Neural Networks. Available online: https://arxiv.org/abs/1506.02626 (accessed on 15 May 2024).
  29. Dettmers, T. 8-BIT Approximations for Parallelism in Deep Learing. In Proceedings of the 4th International Conference on Learning Representations, San Juan, Puerto Rico, 2–4 May 2016. [Google Scholar]
  30. Hinton, G.; Vinyals, O.; Dean, J. Distilling the Knowledge in a Neural Network. Available online: https://arxiv.org/abs/1503.02531 (accessed on 15 May 2024).
  31. Howard, A.G.; Zhu, M.; Chen, B.; Kalenichenko, D.; Wang, W.; Weyand, T.; Andreetto, M.; Adam, H. MobileNets: Efficient Convolutional Neural Networks for Mobile Vision Applications. Available online: https://arxiv.org/abs/1704.04861 (accessed on 15 May 2024).
  32. Zhang, X.; Zhou, X.; Lin, M.; Sun, J. ShuffleNet: An Extremely Efficient Convolutional Neural Network for Mobile Devices. Available online: https://arxiv.org/abs/1707.01083 (accessed on 16 May 2024).
  33. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
  34. Zhang, L.; Yang, F.; Zhang, Y.D.; Zhu, Y.J. Road crack detection using deep convolutional neural network. In Proceedings of the 2016 IEEE International Conference on Image Processing, Phoenix, AZ, USA, 25–28 September 2016. [Google Scholar]
  35. Zou, Q.; Zhang, Z.; Li, Q.; Qi, X.; Wang, Q.; Wang, S. DeepCrack: Learning hierarchical convolutional features for crack detection. IEEE Trans. Image Process. 2019, 28, 1498–1512. [Google Scholar] [CrossRef] [PubMed]
Figure 1. YOLOv8n network architecture.
Figure 1. YOLOv8n network architecture.
Sensors 24 05252 g001
Figure 2. The C2f module of YOLOv8n.
Figure 2. The C2f module of YOLOv8n.
Sensors 24 05252 g002
Figure 3. The SPPF module of YOLOv8n.
Figure 3. The SPPF module of YOLOv8n.
Sensors 24 05252 g003
Figure 4. YOLOv8-CD network architecture.
Figure 4. YOLOv8-CD network architecture.
Sensors 24 05252 g004
Figure 5. The LSKA module.
Figure 5. The LSKA module.
Sensors 24 05252 g005
Figure 6. Block design of LSKA in VAN.
Figure 6. Block design of LSKA in VAN.
Sensors 24 05252 g006
Figure 7. The Ghost module.
Figure 7. The Ghost module.
Sensors 24 05252 g007
Figure 8. The Ghost Bottleneck module.
Figure 8. The Ghost Bottleneck module.
Sensors 24 05252 g008
Figure 9. The GS Bottleneck module.
Figure 9. The GS Bottleneck module.
Sensors 24 05252 g009
Figure 10. The VoV-GSCSP module.
Figure 10. The VoV-GSCSP module.
Sensors 24 05252 g010
Figure 11. The GSConv module.
Figure 11. The GSConv module.
Sensors 24 05252 g011
Figure 12. The new neck and head network in YOLOv8.
Figure 12. The new neck and head network in YOLOv8.
Sensors 24 05252 g012
Figure 13. Detection results: (a) detection of transverse cracks; (b) detection of longitudinal cracks; (c) detection of multiple longitudinal cracks; (d) detection of cracks near road markings.
Figure 13. Detection results: (a) detection of transverse cracks; (b) detection of longitudinal cracks; (c) detection of multiple longitudinal cracks; (d) detection of cracks near road markings.
Sensors 24 05252 g013
Figure 14. A comparison of detection results for different algorithms with the RDD2022 dataset: (a) detection of longitudinal cracks; (b) detection of transverse cracks; (c) detection of mixed cracks; (d) detection of cracks near road markings.
Figure 14. A comparison of detection results for different algorithms with the RDD2022 dataset: (a) detection of longitudinal cracks; (b) detection of transverse cracks; (c) detection of mixed cracks; (d) detection of cracks near road markings.
Sensors 24 05252 g014
Figure 15. A comparison of detection results for different algorithms with the Wall Crack dataset: (a) detection of cracks on concrete walls; (b) detection of multiple small cracks on concrete walls; (c) detection of cracks on concrete walls with influencing factors; (d) detection of cracks in bridges.
Figure 15. A comparison of detection results for different algorithms with the Wall Crack dataset: (a) detection of cracks on concrete walls; (b) detection of multiple small cracks on concrete walls; (c) detection of cracks on concrete walls with influencing factors; (d) detection of cracks in bridges.
Sensors 24 05252 g015
Table 1. Ablation experiment results.
Table 1. Ablation experiment results.
AlgorithmLSKAGhostGSConvF1/%mAP50/%mAP50-95/%GFLOPsFPS
YOLOv8n 78.578.348.78.289
YOLOv8n-L 93.293.570.18.287
YOLOv8n-G 89.790.065.88.283
YOLOv8n-GV 89.289.665.18.379
YOLOv8n-LG 89.389.765.58.282
YOLOv8n-LGV 90.190.367.28.383
YOLOv8n-GGV 92.193.270.08.084
YOLOv8-CD93.293.571.47.988
Table 2. A comparison of detection results for different types of cracks with the RDD2022 dataset.
Table 2. A comparison of detection results for different types of cracks with the RDD2022 dataset.
AlgorithmTypemAP50/%mAP50-95/%
YOLOv8nD0074.447.8
D1075.043.1
D2091.064.6
D4072.739.5
ALL78.348.7
YOLOv8-CDD0089.067.1
D1093.367.4
D2099.185.9
D4092.865.4
ALL93.571.4
Table 3. Results comparison of different algorithms with the RDD2022 dataset.
Table 3. Results comparison of different algorithms with the RDD2022 dataset.
AlgorithmF1/%mAP50/%mAP50-95/%GFLOPsFPS
Faster-RCNN49.451.222.5370.221
Mask-RCNN56.454.825.0110.624
YOLOv757.857.925.613.2114
YOLOv878.478.348.78.289
YOLOv8-CD93.293.571.47.988
Table 4. A comparison of detection results for different types of cracks with the Wall Crack dataset.
Table 4. A comparison of detection results for different types of cracks with the Wall Crack dataset.
AlgorithmTypemAP50/%mAP50-95/%
YOLOv8nD089.770.9
D177.156.8
D278.757.7
D385.868.3
D473.159.2
D584.357.8
ALL81.561.8
YOLOv8-CDD095.682.6
D191.875.2
D292.176.1
D394.582.3
D491.279.4
D597.778.1
ALL93.879.0
Table 5. Results comparison of different algorithms with the Wall Crack dataset.
Table 5. Results comparison of different algorithms with the Wall Crack dataset.
AlgorithmF1/%mAP50/%mAP50-95/%GFLOPsFPS
Faster-RCNN49.651.021.5370.221
Mask-RCNN55.355.627.3110.624
YOLOv756.257.925.613.2114
YOLOv881.881.561.88.289
YOLOv8-CD93.293.879.07.988
Table 6. The results of the cross-validation.
Table 6. The results of the cross-validation.
FoldTraining SetValidation SetPrecision/%Recall/%mAP50/%mAP50-95/%
1Fold 2, 3, 4, 5Fold 191.186.493.377.8
2Fold 1, 3, 4, 5Fold 291.486.092.877.2
3Fold 1, 2, 4, 5Fold 391.885.893.778.0
4Fold 1, 2, 3, 5Fold 491.285.393.177.4
5Fold 1, 2, 3, 4 Fold 592.186.193.677.8
Average 91.585.993.377.6
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Dong, X.; Liu, Y.; Dai, J. Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8. Sensors 2024, 24, 5252. https://doi.org/10.3390/s24165252

AMA Style

Dong X, Liu Y, Dai J. Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8. Sensors. 2024; 24(16):5252. https://doi.org/10.3390/s24165252

Chicago/Turabian Style

Dong, Xuwei, Yang Liu, and Jinpeng Dai. 2024. "Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8" Sensors 24, no. 16: 5252. https://doi.org/10.3390/s24165252

APA Style

Dong, X., Liu, Y., & Dai, J. (2024). Concrete Surface Crack Detection Algorithm Based on Improved YOLOv8. Sensors, 24(16), 5252. https://doi.org/10.3390/s24165252

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop