Next Article in Journal
Rhythms, Patterns and Styles in the Jaw Movement Activity of Beef Cattle on Rangeland as Revealed by Acoustic Monitoring
Previous Article in Journal
Enhanced N-Butanol Sensing Performance of Cr-Doped CeO2 Nanomaterials
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Preprocessing Method for Insulation Pull Rod Defect Dataset Based on the YOLOv5s Object Detection Network

1
Department of Electrical Engineering, Tsinghua University, Beijing 100084, China
2
Shandong Taikai Electrical Insulation Co., Ltd., Taian 271000, China
*
Author to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2025, 25(4), 1209; https://doi.org/10.3390/s25041209
Submission received: 27 January 2025 / Revised: 13 February 2025 / Accepted: 14 February 2025 / Published: 17 February 2025
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Insulation pull rods used in gas-insulated switchgear (GIS) inevitably contain the micro defects generated during production. The intelligent identification method, which requires large datasets with a balanced distribution of defect types, is regarded as the prevailing way to avoid insulation faults. However, the number of defective pull rods is limited, and the occurrence of different types of defects is highly imbalanced in actual production, leading to poor recognition performance. Thus, this work proposes a data preprocessing method for the insulation pull rod defect feature dataset. In this work, the YOLOv5s algorithm is used to detect defects in insulation pull rod images, creating a dataset with five defect categories. Two preprocessing methods for impurities and bubbles are introduced, including copy–paste within images and bounding box corrections for hair-like impurities. The results show that these two methods can specifically enhance small-sized defect targets while maintaining the detection performance for other types of targets. In contrast, the proposed method integrates copy–paste within images with Mosaic data augmentation and corrects bounding boxes for hair-like impurities significantly improving the model’s performance.

1. Introduction

Gas-insulated switchgear (GIS) has been widely used in power grids [1,2], in which the insulation pull rod is employed as a key insulating component in GIS and is responsible for transmitting mechanical motion. Insulation pull rods are typically made from fiber-reinforced epoxy composites, and defects can occur during production, transportation, and installation. The presence of defects may lead to insulation breakdown and further affect the reliable operation of GIS equipment [3,4]. Therefore, accurate detection of insulation pull rod defects is essential to reduce the operational risks of GIS devices and ensure the high reliability of power grid operations.
In recent years, with the rapid development of artificial intelligence and deep learning technologies, deep learning-based object detection techniques have made significant progress. The YOLO (You Only Look Once) series is a typical single-stage detector widely used for object detection. In the power industry, YOLO-based detectors have been applied in various scenarios. Feng et al. [5] introduced the K-means clustering method into the YOLOv5x model to achieve automated detection of insulator defects in power transmission lines. Liu et al. [6] incorporated Ghost convolution and model pruning into the YOLOv5s network to reduce model parameters. This approach significantly improved detection speed without compromising accuracy and was used to detect whether safety harnesses were properly worn during high-altitude operations by power workers. Nguyen [7] proposed using AI-powered drones to automate the inspection process of power facilities. Based on the YOLOv5s network, the system monitored the condition of five types of electrical equipment: crossarms, isolators, insulators, poles, and transformers. Pérez-Aguilar et al. [8] developed an automated computer vision mechanism based on YOLOv5 to detect hot spots in thermal images of electrical substations. Li et al. [9] introduced an image enhancement method based on illumination correction and compensation to overcome lighting challenges. Combined with the YOLOv5 model, this method was used to detect insulator defects. Liang et al. [10] proposed a lightweight non-destructive testing algorithm for welding defect surfaces based on attention mechanism, which solves the problem of it being difficult to efficiently and accurately detect complex defects in the ultrasonic welding process. However, current research mainly focuses on the recognition of large targets, such as insulators and human bodies, and the detection accuracy still needs to be improved when dealing with small-sized defects. The insulation rod defects that need to be detected in this study are generally small-sized defects, and these targeted detection methods need to be studied.
Another study by Li et al. [11] leveraged the YOLOv5s model to detect typical defects in insulation pull rods. They demonstrated that YOLOv5s has a faster inference speed and higher recognition accuracy compared to other mainstream recognition models, including Faster R-CNN, Mask R-CNN, SSD, YOLOv4, etc., in identifying insulation rod defects. However, the dataset primarily consisted of insulation pull rods with highly distinctive features, resulting in decreased detection accuracy for rods with less prominent features encountered in real production procedures.
In practical applications, to maximize the accuracy and efficiency of a model, it is essential to ensure a large and diverse dataset. However, for defect detection in insulation pull rods, directly collecting a large number of images of typical defects is labor-intensive and incurs high production costs. Even if this data collection is feasible, the dataset may still be insufficient for model training. Therefore, data augmentation techniques are required to increase both the quantity and diversity of the dataset. For image datasets, various augmentation methods are available, including image scaling, random cropping, random flipping, color dithering, random erasing, Mixup, CutMix, Mosaic, and others to enhance data diversity [12]. Among these, Mosaic data augmentation is one of the most commonly used techniques for defect detection in insulation pull rod images. In fact, the Mosaic data augmentation method can improve data utilization efficiency, enhancing the diversity and robustness of the dataset, especially when the dataset is small. It has been widely applied in the processing of various image or video datasets [13,14].
However, this data augmentation method has some drawbacks. The generated images tend to have uniform sizes. Since there is an upper limit on the number of pixels in the images used for training, this results in a pixel limit for the images produced by Mosaic data augmentation. Consequently, the clarity of the original images is inevitably reduced, and the image size is decreased. Defect sizes in insulation pull rods vary, with defects such as bubbles and impurities being smaller in size. After applying Mosaic data augmentation, many labeled defects have a size smaller than 3 pixels, leading to poor correspondence between anchor boxes and the defect locations, which significantly hampers the accuracy of the training.
As for the typical defects observed in GIS, each individual defect conveys unique features and contours. White spots, scratches, and cracks are relatively larger in size and have more samples. The size of bubble defects is smaller, with a typical diameter of around 5 mm [15]. Impurity defects, on the other hand, vary greatly in size due to the wide variety of impurities, which include metal particles, hair, threads, and other contaminants [11]. Bubbles and metal impurities have very small sizes, with diameters ranging from 1 mm to 5 mm, and occupy a small proportion of the image in insulation pull rod pictures. Most image processing methods are designed for targets that occupy a relatively larger portion of the image, so when performing data augmentation, it is difficult to specifically enhance these small defects. In fact, the small size of bubbles and impurities may even lead to semantic loss in the image. Semantic loss refers to the degradation of small target features caused by downsampling or excessive enhancement in data augmentation.
In addition, due to the uneven distribution of defects in the actual production of pull rods, the proportion of defects such as white spots is much higher than that of defects such as bubbles. The proportion of imbalanced targets in the original data will be further imbalanced after expansion, resulting in a significant decrease in recognition performance. Kim et al. [16] indicated that the dataset significantly affects the accuracy of object recognition. Issues such as a small dataset size and an imbalanced distribution of different object types can lead to a decline in the model’s recognition performance. Khan et al. [17] proposed that data imbalance is a common problem in many practical applications, where the number of samples in most classes far exceeds that in a few classes. This imbalance can cause the classifier to lean towards the majority class, thereby affecting the recognition performance of the minority class. Therefore, when processing dataset augmentation, it is also necessary to consider the proportion balance of various targets.
Additionally, hair-like impurities are long in length but extremely narrow in width, with complex shapes. When labeling bounding boxes, a large portion of the area in the box is invalid. Therefore, the expanded areas are mostly invalid during data augmentation, which may result in distortion of semantic information.
To address the faced challenges, this study proposes a data preprocessing method that combines Mosaic data augmentation with in-image copy–pasting to enhance the defect features of insulation pull rods. Additionally, the bounding boxes for hair-like impurities are corrected, enabling the diverse shapes of these impurities to be standardized into two forms: diagonal and arc-shaped, thus improving detection accuracy for hair-like impurities. The YOLOv5s model, trained on the augmented dataset using this method, is employed to recognize defects in insulation pull rod images. The results show that, compared to using only Mosaic data augmentation, the model trained on the dataset expanded with this method and improved its mAP@0.5:0.95 from 0.618 to 0.789.

2. YOLOv5s Network Architecture and Procedure

2.1. Overview of YOLOv5s Network

In this study, the YOLOv5 model is used for defect detection, as it is a classic object detection model. Compared to other traditional object detection models, such as SSD and Faster R-CNN, YOLOv5 offers advantages, including fast detection speed, high accuracy, ease of training, and lightweight architecture [18,19]. YOLOv5 is also tailored for different application scenarios and is divided into four variants based on depth and feature map width: YOLOv5s, YOLOv5m, YOLOv5l, and YOLOv5x. Among these, the YOLOv5s model has the smallest depth and feature map width, while also providing faster frame rates and inference speeds [20]. Given that insulation pull rod defect detection requires both high accuracy and fast detection speed, the YOLOv5s model is the most suitable for this task. The YOLOv5s model consists of three main components: the backbone network which is responsible for feature extraction; the neck network which is responsible for feature fusion; and the prediction network. The procedure is shown in Figure 1.
The default training procedure for the YOLOv5s network is as follows: First, the input image is processed by resizing it to the preset dimensions of the network. In this study, the default size used is 640 × 640 pixels. The resized image is then split into its three RGB channels, and each pixel’s information is normalized to a floating-point number in the range of 0 to 1. Next, based on the requirements, the model can optionally use the default Mosaic data augmentation to expand the dataset, enhancing its robustness and diversity. Additionally, during the dataset preprocessing phase, the model calculates adaptive anchor boxes based on the pre-labeled bounding boxes in the dataset in order to determine the optimal anchor boxes for that particular dataset. Once the preprocessing steps are completed, the model begins several epochs of training. In each training epoch, the YOLO algorithm uses the original anchor boxes as references, calculates the differences between the predicted anchor boxes and the actual target locations, and updates the weight matrix using backpropagation. After training is finished, the model outputs the weight matrix corresponding to the best-performing round, which is then used for object detection.

2.2. Three Parts of YOLOv5s Network

2.2.1. Backbone Feature Extraction Network

The main task of the backbone network is to extract features from images, converting the input raw image into multi-level feature maps. This part consists of the Conv module, C3_X module, and SPPF (Spatial Pyramid Pooling Fusion) module. The Conv module includes convolutional layers, batch normalization (BN) layers, and activation functions. The formula for the BN layer is shown in Equation (1), where the subscript k represents the k-th dimension of the sample, xk and yk represent the input and output samples of the BN layer, respectively; βk and γk represent the shift and scale parameters, which are not hyperparameters but are learned through the network; and μk and σk represent the mean and standard deviation of the input samples [21].
y k =   γ k   x k μ k   σ k 2 + ϵ + β k
The convolutional layers extract local features from the image, the BN layers normalize the extracted features, and the activation functions provide the network with nonlinear properties. The C3_X module stacks multiple convolutional operations with residual connections to efficiently extract multi-level information from the input features while reducing information loss during the feature extraction process. The SPPF module is introduced at the end of the backbone network to apply pooling operations of different scales to the convolutional feature maps, then fuses the pooled results to produce output feature maps with a consistent size, which can effectively expand the receptive field of the feature maps [22].

2.2.2. Neck Feature Fusion Network

The neck network adopts a structure combining FPN (Feature Pyramid Network) and PAN (Path Aggregation Network) [23]. FPN uses a top-down approach to upsample the final feature map from the backbone network and fuses it with feature maps of the corresponding scales from earlier layers, enriching the feature representation. After repeating this process, the resulting fused feature maps have dimensions of 40 × 40 × 512 and 80 × 80 × 256. PAN uses a bottom-up strategy to process the lower-level feature maps of FPN through convolution operations, then fuses the results with feature maps of the same scale from FPN. This process is repeated until it reaches the top layer of FPN, producing fused feature maps of 40 × 40 × 256 and 20 × 20 × 512. FPN propagates high-level semantic information downwards, while PAN propagates positional information upwards. Together, they perform feature aggregation at different detection layers.

2.2.3. Prediction Network

The YOLOv5 model’s loss function is divided into three parts: confidence, classification, and localization loss functions. The training process is essentially the process of minimizing the loss function. The localization loss function uses the GIOU loss function [24].
The formula for the YOLOv5 confidence loss function is shown in Equation (2) [25]. Here, o represents the true confidence value, which is 1 when sample i is a positive sample and 0 otherwise; c i ^ represents the predicted confidence value, indicating the probability that the model predicts sample i as a positive sample; and N is the total number of positive and negative samples.
L conf o , c = i o i ln c i ^   + 1 o i ln 1 c i ^ N ,   c i ^ = Sigmoid c i
The formula for the YOLOv5 localization loss function is shown in Equation (3). In this formula, (tx, ty, tw, th) represents the predicted offset of sample i relative to the anchor box, where the coordinate values are not processed by the Sigmoid activation function; l i ^ represents the predicted offset of sample i after being processed by the Sigmoid function; ( g i x , g i y ,  g i w ,  g i h ) represents the true center coordinates and width or height of sample i; ( c i x , c i y ,  p i w ,  p i h ) represents the center coordinates and width or height of the anchor box i; and g i ^ represents the true offset of sample i relative to the anchor box.
L loc l , g = i pos   m x , y , w , h   L GIoU     l i ^ m g i ^ m N pos
The prediction network outputs three feature maps of different scales to detect large, medium, and small objects. YOLOv5s divides the 640 × 640 × 3 input image into N × N grid cells and predicts information for each cell, including the bounding box center coordinates, width, height, classification probabilities, and confidence scores. The bounding box coordinates and size represent the precise position and size of the predicted object, while the confidence score indicates the likelihood that the object exists in the grid cell, with higher values suggesting a higher probability of the object’s presence. The classification probabilities determine the classification information of the predicted object.
First, the confidence score of each predicted bounding box is compared with a set threshold. Predictions with confidence scores below the threshold are discarded, while those above the threshold are considered to contain targets, with the position and size of the detected targets obtained. Next, the Non-Maximum Suppression (NMS) algorithm is applied to filter duplicate bounding boxes for the same target [26]. Finally, the target’s index and category are determined based on the highest classification probability. The detection process of YOLOv5s network is shown in Figure 2.

3. Two Preprocessing Methods for Insulation Pull Rod Defect Dataset

3.1. Copy–Paste Data Augmentation

For small-sized bubble and impurity defects, directly applying Mosaic data augmentation leads to a significant reduction in the size of the defects in the generated images, which in turn results in a marked decrease in training performance. In this study, we use the copy–paste method for augmenting the defect dataset. The approach involves copying a portion of the image from within the bounding box of the original image and pasting it randomly multiple times within the same image, thereby increasing the number of corresponding targets and enhancing the diversity of the dataset. In the original image, the insulation pull rod typically occupies only a central portion, while the top and bottom areas are mostly background, so the pasted regions are likely to fall into the previously inactive areas, ensuring that the defect regions are not obscured.
This data augmentation method is applied only to bubble and impurity defects. For each image, if there are multiple bubble or impurity defects, the copy–paste operation is applied to each defect individually. Other types of defects are left untouched. The same original image can be augmented into multiple processed images, effectively increasing the number of bubble and impurity defect samples.
After performing copy–paste on the image, the resulting images are combined with the original images for Mosaic data augmentation. The Mosaic-augmented images are then mixed with images that only underwent copy–paste augmentation, forming the training dataset for the network. In the neck network, FPN upsamples from top to bottom, while PAN applies convolution to the lower-level feature maps of FPN in a bottom-up manner, and then fuses the convolution results with feature maps of the same scale from FPN. This generates three fused feature maps of sizes 20 × 20 × 1024, 40 × 40 × 512, and 80 × 80 × 256. The procedure of Copy–Paste Data Augmentation Method is shown in Figure 3.
This method is particularly effective for small-sized impurity and bubble defects. After data augmentation, it preserves the large features in the dataset while also retaining the small-sized features introduced by Mosaic. In other words, it adds large and medium-sized defects of the corresponding types to the dataset, significantly increasing the number of these two types of defects, and leading to a more balanced distribution of different defect types in the dataset.

3.2. Correcting Bounding Boxes for Hair-like Impurities

Hair and threads are typical impurities that may occasionally fall into the raw materials during the production of insulation pull rods. These impurities typically feature very fine widths, and long lengths, and are prone to bending into various shapes. During training, YOLOv5s uses a horizontal rectangular bounding box format, which cannot rotate, making it prone to labeling a large number of irrelevant areas when annotating hair-like impurities. This significantly deteriorates the training results.
To address this issue, this study proposes a correction to the annotation method for hair-like impurities. Instead of annotating a single continuous hair strand, the hair is divided into multiple segments. These segments take two main shapes: diagonal and arc-shaped. A diagonal-shaped segment extends from one vertex of the bounding box to the opposite vertex, while an arc-shaped segment extends from one vertex to another vertex on the same side of the box. This method simplifies the diverse shapes of hair-like impurities into two more regular forms, making them more typical and easier to identify. Moreover, during detection, even if a small part of the hair impurity is not detected, the presence of any detected region is sufficient to consider the defect recognized.
The principles of hair impurity bounding box correction include the following: (1) The angle difference between diagonal hair impurities and the diagonal body should not exceed ±15 degrees; (2) The radius of the annotation box for arched hair impurities should not exceed 90 degrees of the corresponding curvature circle; and (3) To avoid the annotation box being too small, the number of segments should not exceed six every 360 degrees.
The process begins by applying the above annotation method to the hair-like impurities in the original defect images of the insulation pull rods. Then, the annotated hair-like impurities are subject to the copy–paste augmentation, and the augmented images are combined with the original images for Mosaic data augmentation. These Mosaic-augmented images, along with the images that only underwent copy–paste augmentation, are then mixed to form the training dataset for the network. After passing through the FPN and PAN feature extraction layers in the YOLOv5s neck network, three fused feature maps with sizes 20 × 20 × 1024, 40 × 40 × 512 and 80 × 80 × 256 are generated. This approach operates solely on hair-like impurities based on the copy–paste augmentation, without making any additional adjustments for other types of defects. The procedure of bounding boxes of hair-like impurities bounding boxes is shown in Figure 4.

4. Experiment Results and Analysis

4.1. Overview

In this study, the YOLOv5s model was trained on multiple datasets to compare its detection performance and validate the effectiveness of the preprocessing methods described earlier. A total of seven datasets were set up, named Dataset 1 through Dataset 7. Each dataset contains 8000 images, with the detection targets being five types of insulation pull rod defects: white spots, impurities, bubbles, cracks, and scratches.
However, in the original dataset, the proportion of various types of defects is very unbalanced, and the specific quantities and proportions of the five types of defects are shown in Table 1.
Before data augmentation, there were a total of 397 original images, of which 350 images were taken as the expanded basis for the training set, including five types of defects. The remaining 47 images will be kept as the validation set. After data augmentation, the original 350 images were expanded to 8000 images. Each target type defect was copied and pasted five times.
As shown in Table 2, these datasets were used to evaluate the performance of the model with different preprocessing techniques.
To evaluate the effectiveness of the data preprocessing methods, mAP@0.5 and mAP@0.5:0.95 are used as evaluation metrics for detection accuracy, where mAP (mean average precision) represents the average detection accuracy across all target types, and AP (average precision) represents the average detection accuracy for a specific target type. The values of mAP and AP can be computed using Equations (4) and (5):
mAP = 1 n A P i
AP = 0 1 P ( R ) d R
In Equation (5), the average precision is the area under the P-R (Precision-Recall) curve, where P(Precision) and R(Recall) are defined by Equations (6) and (7):
P = T P T P + F P
R = T P T P + F N
TP (True Positive) refers to cases where the actual label is positive, and the prediction is also positive. TN (True Negative) refers to cases where the actual label is negative, and the prediction is also negative. FP (False Positive) refers to cases where the actual label is negative, but the prediction is positive. FN (False Negative) refers to cases where the actual label is positive, but the prediction is negative.
F1 score is another important parameter for evaluating the effectiveness of a model, which can be used to comprehensively assess the balance between P and R. The value of F1 score can be computed using Equation (8):
F 1   Score = 2 PR P + R
Based on these definitions, precision reflects the model’s ability to correctly detect positive instances, and recall reflects the proportion of actual positive instances correctly predicted by the model. In this study, the values of mAP@0.5, mAP@0.5:0.95, P, R and F1 score for each defect type are used as the evaluation criteria. Specifically, mAP@0.5 refers to the mAP value when the Intersection over Union (IoU) threshold is set to 0.5, while mAP@0.5:0.95 refers to the average mAP value when the IoU threshold is varied from 0.05 to 0.95 in steps of 0.05.
The Python version used in this study is 3.9.19, and the PyTorch version is 2.3.1.

4.2. Validation of the Effectiveness of Two Methods

4.2.1. Validation of the Effectiveness of the Copy–Paste Method

In this study, seven training datasets were used. By training with Dataset 1, Dataset 2, and Dataset 5, in which no correction to the bounding boxes of hair-like impurities was applied, we can compare whether combining the copy–paste method with Mosaic data augmentation improves the defect detection performance of insulation pull rods. The training was conducted using the same parameters; the starting weight file was the default YOLOv5s model weight file, and the number of training epochs was set to 100. Since the datasets had already undergone data augmentation, no further augmentation was applied during training. The training results are shown in Table 3.
Table 3 presents the training results for three datasets, listing the number of annotated defects of each type after data augmentation, along with the corresponding mAP@0.5 and mAP@0.5:0.95 values. From the table, it can be observed that when all images in the dataset undergo Mosaic data augmentation, the total number of defect samples is the highest, but the distribution is uneven. This is because, in the original data, the number of white spots and impurities is already high; therefore, after augmentation, these two defect types account for 75% of the total samples. Although the number of crack and scratch samples is relatively small, these defects are larger in size and have more distinct features, leading to better detection results. In contrast, the bubble defects, which originally constituted only 4.8% of the dataset, are smaller in size, leading to poorer training results. The detection performance for impurities is also relatively poor.
As the number of augmented images using the copy–paste method increases, the total number of samples decreases, but the numbers of bubble and impurity defect samples decrease less or even increase. The distribution of defects becomes more balanced. It can be observed that the detection performance for bubble defects has significantly improved, while the detection of the other defect types has not noticeably declined. Although the recognition accuracy for white spots, cracks, and scratches in Datasets 2 and 4 is slightly lower than in Dataset 1, this decline is closely related to the reduction in number of samples. In future work, increasing the sample size for these two defect types can help improve their recognition performance.
Based on the Precision, Recall, and F1 score metrics, the precision values of all three datasets are relatively high; however, in Dataset 1, the recall values for bubble and impurity defects are comparatively low. This indicates that in the dataset trained using default Mosaic data augmentation, most of the identified bubble and impurity defects are correctly recognized, but a significant proportion of defects are still missed. In contrast, after applying in-image copy–paste augmentation for bubble and impurity defects, as the number of augmented images increases, the recall for bubbles significantly improves, while the increase in recall for impurities is relatively modest. Consequently, the F1 scores for both bubbles and impurities improve with the increasing number of copy–paste augmented images. Specifically, the F1 score for bubbles increases from 0.585 to 0.913 (Table 3, Dataset 1 and Dataset 5, “Bubble” column), which is comparable to the F1 scores of other defect types; however, the F1 score for impurities remains significantly lower at only 0.793—more than 0.1 below that of other types (Table 3, Dataset 1 and Dataset 5, “Impurity” column).
This suggests that the enhancement in impurity detection is limited, indicating that the proposed method is less effective for irregular defects like impurities, and that alternative approaches are needed to further improve impurity detection performance.

4.2.2. Validation of the Effectiveness of Hair-like Impurity Bounding Boxes Correction

By comparing Datasets 2, 3, and 4, and separately comparing Datasets 5, 6, and 7, we can evaluate whether the bounding box correction of hair-like impurities improves the defect detection performance of the insulation pull rod. The same training parameters were used, with the starting weight file being the default YOLOv5s model weights. The number of training epochs was set to 100, and since each dataset had already undergone data augmentation, no further augmentation was applied during training. The training results are shown in Table 4 and Table 5.
From the results shown in the tables, it is evident that, under otherwise identical conditions, correcting the bounding boxes for hair-like impurities can significantly improve the recognition accuracy of impurity-related defects.
For Dataset 2, after correcting the bounding boxes for hair-like impurities and applying the copy–paste augmentation method, the mAP@0.5:0.95 for impurity defects increased by 8.3%. If the images with corrected bounding boxes are further subjected to Mosaic data augmentation, combined with copy–paste augmentation, the mAP@0.5:0.95 for impurity defects increased by 23.2%, with the mAP@0.5 surpassing 90%.
For Dataset 5, after correcting the bounding boxes for hair-like impurities and applying only the copy–paste augmentation, the mAP@0.5:0.95 for impurity defects increased by 3.3%. When the corrected images were further augmented using Mosaic data augmentation combined with copy–paste method, the mAP@0.5:0.95 for impurity defects increased by 13.8%, and the mAP@0.5 approached 95%.
Based on the Precision, Recall, and F1 score metrics, under otherwise identical conditions, correcting the bounding boxes for hair-like impurities can effectively improve the recall for impurity defects, thereby also enhancing their F1 score. For Dataset 2, the F1 score for impurity defects increased from 0.678 to 0.889 (Table 4, Dataset 2 and Dataset 4, “Impurity” column); for Dataset 5, it increased from 0.793 to 0.913 (Table 5, Dataset 5 and Dataset 7, “Impurity” column). Meanwhile, the Precision, Recall, and F1 scores for other defect types were not significantly affected. Therefore, it can be concluded that correcting the bounding boxes for hair-like impurities enables the model to more accurately detect these impurities, ultimately improving the detection rate of impurity defects.
The results show that correcting the bounding boxes for hair-like impurities, followed by Mosaic data augmentation and combined with copy–paste, performs significantly better than applying only copy–paste after the bounding box correction, while still using the original images for Mosaic data augmentation. Moreover, this method not only improves the recognition accuracy of impurity defects but also enhances the detection of other types of defects to a certain extent. For Dataset 2, the overall mAP@0.5:0.95 for all defect types increased by 13.6%, Precision for all defect types increased by 0.5%, Recall for all defect types increased by 14.2%, and the F1 Score for all defect types increased by 8.2% (Table 4, Dataset 2 and Dataset 4, “All” column). For Dataset 5, the overall mAP@0.5:0.95 for all defect types increased by 8.2%, Precision for all defect types increased by 1.3%, Recall for all defect types increased by 9.7%, and F1 Score for all defect types increased by 5.9% (Table 5, Dataset 5 and Dataset 7, “All” column). The comparison of mAP@0.5:0.95 performance and P/R/F1 score performance cross different data augmentation methods, as Figure 5 shows.
Figure 5 shows a comprehensive comparison of the best-performing datasets in each group. It can be observed that correcting the bounding boxes for hair-like impurities, followed by Mosaic data augmentation combined with copy–paste, resulted in a significant improvement in detection performance compared to the model trained solely with Mosaic data augmentation. The overall mAP@0.5:0.95 for all defect types increased from 61.8% to 78.9%, and a corresponding improvement of 17.1% is observed as well. The overall F1 score for all defect types increased from 0.816 to 0.951, and a corresponding improvement of 13.5% is observed as well (Table 3 and Table 5, Dataset 1 and Dataset 7, “All” column).
The model complexity during training is 7,033,114 parameters, 16.0 GFLOPs. The inference time and GPU memory consumption are listed in Table 6.
It can be found that, after copying and pasting in the image and modifying the hair impurity bounding box, the inference time is not affected much, and the GPU memory consumption is almost unchanged.

4.3. Comparison of Detection Results on Real Photos

To further evaluate the effectiveness of the two data preprocessing methods in improving the defect detection accuracy of insulation rods, Figure 6 presents the detection results of models trained on different datasets using the same set of images. Three test images were selected for comparison. The first image contains multiple defects of white spots and impurities; the second image contains multiple bubble defects; and the third image includes both a large and small hair-like impurity. The models trained on datasets 1, 5, 4, and 7 were named A, B, C, and D, respectively.
From the detection results of the first image, it can be observed that models A and B failed to detect all the impurity defects, with model A having noticeably lower detection confidence. Model C detected all the defects but made a false positive, misclassifying a white spot as both a white spot and an impurity. Model D did not make any false positives, but missed a small impurity defect, and its confidence was slightly lower than that of Model C. In the second image, Model A failed to detect all the bubble defects, while Models B, C, and D successfully detected all bubble defects, with Models C and D showing similar confidence, both higher than Model B. In the third image, Model A completely failed to detect the larger hair impurity, while Models B and C missed the smaller hair-like impurity. Model D successfully detected both hair-like impurities, with better bounding box coverage and higher confidence than Model C.
Comparing these results, it is clear that using only Mosaic data augmentation leads to poorer detection performance. In contrast, combining image copy-pasting with Mosaic augmentation significantly improves the model’s detection accuracy. Additionally, correcting the bounding boxes for hair-like impurities enhances the model’s accuracy in detecting this type of defect. However, the ratio of image copy-pasting to Mosaic-augmented images should not be excessively high, as an overly large number of copy–pasted images may not significantly improve the detection accuracy and could even lead to a decrease in performance.

5. Conclusions

The traditional inspection methods used for insulation rods after production are often subjective, time-consuming, and labor-intensive, which may result in missed detections or errors. This study proposes using a YOLOv5s-based image recognition system for defect detection in insulation pull rods and introduces a combined data augmentation method involving in-image copy & paste and Mosaic augmentation. Additionally, a framework for correcting bounding boxes for hair-like impurities is proposed to improve the model’s accuracy in detecting defects. Based on the above work, the following conclusions can be drawn:
(1)
Comparing with traditional Mosaic data augmentation, the in-image copy–paste technique significantly improves the detection accuracy of small bubble defects. This approach increases the mAP@0.5 for bubble defects from 0.531 to 0.904, increases the mAP@0.5:0.95 from 0.329 to 0.766 and increases the F1 score for bubbles from 0.585 to 0.913 (Table 3, Dataset 1 and Dataset 5, “Bubble” column) It also raises the overall mAP@0.5 from 0.797 to 0.897, raises the overall mAP@0.5:0.95 from 0.618 to 0.707 and raises the overall F1 score from 0.816 to 0.892 (Table 3, Dataset 1 and Dataset 5, “All” column).
(2)
Correcting the bounding boxes for typical hair-like impurities improves the detection accuracy of these defects. Standardizing the diverse shapes of hair-like impurities into two regular forms-arched-type and diagonal-type, combined with the in-image copy–paste and Mosaic data augmentation approach, improves the detection of impurity defects. The mAP@0.5 for hair-like impurities increases from 0.777 to 0.946, mAP@0.5:0.95 rises from 0.593 to 0.731, and the F1 score rises from 0.793 to 0.913(Table 5, Dataset 5 and Dataset 7, “Impurity” column). Meanwhile, the overall mAP@0.5 improves from 0.897 to 0.976, overall mAP@0.5:0.95 increases from 0.707 to 0.789, and the overall F1 score rises from 0.892 to 0.951(Table 5, Dataset 5 and Dataset 7, “All” column).
(3)
The proportion of images generated through the in-image copy–paste method should not be too high, as an excessive number of such images may negatively impact detection accuracy.
There are still several promising directions remaining for future work. Introducing Generative Adversarial Networks (GANs) to synthesize high-resolution defect samples and further alleviate data imbalance is a feasible option [27]. Meanwhile, investigating feature fusion enhancements or hybrid detection approaches will also benefit the improvement of the model.

Author Contributions

Conceptualization, X.L. and C.L.; software, X.L., M.C. and B.L.; validation, X.L., M.C. and B.L.; writing—original draft preparation, X.L. and M.C.; writing—review and editing, X.F., W.Q., F.L. and J.H. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Science Fund Program for Distinguished Young Scholars (Overseas) under grant number 23FAA00813.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Due to the nature of this research, study participants did not agree for their data to be shared publicly, so supporting data are not available.

Acknowledgments

The authors thank the anonymous reviewers and academic editors for their constructive comments and helpful suggestions.

Conflicts of Interest

Authors Meng Cong and Bo Liu were employed by the company Shandong Taikai Electrical Insulation Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Li, C.; Lin, C.; Chen, G.; Tu, Y.; Zhou, Y.; Li, Q.; Zhang, B.; He, J. Field-Dependent Charging Phenomenon of HVDC Spacers Based on Dominant Charge Behaviors. Appl. Phys. Lett. 2019, 114, 203501. [Google Scholar] [CrossRef]
  2. Li, C.; Zhu, Y.; Hu, J.; Li, Q.; Zhang, B.; He, J. Charge Cluster Triggers Unpredictable Insulation Surface Flashover in Pressurized SF6. J. Phys. D Appl. Phys. 2020, 54, 015308. [Google Scholar] [CrossRef]
  3. Li, C.; Yang, Y.; Xu, G.; Zhou, Y.; Jia, M.; Zhong, S.; Gao, Y.; Park, C.; Liu, Q.; Wang, Y.; et al. Insulating Materials for Realising Carbon Neutrality: Opportunities, Remaining Issues and Challenges. High Voltage 2022, 7, 610–632. [Google Scholar] [CrossRef]
  4. Xing, Y.; Wang, Z.; Liu, L.; Xu, Y.; Yang, Y.; Liu, S.; Zhou, F.; He, S.; Li, C. Defects and Failure Types of Solid Insulation in Gas-Insulated Switchgear: In Situ Study and Case Analysis. High Voltage 2022, 7, 158–164. [Google Scholar] [CrossRef]
  5. Feng, Z.; Guo, L.; Huang, D.; Li, R. Electrical Insulator Defects Detection Method Based on YOLOv5. In Proceedings of the 2021 IEEE 10th Data Driven Control and Learning Systems Conference (DDCLS), Suzhou, China, 14–16 May 2021; IEEE: Piscataway, NJ, USA, 2021; pp. 979–984. [Google Scholar]
  6. Liu, L.; Huang, K.; Bai, Y.; Zhang, Q.; Li, Y. Real-Time Detection Model of Electrical Work Safety Belt Based on Lightweight Improved YOLOv5. J. Real-Time Image Process. 2024, 21, 151. [Google Scholar] [CrossRef]
  7. Nguyen, R. Object Detection with YOLOv5 for Electric Utility Asset Inspection Using UAVs. Master’s Thesis, California State University, Long Beach, CA, USA, 2023. [Google Scholar]
  8. Pérez-Aguilar, D.A.; Pérez-Aguilar, J.M.; Pérez-Aguilar, A.P.; Risco-Ramos, R.H.; Malpica-Rodríguez, M.E. Electric Substation Inspection: YOLOv5 in Hotspot Detection Through Thermal Imaging. Ingenius 2024, 31, 43–54. [Google Scholar] [CrossRef]
  9. Li, Y.; Ni, M.; Lu, Y. Insulator Defect Detection for Power Grid Based on Light Correction Enhancement and YOLOv5 Model. Energy Rep. 2022, 8, 807–814. [Google Scholar] [CrossRef]
  10. Liang, F.; Zhao, L.; Ren, Y.; Wang, S.; To, S.; Abbas, Z.; Islam, M.S. LAD-Net: A lightweight welding defect surface non-destructive detection algorithm based on the attention mechanism. Comput. Ind. 2024, 161, 104109. [Google Scholar] [CrossRef]
  11. Li, C.; Hua, Y.; Liu, Y.; Liu, K.; Zhang, S. Intelligent Identification Method of Insulation Pull Rod Defects Based on Intactness-Aware Mosaic Data Augmentation and Fusion of YOLOv5s. High Voltage 2024, 9, 1171–1182. [Google Scholar] [CrossRef]
  12. Nisa, U. Image Augmentation Approaches for Small and Tiny Object Detection in Aerial Images: A Review. Multimed. Tools Appl. 2024. [Google Scholar] [CrossRef]
  13. Teterja, D.; Garcia Rodriguez, J.; Azorin-Lopez, J.; Sebastian-Gonzalez, E.; Nedic, D.; Lekovic, D.; Knezevic, P.; Drajic, D.; Vukobratovic, D. A Video Mosaicing-Based Sensing Method for Chicken Behavior Recognition on Edge Computing Devices. Sensors 2024, 24, 3409. [Google Scholar] [CrossRef] [PubMed]
  14. Li, Z.; Li, Y.; Li, H.; Deng, L.; Yan, R. Surround Sensing Technique for Trucks Based on Multi-Features and Improved Yolov5 Algorithm. Sensors 2024, 24, 2112. [Google Scholar] [CrossRef] [PubMed]
  15. Dai, Y.; Chen, G.; Zhang, Y.; Zhu, Y.; Wang, Z.; Li, W.; Wang, C.; Tu, Y. Electric Field Distribution of Micro Surface Scratch Defects in GIS Insulation Pull Rods. In Proceedings of the 2024 IEEE International Conference on High Voltage Engineering and Applications (ICHVE), Berlin, Germany, 18–22 August 2024; IEEE: Piscataway, NJ, USA, 2024; ISBN 979-8-3503-7498-8. [Google Scholar]
  16. Kim, J.-H.; Kim, N.; Park, Y.W.; Won, C.S. Object Detection and Classification Based on YOLO-V5 with Improved Maritime Dataset. J. Mar. Sci. Eng. 2022, 10, 377. [Google Scholar] [CrossRef]
  17. Hasib, K.M.; Iqbal, M.S.; Shah, F.M.; Mahmud, J.A.; Popel, M.H.; Showrov, M.I.; Ahmed, S.; Rahman, O. A Survey of Methods for Managing the Classification and Solution of Data Imbalance Problem. J. Comput. Sci. 2020, 16, 1546–1557. [Google Scholar] [CrossRef]
  18. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Realtime Object Detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; IEEE: Las Vegas, NV, USA, 2016; pp. 779–788. [Google Scholar]
  19. Bochkovskiy, A.; Wang, C.Y.; Liao, H.Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  20. Wang, X.; Wang, X.; Hu, C.; Dai, F.; Xing, J.; Wang, E.; Du, Z.; Wang, L.; Guo, W. Study on the Detection of Defoliation Effect of an Improved YOLOv5x Cotton. Agriculture 2022, 12, 1583. [Google Scholar] [CrossRef]
  21. Ioffe, S.; Szegedy, C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
  22. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed]
  23. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; IEEE: Salt Lake City, UT, USA, 2018. [Google Scholar]
  24. Rezatofighi, H.; Tsoi, N.; Gwak, J.Y.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; IEEE: Long Beach, CA, USA, 2019. [Google Scholar]
  25. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  26. Qiu, S.; Wen, G.; Deng, Z.; Liu, J.; Fan, Y. Accurate Non-Maximum Suppression for Object Detection in High-Resolution Remote Sensing Images. Remote Sens. Lett. 2018, 9, 237–246. [Google Scholar] [CrossRef]
  27. Rezazadeh, N.; Perfetto, D.; Polverino, A.; De Luca, A.; Lamanna, G. Guided wave-driven machine learning for damage classification with limited dataset in aluminum panel. Struct. Health Monit. 2024, 1–24. [Google Scholar] [CrossRef]
Figure 1. Structural diagram of defect detection model of insulating tie rod based on YOLOv5s.
Figure 1. Structural diagram of defect detection model of insulating tie rod based on YOLOv5s.
Sensors 25 01209 g001
Figure 2. Detection process of YOLOv5s network.
Figure 2. Detection process of YOLOv5s network.
Sensors 25 01209 g002
Figure 3. Illustration of the Copy–Paste Data Augmentation Method. The area marked with a red box indicates the defect target area.
Figure 3. Illustration of the Copy–Paste Data Augmentation Method. The area marked with a red box indicates the defect target area.
Sensors 25 01209 g003
Figure 4. Correction to bounding boxes of hair-like impurities.
Figure 4. Correction to bounding boxes of hair-like impurities.
Sensors 25 01209 g004
Figure 5. Comparison of mAP@0.5:0.95 Performance and P/R/F1 score performance across different data augmentation methods. (a) Comparison of mAP@0.5:0.95 performance across different data augmentation methods; (b) Comparison of P/R/F1 score performance across different data augmentation methods.
Figure 5. Comparison of mAP@0.5:0.95 Performance and P/R/F1 score performance across different data augmentation methods. (a) Comparison of mAP@0.5:0.95 performance across different data augmentation methods; (b) Comparison of P/R/F1 score performance across different data augmentation methods.
Sensors 25 01209 g005
Figure 6. Comparison of detection results on real photos: (1a) Pull Rod 1, Dataset 1; (1b) Pull Rod 1, Dataset 5; (1c) Pull Rod 1, Dataset 4; (1d) Pull Rod 1, Dataset 7; (2a) Pull Rod, Dataset 1; (2b) Pull Rod 2, Dataset 5; (2c) Pull Rod 2, Dataset 4; (2d) Pull Rod 2, Dataset 7; (3a) Pull Rod 3, Dataset 1; (3b) Pull Rod 3, Dataset 5; (3c) Pull Rod 3, Dataset 4; (3d) Pull Rod 3, Dataset 7.
Figure 6. Comparison of detection results on real photos: (1a) Pull Rod 1, Dataset 1; (1b) Pull Rod 1, Dataset 5; (1c) Pull Rod 1, Dataset 4; (1d) Pull Rod 1, Dataset 7; (2a) Pull Rod, Dataset 1; (2b) Pull Rod 2, Dataset 5; (2c) Pull Rod 2, Dataset 4; (2d) Pull Rod 2, Dataset 7; (3a) Pull Rod 3, Dataset 1; (3b) Pull Rod 3, Dataset 5; (3c) Pull Rod 3, Dataset 4; (3d) Pull Rod 3, Dataset 7.
Sensors 25 01209 g006
Table 1. The specific quantity and proportion of the five types of defects.
Table 1. The specific quantity and proportion of the five types of defects.
Defect TypeQuantityProportion
White spot48358.2%
Impurity13816.6%
Bubble404.8%
Crack617.4%
Scratch10813.0%
Table 2. Composition of datasets and preprocessing methods.
Table 2. Composition of datasets and preprocessing methods.
DatasetCopy–PasteHair-like Impurities Bounding Box CorrectionMosaic with Hair-like Impurities Bounding Box CorrectionMosaic without Hair-like Impurities Bounding Box Correction
Dataset 1Adopted/Not Adopted×××
Quantity0/08000
Dataset 2Adopted/Not Adopted××
Quantity2000/06000
Dataset 3Adopted/Not Adopted×
Quantity2000/06000
Dataset 4Adopted/Not Adopted×
Quantity2000/60000
Dataset 5Adopted/Not Adopted××
Quantity4000/04000
Dataset 6Adopted/Not Adopted×
Quantity4000/04000
Dataset 7Adopted/Not Adopted×
Quantity4000/40000
Table 3. Comparison of training results for different data augmentation methods.
Table 3. Comparison of training results for different data augmentation methods.
DatasetWhite SpotImpurityBubbleCrackScratchAll
Dataset 1Number of instances38,85511,23431864769871466,758
mAP@0.50.9580.6640.5310.9070.9280.797
mAP@0.5:0.950.7960.4370.3290.7520.7750.618
Precision0.9480.8940.8880.9660.9620.931
Recall0.9130.5550.4360.8550.8770.727
F1 Score0.9300.6850.5850.9070.9180.816
Dataset 2Number of instances30,02110,40247403520664155,324
mAP@0.50.9550.6590.770.9040.9330.844
mAP@0.5:0.950.7740.4280.5310.7230.7460.641
Precision0.9480.9210.9710.9590.9690.954
Recall0.9070.5360.70.8410.8730.771
F1 Score0.9270.6780.8140.8960.9180.853
Dataset 5Number of instances20,858966664242340449443,782
mAP@0.50.950.7770.9040.910.9430.897
mAP@0.5:0.950.7380.5930.7660.7020.7370.707
Precision0.9540.9240.9810.9560.9660.956
Recall0.9030.6950.8530.8530.8740.836
F1 Score0.9280.7930.9130.9020.9180.892
Table 4. Comparison of training results for datasets with and without hair-like impurity bounding box correction (Datasets 2, 3 and 4).
Table 4. Comparison of training results for datasets with and without hair-like impurity bounding box correction (Datasets 2, 3 and 4).
DatasetWhite SpotImpurityBubbleCrackScratchAll
Dataset 2Number of instances30,02110,40247403520664155,324
mAP@0.50.9550.6590.770.9040.9330.844
mAP@0.5:0.950.7740.4280.5310.7230.7460.641
Precision0.9480.9210.9710.9590.9690.954
Recall0.9070.5360.70.8410.8730.771
F1 Score0.9270.6780.8140.8960.9180.853
Dataset 3Number of instances29,68411,00447283520663255,748
mAP@0.50.9560.7260.7870.9130.9340.863
mAP@0.5:0.950.7750.5110.6350.7370.7650.684
Precision0.9530.9170.9430.9640.9710.95
Recall0.9090.6260.7170.8580.8710.796
F1 Score0.9300.7440.8150.9080.9180.866
Dataset 4Number of instances29,38912,24549663591673957,380
mAP@0.50.9910.9220.9250.9920.9910.964
mAP@0.5:0.950.8170.660.7330.8350.8370.777
Precision0.9690.9310.9370.9780.9810.959
Recall0.9640.8510.8330.9570.9620.913
F1 Score0.9660.8890.8820.9670.9710.935
Table 5. Comparison of training results for datasets with and without hair-like impurity bounding box correction (Datasets 5, 6, and 7).
Table 5. Comparison of training results for datasets with and without hair-like impurity bounding box correction (Datasets 5, 6, and 7).
DatasetWhite SpotImpurityBubbleCrackScratchAll
Dataset 5Number of instances20,858966664242340449443,782
mAP@0.50.950.7770.9040.910.9430.897
mAP@0.5:0.950.7380.5930.7660.7020.7370.707
Precision0.9540.9240.9810.9560.9660.956
Recall0.9030.6950.8530.8530.8740.836
F1 Score0.9280.7930.9130.9020.9180.892
Dataset 6Number of instances20,66110,83064002340447644,707
mAP@0.50.950.8050.8990.910.9390.9
mAP@0.5:0.950.7440.6260.7710.7030.7370.716
Precision0.9670.9410.9870.970.9790.969
Recall0.8910.7150.8490.8330.8590.83
F1 Score0.9270.8130.9130.8960.9150.894
Dataset 7Number of instances20,58311,65965652309452245,638
mAP@0.50.990.9460.9690.9910.9850.976
mAP@0.5:0.950.7980.7310.8160.7990.8040.789
Precision0.9710.9340.9810.9760.9810.969
Recall0.9580.8930.9190.9460.9480.933
F1 Score0.9640.9130.9490.9610.9640.951
Table 6. The inference time and GPU memory consumption of each dataset.
Table 6. The inference time and GPU memory consumption of each dataset.
DatasetInference Time (ms)GPU Memory Consumption
Dataset 122.97.35 G
Dataset 223.07.34 G
Dataset 322.57.35 G
Dataset 425.17.35 G
Dataset 523.67.35 G
Dataset 623.47.35 G
Dataset 721.57.34 G
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, X.; Cong, M.; Liu, B.; Fan, X.; Qin, W.; Liang, F.; Li, C.; He, J. A Preprocessing Method for Insulation Pull Rod Defect Dataset Based on the YOLOv5s Object Detection Network. Sensors 2025, 25, 1209. https://doi.org/10.3390/s25041209

AMA Style

Li X, Cong M, Liu B, Fan X, Qin W, Liang F, Li C, He J. A Preprocessing Method for Insulation Pull Rod Defect Dataset Based on the YOLOv5s Object Detection Network. Sensors. 2025; 25(4):1209. https://doi.org/10.3390/s25041209

Chicago/Turabian Style

Li, Xuetong, Meng Cong, Bo Liu, Xianhao Fan, Weiqi Qin, Fangwei Liang, Chuanyang Li, and Jinliang He. 2025. "A Preprocessing Method for Insulation Pull Rod Defect Dataset Based on the YOLOv5s Object Detection Network" Sensors 25, no. 4: 1209. https://doi.org/10.3390/s25041209

APA Style

Li, X., Cong, M., Liu, B., Fan, X., Qin, W., Liang, F., Li, C., & He, J. (2025). A Preprocessing Method for Insulation Pull Rod Defect Dataset Based on the YOLOv5s Object Detection Network. Sensors, 25(4), 1209. https://doi.org/10.3390/s25041209

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop