Next Article in Journal
Steelmaking Predictive Analytics Based on Random Forest and Semantic Reasoning
Next Article in Special Issue
A Large-Class Few-Shot Learning Method Based on High-Dimensional Features
Previous Article in Journal
Developing an Intelligent Recommendation System for Non-Information and Communications Technology Major University Students
Previous Article in Special Issue
Three-Dimensional Human Pose Estimation with Spatial–Temporal Interaction Enhancement Transformer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Foreign-Object Detection in High-Voltage Transmission Line Based on Improved YOLOv8m

1
School of Information Science and Engineering, Yunnan University, Kunming 650504, China
2
Electric Power Research Institute, Yunnan Power Grid Co., Ltd., Kunming 650214, China
3
Yunnan Key Laboratory of Intelligent Systems and Computing, Kunming 650504, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(23), 12775; https://doi.org/10.3390/app132312775
Submission received: 24 October 2023 / Revised: 20 November 2023 / Accepted: 24 November 2023 / Published: 28 November 2023
(This article belongs to the Special Issue Deep Learning and Computer Vision for Object Recognition)

Abstract

:
The safe operation of high-voltage transmission lines ensures the power grid’s security. Various foreign objects attached to the transmission lines, such as balloons, kites and nesting birds, can significantly affect the safe and stable operation of high-voltage transmission lines. With the advancement of computer vision technology, periodic automatic inspection of foreign objects is efficient and necessary. Existing detection methods have low accuracy because foreign objects attached to the transmission lines are complex, including occlusions, diverse object types, significant scale variations, and complex backgrounds. In response to the practical needs of the Yunnan Branch of China Southern Power Grid Co., Ltd., this paper proposes an improved YOLOv8m-based model for detecting foreign objects on transmission lines. Experiments are conducted on a dataset collected from Yunnan Power Grid. The proposed model enhances the original YOLOv8m by incorporating a Global Attention Module (GAM) into the backbone to focus on occluded foreign objects, replacing the SPPF module with the SPPCSPC module to augment the model’s multiscale feature extraction capability, and introducing the Focal-EIoU loss function to address the issue of high- and low-quality sample imbalances. These improvements accelerate model convergence and enhance detection accuracy. The experimental results demonstrate that our proposed model achieves a 2.7% increase in mAP_0.5, a 4% increase in mAP_0.5:0.95, and a 6% increase in recall.

1. Introduction

High-voltage transmission lines play a crucial role in ensuring the safety and stability of electricity transmission [1]. With the rapid development of the Chinese economy, China’s power generation accounted for approximately 30.34% of the world’s total in 2022, making its high-voltage transmission line network the largest globally [2]. These transmission lines traverse diverse and dynamic environments, often passing through residential and commercial areas, making them susceptible to the attachment of various foreign objects such as balloons, kites, fabrics and plastic waste [3]. Additionally, in natural settings like forests, birds nesting and twigs can challenge the safety of the transmission lines [4]. Suppose foreign objects on the transmission lines are not promptly detected and cleared: in that case, they can lead to accidents such as single-phase grounding and inter-phase short circuits, severely affecting the regular operation of the transmission lines, jeopardizing human safety, and resulting in casualties like accidental electrocution and fatalities [5].
Early detection of foreign objects on transmission lines relied predominantly on manual inspections, a method characterized by high labor intensity and inefficiency. With the development of computer vision technology and the widespread application of unmanned aerial vehicles (UAVs), UAVs can capture the transmission lines’ images. Then, the images can be analyzed using computer vision techniques to automatically identify anomalies, presenting a significant advancement in line inspection methodologies [6,7].
Currently, target detection research on transmission lines has mainly focused on target detection of principal components, foreign-object detection, smoke detection, insulator defect detection, etc. Among them, foreign-object detection mainly focuses on four types of foreign objects, namely balloons, bird nests, kites, and plastic trash. However, in the plateau mountainous environment of Yunnan Province, China, slender twigs and large birds also seriously interfere with the regular operation of high-voltage transmission lines. Therefore, slender twigs and birds are added to our detection database. The existing methods cannot effectively solve the problems faced by the Yunnan Branch of China Southern Power Grid Co., Ltd., so our research has specific value.
This paper’s contributions are as follows:
(1)
Based on the data provided by the Yunnan Power Science Research Institute in the Yunnan Branch of China Southern Power Grid Co., Ltd., a foreign object dataset of power transmission lines was established. This dataset includes six types of objects: trash, twigs, nests, kites, birds, and balloons. In this database, foreign objects are occluded, with multiple types, large-scale changes, and complex backgrounds. The detection accuracy of existing methods is not high.
(2)
We propose a foreign-object-detection model for high-voltage transmission lines based on improved YOLOv8m. The experimental results show that our model has achieved high accuracy and can meet the needs of practical applications.

2. Related Work

2.1. Object-Detection Model

In 2014, Girshick et al. [8] used the “region proposal + convolutional neural network” to replace the traditional method of “sliding window + manually designed features” in object detection, and designed the R-CNN framework, which made a huge breakthrough in object-detection technology. In 2015, Girshick et al. [9] proposed the Fast RCNN algorithm, which further improved RCNN and innovatively proposed multi task loss. They also trained classifiers and bounding box regressors, achieving end-to-end training in the detection stage, greatly improving accuracy and speed. Ren et al. [10] proposed the Faster RCNN algorithm shortly after, introducing an RPN network, making candidate box generation almost cost free. In 2016, Redmon et al. [11] proposed the YOLO algorithm, which is the first single-stage object-detection algorithm in the era of deep learning, and its speed is very fast. This network divides the image into grids and predicts the bounding box and classification probability of each grid area. A single neural network can obtain results from the complete image through one operation, which is conducive to end-to-end optimization of detection performance. In the same year, combining the anchor mechanism of RCNN and the regression idea of YOLO, Liu et al. [12] proposed the SSD algorithm, introducing multi-scale detection methods, and performing detection on feature maps extracted at each scale. Lin et al. [13] proposed RetinaNet in 2017 to investigate the accuracy of single-stage detection methods that lag behind two-stage detection methods. They believe that the imbalance of categories during the training process leads to a disadvantage in the accuracy of single-stage methods. Therefore, they propose Focal Loss to replace traditional cross entropy, improve the weight of background samples, and make the model more inclined towards target samples that are difficult to detect during the training process.
In 2017 and 2018, Redmon et al. launched subsequent versions of YOLO: YOLOv2 [14] and YOLOv3 [15]. The YOLOv2 algorithm introduces BN (batch normalization), multi-scale training, anchor box mechanism, and fine-grained features to improve the YOLOvl algorithm. On the basis of YOLOv2, the YOLOv3 algorithm adopts better backbone network, multi-scale prediction, and nine anchor boxes for detection, which improves the accuracy of the detection algorithm while ensuring real-time performance. In 2020, Bochkovskiy A et al. [16] proposed the YOLOv4 model. In the same year, Glenn et al. [17] proposed YOLOv5, which greatly reduced the number of parameters in the YOLO model, but its accuracy was comparable to YOLOv4. In 2021, Ge Zheng et al. [18] proposed the YOLOX model. In the same year, Wang et al. [19] proposed YOLOR. In 2022, Meituan [20] proposed YOLOv6, which was inspired by the design concept of hardware-aware neural networks. Based on the RepVGG style [21], it designed the reconfigurable and more efficient backbone networks, EfficientRep Backbone and Rep-PAN Neck; it optimized and designed a more concise and effective Efficient Decoupled Head, which further reduces the additional delay overhead caused by general decoupling heads while maintaining accuracy; and the anchor-free mode was adopted, supplemented by the SimOTA [22] label allocation strategy and SIoU [23] boundary box regression loss, to further improve detection accuracy. In the same year, Wang, C, et al. [24] proposed YOLOv7, proposing different network models for different devices. The concept of a gradient propagation path was used to analyze the reparameterization strategies applicable to different layers of networks, and a planned model structure reparameterization was proposed; the author proposes a new label allocation method called the Coarse to Fine (Coarse to Fine) guided label allocation strategy, which allows Aux Head to generate more grids as positive samples and reduce information loss. In terms of model design architecture, E-ELAN was proposed, which enhances the network feature’s learning ability through amplification, confusion, and merging without changing the original gradient path of ELAN [25]. RepVGG style can improve performance through multiple branches during the training process, and inference speed can be accelerated through structural reparameterization. In 2023, Ultratics [26] released YOLOv8, which may have referenced the design concept of YOLOv7 ELAN in its backbone network and Neck section. YOLOv5’s C3 structure was replaced with a C2f structure with richer gradient flow, and different channel numbers were adjusted for different scale models, significantly improving model performance; the Head section has been replaced with the current mainstream decoupling head structure, separating classification and detection heads, and also replaced with Anchor Free from Anchor Based. In terms of loss calculation, the TaskAlignedAssigner positive sample allocation strategy was adopted, and the Distribution Focal Loss was introduced. The data augmentation part of the training introduces the last 10 epochs in YOLOX to turn off Mosiac augmentation, which can effectively improve accuracy.
YOLOv8m is one of the latest models in object detection, exhibiting advantages in detection accuracy and speed. Because the foreign-object-detection model for high-voltage transmission lines will be integrated into the hardware of the drone platform in the later stage, we are considering using the YoloV8m lightweight model and modifying it to make it more suitable for our application.

2.2. Research on Target Detection for Transmission Lines

In the early stages of employing computer vision for detecting foreign objects on transmission lines, traditional digital-image processing techniques were often used, employing non-neural network-based methods to extract defect features from transmission line images [27]. For instance, Cai et al. [28] employed contour detection of the transmission conductor’s shape to determine the presence of foreign objects. Jin et al. [29] proposed an Otsu adaptive threshold segmentation algorithm, improved with morphology, to remove background noise, and applied a gradient method to identify power line edge positions, utilizing the Hough transform to analyze the number of lines to achieve foreign object recognition. Wang et al. [30] introduced a detection method for foreign objects on transmission lines based on line structure perception, segmenting and analyzing the grayscale and line width of acquired line segments in horizontal and vertical directions, to identify foreign objects on the transmission lines within complex backgrounds. However, the rich and diverse details in images of foreign objects on transmission lines make it challenging to describe target features using manually designed single features comprehensively. These methods exhibit relatively low accuracy in identifying foreign objects and struggle to differentiate between anomalies.
With the rapid advancement of deep learning, methodologies for detecting foreign objects on power transmission lines based on deep learning have been proposed. Shi et al. [31] initially employed the Selective Search method to extract candidate regions from pole images. Subsequently, training was conducted using the CaffeNet network model, adjusting and optimizing samples and network parameters through pre-training and re-training, resulting in a detection accuracy of 92.46% for nests. Zhu et al. [32] processed images using the scale histogram-matching method and employed the DFB-NN model, which effectively integrates low-resolution information and high-semantic information, to detect various objects, including vehicle-mounted cranes, tower cranes, forklifts, winding foreign objects, and wildfires, achieving an accuracy of 88.1%. Yang et al. [33] utilized the DenseNet network to replace the penultimate layer network in YOLO v3, achieving average detection accuracies of 94.7% for kites, nests, and trash. Yu et al. [34] used Otsu adaptive threshold segmentation, morphological processing and other methods to extract regions of interest. Then, DenseNet201 was used to extract the features of the regions of interest. Finally, the ECOC-SVM model was trained, and the average accuracy of balloons, kites, plastics and nests reached 93.3%. Yu et al. [35] employed a concatenated fusion strategy to merge features extracted by different networks. They utilized a random-forest classification model to achieve a detection accuracy of 95.8% for balloons, kites, nests, and plastics. Zou et al. [36] replaced the FPN structure of YOLOv5 with the BiFPN structure, integrated the CA attention mechanism into the CSP2_X module, and added a small-object detection layer, resulting in an average detection accuracy of 98.3% for nests, plastics, and kites. Qiu et al. [37] proposed the YOLOv4-EDAM model, utilizing the DnCNN image denoising network to reduce image noise and employing SENet-MobileNetV2 to extract features. Additionally, the model enhanced the network’s feature extraction capability using CBAM-SPP and CBAM-PANet, and utilized NMS to eliminate overlapping boxes, achieving a detection accuracy of 96.71% for nests, kites, balloons and trash. Wu et al. [38] used the ASPP module in YOLOX, incorporated the CBAM attention mechanism, employed the GIOU loss function, and achieved a detection accuracy of 86.57% for nests, balloons, trash and kites. Yu et al. [39] introduced SPD convolution into the YOLOv7 model, optimizing hyperparameters to achieve a detection accuracy of 92.2% for cranes, excavators, bulldozers, tower cranes, trucks and nests. Zhang et al. [40] made improvements based on the YOLOv4 model, generating anchor boxes suitable for this dataset using k-means clustering, replacing the network’s SPP module with an SPPF module, and substituting the SiLU activation function for the Leaky ReLU activation function, achieving an average detection accuracy of 97.57% for balloons, nests, kites, and trash. Tang et al. [41] combined Transformer V2 with YOLOX, utilizing the STCSP feature extraction layer in the backbone network and employing the Hybrid Spatial Pyramid Pooling (HSPP) module and the RepVGGBlock. This approach achieved a detection accuracy of 96.7% for nests, kites, and balloons. In research on transmission line components and environment, Chen et al. [42] improved a YOLOv5s model for key-component detection of power transmission lines. Cheng et al. [43] proposed a detection algorithm called P2E-YOLOv5 to detect a small target. Chen et al. [44] proposed an optimized YOLOv7-tiny model for smoke detection in power transmission lines.
The existing foreign-object detection mainly focuses on four types of foreign objects, namely balloons, bird nests, kites, and plastic trash. However, in the plateau mountainous environment of Yunnan Province, China, slender twigs and large birds also seriously interfere with the regular operation of high-voltage transmission lines. Therefore, slender twigs and birds are added to our detection database. The existing methods cannot effectively solve the problems faced by the Yunnan Branch of China Southern Power Grid Co., Ltd.

3. Datasets

The datasets of foreign objects on power transmission lines utilized in this study were provided by the Electric Power Science Research Institute in Yunnan Branch of China Southern Power Grid Co., Ltd. The image data was collected using various drones, resulting in significant variations in pixel dimensions, ranging from 640 × 640, 856 × 486, 1280 × 720 to 2738 × 2270. The sample images of foreign objects on power transmission lines were categorized into six classes: trash, twig, nest, kite, bird and balloon. Figure 1a represents trash, Figure 1b depicts twig, Figure 1c displays nest, Figure 1d exhibits kite, Figure 1e portrays bird, and Figure 1f illustrates balloon. It is essential to note that the images shown in Figure 1 are not original. Because the foreign objects are small, proportionally, in some original images, we extracted the partial images around foreign objects for a more explicit representation.
Due to detailed latitude and longitude information in some sample images, we have covered this information with black blocks according to data security requirements.
The overall images for each category are presented in Figure 1. It is evident from Figure 1 that, due to varying shooting angles, the images of foreign objects on power transmission lines exhibit challenges such as occlusions, diverse categories of foreign objects, considerable variations in object scales, and complex backgrounds. These factors significantly augment the difficulty in detecting foreign objects on power transmission lines.
Given the infrequent occurrence of foreign objects on power transmission lines, the samples obtained are limited. In the original dataset, the respective counts of actual labels for trash, twig, nest, kite, bird and balloon are 667, 64, 2508, 240, 1244, and 356, totaling 3774 images in the dataset. These samples are insufficient for training neural networks, necessitating dataset augmentation.
As depicted in Figure 2, below, we augmented the datasets through rotations, variations in brightness, salt-and-pepper noise, and occlusions, resulting in an expanded dataset comprising 11,323 images. Following augmentation, the counts of actual labels for trash, twig, nest, kite, bird, and balloon are 2024, 1299, 4017, 2133, 2489, and 1811, respectively. The quantities per category are detailed in Table 1.
In subsequent experiments, to assess the effectiveness of the improved model, we randomly divided both the original and augmented dataset into training set, validation set, and test sets in a 6:2:2 ratio, for experimentation with the proposed model.

4. Proposed Method

Addressing the challenges posed by occlusions, diverse object categories, significant variations in object scales, and complex backgrounds in detecting foreign objects on power transmission lines, this paper proposes an improved model for foreign-object detection based on the YOLOv8m framework. A GAM attention mechanism is incorporated into the backbone to enhance the model’s focus on the foreign object, to mitigate their occlusion. In response to the challenge of multiple object categories and intricate backgrounds, the SPPF module is replaced with the SPPCSPC module to augment the model’s multi-scale feature-extraction capability. A Focal-EIOU loss function is introduced to address the imbalance in high- and low-quality samples, thereby expediting model convergence and elevating detection precision. These enhancements are elaborated sequentially in Section 4.2, Section 4.3 and Section 4.4.

4.1. Enhanced YOLOv8m Model

The improved YOLOv8m model is illustrated in Figure 3, and the enhancements are outlined within the red boxes, elucidated as follows.
In Figure 3, the GAM attention mechanism is introduced into the backbone, a feature detailed in Section 4.2. The SPPF module is replaced by the SPPCSPC module, a modification discussed in Section 4.3. The integration of the Focal-EIoU loss function is presented in Section 4.4.

4.2. GAM (Global Attention Mechanism)

Specifically addressing the issue of occlusion in detecting foreign objects on power transmission lines, we integrate the GAM attention mechanism into the backbone to enhance the model’s focus on the foreign object.
The Global Attention Mechanism (GAM) represents a technique that enriches inter-dimensional interactions across channel and spatial dimensions by preserving channel and spatial information. This augmentation is pivotal in obtaining critical target information.
GAM adopts the channel-spatial sequence of CBAM [45] and entails a redesign of both the channel-attention submodule and the spatial-attention submodule, depicted in Figure 4. In Equation (1), M C represents the handling of the input vector F 1 by the channel-attention submodule, where is the tensor product of the processed vector and the input vector F 1 . In Equation (2), M S signifies the spatial-attention submodule’s processing of the vector F 2 , where is the tensor product of the processed vector and the input vector F 2 .
F 2 = M C ( F 1 ) F 1
F 3 = M S ( F 2 ) F 2
In Figure 5, we delineate the specific processing procedure of the channel-attention submodule depicted in Figure 4. The channel-attention submodule employs a 3D permutation (dimensional transformation) to retain three-dimensional information. Subsequently, a two-layer Multi-Layer Perceptron (MLP) enhances inter-dimensional dependency relationships spanning channel and spatial dimensions. An MLP structure is an encoder–decoder architecture with a reducing factor of r.
In Figure 6, we delineate the specific processing procedure of the spatial-attention submodule depicted in Figure 4. The spatial-attention submodule employs two convolutions to fuse spatial information, to heighten focus on spatial information. Moreover, a reducing factor r is used within the channel-attention submodule to augment attention towards spatial features, as depicted in Figure 6.

4.3. SPPCSPC Module

In response to significant scale changes, multiple types and complex backgrounds of foreign objects, we have replaced the SPPF module with the SPPCSPC module, to augment the model’s ability to extract multi-scale features. The SPPCSPC module uses a multi-scale pyramid pooling (SPP) structure and conventional convolutional operations. This structure amplifies the model’s receptive field and elevates its capacity to extract features across multiple scales, enhancing overall detection capabilities.
The SPPCSPC module, shown in Figure 7, segregates features into two segments. One segment undergoes processing utilizing the SPP structure, employing four max-pooling operations at distinct scales, to capture diverse receptive fields. Consequently, this step amalgamates spatial feature information of varying dimensions and generates fixed-size feature vectors. In parallel, the other segment undergoes standard convolutional operations. Finally, the features obtained from these two segments are concatenated to form the comprehensive feature vector.

4.4. Focal-EIoU Loss Function

To address the imbalance between high- and low-quality samples (the intersection size of different regions on the union of predicted-truth boxes and ground-truth boxes), we add a Focal-EIou loss function. The Focal-EIoU loss function enhances the optimization contributions for high-quality samples with lower quantities when regressing the predicted bounding boxes towards the ground-truth boxes, while simultaneously diminishing the optimization contributions for low-quality samples with higher quantities in this regression process. The Focal-EIoU loss function can effectively address the imbalance between high- and low-quality samples, expediting model convergence and augmenting detection precision.
The EIoU loss function, shown in Figure 8, is represented by Equation (3). The EIoU is a metric for evaluating the dissimilarity between predicted boxes and ground-truth boxes. The Focal-EIoU loss function utilizes this metric to guide the training process, emphasizing high-quality samples for more effective model learning and convergence.
L E I o U = L I o U + L d i s + L a s p = 1 I o U ( A , B ) + ρ 2 ( b , b g t ) ( h c ) 2 + ( w c ) 2 + ρ 2 ( h , h g t ) h c + ρ 2 ( w , w g t ) w c
As illustrated in Equation (3) and Figure 8 above, where 1 I o U ( A , B ) represents the loss function for the predicted box and the ground-truth box, ρ ( ) signifies the Euclidean distance, b and b g t denote the centers of the predicted and ground-truth boxes, respectively, h c and w c denote the height and width of the bounding box, h and h g t denote the height of the predicted and ground-truth boxes, and w and w g t denote the widths of the predicted and ground-truth boxes, respectively.
The EIoU loss function consists of three components: the Intersection over Union (IoU) loss between the predicted and ground-truth boxes, the center-distance loss between the predicted and ground-truth boxes, and the edge-length loss between the predicted and ground-truth boxes. The EIoU loss function directly minimizes the differences in width and height between the predicted and ground-truth bounding boxes, expediting model convergence and enhancing localization accuracy.
After adding Focal to EIoU, EIoU focuses more on high-quality samples (large Intersection over Union between predicted and ground-truth boxes). The comprehensive formula for Focal-EIoU is given by Equation (4).
L F o c a l - E I o U = I o U γ L E I o U
where I o U represents the Intersection over Union for predicted box A and ground-truth box B ( I o U = ( A B ) / ( A B ) ), γ is a parameter controlling the degree of suppression for low-quality samples (small Intersection over Union between predicted and ground-truth boxes), and L E I o U represents the EIoU loss function.

5. Experiment and Analysis

5.1. Experimental Setup

The hardware configuration for our experiment included an Intel(R) Core(TM) i7–10700K CPU, 32 GB of RAM, and an NVIDIA GeForce RTX 2080 Ti GPU. The software environment comprised Windows 10, PyTorch 1.11.0, CUDA 11.3, and PyCharm Community Edition 2021.3.
The momentum was initialized to 0.937, the batch size was 24, the weight decay rate was 0.0005, the initial learning rate was 0.01, and the training epoch was 100. Given the varying sizes of the image samples, all input images were transformed to 640 × 640 pixels.

5.2. Evaluation Metrics

In object detection, Intersection over Union (IoU) represents the area ratio of the intersection to the union between the predicted box and the ground-truth box. TP (True Positive) denotes the sample count where the real category is positive and the model’s prediction is also positive; TN (True Negative) indicates the sample count where the real category is negative, and the model predicts them as negative; FP (False Positive) is the sample count where the real category is negative, but the model predicts them as positive; and FN (False Negative) represents the sample count where the real category is positive, but the model predicts them as negative.
Our ablation experiments used seven metrics: mAP_0.5, mAP_0.5:0.95, precision, recall, parameters, GFloats, and speed. Our comparative experiments used five metrics: mAP_0.5, mAP_0.5:0.95, precision, recall, and speed. The definitions of these metrics are as follows:
  • Precision: the proportion of predicted positive samples that are actually positive (Precision = TP/(TP + FP)).
  • Recall: the proportion of actual positive samples that are predicted as positive (Recall = TP/(TP + FN)).
  • mAP_0.5: precision is calculated for each class when the IoU threshold is 0.5, and the precision of each class is averaged to obtain mAP_0.5.
  • mAP_0.5:0.95: average mAP across different IoU thresholds from 0.5 to 0.95 (with a step of 0.05).
  • Parameter: the model’s parameter count.
  • GFLOPs: model computational complexity. One GFLOPs means that the model requires billions of floating-point operations.
  • Speed: the model’s detection speed. It indicates the milliseconds required for detecting a single image.

5.3. Experimental Results and Analysis

5.3.1. Model Training Comparison

In this experiment, the training epoch was 100. Figure 9 illustrates the comparison of mAP_0.5 during training, Figure 10 presents the comparison of mAP_0.5:0.95 during training, Figure 11 exhibits the comparison of precision during training, Figure 12 showcases the comparison of recall during training, and Figure 13 displays the comparison of cls_loss during training. From these figures, it is evident that our model outperforms the original model across various metrics.
The four curves at the beginning of “Original” are trained using the original dataset, and their models are the original model, incorporating the GAM attention mechanism, continuing to add the SPPCSCP module, and continuing to add the Focal-EIoU loss function, respectively. The other two curves, starting with “dataAugmentation”, are trained using the augmented dataset, and their models are the original model and our improved model, respectively.

5.3.2. Attention Visualization Comparison

Figure 14 compares attention visualization results between the YOLOv8m original model and the addition of the GAM attention mechanism. Incorporating the GAM attention mechanism into our model enhances the focus on the objects and diminishes attention toward the background. The GAM can improve detection confidence and precision.

5.3.3. Ablation Experiments

To demonstrate the effectiveness of the improvement steps, we conducted ablation experiments, and the results are presented in Table 2.
As shown in Table 2, different improved steps can improve mAP_0.5, mAP_0.5:0.95, and recall. After simultaneously incorporating the GAM attention mechanism, SPPCSPC, and Focal-EIoU loss function, the mAP_0.5 increased by 2.7%, mAP_0.5:0.95 improved by 4%, and recall increased by 6%. Remarkably, the recall showed the most significant improvement. In practical applications, secondary manual detection is often performed on detected images. A higher recall can effectively reduce the miss-detection rate, subsequently reducing the workload for detection personnel.

5.3.4. Comparative Experiments with Other Models

To validate the advancement of our improved model, we compared it with mainstream object-detection models, using the same dataset and dataset partition. The comparative experimental results are presented in Table 3.
As depicted in Table 3, one-stage methods generally perform better than two-stage models. Regarding mAP_0.5, mAP_0.5:0.95, and recall metrics, our model stands out, achieving the highest values at 95.5%, 80.4%, and 91.9%, respectively. While not the fastest in terms of speed, it ranks second, and in terms of precision, it holds the third position. Our model demonstrates superior performance compared to other detection models, notably achieving the highest values for mAP_0.5, mAP_0.5:0.95, and recall.

5.3.5. Foreign-Object Detection Result

Figure 15 illustrates the detection results of the original YOLOv8m model, where (a)–(f) correspond to trash, twig, nest, kite, bird and balloon, respectively. Figure 16 presents the detection results of the improved YOLOv8m model, with each subfigure corresponding to the respective categories shown in Figure 15.
In Figure 15 and Figure 16, we sequentially compare the detection results for each subfigure. In Figure 15a and Figure 16a, our improved model enhances the confidence in detecting trash targets. In Figure 15b and Figure 16b, our improved model eliminates duplicate predictions of different parts of the same twig and false detections of a nest, while also achieving a higher confidence in twig detection. In Figure 15c and Figure 16c, our improved model exhibits higher confidence in detecting an obscured nest. In Figure 15d and Figure 16d, our improved model avoids incomplete detection of kite targets and enhances confidence in detecting kite targets. In Figure 15e and Figure 16e, our improved model eliminates duplicate predictions of different parts of the same bird and improves confidence in bird target detection. Finally, in Figure 15f and Figure 16f, our improved model elevates the confidence in detecting balloon targets.

5.3.6. Detection Results for Different Classes

Table 4 presents a comparative analysis of different models in detecting various foreign objects.
From Table 4, it is evident that before data augmentation, our improved model (denoted as ‘Ours’) shows an overall increase in mean precision across all categories in terms of mAP_0.5, mAP_0.5:0.95, and recall when compared to the original YOLOv8m model. However, there is a decrease in precision. Specifically, for mAP_0.5, the ‘twig’ category exhibits the most significant improvement, with other categories also showing improvements or remaining relatively stable. Similarly, in mAP_0.5:0.95, ‘twig’ demonstrates the most substantial advancement, while the other categories also display enhancements or stability. Regarding precision, except for ‘twig’ and ‘balloon’, which experience improvements, the precision of other categories decreases. For recall, all categories witness an increase or remain stable.
Following data augmentation, a comparison between our improved model (‘Ours’) and the original YOLOv8m model reveals an overall increase in mean precision across all categories concerning mAP_0.5, mAP_0.5:0.95, precision, and recall. In mAP_0.5, the precision of each category rises. In mAP_0.5:0.95, the precision of each category shows improvement. Regarding precision, ‘nest’ experiences a decline, but the precision of other categories improves. In recall, ‘nest’ shows a decline, but the precision of other categories sees an increase.

6. Conclusions

This paper proposes a foreign-object-detection model for transmission lines based on improved YOLOv8m. Our model autonomously detects anomalies in aerial images of power transmission lines, reducing the workload associated with anomaly detection and enhancing inspection efficiency.
Addressing the challenges of foreign-object detection along power transmission lines, we use three key improvements: firstly, to mitigate issues related to occlusion, a Global Attention Mechanism (GAM) was introduced, enhancing focus on obscured targets. Secondly, in response to the diversity of foreign-object categories and complex backgrounds, the SPPCSPC module was integrated, boosting the model’s multi-scale feature-extraction capabilities. Thirdly, the Focal-EIoU loss function was introduced to address imbalances between high- and low-quality samples, expediting model convergence and augmenting detection accuracy. The refined YOLOv8m model exhibited significant enhancements, elevating mAP_0.5 from 92.8% to 95.5%, mAP_0.5:0.95 from 76.4% to 80.4%, and recall from 85.9% to 91.9%.
No publicly available datasets for power transmission line foreign-object detection have been provided in previously published literature, thus hindering experimentation with our enhanced model on alternative datasets. The dataset utilized in this study originates from the Electric Power Research Institute, Yunnan Power Grid Co., Ltd., featuring geolocation data in certain images. Upon removing location information and obtaining approval from the aforementioned organization, we plan to release this dataset publicly, facilitating future research in related domains.
Although the upgraded model demonstrates the potential to be applied in practical detection scenarios for foreign objects on power transmission lines, its detection speed is presently insufficient for direct application on terminal CPU devices. Subsequent work will focus on adopting lighter models, such as YOLOv8n, to enhance model detection speed while maintaining detection accuracy.

Author Contributions

Conceptualization, Z.W.; methodology, Z.W.; software, Z.W.; validation, Z.W.; formal analysis, Z.W., G.Y., H.Z., Y.M. (Yi Ma) and Y.M. (Yutang Ma); investigation, Z.W., G.Y., H.Z., Y.M. (Yi Ma) and Y.M. (Yutang Ma); resources, Z.W. and G.Y.; data curation, Z.W.; writing—original draft preparation, Z.W. and G.Y.; writing—review and editing, Z.W. and G.Y.; funding acquisition, G.Y., H.Z., Y.M. (Yi Ma) and Y.M. (Yutang Ma) All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Key R&D Projects of Yunnan Province (Grant No. 202202AD080004), the Natural Science Foundation of China (Grant No. 62061049, 12263008), the Yunnan Provincial Department of Science and Technology-Yunnan University Joint Special Project for Double-Class Construction (Grant No. 202201BF070001-005), and the Application and Foundation Project of the Yunnan Province (Grant No. 202001BB050032).

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data is not publicly available due to the following reason. It is provided by Yunnan Power Grid Co., Ltd. and can only be released after obtaining approval from Yunnan Power Grid Co., Ltd.

Acknowledgments

Thank you to Yunnan Power Grid Co., Ltd. for providing the foreign object dataset.

Conflicts of Interest

Author Zhenyue Wang, Yi Ma and Yutang Ma are employed by the Yunnan Power Grid Co., Ltd. The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Deng, Y.; Tang, Z.; Li, H.; Yang, Z. Real-time Rendering Algorithm Optimization for Large-scale Power Transmission Line Scenes Based on LOD. Comput. Mod. 2017, 257, 115–118. [Google Scholar]
  2. Liu, Z.; Liao, X.; Chen, J.; Jiao, H. A Comprehensive Review on Intelligent Processing of Visible Light Images for Power Overhead Line Inspection. Power Syst. Technol. 2020, 44, 1057–1069. [Google Scholar]
  3. Zhang, H.; Zhou, H.; Li, S.; Li, P. Improving YOLOv3’s Foreign Object Detection Method for Transmission Lines. J. Laser 2022, 43, 82–87. [Google Scholar]
  4. Wan, D.; Zhang, J.; Guo, X. A Transmission Line Foreign Object Detection Method Based on Visual Saliency Analysis. J. Telev. Technol. 2018, 42, 106–110. [Google Scholar]
  5. Yu, G.; Zou, Z.; Fu, X.; Peng, J.; Shi, C. Research on Foreign Object Detection on Transmission Lines Based on UAV Aerial Images. Jiangxi Sci. 2022, 40, 223–228. [Google Scholar]
  6. Liao, X.; Liu, Z.; Yan, Q. A Comprehensive Review on UAV-based Smart Inspection Technology for Power Transmission Lines. J. Fuzhou Univ. 2020, 48, 198–209. [Google Scholar]
  7. Huang, Z.; Wang, Y.; Wang, H.; Gao, C.; Bai, C. Unmanned Aerial Vehicle Intelligent Inspection System for Power Lines Based on Cloud-Fog Edge Heterogeneous Collaboration. China Electr. Power 2020, 53, 161–168. [Google Scholar]
  8. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  9. Girshick, R. Fast R-CNN. In Proceedings of the 2015 IEEE International Conference on Computer Vision and Pattern Recognition, Santiago, Chile, 7–13 December 2015; pp. 1440–1448. [Google Scholar]
  10. Ren, S.; He, k.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef]
  11. Redmon, J.; Divvala, S.; Cirshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015; pp. 779–788. [Google Scholar]
  12. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.; Berg, A. SSD:single shot multibox detector. In Proceedings of the European Conference on Computer Vision and Pattern Recognition, Amsterdam, The Netherlands, 10–16 October 2016; pp. 21–37. [Google Scholar]
  13. Lin, T.; Goyal, P.; Girshick, R.; He, K.; Dollar, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2980–2988. [Google Scholar]
  14. Redmon, J.; Farhadi, A. YOLO900:Better, faster, stronger. In Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  15. Redmon, J.; Farhadi, A. YOLOv3:An incremental improve. In Proceeding of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
  16. Bochkovskiy, A.; Wang, C.; Liao, H. Yolov4: Optimal speed and accuracy of object detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  17. Ultralytics. YOLOv5 Object Detection. Available online: https://github.com/ultralytics/yolov5 (accessed on 24 April 2020).
  18. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOX: Exceeding YOLO Series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  19. Wang, C.; Yeh, I.; Liao, H. You Only Learn One Representation: Unified Network for Multiple Tasks. arXiv 2021, arXiv:2105.04206. [Google Scholar]
  20. Li, C.; Li, L.; Jiang, H.; Weng, K.; Geng, Y.; Li, L.; Ke, Z.; Li, Q.; Cheng, M.; Nie, W.; et al. YOLOv6: A Single-Stage Object Detection Framework for Industrial Applications. arXiv 2022, arXiv:2209.02976. [Google Scholar]
  21. Ding, X.; Zhang, X.; Ma, N.; Han, J.; Ding, G.; Sun, J. RepVGG: Making VGG-style ConvNets Great Again. In Proceeding of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021; pp. 13728–13737. [Google Scholar]
  22. Gevorgyan, Z. SIoU Loss: More Powerful Learning for Bounding Box Regression. arXiv 2022, arXiv:2205.12740. [Google Scholar]
  23. Wang, C.; Bochkovskiy, A.; Liao, H. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceeding of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Oxford, UK, 15–17 September 2023; pp. 7464–7475. [Google Scholar]
  24. Wang, C.; Liao, H.; Yeh, I. Designing Network Design Strategies Through Gradient Path Analysis. arXiv 2022, arXiv:2211.04800. [Google Scholar]
  25. Ultralytics. YOLOv8 Object Detection. [Online]. Available online: https://ultralytics.com/yolov8 (accessed on 17 July 2023).
  26. Liu, Y.; Shao, Z.; Hoffmann, N. Global Attention Mechanism: Retain Information to Enhance Channel-Spatial Interactions. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Online, 19–25 June 2021. [Google Scholar]
  27. Wang, X.; Hu, J.; Sun, C.; Du, L. Online Monitoring of Ice Thickness on Transmission Lines using Image Edge Detection. High Volt. Appar. 2009, 45, 69–73. [Google Scholar]
  28. Cai, W.; Le, J.; Jin, C.; Liu, K. Real-Time Image-Identification-Based Anti-Manmade Misoperation System for Substations. IEEE Trans. Power Deliv. 2023, 27, 1748–1754. [Google Scholar] [CrossRef]
  29. Jin, L.; Yao, C.; Yan, S.; Zhang, W. Recognition of Foreign Objects on Transmission Lines based on Aerial Images. J. Tongji Univ. 2013, 41, 277–281. [Google Scholar]
  30. Wang, W.; Zhang, J.; Han, J.; Liu, L.; Zhu, M. Method for Detecting Line Breaks and Foreign Object Defects on Transmission Lines Based on Drone Images. J. Comput. Appl. 2015, 35, 2404–2408. [Google Scholar]
  31. Shi, P.; Zhang, C.; Zheng, X. Foreign Object Detection on High-voltage Poles based on Deep Learning. J. Chongqing Univ. Sci. Technol. 2020, 22, 83–87. [Google Scholar]
  32. Zhu, J.; Guo, Y.; Yue, F.; Yuan, H.; Rong, M. A Deep Learning Method to Detect Foreign Objects for Inspecting Power Transmission Lines. IEEE Access 2020, 8, 94065–94075. [Google Scholar] [CrossRef]
  33. Yang, J.; Qin, Z.; Pang, X.; He, Z.; Cui, C. A Method for Monitoring and Identifying Intrusions of Foreign Objects on Transmission Lines based on Deep Learning Networks. Power Syst. Prot. Control 2021, 49, 37–44. [Google Scholar]
  34. Yu, Y.; Qiu, Z.; Zhou, Y.; Zhu, X.; Wang, Q. Transmission Line Foreign Object Detection Based on Convolutional Neural Network and ECOC-SVM. Smart Power 2022, 50, 87–92+107. [Google Scholar]
  35. Yu, Y.; Qiu, Z.; Liao, H.; Wei, Z.; Zhu, X.; Zhou, Z. A Method Based on Multi-Network Feature Fusion and Random Forest for Foreign Objects Detection on Transmission Lines. Appl. Sci. 2022, 12, 4982. [Google Scholar] [CrossRef]
  36. Zou, H.; Jiao, L.; Zhang, Z.; Tang, B.; Liu, Z. An Improved YOLO Network for Detection of Small Foreign Objects on Transmission Lines. J. Nanjing Inst. Technol. 2022, 20, 7–14. [Google Scholar]
  37. Qiu, Z.; Zhu, X.; Liao, C.; Qu, W.; Yu, Y. A Lightweight YOLOv4-EDAM Model for Accurate and Real-time Detection of Foreign Objects Suspended on Power Lines. IEEE Trans. Power Deliv. 2023, 38, 1329–1340. [Google Scholar] [CrossRef]
  38. Wu, M.; Guo, L.; Chen, R.; Du, W.; Wang, J.; Liu, M.; Kong, X.; Tang, J. Improved YOLOX Foreign Object Detection Algorithm for Transmission Lines. Wirel. Commun. Mob. Comput. 2022, 2022, 5835693. [Google Scholar] [CrossRef]
  39. Yu, C.; Liu, Y.; Zhang, W.; Zhang, X.; Zhang, Y.; Jiang, X. Foreign Objects Identification of Transmission Line Based on Improved YOLOv7. IEEE Access 2023, 11, 51997–52008. [Google Scholar] [CrossRef]
  40. Zhang, Q.; Zhu, T.; Xiao, S.; Yang, Z.; Zeng, H. Detection of High Voltage Transmission Line Foreign Objects based on Improved YOLOv4 Algorithm. Appl. Sci. 2023, 50, 59–65. [Google Scholar]
  41. Tang, C.; Dong, H.; Huang, Y.; Han, T.; Fang, M.; Fu, J. Foreign object detection for transmission lines based on Swin Transformer V2 and YOLOX. Vis. Comput. 2023, 1432–2315. [Google Scholar] [CrossRef]
  42. Chen, C.; Yuan, G.; Zhou, H.; Ma, Y. Improved YOLOv5s model for key components detecion of power transmission lines. Math. Biosci. Eng. 2023, 20, 7738–7760. [Google Scholar] [CrossRef]
  43. Cheng, Q.; Yuan, G.; Chen, D.; Xu, B.; Chen, E.; Zhou, H. Transmission Lines Small-Target Detection Algorithm Research Based on YOLOv5. Appl. Sci. 2023, 13, 9386. [Google Scholar] [CrossRef]
  44. Chen, C.; Yuan, G.; Zhou, H.; Ma, Y. Optimized YOLOv7-tiny model for smoke detection in power transmission lines. Math. Biosci. Eng. 2023, 20, 19300–19319. [Google Scholar] [CrossRef]
  45. Woo, S.; Park, J.; Lee, J.-Y.; Kwoen, I.S. CBAM: Convolutional Block Attention Module. In Proceedings of the Springer Conference on European Conference on Computer Vision, Munich, Germany, 8–14 September 2018. [Google Scholar]
Figure 1. Collection of Foreign Object Image Datasets for Transmission Lines. (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Figure 1. Collection of Foreign Object Image Datasets for Transmission Lines. (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Applsci 13 12775 g001aApplsci 13 12775 g001b
Figure 2. Example of data augmentation. (a) original. (b) rotations. (c) variations in brightness. (d) salt-and-pepper noise. (e) occlusions.
Figure 2. Example of data augmentation. (a) original. (b) rotations. (c) variations in brightness. (d) salt-and-pepper noise. (e) occlusions.
Applsci 13 12775 g002aApplsci 13 12775 g002b
Figure 3. Improved YOLOv8m Model for Foreign-Object Detection in Transmission Lines.
Figure 3. Improved YOLOv8m Model for Foreign-Object Detection in Transmission Lines.
Applsci 13 12775 g003
Figure 4. GAM structure diagram.
Figure 4. GAM structure diagram.
Applsci 13 12775 g004
Figure 5. Channel-attention submodule diagram.
Figure 5. Channel-attention submodule diagram.
Applsci 13 12775 g005
Figure 6. Spatial-attention submodule diagram.
Figure 6. Spatial-attention submodule diagram.
Applsci 13 12775 g006
Figure 7. SPPCSPC Model diagram.
Figure 7. SPPCSPC Model diagram.
Applsci 13 12775 g007
Figure 8. EIoU loss function diagram.
Figure 8. EIoU loss function diagram.
Applsci 13 12775 g008
Figure 9. Comparison of mAP_0.5 during training.
Figure 9. Comparison of mAP_0.5 during training.
Applsci 13 12775 g009
Figure 10. Comparison of mAP_0.5:0.95 during training.
Figure 10. Comparison of mAP_0.5:0.95 during training.
Applsci 13 12775 g010
Figure 11. Comparison of precision during training.
Figure 11. Comparison of precision during training.
Applsci 13 12775 g011
Figure 12. Comparison of recall during training.
Figure 12. Comparison of recall during training.
Applsci 13 12775 g012
Figure 13. Comparison of cls_loss during training.
Figure 13. Comparison of cls_loss during training.
Applsci 13 12775 g013
Figure 14. Attention visualization comparison. (a) original images. (b) Attention visualization using YOLOv8m. (c) Attention visualization using GAM.
Figure 14. Attention visualization comparison. (a) original images. (b) Attention visualization using YOLOv8m. (c) Attention visualization using GAM.
Applsci 13 12775 g014aApplsci 13 12775 g014b
Figure 15. YOLOv8m original-model detection results and enlarged images (left column: original image of detection results, right column: enlarged image). (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Figure 15. YOLOv8m original-model detection results and enlarged images (left column: original image of detection results, right column: enlarged image). (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Applsci 13 12775 g015aApplsci 13 12775 g015bApplsci 13 12775 g015c
Figure 16. Improved YOLOv8m-model detection results and enlarged images. (left column: original image of detection results, right column: enlarged image). (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Figure 16. Improved YOLOv8m-model detection results and enlarged images. (left column: original image of detection results, right column: enlarged image). (a) trash. (b) twig. (c) nest. (d) kite. (e) bird. (f) balloon.
Applsci 13 12775 g016aApplsci 13 12775 g016bApplsci 13 12775 g016c
Table 1. Total Amount of Each Type of Data.
Table 1. Total Amount of Each Type of Data.
TypeTrashTwigNestKiteBirdBalloon
Original6676425082401244356
Data Augmentation202412994017213324891811
Table 2. Results of Ablation Experiment.
Table 2. Results of Ablation Experiment.
GAMSPPCSPCFocal-EIoUmAP_0.5 (%)mAP_0.5:0.95
(%)
Precision
(%)
Recall
(%)
ParametersGFLPOSSpeed
(ms)
92.876.498.485.925.9 M79.132.3
93.178.595.488.942.5 M85.634.5
93.573.7193.9689.2234 M92.433.7
94.1475.1294.689.925.9 M79.132.4
93.978.393.291.550.6 M98.935.9
93.7582.19890.8634 M92.433.9
94.8984.597.0391.0942.5 M85.634.8
95.580.496.791.950.6 M98.935.9
Table 3. Comparative Experiments with Other Models.
Table 3. Comparative Experiments with Other Models.
ModelmAP_0.5
(%)
mAP_0.5:0.95
(%)
Precision
(%)
Recall
(%)
Speed
(ms)
Faster RCNN86.764.981.983.180.9
Sparse RCNN9066.493.881.179.1
Cascade RCNN86.865.389.177.786.7
YOLOv5m91.968.39586.250.5
YOLOv6m88.6771.891.886.545.7
YOLOv79275.998.285.541.5
YOLOv8m92.876.498.485.932.3
Ours95.580.496.791.935.9
Table 4. Detection Results of Various Foreign Objects.
Table 4. Detection Results of Various Foreign Objects.
ModelForeign ObjectLabelsmAP_0.5 (%)mAP_0.5:0.95
(%)
Precision
(%)
Recall
(%)
YOLOv8mall100892.876.498.485.9
nest51498.785.999.493.3
trash13498.484.910092.5
kite4586.370.199.271.1
twig1476.137.993.264.3
bird24299.589.699.699.2
balloon5997.489.999.394.9
Oursall100895.580.496.791.9
nest51498.787.398.696.3
trash13498.788.498.294.8
kite4587.975.887.380
twig1490.448.397.185.7
bird24299.590.199.499.6
balloon5997.692.899.494.9
Data Augmentationall279897.486.498.694.1
nest100499.188.499.898
trash25593.986.197.890.5
kite45796.181.597.190.3
twig24699.382.398.497.1
bird4879990.498.997.2
balloon34996.989.799.491.7
Data Augmentation + Oursall27989989.599.696.8
nest100499.489.899.597.8
trash255989199.794.5
kite45798.187.198.892.3
twig24699.584.710097.5
bird48799.491.899.499
balloon34999.59310099.4
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Z.; Yuan, G.; Zhou, H.; Ma, Y.; Ma, Y. Foreign-Object Detection in High-Voltage Transmission Line Based on Improved YOLOv8m. Appl. Sci. 2023, 13, 12775. https://doi.org/10.3390/app132312775

AMA Style

Wang Z, Yuan G, Zhou H, Ma Y, Ma Y. Foreign-Object Detection in High-Voltage Transmission Line Based on Improved YOLOv8m. Applied Sciences. 2023; 13(23):12775. https://doi.org/10.3390/app132312775

Chicago/Turabian Style

Wang, Zhenyue, Guowu Yuan, Hao Zhou, Yi Ma, and Yutang Ma. 2023. "Foreign-Object Detection in High-Voltage Transmission Line Based on Improved YOLOv8m" Applied Sciences 13, no. 23: 12775. https://doi.org/10.3390/app132312775

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop