Next Article in Journal
Unity and ROS as a Digital and Communication Layer for Digital Twin Application: Case Study of Robotic Arm in a Smart Manufacturing Cell
Next Article in Special Issue
Research on Integrated Control Strategy for Wind Turbine Blade Life
Previous Article in Journal
Screen-Printed Electrodes as Low-Cost Sensors for Breast Cancer Biomarker Detection
Previous Article in Special Issue
DAEiS-Net: Deep Aggregation Network with Edge Information Supplement for Tunnel Water Stain Segmentation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

UAV Visual and Thermographic Power Line Detection Using Deep Learning

by
Tiago Santos
1,2,*,
Tiago Cunha
2,
André Dias
1,2,
António Paulo Moreira
1,3 and
José Almeida
1,2
1
INESCTEC—Institute for Systems and Computer Engineering, Technology and Science, Rua Dr. Roberto Frias, 4200-465 Porto, Portugal
2
ISEP—School of Engineering, Polytechnic Institute of Porto, Rua Dr. António Bernardino de Almeida 431, 4200-072 Porto, Portugal
3
FEUP—Faculty of Engineering, University of Porto, Rua Dr. Roberto Frias, 4200-465 Porto, Portugal
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(17), 5678; https://doi.org/10.3390/s24175678
Submission received: 25 June 2024 / Revised: 27 August 2024 / Accepted: 29 August 2024 / Published: 31 August 2024
(This article belongs to the Special Issue AI-Assisted Condition Monitoring and Fault Diagnosis)

Abstract

:
Inspecting and maintaining power lines is essential for ensuring the safety, reliability, and efficiency of electrical infrastructure. This process involves regular assessment to identify hazards such as damaged wires, corrosion, or vegetation encroachment, followed by timely maintenance to prevent accidents and power outages. By conducting routine inspections and maintenance, utilities can comply with regulations, enhance operational efficiency, and extend the lifespan of power lines and equipment. Unmanned Aerial Vehicles (UAVs) can play a relevant role in this process by increasing efficiency through rapid coverage of large areas and access to difficult-to-reach locations, enhanced safety by minimizing risks to personnel in hazardous environments, and cost-effectiveness compared to traditional methods. UAVs equipped with sensors such as visual and thermographic cameras enable the accurate collection of high-resolution data, facilitating early detection of defects and other potential issues. To ensure the safety of the autonomous inspection process, UAVs must be capable of performing onboard processing, particularly for detection of power lines and obstacles. In this paper, we address the development of a deep learning approach with YOLOv8 for power line detection based on visual and thermographic images. The developed solution was validated with a UAV during a power line inspection mission, obtaining [email protected] results of over 90.5% on visible images and over 96.9% on thermographic images.

1. Introduction

In the contemporary world, electricity has become one of the most essential commodities for human life. It plays a pivotal role in powering essential household appliances, illumination, and many currently available technologies. It is also crucial for retail, educational, and medical establishments, among others. In the face of ambitious global initiatives such as the commitment to achieve net-zero emissions by 2050 [1], challenges arise around reduction of emissions. Related to this, there is an evident surge in the adoption of renewable energy sources and the electrification of sectors such as transportation and heating. This transition translates into a foreseeable increase in the demand for electricity. In light of this trend, there is no doubt about the need for electricity distribution companies ro guarantee continuous power delivery without fail, as the alternative could lead to catastrophic consequences. Companies need to implement redundancy and backup strategies to avoid power outages. Redundancy can be achieved by using multiple power line routes to the same place; however, the population can reject the installation of new power lines [2]. Another solution is to take care of the already implemented network; for this, predictive maintenance is vital to ensuring timely preventive action.
The inspection of electrical assets is an important maintenance task in which periodic visual and thermal inspections are carried out in order to find failures in power components [3,4,5]. In the past, these inspections were carried out by human patrols [2,6,7,8] or manned aerial vehicles [9,10]; today, it is possible to use Multi-Rotor UAVs equipped with similar sensors that can get closer to the assets, taking advantage of their high maneuverability and capability to operate in harsh environments.
Unmanned Aerial Vehicle (UAV) application scenarios are increasing rapidly due to the current research interest in aerial robotics [11]. There are many different types of UAV configurations that present multiple structural designs, mirroring the vast number of UAV applications. Fixed-wing and multi-rotor UAVs provide a flexible approach for close inspection of structures, resulting in data with high resolution. Electrical power providers have great interest in these capacities [12,13,14,15,16], as they can reduce human risk and simultaneously collect data from different positions, angles, and distances with respect to electric assets. In this sense, it is critical to develop perception systems onboard the UAV that allow it to detect obstacles and electrical power lines during flight. In this paper, we intend to address this problem, focusing specifically on the development of an onboard perception system for a UAV that allows for tracking of electrical power lines. Equipping UAVs with the ability to perceive electric power lines will allow for future implementation of anomaly inspection strategies (i.e., corrosion, hot spots, etc.). Therefore, this paper contributes to ongoing initiatives to enhance the real-time perception capabilities of Unmanned Aerial Vehicles (UAVs) by equipping them with advanced capabilities for secure and resilient autonomous and semi-autonomous operation. This study focuses on using instance segmentation to detect power lines in images of visual and thermographic cameras that have been synchronously triggered to obtain the images at the same time.
The field of power line inspection has been a focal point for researchers, who are continually exploring innovative solutions to improve the accuracy, efficiency, and safety of these critical operations. Significant advancements have been made in remote sensing methods, as highlighted in the comprehensive review by Matikainen et al. [6].
The following literature review concentrates on key sensors, specifically visual and thermographic cameras, as well as LiDar in the context of power transmission line detection. The emphasis is predominantly on visual and thermographic cameras, aligning with the research findings presented in this study. Aerial images captured via UAV allow data on an object to be acquired in great detail due to UAVs’ close flight capability. Visual images are used for power line detection, while thermal images [17,18,19,20] are usually used for fault monitoring (i.e., hot spots) in power line components.
Traditional vision-based power line detection methods are divided into two significant steps, namely, segment detection/extraction and power line segment extraction. In segment extraction, methods used by algorithms include edge extraction, gray image thresholding [21,22,23], line detector masks, ratio line detection [24], the CVK method [25], canny edge detection [26,27], edge drawing [28], and Marr–Hildreth detection [29]. After segment extraction, the Hough transform is typically used as the core of these power line detection algorithms [21,23,25,26,29,30,31], although it has high computational requirements.
With the growth of the deep learning field in recent years, numerous works have sought to solve detection problems using deep learning techniques. Power line detection is one such problem, as predicted by Nguyen [32]. One of the first works to use deep learning for power line detection was presented by Varghese et al. [33], who retrained the last few layers in a pretrained CNN model for multi-class classification of power infrastructures and achieved a 70% F-score for power line detection. Adam Stambler et al. [34] proposed Deep Wire CNN to detect the lines in the image, followed by a triangulation technique to estimate the 3D location of the lines. To combat the limitations of traditional approaches, Nguyen et al. [35] introduced LS-Net, a fast single-shot line segment detector, and applied it to the power line detection problem. Although they used synthetic images, the results showed that the proposed method could outperform other approaches. Motivated by the potential of instance segmentation, Li et al. [36] proposed the CableNet algorithm based on fully convolutional networks, with two improvements to consider the specific appearance characteristics of transmission lines. First, the convolutional layers were configured to better represent continuous long and thin shapes; second, the output consisted of multidimensional feature maps (i.e., segmentation). This resulted in the assignment of a cable ID to each power line. An approach presented by Song et al. [37] involves tagging each power line; first, a Mask Region-based Convolutional Neural Network (R-CNN) algorithm is used for preliminary extraction, followed by a line-fitting algorithm that solves the problem of broken and mis-extracted power lines. Line fitting over the CNN result was used by Diniz et al. [38], with the images processed through a YOLOV4 network to determine the presence ir absence of power lines. The resulting bounding box is isolated and filtered, then the probabilistic Hough transform is used to estimate the lines. YOLOV3 was used by Son et al. [39] to detect power lines to ensure the safety of agricultural spraying drones. Given the difficulty of labeling power lines as objects, this approach based on the tiny-YOLOv3 network uses several continuous boxes of a specific size. The YOLO series of models started in 2015 [40] with the YOLOv1 model. It initially used a single convolutional neural network (CNN) to detect objects in an image; however, some two-stage models were more accurate at the time. Multiple models were developed over the following years in efforts to overcome various limitations and enhance performance [41]. YOLOv8 is the model most recently released by Ultralytics.
Choi et al. presented a work using thermographic and visible images for power line detection [42]. They compared the performance of an existing fast segmentation convolutional neural network (Fast-SCNN) with a modified version in which additional convolution layers and transposed convolution layers were inserted to improve the performance. An interesting conclusion was that using an image contrast enhancement algorithm such as CLAHE [43] can allow for a clearer view of the power lines, thereby improving the detection results. Motivated by the problem of the class imbalance in deep learning, Yang et al. [44] proposed a power line extraction network called (PLE-Net) based on an encoder–decoder framework. They used the U-net network as the baseline and optimized it using an attention and feature enhancement block. PLE-Net was tested with both visible and infrared images.
Damodaran et al. [45] proposed a pipeline that integrates various elements, preprocessing techniques, deep learning models, classification algorithms, and the Hough transform. The images first passed through the Canny edge detector for better segmentation, followed by Otsu thresholding. Deep learning models were subsequently used for feature extraction. RsurgeNet, which uses VGG16 convolution layers along with addition layers and multiplication layers, was created to extract more features; the AlexNet, Vgg16, and ResNet-50 frameworks were also tested for comparison with the results. Three different classification algorithms were tested to classify the features (AdaBoost, Light GBM, and XgBoost). Finally, the Hough transform was employed to extract the semantic information of power lines.
Multiple authors have highlighted the accuracy and speed advantages of YOLOv8, which allow its use in real-time applications [46,47,48]. The above literature review has covered several methods of detecting lines using visible and thermographic cameras; however, none of the works presented concern the issue of integration with a UAV in a real application scenario. All previous authors have focused on the perception component rather than its integration into an actual vehicle that has to perform the mission of tracking power lines in real time. Thus, the research work presented here seeks to address the following lines of inquiry:
  • A processing architecture for a UAV that allows multi-sensory synchronization of acquired information, allowing the power line-following method to be implemented. The solution proposed in Section 2 can also be scaled in the future to integrate new sources of information such as LiDAR [49].
  • A dataset of synchronized and georeferenced visible and thermographic images with labels.
  • A processing pipeline for visible and thermographic images based on YOLOv8.
  • A summary of the experimental results and discussion of topics to be addressed in future research.
Following this motivation, the paper outline is as follows. In Section 2, our multimodal power line detection approach is defined. Section 3 focuses on the processing pipeline implemented for power line detection. Section 4 presents the constructed dataset of power line images and describes the methodology used to train the neural networks. Section 5 shows the results obtained with the implemented approach. Finally, Section 6 draws some conclusions on the work carried out thus far and describes future work.

2. Multimodal Power Line Detection Approach

Conducting power line inspections with UAVs entails a multifaceted approach that extends beyond deploying sensors such as visual or thermographic cameras. To equip a UAV with autonomous navigation capacities, integration of various other crucial technologies is required, including synchronization mechanisms, GPS for accurate positioning, and data processing systems. Figure 1 shows an architecture that mixes current work with future work.
The dashed box represents the companion computer that is connected to the autopilot. Autopilot data are mapped to the ROS environment by the MAVROS package, with MAVLink messages being exchanged between the two modules.
On the companion computer, the system clock is updated using the chrony application, a versatile implementation of the Network Time Protocol (NTP), which uses the GNSS timestamp and the precise pulse per second (PPS) signal to synchronise the system clock with the GNSS reference clock. This approach bases the system clock on a global reference, facilitating a multi-system integration if needed as well as a precise relation between the timestamps of the data.
The orange boxes in the figure represent the primary sensors present in the UAV, composed of the IMU and GNSS, which provide centimeter-level accuracy through the RTK technique and the visible and thermographic camera sensors. The two cameras have an I/O signal line triggered at the same time to capture an image, which is confirmed by data synchronization for correspondence between the trigger and the images. The green Image Processing box, representing the work presented here, takes the images and processes them to detect the power line. The blue boxes show future work, where the pose of the lines is estimated from the detections to serve as input for the line-following maneuver.

3. Implemented Image Processing Approach

The two cameras have different resolutions and FOVs; thus, an image registration technique is used to align and crop the visual image with the thermographic image. Each of these images is then passed to a corresponding neural network previously trained using that type of image. Finally, the results for each image are combined into a single output mask. Figure 2 details the approach used to address this problem.
The use of the homography matrix resolves the image registration problem. Although the homography matrix is typically used to align planes in an image, given the baseline of the cameras and the flight distance to the power lines, the visual result of the alignment is satisfactory, showing the potential of the solution.
Based on the YOLOv8 algorithm, one neural network is trained to detect power lines in thermographic images and another to detect power lines in visual images.

4. Dataset and Data Preparation

The multirotor STORK UAV [50], shown in Figure 3, was used to carry out a flight over the power lines in the field test area. This UAV is used by the Institute of Systems and Computer Engineering Technology and Science (INESC TEC) in the development of several projects in the field of autonomous UAVs [49,50,51,52] due to its payload and adaptability that allow the vehicle to be customized for other application scenarios.
The UAV was equipped with two cameras, as shown in Figure 4, a Teledyne dalsa calibir DXM640 IR-GMZG-4104500 thermographic camera and a Genie Nano C4020 G3-GC11-C4020 visual camera. The thermographic camera has a 640 × 480 resolution with a field of view (FOV) of 24.2° × 18.4°, while the visual camera has a 4112 × 3008 resolution with a FOV of 83.0° × 66.0°. The flight was made over the power lines with the cameras pointing down. Both cameras were synchronized with a frame rate of 4 fps to correlate the visible and thermographic images.
Two flights were conducted, one in the morning and another in the afternoon. A total of 3600 images were acquired, with 1800 captured by each camera. Figure 5 depicts the trajectory of a manual flight.
The location of the flight was chosen based on the presence of a central power tower with five intersections between different power lines corridors, as shown in Figure 6. In these two images, the difference in FOV is very noticeable. The red square on the RGB image represents the region that the thermographic camera captures.
The thermographic camera is capable of output with 14 bits per pixel. Adaptive contrast enhancement is implemented to output with 8 bits per pixel. The 14-bit image is analyzed to determine the minimum and maximum pixel value. The minimum pixel value is assigned to a value of 0 (black) and the maximum pixel value is assigned to 255. All the values in between are mapped between 0 and 255 to form the final grayscale image.
YOLOv8 uses the same format as the previous v5 iteration. For image segmentation, the annotations are different from those for object detection. Each image has a corresponding .txt file, with each line corresponding to each object segmentation. The line starts with the class ID, followed by multiple (x,y) normalized coordinates that describe the form of the segmentation, as shown in Figure 7.
The labeling process was performed in the Roboflow online application, which can augment data in the images to improve the generalization of the model and achieve greater performance on unseen images. In certain images, mainly the visible images, it can be challenging to correctly annotate the power lines due to the background. The dataset comprises 521 thermographic images and 515 visible images for training, with 50 of each type used for validation and 26 used for testing.
As previously stated, YOLOv8 has five pretrained scaled versions, with the -seg suffix used for image segmentation models. Table 1 shows the differences in performance between the five models. This work uses the YOLOv8n and YOLOv8s segmentation models.

5. Results and Discussion

YOLOv8n and YOLOv8s were used to train the models used to detect power lines. There are several key hyperparameters to consider that can affect the model accuracy, including batch size (which uses more memory, but can stabilize training), learning rate (weight update steps, which allows for precise adjustments yet slows down convergence), image size (all images are resized to this dimension), epochs (total number of full passes over the entire dataset), and patience (number of epochs without improvement in validation metrics, used for early stoppage of training).
The training phase was configured to run over 500 epochs, with the two models trained using the same default parameters except for the patience level, which was set as p a t i e n c e = 150 for YOLOv8n and p a t i e n c e = 50 for YOLOv8s. Decreasing the patience allows faster training, but can lead to worse performance.
Both YOLOv8s models were interrupted by the patience parameter, in epoch 229 for the RGB model and epoch 187 for the IR model. For YOLOv8n, the RGB model was interrupted in epoch 379, while the IR model used all 500 epochs. Despite the lower patience value used for YOLOv8n, it took more epochs for training compared to YOLOv8s. In the case of the IR model, the patience parameter was not used, and the model ran for the full 500 epochs.
For each model and camera, Figure 8 shows the training loss (blue) and validation loss (orange), while Figure 9 shows the mean average precision ([email protected]).
After training the models, the validation dataset was used to evaluate their performance. For this task, a fixed confidence threshold of 0.5 was selected for both types of images.
The inference time (milliseconds per image) is a critical aspect to consider, as it shows the processing time for each image. Table 2 shows the summary of the results, with the running times are compared to the previously trained models on an Intel i5-9300H CPU and Nvidia GeForce GTX 1650 GPU. The preprocessing step corresponds to the alignment between the images from the two cameras, while the postprocessing step refers to the combination of the two obtained masks.
As expected, the inference time with the GPU is considerably lower than with the CPU. Although it is expected that the final real-time processing solution onboard the UAV will require the use of a GPU, we present the results obtained with a CPU equivalent to the one currently equipped on the UAV STORK used to produce the dataset. In addition, the difference in performance based on the selected model is notable, with the nano version is approximately two times faster than the small model.
In the labeling phase, there were cases of small or thin lines that were hard for the human eye to see; naturally, these were not annotated in order to avoid misleading the training. However, the result was that the trained models could detect more lines than were annotated in the validation dataset. In the 50 images from each camera presented in the validation dataset, 131 lines were annotated in the visible images and 135 in the thermographic images.
Table 3 shows the results of instance segmentation on the validation dataset, referring to the metrics [email protected] and [email protected]:0.95. From these results, the performance of the YOLOv8n instance seems to be better than that of the YOLOv8s instance, which was unexpected. From the training results in Figure 8 and Figure 9, it can be concluded that training of the YOLOv8n could be improved to achieve better results. However, given the detected instances and the masks produced for the images, the achieved results are satisfactory.
Figure 10 shows an inference example for both models, with the blue lines representing the thermographic image detection and the red lines representing the visible image detection.
Figure 11 shows the obtained results in selected pairs of images. In section A (the tower with intersecting power lines), all lines were correctly detected. In section B, the RGB model was not capable of detection, however, the thermographic model correctly detected all three lines. In section C, the thermographic model failed while the RGB model was capable of detection. In section D, misalignment in the detections can be seen, most likely caused by the image registration technique. Finally, section E shows another successful detection.

6. Conclusions and Future Work

In this study, we have addressed the problem of power line detection by a UAV, proposing a solution based on YOLOv8 segmentation applied to visible and thermographic images synchronously acquired by the STORK UAV. Two flights were conducted in a specific area with five power line spans connected to a single power tower. Due to significant differences in FOV and size between the two image types, a preprocessing step was implemented in which the visual image is cropped and aligned with the thermographic image. YOLOv8n and YOLOV8s scaled versions were trained to detect power lines for each type of image. The experimental results show the efficacy of both models, with the YOLOv8n model achieving 98.4% mAP@50 on thermal images and 91.8% mAP@50 on visible images and the YOLOv8s model achieving 96.9% mAP@50 on thermal images and 90.5% mAP@50 on visible images. The YOLOv8n model performs better than the YOLOv8s model due to the suboptimal training of the YOLOv8s model. Even though a lower patience value was used, the IR model trained over all 500 epochs. Increasing the number of epochs and patience could lead to better results by both models. However, it is important to highlight that both models demonstrated the ability to segment power lines from both visible spectrum and thermographic images. This work has allowed us to validate AI techniques, in particular through the YOLO framework, in the context of line detection (segmentation) in visible and thermographic spectrum cameras and through the conceptualization of a processing architecture that in the future will allow processing onboard the UAV using edge computing. As this is an ongoing research topic, several weaknesses and challenges remain to be addressed. First, the image registration technique should be addressed in order to open up future possibilities involving instance segmentation, as in certain cases misalignment of the detections in the combined mask occurred during image registration. Second, the results revealed instances where detection failed in visible images while succeeding in thermographic images, and vice versa; a combination of the two sources of information could be studied to mitigate possible detection failures, and other sources of information such as LiDAR could be integrated into the proposed architecture. Third, at the time of this study YOLOv8 was the latest iteration of the model released by Ultralytics; however, this field is constantly evolving, with newer models such as YOLOv9 being released. The release of new models does not necessarily mean a clear improvement over the previous versions, as Jaykumaran concluded in his study comparing YOLOv8 and YOLOv9 [54]; nonetheless, emerging AI models such as RT-DERT [55] should be analyzed and tested to seek the best results and the possibility of executing them in real time.

Author Contributions

Conceptualization, T.S. and A.D.; Software, T.S and T.C.; Validation, T.S. and A.D.; Investigation, T.S. and T.C.; Resources, T.S. and A.D.; Writing—original draft, T.S.; Writing—review & editing, T.S., A.D., A.P.M. and J.A.; Project administration, J.A.; Funding acquisition, J.A. All authors have read and agreed to the published version of the manuscript.

Funding

The paper was funding by National Funds through the Portuguese funding agency, FCT—Fundação para a Ciência e a Tecnologia, within project UIDB/50014/2020. DOI 10.54499/UIDB/ 50014/2020.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. International Energy Agency. Net Zero by 2050; International Energy Agency: Paris, France, 2021; p. 224. [Google Scholar] [CrossRef]
  2. Pagnano, A.; Höpf, M.; Teti, R. A Roadmap for Automated Power Line Inspection. Maintenance and Repair. Proc. CIRP 2013, 12, 234–239. [Google Scholar] [CrossRef]
  3. Katrasnik, J.; Pernus, F.; Likar, B. A Survey of Mobile Robots for Distribution Power Line Inspection. IEEE Trans. Power Deliv. 2010, 25, 485–493. [Google Scholar] [CrossRef]
  4. Aggarwal, R.; Johns, A.; Jayasinghe, J.; Su, W. Overview of the condition monitoring of overhead lines. Electr. Power Syst. Res. 2000, 53, 15–22. [Google Scholar] [CrossRef]
  5. Sampedro, C.; Rodriguez-Vazquez, J.; Rodriguez-Ramos, A.; Carrio, A.; Campoy, P. Deep Learning-Based System for Automatic Recognition and Diagnosis of Electrical Insulator Strings. IEEE Access 2019, 7, 101283–101308. [Google Scholar] [CrossRef]
  6. Matikainen, L.; Lehtomäki, M.; Ahokas, E.; Hyyppä, J.; Karjalainen, M.; Jaakkola, A.; Kukko, A.; Heinonen, T. Remote sensing methods for power line corridor surveys. ISPRS J. Photogramm. Remote Sens. 2016, 119, 10–31. [Google Scholar] [CrossRef]
  7. Chen, C.; Yang, B.; Song, S.; Peng, X.; Huang, R. Automatic Clearance Anomaly Detection for Transmission Line Corridors Utilizing UAV-Borne LIDAR Data. Remote Sens. 2018, 10, 613. [Google Scholar] [CrossRef]
  8. Mills, S.J.; Gerardo Castro, M.P.; Li, Z.; Cai, J.; Hayward, R.; Mejias, L.; Walker, R.A. Evaluation of Aerial Remote Sensing Techniques for Vegetation Management in Power-Line Corridors. IEEE Trans. Geosci. Remote Sens. 2010, 48, 3379–3390. [Google Scholar] [CrossRef]
  9. Mirallès, F.; Pouliot, N.; Montambault, S. State-of-the-art review of computer vision for the management of power transmission lines. In Proceedings of the 2014 3rd International Conference on Applied Robotics for the Power Industry, Foz do Iguacu, Brazil, 14–16 October 2014; pp. 1–6. [Google Scholar] [CrossRef]
  10. Whitworth, C.C.; Duller, A.W.G.; Jones, D.I.; Earp, G.K. Aerial video inspection of overhead power lines. Power Eng. J. 2001, 15, 25–32. [Google Scholar] [CrossRef]
  11. Liew, C.F.; DeLatte, D.; Takeishi, N.; Yairi, T. Recent Developments in Aerial Robotics: A Survey and Prototypes Overview. arXiv 2017, arXiv:1711.10085. [Google Scholar] [CrossRef]
  12. Drones for Power Line Inspections. Available online: https://www.utilityproducts.com/line-construction-maintenance/article/16003823/drones-for-power-line-inspections (accessed on 6 December 2019).
  13. A Bird’s-Eye View: Drones in the Power Sector. Available online: https://www.powermag.com/a-birds-eye-view-drones-in-the-power-sector/ (accessed on 6 December 2019).
  14. Robotic Unmanned Drone NDT Inspections for Energy. Available online: https://www.aetosgroup.com/ (accessed on 6 December 2019).
  15. Lages, W.F.; de Oliveira, V.M. A survey of applied robotics for the power industry in Brazil. In Proceedings of the 2012 2nd International Conference on Applied Robotics for the Power Industry (CARPI), Zurich, Switzerland, 11–13 September 2012; pp. 78–82. [Google Scholar] [CrossRef]
  16. Li, X.; Guo, Y. Application of LiDAR technology in power line inspection. IOP Conf. Ser. Mater. Sci. Eng. 2018, 382, 052025. [Google Scholar] [CrossRef]
  17. Wronkowicz, A. Automatic fusion of visible and infrared images taken from different perspectives for diagnostics of power lines. Quant. Infrared Thermogr. J. 2016, 13, 155–169. [Google Scholar] [CrossRef]
  18. Stockton, G.R.; Tache, A. Advances in applications for aerial infrared thermography. In Thermosense XXVIII: Defense and Security Symposium, Orlando, FL, USA, 17–21 April 2006; Miles, J.J., Peacock, G.R., Knettel, K.M., Eds.; International Society for Optics and Photonics, SPIE: Bellingham, WA, USA, 2006; Volume 6205, pp. 99–109. [Google Scholar] [CrossRef]
  19. Jalil, B.; Leone, G.R.; Martinelli, M.; Moroni, D.; Pascali, M.A.; Berton, A. Fault Detection in Power Equipment via an Unmanned Aerial System Using Multi Modal Data. Sensors 2019, 19, 3014. [Google Scholar] [CrossRef]
  20. Shams, F.; Omar, M.; Usman, M.; Khan, S.; Larkin, S.; Raw, B. Thermal Imaging of Utility Power Lines: A Review. In Proceedings of the 2022 International Conference on Engineering and Emerging Technologies (ICEET), Kuala Lumpur, Malaysia, 27–28 October 2022; pp. 1–4. [Google Scholar] [CrossRef]
  21. Zhang, J.; Liu, L.; Wang, B.; Chen, X.; Wang, Q.; Zheng, T. High Speed Automatic Power Line Detection and Tracking for a UAV-Based Inspection. In Proceedings of the 2012 International Conference on Industrial Control and Electronics Engineering, Xi’an, China, 23–25 August 2012; pp. 266–269. [Google Scholar] [CrossRef]
  22. Sharma, H.; Bhujade, R.; Adithya, V.; Balamuralidhar, P. Vision-based detection of power distribution lines in complex remote surroundings. In Proceedings of the 2014 Twentieth National Conference on Communications (NCC), Kanpur, India, 28 February–2 March 2014; pp. 1–6. [Google Scholar] [CrossRef]
  23. Yang, T.W.; Yin, H.; Ruan, Q.Q.; Han, J.D.; Qi, J.T.; Yong, Q.; Wang, Z.T.; Sun, Z.Q. Overhead power line detection from UAV video images. In Proceedings of the 2012 19th International Conference on Mechatronics and Machine Vision in Practice (M2VIP), Auckland, New Zealand, 28–30 November 2012; pp. 74–79. [Google Scholar]
  24. Yan, G.; Li, C.; Zhou, G.; Zhang, W.; Li, X. Automatic Extraction of Power Lines From Aerial Images. IEEE Geosci. Remote Sens. Lett. 2007, 4, 387–391. [Google Scholar] [CrossRef]
  25. Golightly, I.; Jones, D. Visual control of an unmanned aerial vehicle for power line inspection. In Proceedings of the ICAR ’05, 12th International Conference on Advanced Robotics, Seattle, WA, USA, 18–20 July 2005; pp. 288–295. [Google Scholar] [CrossRef]
  26. Gerke, M.; Seibold, P. Visual inspection of power lines by U.A.S. In Proceedings of the 2014 International Conference and Exposition on Electrical and Power Engineering (EPE), Iasi, Romania, 16–18 October 2014; pp. 1077–1082. [Google Scholar] [CrossRef]
  27. Cerón, A.; B., I.F.M.; Prieto, F. Power line detection using a circle based search with UAV images. In Proceedings of the 2014 International Conference on Unmanned Aircraft Systems (ICUAS), Orlando, FL, USA, 27–30 May 2014; pp. 632–639. [Google Scholar] [CrossRef]
  28. Santos, T.; Moreira, M.; Almeida, J.; Dias, A.; Martins, A.; Dinis, J.; Formiga, J.; Silva, E. PLineD: Vision-based power lines detection for Unmanned Aerial Vehicles. In Proceedings of the 2017 IEEE International Conference on Autonomous Robot Systems and Competitions (ICARSC), Coimbra, Portugal, 26–28 April 2017; pp. 253–259. [Google Scholar] [CrossRef]
  29. Tong, W.G.; Li, B.S.; Yuan, J.S.; Zhao, S.T. Transmission line extraction and recognition from natural complex background. In Proceedings of the 2009 International Conference on Machine Learning and Cybernetics, Baoding, China, 12–15 July 2009; pp. 12–15. [Google Scholar]
  30. Li, Z.; Liu, Y.; Hayward, R.; Zhang, J.; Cai, J. Knowledge-based power line detection for UAV surveillance and inspection systems. In Proceedings of the 2008 23rd International Conference Image and Vision Computing New Zealand, Christchurch, New Zealand, 26–28 November 2008; pp. 1–6. [Google Scholar] [CrossRef]
  31. Du, S.; Tu, C. Power line inspection using segment measurement based on HT butterfly. In Proceedings of the 2011 IEEE International Conference on Signal Processing, Communications and Computing (ICSPCC), Xi’an, China, 14–16 September 2011; pp. 1–4. [Google Scholar] [CrossRef]
  32. Nguyen, V.N.; Jenssen, R.; Roverso, D. Automatic autonomous vision-based power line inspection: A review of current status and the potential role of deep learning. Int. J. Electr. Power Energy Syst. 2018, 99, 107–120. [Google Scholar] [CrossRef]
  33. Varghese, A.; Gubbi, J.; Sharma, H.; Balamuralidhar, P. Power infrastructure monitoring and damage detection using drone captured images. In Proceedings of the 2017 International Joint Conference on Neural Networks (IJCNN), Anchorage, AK, USA, 14–19 May 2017; pp. 1681–1687. [Google Scholar] [CrossRef]
  34. Stambler, A.; Sherwin, G.; Rowe, P. Detection and Reconstruction of Wires Using Cameras for Aircraft Safety Systems. In Proceedings of the 2019 International Conference on Robotics and Automation (ICRA), Montreal, QC, Canada, 20–24 May 2019; pp. 697–703. [Google Scholar] [CrossRef]
  35. Nguyen, V.N.; Jenssen, R.; Roverso, D. LS-Net: Fast Single-Shot Line-Segment Detector. arXiv 2019, arXiv:1912.09532. [Google Scholar] [CrossRef]
  36. Li, B.; Chen, C.; Dong, S.; Qiao, J. Transmission line detection in aerial images: An instance segmentation approach based on multitask neural networks. Signal Process. Image Commun. 2021, 96, 116278. [Google Scholar] [CrossRef]
  37. Song, J.; Qian, J.; Li, Y.; Liu, Z.; Chen, Y.; Chen, J. Automatic Extraction of Power Lines from Aerial Images of Unmanned Aerial Vehicles. Sensors 2022, 22, 6431. [Google Scholar] [CrossRef]
  38. Diniz, L.; Faria Pinto, M.; Melo, A.; Honorio, L. Visual-based Assistive Method for UAV Power Line Inspection and Landing. J. Intell. Robot. Syst. 2022, 106, 41. [Google Scholar] [CrossRef]
  39. Son, H.S.; Kim, D.K.; Yang, S.H.; Choi, Y.K. Real-Time Power Line Detection for Safe Flight of Agricultural Spraying Drones Using Embedded Systems and Deep Learning. IEEE Access 2022, 10, 54947–54956. [Google Scholar] [CrossRef]
  40. Redmon, J.; Divvala, S.K.; Girshick, R.B.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. arXiv 2016, arXiv:1506.02640. [Google Scholar] [CrossRef]
  41. Terven, J.; Cordova-Esparza, D. A Comprehensive Review of YOLO: From YOLOv1 to YOLOv8 and Beyond. arXiv 2023, arXiv:2304.00501. [Google Scholar] [CrossRef]
  42. Choi, H.; Koo, G.; Kim, B.J.; Kim, S.W. Real-time Power Line Detection Network using Visible Light and Infrared Images. In Proceedings of the 2019 International Conference on Image and Vision Computing New Zealand (IVCNZ), Dunedin, New Zealand, 2–4 December 2019; pp. 1–6. [Google Scholar]
  43. Zuiderveld, K. Contrast Limited Adaptive Histogram Equalization. In Graphics Gems IV; Academic Press Professional, Inc.: Cambridge, MA, USA, 1994; pp. 474–485. [Google Scholar]
  44. Yang, L.; Fan, J.; Huo, B.; Li, E.; Liu, Y. PLE-Net: Automatic power line extraction method using deep learning from aerial images. Expert Syst. Appl. 2022, 198, 116771. [Google Scholar] [CrossRef]
  45. Satheeswari Damodaran, L.S.; Swaroopan, N.J. Overhead power line detection from aerial images using segmentation approaches. Automatika 2024, 65, 261–288. [Google Scholar] [CrossRef]
  46. Kaliappan, V.K.; V, M.S.; Shanmugasundaram, K.; Ravikumar, L.; Hiremath, G.B. Performance Analysis of YOLOv8, RCNN, and SSD Object Detection Models for Precision Poultry Farming Management. In Proceedings of the 2023 IEEE 3rd International Conference on Applied Electromagnetics, Signal Processing, & Communication (AESPC), Bhubaneswar, India, 24–26 November 2023; pp. 1–6. [Google Scholar] [CrossRef]
  47. Sapkota, R.; Ahmed, D.; Karkee, M. Comparing YOLOv8 and Mask RCNN for object segmentation in complex orchard environments. arXiv 2024, arXiv:2312.07935. [Google Scholar] [CrossRef]
  48. Sohan, M.; Sai Ram, T.; Rami Reddy, C.V. A Review on YOLOv8 and Its Advancements. In Data Intelligence and Cognitive Informatics, Proceedings of the ICDICI 2023: 4th International Conference on Data Intelligence and Cognitive Informatics, Tirunelveli, India, 27–28 June 2023; Jacob, I.J., Piramuthu, S., Falkowski-Gilski, P., Eds.; Springer: Singapore, 2024; pp. 529–545. [Google Scholar]
  49. Azevedo, F.; Dias, A.; Almeida, J.; Oliveira, A.; Ferreira, A.; Santos, T.; Martins, A.; Silva, E. Lidar-based real-time detection and modeling of power lines for unmanned aerial vehicles. Sensors 2019, 19, 1812. [Google Scholar] [CrossRef] [PubMed]
  50. Dias, A.; Almeida, J.; Oliveira, A.; Santos, T.; Martins, A.; Silva, E. Unmanned Aerial Vehicle for Wind-Turbine Inspection. Next Step: Offshore. In Proceedings of the OCEANS 2022, Hampton Roads, VA, USA, 17–20 October 2022; pp. 1–6. [Google Scholar] [CrossRef]
  51. Silva, P.; Dias, A.; Pires, A.; Santos, T.; Amaral, A.; Rodrigues, P.; Almeida, J.; Silva, E. 3D Reconstruction of historical sites using an UAV. In Proceedings of the CLAWAR, Moscow, Russia, 24–26 August 2020. [Google Scholar]
  52. Dias, A.; Mucha, A.P.; Santos, T.; Pedrosa, D.; Amaral, G.; Ferreira, H.; Oliveira, A.; Martins, A.; Almeida, J.; Almeida, C.M.; et al. ROSM—Robotic Oil Spill Mitigations. In Proceedings of the OCEANS 2019, Marseille, France, 17–20 June 2019; pp. 1–7. [Google Scholar] [CrossRef]
  53. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLO. Available online: https://github.com/ultralytics/ultralytics (accessed on 31 December 2023).
  54. Jaykumaran. YOLOv9 Instance Segmentation on Medical Dataset. Available online: https://learnopencv.com/yolov9-instance-segmentation-on-medical-dataset#aioseo-conclusion (accessed on 30 July 2024).
  55. Zhao, Y.; Lv, W.; Xu, S.; Wei, J.; Wang, G.; Dang, Q.; Liu, Y.; Chen, J. DETRs Beat YOLOs on Real-time Object Detection. arXiv 2024, arXiv:2304.08069. [Google Scholar] [CrossRef]
Figure 1. Overview of the software and hardware architecture with data flow.
Figure 1. Overview of the software and hardware architecture with data flow.
Sensors 24 05678 g001
Figure 2. Processing pipeline.
Figure 2. Processing pipeline.
Sensors 24 05678 g002
Figure 3. STORK UAV during an inspection mission.
Figure 3. STORK UAV during an inspection mission.
Sensors 24 05678 g003
Figure 4. Onboard thermographic and visual camera on STORK UAV.
Figure 4. Onboard thermographic and visual camera on STORK UAV.
Sensors 24 05678 g004
Figure 5. UAV trajectory (green line) during inspection of the power lines.
Figure 5. UAV trajectory (green line) during inspection of the power lines.
Sensors 24 05678 g005
Figure 6. Tower with five intersecting power lines: (a) RGB image, with red square representing the placement of thermographic image; (b) thermographic image.
Figure 6. Tower with five intersecting power lines: (a) RGB image, with red square representing the placement of thermographic image; (b) thermographic image.
Sensors 24 05678 g006
Figure 7. Segmentation points annotation example.
Figure 7. Segmentation points annotation example.
Sensors 24 05678 g007
Figure 8. Model training loss (blue line) and validation loss (orange line): (a) YOLOV8n RGB model, (b) YOLOV8s RGB model, (c) YOLOV8n IR model, (d) YOLOV8s IR model.
Figure 8. Model training loss (blue line) and validation loss (orange line): (a) YOLOV8n RGB model, (b) YOLOV8s RGB model, (c) YOLOV8n IR model, (d) YOLOV8s IR model.
Sensors 24 05678 g008
Figure 9. Model mean average precision ([email protected]): (a) YOLOV8n RGB model, (b) YOLOV8s RGB model, (c) YOLOV8n IR model, (d) YOLOV8s IR model.
Figure 9. Model mean average precision ([email protected]): (a) YOLOV8n RGB model, (b) YOLOV8s RGB model, (c) YOLOV8n IR model, (d) YOLOV8s IR model.
Sensors 24 05678 g009
Figure 10. Example model predictions: (a) RGB image, (b) thermographic image, (c) YOLOV8n detection mask, (d) YOLOV8s detection mask.
Figure 10. Example model predictions: (a) RGB image, (b) thermographic image, (c) YOLOV8n detection mask, (d) YOLOV8s detection mask.
Sensors 24 05678 g010
Figure 11. Detection examples on multiple images.
Figure 11. Detection examples on multiple images.
Sensors 24 05678 g011
Table 1. YOLOv8 pretrained segment models [53].
Table 1. YOLOv8 pretrained segment models [53].
ModelSize
(Pixels)
mAPbox
50–95
mAPmask
50–95
Speed
CPU ONNX
(ms)
Speed
A100 TensorRT
(ms)
Params
(M)
FLOPs
(B)
YOLOv8n-seg64036.730.596.11.213.412.6
YOLOv8s-seg64044.636.8155.71.4711.842.6
YOLOv8m-seg64049.940.8317.02.1827.3110.2
YOLOv8l-seg64052.342.6572.42.7946.0220.5
YOLOv8x-seg64053.443.4712.14.0271.8344.1
Table 2. Runtime comparison of the models (ms/image) on an Intel i5 9300H CPU and Nvidia GeForce GTX 1650 GPU.
Table 2. Runtime comparison of the models (ms/image) on an Intel i5 9300H CPU and Nvidia GeForce GTX 1650 GPU.
ModelGPU
ThermalVisible
Pre-Process (ms)Inference (ms)Post-Process (ms)Pre-Process (ms)Inference (ms)Post-Process (ms)
YOLOv8n2.58.30.92.47.50.9
YOLOv8s2.516.60.82.616.61.0
ModelCPU
ThermalVisible
Pre-process (ms)Inference (ms)Post-process (ms)Pre-process (ms)Inference (ms)Post-process (ms)
YOLOv8n1.999.10.32.199.20.3
YOLOv8s2.1213.20.22.2210.20.3
Table 3. Instance segmentation results on the validation dataset.
Table 3. Instance segmentation results on the validation dataset.
Thermal Images
ModelDetected
Instances
BoxMask
[email protected]
(%)
mAP@[0.5:0.95]
(%)
[email protected]
(%)
mAP@[0.5:0.95]
(%)
YOLOv8n14198.493.098.465.2
YOLOv8s14196.988.996.962.3
Visible Images
ModelDetected
Instances
BoxMask
[email protected]
(%)
mAP@[0.5:0.95]
(%)
[email protected]
(%)
mAP@[0.5:0.95]
(%)
YOLOv8n13597.086.791.847.3
YOLOv8s13595.785.590.546.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Santos, T.; Cunha, T.; Dias, A.; Moreira, A.P.; Almeida, J. UAV Visual and Thermographic Power Line Detection Using Deep Learning. Sensors 2024, 24, 5678. https://doi.org/10.3390/s24175678

AMA Style

Santos T, Cunha T, Dias A, Moreira AP, Almeida J. UAV Visual and Thermographic Power Line Detection Using Deep Learning. Sensors. 2024; 24(17):5678. https://doi.org/10.3390/s24175678

Chicago/Turabian Style

Santos, Tiago, Tiago Cunha, André Dias, António Paulo Moreira, and José Almeida. 2024. "UAV Visual and Thermographic Power Line Detection Using Deep Learning" Sensors 24, no. 17: 5678. https://doi.org/10.3390/s24175678

APA Style

Santos, T., Cunha, T., Dias, A., Moreira, A. P., & Almeida, J. (2024). UAV Visual and Thermographic Power Line Detection Using Deep Learning. Sensors, 24(17), 5678. https://doi.org/10.3390/s24175678

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop