Skip to Content
EnergiesEnergies
  • Article
  • Open Access

8 April 2025

Fast Monocular Measurement via Deep Learning-Based Object Detection for Real-Time Gas-Insulated Transmission Line Deformation Monitoring

,
,
,
,
,
and
1
Department of Mechanical Engineering, North China Electric Power University, Baoding 071003, China
2
Hebei Engineering Research Center for Advanced Manufacturing & Intelligent Operation and Maintenance of Electric Power Machinery, North China Electric Power University, Baoding 071003, China
3
Hebei Key Laboratory of Electric Machinery Health Maintenance & Failure Prevention, North China Electric Power University, Baoding 071003, China
4
The State Key Laboratory of Rolling and Automation, Northeastern University, Shenyang 110819, China

Abstract

Deformation monitoring of Gas-Insulated Transmission Lines (GILs) is critical for the early detection of structural issues and for ensuring safe power transmission. In this study, we introduce a rapid monocular measurement method that leverages deep learning for real-time monitoring. A YOLOv10 model is developed for automatically identifying regions of interest (ROIs) that may exhibit deformations. Within these ROIs, grayscale data is used to dynamically set thresholds for FAST corner detection, while the Shi–Tomasi algorithm filters redundant corners to extract unique feature points for precise tracking. Subsequent subpixel refinement further enhances measurement accuracy. To correct image tilt, ArUco markers are employed for geometric correction and to compute a scaling factor based on their known edge lengths, thereby reducing errors caused by non-perpendicular camera angles. Simulated experiments validate our approach, demonstrating that combining refined ArUco marker coordinates with manually annotated features significantly improves detection accuracy. Our method achieves a mean absolute error of no more than 1.337 mm and a processing speed of approximately 0.024 s per frame, meeting the precision and efficiency requirements for GIL deformation monitoring. This integrated approach offers a robust solution for long-term, real-time monitoring of GIL deformations, with promising potential for practical applications in power transmission systems.

1. Introduction

Gas-Insulated Transmission Lines (GILs) are high-voltage, high-current power transmission systems that utilize SF6 gas or SF6/N2 gas mixtures for insulation, with conductors and outer casings arranged coaxially. Compared to conventional overhead and cable transmission lines, GILs offer advantages such as higher transmission capacity, reduced transmission losses, environmental friendliness, and versatility in application, which have facilitated their widespread global adoption [1,2,3]. Deformation data on GIL pipelines under operational loads serve as critical indicators for assessing structural health and preventing structural failures. Therefore, developing effective deformation monitoring methods for GIL pipelines is of paramount engineering significance for ensuring the safe and reliable operation of both GIL systems and the broader power transmission infrastructure.
Traditional deformation monitoring methods for GILs involve direct mechanical contact of sensitive measurement components, such as resistive strain gauges and fiber optic sensors, with the surface of the target structure to acquire deformation data at specific points. However, these approaches are characterized by low measurement efficiency, vulnerability to strong electromagnetic interference, and high implementation complexity. Consequently, they are inadequate for large-scale and long-term deformation monitoring applications [4,5,6].
Recent advancements in computer hardware and image processing technologies have facilitated the gradual integration of machine vision into structural displacement monitoring for power transmission lines [7]. Yang et al. [8] utilized monocular vision technology to extract the contours of transmission towers and accurately measure their degree of inclination. Wang et al. [9] developed a sag variation monitoring method for transmission lines under icing conditions utilizing binocular vision. Zhang et al. [10] introduced a Region of Interest (ROI) key point method based on the geometric configuration of transmission towers to obtain displacement information of tower structures. Li et al. [11] combined deep learning approaches to propose a real-time safety clearance monitoring method for transmission corridors. Fauzan et al. [12] proposed a rapid method for measuring GIL deformation that achieves a mean absolute error of less than 3 mm. In their approach, artificial targets affixed to the surface of the GIL facility serve as monitoring points, with the displacement of these targets indicating the extent of deformation. However, since the pixel coordinates of these targets are obtained through a semi-manual process and the measurements are conducted offline, the procedure is relatively complex and labor-intensive. Future studies should focus on enhancing detection efficiency and automation while maintaining the required accuracy, thereby reducing both measurement complexity and personnel involvement.
Fundamentally, the studies mentioned above primarily focused on employing machine vision for displacement detection or monitoring in power transmission line structures. However, research on machine vision-based deformation monitoring methods for GIL pipelines remains in its nascent stages. Determining the pixel coordinates of markers in an image can be framed as an object detection problem. Among traditional detection methods, template matching has long been recognized as an effective approach. Xu et al. [13] applied monocular vision techniques for multi-point displacement measurements on bridges. By estimating camera parameters from point pairs with known coordinates and tracking bridge feature points through template matching, they achieved millimeter-level precision in capturing bridge vibrations. Similarly, Brownjohn et al. [14] converted displacement data at bridge measurement points by using the distance between the camera’s optical center and the measurement plane (i.e., line features), also employing template matching for target tracking and identification. Yang et al. [15] further improved target localization in complex backgrounds by introducing a dynamic template updating strategy.
While template matching effectively extracts targets, its application in GIL deformation monitoring, which requires multiple manual markers, introduces additional complexity in determining marker positions. Moreover, template matching is relatively slow and computationally demanding, as it relies on pre-prepared template images. During target motion, factors such as noise, illumination fluctuations, and background variations may cause mismatches, thereby necessitating increased manual intervention. A rapid method for rough localization of the target region could substantially accelerate both detection and tracking. In recent years, the deep learning-based YOLO algorithm has demonstrated high adaptability and detection efficiency across diverse scenarios, satisfying the requirements of such applications. Wu et al. [16] analyzed object detection algorithms in traditional remote sensing imagery and noted that template matching shows limited robustness under conditions of rotation or scaling. They subsequently refined the YOLOv8 framework to improve detection accuracy for multi-scale objects and small targets. Li et al. [17] combined the YOLO-T algorithm with manual marking and monocular vision-based localization to enable underwater robots to detect and locate targets. Yang et al. [18] applied an improved YOLOv3 model for the coarse localization of tea bud pixel coordinates by using the two coordinate points provided by the detection bounding box to identify the region of interest, extracting it as a sub-image and applying subsequent image processing techniques for precise localization.
The YOLO detection model allows rapid coarse localization of manual markers, significantly enhancing both detection speed and accuracy—even achieving real-time performance. These markers yield robust corner features that serve as reliable tracking targets. Consequently, after extracting the sub-image containing a marker, further extraction of corner features facilitates subsequent tracking and measurement. For instance, the classical Oriented FAST and Rotated BRIEF (ORB) [19] algorithm uses the Features from Accelerated Segment Test (FAST) to extract corner features, making it widely applicable in scenarios that prioritize real-time performance over ultra-high precision. However, the conventional FAST algorithm is constrained by its reliance on manually specified thresholds, which may not be optimal for detecting corners across an entire image, thus reducing detection accuracy. To address this limitation, Li et al. [20] optimized the pixel comparison order using decision trees and applied non-maximum suppression to filter feature points, thereby enhancing detection performance.
The proposed research approach involves first performing rapid coarse localization of the marker region and extracting it as a sub-image, followed by the extraction of corner features for tracking. This method not only enables real-time localization but also supports automatic tracking of corner points. In monocular measurement, the calculation of the actual displacement of a target is often performed using methods such as the proportional factor approach. This technique involves converting the physical displacement by leveraging known parameters, such as the distance between the measurement point and the target or identifiable texture features present on the target itself [21,22]. These techniques convert physical displacements by utilizing the known distance between measurement points and the target or by leveraging known texture features on the target surface. However, both methods are adversely affected by the misalignment between the camera optical axis and the measurement plane, which can significantly impact measurement accuracy. Research conducted under laboratory conditions, as reported in reference [22], indicates that the proportional factor method is particularly sensitive to the angle between the camera’s optical axis and the target plane. It is recommended that this angle be maintained within 10°, and the distance from the camera to the target should not exceed 3.7 m to minimize measurement errors. To address the limitations imposed by the camera’s placement angle, image tilt correction is performed to align the image with an ideal imaging plane that is perpendicular to the optical axis. Nevertheless, a significant limitation of monocular measurements is their inability to capture out-of-plane displacements, which restricts the camera’s measurement pose. Instead of constraining the camera’s pose during monitoring, image correction techniques can address this issue. For example, Shan et al. [23] employed prior instrument information to correct perspective distortions, thereby improving the accuracy of their automatic meter reading algorithm. Consequently, when measuring marker displacements using monocular vision, incorporating image tilt correction techniques supported by prior information warrants further investigation and validation.
Issues such as the compatibility between GIL pipeline deformation characteristics and machine vision techniques, as well as the real-time automation of the monitoring process, remain areas that require further exploration. To address the limitations of traditional deformation monitoring methods for GIL, we propose an automatic, real-time monocular measurement approach tailored to the specific deformation characteristics of GIL pipelines. Experimental validation demonstrates the accuracy, timeliness, and robustness of the proposed method, thereby providing a feasible and effective monitoring solution for GIL pipeline deformations. Compared to previous studies, our proposed method reduces measurement complexity and costs while significantly accelerating deformation monitoring, achieving real-time performance without compromising accuracy. Furthermore, by integrating the Shi–Tomasi algorithm with adaptive threshold computation, our approach eliminates the need for manual intervention and enhances the detection precision of the traditional FAST algorithm for automatic extraction and tracking of marker features. Finally, by leveraging the physical information provided by ArUco codes, we implement image tilt correction to eliminate additional errors caused by non-perpendicular alignment between the camera’s optical axis and the measurement plane in monocular setups.

2. Design of a Monocular Measurement System for GIL Deformation

Due to the highly repetitive texture features on the surfaces of GIL pipeline components, texture-based feature extraction and matching methods, such as Scale-Invariant Feature Transform (SIFT) [24], Speeded-Up Robust Features (SURF) [25], and Oriented FAST and Rotated BRIEF (ORB) [26], often result in repetitive local feature matches. This repetition impedes tracking algorithms from accurately identifying and following individual feature points. Implementing artificial markers effectively addresses this issue by simplifying the measurement process, transforming GIL pipeline deformation monitoring into the calculation of displacements for unique feature points within the markers. Additionally, spraying markers onto the pipeline surface is preferable to affixing them directly, as it prevents positional shifts of the markers caused by pipeline deformation.
A monocular measurement method is employed to monitor GIL pipeline deformations, with the specific technical approach illustrated in Figure 1. The monitoring system’s hardware includes artificial markers, ArUco markers, a computer, and a camera, while the software algorithms are developed using Python 3.8 and Pytorch 2.0.0 deep learning platforms. For practical operational convenience, printed artificial markers are used during the monitoring process. The effectiveness of the proposed method is validated through displacement and twist angle measurements of a feature point on the GIL pipeline surface, demonstrating its feasibility for accurate and real-time deformation monitoring.
Figure 1. A real-time monitoring method for GIL deformation: overview and pipeline using YOLOv10 object detection and sub-pixel corner detection.
Initially, images containing markers are collected to create a dataset for training the YOLOv10 model. This enables the rapid and automatic localization of markers within the entire image, providing their positional information. Subsequently, sub-images, each containing a single marker, are extracted and processed to identify the coordinates of the target corner point within each marker that requires tracking. To address the loss of depth information inherent in monocular measurements, the known physical edge length of the ArUco marker is utilized. Additionally, the coordinates of the four corners of the ArUco marker are employed to calculate the projection transformation matrix, which is applied to correct image tilt and reduce measurement errors caused by the lack of perpendicularity between the camera’s optical axis and the imaging plane. Finally, by integrating the positional information of the marker detected by the YOLO model, the coordinates of the target corner point in the marker within the cropped sub-image, the known physical dimensions of the ArUco marker, and the perspective transformation matrix, the actual displacement of each corner point relative to the reference position can be calculated.

4. Experimental Analysis of GIL Deformation Measurements

As illustrated in Figure 10a, the measurement point arrangement for evaluating axial strain in pipelines involves selecting a smaller arc segment, AB, on the cross-section. In practical deformation measurements, such an approach can be adopted. Within this segment, n cross-sectional characteristic points and one end characteristic point are positioned, with the end characteristic point assumed to experience no strain. When the pipeline undergoes overall axial rigid body displacement, the rigid displacement is eliminated using the end characteristic point, resulting in the axial displacements at the n characteristic points attributable to pipeline deformation. Building on the findings of reference [22], Equation (15) is employed to define the extent of arc segment AB, thereby assuming that within this segment, the cross-sectional characteristic points lie within the same plane.
L m R sin 10 °
Figure 10. Schematic diagram of measurement point arrangement for pipeline deformation assessment. (a) Axial deformation measurement (b) Torsional deformation measurement.
In Equation (15), R represents the radius of the circle corresponding to arc segment AB, and Lm denotes the perpendicular distance from the midpoint of arc AB to chord AB. As illustrated in Figure 10b, the plane in which the measurement point induces the rotational angle coincides with the pipeline’s cross-sectional plane. When assessing the circumferential torsional deformation of a pipeline, it is assumed that no axial displacement occurs during the measurement period.

4.1. Axial Deformation Measurement at a Single Feature Point

To evaluate the proposed methodology, a simulation experiment was conducted using a pipeline structure similar to the GIL pipeline. In the simulation experiment, the displacement of a measurement point was measured and verified. To facilitate the validation of visual measurement results, limiting blocks were employed to constrain the pipeline’s displacement along the axial direction. During the experiment, axial force was applied to the simulated pipeline, inducing axial displacement. The deformation monitoring process was simulated by tracking the displacement changes at the target corner point. The experimental simulation setup is shown in Figure 11. To verify the accuracy of the proposed method, data acquisition was performed using both a displacement sensor and a camera. The displacement sensor operated at a sampling frequency of 600 Hz. To ensure synchronization, the camera is fixed in position throughout each measurement session. Initially, while the pipeline is stationary, a reference image is captured. Next, the pipeline is subjected to a controlled axial displacement, and the system is allowed to reach a new stationary state. Once the sensor’s displacement reading stabilizes, a subsequent image is taken. The displacement relative to the initial state is then determined by comparing the two images. Under these stationary conditions, the visual measurement aligns with the sensor’s reading. This ensures an accurate correspondence between the displacement measured by the vision system and that recorded by the sensor. Using this method, a total of 10 sets of images were obtained, with the initial image corresponding to a displacement of 0. Consequently, nine sets of displacement comparisons were acquired. This synchronization was critical for aligning the data streams and enabling a precise comparison of the measurement results. This experimental setup allowed for a controlled assessment of the proposed method’s precision and reliability under conditions mimicking real-world pipeline deformation scenarios.
Figure 11. Simulation experimental platform. (a) Schematic diagram of the operation. (b) Equipment arrangement.
The experiment was conducted four times to validate the robustness of the proposed measurement method. For each trial, varying magnitudes and directions of force were applied to the pipeline. Additionally, the pipeline was rotated appropriately to alter its orientation, thereby creating different angles between the camera and the measurement marker plane, as well as varying the distance between the camera and the target. These variations introduced diverse measurement conditions, enabling a comprehensive assessment of the method’s adaptability. The measurement results, shown in Figure 12, illustrate the displacement from the initial position. The data reflects the performance of the proposed method under varying experimental conditions, demonstrating its ability to adapt to changes in camera orientation and distance while maintaining measurement accuracy.
Figure 12. Results of simulation measurements.
To verify the accuracy of the visual measurement method, the measurements obtained from the wire-type displacement sensor were used as the reference. For a more intuitive presentation of the results, key metrics from the four experimental trials are summarized in Table 2. These metrics include the maximum absolute error (MaxAE), the mean absolute error (MAE1), the computation time per frame, and the root mean square error (RMSE). In a measurement, the amount of collected data is n. Let yi (i = 1, 2, …, n) represent the measured values obtained by the visual method, and yi′ represent the true values corresponding to the measured values. Then the MaxAE, MAE1, and RMSE of this set of data are determined by the following equations:
MaxAE = max ( y i y i , i = 1 , 2 , n ) MAE 1 = 1 n i = 1 n y i y i R M S E = 1 n i = 1 n y i y i 2
Table 2. Statistical analysis of visual measurement errors and time consumption.
The experimental results indicate that variations in applied forces, angles, and distances did not significantly impact measurement accuracy. The observed measurement errors remained consistently low, demonstrating the stability and reliability of the proposed method. The method proposed in reference [12] utilizes a multi-camera setup for 3D coordinate measurements, which monitors the displacement of markers affixed to the pipeline to assess GIL deformation, achieving a mean absolute error of less than 3 mm. In contrast, the monocular vision measurement method proposed in this study is simpler and more convenient. As shown in Table 2, the proposed method achieved a mean absolute error of no more than 1.337 mm across multiple measurements, with the minimum average absolute error being 0.265 mm. These results highlight the high accuracy of the proposed method. By flexibly positioning artificial markers, the method can adapt to various deformation scenarios in GIL pipelines. Moreover, improving the precision of the artificial markers and enhancing camera performance could further increase the accuracy of the measurements. Additionally, the average computation time per frame is approximately 0.024 s, demonstrating that the method is capable of real-time continuous measurement. The calculated corner coordinates of the artificial markers are unique, indicating that the measurement system achieves automated measurements, reduces human intervention, and offers significant practical value for real-world applications.

4.2. Torsional Deformation Measurements

To simulate the torsional deformation of actual pipelines, we measure the rotational angle of a rigid body about its axis. Specifically, we focus on measuring the rotation angle at a particular point on the cross-section. By manually applying torque at one end and utilizing the limit blocks illustrated in Figure 11, we ensure that the pipeline undergoes pure rigid-body rotation. An operational schematic is presented in Figure 13.
Figure 13. Operational schematic.
Additional experimental equipment includes a marker fixation device, protractor, string, and metal rings, with the experimental setup arranged as illustrated in Figure 14. The T-shaped welding platform shown in Figure 14a positions the line segment CD along the diameter of the end face circle. One end of the string is connected to the metal ring, while the other end is secured at point C. Under the influence of gravity, the string remains taut and vertically aligned, facilitating the measurement of the actual rotation angle. In the marker fixation device depicted in Figure 14b, point E corresponds to the upper-left corner of the ArUco marker, and point F represents the target corner point of the extracted marker. The mechanical structure ensures that both line segments EF and CD lie along the diameter of the end face circle.
Figure 14. Experimental apparatus layout. (a) End of the pipeline. (b) Marker fixation device.
To facilitate the validation of the visual measurement accuracy for rotational angles, a static measurement approach was employed. As illustrated in Figure 13, the rotational angle was measured from a completely stationary state A to a completely stationary state B. The geometric relationships involved in calculating the rotational angle during visual measurements are depicted in Figure 15.
Figure 15. Geometric relationships in rotation angle calculation.
Due to the geometric configuration, the angles α and β are equal. Therefore, measuring the β value at point C suffices to verify the accuracy of the visually measured rotation angle. Within the image pixel coordinate system under ouv, the rotation angle α is computed using Equation (17).
α = arccos F 1 E 1 F 2 E 2 F 1 E 1 F 2 E 2
The vectors F1E1 and F2E2 are calculated through points F and E in the pixel coordinates of the ouv coordinate system.
The camera was strategically positioned approximately 1.7 m from the measurement plane. Images of the pipe end face were captured under seven distinct states, with each state corresponding to a specific rotational angle increment. The coordinates of points E and F for each state are recorded in Table 3, with all values presented to two decimal places.
Table 3. Pixel coordinates of points E and F.
In Table 3, the pixel coordinates of point E remain unchanged before and after calibration. The coordinates of points E and F in states 1–6 are corrected using the correction matrix M of the 0-state image. The coordinates of point E in state 0 are unchanged before and after correction. This stability is attributed to the image tilt correction theory discussed in Section 3.3, where point E serves as the reference point for the projection transformation. Utilizing Equation (17), the angles between each pair of states were calculated (α). The angles before image calibration (pre-α) and their absolute errors (pre-AE) were computed, as well as the angles after image calibration (post-α) and their absolute errors (post-AE). The calculated results, along with the actual measurements, are summarized in Table 4, with all values rounded to two decimal places. Table 4 presents the angle value.
Table 4. Angle Values.
To clearly illustrate the variations in the measurement results, the data from Table 4 are depicted in Figure 16. As shown in Figure 16, using the experimentally measured range of rotation angles as an example, directly converting the actual rotation angle from the pixel coordinates provides relatively accurate results, with the maximum deviation from the actual value being less than 0.5°. This finding also suggests that the camera’s optical axis is nearly perpendicular to the measurement plane. After applying the correction, the measured rotation angles deviate only slightly from the actual values. The largest error, 0.73°, occurs between State 1 and State 2, indicating that this tilt correction technique is also effective for measuring torsional deformation angles in GIL. In reference [35], the authors employed binocular digital image correlation technology to measure the shear angle of torsional deformation in carbon steel cylinders, achieving an MAE of 0.35°, which adequately meets measurement needs. Based on the measurement data from the reference, this study introduces a new measurement method for the torsional deformation of GIL, achieving an MAE of 0.31° within a certain measurement range. This method is considered to be an effective measurement approach to some extent, although the specific applicable measurement range requires further investigation.
Figure 16. Comparison of measured angles.

5. Conclusions

Based on the characteristics of GIL pipelines, this study integrates the constructed YOLOv10 model, high-precision corner detection methods, and image tilt correction techniques to propose an automated, real-time monitoring method for GIL pipeline deformation. The key conclusions of the research are as follows:
(1)
To address the limitations of traditional FAST corner detection methods, an adaptive threshold calculation method was developed. Combined with the Shi–Tomasi algorithm, redundant corner points were effectively removed, enabling the automated extraction and tracking of the target corner point at the center of the marker. Experiments verified the effectiveness of the proposed image tilt correction theory in monocular measurement and analyzed the impact of subpixel refinement on measurement error. The results show that, after image correction, measurement accuracy became insensitive to the camera’s orientation. The pixel-level measurement method achieved a mean absolute error (MAE) of 1.474 mm with a standard deviation of 1.0178 mm, while the subpixel-level measurement method improved the MAE to 0.791 mm with a standard deviation of 0.2953 mm. This demonstrates that the subpixel-level method offers higher measurement precision and better data stability, making it advantageous for practical measurement scenarios.
(2)
Simulated deformation experiments verified the robustness, precision, real-time capability, and automation of the proposed monitoring system. The results indicate that the system is insensitive to changes in pipeline loading conditions and camera orientation. The average absolute error of the measurements did not exceed 1.337 mm, with the minimum average absolute error being 0.265 mm, meeting the required measurement accuracy. Furthermore, the computation time per frame was approximately 0.024 s, enabling real-time, automated deformation monitoring.
(3)
The proposed GIL pipeline deformation measurement method was validated through torsional deformation experiments. When the camera was positioned approximately 1.7 m from the measurement plane, the maximum average absolute error did not exceed 0.31°. It is important to note that, regardless of whether axial or torsional deformation is being measured, the projective transformation matrix M should always be calculated from the first image acquired during each measurement session. This same matrix should then be applied to all subsequent images, ensuring consistent and accurate correction.
In this study, we selected a camera with less distortion. Although the current work does not delve deeply into intrinsic camera calibration, future research will explore comprehensive calibration methods to further enhance measurement accuracy without increasing manual intervention. Additionally, further optimization of the YOLO-based model will also be pursued to enhance its specificity for GIL deformation, improve the speed of region of interest extraction, and reduce computational overhead. Exploring multi-camera systems and alternative AI techniques will also be considered to more effectively capture three-dimensional deformations.

Author Contributions

Study conceptualization: G.Y., W.Y. and M.W. Methodology development: G.Y., H.H. and M.W. Code implementation and results analyses: G.Y. and Q.W. Data Acquisition: G.Y. and E.L. Writing of the manuscript: G.Y., W.Y. and Q.W. Study supervision: W.Y., M.W. and J.S. Funding acquisition: Q.W. and J.S. All authors revised and commented on the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This study was financially supported by the National Natural Science Foundation of China (Grant No.: U21A20117), and the Fundamental Research Funds for the Central Universities (Grant No.: 2023MS134).

Data Availability Statement

Data will be made available on request.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Xiao, S.; Zhang, X.; Tang, J.; Liu, S. A review on SF6 substitute gases and research status of CF3I gases. Energy Rep. 2018, 4, 486–496. [Google Scholar] [CrossRef]
  2. Magier, T.; Tenzer, M.; Koch, H. Direct Current Gas-Insulated Transmission Lines. IEEE Trans. Power Deliv. 2018, 33, 440–446. [Google Scholar] [CrossRef]
  3. Tu, Y.; Chen, G.; Li, C.; Wang, C.; Ma, G.; Zhou, H.; Ai, X.; Cheng, Y. ±100-kV HVDC SF6/N2 Gas-Insulated Transmission Line. IEEE Trans. Power Deliv. 2020, 35, 735–744. [Google Scholar] [CrossRef]
  4. Shu, Q.; Huang, Z.; Yuan, G.; Ma, W.; Ye, S.; Zhou, J. Impact of wind loads on the resistance capacity of the transmission tower subjected to ground surface deformations. Thin-Walled Struct. 2018, 131, 619–630. [Google Scholar] [CrossRef]
  5. Fu, X.; Du, W.-L.; Li, H.-N.; Li, G.; Dong, Z.-Q.; Yang, L.-D. Stress state and failure path of a tension tower in a transmission line under multiple loading conditions. Thin-Walled Struct. 2020, 157, 107012. [Google Scholar] [CrossRef]
  6. Tan, X.; Poorghasem, S.; Huang, Y.; Feng, X.; Bao, Y. Monitoring of pipelines subjected to interactive bending and dent using distributed fiber optic sensors. Autom. Constr. 2024, 160, 105306. [Google Scholar] [CrossRef]
  7. Yang, L.; Fan, J.; Liu, Y.; Li, E.; Peng, J.; Liang, Z. A Review on State-of-the-Art Power Line Inspection Techniques. IEEE Trans. Instrum. Meas. 2020, 69, 9350–9365. [Google Scholar] [CrossRef]
  8. Yang, Y.; Wang, M.; Wang, X.; Li, C.; Shang, Z.; Zhao, L. A Novel Monocular Vision Technique for the Detection of Electric Transmission Tower Tilting Trend. Appl. Sci. 2023, 13, 407. [Google Scholar]
  9. Wang, D.; Yue, J.; Li, J.; Xu, Z.; Zhao, W.; Zhu, R. Research on sag monitoring of ice-accreted transmission line arcs based on stereovision technique. Electr. Power Syst. Res. 2023, 225, 109794. [Google Scholar] [CrossRef]
  10. Zhang, K.; Liu, J.; Li, Y.; Sun, C.; Zhang, L. Research on Improving Denoising Performance of ROI Computer Vision Method for Transmission Tower Displacement Identification. Energies 2023, 16, 539. [Google Scholar] [CrossRef]
  11. Li, J.; Zheng, H.; Liu, P.; Liang, Y.; Shuang, F.; Huang, J. Safety monitoring method for powerline corridors based on single-stage detector and visual matching. High. Volt. 2024, 9, 805–815. [Google Scholar] [CrossRef]
  12. Fauzan, K.N.; Suwardhi, D.; Murtiyoso, A.; Gumilar, I.; Sidiq, T.P. Close-range photogrammetry method for SF6 Gas Insulated Line (GIL) deformation monitoring. Int. Arch. Photogramm. Remote Sens. Spatial Inf. Sci. 2021, XLIII-B2-2021, 503–510. [Google Scholar] [CrossRef]
  13. Xu, Y.; Brownjohn, J.; Kong, D. A non-contact vision-based system for multipoint displacement monitoring in a cable-stayed footbridge. Struct. Control Health Monit. 2018, 25, e2155. [Google Scholar] [CrossRef]
  14. Brownjohn, J.M.W.; Xu, Y.; Hestr, D. Vision-based bridge deformation monitoring. Front. Built Environ. 2017, 3, 23. [Google Scholar] [CrossRef]
  15. Yang, Y.; Ma, T.; Du, S. Robust template matching with angle location using dynamic feature pairs updating. Appl. Soft Comput. 2019, 85, 105804. [Google Scholar] [CrossRef]
  16. Wu, T.; Dong, Y. YOLO-SE: Improved YOLOv8 for remote sensing object detection and recognition. Appl. Sci. 2023, 13, 12977. [Google Scholar] [CrossRef]
  17. Li, Y.; Liu, W.; Li, L.; Zhang, W.; Xu, J.; Jiao, H. Vision-based target detection and positioning approach for underwater robots. IEEE Photonics J. 2022, 15, 8000112. [Google Scholar] [CrossRef]
  18. Yang, H.; Chen, L.; Chen, M.; Ma, Z.; Deng, F.; Li, M.; Li, X. Tender tea shoots recognition and positioning for picking robot using improved YOLO-V3 model. IEEE Access 2019, 7, 180998–181011. [Google Scholar] [CrossRef]
  19. Bansal, M.; Kumar, M.; Kumar, M. 2D object recognition: A comparative analysis of SIFT, SURF and ORB feature descriptors. Multimed. Tools Appl. 2021, 80, 18839–18857. [Google Scholar] [CrossRef]
  20. Li, Y.; Zheng, W.; Liu, X.; Mou, Y.; Yin, L.; Yang, B. Research and improvement of feature detection algorithm based on FAST. Rend. Lincei 2021, 32, 775–789. [Google Scholar] [CrossRef]
  21. Zhuang, Y.; Chen, W.; Jin, T.; Chen, B.; Zhang, H.; Zhang, W. A Review of Computer Vision-Based Structural Deformation Monitoring in Field Environments. Sensors 2022, 22, 3789. [Google Scholar] [CrossRef] [PubMed]
  22. Choi, I.; Kim, J.; Kim, D. A target-less vision-based displacement sensor based on image convex hull optimization for measuring the dynamic response of building structures. Sensors 2016, 16, 2085. [Google Scholar] [CrossRef] [PubMed]
  23. Shan, W.; Bai, F.; Xu, Y.; Gao, X.; Li, P. Perspective deformation correction for circular pointer meter based on prior indication structure feature. Measurement 2024, 229, 114423. [Google Scholar] [CrossRef]
  24. Barath, D. On Making SIFT Features Affine Covariant. Int. J. Comput. Vis. 2023, 131, 2316–2332. [Google Scholar] [CrossRef]
  25. Yang, Z.; Shen, D.; Yap, P.-T. Image mosaicking using SURF features of line segments. PLoS ONE 2017, 12, e0173627. [Google Scholar] [CrossRef]
  26. Damaneh, M.M.; Mohanna, F.; Jafari, P. Static hand gesture recognition in sign language based on convolutional neural network with feature extraction method using ORB descriptor and Gabor filter. Expert. Syst. Appl. 2023, 211, 118559. [Google Scholar] [CrossRef]
  27. Wang, A.; Chen, H.; Liu, L.; Chen, K.; Lin, Z.; Han, J.; Ding, G. YOLOv10: Real-Time End-to-End Object Detection. arXiv 2024, arXiv:2405.14458. [Google Scholar]
  28. Lee, Y.H.; Chen, T.C.; Liang, H.C.; Liao, J.X. Algorithm and Architecture Design of FAST-C Image Corner Detection Engine. IEEE Trans. Very Large Scale Integr. (VLSI) Syst. 2021, 29, 788–799. [Google Scholar] [CrossRef]
  29. Hong, Q.; Jiang, H.; Xiao, P.; Du, S.; Li, T. A Parallel Computing Scheme Utilizing Memristor Crossbars for Fast Corner Detection and Rotation Invariance in the ORB Algorithm. IEEE Trans. Comput. 2024, 74, 996–1010. [Google Scholar] [CrossRef]
  30. Bansal, M.; Kumar, M.; Kumar, M.; Kumar, K. An efficient technique for object recognition using Shi-Tomasi corner detection algorithm. Soft Comput. 2021, 25, 4423–4432. [Google Scholar] [CrossRef]
  31. Zhu, Q.; Wu, B.; Wan, N. A sub-pixel location method for interest points by means of the Harris interest strength. Photogramm. Rec. 2007, 22, 321–335. [Google Scholar] [CrossRef]
  32. Xie, X.; Ge, S.; Xie, M.; Hu, F.; Jiang, N. An improved industrial sub-pixel edge detection algorithm based on coarse and precise location. J. Ambient. Intell. Humaniz. Comput. 2020, 11, 2061–2070. [Google Scholar] [CrossRef]
  33. Liu, Y.; Su, X.; Guo, X.; Suo, T.; Yu, Q. A novel concentric circular coded target, and its positioning and identifying method for vision measurement under challenging conditions. Sensors 2021, 21, 855. [Google Scholar] [CrossRef]
  34. Becerra, L.L.; Ferrua, J.A.; Drake, M.J.; Kumar, D.; Anders, A.S.; Wang, E.N.; Preston, D.J. Active fume hood sash height monitoring with audible feedback. Energy Rep. 2018, 4, 645–652. [Google Scholar] [CrossRef]
  35. Wang, L.; Feng, Y. An Adaptive Window Shape-Based DIC Method for Large Torsional Deformation Measurement. IEEE Trans. Instrum. Meas. 2023, 72, 5004510. [Google Scholar] [CrossRef]
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.