Next Article in Journal
Research on the Analytical Model of Improved Magnetic Flux Leakage Signal for the Local Stress Concentration Zone of Pipelines
Previous Article in Journal
Measuring Setup for Experimental Research of Two-Coordinate Fiber-Optic Acceleration Sensors with Cylindrical Lenses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Real-Time Φ-OTDR Vibration Event Recognition Based on Image Target Detection

1
Data and Target Engineering Institute, PLA Strategic Support Force Information Engineering University, Zhengzhou 450001, China
2
Research Institute for National Defense Engineering of Academy of Military Science PLA, Luoyang 471023, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(3), 1127; https://doi.org/10.3390/s22031127
Submission received: 18 January 2022 / Revised: 30 January 2022 / Accepted: 31 January 2022 / Published: 2 February 2022
(This article belongs to the Section Sensing and Imaging)

Abstract

:
Accurate and fast identification of vibration signals detected based on the phase-sensitive optical time-domain reflectometer (Φ-OTDR) is crucial in reducing the false-alarm rate of the long-distance distributed vibration warning system. This study proposes a computer vision-based Φ-OTDR multi-vibration events detection method in real-time, which can effectively detect perimeter intrusion events and reduce personnel patrol costs. Pulse accumulation, pulse cancellers, median filter, and pseudo-color processing are employed for vibration signal feature enhancement to generate vibration spatio-temporal images and form a customized dataset. This dataset is used to train and evaluate an improved YOLO-A30 based on the YOLO target detection meta-architecture to improve system performance. Experiments show that using this method to process 8069 vibration data images generated from 5 abnormal vibration activities for two types of fiber optic laying scenarios, buried underground or hung on razor barbed wire at the perimeter of high-speed rail, the system [email protected] is 99.5%, 555 frames per second (FPS), and can detect a theoretical maximum distance of 135.1 km per second. It can quickly and effectively identify abnormal vibration activities, reduce the false-alarm rate of the system for long-distance multi-vibration along high-speed rail lines, and significantly reduce the computational cost while maintaining accuracy.

1. Introduction

Phase-sensitive optical time-domain reflectometer (Φ-OTDR) is a simple and effective method to measure single-mode fiber vibration [1,2], which has the advantages of distributed detection capability up to 100 km at a single station [3], as well as high resolution, anti-electromagnetic interference, corrosion resistance, low energy loss, flame and explosion resistance [4], and can be combined with Raman pumping amplification to achieve long-range distributed abnormal vibration location inspection [5], which makes Φ-OTDR extensively employed in perimeter security and oil pipeline monitoring.
However, the long-range detection and high sensitivity of Φ-OTDR will inevitably result in high nuisance-alarm rates (NARs). In the case of vibration detection along the fence and the ground near the fence of the high-speed railway (HSR) at this study, noise mainly comes from the frequency drift and linewidth range of the optical devices, thermal noise, and scattering noise of electronic devices. The interference signals are caused by the airflow in the detection environment. Based on the hardware realization of long-distance vibration detection, the subsequent signal processing and vibration classification problems become the essential factors for Φ-OTDR to be able to identify abnormal vibration signals precisely, and the precision of the identification significantly affects the eventual false alarm probability of the system. For the nonstationary vibration signals detected by Φ-OTDR, wavelet noise reduction can be used for long-range slow time-domain perturbation signal purification to achieve vibration localization [5]. Using wavelet packet decomposition to extract the features of each energy layer of vibration signal, combined with an artificial neural network classifier, high-precision recognition for vibration signals can be achieved [6]. Short-time Fourier transform is also a typical feature extractor, and combined with a Gaussian mixture model or convolutional neural network can complete distinct intrusion activity detection and classification [7,8].
The above identification methods extract the slow time-domain signal features of a vibration location and then distinguish various vibration activities by a well-designed classifier. Nevertheless, since the fiber strain and refractive index near the vibration activity will vary, it is complicated to determine the vibration source location in practical engineering applications. The deviation of vibration source location selection will also decrease the recognition rate. In contrast, morphological feature extraction based on spatio-temporal domain images can improve the recognition rate by adding contextual features [9]. When the morphologic features of time-space domain signals are extracted and obtained in the experiment, the relevance vector machine (RVM) can be used to achieve a relatively short recognition time of below 1 s for vibration behavior [10]. In addition, convolutional neural networks (CNNs) can effectively classify a single vibration activity image without expert knowledge [11,12].
For 100-km level detection, however, various vibration activities may occur simultaneously in a single spatio-temporal domain image, and the real-time capability of the system cannot be guaranteed on account of the apparent increment in total recognition time. This problem necessitates the detection of multiple vibrational activity classes in a single vibration alarm event spatio-temporal image by the target detection algorithm of computer vision. Convolutional neural network-based target detectors are distinguished into two categories: region-based CNNs (R-CNNs) [13] and their variants [14,15] belong to two-stage target detectors, which utilize the region-based proposal sliding window method to monitor targets in images; in contrast to two-stage detectors, single-step detectors, such as You Only Look Once (YOLO) detectors [16] and their variants [17,18,19], employ down-sampling components and concatenation components or a path aggregation network to accomplish feature extraction of the original image in different dimensions, directly mapping from feature maps in different sizes to target probabilities, bounding box locations, and target classification probabilities, which can increase the number of detected frames per second with comparable performance [19]. YOLO detectors are the State-Of-The-Art (SOTA) object detection algorithms in deep learning [20,21], and are real-time and efficient in practical applications for portable embedded devices and have significant advantages in terms of accuracy and processing speed [22]. While the feature extraction network in the existing YOLO detection framework is designed for publicly available target detection datasets [23,24], the corresponding feature extraction network needs to be improved for a customized dataset.
Considering the Φ-OTDR high-speed railway perimeter security detection environment, this paper aims to complete monitoring of multiple vibration images in finite time and complete multiple vibration target identification and localization in single image detection. High-speed rail perimeter security is mainly threatened by climbing or jumping over the fence. Early warning threats include people walking on the ground near the fence, manual excavation, construction machinery, and interference, primarily from the shaking of optic fiber caused by airflow due to natural wind or high-speed rail. Single-mode optical fiber is laid on the razor barbed wire on the fence of the high-speed railway station to detect the vibration signals of people climbing or jumping over the fence and the shaking caused by airflow. About 1 km long single-mode optical fiber is buried at 100 m from the initial fiber position and 20 cm underground to detect the early warning activities of people walking, manual excavation, and mechanical construction close to the fence. The raw signals acquired by Φ-OTDR are purified by signal pre-processing and image pre-processing to form the ultimate dataset. An improved YOLO model is trained and evaluated to accurately and quickly recognize abnormal vibration events. The contributions of this study are: (1) The effect of parameter adjustment in the signal pre-processing module on inspection is illustrated from the perspective of statistical theory; (2) An improved YOLO model for high-speed railway perimeter abnormal vibration detection is proposed to achieve single detection of multiple vibration targets; (3) The effect of different feature extraction networks on the detection performance of the improved YOLO model is tested and discussed; (4) The detection performance of the improved YOLO model with other State-of-The-Art (SOTA) models has been compared and summarized.
This paper is organized as follows: Section 2 describes the entire system structure and analyzes the statistical characteristics of the raw signal acquired by Φ-OTDR. Section 3 presents the principle of demodulating vibration signals from raw signals and proposes signal pre-processing methods to improve the signal-to-noise ratio (SNR) and detection probability, thereby illustrating the signal pre-processing procedure. Sequentially, it introduces the way to acquire vibration data and generate the spatio-temporal images to be identified. Section 4 discusses the experimental dataset generation method and the components of the multi-target detection system, the effects of different feature extraction network components on the detection performance of the data images, the basis for model improvement, and the evaluation metrics. Section 5 focuses on the experimental procedure, examines and compares the impact of different feature extraction networks on recognition, and compares the performance of the improved YOLO model with other SOTA models. The last section concludes the paper.

2. Baseline System

2.1. System Overview

For the high-speed railway perimeter detection environment with many noise and interference signals, to quickly and accurately identify the vibration source class and thus reduce the false-alarm rate, it is necessary first to design the signal pre-processing algorithm and set the corresponding signal acquisition parameters based on the optical pulse detection principle. Up to this point, the system has removed most of the noise to improve the SNR of the Φ-OTDR detection signal and effectively detect vibration activities. In practice, airflow can generate interference vibration signals, so such interference signals need to be effectively identified and excluded from the necessary alarm to improve the reliability of the overall system alarm. In this regard, it is essential to generate images from the pre-processed data matrix, enhance the vibration target features, and finally use an improved image target detection algorithm to identify all classes of vibration activities accurately.
As shown in Figure 1, the real-time Φ-OTDR vibration events detection system contains three components: (1) Abnormal vibration detection environment, which is the environment for abnormal vibration detection at the perimeter of the high-speed railway. The primary monitoring events are the five kinds of vibration signals mentioned above. The single-mode optical fiber is laid in the detection environment by burying underground or hanging on the razor barbed wire to realize the distributed real-time detection of abnormal vibration. (2) In the Φ-OTDR system, light probe pulses generated by a coherent laser source with narrow linewidth are injected into the single-mode fiber through an optical circulator. Due to the elastic scattering effect, the backscattered light interferes with each scattering center in the fiber to generate Rayleigh backscattered light. The backscattered light is converted into a digital signal by the data acquisition (DAQ) card through the photodetector (PD). Then, the digital signal is transmitted to the algorithm processing module. (3) The algorithm processing module functions as follows:
  • The original digital signal is processed by the signal processing unit to accomplish pulse accumulation, pulse cancellers, maximum value processing to generate the primary data matrix, and then determine the location of the vibration signal appearing on the detection fiber and the corresponding time according to the statistical detection principle;
  • On this basis, the data matrix is converted into grayscale images and then processed by median filtering and pseudo-color to generate the final spatio-temporal image of vibration data to be detected;
  • The images are detected by the improved YOLO target detector, which first completes the multi-scale feature conversion by the feature extraction network, then uses the classifier to achieve the detection of multiple vibration targets in a single image, and eventually improves the detection accuracy and reduces the false-alarm rate of the system.
In this application, the raw signal acquired by Φ-OTDR is processed by the signal processing unit and spatio-temporal image generation module to generate a customized dataset. The dataset is subsequently divided into the training set, validation set, and test set. The training and validation sets are employed to train and modify the improved YOLO target detection model. The test set is applied to test and evaluate the model performance, then select the optimal model.
On this foundation, the spatio-temporal image, which is generated from real-time vibration data, is transmitted to the improved optimal model to obtain the classes of multiple vibration activities and their locations in the image, thereby reducing the false-alarm rate and ensuring the real-time effectiveness of the distributed fiber optic perimeter security system.

2.2. Φ-OTDR Raw Signal Statistical Characteristics

The signal processing unit utilizes the statistical hypothesis detection principle to obtain the location and corresponding time of all vibration events on the detection fiber, and uses signal processing methods to improve the signal-to-noise ratio to enhance the accuracy and precision of subsequent pattern recognition. Φ-OTDR backscattered light is generated by the backscattering interference of coherent pulsed light from 2 N   + 1 , elastic scattering centers in the region covered by half of the optical pulse width; refer to Figure 2a.
Define the attenuation coefficient of the fiber as α , after Acousto-optic Modulator (AOM) modulation and Erbium Doped Fiber Amplifier (EDFA) amplification, a sequence of coherent optical pulses of periodic width T c / n is injected into a single-mode fiber with group refractive index n , the optical field is A 0 e x p ( j 2 π F t 1 + θ t 2 ) , t 2 records the slow-time domain pulse emission time, and pulse repetition interval (PRI) indicates the difference in emission time of neighbor pulses. The specific backscattering process can be analyzed from the change in position of the light pulse on the fiber at different moments in a single detection.
Considering t 2 as the starting time, the backscattered light from the light pulse at the moment t 1 in the fast-time domain is converted into an electrical signal by a PD, and then quantized and saved by a DAQ. The integrated reflection coefficient σ at position t 1 c / n is derived from the combined effect of the 2 N   + 1 scattering centers in range l t 1 t 1 c / n , 2 t 1 c + T c / 2 n . Specifically, at the moment t , the central position of the light pulse propagates to t 1 c / n , and the backscattered light from the scattering center at 2 t 1 c + T c / 2 n begins to scatter in the opposite direction of the light pulse propagation. t + Δ t moment, the above backscattered light propagates to 2 t 1 c + T c 2 Δ t c / 2 n , during which the backscattered light is the superposition of a series of small Rayleigh scattering centers within Δ t c / n . When the pulsed light propagates half the light pulse width, i.e., the tail end of the pulsed light propagates to t 1 c / n , the backscattered light at this point is superposed by the backscattered light from 2 N   + 1 scattering centers at different spatial positions within half the light pulse width l t 1 , and there is no further superposition. In the range of l t 1 , the reflectivity corresponding to the i th scattering center is σ i , the distance to the t 1 c / n position is l i , the echo phase is e x p ( j 2 π F t 1 2 n l i / c + θ t 2 ) , and θ t 2 indicates the initial phase of the pulsed light. The final superposed echo signal ζ t 1 , t 2 at the t 1 c / n position can be expressed as:
ζ t 1 , t 2 = A 0 i = N N σ i e x p α l i t 1 e x p j 2 π F t 1 2 n l i t 1 / c + θ t 2 = A 0 e x p ( α l t 1 ¯ ) e x p ( j 2 π F t 1 + θ t 2 ) i = N N σ i e x p ( j 4 π n l i / λ )
where λ represents light wavelength. Equation (1) reveals that the relative positions l i of all scattering centers in the half-wavelength region are interrelated with the echo signal ζ t 1 , t 2 . In the presence of vibrations in the area, the vibration-induced fiber strain leads to changes in l i and refractive index n at different positions, which in turn cause shifts in the backscattered light phase and eventually change ζ t 1 , t 2 . As demonstrated in Figure 2b, both the variation of the relative position of the scattering center within the half-light pulse width l t 1 and the change of the refractive index n lead to the shift of the echo signal at t 1 c / n . Furthermore, when the scattering centers within the vibration area L are disturbed, the vibration interferes with the backscattered light phase, leading to a shift in the echo signal within the range T c / 2 n + L . The location of the vibration occurrence is judged by detecting the area where the echo signal shift is located.
For analysis of the echo signal, define ξ = e x p ( j 2 π F t 1 + θ t 2 ) i = N N σ i e x p ( j 4 π n l i / λ ) = σ t 1 , t 2 e x p ( j ϕ t 1 , t 2 ) , ϕ i = 4 π n l i / λ . According to Euler’s formula, the integrated reflection coefficient σ t 1 , t 2 and the integrated phase fluctuation ϕ t 1 , t 2 can be expressed as:
σ t 1 , t 2 = ( i = N N σ i   sin ϕ i ) 2 + ( i = N N σ i cos ϕ i ) 2 ϕ t 1 , t 2 = arctan ( i = N N σ i   sin ϕ i / i = N N σ i cos ϕ i )
According to the central limit theorem, the real and imaginary parts of i = N N σ i e x p ( j 4 π n l i / λ ) are statistically independent Gaussian random variables, and the echo phase is uniformly distributed, so ξ is Rayleigh distributed, and σ t 1 , t 2 is exponentially distributed [25]. It is observed that the phase jitter θ t 2 of the pulsed light at different moments does not affect the final acquisition results, avoiding the effect of phase noise introduced by the laser and the AOM [26].
The photodetector output electrical signal can be expressed as the following Equation:
I t 1 , t 2 = ρ ζ ζ * = ρ A 0 2 e 2 α l ( t 1 ) ¯ ξ 2 = ρ A 0 2 e 2 α l ( t 1 ) ¯ σ t 1 , t 2
where the photoelectric conversion rate ρ , the initial optical power A 0 2 and the light decay rate α can be considered as fixed values, while the variation of the collected vibration data is from the variation of the integrated reflection coefficient σ t 1 , t 2 . The electrical signals received by the DAQ in the absence of vibration signals are the coherent noise power at each location along the fiber.

3. Raw Signal Pre-Processing and Image Generation

3.1. Analysis of Vibration Detection Principles

According to the analysis in Section 2.2, the real and imaginary parts of the optical echo signal z at the position t 1 c / n of the fiber can be considered as statistically independent zero-mean Gaussian random variables when the fiber is devoid of vibration in the range l t 1 . The backscattered signal z is the sum of the Gaussian white noise sample w and the target component s in the presence of a vibrational signal in this range, where s is a complex constant of unknown amplitude σ s and phase ϕ s , s = σ s e x p ( j ϕ s ) , during one sweep. The signal z received by the photodetector in one inspection decision can be expressed as the following two hypotheses:
H 0 : z = w         w ~ C N 0 , δ w 2 t 1 / 2 H 1 : z = s + w                   s = σ s e j ϕ s
where δ w 2 t 1 is related to the coherent noise at t 1 c / n . The null hypothesis H 0 indicates that there is no vibration event within the corresponding detection location, and the probability density function (PDF) of y = z is Rayleigh distributed. The PDF of y in the alternative hypothesis H 1 is the Rice distribution, cf. Equation (5), where I 0 · is the modified Bessel function of the first kind:
p y y | σ s , H 1 = 2 y δ w 2 e x p y 2 + σ s δ w 2 I 0 2 σ s y δ w 2 y 0 0 y < 0
For a multi-pulse sampling system with a noncoherent integration of M samples, a modified Bessel function of the first kind approximation leads to the statistic y , where λ is the detection threshold satisfying a fixed false-alarm rate:
y = m = 0 M 1 y m 2 H 1 H 0 δ w 4 σ s ln ( λ + σ s δ w 2 ) = λ
The presence of a vibration activity is detected by determining whether the sum of the squares of the acquired signal amplitudes, y , exceeds the threshold λ . If H 0 best accounts for the data, and the false alarm probability P FA is a fixed value, the threshold λ is only related to the number of noncoherent integration samples, M , and coherent noise δ w 2 . If H 1 best accounts for the data, the integrated reflection coefficient σ s , which is independent in all M samples of a single detection decision, shows different distribution characteristics p σ σ s for different vibration sources [7,27]. The PDFs of the final statistic y corresponding to the different vibration modes can be found using characteristic functions. The detection probability of different vibrations is the integral value of the corresponding PDF within the range exceeding the threshold.
Either improving the system signal-to-noise ratio or increasing the noncoherent integration samples number M can lead to an increase in the detection probability P D of the system for vibration signals. High-frequency noise can be suppressed by pulse cancellers to improve the system detection probability from the perspective of improving the signal-to-noise ratio. In addition, increasing the noncoherent integration samples number M can improve the detection performance with a constant SNR.

3.2. Signal Pre-Processing Procedure

In the signal acquisition process, there are M p detection pulses contained within one detection sweep. The backscattered light is converted to an electrical signal I after PD, and the detection data collected by DAQ at the fixed position t 1 c / n for a single pulse is x = I t 1 , t 2 = ρ A 0 2 e x p 2 α l t 1 ¯ σ t 1 , t 2 . An array of detection data X = x 0 ,   x 1 , x m , x M p 1 can be obtained at the fixed fiber position in one detection sweep. The number of noncoherent integration samples is M , the pulse canceller interval is M g , and detection data array is converted into M c = M p M 1 M g statistics X = x 0 ,   x 1 , x i , x M c 1 , where:
x i = m = i M 1 + i ( x m x m + M g ) ,   i 0 , M c 1
as the final detection value x m a x = max X is taken and different threshold values are set at different locations. The vibration signal is considered to exist if x m a x exceeds the threshold. The above steps are repeated for different detection locations within one sweep, thus the vibration detection is completed from M p pulses to one whole fiber. Moreover, to improve the reliability of the detection decision, the decision rule is defined as a vibration activity being detected at least K times in multiple decisions before it is finally determined as a valid vibration activity, called coincidence detection, which can further improve the detection probability and reduce the false-alarm rate. Vibration detection performance can be optimized by adjusting M p , M g , M and K in the signal pre-processing procedure. After setting the parameters, the data matrix collected in real-time by the system injecting light pulses periodically into the optical fiber can generate the corresponding grayscale images through numerical mapping for later image pre-processing and target detection.

3.3. Median Filtering and Pseudo-Color Processing

As described in Section 3.2, one array of detection data was obtained from M p detection pulses, and the vibration spatio-temporal data matrix can be generated during continuous system detection. While there is still a certain probability of false alarm after effectively improving the detection probability, discrete noise will appear if the spatio-temporal image is generated directly. Before obtaining the spatio-temporal images, the grayscale images generated from the data matrix need to be pre-processed to reduce the impact of image noise on later multi-target detection, enabling the targets in the images to be better recognized, and the classification accuracy to be improved [28]. The system uses median filtering to smooth the image pixel values to achieve the effect of denoising [29]. The median filter takes the median value of adjacent pixels as the output of the current pixel to suppress high-frequency noise before image detection. Afterward, the data image is processed using the pseudo-color image fusion method with low computational cost to achieve image enhancement by assigning different colors to the gray levels in the image [30]. The fused image can enhance the performance of object segmentation, feature extraction, and target object detection [31]. The system generates in real-time the final spatio-temporal images to be detected after the process described in Section 3.2 and Section 3.3.

3.4. Vibration Data Collection

Fiber optic sensing technology overcomes the shortcomings of high-speed rail tracking circuits easily damaged by lightning. Not only can it obtain the running status of the train in real-time [32], but it also can play a well-assisted role in high-speed railway perimeter intrusion detection [33]. The abnormal vibration to be detected by the Φ-OTDR high-speed railway perimeter security warning system originates primarily from construction machinery, manual excavation, walking near fences, climbing fences, airflow induced by natural wind, or high-speed rail passing by. The first three types of early warning need to be monitored by buried optical fiber cable. The last two types are monitored by optical fiber cable hung on razor barbed wire, where fence climbing needs to be an alarm in real-time, and vibration caused by airflow does not require an alarm.
The fiber optic sensing device uses a narrow linewidth (~1.69 kHz) laser at 1550 nm with a typical output power of 13 dBm. Optical frequency shifts 200 MHz by controlling the acousto-optic modulator with 50 ns pulse width. Detailed parameters of the device signal processing are as follows: the optical pulse repetition frequency is 5 kHz, the sampling rate is 250 MS/s for the DAQ, M p as the number of samples in one single sweep is 1000, the noncoherent integration samples number M is 880, and the pulse canceller interval M g is 100. One pixel corresponds to a distance dimension of 0.4 m and a time dimension of 0.2 s in the generated spatio-temporal vibration image. Different threshold adjustment factors and K values were set at different monitoring areas of the security perimeter of the HSR station to reduce the effect of coherent noise.
The system collected vibration data images of different classes at four ground detection areas and multiple razor barbed wire detection areas in 30 consecutive days, and vibration data images of changes over the spatio-temporal domain were generated in real-time after signal pre-processing and image processing. The fence climbing and airflow activities were sensed by monitoring the optical fiber cable hung on razor barbed wire. Three kinds of activities nearby the fence, namely construction machinery, manual excavation, and walking, were sensed by monitoring the optical fiber cable laid about 20 cm underground and parallel to the fence at a distance of 0–1 km from the starting position of the fiber optic. The experimenter collected data images by walking within the warning area directly above the buried optical fiber cable nearby the fence, and also simulated construction machinery and manual excavation activities by compacting the ground using a pneumatic hammer and hitting the ground using a shovel at a distance of about 0.4 m perpendicular to the buried optical fiber cable laying direction, respectively. Vibration data for the above 3 types of activities were collected by the real activities of the experimenters. Razor barbed wire binding optical fiber cable was shaken with a wooden stick to simulate climbing over the fence. The experimenters used a larger force with a wooden stick in the actual data collection to simulate a real climbing event in order to try to achieve a more realistic simulation. Considering the real-time requirement of the alarm and the authenticity of vibration activities, the duration of a construction simulation activity is 3–6 s, the duration of a person walking simulation activity is 3–15 s, and the duration of a simulation activity of climbing over the fence is 5–18 s. Furthermore, the razor barbed wire vibration data caused by the airflow of the natural wind or HSR passing by was recorded continuously by the system for a variable duration.
Figure 3 shows the grayscale image of the vibration signal generated by the signal matrix after median filtering. The seven grayscale images show the seven typical event manifestations of the five classes of vibration activities. It can be visualized that the spatio-temporal images can well illustrate the vibration events. The width and height pixel values of the images are marked sequentially, where width indicates the vibration distribution along the fiber distance dimension and height represents the vibration duration. Figure 4 shows the pseudo-color images of the grayscale photographs of the seven different vibrational events after image enhancement processing to visually highlight the effect of pseudo-color processing.

4. Experimental Image Dataset and Methodology

The data pre-processing procedure can reduce the false-alarm rate and detect the location and time of vibration occurrence. However, in the actual detection environment of HSR perimeter security, there will still be vibrations from the natural wind or from the HSR passing by, thus resulting in false alarms. Such vibrations not within the alarm classes require an improved YOLO target detection model training with the customized dataset to recognize different patterns and remove non-essential alarms.

4.1. Image Data Augmentation

Convolutional neural network target detection systems have superior performance compared to traditional computer vision algorithms. However, overfitting problems can lead to performance degradation, which can typically be avoided by regularization methods [34] or increasing the amount of data. Data augmentation methods can increase the number of additional samples in the dataset, thus improving the generalization ability of the system, and effectively enhancing its performance for small target detection [35], including rotation, horizontal flipping, resizing, hue, saturation alteration, etc.

4.2. Dataset Generation

Four thousand three hundred eighty-one vibration data of 5 classes at different times and locations were collected by the method in Section 3.4.
The image dataset size was increased using the image data augmentation techniques in Section 4.1 to improve the robustness and generalization ability of the system, and each spatio-temporal domain vibration image generated a corresponding dataset image of size 608 pixels × 608 pixels to finally build a dataset of HSR perimeter security fiber optic sensing vibration images and generate the corresponding YOLO format annotation files. A total of 8069 images were eventually used for the experiments, as shown in Table 1, and the dataset was randomly assigned into a training set, test set, and validation set in accordance with the ratio of 8:1:1. In addition, the five vibration activities are divided into three alarm event types according to the practical application. Vibrations caused by airflow are natural phenomena and occur frequently, and such events do not require an alarm. When there is human activity or mechanical activity in the monitoring area near the fence, the detection system needs to send an early warning message to alert security personnel to the vibration changes in the area. As a critical monitoring vibration activity, climbing over the fence activities need the real-time alarm to inform personnel to rush to the scene to deal with such violations.

4.3. Target Detection Meta-Architecture Analysis

The target detection module is an important and integral part of the entire system. Its primary accomplishments include detection of target number in spatio-temporal data images, recognition of vibration pattern, and annotation of the corresponding bounding box, i.e., the position of the target in the image. This subsection focuses on the training process of the YOLO target detection model and the factors affecting the recognition results to provide a basis for the improvement of the multi-target detection model.
The YOLO target detection meta-architecture training process can be divided into three parts:
  • Backbone and Neck are the most dominant feature extraction networks of the model, and these parts extract feature maps at different scales by linear and nonlinear operations;
  • The detection Head contains multiple scales to be detected, the number of which depends on the number of different scales of the feature map after the aforementioned downsampling feature extraction, and the feature map width and height on each scale depends on the number of times the original spatio-temporal image has been downsampled;
  • The loss function defined by the YOLO detector includes three types of loss values: target confidence loss, localization loss [36,37], and classification loss. The difference between system detection results and prior knowledge in target detection and localization can be judged by these three types of loss scores.
As illustrated in Figure 5, the spatio-temporal vibration data images in the training set are fed into the Backbone. The feature maps at different scales are extracted from the Backbone and Neck sections by the convolutional unit to complete the downsampling. It is worth mentioning that the activation function in the convolutional unit allows the convolutional network to extract nonlinear features while avoiding the vanishing gradient problem. Currently, three activation functions are used in typical convolutional units frequently. Leaky-ReLU is simple and easy to implement with low computational cost. Unlike the Leaky-ReLU function, the non-monotonic and smooth nature of Swish [38], which is defined as f x = x s i g m o i d x , provides some robustness to different initialization and learning rates, and it is beneficial for model optimization and generalization ability, further improving the classification accuracy. Similar to Swish, Mish [39] is a smooth and non-monotonic activation function. Although Mish can improve the detection accuracy of neural networks, it increases the training time. Feature maps at different scales will enhance the discriminability and robustness of target features in the Neck section by expanding the receptive field. Furthermore, this part will accomplish feature propagation at different scales through a path aggregation network consisting of several bottom-up paths and several top-down paths to achieve feature enhancement [40,41]. In addition, CSPNet [42], Focus module, and Spatial Pyramid Pooling (SPP) network [43] are among the feature extraction network optional modules for feature enhancement.
The detection Head subsequent to the feature extraction network maps the depth of the feature map at different scales to c l a s s e s + 5 × 3 , where c l a s s e s indicates the number of types to be detected, 5 refers to an object confidence value and four values represent the location of the bounding box, and 3 denotes the number of prior anchor boxes at each scale. In this paper, the depth of the feature map is 30 for each scale.
The target confidence loss, localization loss, and classification loss are obtained by calculating the difference between the results of the detection system and the prior knowledge, updating the sensitivity map and the parameters of each convolutional layer of the detector using the sum of these three types of losses and the backpropagation algorithm, iterating the above steps until convergence, and until the detector can meet the detection requirements on the test set, thus completing the entire training process.
It should be remarked that the localization loss can be chosen from different Intersection over Unions (IoUs): Generalized IoU (GIoU) [36], Distance-IoU, and Complete IoU (CIoU) loss [37]. Among them, the CIoU loss considers the three geometric factors of overlap area, central point distance, and aspect ratio, thereby leading to faster convergence and better performance.
Based on the above analysis, the application of different feature extractors is based mainly on the complexity of the detection problem itself. The feature extraction network structure affects the final detection performance during the training process. The long-distance fiber optic vibration monitoring system needs to monitor the abnormal vibration at each region in real-time and reduce the false-alarm rate of the system at the same time. Therefore, it is required to consider both the real-time ability of system inference and the requirement of multi-target detection accuracy in designing the feature extraction network of the system.

4.4. Network Construction

In order to achieve high-precision detection of vibration activity in real-time using Φ-OTDR, four different types of feature extraction networks designed for fiber optic vibration images were implemented based on the analysis of the feature extraction network in Section 4.3, as presented in Table 2. The improved models were trained and tested in experiments using the image dataset in Section 4.2.
The design of the feature extraction network depends mainly on the complexity of the detection problem itself, and the main factors to be considered in the design include the type and number of feature extraction layers. Generally, the deeper the network is, the easier it is to obtain more abstract features. Yet, at the same time, it will increase model complexity and the number of parameters, which will affect the computational speed and real-time recognition ability of the system. The relationship between detection accuracy and model complexity can be analyzed by choosing different feature extraction components and different combinations of approaches in the feature extraction networks described above, and a suitable improved model can be selected using the evaluation metrics specified in Section 4.5. The feature extraction network consists of a combination of different feature extraction modules in the Backbone and Neck, where the CBA block contains a convolutional layer, a batch normalization layer, and an activation function, and the CSPNet block consists of two paths that are connected together by a concat layer to form a feature map after the features are extracted by the CBA block and several ResBlocks.
The improved YOLO-A30 network architecture is shown in Figure 6. YOLOv3-tiny and YOLOv3 [18] were trained as control groups to compare the prediction performance of different feature extraction networks using the YOLO meta-architecture.

4.5. Evaluation Metrics

Precision, recall, F1 score, mean average precision (mAP), inference time, and parameters (M) were used as the evaluation metrics to evaluate the detection effectiveness of different networks in a comprehensive manner.
The definitions of precision, recall, F1 score, average precision (AP), and mean average precision (mAP) are expressed by Equations (8)–(12). True positive (TP) indicates the amount of target data where the positive class of the outcome of the model detection is consistent with the actual class in the dataset. False positive (FP) represents the amount of data where the positive class of the outcome of the model detection is inconsistent with the actual class in the dataset, i.e., the number of false detections. False negative (FN) then denotes the number of target data in the dataset with actual classes but detected by the model as other classes, i.e., the number of missed detections. Precision and recall indicate the confidence of the model in target detection and the ability of the model to detect all targets, respectively. The precision and recall metrics of the model can be comprehensively evaluated using F1. Under a fixed IoU, the given target will get different precision and recall values according to different confidence, and the continuous curve generated by interpolating the precision and recall values is the precision-recall (PR) curve. Average precision (AP) shows the comprehensive performance of the model at different confidences by the area under the curve (AUC) on the PR curve for a given target class, and a higher AP value represents better model detection performance. Each IoU corresponds to a different AP, and the [email protected] corresponds to the AP when IoU takes 0.5, AP@[.5:.95] corresponds to the average AP for IoU from 0.5 to 0.95 with a step size of 0.05. The mean average precision (mAP) evaluates the mean of the average precision of the model on all detection classes. Eventually, the inference time describes the time required by the model for single image detection. Usually, the fewer model parameters, the smaller the GPU memory consumption, the shorter the inference time, and the more it can be implemented on embedded devices.
precision   = TP TP + FP
Recall = TP TP + FN
F 1 = 2 ×   Precision   ×   Recall     Precision   +   Recall  
AP class = i     confidence   precision i recall , class , iou
mAP = 1 N AP i

5. Experimental Procedure

5.1. Experimental Setup

The phase-sensitive optical time-domain reflectometer is used as a vibration signal collector in the monitoring area, as described in Section 2.1. The pulsed light is amplified by EDFA and injected into the fiber through an isolator and circulator. The backscattered light is detected by a photodetector, and the collected raw data is formed into spatio-temporal images after signal pre-processing, median filtering, and pseudo-color technique. The data images of different vibration events are collected, and the ultimate dataset is formed through data augmentation. The parameters for equipment data acquisition and pre-processing, the method of single-mode fiber deployment, and the method of data acquisition are specified in Section 3.4.
This study combines different network modules based on the YOLO algorithm framework to implement four feature extraction networks, using the customized dataset to train the models. The specifics of the feature extraction networks are elaborated in Section 4.4. The prediction performance of the improved models is compared comprehensively using the previously described evaluation metrics. The experimental procedure is shown in Figure 7. Considering the practicality and computational complexity, the feature extraction networks in this paper use the Swish activation function, CIoU localization loss and YOLOv3 detection Head, and combine different feature extraction components to achieve YOLO model improvement. The experimental contents of the fiber optic sensing vibration recognition system contain the following:
  • The YOLO target detection models composed of different feature extraction networks are trained using the training set, the training effect is tested using the validation set, and the optimal model in each type is selected; the performance metrics of the selected models in each type as described in Section 4.5 are measured using the test set, and the one with a better balance of various metrics is chosen as the final improved model;
  • Various combinations of vibration type data in the test set are selected to simulate different mixed vibration activities to test and discuss the qualitative detection results of the final improved model;
  • The metrics of the final improved model and other SOTA models arere counted using the test set and comparisons are made.

5.2. Network Training

Using YOLOv3-tiny and YOLOv3 as control groups, the corresponding performance metrics are obtained by varying the feature extraction network of the YOLO meta-architecture. The training set is used for model parameter updates, the sum of the three types of losses described in Section 4.3 is calculated for each iteration of the optimization process during the experimental training, which in turn updates the model parameters by the back-propagation algorithm, and the model performance metrics are counted using the validation set at the end of each epoch during the training process, while the system hyperparameters are kept consistent. Considering the model generalization ability and overfitting problem, each feature extraction network is trained at least three times and 100 epochs are run each time, from which the model with the best-combined performance of [email protected] and AP@[.5:.95] in the validation set is selected, and the model performance is tested by the test set. The model training process was run on an Ubuntu 18.04 operating system with an NVIDIA GTX 3090 GPU with 24-GB onboard memory, accelerated by CUDA (version 11.1) and cuDNN (version 8.0). YOLO meta-architecture is based on Python (version 3.8) and the PyTorch library (version 1.8.0) released by Facebook AI.
Figure 8 shows the training loss and validation loss of different feature extraction networks after 100 epochs of training. The training losses decrease gradually during the training process, which indicates that the networks learn model parameters from the training set effectively without overfitting. YOLOv3 and the four types of improved models each have significantly lower losses than YOLOv3-tiny, with the training loss ranging around 0.03 and the validation loss ranging around 0.025. It is observed that although the difference in training loss between YOLO-A30, YOLO-A78, and YOLOv3 is slight, YOLO-A30 validation loss is much lower than the other models, which reveals that YOLO-A30 performs better than the other models in the validation set.
Figure 9 shows the PR curves for YOLOv3-tiny and YOLO-A30 with an IoU threshold of 0.5. The [email protected] values for the five detection events are represented by the area under the corresponding curves. It can be found from the PR curves that the detection performance gap between the improved YOLO-A30 and the control YOLOv3-tiny on airflow events is significantly larger than for the other activities. Other performance metrics of the models can be compared in Table 3.
Table 3 evaluates the prediction performance, the number of parameters, and the inference time of each model from seven metrics. For HSR perimeter abnormal vibration monitoring, fence climbing is the primary monitoring activity, which requires no missing alarms and high confidence of alarm. Vibration caused by airflow does not need to be alarmed, whereas due to the frequent commuting of high-speed rail and the windy field environment in which the high-speed rail line is located, the false-alarm rate of the system will be too high if the airflow cannot be identified precisely. However, the low recall of airflow can instead be considered automatic filtering of such events. Accordingly, while evaluating the overall performance, the F1 score of the fence climbing activity and the precision of the airflow event should be highlighted.
In the precision column, YOLO-A30 has the highest precision of 99.2%, which is 0.4% and 0.6% higher than YOLO-A78 and YOLOv3, which have more parameters, respectively. Its precision is the highest in the events of construction machinery, walking, and especially airflow, which is 0.5% and 36.7% higher than YOLOv3 and YOLOv3-tiny, respectively. Although YOLO-RP achieves 100.0% and 99.9% precision for manual excavation and fence climbing, its precision for airflow is only 86.4%.
The recall represents the probability that the model does not have missed detection, and all models achieve 100.0% in walking and fence climbing; for the crucial monitoring of fence climbing, especially, it is essential to achieve low missed-detection rate for the practical application of the model. For airflow events that do not require an alarm, the recall rate does not affect the system performance, so the effect of airflow recall should be removed when discussing the recall metrics. After removing the effect of airflow, the average recall of YOLO-A30 is 99.8%, reaching the highest value along with YOLO-RCP, which is higher than YOLOv3-tiny and YOLOv3 by 0.2% and 0.1%, respectively.
The F1 score can comprehensively evaluate the precision and recall of the model for detection activities, and the F1 score of YOLO-A30 is the highest at 99.2%. The F1 scores for construction machinery and walking activities are 99.1% and 99.8%, respectively. The F1 score for fence climbing is obtained in YOLO-RP with a maximum of 99.9%, which is 0.1% higher than that of YOLO-A30.
Focusing on the columns [email protected] and AP@[.5:.95] with reference to YOLO-RP, YOLO-RCP, and YOLO-A30, although the number of parameters decreases in sequence, the mAP can be significantly improved with the addition of different components of the feature extraction network. The [email protected] of YOLO-A30 is 0.1% and 4.1% higher than that of YOLOv3 and YOLOv3-tiny, and the AP@[.5:.95] is 1% and 9.4% higher, respectively, while the number of YOLOv3 parameters is 8.58 and 7.10 times greater than that of YOLO-A30 and YOLOv3-tiny. It can be noticed that using the same network components, increasing the number of parameters can improve the mAP value of the model by comparing YOLOV3 and YOLOV3-tiny performance metrics. However, there is a limit to the mAP gain from the increase in the number of parameters. YOLO-A30 and YOLO-A78 use the same network components, but with the increase in the number of parameters, [email protected] stays the same, and AP@[.5:.95] only increases by 0.1%. On balance, the YOLO-A30 with the fewest number of parameters performs better on the mAP metric.
When the system monitors abnormal vibration, the image target detection needs to achieve long-distance detection range and high timeliness, which is strongly related to the size of the spatio-temporal image to be detected and the inference speed. In this experiment, the length of the single image to be detected is 608 pixels, corresponding to the actual fiber distance of 243.2 m. Taking the YOLO-A30 with an inference speed of 1.8 ms as an example, the maximum multi-vibration events within 135.1 km can be detected in 1 s.
Comparing the inference speed of each model, it can be realized that the types of network components and the number of parameters used in the feature extraction network have an impact on the inference speed of the network. When designing a feature extraction network, one cannot just increase the number of network modules and parameters to improve detection performance while neglecting its impact on detection speed.

5.3. Qualitative Results

The above metrics illustrate the performance of YOLO-A30 on the test set. Benefiting from the distributed nature of fiber optic sensing, the vibration signals generated by different activities occurring simultaneously at different locations of the fiber do not affect each other. The combinations for different vibration classes in the test set are picked to simulate six different kinds of mixed vibration events to visualize the results of multi-target detection of YOLO-A30 in specific scenarios. The test parameters are IoU > 0.5, Confidence > 0.4, as depicted in Figure 10. The image width represents the range of influence in the fiber distance dimension corresponding to the multi-vibration event in meters. The image height indicates the duration of the multi-vibration event in seconds. Although repeated detection occurs when the vibration signal is too dense in the buried fiber state, the system can effectively distinguish multiple vibration classes in the same detection range.
Figure 10a,b separately depict the two construction early warning activities detected by the system in the buried fiber state and their respective detection confidence. When multiple mechanical construction events occur in the monitoring area, YOLO-A30 can effectively identify the vibration activity class of the event and divide the multiple vibration activities. When two classes of construction activities occur simultaneously at different locations, the system will automatically distinguish the different classes of activities and label them by the bounding box.
Figure 10c–e display three scenarios where vibration is detected simultaneously in the two cases of fiber buried underground and hung on razor barbed wire. Due to the distributed nature of the fiber optic sensor, the system can still accurately detect construction events in the buried fiber state in the presence of airflow interference. Whether there is a continuous behavior such as people walking and climbing over the fence or construction vibration disturbances, the model can identify such a crucial monitoring activity as climbing over the fence.
Figure 10f illustrates the extreme case when airflow interference and fence climbing occur under the fiber optic hanging condition in the same area. The system can eliminate the interference and identify fence climbing activity to improve the overall detection performance.

5.4. Comparison with Other State-of-the-Art Models

The performance of YOLO-A30 is compared with other SOTA models on the dataset described in Section 4.2. The [email protected], F1 score, airflow precision, climbing fence F1 score, and inference time of 6 models, Faster R-CNN [14], SSD [17], YOLOv3-tiny, YOLOv3 [18], YOLOv4 [19], and the improved YOLO-A30, are shown in Figure 11, considering the high-speed rail application environment.
YOLO-A30 has the highest F1 score of 99.2%, indicating that it can detect the five classes of abnormal vibration activities in the dataset relatively accurately with a low missed-detection rate. F1 scores of YOLOv3 and SSD are only lower than YOLO-A30, each reaching 99.1% and 98.7%, both with excellent performance, but both are more than three times slower than YOLO-A30 in terms of detection speed, which means the longest distance these two models can monitor is much shorter than the monitoring distance of YOLO-A30. Over 95% [email protected] are achieved for all models, with YOLO-A30 and SSD both reaching a maximum of 99.5%, 4.1% higher than the lowest, YOLOv3-tiny. The precision of airflow detection, which is the most frequent vibration interference event in the actual application environment for high-speed rail, is also noticed separately in the figure, and the precision of this event monitoring will directly affect the final false-alarm rate of the system. One explanation for the considerable variation in airflow precision across models is that airflow events do not have a fixed shape and aspect ratio in the spatio-temporal data image, which leads to particular difficulty in detection and requires a high feature extraction capability for the model network, which makes this metric a critical factor in differentiating model performance. The airflow precision of YOLO-A30 is 98.7%, which is 0.5% higher than YOLOv3 and significantly higher than other target detection models. The SSD reaches 100% in the fence climbing F1 score, which is 0.2% higher than the YOLO-A30, but it lags behind the YOLO-A30 in all other metrics; in particular, the differences in airflow precision and inference time are 7.2% and 4.3 ms, respectively. The inference time of Faster R-CNN [14] as a two-stage detection is 30.1 ms, which is much longer than other single-stage detection models, which also shows from one side that the single-stage detection model is more appropriate for long-distance monitoring requirements.
Overall, YOLO-A30 performs excellently in the main metrics compared to other SOTA models. It can avoid false alarms of fiber optic sensors due to airflow to a certain extent and detect five key monitoring events simultaneously, providing a possible solution for high-speed rail perimeter security automation.

6. Conclusions

In this paper, a real-time multi-vibration events detection method based on an improved YOLO model for Φ-OTDR optical fiber buried underground and hung on razor barbed wire dual detection environments is proposed for high-speed railway perimeter abnormal vibration detection, aiming to realize long-distance perimeter security automation and reduce system false-alarm rate and security cost. Based on the statistical characteristics of the acquired raw signals and the signal detection theory, it is shown how pulse accumulation, pulse cancellers, and coincidence detection can improve the detection performance of the system. Following this, the signal pre-processed data matrix is median filtered and pseudo-color processed to achieve feature enhancement and generate a spatio-temporal vibration image dataset with 8096 images of five vibration activities for training, validation, and testing. Improving the feature extraction network of the YOLO target detection model, specifically using different network components and network structures, aims to discover a deep learning architecture more appropriate for fiber optic vibration monitoring tasks in the application environment of the high-speed railway. The experimental results demonstrate that the improved YOLO-A30 model achieves a satisfactory balance in precision corresponding to the false-alarm rate of the system, recall related to the missed-detection rate of the system, and inference speed, which exhibits a better detection performance than other experimental models. This study illustrates that the method can automatically perform Φ-OTDR acquisition signal pre-processing and effectively extract vibration event features to identify abnormal vibration, which reduces the false-alarm rate in the practical application of distributed fiber optic sensing system and makes it possible to apply this method in high-speed railway perimeter security.
Future work will focus on increasing the detection distance of a single device through multi-channel or other Φ-OTDR techniques and optimizing the existing results, including validating model accuracy and performance in the actual scene, evaluating the feasibility, mining for more possible vibration events, and verifying the adaptation of the model to recognize similar events to improve the robustness and adaptability of the model, as well as developing embedded devices for practical applications.

Author Contributions

All authors contributed substantially to the manuscript. Conceptualization, N.Y., Y.Z. and J.C.; methodology, N.Y.; software, N.Y.; validation, N.Y.; formal analysis, N.Y.; investigation, N.Y. and J.C.; resources, N.Y., Y.Z. and J.C.; data curation, N.Y.; writing—original draft preparation, N.Y.; writing—review and editing, Y.Z. and J.C.; visualization, N.Y.; supervision, Y.Z.; project administration, Y.Z.; funding acquisition, Y.Z. and J.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the China Postdoctoral Fund (Grant No. 2017M623349).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author.

Acknowledgments

We thank the reviewers for their valuable feedback and the editors for their devoted guidance.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Juarez, J.C.; Maier, E.W.; Kyoo Nam, C.; Taylor, H.F. Distributed fiber-optic intrusion sensor system. J. Lightwave Technol. 2005, 23, 2081–2087. [Google Scholar] [CrossRef]
  2. Juarez, J.C.; Taylor, H.F. Field test of a distributed fiber-optic intrusion sensor system for long perimeters. Appl. Opt. 2007, 46, 1968–1971. [Google Scholar] [CrossRef] [PubMed]
  3. Wang, Z.N.; Zeng, J.J.; Li, J.; Fan, M.Q.; Wu, H.; Peng, F.; Zhang, L.; Zhou, Y.; Rao, Y.J. Ultra-long phase-sensitive OTDR with hybrid distributed amplification. Opt. Lett. 2014, 39, 5866–5869. [Google Scholar] [CrossRef] [PubMed]
  4. Bao, X.; Zhou, D.P.; Baker, C.; Chen, L. Recent Development in the Distributed Fiber Optic Acoustic and Ultrasonic Detection. J. Lightwave Technol. 2017, 35, 3256–3267. [Google Scholar] [CrossRef]
  5. Peng, F.; Wu, H.; Jia, X.-H.; Rao, Y.-J.; Wang, Z.-N.; Peng, Z.-P. Ultra-long high-sensitivity Φ-OTDR for high spatial resolution intrusion detection of pipelines. Opt. Express 2014, 22, 13804–13810. [Google Scholar] [CrossRef]
  6. Wu, H.; Qian, Y.; Zhang, W.; Tang, C. Feature extraction and identification in distributed optical-fiber vibration sensing system for oil pipeline safety monitoring. Photonic Sens. 2017, 7, 305–310. [Google Scholar] [CrossRef] [Green Version]
  7. Tejedor, J.; Martins, H.F.; Piote, D.; Macias-Guarasa, J.; Pastor-Graells, J.; Martin-Lopez, S.; Guillén, P.C.; Smet, F.D.; Postvoll, W.; González-Herráez, M. Toward Prevention of Pipeline Integrity Threats Using a Smart Fiber-Optic Surveillance System. J. Lightwave Technol. 2016, 34, 4445–4453. [Google Scholar] [CrossRef] [Green Version]
  8. Xu, C.; Guan, J.; Bao, M.; Lu, J.; Ye, W. Pattern recognition based on time-frequency analysis and convolutional neural networks for vibrational events in φ-OTDR. Opt. Eng. 2018, 57, 016103. [Google Scholar] [CrossRef]
  9. Tejedor, J.; Macias-Guarasa, J.; Martins, H.F.; Piote, D.; Pastor-Graells, J.; Martin-Lopez, S.; Corredera, P.; Gonzalez-Herraez, M. A Novel Fiber Optic Based Surveillance System for Prevention of Pipeline Integrity Threats. Sensors 2017, 17, 355. [Google Scholar] [CrossRef]
  10. Sun, Q.; Feng, H.; Yan, X.; Zeng, Z. Recognition of a Phase-Sensitivity OTDR Sensing System Based on Morphologic Feature Extraction. Sensors 2015, 15, 15179–15197. [Google Scholar] [CrossRef] [Green Version]
  11. Sun, Q.; Li, Q.; Chen, L.; Quan, J.; Li, L. Pattern recognition based on pulse scanning imaging and convolutional neural network for vibrational events in Φ-OTDR. Optik 2020, 219, 165205. [Google Scholar] [CrossRef]
  12. Shi, Y.; Wang, Y.; Zhao, L.; Fan, Z. An Event Recognition Method for Φ-OTDR Sensing System Based on Deep Learning. Sensors 2019, 19, 3421. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  13. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  14. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards Real-Time Object Detection with Region Proposal Networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  15. He, K.; Gkioxari, G.; Dollár, P.; Girshick, R. Mask R-CNN. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  16. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You Only Look Once: Unified, Real-Time Object Detection. In Proceedings of the 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  17. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. SSD: Single Shot MultiBox Detector. arXiv 2016, arXiv:1512.02325. [Google Scholar]
  18. Redmon, J.; Farhadi, A. YOLOv3: An Incremental Improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  19. Bochkovskiy, A.; Wang, C.-Y.; Liao, H.-Y.M. YOLOv4: Optimal Speed and Accuracy of Object Detection. arXiv 2020, arXiv:2004.10934. [Google Scholar]
  20. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object Detection With Deep Learning: A Review. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3212–3232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  21. Wu, X.; Sahoo, D.; Hoi, S.C.H. Recent advances in deep learning for object detection. Neurocomputing 2020, 396, 39–64. [Google Scholar] [CrossRef] [Green Version]
  22. Guo, F.; Qian, Y.; Shi, Y. Real-time railroad track components inspection based on the improved YOLOv4 framework. Autom. Constr. 2021, 125, 103596. [Google Scholar] [CrossRef]
  23. Everingham, M.; Eslami, S.M.A.; Van Gool, L.; Williams, C.K.I.; Winn, J.; Zisserman, A. The Pascal Visual Object Classes Challenge: A Retrospective. Int. J. Comput. Vis. 2015, 111, 98–136. [Google Scholar] [CrossRef]
  24. Lin, T.-Y.; Maire, M.; Belongie, S.; Hays, J.; Perona, P.; Ramanan, D.; Dollár, P.; Zitnick, C.L. Microsoft COCO: Common Objects in Context. arXiv 2014, arXiv:1405.0312, 740–755. [Google Scholar]
  25. Healey, P. Fading in heterodyne OTDR. In Electronics Letters; Institution of Engineering and Technology: London, UK, 1984; Volume 20, pp. 30–32. [Google Scholar] [CrossRef]
  26. Izumita, H.; Koyamada, Y.; Furukawa, S.; Sankawa, I. Stochastic amplitude fluctuation in coherent OTDR and a new technique for its reduction by stimulating synchronous optical frequency hopping. J. Lightwave Technol. 1997, 15, 267–278. [Google Scholar] [CrossRef]
  27. Fedorov, A.K.; Anufriev, M.N.; Zhirnov, A.A.; Stepanov, K.V.; Nesterov, E.T.; Namiot, D.E.; Karasik, V.E.; Pnev, A.B. Note: Gaussian mixture model for event recognition in optical time-domain reflectometry based sensing systems. Rev. Sci. Instrum. 2016, 87, 036107. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  28. Maini, R.; Aggarwal, H. A Comprehensive Review of Image Enhancement Techniques. arXiv 2010, arXiv:1003.4053. [Google Scholar]
  29. Zheng, D.; Da, F.; Kemao, Q.; Seah, H.S. Phase-shifting profilometry combined with Gray-code patterns projection: Unwrapping error removal by an adaptive median filter. Opt. Express 2017, 25, 4700–4713. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, J.; Lu, S.; Chen, J. Application of Pseudo-color Image Feature-Level Fusion in Nondestructive Testing of Wire Ropes. J. Fail. Anal. Prev. 2020, 20, 1541–1553. [Google Scholar] [CrossRef]
  31. Zhu, Q.; Zheng, H.; Wang, Y.; Cao, Y.; Guo, S. Study on the Evaluation Method of Sound Phase Cloud Maps Based on an Improved YOLOv4 Algorithm. Sensors 2020, 20, 4314. [Google Scholar] [CrossRef]
  32. Peng, F.; Duan, N.; Rao, Y.; Li, J. Real-Time Position and Speed Monitoring of Trains Using Phase-Sensitive OTDR. IEEE Photonics Technol. Lett. 2014, 26, 2055–2057. [Google Scholar] [CrossRef]
  33. Meng, H.; Wang, S.; Gao, C.; Liu, F. Research on Recognition Method of Railway Perimeter Intrusions Based on Φ-OTDR Optical Fiber Sensing Technology. IEEE Sens. J. 2021, 21, 9852–9859. [Google Scholar] [CrossRef]
  34. Ghiasi, G.; Lin, T.-Y.; Le, Q.V. DropBlock: A regularization method for convolutional networks. arXiv 2018, arXiv:1810.12890. [Google Scholar]
  35. Kisantal, M.; Wojna, Z.; Murawski, J.; Naruniec, J.; Cho, K. Augmentation for small object detection. arXiv 2019, arXiv:1902.07296. [Google Scholar]
  36. Rezatofighi, H.; Tsoi, N.; Gwak, J.; Sadeghian, A.; Reid, I.; Savarese, S. Generalized Intersection Over Union: A Metric and a Loss for Bounding Box Regression. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 658–666. [Google Scholar]
  37. Zheng, Z.; Wang, P.; Liu, W.; Li, J.; Ye, R.; Ren, D. Distance-IoU Loss: Faster and Better Learning for Bounding Box Regression. arXiv 2019, arXiv:1911.08287. [Google Scholar] [CrossRef]
  38. Ramachandran, P.; Zoph, B.; Le, Q.V. Swish: A Self-Gated Activation Function. arXiv 2017, arXiv:1710.05941. [Google Scholar]
  39. Misra, D. Mish: A self regularized non-monotonic activation function. arXiv 2019, arXiv:1908.08681. [Google Scholar]
  40. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature Pyramid Networks for Object Detection. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar]
  41. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path Aggregation Network for Instance Segmentation. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  42. Wang, C.Y.; Liao, H.Y.M.; Wu, Y.H.; Chen, P.Y.; Hsieh, J.W.; Yeh, I.H. CSPNet: A New Backbone that can Enhance Learning Capability of CNN. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), Seattle, WA, USA, 14–19 June 2020; pp. 1571–1580. [Google Scholar]
  43. He, K.; Zhang, X.; Ren, S.; Sun, J. Spatial Pyramid Pooling in Deep Convolutional Networks for Visual Recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 1904–1916. [Google Scholar] [CrossRef] [PubMed] [Green Version]
Figure 1. Overall system operation structure.
Figure 1. Overall system operation structure.
Sensors 22 01127 g001
Figure 2. Φ-OTDR backscattering principle: (a) Superposition principle of backscattered light from all scattering centers within a half-light pulse width; (b) The effect of vibration on the echo signal, where the variable scattering center is marked in yellow.
Figure 2. Φ-OTDR backscattering principle: (a) Superposition principle of backscattered light from all scattering centers within a half-light pulse width; (b) The effect of vibration on the echo signal, where the variable scattering center is marked in yellow.
Sensors 22 01127 g002
Figure 3. Grayscale images of different vibration events: (a) Natural wind; (b) HSR passing by; (c) Walking and climbing over fence simultaneously; (d) Climbing over fence; (e) Walking near fence; (f) Manual excavation; (g) Construction machinery.
Figure 3. Grayscale images of different vibration events: (a) Natural wind; (b) HSR passing by; (c) Walking and climbing over fence simultaneously; (d) Climbing over fence; (e) Walking near fence; (f) Manual excavation; (g) Construction machinery.
Sensors 22 01127 g003
Figure 4. Data images of vibration events after pseudo-color processing: (a) Natural wind; (b) HSR passing by; (c) Walking and climbing over fence simultaneously; (d) Climbing over fence; (e) Walking near fence; (f) Manual excavation; (g) Construction machinery.
Figure 4. Data images of vibration events after pseudo-color processing: (a) Natural wind; (b) HSR passing by; (c) Walking and climbing over fence simultaneously; (d) Climbing over fence; (e) Walking near fence; (f) Manual excavation; (g) Construction machinery.
Sensors 22 01127 g004
Figure 5. YOLO target detection meta-architecture training process.
Figure 5. YOLO target detection meta-architecture training process.
Sensors 22 01127 g005
Figure 6. Feature extraction network and main components: (a) Improved YOLO-A30 network architecture; (b) CBA block component; (c) CSPNet block component.
Figure 6. Feature extraction network and main components: (a) Improved YOLO-A30 network architecture; (b) CBA block component; (c) CSPNet block component.
Sensors 22 01127 g006
Figure 7. Experimental procedure of the proposed optical fiber sensing vibration identification method based on deep learning.
Figure 7. Experimental procedure of the proposed optical fiber sensing vibration identification method based on deep learning.
Sensors 22 01127 g007
Figure 8. Training loss and validation loss of different models: (a) Training loss of six experimental models; (b) Validation loss of six experimental models.
Figure 8. Training loss and validation loss of different models: (a) Training loss of six experimental models; (b) Validation loss of six experimental models.
Sensors 22 01127 g008
Figure 9. PR curve with IoU threshold of 0.5: (a) YOLOv3-tiny; (b) YOLO-A30.
Figure 9. PR curve with IoU threshold of 0.5: (a) YOLOv3-tiny; (b) YOLO-A30.
Sensors 22 01127 g009
Figure 10. Effectiveness of six different mixed vibration events identification: (a) Multiple construction machines; (b) Multiple construction machines and manual excavation; (c) Airflow interference exists nearby the buried fiber optic site where the early warning activities occurred; (d) Climbing over the fence behavior and construction machinery activities occur in the same monitoring area; (e) People walking in the buried fiber optic area and climbing over the fence; (f) Person climbing over the fence during airflow interference.
Figure 10. Effectiveness of six different mixed vibration events identification: (a) Multiple construction machines; (b) Multiple construction machines and manual excavation; (c) Airflow interference exists nearby the buried fiber optic site where the early warning activities occurred; (d) Climbing over the fence behavior and construction machinery activities occur in the same monitoring area; (e) People walking in the buried fiber optic area and climbing over the fence; (f) Person climbing over the fence during airflow interference.
Sensors 22 01127 g010aSensors 22 01127 g010b
Figure 11. Performance metrics of YOLO-A30 and other SOTA models on focused activities.
Figure 11. Performance metrics of YOLO-A30 and other SOTA models on focused activities.
Sensors 22 01127 g011
Table 1. Distribution of each vibration pattern in the described dataset.
Table 1. Distribution of each vibration pattern in the described dataset.
Fiber Optic Laying MethodActivityActivity SimulationAmount of Collected DataCollection
Location
Duration of a CollectionTotal Data Amount with Data
Augmentation
Alarm Type
Optical fiber cable buried undergroundConstruction machineryPneumatic hammer13570–1 km3–6 s2714Early
Warning
Manual
excavation
Shovel type construction10510–1 km3–6 s2102Early
Warning
WalkingPeople walking within the warning area nearby the fence1370–1 km3–15 s819Early
Warning
Optical fiber cable hung on razor barbed wireClimbing over fenceShaking Razor Barbed Wire2001–5 km5–18 s798Real-time alarm
AirflowNatural wind or high-speed rail passing by16361–5 kmVariable1636False
alarm
Table 2. Six types of feature extraction networks composed of different network components.
Table 2. Six types of feature extraction networks composed of different network components.
Model Backbone Neck HeadNumber of Convolutional Layers
FocusResBlockCSPNetSPP blockFPNPANetYOLO
YOLOV3-tiny13
YOLOV352
YOLO-RP30
YOLO-RCP30
YOLO-A3030
YOLO-A7878
Table 3. Models performance metrics statistics.
Table 3. Models performance metrics statistics.
ModelClass Precision
(%)
Recall without Airflow
(%)
F1
(%)
[email protected]
(%)
mAP@[.5:.95]
(%)
Parameters (MB)Inference
(ms)
YOLOV3-tinyAll91.6 99.695.1 95.4 59.9 17.41.3
Construction machinery98.2 99.0 98.6 99.4 58.1
Manual excavation99.0 99.2 99.1 99.6 58.1
Walking99.4 100.0 99.7 99.5 75.2
Climbing fence99.5 100.0 99.7 99.6 69.4
Airflow62.0 95.775.2 78.8 38.9
YOLOV3All98.6 99.7 99.1 99.4 68.3 123.56.2
Construction machinery97.2 99.0 98.1 99.3 62.8
Manual excavation98.9 99.6 99.2 99.6 58.5
Walking99.4 100.0 99.7 99.5 81.4
Climbing fence99.3 100.0 99.6 99.5 71.1
Airflow98.2 100.099.1 99.2 67.4
YOLO-RPAll96.7 99.7 98.0 98.8 68.6 201.8
Construction machinery97.8 99.0 98.4 99.5 64.0
Manual excavation100.0 99.8 99.999.6 59.8
Walking99.4 100.0 99.7 99.5 82.3
Climbing fence99.9 100.0 99.999.6 73.2
Airflow86.4 98.191.9 95.8 63.5
YOLO-RCPAll97.0 99.898.1 98.9 68.2 17.81.7
Construction machinery97.6 99.4 98.5 99.5 63.5
Manual excavation99.6 99.8 99.7 99.5 60.5
Walking99.4 100.0 99.7 99.5 82.2
Climbing fence98.8 100.0 99.4 99.5 71.7
Airflow89.7 96.993.2 96.1 63.3
YOLO-A30All99.299.899.299.569.3 14.41.8
Construction machinery98.8 99.499.199.5 62.5
Manual excavation99.5 99.699.5 99.4 59.3
Walking99.6 100.099.899.5 83.5
Climbing fence99.6 100.099.8 99.6 71.7
Airflow98.796.997.8 99.4 69.6
YOLO-A78All98.8 99.799.1 99.569.416.42.3
Construction machinery98.6 99.799.199.6 61.6
Manual excavation99.0 99.299.1 99.5 59.6
Walking99.5 100.099.7 99.5 82.3
Climbing fence99.5 100.099.7 99.5 73.4
Airflow97.6 98.898.2 99.2 70.2
The key performance metrics with the best results are highlighted.
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yang, N.; Zhao, Y.; Chen, J. Real-Time Φ-OTDR Vibration Event Recognition Based on Image Target Detection. Sensors 2022, 22, 1127. https://doi.org/10.3390/s22031127

AMA Style

Yang N, Zhao Y, Chen J. Real-Time Φ-OTDR Vibration Event Recognition Based on Image Target Detection. Sensors. 2022; 22(3):1127. https://doi.org/10.3390/s22031127

Chicago/Turabian Style

Yang, Nachuan, Yongjun Zhao, and Jinyang Chen. 2022. "Real-Time Φ-OTDR Vibration Event Recognition Based on Image Target Detection" Sensors 22, no. 3: 1127. https://doi.org/10.3390/s22031127

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop