Next Article in Journal
Study on Gesture Recognition Method with Two-Stream Residual Network Fusing sEMG Signals and Acceleration Signals
Previous Article in Journal
Determination of Self-Heating in Silicon Photomultipliers
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Research on a Sound Source Localization Method for UAV Detection Based on Improved Empirical Mode Decomposition

School of Mechanical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(9), 2701; https://doi.org/10.3390/s24092701
Submission received: 22 March 2024 / Revised: 14 April 2024 / Accepted: 20 April 2024 / Published: 24 April 2024
(This article belongs to the Section Vehicular Sensing)

Abstract

:
To address the challenge of accurately locating unmanned aerial vehicles (UAVs) in situations where radar tracking is not feasible and visual observation is difficult, this paper proposes an innovative acoustic source localization method based on improved Empirical Mode Decomposition (EMD) within an adaptive frequency window. In this study, the collected flight signals of UAVs undergo smoothing filtering. Additionally, Robust Empirical Mode Decomposition (REMD) is applied to decompose the signals into Intrinsic Mode Function (IMF) components for spectrum analysis. We introduce a sliding frequency window with adjustable bandwidth, which is automatically determined using a Grey Wolf Optimizer (GWO) with a sliding index. This window is used to lock and extract specific frequencies from the IMFs. Based on predefined criteria, the extracted IMF components are reconstructed, and trigger signal times are analyzed and recorded from these reconstructed IMFs. The time differences between sensor receptions are then calculated. Furthermore, this study introduces the Chan–Taylor localization algorithm based on weighted least squares. This advanced algorithm takes sensor time delay parameters as input and solves a set of nonlinear equations to determine the target’s location. Simulations and real-world signal tests are used to validate the robustness and performance of the proposed method. The results indicate that the localization error remains below 5% within a 15 m × 15 m measurement area. This provides an efficient and real-time method for detecting the location of small UAVs.

1. Introduction

In military, civilian, and anti-terrorism sectors, precise detection of micro-UAV locations is crucial, yet traditional detection methods are costly and complex to operate. Passive acoustic detection systems, especially those employing the Time Difference of Arrival (TDOA) algorithm, offer a low-cost and efficient solution to this challenge [1]. However, there are three major obstacles in practical applications: accuracy issues in low signal-to-noise ratio (SNR) environments, difficulties in solving nonlinear equation systems, and limitations in long-range detection.
To address the low SNR problem, this paper introduces the Robust Empirical Mode Decomposition (REMD) algorithm. It incorporates robustness criteria or regularization terms, such as the Self-Adaptive Sifting Stopping Criterion (SSSC). By improving the endpoint effects of traditional EMD [2], REMD significantly enhances the stability and accuracy of signal processing, especially in complex and noise-mixed signal environments. The REMD method can adapt to various complex signal characteristics and effectively avoid noise interference and outlier influences, resulting in more accurate and reliable reconstructed signals. This solves the defect of EMD, which performs well in processing nonlinear and non-stationary signals but may be affected when facing complex signals or noise.
To address the difficulty in solving nonlinear equation systems, we combine the advantages of the Chan algorithm and the Taylor algorithm, proposing a novel Chan–Taylor fusion algorithm [3]. This approach not only reduces the difficulty of localization but also significantly improves localization accuracy, especially in noisy environments. To expand the detection range and increase the significance of time delay differences, we adopt a three-element discrete distributed array. By pre-measuring the specific coordinates of the sensors and combining them with the time delay estimation values obtained through REMD processing, we can accurately determine the target location using the Chan–Taylor fusion algorithm.
The structure of this paper is arranged as follows: Firstly, we introduce the division of far and near sound fields and the sound source localization model of the three-element discrete array. Secondly, we elaborate on the REMD time delay analysis method and then we delve into the specific mathematical methods of the Chan–Taylor fusion algorithm. Afterward, we present the simulation results of the REMD and Chan–Taylor algorithms. Finally, we summarize the full text. Through this structured arrangement, we aim to provide readers with a comprehensive and in-depth understanding of the applications and advantages of passive acoustic detection systems in micro-UAV localization. Figure 1 is a general flowchart of our approach to the methodology explained in this paper.

2. Sound Source Localization Model

2.1. Delineation of Near-Field and Far-Field Models

Assume that d is the array spacing, λ m is the minimum wavelength, L is the distance from the sound source to the microphone, and D = n d is the size of the receiver array. The waveform of the sound source arriving at the microphone array at this point is considered to be a spherical waveform, which is consistent with the near-field model, and vice versa, for a plane wave, which is consistent with the far-field model, when L < 2 D 2 λ m . The near-field model is suitable for close sound sources, but considering the amplitude difference and phase difference in sound waves, it can more accurately describe the characteristics of sound wave propagation, thus improving the positioning accuracy, while the far-field model is the opposite [4,5]. Therefore, the near-field model is chosen in this paper, and its applicable area is improved by increasing the array aperture.
The near-field model requires three or more microphones for localization, using a three-microphone uniform linear microphone array as an example shown in the Figure 2. The first microphone m 1 is chosen as the reference microphone, and τ 12 and τ 13 denote the time delay between the acoustic signal arriving at the first microphone m 1 and the second microphone m 2 and the third microphone m 3 , respectively. Then, the time delay τ 12 and τ 13 can be expressed as:
τ 12 = r 2 r 1 c
τ 13 = r 3 r 1 c
Note: r i ( i = 1 , 2 , 3 ) is the distance from the point source to the i microphone, and c is the velocity of the propagation of the acoustic signal in air (generally taken as an approximation of 340 m/s).

2.2. Discrete Distributed Microphone Arrays

The acoustic signals received by microphone arrays based on the near-field model are spherical waves with a high correlation among signals. By increasing the aperture of the array, the reception range of a spherical wave can be enlarged, i.e., the practical application area of the microphone array can be increased. Therefore, a discrete distributed microphone array is selected, as shown in Figure 3.
In the coordinate system O x y , the position of the N microphones are m 1 X 1 , Y 1 , m 2 X 2 , Y 2 , m 3 X 3 , Y 3 m n X n , Y n . The array of N microphones can be flexibly arranged. T X , Y is the true position of the point source, the distance from T to the origin of the O x y coordinate system is the drop-off distance R , the distance to each microphone m i is r i ( i = 1 , 2 , 3 ) , and the time at which the signal arrives at the reference microphone m 1 is t . According to the geometrical relationship, the following can be obtained [6]:
X 2 + Y 2 = R 2 ( X X 1 ) 2 + ( Y Y 1 ) 2 = r 1 2 = ( t c ) 2 ( X X 2 ) 2 + ( Y Y 2 ) 2 = r 2 2 = ( R + τ 12 c ) 2 ( X X 3 ) 2 + ( Y Y 3 ) 2 = r 3 2 = ( R + τ 13 c ) 2 ( X X n ) 2 + ( Y Y n ) 2 = r n 2 = ( R + τ 1 n c ) 2
From Formula (3), we can obtain the hyperbolic intersection equations for the discrete distributed microphone array. By solving this system of hyperbolic equations, we can obtain the location information of the target. This system of equations is nonlinear because it contains square roots and squared terms. To solve this system of equations, we usually need to use nonlinear optimization algorithms, such as Newton’s method, the gradient descent method, or the least squares method. Specifically, we can define an objective function, which is the sum of the squared differences between the left and right sides of all equations, as shown in Formula (4) [7].
F ( X s , Y s ) = n = 2 N ( ( X s X n ) 2 + ( Y s Y n ) 2 ( X s X 1 ) 2 + ( Y s Y 1 ) 2 c τ 1 n ) 2
Our goal is to find the value of ( X s , Y s ) that minimizes F ( X s , Y s ) . Additionally, the selection of the reference microphone is not fixed and can be randomly chosen as the i t h microphone, which can then be combined with the other N 1 microphones.

3. Time Delay Analysis of Sound Source Signals

3.1. Frequency Window Robust Empirical Mode Decomposition Method

3.1.1. Empirical Mode Decomposition (EMD)

The Empirical Mode Decomposition method (EMD) is a method of signal decomposition based on the time scale characteristics of the data itself without any predefined basis functions. Theoretically, it can achieve time resolution and frequency resolution in any dimension, providing a new method for analyzing and guiding smooth and nonlinear signals [8]. Based on its own characteristics, EMD makes the method theoretically applicable to the decomposition of any type of signal, and it has very obvious advantages and high signal-to-noise ratios in dealing with non-smooth and nonlinear data [9].
The traditional EMD decomposition steps can be categorized as follows:
(1)
Find all extreme points of the original signal.
(2)
Fit the upper and lower envelopes e max ( n ) and e min ( n ) by using cubic spline curves.
(3)
Find the mean of the upper and lower envelopes and draw the mean envelope m i ( n ) .
m i ( n ) = e max ( n ) + e min ( n ) 2
(4)
Obtain the middle signal h i ( n ) as the original signal minus the mean envelope.
h i ( n ) = x i ( n ) m i ( n )
During the generation of intermediate signals, a conditional judgement on h i ( n ) is required as follows:
  • In the interval, the number of extreme points and the number of points past zero must be equal or differ by at most one.
  • At any time, the average value of the upper envelope formed by the local maxima and the lower envelope formed by the local minima is zero.
When the above conditions are met, an IMF component can be generated, and the algorithm updates the original signal and repeats the process.
c 1 ( n ) = I M F 1 ( n ) = h i ( n )
r 1 ( n ) = x ( n ) c 1 ( n )
Note: c 1 ( n ) is the first-order IMF component and r 1 ( n ) denotes the updated signal.

3.1.2. Robust Empirical Mode Decomposition with Adaptive Frequency Windows

This paper proposes the use of adaptive frequency windows for locking and extraction in the signal spectrum to solve the problem of missing spectrum boundaries when segmenting the original signal. The schematic process is shown in Figure 4. The frequency window shown in the figure is f a , f b , where f a , f b are the center frequencies of the front and back stopbands of the window. The shaded area is the segmented fragment transition region with a width of 2 π [10]. After initially setting a sliding window width, the sliding window is used to traverse the decomposition frequencies in sequence. The bandwidth range of the frequency window is variable. While traversing, the frequency amplitude within the sliding window f a , f b is solved in parallel. When the calculated amplitude exceeds the set threshold, the sliding window locates the conforming frequency band and outputs the traversal result data.
As the traditional EMD algorithm adopts cubic spline interpolation in the decomposition, the algorithm has endpoint effects when decomposing the IMF. This leads to signal distortion, and there is a large error in the decomposition of the obtained IMF time axis. The principle of the algorithm leads to modal aliasing phenomenon in the decomposed IMF. When the signal-to-noise ratio of the original signal is low, the signal-to-noise ratio of the decomposed IMF is also low, and the effect becomes larger in further decomposition. Since the UAV flight sound is still mixed with noise in different frequency intervals during the process of being received by the sensors, it is necessary to improve the traditional EMD algorithm to adapt to the scenarios applied in this paper.
Aiming at the phenomena existing in the traditional EMD, combined with the application requirements of sound source localization, this paper adopts the improved Empirical Mode Decomposition algorithm, i.e., Robust Empirical Mode Decomposition (REMD).The main idea of REMD is to reduce the influence of noise in the process of signal decomposition and to improve the stability and reliability of the decomposition through the introduction of some robustness criteria or regularization terms [11]. By adjusting the regularization parameter or robustness criteria in the algorithm, the influence of noise can be reduced while retaining the signal characteristics. When the adaptive frequency window locks and splits the original signal, the high and low frequency signals in the split signal are decomposed by weights, and the decomposed IMF components are reconstructed, so as to obtain the reconstructed signal with a high signal-to-noise ratio. The specific implementation process is shown in Figure 5.
An REMD decomposition of the segmented signal x ( n ) is performed, and the decomposition yields n IMF components.
x ( n ) = i = 1 N c i ( n ) + r N ( n )
Note:   c i ( n ) is the improved robust intrinsic modal function and r N ( n ) is the residual term.
1.
Filter the components with higher signal-to-noise ratio according to the following reconstruction criterion:
(1)
Correlation coefficient
The correlation coefficient is a statistical indicator of the close connection among response variables. In this paper, Pearson’s correlation coefficient, which is commonly used in statistics, is used [12]. The degree of correlation between X and Y is described by a value in the interval [−1, 1]; the larger the value of the coefficient, the higher the correlation, and vice versa. After performing REMD decomposition, the algorithm automatically calculates the correlation coefficients between the IMF components and the segmented signals as a way to differentiate the useful components and retain them. Let two samples be X and Y . The following equation represents the Pearson correlation coefficient ( ρ x , y ) of two continuous variables ( x , y ) , which is equal to the product of the covariance cov ( X , Y ) between them divided by their respective standard deviations ( σ x , σ y ) . Variables close to 0 are considered uncorrelated and those close to 1 or −1 are said to be strongly correlated.
ρ X , Y = cov ( X , Y ) σ X σ Y = E ( X μ x ) ( Y μ y σ X σ Y = E ( X Y ) E ( X ) E ( Y ) E ( X 2 ) E 2 ( X ) E ( Y 2 ) E 2 ( Y ) ρ X , Y = N X Y X Y N X 2 ( X ) 2 N Y 2 ( Y ) 2 ρ X , Y = X Y X Y N ( X 2 ( X ) 2 N ) ( Y 2 ( Y ) 2 N )
(2)
Amplitude ratio
The mathematical expression for the energy ratio coefficient is
ε = E I M F ( i ) E x
where E x is the total energy of the IMF component; E I M F ( i ) is the energy of different IMFs; and ε is the energy ratio coefficient.
2.
According to the proportional weighting method, take the maximum correlation coefficient as I M F max ( n ) and the minimum correlation coefficient as I M F min ( n ) , subtract the two components and compute the average value, add the value to the segmented signal x ( n ) , and reconstruct it to obtain a signal with high signal-to-noise ratio.
x ( n ) = x ( n ) + I M F max ( n ) I M F min ( n ) 2
3.
Analyze the trigger time values in the reconstructed signal.

3.1.3. Time Delay Estimation Method Based on the Comparison of Amplitude Mean Values of Reconstructed Signals

The time delay estimation method based on the comparison of amplitude mean values of reconstructed signals involves calculating the amplitude α at each sampling point of a pure signal and then determining its average value α ¯ . By comparing the amplitude mean values of the reconstructed signal and the reference signal, the time delay of the signal can be inferred. If there is a shift in the amplitude mean of the reconstructed signal relative to the reference signal, this offset can serve as an estimate τ i j for the time delay. The effectiveness of this approach relies on the characteristics of the signal and the noise level. However, the use of improved REMD (Robust Empirical Mode Decomposition) can circumvent this issue.
To simplify the calculation, Figure 5 illustrates that this study only focuses on the amplitude within the first peak of the reconstructed signal to calculate the mean amplitude value. One of the reconstructed REMD signals from a microphone is selected as the reference signal, and its mean amplitude is compared with that of other microphone signals. The offset of the same amplitude value is chosen as the time delay estimation. When the ratio of the mean amplitude value α ¯ i , α ¯ i 1 at time T i to that at time T i 1 reaches its maximum, we consider that τ i , i 1 = T i T i 1 .

3.2. GWO-Based Adaptive Optimization

The Grey Wolf Optimization (GWO) algorithm is inspired by grey wolf packs in nature, and the GWO algorithm simulates the hierarchy and hunting mechanism of grey wolf packs in nature [13]. The Grey Wolf Optimization algorithm does not require complex parameter tuning, and, at the same time, it has a strong global search ability, which can quickly find the optimal solution. This article seeks the optimal fitness function of the adaptive frequency window with the help of the advantages of this algorithm. GWO classifies the entire wolf population into Alpha, Beta, Delta, and Omega wolves according to the fitness curve, with wolves decreasing in rank in the following update process:
(1)
Calculate the distance between an individual grey wolf and its prey and update the grey wolf’s position at all times:
D ( t ) = C X p ( t ) X ( t )
X ( t + 1 ) = X p ( t ) A D ( t )
where t is the number of moment iterations; D ( t ) is the distance between the individual grey wolf and the prey; X p is the position vector of the prey; X is the position vector of the individual grey wolf; and A , C is the coefficient vector, which is calculated as:
A = 2 α r 1 α
C = 2 r 2
α = 2 ( 1 t / t max )
where r 1 , r 2 is the random vector in the interval [0, 1], t max is the maximum number of iterations, and a is the convergence factor.
Grey wolves are able to identify the location of their prey and surround them [14]. When a grey wolf identifies the location of a target,   β and δ guide the pack to encircle the prey under the leadership of α . Constantly update their positions using their tracking mathematical model expressions as follows:
D α = C 1 X α ( t ) X ( t )
D β = C 2 X β ( t ) X ( t )
D δ = C 2 X δ ( t ) X ( t )
X 1 = X α A 1 ( D α )
X 2 = X β A 2 ( D β )
X 3 = X δ A 3 ( D δ )
X ( t + 1 ) = X 1 + X 2 + X 3 3
where D α , D β and D δ denote the distance between α , β and δ and other individuals, respectively; X α , X β and X δ represent the current positions of α , β and δ , respectively; C 1 , C 2 and C 3 are random vectors; and X is the current position of the grey wolf. Equations (21)–(23) denote the step length and direction of individual w in the wolf pack towards e, f, and g, respectively, and Equation (24) denotes the final position of W .
The coefficient vector A in Equation (15) represents a random value in the interval α , α , and during the attack process of the wolf pack, the value of a is gradually reduced, while the range of fluctuation in A is also reduced. When A < 1 , the wolf pack approaches and captures the prey, i.e., the optimal solution is obtained, and when A > 1 , the grey wolf separates from the prey and searches for a more suitable target [15]. In this paper, the noise ratio in the frequency domain is used as the fitness function of the frequency window, and the GWO algorithm is used to adaptively determine the window position. Figure 6 shows the schematic of the GWO adaptive approach.

4. Improved Sound Source Localization Algorithms

4.1. TDOA Algorithm

The TDOA algorithm works by measuring the time differences between signals received by different receivers and uses these time differences to calculate the position of the transmitting source. Specifically, it is assumed that there are three or more receivers with known relative positions to each other and that when the source transmits a signal, each receiver records the time at which the signal is received. Since the velocity of signal propagation is known, this time difference and velocity information can be used to construct a system of localization hyperbolic equations to calculate the position of the signal source. However, this hyperbolic system of equations tends to be nonlinear, and the obtained solution is often imaginary, or there is no solution, which needs to be solved by transforming the nonlinear system of equations into a linear system of equations. The mainstream methods are the Fang algorithm, the Chan algorithm, and the Taylor algorithm.
In this paper, the Chan algorithm and the Taylor algorithm are selected, and five different sizes of Gaussian white noise are selected for simulation. Figure 7 represent the Chan algorithm measurement noise standard deviation, the Taylor algorithm measurement noise standard deviation, and the Chan and Taylor algorithm measurement noise standard deviation, respectively.
The simulation results in Figure 7 show that the RMSE (Root Mean Square Error) of the Chan algorithm and the Taylor algorithm are basically the same when the noise is small. As the noise of the measurements becomes bigger and bigger, the Taylor algorithm’s RMSE is obviously better than that of the Chan algorithm.

4.2. Improving the Chan–Taylor Localization Algorithm

As the Chan algorithm can make full use of the information of the TDOA measurement value of each base station, it can effectively reduce the negative impact of the large measurement error of individual base stations on the positioning, and the Taylor algorithm obtains higher positioning accuracy when the initial position selection accuracy is higher. Therefore, in this paper, the Chan algorithm and the Taylor algorithm are fused to locate the position of experimental micro-UAV, and five different sizes of Gaussian white noise are selected for simulation.
The principle of the Chan–Taylor joint positioning algorithm is to use the Chan algorithm to estimate the initial coordinates of the target x 0 , y 0 . After the Chan algorithm, Taylor iterative calculations are carried out to constantly update the position label obtained by the method that more closely resembles the true position x , y . In this paper, the Chan–Taylor algorithm is selected and combined with the weighted least squares method for solving. The detailed derivation process is shown in Appendix B; here, we simply show the algorithm processing steps.
(1)
In the first step, apply WLS estimation to the nonlinear equations, to obtain a quadratic equation in terms of r 1
.
a r 1 2 + b r 1 + c = 0
Then, solve the equation for the two roots available, omit the invalid solutions, and substitute the valid solutions back into Equation (12) to obtain the estimated coordinates x 0 , y 0 of the location point T .
(2)
Use the estimated coordinates x 0 , y 0 obtained by the Chan algorithm as the initial values of Taylor’s series expansion method. Performing a Taylor expansion of f i ( x , y , x i , y i ) at x 0 , y 0 and translate into matrix form:
ψ = h i G i δ
Finally, solve the weighted least squares solution of the above equation to obtain the final target coordinates:
δ = ε x ε y = G i T Q 1 G i 1 G i T Q 1 h i
where T x , y is the final positioning coordinate of the target that is infinitely close to the true value after iterative calculation.

5. Verification and Analysis

In this section, we test the performance and robustness of the proposed localization algorithm. In Section 5.1, we compare the EMD with the improved algorithm proposed in this paper by loading simulated signals under ideal conditions. We draw conclusions by comparing the IMF curves of the two algorithms. Section 5.2 verifies the actual performance of the strategy and algorithm proposed in this paper using signals collected in a real environment, thus demonstrating the accuracy and feasibility of the entire sound source localization algorithm proposed in this paper for drone localization scenarios.

5.1. Algorithm Stability Validation

Based on the analysis of the flight characteristics of drones, the signal generated by the rotating blades during drone acceleration can be regarded as a short-term impulse signal. This signal is characterized by a peak value appearing after a period of time, which gradually weakens over time. Because of the complex surrounding environment of the experiment, the drone flight signal may be interfered with by certain high-frequency vibration signals in the air and soil. As a result, the received signal may contain high-energy high-frequency vibration signals and low-amplitude random white noise signals. Therefore, we created a similar environment in the laboratory, where the hardware equipment consists of a DAQ122 acquisition card and a discrete microphone array composed of three MEMS digital microphones. The signal collected by the host computer is shown in the figure. We compared the EMD algorithm with the proposed algorithm using this analog signal. The simulation was performed using Matlab (Version: R2022b, created by MathWorks in Natick, MA, USA), and the simulation results are shown in Figure 8, Figure 9, Figure 10 and Figure 11.
The IMF diagrams decomposed by the two algorithms are both listed in Appendix A. By comparing the results in the appendix, it is concluded that the orthogonality between each IMF obtained after REMD of the collected signals is higher than that of the IMF decomposed by EMD, which makes the decomposed IMF more independent. At the same time, after improving robustness, the IMF decomposed by REMD fluctuates significantly, and the IMF decomposed at low frequencies has more convergent stability. There is no signal aliasing in each IMF, which can more accurately capture the local characteristics of the signal, thus greatly improving the performance of later signal reconstruction and adaptive sliding window positioning.

5.2. Real Signal Simulation Experiment

Within the test area, to eliminate the influence of human manipulation on the drone’s flight position, a system verification test was conducted using pre-recorded audio of a drone’s accelerating ascent sound signal to simulate drone flight. In the experiment, the actual positions of the sensors were preset and further corrected using differential GPS for positional coordinates. The recorded drone flight sound signal was then amplified and played at a designated point. The microphone array in the test site collected and processed these signals. After obtaining the time delay values, the TDOA algorithm was utilized for positioning. The system recorded each solution result and compared the positioning outcomes with the preset real target points to evaluate positioning accuracy and system robustness. Figure 12 shows the microphone placement at the test site.
A total of 10 valid tests were conducted in the experiment. Figure 13 shows a schematic diagram of the target positions and sensor positions corresponding to the ten tests. The same drone flight sound signal was used at all playback points, and the signal decibel was unified. Figure 14 shows the pre-collected drone flight sound signals collected by three microphone sensors in the first experiment. According to pre-experiment tests, the signals collected by the sensors in the experiment were basically consistent with the pre-collected drone flight signals. Therefore, they could be used as experimental simulation signals. Meanwhile, there were slight environmental noises such as wind, people walking, and vehicles driving around during the experiment. Table 1 is a record of the trigger times for the signals collected by the system analysis, and Table 2 is a record of the ten tests for the real landing point ( x , y ) and the system positioning estimate landing point ( x * , y * ) , which are used to calculate the positioning error δ .
As can be seen in Figure 14, because of the limited distance of the microphones, the time received by each microphone is very close. For the positioning algorithm adopted in this paper, the signals processed without any treatment or using traditional signal decomposition algorithms cannot guarantee the positioning accuracy of the drone. Next, we decompose the signals collected by the three microphones.
As shown in Figure 15, the amplitudes of the IMFs decomposed by the algorithm in this paper for the three microphones are the same, indicating that the filtering approach adopted in this paper completely separates the mixed noise in the acoustic wave. It can also be seen that the filtering used does not clip the useful signal at the top or bottom. The signal spectrum obtained by FFT transformation is shown in Figure 16. A sliding frequency window with variable bandwidth is used for corresponding frequency matching and extraction. The reconstructed signal filtered by the screening criteria is shown in Figure 17. By analyzing the data and using the algorithm to calculate the time difference between different sensor receptions, the trigger signal moments in the extracted IMF components are determined and recorded.
Following this process, the time delay results obtained from the remaining nine experiments are recorded in Table 1. Meanwhile, the estimated position coordinates are compared with the actual measured coordinates using the positioning algorithm presented in this paper, and the calculated error values are recorded in Table 2.
In Table 1 and Table 2, it can be seen that in the same experimental environment, after analysis using the improved Empirical Mode Decomposition sound source localization algorithm with an adaptive frequency window, the positioning estimation error is below 5% within the range where signals can be collected. This indicates that using the algorithm presented in this paper, the time delay estimation value resolved from the decomposed signal greatly improves the success rate of position estimation. This algorithm performs well in both near and far fields and even in the presence of random noise. However, during the experiment, it was found that the positioning accuracy was still not high enough in some locations. The specific reason is that because of product packaging and other factors, the omnidirectional microphone does not have exactly the same sensitivity in all 360-degree directions, which interferes with the selection of relevant points for time delay estimation. Thus, subsequent algorithms still need to make corresponding compensations for special values.

6. Conclusions

In this paper, a drone sound source localization system based on the Empirical Mode Decomposition (EMD) algorithm is constructed. In the presence of noise aliasing, the system can separate signals of different frequency bands and recombine them. Using frequency domain analysis of drone flight signals before the experiment, the complete signal of the drone flight can be effectively decomposed, thus improving the accuracy of time delay estimation. This also improves the rapid performance degradation of the Chan–Taylor algorithm when the time delay estimation accuracy is insufficient. At the same time, a distributed sensor array is adopted, which can be freely arranged, and a mathematical geometric positioning model for random arrangement of N microphones is derived, expanding the positioning area and flexibility.
However, this paper did not conduct experiments on real drone flights in actual test sites. Instead, the test conditions were carried out in a semi-open field environment without multipath interference. Our next research goal is to further improve the performance of the EMD algorithm when drone flight signals are reflected and aliased multiple times.

Author Contributions

Conceptualization, T.C. and Z.Y.; Methodology, T.C. and Z.Y.; Software, T.C.; Validation, T.C.; Formal analysis, T.C.; Investigation, T.C. and Z.Y.; Resources, T.C. and Z.Y.; Data curation, T.C. and Z.Y.; Writing—original draft, T.C.; Writing—review & editing, J.Y.; Visualization, J.Y.; Supervision, J.Y.; Project administration, J.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This work was sponsored by Nanjing University of Science and Technology.

Data Availability Statement

Data on supporting conclusions have been charted and presented in the article.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Diagram Representation of the Simulated Signal Simulation

  • Traditional EMD decomposition of the simulated signals.
Sensors 24 02701 g0a1aSensors 24 02701 g0a1b
2.
A modified REMD decomposition of the simulated signals.
Sensors 24 02701 g0a2aSensors 24 02701 g0a2b

Appendix B. The Specific Derivation Process of the Chan–Taylor Algorithm

The specific algorithmic process is as follows:
(1)
Initial estimation of the target’s true position using the Chan algorithm [16]. The position of microphone m i is known to be ( x i , y i ) i = 1 , 2 , 3 . Assuming that the true position of the target is T ( x , y ) , the following can be determined:
r i = ( x i x ) 2 + ( y i y ) 2
r i , 1 = r i r 1 = x i x 2 + y i y 2 x 1 x 2 + ( y 1 y ) 2
Let K i = x i 2 + y i 2 , according to Equation (A1), we have:
r i 2 = x i 2 + y i 2 = K i 2 x i x 2 y i y + x 2 + y 2
According to Equation (A2), we have:
r i 2 = ( r i , 1 + r i ) 2
Substitute Equation (A3) into Equation (A4) and let x i , 1 = x i x 1 :
r i , 1 2 + 2 r i , 1 r 1 = ( K i K 1 ) 2 x i , 1 x 2 y i , 1 y
This step eliminates the squared terms of the unknowns and preserves a series of linear equations which, when i = 1 , 2 , 3 , gives the following expression:
2 x 2 , 1 x + 2 y 2 , 1 y = ( K 2 K 1 ) ( r 2 , 1 2 + 2 r 2 , 1 r 1 ) 2 x 3 , 1 x + 2 y 3 , 1 y = ( K 3 K 1 ) ( r 3 , 1 2 + 2 r 3 , 1 r 1 )
The following solution can be obtained:
x = y 2 , 1 r 3 , 1 2 y 3 , 1 r 2 , 1 2 + y 3 , 1 ( K 2 K 1 ) y 2 , 1 ( K 3 K 1 ) + 2 ( y 2 , 1 r 3 , 1 y 3 , 1 r 2 , 1 ) r 1 2 ( x 2 , 1 y 3 , 1 x 3 , 1 y 2 , 1 ) y = x 2 , 1 r 3 , 1 2 x 3 , 1 r 2 , 1 2 + x 3 , 1 ( K 2 K 1 ) x 2 , 1 ( K 3 K 1 ) + 2 ( x 2 , 1 r 3 , 1 x 3 , 1 r 2 , 1 ) r 1 2 ( x 3 , 1 y 2 , 1 x 2 , 1 y 3 , 1 )
Let
p 1 = x , p 2 = y
q 1 r 1 = 2 ( y 2 , 1 r 3 , 1 y 3 , 1 r 2 , 1 ) r 1 2 ( x 2 , 1 y 3 , 1 x 3 , 1 y 2 , 1 ) q 2 r 1 = 2 ( x 2 , 1 r 3 , 1 x 3 , 1 r 2 , 1 ) r 1 2 ( x 3 , 1 y 2 , 1 x 2 , 1 y 3 , 1 )
Then, Equations (A7) and (A8) can be simplified to the following calculation equation:
x = p 1 + q 1 r 1 y = p 2 + q 2 r 1
Substituting Equation (A9) into Equation (A5), such that
a = ( q 1 2 + q 2 2 1 ) b = 2 [ q 1 ( x 1 p 1 ) + q 2 ( y 1 p 2 ) ] c = [ ( x 1 p 1 ) 2 + ( y 1 p 2 ) 2 ]
A quadratic equation about r 1 can be obtained:
a r 1 2 + b r 1 + c = 0
Solve the equation for the two roots available, omit the invalid solutions, and substitute the valid solutions back into Equation (12) to obtain the estimated coordinates x 0 , y 0 of the location point T x , y .
(2)
Substitute the estimated coordinates x 0 , y 0 obtained by the Chan algorithm as the initial values of the Taylor series expansion method. Let the true coordinate value of the target be x , y :
x = x 0 + ε x y = y 0 + ε y
Then, the Taylor series expansion of the function f i ( x , y , x i , y i ) at x 0 , y 0 is as follows:
f i ( x , y , x i , y i ) = f i ( x 0 , y 0 , x i , y i ) + ( ε x x + ε y y ) f i ( x 0 , y 0 , x i , y i ) + 1 2 ! ( ε x x + ε y y ) 2 f i ( x 0 , y 0 , x i , y i ) + + 1 n ! ( ε x x + ε y y ) 2 f i ( x 0 , y 0 , x i , y i ) +
Substituting Equations (A1) and (A2) into Equation (A12) gives:
f i ( x , y , x i , y i ) = r i r j = ( x i x ) 2 + ( y i y ) 2 ( x j x ) 2 + ( y j y ) 2
By performing a Taylor expansion of f i ( x , y , x i , y i ) at x 0 , y 0 and neglecting components above second order, we have:
f i ( x , y , x i , y i ) = f i ( x 0 , y 0 , x i , y i ) + ( ε x x + ε y y ) f i ( x 0 , y 0 , x i , y i )
Translated into matrix form, we obtain:
ψ = h i G i δ
Notes: ψ is the error vector.
Then,
h i = r 2 , 1 ( r 2 r 1 ) r 3 , 1 ( r 3 r 1 ) r L , 1 ( r L r 1 )
G i = x 1 x r 1 x 2 x r 2 y 1 y r 1 y 2 y r 2 x 1 x r 1 x 3 x r 3 y 1 y r 1 y 3 y r 3 x 1 x r 1 x L x r L y 1 y r 1 y L y r L
(3)
Solve the weighted least squares solution of the above equation:
δ = ε x ε y = G i T Q 1 G i 1 G i T Q 1 h i
where Q is the covariance matrix of the noise measurements, and in the next iteration of the computation, let x = x 0 + ε x , y = y 0 + ε y . The above computation process is repeated until the error stops when it meets the set threshold, i.e., ε x + ε y < η .
The x , y obtained at this point is the final positioning coordinate of the target that is infinitely close to the true value after iterative calculation.

References

  1. Chen, P.; Chen, Y.; Wang, W. A review of acoustic localization techniques for unmanned aircraft. J. South China Univ. Technol. (Nat. Sci. Ed.) 2022, 50, 109–123. [Google Scholar]
  2. Zou, Y.; Jia, Y.; Liu, S. An algorithm for human-vehicle ground vibration signal recognition based on improved EMD decomposition. China Test. 2022, 48, 85–94. [Google Scholar]
  3. Li, S.C.; Wu, J.-F. 5G indoor positioning algorithm based on Chan-Taylor and optimized BP neural network. Chin. J. Inert. Technol. 2023, 31, 806–813, 822. [Google Scholar] [CrossRef]
  4. Wang, F. Research on Three-Dimensional Spatial Sound Source Localization System Based on Microphone Array. Master’s Thesis, Hebei University of Engineering, Handan, China, 2010. [Google Scholar] [CrossRef]
  5. Lu, G.Q. Research on Sound Source Target Search Based on Swarm Robot Collaboration. Master’s Thesis, Hebei University of Technology, Handan, China, 2021. [Google Scholar]
  6. Cho, K.; Nishiura, T.; Yamashita, Y. A Study on Multiple Sound Source Localization with a Distributed Microphone System. In Proceedings of the 10th INTERSPEECH 2009 Conference, Brighton, UK, 6–10 September 2009. [Google Scholar]
  7. Zheng, J.; Zhang, B.; Xiong, C. Acoustic localization method for projectile drop observation based on composite dual-base array. J. Ballist. 2016, 28, 68–74. [Google Scholar] [CrossRef]
  8. Chen, Z.; Luo, W.; Chang, J.; Zhong, H.; Wu, Z.; Zheng, Y. EMD-based neural network air-coupled ultrasonic oil storage tank level detection. China Test. 2021, 47, 9–14. [Google Scholar] [CrossRef]
  9. Langston Technology (Beijing) Co. A Transmission System Fault Diagnosis Method Based on Multi-Information Fusion. Chinese Patent CN202010199373.1, 12 November 2021. [Google Scholar]
  10. Deng, F.; Qiang, Y.; Yang, S.; Hao, R.; Liu, Y. An adaptive frequency window empirical wavelet transform for rolling bearing fault diagnosis. J. Xi’an Jiaotong Univ. 2018, 52, 22–29. [Google Scholar] [CrossRef]
  11. Zhao, Z.; Liu, H.; Zhang, H. Adaptive regularization term to remove multiplicative noise. Electron. Sci. Technol. 2011, 24, 1–3, 8. [Google Scholar] [CrossRef]
  12. Yang, L.; Fan, D.; Zhou, R. Gearbox fault diagnosis based on EEMD combined with quadratic wavelet packet noise reduction. J. N. Cent. Univ. (Nat. Sci. Ed.) 2016, 37, 262–267. [Google Scholar] [CrossRef]
  13. Xia, Y.; Wang, C.; Feng, X.; Cai, M. Prediction of lubricant friction and wear performance based on GRNN optimized by Gray Wolf algorithm. J. Tribol. 2023, 43, 947–955. [Google Scholar] [CrossRef]
  14. Hangzhou Xinhe Shengshi Technology Co. A BP Neural Network Optimization Method Based on Improved Gray Wolf Algorithm. Chinese Patent CN202010496667.0, 29 September 2020. [Google Scholar]
  15. Song, J. Research on Underdetermined Blind Source Separation Algorithm of Hydroacoustic Signal Based on Sparse Component Analysis. Master’s Thesis, Qingdao University of Science and Technology, Qingdao, China, 2023. [Google Scholar]
  16. Zhu, Y. Analysis and Processing of Seismic Wave Signals in Explosion Point Localization. Master’s Thesis, North Central University, Taiyuan, China, 2021. [Google Scholar]
Figure 1. Positioning flowchart.
Figure 1. Positioning flowchart.
Sensors 24 02701 g001
Figure 2. Near-field modeling of uniform line arrays.
Figure 2. Near-field modeling of uniform line arrays.
Sensors 24 02701 g002
Figure 3. Discrete distributed microphone array model.
Figure 3. Discrete distributed microphone array model.
Sensors 24 02701 g003
Figure 4. Adaptive frequency window.
Figure 4. Adaptive frequency window.
Sensors 24 02701 g004
Figure 5. Schematic diagram of the time delay estimation method based on amplitude mean value comparison.
Figure 5. Schematic diagram of the time delay estimation method based on amplitude mean value comparison.
Sensors 24 02701 g005
Figure 6. Schematic diagram of the GWO adaptive algorithm.
Figure 6. Schematic diagram of the GWO adaptive algorithm.
Sensors 24 02701 g006
Figure 7. Different algorithms used to measure standard deviation.
Figure 7. Different algorithms used to measure standard deviation.
Sensors 24 02701 g007
Figure 8. Site layout.
Figure 8. Site layout.
Sensors 24 02701 g008
Figure 9. Simulated acoustic signal time domain diagram.
Figure 9. Simulated acoustic signal time domain diagram.
Sensors 24 02701 g009
Figure 10. IMF decomposition of the EMD algorithm for simulating acoustic signals.
Figure 10. IMF decomposition of the EMD algorithm for simulating acoustic signals.
Sensors 24 02701 g010
Figure 11. IMF decomposition of the REMD algorithm for simulating acoustic signals.
Figure 11. IMF decomposition of the REMD algorithm for simulating acoustic signals.
Sensors 24 02701 g011
Figure 12. Map of sensor positions during the test.
Figure 12. Map of sensor positions during the test.
Sensors 24 02701 g012
Figure 13. Schematic diagram of the corresponding target points and sensor positions in the test.
Figure 13. Schematic diagram of the corresponding target points and sensor positions in the test.
Sensors 24 02701 g013
Figure 14. Waveforms of the signals collected by the three array microphones.
Figure 14. Waveforms of the signals collected by the three array microphones.
Sensors 24 02701 g014
Figure 15. REMD decomposition of the signals collected by the three array microphones.
Figure 15. REMD decomposition of the signals collected by the three array microphones.
Sensors 24 02701 g015aSensors 24 02701 g015b
Figure 16. Spectrogram.
Figure 16. Spectrogram.
Sensors 24 02701 g016
Figure 17. Reconstruction of the IMF trigger signal moment chart: (ac) denote the IMF decomposition process of the signal, respectively, and c is the final IMF decomposition result.
Figure 17. Reconstruction of the IMF trigger signal moment chart: (ac) denote the IMF decomposition process of the signal, respectively, and c is the final IMF decomposition result.
Sensors 24 02701 g017aSensors 24 02701 g017b
Table 1. Time estimation results after REMD decomposition.
Table 1. Time estimation results after REMD decomposition.
Serial NumberMic(i) Trigger Moment/sMic2, Mic3, and Mic1 Time Difference/s
123α1α2
11.71231.72171.72820.009410.01592
21.62721.630731.641210.003530.01401
31.72331.730831.738510.007530.01521
41.23411.248081.248080.013980.01398
51.17520.189721.1816310.014520.006431
62.38122.393512.381310.012310.00011
71.42511.432071.416960.00697−0.00814
82.31272.31282.298790.0001−0.01391
91.45271.446361.33037−0.00634−0.01498
101.81621.803871.80387−0.01233−0.01233
Table 2. Test drop monitoring results.
Table 2. Test drop monitoring results.
Serial NumberTrue Point of ImpactThe Estimated Position δ / %
x / m y / m x * / m y * / m
1302.99350.00984.48
22.5981.52.616361.516432.41
31.52.5981.498532.610341.00
40303.021332.13
5−1.52.598−1.489132.622741.61
6−2.5981.5−2.590351.495640.88
7−30−3.027180.001686322.72
8−2.598−1.5−2.59451−1.497750.41
9−1.5−2.598−1.46837−2.750189.77
100−3−0.0001569−3.068544.85
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, T.; Yu, J.; Yang, Z. Research on a Sound Source Localization Method for UAV Detection Based on Improved Empirical Mode Decomposition. Sensors 2024, 24, 2701. https://doi.org/10.3390/s24092701

AMA Style

Chen T, Yu J, Yang Z. Research on a Sound Source Localization Method for UAV Detection Based on Improved Empirical Mode Decomposition. Sensors. 2024; 24(9):2701. https://doi.org/10.3390/s24092701

Chicago/Turabian Style

Chen, Tao, Jiyan Yu, and Zhengpeng Yang. 2024. "Research on a Sound Source Localization Method for UAV Detection Based on Improved Empirical Mode Decomposition" Sensors 24, no. 9: 2701. https://doi.org/10.3390/s24092701

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop