Next Article in Journal
Could Caregivers’ Stressful Care Experiences Be Related to Their Attitudes towards Advance Care Planning? A Cross-Sectional Descriptive Study
Previous Article in Journal
Exploring an Efficient Remote Biomedical Signal Monitoring Framework for Personal Health in the COVID-19 Pandemic
Previous Article in Special Issue
Quantification of the Psychoacoustic Effect of Noise from Small Unmanned Aerial Vehicles
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Assessment of Sound Source Tracking Using Multiple Drones Equipped with Multiple Microphone Arrays

1
Department of Systems and Control Engineering, School of Engineering, Tokyo Institute of Technology, 2-12-1 Ookayama, Meguro-ku, Tokyo 152-8552, Japan
2
Honda Research Institute Japan Co., Ltd., 8-1 Honcho, Wako, Saitama 351-0188, Japan
*
Author to whom correspondence should be addressed.
Int. J. Environ. Res. Public Health 2021, 18(17), 9039; https://doi.org/10.3390/ijerph18179039
Submission received: 1 July 2021 / Revised: 14 August 2021 / Accepted: 23 August 2021 / Published: 27 August 2021
(This article belongs to the Special Issue Drone Noise: A New Public Health Challenge)

Abstract

:
Drone audition techniques are helpful for listening to target sound sources from the sky, which can be used for human searching tasks in disaster sites. Among many techniques required for drone audition, sound source tracking is an essential technique, and thus several tracking methods have been proposed. Authors have also proposed a sound source tracking method that utilizes multiple microphone arrays to obtain the likelihood distribution of the sound source locations. These methods have been demonstrated in benchmark experiments. However, the performance against various sound sources with different distances and signal-to-noise ratios (SNRs) has been less evaluated. Since drone audition often needs to listen to distant sound sources and the input acoustic signal generally has a low SNR due to drone noise, making a performance assessment against source distance and SNR is essential. Therefore, this paper presents a concrete evaluation of sound source tracking methods using numerical simulation, focusing on various source distances and SNRs. The simulated results captured how the tracking performance will change when the sound source distance and SNR change. The proposed approach based on location distribution estimation tended to be more robust against distance increase, while existing approaches based on directional estimation tended to be more robust against decreasing SNR.

1. Introduction

Drones have recently been required to be used at disaster sites. By attaching auditory sensors, drones can obtain auditory information, and this field of auditory information processing for drones is referred to as “drone audition”. Drone audition techniques enable drones to find targets in low light conditions and have higher quality scene analysis with sound recognition techniques. In particular, microphone arrays are commonly used since they can estimate sound source directions by calculating the time difference of arrival between microphones. One example of drone audition applications is to carry out people searching tasks [1,2,3]. By estimating the sound direction of people calling for help, drones will be able to find them even if they are covered in rubble. The main goal of this people searching task is to localize a stationary sound source (the person to be rescued), but only localizing stationary sound sources is not enough to accomplish this task. In addition to localization of stationary sources, tracking moving sound sources and sound source separation are also required.
Since different sound sources can exist in the disaster site, drones will also hear non-target sound sources (such as rescue staff, heavy machinery, and so on), and the input to a microphone array is often a mixture of several acoustic signals. In order to enable the drone to recognize which sound source to target, identifying the movement and individual acoustic signal of each sound source is necessary. If the drone realizes that a human-like audio signal has been emitted from a stationary sound source, it is likely to be the target sound source. Besides, research shows that the information of the source location is helpful for sound source separation techniques [4,5], which demonstrates the importance of tracking moving sound sources in drone audition. Therefore, several sound source tracking methods for drones has been proposed [1,6,7,8]. Authors also have proposed a sound source tracking method that utilizes multiple microphone arrays [9,10]. By integrating the spatial spectra computed from each microphone array, the likelihood distribution of the sound source location, which is useful for sound source tracking, can be obtained. Many sound source tracking methods have also been proposed, not only for drones. For example, microphones can be equipped around a room to track human activities [11,12]. Similar to drones, land-based robots can also benefit from microphone arrays, so they can detect and localize sound sources and recognize the environment of the surroundings [13,14,15,16,17]. These tracking methods can be applied to drone systems. However, it is questionable whether these tracking methods are affected by (1) a long distance to the target sound source and (2) significant drone noise. Most research evaluates tracking performance by conducting a few experiments and does not provide an exhaustive evaluation of distance or SNR. For example, Lauzon et al. proposed a sound source tracking method for drone detection and showed its effectiveness through a single outdoor experiment, but tracking performance against various scenarios was not examined [18]. Brandstein et al. also proposed a sound source tracking method and demonstrated experiments to evaluate tracking performance against various source distances. However, the source distance varied between 1 and 3 m, which is not sufficient in terms of drone applications [19]. Since experiments using drones take time and involve high costs, it is difficult to perform exhaustive evaluations through real-world experiments. Therefore, this paper aims to assess the tracking performance of tracking methods by performing numerical simulations for various scenarios, especially various source distances and SNRs.
The rest of the paper is organized as follows. Related work is mentioned in Section 2 and the proposed method [10] is explained in Section 3. The evaluation via numerical simulation is described in Section 4. Finally, the conclusion is written in Section 5.

2. Related Work

In the field of acoustic signal processing, microphone arrays are well used to localize the direction or location of the sound source. To track sound source locations using microphone arrays, several measures have been taken. One idea is to estimate the distance between a microphone array and a sound source by applying Bayesian filtering to estimated directions [13,14,20]. However, sound source tracking using a single microphone array is difficult since estimation converging to the ground truth takes time, which is concerning when the sound source is moving.
Therefore, using multiple microphone arrays is much more popular when it comes to tracking sound source locations. As well as using a single microphone array, applying Kalman filtering or particle filtering to estimated directions obtained from multiple microphone arrays shows high tracking performance [18]. Another way is to use triangulation to obtain the sound source location. One of the biggest benefits of using multiple microphone arrays is that we can explicitly obtain the sound source location through triangulation. Since the approximate location can be obtained from one direction estimation, convergence to the ground truth is much faster. Therefore, there are many studies that utilize triangulation for both localizing stationary sound sources [21,22] and tracking moving sound sources [9,23]. If the sound source is assumed to be on the ground, calculating the intersection point between the sound source direction and the ground surface is also a good way to localize the sound source [1,6]. However, when it comes to drone applications, severe ego-noise and wind noise may distort estimation results of the direction, which will negatively affect triangulation methods. Since both commonly used direction estimation and triangulation have discreteness in their calculation, a little distortion will turn to a large tracking error.
Therefore, we introduce a sound source tracking method that estimates the location distribution of the sound source without using triangulation. Generally, direction estimation methods first calculate the likelihood distribution of the direction and then search for the direction with the largest likelihood. In previous work, we integrated likelihood distributions obtained from each microphone array and obtained the location likelihood distribution instead of estimating directions and computing triangulation points [10]. In this way, the information of the sound source location is represented in a location distribution, which is a more continuous representation than a group of triangulation points, and we expect that it will be more robust to drone noise. We apply this location distribution to particle filtering techniques in order to track moving sound sources. In this paper, we will evaluate the tracking performance of existing work [18,23] and the proposed method named particle filtering with integrated MUSIC (PAFIM) [10] via numerical simulation in order to understand the performance against different source distances and different SNRs.

3. Method

This section explains PAFIM [10], a sound source tracking method based on location likelihood estimation. The likelihood distribution of the location is obtained by integrating the likelihood distribution of the sound source direction.

3.1. Settings

We consider tracking a sound source by multiple microphone arrays. We assume that N microphone arrays are mounted to drones and each microphone array is numbered as follows.
MA 1 , , MA n , , MA N
Each microphone array is mounted to a drone, and the state of MA n is described as below, and assumed to be known.
m n , k = x n , k , y n , k , z n , k , ϕ n , k , θ n , k , ψ n , k T
where k indicates the time step, x n , k , y n , k , and z n , k indicate the center of MA n in three-dimensional coordinates and ϕ n , k , θ n , k , and ψ n , k indicate the three-dimensional rotational angles of MA n . Each microphone array consists of M microphones. We assume the sound source to be a point source, and its location is described as below.
e k = x e , k , y e , k , z e , k T
The problem addressed in this paper is to estimate the sound source trajectory by repeatedly estimating the sound source location e k from the state of all N microphone arrays and recorded sound signal s 1 , , e N R M .

3.2. System Outline

This method is based on the integration of direction likelihood distributions obtained by source direction estimation. In most cases, direction estimation methods calculate the likelihood P ( ϕ , θ ) for each azimuth ϕ and elevation θ , and assume the direction with the largest P ( ϕ , θ ) is the sound source direction. The approach of this method is to estimate the sound source location by converting the direction likelihood P ( ϕ , θ ) of each microphone array into the likelihood of a three-dimensional location. Figure 1 illustrates the procedure of the proposed tracking method. This method estimates the sound source trajectory through particle filtering, with each particle being given the location likelihood as its weight.

3.3. Estimation of Direction Likelihood Distribution

Some indicators can be regarded as likelihoods for sound source direction. The cross-power spectrum phase (CSP) coefficients are used in the CSP method [24], which is a method for estimating the time difference of arrival (TDOA) with two microphones, and the spatial spectrum obtained from the delay-and-sum beamformer includes scalar quantities whose parameters are the sound source direction, and they generally have a peak in the direction of the source [20]. In this paper, we use a MUSIC spectrum, which produces sharp peaks in the direction of the sound source as the direction likelihood. The MUSIC (multiple signal classification) method analyses the eigenspace of the spatial correlation matrix and estimates the source azimuth and elevation by using the orthogonality between the subspaces of the target sound source and the noise [25]. Let a ( ω , ϕ , θ ) C M be the transfer function of a sound signal of a frequency component ω from a direction ( ϕ , θ ) , then the spatial spectrum, which is known as the MUSIC spectrum, can be expressed as
P ( ϕ , θ ) = 1 ω H ω L + 1 ω = ω L ω H a ( ω , ϕ , θ ) H a ( ω , ϕ , θ ) a ( ω , ϕ , θ ) H E N ( ω ) E N ( ω ) H a ( ω , ϕ , θ )
where E N is a matrix consisting of eigenvectors of the noise subspace of the spatial correlation matrix. In general, the sound source direction is considered to be the direction where the MUSIC spectrum has peaks when estimating the direction. In this paper we regard the MUSIC spectrum as a direction likelihood of the sound source and we integrate the MUSIC spectra obtained from each microphone array and convert them into a location likelihood of the sound source.

3.4. Converting to Location Likelihood Distribution

Let P n ( ϕ , θ ) be the MUSIC spectrum calculated from MA n . We express the likelihood distribution of the sound location by simply summing P n ( ϕ , θ ) . Let an arbitrary three-dimensional location be x, and the direction from point x to MA n be ( ϕ n , θ n ) . Then, the location likelihood L at a location x is described as follows.
L ( x ) = n P n ( ϕ ˜ n round , θ ˜ n round )
ϕ ˜ n round = round ( ϕ ˜ n ) , θ ˜ n round = round ( θ ˜ n )
cos ϕ ˜ n cos ϕ ˜ n sin ϕ ˜ n cos ϕ ˜ n sin ϕ ˜ n = R n 1 cos ϕ n cos ϕ n sin ϕ n cos ϕ n sin ϕ n
where round ( · ) is a function that rounds the direction according to the resolution of transfer function a ( ω , ϕ , θ ) and R n is the rotation matrix representing the posture of MA n , which can be defined by ( ϕ n , k , θ n , k , ψ n , k ) . In other words, location likelihood L ( x ) is the summation of direction likelihoods corresponding to directions towards x seen from each microphone array.

3.5. Tracking Based on Location Likelihood Distribution

Since we can obtain the location likelihood of an arbitrary 3D location x, it is able to track the sound source location by applying this location likelihood L ( x ) to particle filtering. Let I be the number of particles and x k i , w k i be, respectively, the state and weight of the i-th particle at a time step k. The state x k i includes the location and velocity of particle i.
x k i = x k i , y k i , z k i , x ˙ k i , y ˙ k i , z ˙ k i T

3.6. Initialization of Particle Filter

Initial particles are sampled from the following distribution.
x 0 i N ( μ 0 , Σ 0 )
μ 0 = μ 0 , p o s , 0 , 0 , 0 T
Σ 0 = σ p o s 2 I O O σ v e l 2 I
where μ 0 , p o s is the mean of the initial distribution. μ 0 , p o s can be determined if there is initial information of the sound source, or it can be derived from triangulation based on direction estimation [9]. After calculating triangulation points, μ 0 , p o s is obtained by taking the average of all triangulation points.

3.7. Particle Update

We use an excitation-damping model for the prediction model [18].
x k i = F x k 1 i + H v
F = I T I O a I , H = O b I
v N ( 0 , I )
where I R 3 × 3 is an identity matrix and O R 3 × 3 is a zero matrix. Variables a and b determine the ratio of velocity to carry on from the previous time step and the excitation of particles, respectively. Each particle gains weight proportional to L ( x k , p o s i ) , hence
w k i = w k 1 i L ( x k , p o s i ) i L ( x k , p o s i )
where x k , p o s i = [ x k i , y k i , z k i ] T . Resampling would be necessary if effective particles are less than a threshold N t h r . Hence, when
1 i ( w k i ) 2 N t h r
is satisfied, the weight of each particle should reset to 1 / I .

3.8. Location Estimation

Finally, the estimated sound source location at time step k is obtained by taking the weighted average of the particles.
x ^ e , k = i = 1 w k i x i k

4. Evaluation

In this section, we perform a numerical simulation to evaluate the tracking performance of several tracking methods. When performing source tracking, various challenges are taken into account such as source distance, SNR, presence of obstacles, microphone array placement, and self-positioning of the drone. However, it is difficult to evaluate all the factors at once because it is difficult to identify which factor is responsible for the performance change. Hence, factors to be evaluated should be narrowed. In outdoor environments the range of the sound source that can be heard can be very wide. Hence, how far the sound source tracking can succeed is a big interest. Besides, since loud drone noise will be input to the microphones, robustness against low SNR is also a big concern. Therefore, in this evaluation, we focus only on the two important factors: source distance and SNR.

4.1. Evaluation Outline

We performed a numerical simulation via MATLAB to evaluate the performance of PAFIM and existing methods in sound source tracking. We considered a scenario of sound source tracking with two drones surrounding a single sound source (see Figure 2). Each drone has two microphone arrays equipped as shown in Figure 3 and each microphone array consists of 16 microphones placed spherically (See Figure 4). Microphones record acoustic signals at 16 kHz, 24 bits, and the transfer function used to calculate the MUSIC spectrum has a resolution of 5 degrees. The simulated recording is calculated by a definitive transfer function from the sound source, and noise during the sound transmission is omitted. Both drones are hovering still at a 30 m height, which means they do not move through the simulation.
In this simulation, we focused on the concern in relation to tracking performance with respect to the change in source distance and the change in signal-to-noise ratio (SNR). Therefore, we evaluated the performance of tracking methods by moving the source distance and SNR independently, which means the SNR does not correlate with the source distance. We moved the horizontal distance between the sound source and drones from 10 m to 80 m in increments of 10 m, and we varied the SNR from −60 dB to 10 dB. The range of distance was decided as above since a simulated rescue demonstration has been performed against a sound source 20 to 30 m away [6], and we wanted to evaluate performance limits in terms of distance. The range of SNR was decided by reports of indoor experiments showing that the SNR could be −15 to −25 dB [8,26]. In addition, outdoor environments could make SNR lower, since drones for outdoors tend to make louder noise and wind noise input to the microphones. Drone noise is added to the input by adding prerecorded drone noise to the recorded signal, and the SNR is set by adjusting the amplitude of the drone noise. Drone noise is assumed to be emitted from the red dots illustrated in Figure 3. As the horizontal distance increases, the tracking result is expected to worsen due to the direction estimation’s discreteness. As the SNR decreases, the tracking result is expected to worsen since the direction estimation will be affected by low SNR. In order to evaluate the tracking performance for various source motions, simulations were performed for three types of motion: stationary, circular, and random walk. Stationary motion does not permit the sound source to move at the origin. Circular motion lets the sound source move in a circle with a radius of 5 m and the circle center is set to the origin. While the sound source is moving in a circle, the speed of the sound source is constant, and it takes 10 s to go around the circle. Random walk motion lets the sound source move randomly with the following behavior starting from the origin.
x e , k + 1 = x e , k + v k + 1 cos ϕ k + 1 sin ϕ k + 1 0
v k + 1 = v k + T Δ v , Δ v 0.5 N ( 0 , 1 )
ϕ k + 1 = ϕ k + T Δ ϕ , Δ ϕ 1000 × π 180 × N ( 0 , 1 )
For the sound source, we prepared thirty sound clips (ten male voices, ten female voices, and ten white noise clips). Human voice clips were made from a corpus that reads Japanese news [27]. Each tracking method estimated the sound source location every T = 0.2 s during 10 s. For all tracking methods, we utilized the MUSIC method for direction estimation (or calculation of location likelihood) in this simulation, although some methods use different direction estimation methods. In the MUSIC method, we used a transfer function with a resolution of 5 degrees, and the range of frequency was ω L = 200 Hz to ω H = 2200 Hz. Results obtained in this simulation are described in the next subsection.

4.2. Simulation Results and Discussion

We mainly evaluate tracking performance by looking at the tracking error. We define the tracking error as the Euclidean distance between the estimated location and the ground truth. We also use root-mean-square error (RMSE), which is defined as below, to evaluate the error of the entire trajectory, where K is the number of time steps in the simulation.
RMSE = 1 K k = 1 K error k 2
Table 1 summarizes the figures showing the tracking errors. Each figure shows the simulated RMSE of each method. Basically, each figure shows the RMSE for each distance and SNR, although for the sake of clarity, figures for a specific distance (=30 m) and for a specific SNR (=−20 dB) are also shown. Given these results, we discuss the tracking performance of the three methods. Through the simulations, significant results were obtained in terms of source distance, SNR, and tracking methods.
Table 1. Table of figures showing RMSE of simulation results for each source movement and sound type.
Table 1. Table of figures showing RMSE of simulation results for each source movement and sound type.
FemaleMaleWhite Noise
StationaryFigure 5OverallFigure 6Figure 7
Fixed the distance at 30 mFigure 8
Fixed the SNR at −20 dBFigure 9
CircleFigure 10OverallFigure 11Figure 12
Fixed the distance at 30 mFigure 13
Fixed the SNR at −20 dBFigure 14
Random walkFigure 15OverallFigure 16Figure 17
Fixed the distance at 30 mFigure 18
Fixed the SNR at −20 dBFigure 19
Figure 5. RMSE of tracking a stationary sound source emitting female voice.
Figure 5. RMSE of tracking a stationary sound source emitting female voice.
Ijerph 18 09039 g005
Figure 6. RMSE of tracking a stationary sound source emitting male voice.
Figure 6. RMSE of tracking a stationary sound source emitting male voice.
Ijerph 18 09039 g006
Figure 7. RMSE of tracking a stationary sound source emitting white noise.
Figure 7. RMSE of tracking a stationary sound source emitting white noise.
Ijerph 18 09039 g007
Figure 8. RMSE of tracking a stationary sound source emitting male voice (focusing on source horizontal distance = 30 m).
Figure 8. RMSE of tracking a stationary sound source emitting male voice (focusing on source horizontal distance = 30 m).
Ijerph 18 09039 g008
Figure 9. RMSE of tracking a stationary sound source emitting male voice (focusing on SNR = −20 dB).
Figure 9. RMSE of tracking a stationary sound source emitting male voice (focusing on SNR = −20 dB).
Ijerph 18 09039 g009
Figure 10. RMSE of tracking a sound source moving in a circle emitting female voice.
Figure 10. RMSE of tracking a sound source moving in a circle emitting female voice.
Ijerph 18 09039 g010
Figure 11. RMSE of tracking a sound source moving in a circle emitting male voice.
Figure 11. RMSE of tracking a sound source moving in a circle emitting male voice.
Ijerph 18 09039 g011
Figure 12. RMSE of tracking a sound source moving in a circle emitting white noise.
Figure 12. RMSE of tracking a sound source moving in a circle emitting white noise.
Ijerph 18 09039 g012
Figure 13. RMSE of tracking a sound source moving in a circle emitting male voice (focusing on source horizontal distance = 30 m).
Figure 13. RMSE of tracking a sound source moving in a circle emitting male voice (focusing on source horizontal distance = 30 m).
Ijerph 18 09039 g013
Figure 14. RMSE of tracking a sound source moving in a circle emitting male voice (focusing on SNR = −20 dB).
Figure 14. RMSE of tracking a sound source moving in a circle emitting male voice (focusing on SNR = −20 dB).
Ijerph 18 09039 g014
Figure 15. RMSE of tracking a randomly moving sound source emitting female voice.
Figure 15. RMSE of tracking a randomly moving sound source emitting female voice.
Ijerph 18 09039 g015
Figure 16. RMSE of tracking a randomly moving sound source emitting male voice.
Figure 16. RMSE of tracking a randomly moving sound source emitting male voice.
Ijerph 18 09039 g016
Figure 17. RMSE of tracking a randomly moving sound source emitting white noise.
Figure 17. RMSE of tracking a randomly moving sound source emitting white noise.
Ijerph 18 09039 g017
Figure 18. RMSE of tracking a randomly moving sound source emitting male voice (focusing on source horizontal distance = 30 m). Tracking results of Potamitis’s method did not converged when the SNR is less than −40 dB.
Figure 18. RMSE of tracking a randomly moving sound source emitting male voice (focusing on source horizontal distance = 30 m). Tracking results of Potamitis’s method did not converged when the SNR is less than −40 dB.
Ijerph 18 09039 g018
Figure 19. RMSE of tracking a randomly moving sound source emitting male voice (focusing on SNR = −20 dB).
Figure 19. RMSE of tracking a randomly moving sound source emitting male voice (focusing on SNR = −20 dB).
Ijerph 18 09039 g019

4.2.1. Discussion: Distance

For any source movement, the tracking error increases as the source distance increases. We believe that the main cause of this phenomenon is the resolution of the transfer function a ( ϕ , θ ) , which is used in every method. As in the simulation, if the resolution is five degrees, the estimation error at a point l m away will be π l / 36 m if the source direction estimation is off by one element. For example, if the sound source is 50 m away, the source location estimation error will be about 4.4 m. This tendency can be seen in the RMSE of PAFIM from Figure 9 and Figure 14.

4.2.2. Discussion: SNR

Similar to distance, tracking error increases as the SNR decreases. Besides, this increase is more noticeable than the increase due to distance. This is because the direction estimation error is largely affected by SNR rather than the source distance, which could be seen in Figure 20 and Figure 21. Both figures illustrate the direction estimation error of one microphone array by box plot. The direction estimation error significantly deteriorates when the SNR gets lower than −40 dB. This can be taken as a threshold to determine the level of effectiveness of the source direction estimation. When the SNR surpasses this threshold, the RMSE of each method rises sharply. In this simulation, we used the SEVD-MUSIC method, which is the simplest MUSIC method that does not use noise reduction algorithms. Therefore, SEVD-MUSIC has difficulty outputting a proper MUSIC spectrum when the SNR is lower than 0 dB. In this simulation, drone noise is set to be output from the red dots in Figure 3 while the target sound source emits from a single point. It seems that even if the SNR is lower than 0 dB since the drone noise emits from several points, the directionality has blurred compared to the target signal, which has lowered the threshold SNR.

4.2.3. Discussion: Overall Performance

In this section, we assess the overall performance by comparing the simulated methods. Since the nature of tracking error differs between the change in distance and SNR, the nature of the source tracking error for each tracking method also differs. When the source distance is large, the tracking errors of PAFIM [10] and method [18] do not get larger compared to method [23]. This is considered because the source location is represented in the form of a location distribution rather than a discrete representation such as triangulation points or source directions. The discreteness of triangulation points leads to large tracking error because even if the direction estimation error is minimum (=5 degrees) the location error of the triangulation points will be about 3 to 5 m and this gets worse as the source distance increases. However, PAFIM reduces the increase in tracking error with increasing distance by using all the microphone arrays to compute a location distribution. Comparing PAFIM and method [18], PAFIM has performed slightly better tracking against increasing distance since PAFIM utilizes the entire MUSIC spectrum (rather continuous information) while method [18] only uses the direction that takes the peak of the MUSIC spectrum (rather discrete information).
In terms of SNR, we see that PAFIM has a larger error than other methods when the SNR gets lower, and method [23], which uses triangulation, sometimes results in uncomputable error since direction estimation fails and triangulation cannot be performed. Two main reasons are considered as to why PAFIM fails in sound source tracking. Since PAFIM tries to reduce discreteness by estimating the source location distribution, a non-informative distribution will be obtained in a situation where direction estimation fails repeatedly, which means the likelihood distribution of the source direction is distorted. Regardless of the iterative estimation, if the source distribution is highly skewed, it will not converge to the correct distribution. However, methods with triangulation or direction estimation have a chance to capture the correct location or direction of the sound source at some time step, which can lead the tracking to the correct trajectory. This might be one of the reasons why method [18,23] has a smaller RMSE than PAFIM. Another reason for PAFIM’s vulnerability to SNR is that PAFIM does not have a reliable initialization method and has to rely on triangulation in the first step. As mentioned before, triangulation is not reliable when the SNR is too low, and it is difficult for particle filters to get to the correct trajectory when the initial state is too far from the ground truth. However, PAFIM only shows the worst results when the SNR is lower than the assumptions of the direction estimation method. If the direction estimation method can withstand the SNR, we see that PAFIM can outperform other methods from Figure 8, Figure 13 and Figure 18.
In short, numerical simulations show that PAFIM is strong against distant sound sources compared to existing methods since PAFIM tries to reduce the discreteness of direction estimation. Since drones generally perform their tasks in vast fields, this is a key strength for drone audition. When it comes to low SNR, PAFIM has the largest tracking error since it is hardly possible to obtain a feasible location distribution. However, SNR observed in real-world experiments is about −20 dB, for which PAFIM had the least tracking error. In summary, PAFIM was found to be more robust to increasing distance than the compared methods if the drone noise was not less than −20 dB.

5. Conclusions

In this paper, we evaluated the performance of sound source tracking among several tracking methods that can be used in drone audition. In the numerical simulation, we focused on the performance difference against source distance and SNR. Simulation results address that the nature of tracking error is different between changes in distance and SNR, and this difference brings to light the advantages and disadvantages of tracking methods. The proposed method (PAFIM), which integrates MUSIC spectra to capture the sound source location distribution, has been found to be robust when the target sound source is far from the drones. However, methods that utilize discrete information such as directions and triangulation points are sensitive to low SNR since there is a chance to obtain the correct direction even if the MUSIC spectrum is noisy, while PAFIM fails to capture the location distribution with low SNR. However, when the SNR is about −20 dB or more, PAFIM has still shown the lowest RMSE. In this paper, the source distance and SNR were changed independently, but generally, the SNR increases if the drone gets nearer to the sound source. Therefore, PAFIM can be improved if the drones can be maneuvered with proper action planning that can get close enough to target sound sources and cover unobserved areas. Based on this concept, assessment of how placement of microphone arrays will affect the tracking and action planning for drones is one of the issues to be addressed in the future. In addition, drone noise reduction methods were not applied in the simulations. Due to the fact that PAFIM does not involve conventional source direction estimation, the benefit of denoising methods to PAFIM is unclear. It is also a future task to evaluate how much the PAFIM is improved by noise removal by comparing it with other methods.

Author Contributions

Conceptualization, T.Y. and K.N. (Kazuhiro Nakadai); methodology, T.Y.; software, T.Y.; validation, T.Y.; formal analysis, T.Y.; investigation, T.Y.; resources, T.Y.; data curation, T.Y.; writing—original draft preparation, T.Y.; writing—review and editing, T.Y., K.I., K.N. (Kenji Nishida), and K.N. (Kazuhiro Nakadai); visualization, T.Y.; supervision, K.N. (Kazuhiro Nakadai); project administration, T.Y.; funding acquisition, K.I., K.N. (Kenji Nishida), and K.N. (Kazuhiro Nakadai). All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by JSPS KAKENHI Grant No. JP19K12017, JP19KK0260 and JP20H00475.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Restrictions apply to the availability of these data. Data was obtained from The Acoustical Society of Japan (2006): ASJ Japanese Newspaper Article Sentences Read Speech Corpus (JNAS) and are available at https://doi.org/10.32130/src.JNAS with the permission of The Acoustical Society of Japan.

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
SNRSignal-to-noise ratio
MUSICMultiple signal classification
PAFIMParticle filtering with integrated MUSIC
RMSERoot-mean-square error

References

  1. Hoshiba, K.; Washizaki, K.; Wakabayashi, M.; Ishiki, T.; Kumon, M.; Bando, Y.; Gabriel, D.; Nakadai, K.; Okuno, H. Design of UAV-embedded microphone array system for sound source localization in outdoor environments. Sensors 2017, 17, 2535. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  2. Nonami, K.; Hoshiba, K.; Nakadai, K.; Kumon, M.; Okuno, H.G.; Tanabe, Y.; Yonezawa, K.; Tokutake, H.; Suzuki, S.; Yamaguchi, K.; et al. Recent R&D technologies and future prospective of flying robot in tough robotics challenge. In Disaster Robotics; Springer: Berlin/Heidelberg, Germany, 2019; pp. 77–142. [Google Scholar]
  3. Nakadai, K.; Kumon, M.; Okuno, H.G.; Hoshiba, K.; Wakabayashi, M.; Washizaki, K.; Ishiki, T.; Gabriel, D.; Bando, Y.; Morito, T.; et al. Development of microphone-array-embedded UAV for search and rescue task. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 5985–5990. [Google Scholar]
  4. Nakajima, H.; Nakadai, K.; Hasegawa, Y.; Tsujino, H. Adaptive step-size parameter control for real-world blind source separation. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Las Vegas, NV, USA, 31 March–4 April 2008; pp. 149–152. [Google Scholar]
  5. Sugiyama, O.; Uemura, S.; Nagamine, A.; Kojima, R.; Nakamura, K.; Nakadai, K. Outdoor acoustic event identification with DNN using a quadrotor-embedded microphone array. J. Robot. Mechatronics 2017, 29, 188–197. [Google Scholar] [CrossRef]
  6. Wakabayashi, M.; Washizaka, K.; Hoshiba, K.; Nakadai, K.; Okuno, H.G.; Kumon, M. Design and Implementation of Real-Time Visualization of Sound Source Positions by Drone Audition. In Proceedings of the IEEE/SICE International Symposium on System Integration (SII), Honolulu, HI, USA, 12–15 January 2020; pp. 814–819. [Google Scholar]
  7. Furukawa, K.; Okutani, K.; Nagira, K.; Otsuka, T.; Itoyama, K.; Nakadai, K.; Okuno, H.G. Noise correlation matrix estimation for improving sound source localization by multirotor UAV. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 3943–3948. [Google Scholar]
  8. Strauss, M.; Mordel, P.; Miguet, V.; Deleforge, A. DREGON: Dataset and methods for UAV-embedded sound source localization. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 1–8. [Google Scholar]
  9. Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Sound Source Tracking by Drones with Microphone Arrays (forthcoming). In Proceedings of the IEEE/SICE International Symposium on System Integration, Honolulu, HI, USA, 12–15 January 2020. [Google Scholar]
  10. Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Sound Source Tracking Using Integrated Direction Likelihood for Drones with Microphone Arrays. In Proceedings of the IEEE/SICE International Symposium on System Integration (SII), Fukushima, Japan, 11–14 January 2021; pp. 394–399. [Google Scholar]
  11. Nakadai, K.; Nakajima, H.; Yamada, K.; Hasegawa, Y.; Nakamura, T.; Tsujino, H. Sound source tracking with directivity pattern estimation using a 64 ch microphone array. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems, Edmonton, AB, Canada, 2–6 August 2005; pp. 1690–1696. [Google Scholar]
  12. Nakadai, K.; Nakajima, H.; Murase, M.; Kaijiri, S.; Yamada, K.; Nakamura, T.; Hasegawa, Y.; Okuno, H.G.; Tsujino, H. Robust tracking of multiple sound sources by spatial integration of room and robot microphone arrays. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 4. [Google Scholar]
  13. Portello, A.; Bustamante, G.; Danès, P.; Piat, J.; Manhes, J. Active localization of an intermittent sound source from a moving binaural sensor. In Proceedings of the Forum Acuticum, Krakow, Poland, 7–12 September 2014. [Google Scholar]
  14. Nguyen, Q.V.; Colas, F.; Vincent, E.; Charpillet, F. Localizing an intermittent and moving sound source using a mobile robot. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, 9–14 October 2016; pp. 1986–1991. [Google Scholar]
  15. Evers, C.; Dorfan, Y.; Gannot, S.; Naylor, P.A. Source tracking using moving microphone arrays for robot audition. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 6145–6149. [Google Scholar]
  16. Evers, C.; Habets, E.A.P.; Gannot, S.; Naylor, P.A. DoA Reliability for Distributed Acoustic Tracking. IEEE Signal Process. Lett. 2018, 25, 1320–1324. [Google Scholar] [CrossRef] [Green Version]
  17. Sasaki, Y.; Tanabe, R.; Takemura, H. Probabilistic 3D sound source mapping using moving microphone array. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Korea, 9–14 October 2016; pp. 1293–1298. [Google Scholar]
  18. Lauzon, J.; Grondin, F.; Létourneau, D.; Desbiens, A.L.; Michaud, F. Localization of RW-UAVs using particle filtering over distributed microphone arrays. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 2479–2484. [Google Scholar]
  19. Brandstein, M.S.; Silverman, H.F. A practical methodology for speech source localization with microphone arrays. Comput. Speech Lang. 1997, 11, 91–126. [Google Scholar] [CrossRef] [Green Version]
  20. Valin, J.M.; Michaud, F.; Rouat, J. Robust 3D Localization and Tracking of Sound Sources Using Beamforming and Particle Filtering. In Proceedings of the IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 4. [Google Scholar]
  21. Ishi, C.T.; Even, J.; Hagita, N. Speech activity detection and face orientation estimation using multiple microphone arrays and human position information. In Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 5574–5579. [Google Scholar]
  22. Gabriel, D.; Kojima, R.; Hoshiba, K.; Itoyama, K.; Nishida, K.; Nakadai, K. 2D sound source position estimation using microphone arrays and its application to a VR-based bird song analysis system. Adv. Robot. 2019, 33, 403–414. [Google Scholar] [CrossRef]
  23. Potamitis, I.; Chen, H.; Tremoulis, G. Tracking of multiple moving speakers with multiple microphone arrays. IEEE Trans. Speech Audio Process. 2004, 12, 520–529. [Google Scholar] [CrossRef]
  24. Knapp, C.; Carter, G. The generalized correlation method for estimation of time delay. IEEE Trans. Acoust. Speech Signal Process. 1976, 24, 320–327. [Google Scholar] [CrossRef] [Green Version]
  25. Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef] [Green Version]
  26. Wang, L.; Cavallaro, A. Acoustic Sensing From a Multi-Rotor Drone. IEEE Sens. J. 2018, 18, 4570–4582. [Google Scholar] [CrossRef]
  27. Itou, K.; Yamamoto, M.; Takeda, K.; Takezawa, T.; Matsuoka, T.; Kobayashi, T.; Shikano, K.; Itahashi, S. JNAS: Japanese speech corpus for large vocabulary continuous speech recognition research. J. Acoust. Soc. Jpn. (E) 1999, 20, 199–206. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Procedure of proposed method (for one time step).
Figure 1. Procedure of proposed method (for one time step).
Ijerph 18 09039 g001
Figure 2. Top view of simulation scenario (horizontal distance = 30 m, sound source movement = circle).
Figure 2. Top view of simulation scenario (horizontal distance = 30 m, sound source movement = circle).
Ijerph 18 09039 g002
Figure 3. Drone noise source (red dots) in the simulation.
Figure 3. Drone noise source (red dots) in the simulation.
Ijerph 18 09039 g003
Figure 4. Microphone placement in a microphone array.
Figure 4. Microphone placement in a microphone array.
Ijerph 18 09039 g004
Figure 20. Box plot of the direction estimation error for the lower right microphone array in Figure 2. Considered scenario is tracking a circular motion sound source emitting a male voice.
Figure 20. Box plot of the direction estimation error for the lower right microphone array in Figure 2. Considered scenario is tracking a circular motion sound source emitting a male voice.
Ijerph 18 09039 g020
Figure 21. Direction estimation error shown in Figure 20 focusing on distance = 30 m.
Figure 21. Direction estimation error shown in Figure 20 focusing on distance = 30 m.
Ijerph 18 09039 g021
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Assessment of Sound Source Tracking Using Multiple Drones Equipped with Multiple Microphone Arrays. Int. J. Environ. Res. Public Health 2021, 18, 9039. https://doi.org/10.3390/ijerph18179039

AMA Style

Yamada T, Itoyama K, Nishida K, Nakadai K. Assessment of Sound Source Tracking Using Multiple Drones Equipped with Multiple Microphone Arrays. International Journal of Environmental Research and Public Health. 2021; 18(17):9039. https://doi.org/10.3390/ijerph18179039

Chicago/Turabian Style

Yamada, Taiki, Katsutoshi Itoyama, Kenji Nishida, and Kazuhiro Nakadai. 2021. "Assessment of Sound Source Tracking Using Multiple Drones Equipped with Multiple Microphone Arrays" International Journal of Environmental Research and Public Health 18, no. 17: 9039. https://doi.org/10.3390/ijerph18179039

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop