Next Article in Journal
Numerical Investigation of the Seabed Dynamic Response to a Perforated Semi-Circular Breakwater
Previous Article in Journal
Ship Bridge OOW Activity Status Detection Using Wi-Fi Beamforming Feedback Information
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Angle of Arrival Estimator Utilizing the Minimum Number of Omnidirectional Microphones

System Engineering Department, Sejong University, Seoul 05006, Republic of Korea
J. Mar. Sci. Eng. 2024, 12(6), 874; https://doi.org/10.3390/jmse12060874
Submission received: 17 April 2024 / Revised: 10 May 2024 / Accepted: 24 May 2024 / Published: 24 May 2024

Abstract

:
In sound signal processing, angle of arrival indicates the direction from which a propagating sound signal arrives at a point where multiple omnidirectional microphones are positioned. Considering a small underwater platform (e.g., underwater unmanned vehicle), this article addresses how to estimate a non-cooperative target’s signal direction utilizing the minimum number of omnidirectional microphones. It is desirable to use the minimum number of microphones, since one can reduce the cost and size of the platform by using small number of omnidirectional microphones. Suppose that each microphone measures a real-valued sound signal whose speed and frequency information are not known in advance. Since two microphones cannot determine a unique AOA solution, this study presents how to estimate the angle of arrival using a general configuration composed of three omnidirectional microphones. The effectiveness of the proposed angle of arrival estimator utilizing only three microphones is demonstrated by comparing it with the state-of-the-art estimation algorithm through computer simulations.

1. Introduction

In underwater environments, electromagnetic signal is easily dissipated; thus, sound is mainly used for underwater target localization. In sound signal processing, angle of arrival (AOA) indicates the direction from which a propagating sound signal arrives at a point where multiple omnidirectional microphones are positioned. Microphones are used to measure a non-cooperative target’s signal in a passive manner. It is argued that AOA is desirable, since it does not require active ping generation. Thus, AOA is power-efficient and the non-cooperative target cannot detect the presence of passive microphones.
In underwater environments, sound speed can change according to various environmental effects (e.g., water temperature and salinity) [1]. Moreover, as microphones measure the signal of a non-cooperative target, the target signal frequency may not be known to a microphone.
Thus, this paper considers the case where each omnidirectional microphone measures a real-valued sound signal whose speed and frequency information are not known in advance. The problem is to find the target signal’s direction relative to the array position.
Considering a small underwater platform (e.g., underwater unmanned vehicle), this study addresses how to estimate the signal direction utilizing the minimum number of microphones. By integrating the AOA measurements of the target’s sound signal, one or more underwater platforms can estimate the target position without being detected by the target [2,3,4,5,6].
It is desirable to use the minimum number of microphones, since one can reduce the cost and size of the platform by using a small number of microphones. Every omnidirectional microphone collects signals uniformly in all directions. Every microphone is connected to every other microphone. Then, readings of all microphones are processed for AOA estimation.
There are many papers on AOA estimation based on signal measurements at multiple microphones. Various AOA estimators (Minimum Variance Distortionless Response (MVDR) beamformer [7,8,9], Estimation of Signal Parameters via Rotational Invariance Technique (ESPRIT) [10], or MUltiple SIgnal Classification (MUSIC) [11,12,13]) utilize phase measurements at each microphone for estimating the signal direction.
In the field of array signal processing, the MUSIC estimator is a classical spectrum estimation algorithm. MUSIC [11,12,13] estimates the autocorrelation matrix utilizing an eigenspace method. MUSIC [11,12,13] is based on the idea that the signal subspace is orthogonal to the noise subspace. The reference [14] addressed an adaptive beamformer to achieve high performance in the case of low input signal-to-noise ratio (SNR). The reference [15] integrated a MUSIC-based AOA estimation method that applies to both full arrays and sparse arrays. In reference [16], the authors analyzed the MUSIC estimator in the uniform linear array (ULA) and studied various factors which affect the estimation performance. The authors of [17] modified the classic MUSIC to be competent in unknown non-uniform noisy environments. For estimating the AOA of coherent signals in the ULA, MUSIC was modified by reconstructing a noise subspace [18]. In [19], a DOA estimation method based on spatial difference and a modified projection subspace algorithm was proposed for handling serious misalignment in the AOA estimation of multi-path signals under the background of impulse noise. The reference [20] introduced a modified MUSIC estimator to compute the AOA of multiple radio frequency signals, considering an antenna array with an imperfectly calibrated array response. In [21], the authors studied how to optimize parameters of the MUSIC estimator in order to improve the estimation performance.
To the best of our knowledge, every AOA estimator in the literature used more than three microphones. However, it is desirable to reduce the number of microphones, as we consider the cost and size of the platform.
Since two microphones cannot determine a unique AOA solution, this article presents how to estimate the AOA using a general configuration composed of only three microphones. The outperformance of the proposed AOA estimator utilizing only three microphones is demonstrated by comparing it with the MUSIC estimator through computer simulations. One proves that the proposed AOA estimator outperforms the MUSIC estimator considering both estimation accuracy and computation time.
This study is organized as follows. Section 2 introduces the definitions and assumptions. Section 3 introduces the proposed AOA estimator using only three microphones. MATLAB simulations are addressed in Section 4. Section 5 provides the discussion. Conclusions are addressed in Section 6.

2. Definitions and Assumptions

Let s ( ) = sin ( ) and c ( ) = cos ( ) for notation simplicity. Let a t a n 2 ( y , x ) denote the angle (phase) of x + j y . a t a n 2 ( y , x ) exists in the interval [ π , π ]. Considering a list, L , m a x ( L ) denotes an element with the maximum value in L . Also, m i n ( L ) denotes an element with the minimum value in L . For instance, m i n ( [ 1 , 2 , 3 ] ) = 1 and m a x ( [ 1 , 2 , 3 ] ) = 3 .
Let S i ( i { 1 , 2 , 3 } ) define the 2D coordinates of the i-th microphone. The microphone configuration with three microphones is depicted in Figure 1. In Figure 1, r defines the distance from every microphone to the origin of the frame. For avoiding the phase wrapping case, this paper assumes that r is less than half of the wavelength. In MATLAB simulations, we set r as 0.4 of the wavelength.
This study considers a target which is sufficiently far from the microphones. Let u define the unit vector from the origin to the target. Let ϕ define the azimuth angle of the target signal, such that π < ϕ π . Here, ϕ is measured from the x-axis of the microphone configuration. See Figure 1. We have
u = ( c ( ϕ ) , s ( ϕ ) ) T .
Our problem is to estimate the AOA ϕ based on signal measurements at three microphones. Recall that every microphone is synchronized to every other microphone.
Let P i define the projection of S i onto u , as depicted in Figure 1. We have
P i = S i · u .
Here, (·) operator indicates the inner product operation, defined as
S i · u = S i u cos ( a )
where a is the angle formed by two vectors u and S i . Also, u = 1 , since u is a unit vector.
Fourier analysis converts a signal from its original time domain to a representation in the frequency domain. DC offset is first removed from the time domain signal at S i . By applying Fast Fourier Transform (FFT) to the time domain signal, one computes the signal amplitude at different frequencies in the frequency domain. Then, in the frequency domain, one finds a frequency where the signal amplitude is maximized. Let f i define the frequency where the signal amplitude is maximized.
Let a n g l e ( f i ) denote the angle (phase) at f i ( i { 1 , 2 , 3 } ). Let D i ( i { 1 , 2 , 3 } ) define the signal delay measurement using the i-th microphone. Using a n g l e ( f i ) , we can compute the signal delay D i as
D i = a n g l e ( f i ) 2 π f s .
Here, f s denotes the sampling frequency.
Let C denote the signal speed. Then, we have
D i = S i · u C .
This implies that in the case where S i · u > 0 , D i > 0 . In addition, in the case where S i · u < 0 , D i < 0 . For instance, in Figure 1, S 3 · u < 0 . Thus, D 3 < 0 . On the other hand, if S 1 · u > 0 , then D 1 > 0 .

3. AOA Estimator Using Three Microphones

Before presenting the proposed AOA estimator using three microphones, we show that two microphones cannot determine a unique AOA solution. Suppose that we have only two microphones, S 1 and S 2 , as plotted in Figure 2. Then, we draw a straight infinite line connecting these two microphones. In Figure 2, dotted arrows indicate the signal direction at each microphone. Utilizing the phase differences at the two microphones, we cannot determine whether the target exists to the left or to the right of this line. Therefore, we require at least three microphones for determining a unique AOA solution.
Using (3), we obtain
D 2 D 1 = S 2 · u C S 1 · u C .
Using (1) and Figure 1, we further obtain
D 2 D 1 r = c ( 2 π 3 ) c ( ϕ ) + s ( 2 π 3 ) s ( ϕ ) C c ( ϕ ) C .
Using (3), we have
D 2 D 3 = S 2 · u C S 3 · u C .
Using (1) and Figure 1, we further obtain
D 2 D 3 r = c ( 2 π 3 ) c ( ϕ ) + s ( 2 π 3 ) s ( ϕ ) C c ( 2 π 3 ) c ( ϕ ) s ( 2 π 3 ) s ( ϕ ) C .
(7) and (9) lead to
C r ( D 2 D 1 , D 2 D 3 ) T = M ( c ( ϕ ) , s ( ϕ ) ) T ,
where
M = c ( 2 π 3 ) 1 s ( 2 π 3 ) 0 2 s ( 2 π 3 ) .
Then, (10) leads to
r C ( c ( ϕ ) , s ( ϕ ) ) T = M 1 ( D 2 D 1 , D 2 D 3 ) T .
Using the fact that ( c ( ϕ ) , s ( ϕ ) ) = 1 , we estimate ( c ( ϕ ) , s ( ϕ ) ) as
( c ( ϕ ) , s ( ϕ ) ) T = M 1 ( D 2 D 1 , D 2 D 3 ) T M 1 ( D 2 D 1 , D 2 D 3 ) T .
We then estimate ϕ as
ϕ = a t a n 2 ( s ( ϕ ) , c ( ϕ ) ) .
Note that we do not have to access either r or C, since ( c ( ϕ ) , s ( ϕ ) ) T is derived using (13). Recall that using the phase at f i ( i { 1 , 2 , 3 } ), we can compute the signal delay D i .
It is acknowledged that (13) is singular when D 1 = D 2 = D 3 . In this case, the signal delay of every microphone is identical to that of any other microphone. In this case, one cannot estimate the bearing angle ϕ under (14). Next, we present the requirement to enable a microphone configuration with three microphones to estimate ϕ .

3.1. AOA Estimator with a General Configuration Composed of Three Microphones

In this subsection, we consider an AOA estimator which has a general microphone configuration with three microphones, which is distinct from Figure 1. This paper considers a general configuration composed of three microphones. A general microphone configuration considered in this subsection is depicted in Figure 3. In this figure, A j denotes the angle of the j-th microphone measured in the counter-clockwise direction starting from the x-axis of the microphone configuration. Moreover, r j denotes the distance between the center and the j-th microphone.
Considering a general microphone configuration with three microphones, r j r k is feasible, where j k . To avoid the phase wrapping case, this paper assumes that m a x ( [ r 1 , r 2 , r 3 ] ) is less than half of the wavelength.
Figure 4 shows a singular microphone configuration where D 1 = D 2 = D 3 . In this configuration, P 1 = P 2 = P 3 . Thus, one cannot determine whether the signal direction is u or u . Therefore, a microphone configuration with three microphones must satisfy
( D 1 D 2 ) 2 + ( D 2 D 3 ) 2 0 .
Suppose that (15) is satisfied for a microphone configuration with three microphones. Using (3), we obtain
D 2 D 1 = S 2 · u C S 1 · u C .
Using (1) and Figure 3, we further obtain
D 2 D 1 = r 2 ( c ( A 2 ) c ( ϕ ) + s ( A 2 ) s ( ϕ ) ) C r 1 c ( ϕ ) C .
Using (3), we have
D 2 D 3 = S 2 · u C S 3 · u C .
Using (1) and Figure 3, we further obtain
D 2 D 3 = r 2 ( c ( A 2 ) c ( ϕ ) + s ( A 2 ) s ( ϕ ) ) C r 3 ( c ( A 3 ) c ( ϕ ) + s ( A 3 ) s ( ϕ ) ) C .
(17) and (19) lead to
C ( D 2 D 1 , D 2 D 3 ) T = G ( c ( ϕ ) , s ( ϕ ) ) T ,
where
G = r 2 c ( A 2 ) r 1 r 2 s ( A 2 ) r 2 c ( A 2 ) r 3 c ( A 3 ) r 2 s ( A 2 ) r 3 s ( A 3 ) .
Then, (20) leads to
1 C ( c ( ϕ ) , s ( ϕ ) ) T = G 1 ( D 2 D 1 , D 2 D 3 ) T .
Using the fact that ( c ( ϕ ) , s ( ϕ ) ) = 1 , we estimate ( c ( ϕ ) , s ( ϕ ) ) as
( c ( ϕ ) , s ( ϕ ) ) T = G 1 ( D 2 D 1 , D 2 D 3 ) T G 1 ( D 2 D 1 , D 2 D 3 ) T .
In order to satisfy the existence of G 1 , G must be invertible. For making G invertible, the condition number of G must be as close to one as possible. As the condition number of G increases, G becomes ill-conditioned. Hence, the condition number of G in (21) is applied as the observability index of the AOA estimator.
Once the condition number of G in (21) is less than a certain threshold, T h r e s , and (15) is met, we then estimate ϕ using the proposed AOA estimator (23). Note that we do not have to access either r or C, since ( c ( ϕ ) , s ( ϕ ) ) T is derived using (23). Recall that using the phase at f i ( i { 1 , 2 , 3 } ), we can compute the signal delay D i .

3.2. Computation Load Analysis

In the proposed AOA estimator, FFT is used to find a frequency where the signal amplitude is maximized. The computation load of FFT is O ( N × l o g N ) , where N denotes the data size. Reference (14) is used to estimate the target’s bearing angle. Once FFT is done, the computation of (14) does not depend on the data size, N. Thus, the computation load of the proposed AOA estimator is O ( N × l o g N ) . In the next section, MATLAB simulations show that the proposed AOA estimator outperforms MUSIC estimators considering both computation load and estimation accuracy.

4. MATLAB Simulations

MATLAB simulations (version: MATLAB R2022a) are applied to demonstrate the effectiveness of the proposed AOA estimator with only three microphones. The sampling frequency is f s = 500,000 Hz, and the signal length is set as only 50 samples. Once the condition number of G in (21) is less than T h r e s = 2 and (15) is met, then we estimate the AOA ϕ using the proposed AOA estimator (23).
Suppose that each microphone measures a real-valued sound signal whose frequency information or speed are not known in advance. The measured signal is a sinusoidal signal with frequency f = 10,000 Hz. Note that the signal frequency f is not known in advance, since one considers a non-cooperative target. The signal speed C is 1400 m/s, but C is not known in advance. In underwater environments, sound speed can change according to various environmental effects (e.g., water temperature and salinity) [1]. In the AOA estimation, a wrong signal speed, e.g., C w = 1200 m/s, is used.
Let s i g [ n ] ( n 1 , 2 , , 100 ) denote the real-valued signal sampled at sampling index n. As the real-valued signal, we use
s i g [ n ] = c ( 2 π f n t s ) + G ,
where t s = 1 / f s indicates the sampling period. Considering measurement noise, G indicates a Gaussian noise having zero mean and standard deviation σ G = 0.5 . This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.5 ) = 10 l o g ( 2 ) in dB.
Recall that the signal delay D i ( i { 1 , 2 , 3 } ) was addressed in (3). Then, the signal at each microphone S i ( i { 1 , 2 , 3 } ) is modeled using
s i g i [ n ] = c ( 2 π f n t s + 2 π f s D i ) + G .
Note that c ( 2 π f n t s + 2 π f s D i ) is the real part of a complex-valued exponential number e j 2 π f n t s × e j 2 π f s D i . Since D i = S i · u C under (3), we have
e j 2 π f s D i = e j 2 π f s C S i · u .
A vector with 3 rows, whose i-th row ( i { 1 , 2 , 3 } ) is e j 2 π f s D i in (26), is termed the steering vector in MUSIC estimators [11,12,13,15].
For robust verification of the proposed estimator using three microphones, we use 100 Monte Carlo (MC) simulations. Recall that ϕ t denotes the true bearing angle. Let ϕ ^ [ m ] denote an estimate of the bearing angle ϕ in the m-th MC simulation ( m { 1 , 2 , , 100 } ). Then, we define a v g E r r (in degrees) as
a v g E r r = 1 100 m = 1 100 ( ϕ ^ [ m ] ϕ ) .
We define s t d E r r (in degrees) as
s t d E r r = 1 100 m = 1 100 ( ϕ ^ [ m ] ϕ ) 2 .
The computation time required for all MC simulation is termed C o m p u t e T i m e in seconds.

4.1. AOA Estimation Using a Sensor Configuration as Plotted in Figure 1

As the first computer simulation scenario, we use the microphone configuration as plotted in Figure 1. In addition, r, the radius of microphone configuration, is 0.4 of the wavelength. Here, the wavelength, λ , is given as λ = C f . See that r is less than half of the wavelength, for removing aliasing in the AOA estimator.
Recall that the condition number of G in (21) is applied as the observability index of the AOA estimator. For making G invertible, the condition number of G must be as close to one as possible. The condition number of G is 1.73, as we use the microphone configuration in Figure 1. Observe that this condition number is less than T h r e s = 2 .
Figure 5 shows the signal strength at every microphone. The target’s true bearing angle ϕ is −180 degrees. Under the proposed AOA estimator, we get a v g E r r = 0.2 degrees. In addition, we get s t d E r r = 1.3 degrees. C o m p u t e T i m e is 0.04 s. Observe that the proposed estimation is accurate and fast.

4.1.1. Change the Target’s Bearing Angle

For comparison with the proposed AOA estimator using only three microphones, this study uses the MUSIC estimator [11,12,13]. Considering the case where only three microphones are used, one shows that the proposed estimator outperforms the MUSIC estimator considering both estimation accuracy and computation time.
The MUSIC estimator [11,12,13] estimates the autocorrelation matrix utilizing an eigenspace method. In the MUSIC estimator, AOA search is used with step size (0.5 degree) in the range of (−180, −179.5, …, 179.5, 180) in degrees. One can decrease the step size to improve the estimation accuracy. However, decreasing the step size increases the AOA computation load; thus, there is a trade-off between decreasing the step size and the computation load.
We change the target’s bearing angle gradually and test the performance of the proposed AOA estimator ( P r o ) and the MUSIC estimator ( M U ). Table 1 depicts the simulation results. Considering measurement noise, (24) uses σ G = 0.1 . This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.1 ) = 10 in dB.
The unit for angle measurements in the table is degrees. C o m p u t e T i m e for all P r o MC simulations in Table 1 is 0.6 s, and C o m p u t e T i m e for all M U MC simulations in Table 1 is 2 s.
For all angles in Table 1, m i n ( P r o ( a v g E r r ) ) = 0.04 and m a x ( P r o ( a v g E r r ) ) = 0.04 . For all angles in Table 1, m i n ( M U ( a v g E r r ) ) = 134 and m a x ( M U ( a v g E r r ) ) = 136 . Table 1 shows that the proposed AOA estimator outperforms the MUSIC estimator considering both accuracy and computation time.

4.1.2. Change the Noise Level

We change the noise level σ G in (24) and test the performance of the proposed AOA estimator. We change the noise level σ G in (24) to 1. This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 1 ) = 0 in dB.
Table 2 depicts the simulation results, as we set SNR as zero. C o m p u t e T i m e for all P r o simulations in Table 2 is 0.6 s and C o m p u t e T i m e for all M U simulations in Table 2 is 2 s.
For all angles in Table 2, m i n ( P r o ( a v g E r r ) ) = 0.5 and m a x ( P r o ( a v g E r r ) ) = 1.8 . For all angles in Table 2, m i n ( M U ( a v g E r r ) ) = 128 and m a x ( M U ( a v g E r r ) ) = 125 . Table 2 shows that the proposed AOA estimator outperforms the MUSIC estimator considering both accuracy and computation time.

4.2. AOA Estimation Using a General Configuration with Three Microphones

In this subsection, we consider a general configuration with three microphones. Consider the case where A 2 = π / 2 and A 3 = π / 2 in Section 3.1. Moreover, we set r 1 = 0.4 λ , r 2 = r 1 / 2 , and r 3 = r 1 / 3 . For avoiding the phase wrapping case, this paper assumes that m a x ( [ r 1 , r 2 , r 3 ] ) is less than half of the wavelength.
See Figure 6 for this general configuration. Recall that the condition number of G in (21) is applied as the observability index of the AOA estimator. For making G invertible, the condition number of G must be as close to one as possible. The condition number of G is 1.76, as we use the microphone configuration in Figure 6. Observe that this condition number is less than T h r e s = 2 .
The true signal speed C is 1400 m/s, but C is not known in advance. Thus, in each MC simulation, the proposed AOA estimator sets a random number in the interval [0, 1400] as a wrong signal speed C w .
In this subsection, we change the target’s bearing angle gradually and test the performance of the proposed AOA estimator ( G e n e r a l ) with this general microphone configuration. Considering the general microphone configuration, Table 3 depicts the simulation results. Considering measurement noise, (24) uses σ G = 0.1 . This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.1 ) = 10 in dB.
The unit for angle measurements in the table is degrees. C o m p u t e T i m e for all G e n e r a l MC simulations in Table 3 is 1 s. For all angles in Table 3, m i n ( G e n e r a l ( a v g E r r ) ) = 0.08 , and m a x ( G e n e r a l ( a v g E r r ) ) = 0.1 . Note that G e n e r a l outperforms M U in Table 1.
Considering the general microphone configuration with three microphones, Table 4 depicts the simulation results. Considering measurement noise, (24) uses σ G = 1 . This implies that the signal-to-noise ratio (SNR) is 0. The unit for angle measurements in the table is degrees. C o m p u t e T i m e for all G e n e r a l MC simulations in Table 4 is 1 s. For all angles in Table 4, m i n ( G e n e r a l ( a v g E r r ) ) = 0.9 and m a x ( G e n e r a l ( a v g E r r ) ) = 0.9 . Note that G e n e r a l outperforms M U in Table 2.

4.3. AOA Estimation Using Random Configurations with Three Microphones

In this subsection, we consider random configurations with three microphones. In each MC simulation, we set A 2 in Section 3.1 as a random number in the interval [ π , π ] . Also, in each MC simulation, we set A 3 in Section 3.1 as a random number in the interval [ π , π ] . We set r 1 = 0.4 λ . In each MC simulation, r 2 = r 1 r a n d and r 3 = r 1 r a n d , where r a n d returns a random number in the interval [ 0 , 1 ] . In this way, we generate a random microphone configuration at each MC simulation.
For making G in (21) invertible, the condition number of G must be as close to one as possible. Thus, in the case where the condition number of G is less than T h r e s = 2 and (15) is met, the associated MC simulation uses the randomly generated microphone configuration.
The true signal speed C is 1400 m/s, but C is not known in advance. Thus, in each MC simulation, the proposed AOA estimator sets a random number in the interval [0, 1400] as a wrong signal speed C w .
We change the target’s bearing angle gradually and test the performance of the proposed AOA estimator ( R a n d o m ) with a randomly generated microphone configuration. Considering a random microphone configuration, Table 5 depicts the simulation results. Considering measurement noise, (24) uses σ G = 0.1 . This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.1 ) = 10 in dB.
C o m p u t e T i m e for all R a n d o m MC simulations in Table 5 is 1 s. For all angles in Table 5, m i n ( R a n d o m ( a v g E r r ) ) = 0.1186 , and m a x ( R a n d o m ( a v g E r r ) ) = 0.0867 . Note that R a n d o m outperforms M U in Table 1.
Next, we check the effect of loosening the requirements for microphone configuration. For making G invertible, the condition number of G must be as close to one as possible. In the case where the condition number of G is less than T h r e s = 200 and (15) is met, the associated MC simulation uses the randomly generated microphone configuration. Using a large T h r e s implies that we use loose requirements for microphone configuration.
C o m p u t e T i m e for all R a n d o m MC simulations in Table 6 is 1 s. For all angles in Table 6, m i n ( R a n d o m ( a v g E r r ) ) = 1.2953 , and m a x ( R a n d o m ( a v g E r r ) ) = 0.8627 . Compared to Table 5, the estimation accuracy decreased in Table 6, since the requirements for microphone configuration are loose in Table 6. This shows that the condition number of G is suitable for the observability index of the AOA estimator.

5. Discussion

This study handles the case where only a single sound source exists. In practice, there can be multiple signal sources [22,23,24]. Under the assumption that sound sources rarely overlap in the time-frequency domain, one can apply the proposed AOA scheme for estimating the AOA of every sound source. Thereafter, delay-and-sum beamforming in [25,26,27] can separate sound arriving from an estimated sound direction.

6. Conclusions

Considering a small underwater platform (e.g., underwater unmanned vehicle), this article addresses how to estimate the signal direction utilizing the minimum number of omnidirectional microphones. Suppose that each omnidirectional microphone measures a real-valued sound signal whose speed and frequency information are not known in advance. This paper addresses how to estimate the AOA of the incoming signal utilizing only three omnidirectional microphones. This study further presents how to estimate the AOA using a general configuration composed of three microphones.
The effectiveness of the proposed AOA estimator with only three omnidirectional microphones is demonstrated by comparing it with the MUSIC algorithm under computer simulations. Considering the case where only three microphones are used, this paper shows that the proposed estimator outperforms the MUSIC estimator considering both estimation accuracy and computation time.
In the future, we will extend the proposed AOA estimator to 3D environments. In 3D environments, one needs more than three omnidirectional microphones to estimate the elevation and azimuth of a 3D non-cooperative target. We will study the minimum number of omnidirectional microphones for estimating the elevation and azimuth of a 3D non-cooperative target.

Funding

This work was supported by the Korea Research Institute for defense Technology planning and advancement (KRIT) grant funded by the Korea government (DAPA (Defense Acquisition Program Administration)) (No. KRIT-CT-22-052, Physics-guided Intelligent Sonar Signal Detection Research Laboratory, 2024).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The author declares no conflicts of interest.

References

  1. Li, B.; Zhai, J. A Novel Sound Speed Profile Prediction Method Based on the Convolutional Long-Short Term Memory Network. J. Mar. Sci. Eng. 2022, 10, 572. [Google Scholar] [CrossRef]
  2. Nardone, S.C.; Aidala, V.J. Observability Criteria for Bearings-Only Target Motion Analysis. IEEE Trans. Aerosp. Electron. Syst. 1981, AES-17, 162–166. [Google Scholar] [CrossRef]
  3. Fascista, A.; Ciccarese, G.; Coluccia, A.; Ricci, G. Angle of Arrival-Based Cooperative Positioning for Smart Vehicles. IEEE Trans. Intell. Transp. Syst. 2018, 19, 2880–2892. [Google Scholar] [CrossRef]
  4. Peach, N. Bearings-only tracking using a set of range-parameterised extended Kalman filters. IEE Proc. Control Theory Appl. 1995, 142, 73–80. [Google Scholar] [CrossRef]
  5. Kim, J. Obstacle Information Aided Target Tracking Algorithms for Angle-Only Tracking of a Highly Maneuverable Target in Three Dimensions. IET Radar Sonar Navig. 2019, 13, 1074–1080. [Google Scholar] [CrossRef]
  6. Ristic, B.; Arulampalam, M.S. Tracking a manoeuvring target using angle-only measurements: Algorithms and performance. Signal Process. 2003, 83, 1223–1238. [Google Scholar] [CrossRef]
  7. Kim, H.; Kang, K.; Shin, J.W. Factorized MVDR Deep Beamforming for Multi-Channel Speech Enhancement. IEEE Signal Process. Lett. 2022, 29, 1898–1902. [Google Scholar] [CrossRef]
  8. Yazdi, N.; Todros, K. Measure-Transformed MVDR Beamforming. IEEE Signal Process. Lett. 2020, 27, 1959–1963. [Google Scholar] [CrossRef]
  9. Ibrahim, E.N.; Khalil, E. Improve the robustness of MVDR beamforming method based on steering vector estimation and sparse constraint. In Proceedings of the 2019 International Symposium on Advanced Electrical and Communication Technologies (ISAECT), Rome, Italy, 27–29 November 2019; pp. 1–5. [Google Scholar]
  10. Li, W.; Tang, Q.; Huang, C. A New Close Form Location Algorithm with AOA and TDOA for Mobile User. Wirel. Pers. Commun. 2017, 97, 3061–3080. [Google Scholar] [CrossRef]
  11. Tota, R.; Hossain, M.S. Three Dimentional Multiple Near-field Source Localization Based on MUSIC Algorithm to Increase the Localization Accuracy of Optimal Beamformer. In Proceedings of the 2021 3rd International Conference on Electrical & Electronic Engineering (ICEEE), Rajshahi, Bangladesh, 22–24 December 2021; pp. 9–12. [Google Scholar]
  12. Santos, E.; Zoltowski, M. Spatial power spectrum estimation based on a MVDR-MMSE-MUSIC hybrid beamformer. In Proceedings of the Proceedings. (ICASSP ’05). IEEE International Conference on Acoustics, Speech, and Signal Processing, 2005, Philadelphia, PA, USA, 23–23 March 2005; Volume 4, pp. iv/809–iv/812. [Google Scholar]
  13. Ahmed, I.; Anirudh Koushik, B.; Charan Kumar, G.; Neethu, S. Simulation of Direction of Arrival Using MUSIC Algorithm and Beamforming using Variable Step Size LMS Algorithm. In Proceedings of the 2022 IEEE Microwaves, Antennas, and Propagation Conference (MAPCON), Bangalore, India, 12–16 December 2022; pp. 993–997. [Google Scholar]
  14. Lei, S.; Wu, Y.; Wang, X.; Wang, Z.; Yang, S. Adaptive Beamformer Based on Effectiveness of Reconstruction. In Proceedings of the 2019 IEEE 2nd International Conference on Electronic Information and Communication Technology (ICEICT), Harbin, China, 20–22 January 2019; pp. 353–357. [Google Scholar]
  15. Adhikari, K.; Drozdenko, B. Symmetry-Imposed Rectangular Coprime and Nested Arrays for Direction of Arrival Estimation With Multiple Signal Classification. IEEE Access 2019, 7, 153217–153229. [Google Scholar] [CrossRef]
  16. Zhou, X.; Zhu, F.; Jiang, Y.; Zhou, X.; Tan, W.; Huang, M. The Simulation Analysis of DOA Estimation Based on MUSIC Algorithm. In Proceedings of the 2020 5th International Conference on Mechanical, Control and Computer Engineering (ICMCCE), Harbin, China, 25–27 December 2020; pp. 1483–1486. [Google Scholar]
  17. Wen, F.; Javed, U.; Yang, Y.; He, D.; Zhang, Y. Improved subspace direction-of-arrival estimation in unknown nonuniform noise fields. In Proceedings of the 2016 Fourth International Conference on Ubiquitous Positioning, Indoor Navigation and Location Based Services (UPINLBS), Shanghai, China, 2–4 November 2016; pp. 230–233. [Google Scholar]
  18. Mao, Z.; Li, B.; Dong, L.; Qiao, Y.; Sun, H.; Li, Y. An Effective Algorithm for Direction-of-Arrival Estimation of Coherent Signals with ULA. In Proceedings of the 2023 5th International Conference on Natural Language Processing (ICNLP), Guangzhou, China, 24–26 March 2023; pp. 136–140. [Google Scholar]
  19. Lin, B.; Hu, G.; Zhou, H.; Zheng, G.; Song, Y. The DOA Estimation Method for Low-Altitude Targets under the Background of Impulse Noise. Sensors 2022, 22, 4853. [Google Scholar] [CrossRef] [PubMed]
  20. Kintz, A.L.; Gupta, I.J. A Modified MUSIC Algorithm for Direction of Arrival Estimation in the Presence of Antenna Array Manifold Mismatch. IEEE Trans. Antennas Propag. 2016, 64, 4836–4847. [Google Scholar] [CrossRef]
  21. Mohanna, M.; Rabeh, M.L.; Zieur, E.M.; Hekala, S. Optimization of MUSIC algorithm for angle of arrival estimation in wireless communications. NRIAG J. Astron. Geophys. 2013, 2, 116–124. [Google Scholar] [CrossRef]
  22. Florio, A.; Avitabile, G.; Coviello, G. Multiple Source Angle of Arrival Estimation Through Phase Interferometry. IEEE Trans. Circuits Syst. II Express Briefs 2022, 69, 674–678. [Google Scholar] [CrossRef]
  23. Cai, X.; Sarabandi, K. A Fast Analytic Multiple-Sources Angle-of-Arrival Estimation Algorithm for Automotive MIMO Radars. In Proceedings of the 2020 IEEE International Symposium on Antennas and Propagation and North American Radio Science Meeting, Montreal, QC, Canada, 5–10 July 2020; pp. 1301–1302. [Google Scholar]
  24. Cobos, M.; Lopez, J.J.; Spors, S. A sparsity-based approach to 3D binaural sound synthesis using time-frequency array processing. EURASIP J. Adv. Signal Process 2010, 2010, 415840. [Google Scholar] [CrossRef]
  25. Xia, H.; Ma, Y.; Yang, K.; Cao, R.; Chen, P.; Li, H. Delay-and-sum beamforming based on the diagonal reducing method. In Proceedings of the OCEANS 2017—Aberdeen, Aberdeen, UK, 19–22 June 2017; pp. 1–5. [Google Scholar]
  26. Van Veen, B.; Buckley, K. Beamforming: A versatile approach to spatial filtering. IEEE ASSP Mag. 1988, 5, 4–24. [Google Scholar] [CrossRef] [PubMed]
  27. Hussain, A.; Chellappan, K.; Zamratol M., S. Evaluation of multichannel speech signal separation with beamforming techniques. In Proceedings of the 2014 IEEE Conference on Biomedical Engineering and Sciences (IECBES), Kuala Lumpur, Malaysia, 8–10 December 2014; pp. 766–771. [Google Scholar]
Figure 1. The microphone configuration with three microphones. u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u .
Figure 1. The microphone configuration with three microphones. u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u .
Jmse 12 00874 g001
Figure 2. There are two microphones, S 1 and S 2 . A straight infinite line connects these two microphones. Dotted arrows indicate the signal direction at each microphone. Utilizing the phase differences at these two microphones, we cannot determine whether the target exists to the left or to the right of this line.
Figure 2. There are two microphones, S 1 and S 2 . A straight infinite line connects these two microphones. Dotted arrows indicate the signal direction at each microphone. Utilizing the phase differences at these two microphones, we cannot determine whether the target exists to the left or to the right of this line.
Jmse 12 00874 g002
Figure 3. A general configuration with three microphones. u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u . A j denotes the angle of the j-th microphone measured in the counter-clockwise direction starting from the x-axis of the microphone configuration. Moreover, r j denotes the distance between the center and the j-th microphone.
Figure 3. A general configuration with three microphones. u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u . A j denotes the angle of the j-th microphone measured in the counter-clockwise direction starting from the x-axis of the microphone configuration. Moreover, r j denotes the distance between the center and the j-th microphone.
Jmse 12 00874 g003
Figure 4. A singular microphone configuration where D 1 = D 2 = D 3 . u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u . A j denotes the angle of the j-th microphone measured in the counter-clockwise direction starting from the x-axis of the microphone configuration. Moreover, r j denotes the distance between the center and the j-th microphone. One cannot determine whether the signal direction is u or u .
Figure 4. A singular microphone configuration where D 1 = D 2 = D 3 . u defines the unit vector from the origin to the target. ϕ defines the bearing angle of the target signal, such that π < ϕ π . S i ( i { 1 , 2 , 3 } ) defines the 2D coordinates of the i-th microphone. P i defines the projection of S i onto u . A j denotes the angle of the j-th microphone measured in the counter-clockwise direction starting from the x-axis of the microphone configuration. Moreover, r j denotes the distance between the center and the j-th microphone. One cannot determine whether the signal direction is u or u .
Jmse 12 00874 g004
Figure 5. The signal strength at every microphone ( σ G = 0.5 ). This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.5 ) = 10 l o g ( 2 ) in dB.
Figure 5. The signal strength at every microphone ( σ G = 0.5 ). This implies that the signal-to-noise ratio (SNR) is 10 l o g ( 1 0.5 ) = 10 l o g ( 2 ) in dB.
Jmse 12 00874 g005
Figure 6. This figure depicts a general microphone configuration. We set A 2 = π / 2 and A 3 = π / 2 in Section 3.1. Moreover, we set r 1 = 0.4 λ , r 2 = r 1 / 2 , and r 3 = r 1 / 3 .
Figure 6. This figure depicts a general microphone configuration. We set A 2 = π / 2 and A 3 = π / 2 in Section 3.1. Moreover, we set r 1 = 0.4 λ , r 2 = r 1 / 2 , and r 3 = r 1 / 3 .
Jmse 12 00874 g006
Table 1. Algorithm comparison ( S N R = 10 ).
Table 1. Algorithm comparison ( S N R = 10 ).
ϕ Pro ( avgErr ) Pro ( stdErr ) MU ( avgErr ) MU ( stdErr )
−1800.040.2−0.010.4
−160−0.020.2−1340.5
−1400.030.21340.5
−1200.030.20.020.4
−1000.040.2−1340.4
−80−0.030.21340.3
−60−0.040.2−0.050.4
−40−0.0070.2−1340.3
−200.0010.21340.4
0−0.0040.2−0.050.4
200.010.2−1340.3
400.0090.21340.4
600.010.20.020.3
80−0.010.2−1340.3
100−0.0090.21340.3
120−0.0040.2−0.010.4
1400.010.2−1340.3
1600.020.21360.4
Table 2. Algorithm comparison ( S N R = 0 ).
Table 2. Algorithm comparison ( S N R = 0 ).
ϕ Pro ( avgErr ) Pro ( stdErr ) MU ( avgErr ) MU ( stdErr )
−1800.172.5−0.532
−1600.092.5−12145
−1400.12.311553
−1201.82.5−0.139
−100−0.52.8−11854
−800.292.511950
−600.152.6−625
−400.42.7−12853
−200.22.512557
0−0.32.6−119
200.012.2−12050
400.0032.911960
60−0.52.7−127
800.32.9−11965
100−0.32.611965
120−0.22.9223
1400.012.6−11058
1600.12.510961
Table 3. Algorithm evaluation ( S N R = 10 ).
Table 3. Algorithm evaluation ( S N R = 10 ).
ϕ General ( avgErr ) General ( stdErr )
−1800.030.5
−160−0.0090.5
−140−0.020.5
−120−0.040.4
−1000.070.3
−800.040.4
−60−0.040.4
−400.070.4
−200.10.4
00.010.4
200.070.4
40−0.080.4
600.050.4
800.020.4
100−0.060.3
120−0.080.5
1400.10.5
1600.0080.5
Table 4. Algorithm evaluation ( S N R = 0 ).
Table 4. Algorithm evaluation ( S N R = 0 ).
ϕ General ( avgErr ) General ( stdErr )
−180−0.175
−160−0.048
−1400.94
−120−0.44
−1000.63
−80−0.0075
−600.84
−400.55
−20−0.75
00.95
200.79
400.25
600.054
80−0.33
100−0.14
120−0.94
140−0.25
160−0.25
Table 5. Algorithm evaluation ( S N R = 10 ).
Table 5. Algorithm evaluation ( S N R = 10 ).
ϕ Random ( avgErr ) Random ( stdErr )
−1800.0060.8
−160−0.00070.6
−140−0.07310.6
−1200.04280.5
−100−0.01440.5
−80−0.02510.4
−60−0.05520.5
−40−0.09550.7
−200.07880.7
00.07130.6
20−0.11470.7
40−0.11860.5
600.07080.5
80−0.01360.4
1000.06310.4
1200.08670.5
140−0.02500.5
160−0.03910.6
Table 6. Algorithm evaluation of loose requirements for microphone configuration ( S N R = 10 ).
Table 6. Algorithm evaluation of loose requirements for microphone configuration ( S N R = 10 ).
ϕ Random ( avgErr ) Random ( stdErr )
−180−0.93356
−160−0.19445
−140−0.44415
−120−0.17363
−1000.04181
−800.11431
−60−0.34004
−40−0.01722
−20−0.27717
0−0.62916
20−1.29537
400.86273
60−0.03111
80−0.44954
100−0.21531
120−0.17632
1400.03532
1600.41475
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Kim, J. Angle of Arrival Estimator Utilizing the Minimum Number of Omnidirectional Microphones. J. Mar. Sci. Eng. 2024, 12, 874. https://doi.org/10.3390/jmse12060874

AMA Style

Kim J. Angle of Arrival Estimator Utilizing the Minimum Number of Omnidirectional Microphones. Journal of Marine Science and Engineering. 2024; 12(6):874. https://doi.org/10.3390/jmse12060874

Chicago/Turabian Style

Kim, Jonghoek. 2024. "Angle of Arrival Estimator Utilizing the Minimum Number of Omnidirectional Microphones" Journal of Marine Science and Engineering 12, no. 6: 874. https://doi.org/10.3390/jmse12060874

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop