Next Article in Journal
The Development of Digital Image Colorimetric Quantitative Analysis of Multi-Explosives Using Polymer Gel Sensors
Previous Article in Journal
Closing Connectivity Gap: An Overview of Mobile Coverage Solutions for Not-Spots in Rural Zones
Previous Article in Special Issue
Multirate Audio-Integrated Feedback Active Noise Control Systems Using Decimated-Band Adaptive Filters for Reducing Narrowband Noises
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sound Source Localization Using a Convolutional Neural Network and Regression Model

Department of Electrical Engineering, National Taipei University of Technology, Taipei 10608, Taiwan
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(23), 8031; https://doi.org/10.3390/s21238031
Submission received: 15 October 2021 / Revised: 25 November 2021 / Accepted: 29 November 2021 / Published: 1 December 2021
(This article belongs to the Special Issue Intelligent Acoustic Sensors and Its Applications)

Abstract

:
In this research, a novel sound source localization model is introduced that integrates a convolutional neural network with a regression model (CNN-R) to estimate the sound source angle and distance based on the acoustic characteristics of the interaural phase difference (IPD). The IPD features of the sound signal are firstly extracted from time-frequency domain by short-time Fourier transform (STFT). Then, the IPD features map is fed to the CNN-R model as an image for sound source localization. The Pyroomacoustics platform and the multichannel impulse response database (MIRD) are used to generate both simulated and real room impulse response (RIR) datasets. The experimental results show that an average accuracy of 98.96% and 98.31% are achieved by the proposed CNN-R for angle and distance estimations in the simulation scenario at SNR = 30 dB and RT60 = 0.16 s, respectively. Moreover, in the real environment, the average accuracies of the angle and distance estimations are 99.85% and 99.38% at SNR = 30 dB and RT60 = 0.16 s, respectively. The performance obtained in both scenarios is superior to that of existing models, indicating the potential of the proposed CNN-R model for real-life applications.

1. Introduction

Localization technologies are widely used in everyday applications, such as navigation, human–computer interaction, surveillance, rescue, and smart monitoring [1,2]. Global positioning system (GPS) is the most frequently used technology for outdoor positioning [3,4]. However, GPS accuracy is degraded when it is used in indoor environments due to obstacles blocking the signal’s propagation [5,6]. Consequently, a number of technologies, such as infrared (IR), Bluetooth, and Wi-Fi, have been developed to address the challenge of indoor positioning. These technologies have become widely used for indoor localization and positioning in recent years [7]. The propagation path of radio signals can be line-of-sight (LOS) or non-line-of-sight (NLOS) in indoor environments [8]. However, the signals of indoor positioning technologies must be propagated in LOS conditions in order to produce accurate location estimates [9]. Although IR offers high localization accuracy, its signal can be easily obscured by obstacles [10]. Bluetooth and Wi-Fi have the advantage of strong penetrating power, which can penetrate through indoor obstacles [11,12]. Nevertheless, Bluetooth is disadvantaged by its short range, and Wi-Fi requires high costs of hardware installation and maintenance [13]. Sound has the advantages of strong penetrating power, simple construction, and low cost [14]. Additionally, sound includes a tone, timbre, and other features, which make it more effective than other technologies [15]. For example, the frequency of sound emitted from different locations can be distinguished efficiently, and multiple sound sources can be located at the same time. Therefore, sound source localization (SSL) has attracted much attention in recent years [16,17,18].
Currently, two types of sound source localization methods are generally used in the literature. First, the microphone array methods use the microphone array as a receiving end to determine the direction of the sound source. The microphone arrays can be divided into linear arrays, circular arrays, and distributed arrays. Second, human ear analysis methods identify the sound source via simulating the signal received by the human ear. It was shown in [19,20,21,22,23,24] that binaural beamforming-based methods can achieve high noise reduction and sound sources preservation and localization. Microphone array-based methods can be further divided into four approaches under different acoustic characteristics as follows [25,26,27,28]:
  • Beamforming: calculate the input signal power, phase, and amplitude of each receiving point through beamforming technology, and calculate the azimuth angle of the sound source with the greatest probability.
  • Time difference of arrival (TDOA): the time difference between the signals’ arrival at two or more receiving points is combined with the spatial information of these receiving points to infer the azimuth of the sound source [29].
  • High-resolution spectrum estimation (HRSE): the signal at the receiving point is used to calculate the correlation between the spatial and spectral characteristics to obtain the azimuth angle of the sound source [30].
  • Neural network (NN): train a NN model using a large amount of data to find audio patterns for multiple acoustic sources localization [31].
Recently, various deep neural networks (DNNs) were employed for sound source localization. Chakrabarty et al. [32] proposed a CNN-based supervised learning (CNN-SL) approach to estimate the direction of arrival (DOA) of multiple speakers. The phase component of the STFT coefficients of the received microphone signals are directly fed into the CNN, and the features for DOA estimates are learned during the training process. The ability of the DOA estimation method to accurately adapt to unseen acoustic conditions is pretty robust. However, this method is highly dependent on the time-varying source signal [33]. Yiwere et al. [34] presented a sound source distance estimation (SSDE) approach by using a convolutional recurrent neural network (CRNN). The CRNN is trained using log-scaled mel spectrograms extracted from single-channel audio signals as input features. The transformation of the audio signals to images allows the convolutional layers of the network to extract distance-dependent features from the audio signals. The experimental results showed that the CRNN model can achieve a high level of accuracy. Another interesting research work [35] proposed an indoor sound source regional localization method based on a convolutional neural network (CNN). The sound source signal is converted into a spectral map and fed into the CNN for regional localization. The simulation results showed that the CNN can bring better robustness and generalization with different SNRs. Pang et al. [36] introduced a binaural sound source localization (SSL) method based on time–frequency CNN (TF-CNN) with multitask learning to simultaneously localize azimuth and elevation under various acoustic conditions. The IPD and interaural level difference (ILD) are first extracted from the received binaural signals, then each or both of them are fed to the SSL neural network. The experimental results illustrated that the proposed method can achieve comparable localization performance. Nevertheless, such methods are restricted to certain ranges or areas.
This research aims to construct an indoor localization model based on the characteristics of the sound spectrum, which can estimate the azimuth angle and distance of the indoor speaker. The CNN is used to automatically extract features and increase their versatility and robustness by training the model to resist noise. Previous works used classification functions to normalize the output of a CNN to a probability distribution over the output target. However, the output of classification functions is a discrete value, and hence, it does not predict the exact value in the case of continuous variables. Unlike the previous studies, our CNN uses a regression function instead of a classification function because it is better suited for the continuous variable output. Additionally, this research uses the Pyroomacoustics [37] platform to quickly construct a virtual three-dimensional space and generate a room impulse response (RIR) with spatial sound signals. Moreover, real space signals are synthesized with a multi-channel impulse response database [38]. In addition, the signal dataset is converted into a time-frequency domain signal through STFT and then is converted to an IPD feature map to be fed to the CNN model. Finally, the distribution of output values with the regression model are observed to find the best configuration of the model through training and evaluate the performance of the model in different environments.

2. Proposed Methods

The overall flow chart of the proposed sound source localization system is demonstrated in Figure 1. The sound database signal is firstly convolved with the real and simulated RIR to obtain a new signal with spatial effect. Then the STFT of the new signal is obtained, and the IPD features are extracted. Finally, the CNN-R model is trained on the IPD features to estimate the angle and distance of the sound source. Notably, the IPD image sets are divided into 70% training set, 10% validation set, and 20% test set.

2.1. Data Collection

This research uses the CMU_ARCTIC database, which is a speech database in CMU_ARCTIC speech synthesis databases [39], established by the Language Technologies Institute of Carnegie Mellon University, USA. This database is mainly used in the research of speech synthesis. The content of the corpus database was selected by the non-copyright center of Project Gutenberg, which is about 1150 sentences. An audio of two males and two females with American English accents were collected. The recording format is 16 bits, the sampling rate is 32 KHz, and the length of each sentence is 3 seconds. The database has a total of 4528 audio files.

2.1.1. Simulated Room Database

In this research, we built our own simulated spatial dataset, and the RIR was simulated by using the Pyroomacoustics Python platform [37]. The length and the width of the generated space are 5 × 5, 5 × 6, 6 × 6, 7 × 6, and 7 × 7 (m 2 ), and the height is 2.5 m. The position of the two microphones is (x = (width/2) ± 0.3, y = 1, z = 1). Figure 2 shows an example of a 5 × 5 space. The microphones are located at (2.2, 1, 1) and (2.8, 1, 1). The sound source point is 1 and 2 m from the midpoint of the two microphones. The angle is distributed from 0° to 180°, where every 15° is a step distance. In total, there are 26 source points. The sound database adopts the CMU_ARCTIC database. A total of 100 audio files are taken from the corpus of 4 participants. Convolution operations are performed at each sampling point to generate spatial sound effects, and we adjust RT60 and SNR to achieve data diversity.

2.1.2. Real Room Database

This research uses the real room response database generated by the Multichannel Impulse Response Database. The database was constructed by the Institute of Communication Systems and Data Processing at RWTH Aachen University, Germany. These data can produce a variety of reverberation levels by changing the spatial scene. The database mainly has three different reverberation levels with reverberation times (RT60) of 0.16 s, 0.36 s, and 0.61 s, respectively [38]. On different grids in space, in the angular range of 90° to −90°, each 15° step is taken as a measurement point, and each measurement point is 1 m or 2 m away from the microphone array. There are 26 measurement points in total, as shown in Figure 3.

2.2. Data Processing

After the simulation and real space datasets are generated, data processing is still needed before the data can be fed to the model for training. First, the sound signal is transformed to the time-frequency domain signals through STFT, then an IPD feature map is produced by using the Hann window as a sliding window for STFT. The Hann window size is 2048 sampling points with 512 overlapping sampling points and a sampling rate of 14,400 Hz. IPD is a feature formed by converting two audio signals into frequency spectra and subtracting each phase. The intensity of the left and right phases can be observed. Unlike using two spectra for training, one IPD makes training faster. Hann’s window is used because it shows superior performance in random signals. The conversion formula is shown in Formula (1):
Y m ( ω , k ) = A m ( ω , k ) e ( j φ m ( ω , k ) ) , m = l , r
where A m ( ω , k ) and φ m ( ω , k ) are the components of amplitude and phase at frequency ω and time k, respectively. l and r are the left and right channels, respectively.
The IPD features are obtained based on the phase difference between the two receiving ends, and its formula can be expressed as follows:
φ ( ω , k ) = Y l ( ω , k ) Y r ( ω , k )
where Y l ( ω , k ) and Y r ( ω , k ) are the left and right receiving signals. The IPD can be obtained by subtracting its phase components. In other words, the IPD is computed as the difference of the phase angles, and phase unwrapping is used on the phase image. Figure 4 is an example of the actual output of the IPD.
The dataset of the simulated IPD includes 400 audio records, 5 spatial sizes (5 × 5, 5 × 6, 6 × 6, 7 × 6, and 7 × 7 m 2 ), 26 sampling points, 5 SNR (0, 5, 10, 20, and 30 dB), 3 RT60 (0.16, 0.36, and 0.61 s), and a total of 780,000 images. The dataset of Real IPD includes: 400 audio records, 1 spatial size (6 × 6 m 2 ), 26 sampling points, 5 SNR (0, 5, 10, 20, and 30 dB), 3 RT60 (0.16, 0.36, and 0.61 s), and a total of 156,000 images. The nature of the noise is independent Gaussian white noise added to each channel, and it is computed as follows:
S N R = 10 l o g 10 S i g n a l P o w e r N o i s e P o w e r d B

2.3. The Proposed CNN-R Architecture

The proposed model is composed of two architectures, including CNN and regression architecture. First, the resolution of the original data is reduced from 1395 × 906 to 32 × 32 before training. The CNN is constructed by two convolutional layers. The first layer has 128 kernels and the size of the kernel is 77, the padding is same, and the stride is 1. The second convolutional layer has 16 kernels of the size, 7 × 7, the padding is the same, and the stride is 1. The ReLu activation function is used after each convolutional layer. The feature map output by the last layer of CNN is flattened by the fully connected (FC) layer and then is used to train the model. The regression model has three layers in total. The number of kernels in the first and second layers is 512, and the number of kernels in the third layer is set to 1 to generate the final output. The activation function used between each layer is linear. Figure 5 shows the overall architecture of the proposed CNN-R model used in this research. Additionally, Table 1 shows the experimental training setting.
The proposed CNN-R architecture is kept as small as possible to avoid overfitting. The simpler architecture is less likely to overfit [40]. The choice and numbers of layers were decided by trial and error. The main criteria were accuracy and MAE. When the accuracy and MAE are not improving on the validation dataset during the training stage, the training process is interrupted after a certain number of epochs, and then the structure of the model is modified. This process is repeated until the new structure produces satisfactory results. In the CNN part, the model started with one convolutional layer and one ReLu layer; however, the results were not satisfactory. Hence, we added one more convolutional layer followed by the ReLu layer, and the results were impressive. Additionally, the regression part, including Dense and the linear activation function, were used because the output of the FC layer is a 1D continuous variable output. Finally, a Dense layer was used to produce the value of angle and distance.

3. Experimental Results

The generated IPD dataset was divided into three parts: training, validation, and testing. We used the validation dataset to monitor the performance of our model during the training and adjust the model hyperparameters to find the optimal configuration. All the experiments were performed using a PC with Intel Core i7-7700, CPU 3.6 GHz, and 32 GB of RAM. The GPU was NVIDIA GeForce RTX 2070 with 8 GB of memory. The model was implemented in Python with TensorFlow.

3.1. Performance Evaluation Metrics

In this study, the regression model is used to evaluate the overall performance by comparing the difference between the predicted value and the actual value. According to the distribution of output values, the precision and accuracy (Acc.) are interpreted as follows:
  • High precision and high accuracy: the distribution of predicted values is concentrated and close to the target value.
  • High precision and low accuracy: the distribution of the predicted values is concentrated, but far from the target value.
  • Low precision and high accuracy: the distribution of the predicted values is more scattered, but around the target value.
  • Low precision and low accuracy: the distribution of the predicted values is more scattered and far from the target value.
Figure 6 illustrates the relationship between precision and accuracy with respect to the distribution of the predicted values. In this research, the mean absolute error (MAE) is used to measure the accuracy. MAE calculates the difference between the predicted value and the actual value. The formula is shown in (4):
M A E = 1 N i = 1 N | y i y i ^ |
Because the output of the regression model is a continuous value, this research uses the formula in (5) to evaluate the accuracy of the proposed CNN-R:
A c c u r a c y = N f i n e N T
where N T is the total times of experiments, N f i n e is the number of the error which is less than the unit scale divided by 2. In this research, the angle estimation unit scale is 15° and the distance unit scale is 1 meter; therefore, the angle is 7.5° (15/2) and the distance estimation is 0.5 meter (1/2) as the baseline value. When the predicted value is less than the baseline value, it is considered to be a correct value. In this research, the experiments are mainly divided into two parts as follows:
  • Experiment 1 in a simulated environment consists of two parts: (i) a single acoustic environment is used to train the model for angle and distance estimation, and (ii) a multiple acoustic environment is used to train the model for angle and distance estimation.
  • Experiment 2 uses a real spatial sound dataset to train the model for angle and distance estimation.

3.2. Experiment 1

3.2.1. Model Performance in a Single Acoustic Environment

In the experiment, the goal is to show the ability of the proposed CNN-R architecture to make correct predictions in different room dimensions under the same RT60. The use of different room dimensions is to avoid data overfitting and validate the performance of CNN-R. Additionally, the same RT60 is used to avoid environmental parameter changes. The experimental results show that the proposed CNN-R can be generalized and used in multiple acoustic environments. Table 2 shows the single acoustic environment configuration. The training set room includes 5 × 5, 6 × 5, 6 × 7, and 7 × 7 (m 2 ). The SNRs are 0 dB and 5 dB, respectively. The RT60 is 0.16 s. On the other hand, the testing set room is 6 × 6 (m 2 ). The SNRs are 10 dB, 20 dB, and 30 dB, respectively. The RT60 is 0.16 s.
Table 3 and Table 4 show the model performance for angle and distance estimation in the single acoustic environment under three SNR scenarios, respectively.
Figure 7 shows the average of accuracy and MAE for the angle and distance estimation in the single acoustic environment under SNR = 10 dB, 20 dB, 30 dB, and RT60 = 0.16 s. In the single acoustic environment, the accuracy of the angle and distance estimation increases as the SNR increases. When the SNR is greater than 20 dB, the angle and distance accuracy can reach 99.42% and 96.12%, respectively. Additionally, the MAE is reduced to 0.8° and 0.18 m, and the RMSE is reduced to 1.32° and 0.14 m. The accuracy of the angle estimation model in each SNR is better than the distance estimation model.

3.2.2. Model Performance in a Multiple Acoustic Environment

Table 5 shows the multiple acoustic environment configuration. The training set room includes 5 × 5, 6 × 5, 6 × 7, and 7 × 7 (m 2 ). The SNRs are 0 dB, 5 dB, and 10 dB, respectively. The RT60 are set to 0.16 s, 0.36 s, and 0.61 s, respectively. In order to be different from the training environment, the testing set room is 6 × 6 (m 2 ). The SNRs are set to 10 dB, 20 dB, and 30 dB, respectively. The RT60 are 0.16 s, 0.36 s, and 0.61 s, respectively, which are the same as the training set.
Table 6 shows the model performance for angle estimation in the multiple acoustic environment under SNR = 10, 20, 30 dB, and three RT60 scenarios. Table 7 shows the model performance for distance estimation in the multiple acoustic environment under SNR = 10, 20, and 30 dB, and three RT60 scenarios.
Figure 8 and Figure 9 show the model performance for the angle and distance estimation in the multiple acoustic environments under SNR = 10 dB, 20 dB, 30 dB, and RT60 = 0.16 s, 0.36 s, 0.61 s, respectively. In the multiple acoustic environment, two acoustic spaces, RT60 = 0.36 s and RT60 = 0.61 s, were added. The overall accuracy increases with the increase of SNR and the decrease of RT60. When RT60 = 0.61 s, the performance of angle estimation is not satisfactory, and the average accuracy is 61.19%. However, if RT60 is reduced to 0.36 s, the accuracy can be greatly increased by about 20%. Moreover, MAE and RMSE drop sharply at RT60 = 0.16 s. Nevertheless, the best performance for angle estimation is achieved when SNR = 30 dB, MAE = 1.96, and RMSE = 1.64, where the accuracy is 98.96%.

3.3. Experiment 2

The Model Performance in a Real Acoustic Environment

Table 8 shows the real acoustic environment configuration. The training set room is 6 × 6 (m 2 ). SNR are 0 dB, 5 dB, and 10 dB. The RT60 are 0.16 s, 0.36 s, and 0.61 s. The test set room is 6 × 6 (m 2 ). The SNR are 10 dB, 20 dB, and 30 dB. The RT60 are 0.16 s, 0.36 s, and 0.61 s, which is similar to the training set.
Table 9 shows the model performance for distance estimation in the real acoustic environment under SNR = 10, 20, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s, respectively.
Table 10 shows the model performance for angle estimation in the real acoustic environment under SNR = 10, 20, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s, respectively.
Table 11 shows the average Acc and MAE of the proposed model for angle and distance estimation in real acoustic environments, where SNR = 10 dB, 20 dB, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s, respectively. In a real acoustic environment, the angle estimation accuracy increases and the error decreases as SNR increases and RT60 decreases. Moreover, when the SNR is greater than 20 dB, the accuracy obtained is higher than 96%, and the MAE is less than 1.7°. The accuracy of distance estimation is also improved with the increase in SNR. Overall, the accuracy is higher than 95% when SNR = 20 dB and 30 dB. The Acc. and MAE of each RT60 are stable when SNR is greater than 20 dB. Table 12 shows the accuracy of CNN-R for angle and distance estimation compared to other methods based on the multi-channel impulse response database [38].
The training–validation loss curves for the proposed CNN-R in a single acoustic environment, multiple acoustic environment, and real acoustic environment are shown in Figure 10. Unlike the single acoustic environment and multiple acoustic environment, the loss in real acoustic environment gradually reduces and slowly converges as the number of epochs increases. Moreover, note in Figure 10c that the training loss curve and validation loss curve behave similarly, which implies that the proposed CNN-R model can be generalized and does not suffer from overfitting.

4. Discussion

This research aims to establish a general sound localization model. The results of a single acoustic environment in Experiment 1 show that under different room conditions, the test model can still effectively estimate the angle and distance in a single acoustic environment, and it will be more accurate as the SNR increases. In the multiple acoustic environments, good estimation performance can also be obtained under different room conditions. When RT60 = 0.61 s, the accuracy is relatively insufficient. However, as the SNR increases, the accuracy can be effectively improved. The model proposed in this research has the best performance in the simulated room where the RT60 is less than 0.36 s and the SNR is greater than 20 dB. In addition, in the real acoustic environment of Experiment 2, the overall accuracy is enhanced significantly, verifying the practicability of our proposed model in a real acoustic environment. The experimental results show that the MAE of the model for the angle estimation is smaller than the distance estimation, which means that the error between the predicted value and the actual value is small. Nonetheless, the RMSE of the model for angle estimation is greater than the distance estimation, which means that a small number of predicted values has large variations; hence, the model for angle estimation has high accuracy. However, the precision is low. On the other hand, the model performance for distance estimation has high accuracy and high precision.
Comparing the results of the proposed CNN-R in the multiple acoustic environment in Experiment 1 with the results in the real acoustic environment in Experiment 2, it can clearly be seen that under the same environmental acoustic parameters, the accuracy of the model trained in the real environment is higher than that of the simulated acoustic environment. The reason for this result is that when generating simulated room sound effects, the only parameters we can adjust are SNR and RT60. However, in the real environment, the parameters that affect sound propagation are more complex. Therefore, the model trained with the simulation dataset has insufficient features, which affects the learning of the model, resulting in a decrease in accuracy. The experimental results show that the accuracy of the distance estimation is better than that of the angle estimation. The reason is that there are 13 target values for the angle estimation and only 2 target values for the distance estimation, which increases the complexity of the angle estimation model weight training and makes the weight distribution uneven.
Taking Table 3 and Table 4 as an example, when SNR = 10 dB, the accuracy of the angle estimation is between 71% and 100%. The accuracy close to 90° is higher, and the accuracy close to 0° or 180° decreases on both sides. The accuracy of the distance estimation is distributed between 87.08% and 94.15%, and the distribution of the accuracy of the distance estimation is more concentrated than that of the angle estimation. Moreover, the Acc. is low and MAE is high due to the small number of training samples in the single acoustic environment compared to the multiple acoustic environment. Additionally, in general, the accuracy drops significantly when the value of RT60 increases, except when the angle is 90 degrees in the multiple acoustic environment. One limitation of the proposed model might be the offline design. Future work will focus on improving the proposed model for real-time positioning. Additionally, the proposed model still needs further enhancement for multiple sound source localization.

5. Conclusions and Future Works

In this paper, an original sound source localization model was developed by combining a convolutional neural network and a regression model (CNN-R). Simulated and real sound datasets were generated to perform the experiments. Initially, the sound signals were transformed into time-frequency signals through STFT, and then IPD feature maps were calculated from the time-frequency signals. These were then fed into the CNN-R model for a series of experiments. The evaluation metrics of Acc., MAE, and RMSE were used to evaluate the performance of the proposed model. The experimental results in the simulated acoustic scenarios showed that the proposed model can effectively estimate the angles and distances in a single or multiple acoustic environments under different spatial conditions. When SNR is greater than 10 dB and RT60 is less than 0.61s, the accuracy of the angle and distance estimations can reach, on average, more than 95%. Additionally, when SNR = 30 dB and RT60 = 0.16 s, the accuracies of the angle and distance estimations can reach 98.96% and 98.31%, respectively. On the other hand, the experimental results in the real acoustic scenarios showed that when the SNR is greater than 20 dB, the accuracy of the angle and distance estimation exceeds 96%. Furthermore, when SNR = 30 dB and RT60 = 0.16 s, the accuracies of the angle and distance estimations reach 99.85% and 99.38%, respectively. In comparison to the existing methods, the experimental results also showed that the proposed CNN-R outperforms the existing methods in terms of the angle and distance estimation accuracies. Future work will study the combination of other acoustic features, such as ILD, to make the features richer. Moreover, the impact of more acoustic environments on the accuracy will also be investigated.

Author Contributions

Conceptualization, T.-H.T., Y.-T.L. and Y.-L.C.; methodology, Y.-L.C. and Y.-T.L.; software, Y.-T.L.; validation, M.A., T.-H.T. and Y.-L.C.; formal analysis, Y.-L.C. and M.A.; investigation, Y.-T.L. and M.A.; data curation, Y.-T.L.; writing—review and editing, T.-H.T., M.A. and Y.-L.C.; supervision, T.-H.T. and M.A. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the Ministry of Science and Technology, Taiwan, No. MOST 110-2221-E-027-089.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Morar, A.; Moldoveanu, A.; Mocanu, I.; Moldoveanu, F.; Radoi, I.E.; Asavei, V.; Gradinaru, A.; Butean, A. A comprehensive survey of indoor localization methods based on computer vision. Sensors 2020, 20, 2641. [Google Scholar] [CrossRef]
  2. Bai, Y.; Lu, L.; Cheng, J.; Liu, J.; Chen, Y.; Yu, J. Acoustic-based sensing and applications: A survey. Comput. Netw. 2020, 181, 107447. [Google Scholar] [CrossRef]
  3. Zhu, R.; Wang, Y.; Cao, H.; Yu, B.; Gan, X.; Huang, L.; Zhang, H.; Li, S.; Jia, H.; Chen, J. RTK/pseudolite/LAHDE/IMU-PDR integrated pedestrian navigation system for urban and indoor environments. Sensors 2020, 20, 1791. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  4. Cengiz, K. Comprehensive Analysis on Least-Squares Lateration for Indoor Positioning Systems. IEEE Internet Things J. 2020, 8, 2842–2856. [Google Scholar] [CrossRef]
  5. Uzun, A.; Ghani, F.A.; Ahmadi Najafabadi, A.M.; Yenigün, H.; Tekin, İ. Indoor Positioning System Based on Global Positioning System Signals with Down-and Up-Converters in 433 MHz ISM Band. Sensors 2021, 21, 4338. [Google Scholar] [CrossRef] [PubMed]
  6. Bhat, S.J.; Venkata, S.K. An optimization based localization with area minimization for heterogeneous wireless sensor networks in anisotropic fields. Comput. Netw. 2020, 179, 107371. [Google Scholar] [CrossRef]
  7. Zhu, X.; Qu, W.; Qiu, T.; Zhao, L.; Atiquzzaman, M.; Wu, D.O. Indoor intelligent fingerprint-based localization: Principles, approaches and challenges. IEEE Commun. Surv. Tutor. 2020, 22, 2634–2657. [Google Scholar] [CrossRef]
  8. Jo, H.J.; Kim, S. Indoor smartphone localization based on LOS and NLOS identification. Sensors 2018, 18, 3987. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  9. Gu, Y.; Lo, A.; Niemegeers, I. A survey of indoor positioning systems for wireless personal networks. IEEE Commun. Surv. Tutor. 2009, 11, 13–32. [Google Scholar] [CrossRef] [Green Version]
  10. Ye, T. Algorithms for Indoor Localization Based on IEEE 802.15. 4-2011 UWB and Inertial Sensors. Ph.D. Thesis, University College Cork, Cork, Ireland, 2015. [Google Scholar]
  11. Li, W.; Su, Z.; Zhang, K.; Benslimane, A.; Fang, D. Defending malicious check-in using big data analysis of indoor positioning system: An access point selection approach. IEEE Trans. Netw. Sci. Eng. 2020, 7, 2642–2655. [Google Scholar] [CrossRef]
  12. Maheepala, M.; Kouzani, A.Z.; Joordens, M.A. Light-based indoor positioning systems: A review. IEEE Sens. J. 2020, 20, 3971–3995. [Google Scholar] [CrossRef]
  13. Mirowski, P.; Ho, T.K.; Yi, S.; MacDonald, M. SignalSLAM: Simultaneous localization and mapping with mixed WiFi, Bluetooth, LTE and magnetic signals. In Proceedings of the International Conference on Indoor Positioning and Indoor Navigation, IEEE, Montbeliard, France, 28–31 October 2013; pp. 1–10. [Google Scholar]
  14. Chen, X.; Sun, H.; Zhang, H. A new method of simultaneous localization and mapping for mobile robots using acoustic landmarks. Appl. Sci. 2019, 9, 1352. [Google Scholar] [CrossRef] [Green Version]
  15. Yang, K.; Wang, K.; Bergasa, L.M.; Romera, E.; Hu, W.; Sun, D.; Sun, J.; Cheng, R.; Chen, T.; López, E. Unifying terrain awareness for the visually impaired through real-time semantic segmentation. Sensors 2018, 18, 1506. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  16. Sun, Y.; Chen, J.; Yuen, C.; Rahardja, S. Indoor sound source localization with probabilistic neural network. IEEE Trans. Ind. Electron. 2017, 65, 6403–6413. [Google Scholar] [CrossRef]
  17. Rahamanand, A.; Kim, B. Sound source localization in 2D using a pair of bio–inspired MEMS directional microphones. IEEE Sens. J. 2020, 21, 1369–1377. [Google Scholar] [CrossRef]
  18. Park, Y.; Choi, A.; Kim, K. Single-Channel Multiple-Receiver Sound Source Localization System with Homomorphic Deconvolution and Linear Regression. Sensors 2021, 21, 760. [Google Scholar] [CrossRef]
  19. Zohourian, M.; Enzner, G.; Martin, R. Binaural speaker localization integrated into an adaptive beamformer for hearing aids. IEEE ACM Trans. Audio Speech Lang. Process. 2017, 26, 515–528. [Google Scholar] [CrossRef]
  20. Amini, J.; Hendriks, R.C.; Heusdens, R.; Guo, M.; Jensen, J. Asymmetric coding for rate-constrained noise reduction in binaural hearing aids. IEEE ACM Trans. Audio Speech Lang. Process. 2018, 27, 154–167. [Google Scholar] [CrossRef]
  21. Koutrouvelis, A.I.; Hendriks, R.C.; Heusdens, R.; Jensen, J. A convex approximation of the relaxed binaural beamforming optimization problem. IEEE ACM Trans. Audio Speech Lang. Process. 2018, 27, 321–331. [Google Scholar] [CrossRef]
  22. Koutrouvelis, A.I.; Hendriks, R.C.; Heusdens, R.; Jensen, J. Relaxed binaural LCMV beamforming. IEEE ACM Trans. Audio Speech Lang. Process. 2016, 25, 137–152. [Google Scholar] [CrossRef] [Green Version]
  23. Jeffet, M.; Shabtai, N.R.; Rafaely, B. Theory and perceptual evaluation of the binaural reproduction and beamforming tradeoff in the generalized spherical array beamformer. IEEE ACM Trans. Audio Speech Lang. Process. 2016, 24, 708–718. [Google Scholar] [CrossRef]
  24. Marquardt, D.; Hohmann, V.; Doclo, S. Interaural coherence preservation in multi-channel Wiener filtering-based noise reduction for binaural hearing aids. IEEE ACM Trans. Audio Speech Lang. Process. 2015, 23, 2162–2176. [Google Scholar] [CrossRef]
  25. Rascon, C.; Meza, I. Localization of sound sources in robotics: A review. Robot. Auton. Syst. 2017, 96, 184–210. [Google Scholar] [CrossRef]
  26. Cobos, M.; Antonacci, F.; Alexandridis, A.; Mouchtaris, A.; Lee, B. A survey of sound source localization methods in wireless acoustic sensor networks. Wirel. Commun. Mob. Comput. 2017, 2017. [Google Scholar] [CrossRef]
  27. Argentieri, S.; Danes, P.; Souères, P. A survey on sound source localization in robotics: From binaural to array processing methods. Comput. Speech Lang. 2015, 34, 87–112. [Google Scholar] [CrossRef] [Green Version]
  28. Pak, J.; Shin, J.W. Sound localization based on phase difference enhancement using deep neural networks. IEEE ACM Trans. Audio Speech Lang. Process. 2019, 27, 1335–1345. [Google Scholar] [CrossRef]
  29. Scheuing, J.; Yang, B. Disambiguation of TDOA estimation for multiple sources in reverberant environments. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 1479–1489. [Google Scholar] [CrossRef]
  30. Farmani, M. Informed Sound Source Localization for Hearing Aid Applications. Ph.D. Thesis, Aalborg University, Aalborg, Denmark, 2017. [Google Scholar]
  31. Youssef, K.; Argentieri, S.; Zarader, J.L. A learning-based approach to robust binaural sound localization. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, IEEE, Tokyo, Japan, 3–7 November 2013; pp. 2927–2932. [Google Scholar]
  32. Chakrabarty, S.; Habets, E.A. Multi-speaker DOA estimation using deep convolutional networks trained with noise signals. IEEE J. Sel. Top. Signal Process. 2019, 13, 8–21. [Google Scholar] [CrossRef] [Green Version]
  33. Hu, Y.; Samarasinghe, P.N.; Gannot, S.; Abhayapala, T.D. Semi-supervised multiple source localization using relative harmonic coefficients under noisy and reverberant environments. IEEE ACM Trans. Audio Speech Lang. Process. 2020, 28, 3108–3123. [Google Scholar] [CrossRef]
  34. Yiwere, M.; Rhee, E.J. Sound source distance estimation using deep learning: An image classification approach. Sensors 2020, 20, 172. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  35. Zhang, X.; Sun, H.; Wang, S.; Xu, J. A new regional localization method for indoor sound source based on convolutional neural networks. IEEE Access 2018, 6, 72073–72082. [Google Scholar] [CrossRef]
  36. Pang, C.; Liu, H.; Li, X. Multitask learning of time-frequency CNN for sound source localization. IEEE Access 2019, 7, 40725–40737. [Google Scholar] [CrossRef]
  37. Scheibler, R.; Bezzam, E.; Dokmanić, I. Pyroomacoustics: A python package for audio room simulation and array processing algorithms. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 351–355. [Google Scholar]
  38. Hadad, E.; Heese, F.; Vary, P.; Gannot, S. Multichannel audio database in various acoustic environments. In Proceedings of the 2014 14th International Workshop on Acoustic Signal Enhancement (IWAENC), IEEE, Juan les Pins, France, 8–11 September 2014; pp. 313–317. [Google Scholar]
  39. Kominek, J.; Black, A.W.; Ver, V. CMU ARCTIC databases for speech synthesis. In Proceedings of the Fifth ISCA ITRW on Speech Synthesis, Pittsburgh, PA, USA, 14–16 June 2004; pp. 223–224. [Google Scholar]
  40. Chollet, F. Deep Learning with Python; Simon and Schuster: Shelter Island, NY, USA, 2021. [Google Scholar]
Figure 1. Flow chart of the proposed system.
Figure 1. Flow chart of the proposed system.
Sensors 21 08031 g001
Figure 2. Simulated space configuration diagram.
Figure 2. Simulated space configuration diagram.
Sensors 21 08031 g002
Figure 3. Multichannel Impulse Response Database measurement space configuration diagram [38].
Figure 3. Multichannel Impulse Response Database measurement space configuration diagram [38].
Sensors 21 08031 g003
Figure 4. An example of the actual output of IPD where the color map represents the angle (°).
Figure 4. An example of the actual output of IPD where the color map represents the angle (°).
Sensors 21 08031 g004
Figure 5. The architecture of CNN-R.
Figure 5. The architecture of CNN-R.
Sensors 21 08031 g005
Figure 6. Performance evaluation criteria where yellow dots are the predicted values. (1) is high accuracy and high precision, which is the best scenario. (2) is low accuracy and high precision. (3) is high accuracy and low precision. (4) is low accuracy and low precision, which is the worst scenario.
Figure 6. Performance evaluation criteria where yellow dots are the predicted values. (1) is high accuracy and high precision, which is the best scenario. (2) is low accuracy and high precision. (3) is high accuracy and low precision. (4) is low accuracy and low precision, which is the worst scenario.
Sensors 21 08031 g006
Figure 7. The average accuracy and MAE of (a) angle and (b) distance estimation by CNN-R in a single acoustic environment, where SNR = 10 dB, 20 dB, 30 dB, and RT60 = 0.16 s.
Figure 7. The average accuracy and MAE of (a) angle and (b) distance estimation by CNN-R in a single acoustic environment, where SNR = 10 dB, 20 dB, 30 dB, and RT60 = 0.16 s.
Sensors 21 08031 g007
Figure 8. The average ACC and MAE of angle estimation by CNN-R in a multiple acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and (a) RT60 = 0.16 s, (b) RT60 = 0.36 s, and (c) RT60 = 0.61 s.
Figure 8. The average ACC and MAE of angle estimation by CNN-R in a multiple acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and (a) RT60 = 0.16 s, (b) RT60 = 0.36 s, and (c) RT60 = 0.61 s.
Sensors 21 08031 g008
Figure 9. The average ACC and MAE of distance estimation by CNN-R in a multiple acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and (a) RT60 = 0.16 s, (b) RT60 = 0.36 s, and (c) RT60 = 0.61 s.
Figure 9. The average ACC and MAE of distance estimation by CNN-R in a multiple acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and (a) RT60 = 0.16 s, (b) RT60 = 0.36 s, and (c) RT60 = 0.61 s.
Sensors 21 08031 g009
Figure 10. The training–validation loss curves of CNN-R, where (a) is the performance of CNN-R in a single acoustic environment. (b) is the performance of CNN-R in a multiple acoustic environment. (c) is the performance of CNN-R in a real acoustic environment.
Figure 10. The training–validation loss curves of CNN-R, where (a) is the performance of CNN-R in a single acoustic environment. (b) is the performance of CNN-R in a multiple acoustic environment. (c) is the performance of CNN-R in a real acoustic environment.
Sensors 21 08031 g010
Table 1. The settings of the training option used in the single acoustic environment.
Table 1. The settings of the training option used in the single acoustic environment.
HyperparametersConfigurations
OptimizerAdam
Loss FunctionMAE
Learning Rate0.001
Decay 10 3
Execution EnvironmentGPU
Batch Size64
Table 2. A single acoustic environment configuration.
Table 2. A single acoustic environment configuration.
Training SetTest Set
Room size (m2)5 × 5, 6 × 5, 6 × 7, 7 × 76 × 6
SNR (dB)0, 510, 20, 30
RT60 (s)0.160.16
Table 3. Performance of angle estimation by CNN-R in a single acoustic environment.
Table 3. Performance of angle estimation by CNN-R in a single acoustic environment.
SNR = 10 dBSNR = 20 dBSNR = 30 dB
Angle (°) Acc. (%) MAE (°) Acc. (%) MAE (°) Acc. (%) MAE (°)
071.0015.6790.504.2899.001.68
1587.503.1398.001.43100.000.91
3083.505.0194.002.0897.001.66
4588.504.3297.001.60100.000.67
6099.001.02100.000.54100.000.57
75100.000.53100.000.32100.000.34
90100.000.47100.000.22100.000.18
10599.500.69100.000.43100.000.53
120100.000.72100.000.40100.000.40
13585.505.2797.501.4999.500.52
15082.004.5991.502.4697.501.62
16594.502.9599.500.86100.000.56
18080.007.8597.001.5399.500.81
Average90.084.0297.311.3699.420.8
Table 4. Performance of distance estimation by CNN-R in a single acoustic environment.
Table 4. Performance of distance estimation by CNN-R in a single acoustic environment.
SNR = 10 dBSNR = 20 dBSNR = 30 dB
Distance (m) Acc. (%) MAE (m) Acc. (%) MAE (m) Acc. (%) MAE (m)
187.080.2595.690.1896.920.17
294.150.2196.080.1895.310.19
Average90.620.2395.880.1896.120.18
Table 5. A multiple acoustic environment configuration.
Table 5. A multiple acoustic environment configuration.
Training SetTesting Set
Room size (m 2 )5 × 5, 6 × 5, 6 × 7, 7 × 76 × 6
SNR (dB)0, 5, 1010, 20, 30
RT60 (s)0.16, 0.36, 0.610.16, 0.36, 0.61
Table 6. Performance of angle estimation by CNN-R in a multiple acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
Table 6. Performance of angle estimation by CNN-R in a multiple acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
RT60 = 0.16 sRT60 = 0.36 sRT60 = 0.61 s
Angle (°) SNR Acc. (%) MAE (°) Acc. (%) MAE (°) Acc. (%) MAE (°)
01080.5005.3285.0004.3259.5010.69
2093.0002.6697.5002.6981.5005.57
3099.0002.3999.5002.0287.0004.29
151096.0002.8280.0005.1731.5017.70
2097.5002.6782.5004.3835.5018.60
3099.5002.1593.5003.8844.0015.74
301091.0003.2458.0008.5544.5021.34
2096.5002.4761.0008.3148.5024.91
3096.0002.1963.5006.5448.5025.24
451089.5003.4579.5006.0645.0019.54
2096.5002.1693.5003.1658.5014.38
3099.5001.5695.5002.5665.0012.35
601086.5003.9080.0004.9260.5008.24
2098.0002.1794.5003.6363.5007.24
30100.001.5697.0003.4360.5006.92
751095.5003.1080.5905.0571.0005.89
20100.001.5798.0002.2481.0004.80
30100.001.1696.5002.0081.0004.74
9010100.000.8599.5000.95100.000.78
20100.000.52100.000.50100.000.52
30100.000.51100.000.44100.000.41
1051098.5001.6684.5004.1458.0009.04
2099.5001.0297.5001.7961.5008.28
30100.001.0699.5001.5971.5006.19
1201087.0003.7376.5007.4043.0014.05
2098.0001.8184.0004.5949.5010.04
3099.5001.3795.5002.5651.5008.14
1351083.5003.8064.0009.9329.0032.05
2095.0002.1979.0007.1244.5024.60
3099.5001.3881.0004.9549.5019.69
1501093.5002.9788.5003.4558.0011.81
2095.5002.4694.0002.5061.5011.39
3099.5001.9997.0002.0765.0010.48
1651098.0002.6277.0005.0460.5013.61
20100.001.8475.5005.0145.5010.05
30100.001.7981.0004.7651.5007.10
1801072.0006.4269.0008.3933.5022.87
2098.5002.3185.0004.2752.0017.15
3094.0002.8587.5003.8056.5013.37
Average1090.1203.3778.6205.6453.3814.43
2097.5401.9987.8503.8660.2312.12
3098.9601.6991.3103.1263.9610.36
Table 7. Performance of distance estimation by CNN-R in a multiple acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
Table 7. Performance of distance estimation by CNN-R in a multiple acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
RT60 = 0.16 sRT60 = 0.36 sRT60 = 0.61 s
Distance (m) SNR Acc. (%) MAE Acc. (%) MAE Acc. (%) MAE
11086.0000.2684.3100.2976.4600.34
2096.7700.1794.7700.2085.0800.27
3098.4600.1698.3800.1689.0800.24
21092.0000.2492.6200.2385.9200.26
2095.9200.1996.3800.1885.6900.26
3098.1500.1897.6900.1882.3100.30
Average1089.0000.2588.4600.2681.1900.30
2096.3500.1895.5800.1985.3800.27
3098.3100.1798.0400.1785.6900.27
Table 8. The real acoustic environment configuration.
Table 8. The real acoustic environment configuration.
Training SetTest Set
Room size (m 2 )6 × 66 × 6
SNR (dB)0, 5, 1010, 20, 30
RT60 (s)0.16, 0.36, 0.610.16, 0.36, 0.61
Table 9. Performance of distance estimation by CNN-R in a real acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
Table 9. Performance of distance estimation by CNN-R in a real acoustic environment at SNRs of 10, 20, and 30 dB, respectively.
RT60 = 0.16 sRT60 = 0.36 sRT60 = 0.61 s
Distance (m) SNR Acc. (%) MAE (°) Acc. (%) MAE (°) Acc. (%) MAE (°)
11088.0800.2595.3100.2094.6200.21
2098.5400.1599.3100.1499.2300.14
3099.9200.1499.9200.1399.8500.13
21090.2300.2589.4600.2492.2300.23
2097.6200.1797.8500.1697.3100.16
3098.8500.1599.6900.1499.1500.15
Average1089.1500.2592.3800.2293.4200.22
2098.0800.1698.5800.1598.2700.15
3099.3800.1499.8100.1399.5000.14
Table 10. Performance of angle estimation by CNN-R in a real acoustic environment at SNRs of 10, 20, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s.
Table 10. Performance of angle estimation by CNN-R in a real acoustic environment at SNRs of 10, 20, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s.
RT60 = 0.16 sRT60 = 0.36 sRT60 = 0.61 s
Angle (°) SNR Acc. (%) MAE (°) Acc. (%) MAE (°) Acc. (%) MAE (°)
01088.0004.0591.5004.0379.0010.15
2098.5001.6699.5001.2897.0002.62
30100.000.7999.5000.7799.5001.53
151080.0004.6289.0005.0892.5003.40
2095.0002.3695.0002.4899.0001.36
3099.0001.2999.5001.0899.5000.97
301095.0002.3581.0010.3677.5012.64
2099.5000.5697.5001.8195.0002.08
30100.000.37100.000.4499.0000.55
451087.0004.4991.5003.0880.5007.85
2099.5000.6499.5000.4697.5000.78
30100.000.32100.000.27100.000.30
601086.0005.8786.0006.1283.0007.33
2099.5000.5499.5000.4698.5000.88
30100.000.28100.000.29100.000.27
7510100.001.4678.5010.9782.5007.85
20100.000.3898.0001.5299.0000.90
30100.000.3296.5000.40100.000.35
901088.0004.2985.0005.0372.5012.51
2099.5000.8199.0000.7793.5002.50
30100.000.38100.000.4099.0000.54
1051086.0005.9878.5011.3661.5016.93
2098.5000.8297.5001.4995.5001.78
30100.000.28100.000.26100.000.32
1201093.0002.7486.0004.1278.0010.82
2098.5000.5898.0001.4596.5001.65
30100.000.21100.000.23100.000.29
1351085.5005.8087.5005.4382.5006.41
2098.5000.8198.0001.0498.5001.17
30100.000.26100.000.40100.000.48
1501082.5008.7491.0002.9078.5009.86
2094.0002.92100.000.5592.5002.60
3099.0000.55100.000.2999.5000.38
1651092.0002.8791.0003.0277.5009.77
2099.5000.9199.5001.0197.0001.41
30100.000.5399.5000.68100.000.59
1801076.5005.8481.5005.9982.0009.24
2092.5002.3799.0001.0297.5001.61
30100.000.72100.000.45100.001.12
Average1087.3804.5586.0006.0279.0409.06
2097.9201.1898.4601.1896.6901.64
3098.8500.4899.8500.4699.7300.59
Table 11. The average Acc. and MAE of angle and distance estimation by CNN-R in a real acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s, respectively.
Table 11. The average Acc. and MAE of angle and distance estimation by CNN-R in a real acoustic environment where SNR = 10 dB, 20 dB, and 30 dB, and RT60 = 0.16 s, 0.36 s, and 0.61 s, respectively.
AngleDistance
RT60 (s) SNR (dB) Acc. (%) MAE (°) SNR (dB) Acc. (%) MAE (m)
0.161087.3804.551089.1500.28
2097.9201.182098.0800.16
3099.8500.483099.3800.14
0.361086.0006.021092.3800.22
2098.4601.182098.5800.15
3099.8500.463099.8100.13
0.611079.0409.601093.4200.22
2096.6901.642098.2700.15
3099.7300.593099.5000.14
Table 12. Comparative results of angle and distance estimation based on the multi-channel impulse response database in a real acoustic environment at SNR = 30 dB and RT60 = 0.16 s.
Table 12. Comparative results of angle and distance estimation based on the multi-channel impulse response database in a real acoustic environment at SNR = 30 dB and RT60 = 0.16 s.
MethodAverage Angle (0–180°) Acc.Average Distance (1–2 m) Acc.
CNN-SL [32]90.25%88.85%
CRNN [34]87.37%85.64%
CNN [35]98.51%98.09%
TF-CNN [36]95.18%94.66%
CNN-R99.85%99.38%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tan, T.-H.; Lin, Y.-T.; Chang, Y.-L.; Alkhaleefah, M. Sound Source Localization Using a Convolutional Neural Network and Regression Model. Sensors 2021, 21, 8031. https://doi.org/10.3390/s21238031

AMA Style

Tan T-H, Lin Y-T, Chang Y-L, Alkhaleefah M. Sound Source Localization Using a Convolutional Neural Network and Regression Model. Sensors. 2021; 21(23):8031. https://doi.org/10.3390/s21238031

Chicago/Turabian Style

Tan, Tan-Hsu, Yu-Tang Lin, Yang-Lang Chang, and Mohammad Alkhaleefah. 2021. "Sound Source Localization Using a Convolutional Neural Network and Regression Model" Sensors 21, no. 23: 8031. https://doi.org/10.3390/s21238031

APA Style

Tan, T. -H., Lin, Y. -T., Chang, Y. -L., & Alkhaleefah, M. (2021). Sound Source Localization Using a Convolutional Neural Network and Regression Model. Sensors, 21(23), 8031. https://doi.org/10.3390/s21238031

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop