Placement Planning for Sound Source Tracking in Active Drone Audition
Abstract
:1. Introduction
2. Related Work
2.1. Definition of Active Drone Audition
2.2. Sound Source Tracking Using Microphone Arrays
2.3. Robot/Drone Activities in Robot/Drone Audition
3. Method
3.1. MUSIC Method
3.2. Particle Filtering with Integrated MUSIC (PAFIM)
3.3. Active Placement Optimization
- a.
- Microphone arrays should be allocated only to sound sources they can hear if there are multiple sound sources.
- b.
- Pairs of array-to-source directions should be orthogonal with each other to have stable location estimation.
- c.
- Microphone arrays should be close to the sound source to maintain high SNR.
- d.
- Drones should make small movements as possible.
- 1.
- Initialize parameters (See Section 3.3.4)
- 2.
- Calculate MUSIC spectrum by MUSIC method [9]
- 3.
- Calculate probability of observation from Equation (8) and assign microphone arrays to sources to be tracked according to probability of observation.
- 4.
- Perform sound source tracking for each sound source with the corresponding microphone array group.
- 5.
- Calculate next drone positions by estimated source locations with Equation (11).
Algorithm 1 Proposed algorithm (For one time step). |
Require:
|
3.3.1. Probability of Observation Update
3.3.2. Sound Source Tracking
3.3.3. Optimization of Drone Placement
3.3.4. Initialization
4. Evaluation
4.1. Single-Source Tracking
4.1.1. Simulation Settings
- Inner
- Place at the inner part of the source trajectory.
- Rand1, 2
- Uniformly random placements within the area .
- Surround
- Place at the outer part of the source trajectory.
- MA1
- MA2
- MA3
- Potamitis04
- Calculates the triangulation points based on estimated directions from each microphone array and applies Kalman filtering to the average point of them [29].
- Lauzon17
- Tracks the sound source by particle filtering where the weight of particles are determined by the angular difference between the direction estimation and the expectation [31].
- Yamada20
- Converts multiple triangulation points into a Gaussian mixture and applies Gaussian sum filtering to track the sound source [11].
- Yamada21
- PAFIM introduced in Section 3.2.
4.1.2. Results and Discussion
4.2. Multiple Source Tracking
4.2.1. Simulation Settings
4.2.2. Results and Discussion
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
Abbreviations
SNR | Signal-to-noise ratio |
MUSIC | Multiple signal classification |
PAFIM | Particle filtering with integrated MUSIC |
RMSE | Root-mean-square error |
References
- Brown, G.J.; Cooke, M. Computational auditory scene analysis. Comput. Speech Lang. 1994, 8, 297–336. [Google Scholar]
- Okuno, H.G.; Nakadai, K. Robot audition: Its rise and perspectives. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, South Brisbane, QLD, Australia, 19–24 April 2015; pp. 5610–5614. [Google Scholar]
- Hoshiba, K.; Washizaki, K.; Wakabayashi, M.; Ishiki, T.; Kumon, M.; Bando, Y.; Gabriel, D.; Nakadai, K.; Okuno, H. Design of UAV-embedded microphone array system for sound source localization in outdoor environments. Sensors 2017, 17, 2535. [Google Scholar]
- Nakadai, K.; Kumon, M.; Okuno, H.G.; Hoshiba, K.; Wakabayashi, M.; Washizaki, K.; Ishiki, T.; Gabriel, D.; Bando, Y.; Morito, T.; et al. Development of microphone-array-embedded UAV for search and rescue task. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), IEEE, Vancouver, BC, Canada, 24–28 September 2017; pp. 5985–5990. [Google Scholar]
- Hoshiba, K.; Sugiyama, O.; Nagamine, A.; Kojima, R.; Kumon, M.; Nakadai, K. Design and assessment of sound source localization system with a UAV-embedded microphone array. J. Robot. Mechatronics 2017, 29, 154–167. [Google Scholar]
- Sibanyoni, S.V.; Ramotsoela, D.T.; Silva, B.J.; Hancke, G.P. A 2-D acoustic source localization system for drones in search and rescue missions. IEEE Sensors J. 2018, 19, 332–341. [Google Scholar]
- Van Veen, B.D.; Buckley, K.M. Beamforming: A versatile approach to spatial filtering. IEEE ASSP Mag. 1988, 5, 4–24. [Google Scholar]
- Knapp, C.; Carter, G. The generalized correlation method for estimation of time delay. IEEE Trans. Acoust. Speech Signal Process. 1976, 24, 320–327. [Google Scholar] [CrossRef] [Green Version]
- Schmidt, R. Multiple emitter location and signal parameter estimation. IEEE Trans. Antennas Propag. 1986, 34, 276–280. [Google Scholar] [CrossRef] [Green Version]
- Wakabayashi, M.; Washizaka, K.; Hoshiba, K.; Nakadai, K.; Okuno, H.G.; Kumon, M. Design and Implementation of Real-Time Visualization of Sound Source Positions by Drone Audition. In Proceedings of the 2020 IEEE/SICE International Symposium on System Integration (SII), Honolulu, HI, USA, 12–15 January 2020; pp. 814–819. [Google Scholar]
- Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Sound Source Tracking by Drones with Microphone Arrays. In Proceedings of the 2020 IEEE/SICE International Symposium on System Integration, Honolulu, HI, USA, 12–15 January 2020. [Google Scholar]
- Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Sound Source Tracking Using Integrated Direction Likelihood for Drones with Microphone Arrays. In Proceedings of the 2020 IEEE/SICE International Symposium on System Integration (SII), Honolulu, HI, USA, 12–15 January 2020; pp. 394–399. [Google Scholar]
- Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Assessment of sound source tracking using multiple drones equipped with multiple microphone arrays. Int. J. Environ. Res. Public Health 2021, 18, 9039. [Google Scholar]
- Kagami, S.; Thompson, S.; Sasaki, Y.; Mizoguchi, H.; Enomoto, T. 2D sound source mapping from mobile robot using beamforming and particle filtering. In Proceedings of the 2009 IEEE International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 3689–3692. [Google Scholar]
- Sasaki, Y.; Tanabe, R.; Takemura, H. Probabilistic 3D sound source mapping using moving microphone array. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 1293–1298. [Google Scholar] [CrossRef]
- Misra, P.; Kumar, A.A.; Mohapatra, P.; Balamuralidhar, P. Droneears: Robust acoustic source localization with aerial drones. In Proceedings of the 2018 IEEE International Conference on Robotics and Automation (ICRA), Brisbane, QLD, Australia, 21–25 May 2018; pp. 80–85. [Google Scholar]
- Atmoko, H.; Tan, D.; Tian, G.; Fazenda, B. Accurate sound source localization in a reverberant environment using multiple acoustic sensors. Meas. Sci. Technol. 2008, 19, 024003. [Google Scholar]
- Ishi, C.T.; Even, J.; Hagita, N. Using multiple microphone arrays and reflections for 3D localization of sound sources. In Proceedings of the 2013 IEEE/RSJ International Conference on Intelligent Robots and Systems, Tokyo, Japan, 3–7 November 2013; pp. 3937–3942. [Google Scholar]
- Ishi, C.T.; Even, J.; Hagita, N. Speech activity detection and face orientation estimation using multiple microphone arrays and human position information. In Proceedings of the 2015 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Hamburg, Germany, 28 September–2 October 2015; pp. 5574–5579. [Google Scholar] [CrossRef]
- Gabriel, D.; Kojima, R.; Hoshiba, K.; Itoyama, K.; Nishida, K.; Nakadai, K. Design and assessment of multiple-sound source localization using microphone arrays. In Proceedings of the 2019 IEEE/SICE International Symposium on System Integration (SII), IEEE, Paris, France, 14–16 January 2019; pp. 199–204. [Google Scholar]
- Gabriel, D.; Kojima, R.; Hoshiba, K.; Itoyama, K.; Nishida, K.; Nakadai, K. 2D sound source position estimation using microphone arrays and its application to a VR-based bird song analysis system. Adv. Robot. 2019, 33, 403–414. [Google Scholar]
- Kaneko, S.; Gamper, H. Large-scale simulation of bird localization systems in forests with distributed microphone arrays. JASA Express Lett. 2022, 2, 101201. [Google Scholar]
- Washizaki, K.; Wakabayashi, M.; Kumon, M. Position estimation of sound source on ground by multirotor helicopter with microphone array. In Proceedings of the 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Daejeon, Republic of Korea, 9–14 October 2016; pp. 1980–1985. [Google Scholar]
- Wakabayashi, M.; Okuno, H.G.; Kumon, M. Multiple sound source position estimation by drone audition based on data association between sound source localization and identification. IEEE Robot. Autom. Lett. 2020, 5, 782–789. [Google Scholar]
- Valin, J.M.; Michaud, F.; Rouat, J. Robust 3D Localization and Tracking of Sound Sources Using Beamforming and Particle Filtering. In Proceedings of the 2006 IEEE International Conference on Acoustics Speech and Signal Processing Proceedings, Toulouse, France, 14–19 May 2006; Volume 4, p. IV. [Google Scholar] [CrossRef] [Green Version]
- Portello, A.; Bustamante, G.; Danès, P.; Piat, J.; Manhes, J. Active localization of an intermittent sound source from a moving binaural sensor. In Proceedings of the Forum Acuticum, Krakow, Poland, 7–12 September 2014. [Google Scholar]
- Evers, C.; Dorfan, Y.; Gannot, S.; Naylor, P.A. Source tracking using moving microphone arrays for robot audition. In Proceedings of the 2017 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 6145–6149. [Google Scholar] [CrossRef]
- Brandstein, M.S.; Silverman, H.F. A practical methodology for speech source localization with microphone arrays. Comput. Speech Lang. 1997, 11, 91–126. [Google Scholar]
- Potamitis, I.; Chen, H.; Tremoulis, G. Tracking of multiple moving speakers with multiple microphone arrays. IEEE Trans. Speech Audio Process. 2004, 12, 520–529. [Google Scholar]
- Busset, J.; Perrodin, F.; Wellig, P.; Ott, B.; Heutschi, K.; Rüuhl, T.; Nussbaumer, T. Detection and tracking of drones using advanced acoustic cameras. In Unmanned/Unattended Sensors and Sensor Networks XI; and Advanced Free-Space Optical Communication Techniques and Applications; SPIE: Bellingham, WA, USA, 2015; Volume 9647, pp. 53–60. [Google Scholar]
- Lauzon, J.; Grondin, F.; Létourneau, D.; Desbiens, A.L.; Michaud, F. Localization of RW-UAVs using particle filtering over distributed microphone arrays. In Proceedings of the 2017 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Vancouver, BC, Canada, 24–28 September 2017; pp. 2479–2484. [Google Scholar] [CrossRef]
- Evers, C.; Habets, E.A.P.; Gannot, S.; Naylor, P.A. DoA Reliability for Distributed Acoustic Tracking. IEEE Signal Process. Lett. 2018, 25, 1320–1324. [Google Scholar] [CrossRef] [Green Version]
- Gazor, S.; Grenier, Y. Criteria for positioning of sensors for a microphone array. IEEE Trans. Speech Audio Process. 1995, 3, 294–303. [Google Scholar]
- Rabinkin, D.V.; Renomeron, R.J.; French, J.C.; Flanagan, J.L. Optimum microphone placement for array sound capture. In Advanced Signal Processing: Algorithms, Architectures, and Implementations VII; SPIE: Bellingham, WA, USA, 1997; Volume 3162, pp. 227–239. [Google Scholar]
- Akbarzadeh, V.; Gagne, C.; Parizeau, M.; Argany, M.; Mostafavi, M.A. Probabilistic sensing model for sensor placement optimization based on line-of-sight coverage. IEEE Trans. Instrum. Meas. 2012, 62, 293–303. [Google Scholar]
- Song, B.; Roy-Chowdhury, A.K. Robust tracking in a camera network: A multi-objective optimization framework. IEEE J. Sel. Top. Signal Process. 2008, 2, 582–596. [Google Scholar]
- Ma, N.; May, T.; Wierstorf, H.; Brown, G.J. A machine-hearing system exploiting head movements for binaural sound localisation in reverberant conditions. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 2699–2703. [Google Scholar]
- Tourbabin, V.; Barfuss, H.; Rafaely, B.; Kellermann, W. Enhanced robot audition by dynamic acoustic sensing in moving humanoids. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), IEEE, South Brisbane, QLD, Australia, 19–24 April 2015; pp. 5625–5629. [Google Scholar]
- Bustamante, G.; Danés, P.; Forgue, T.; Podlubne, A. Towards information-based feedback control for binaural active localization. In Proceedings of the 2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 6325–6329. [Google Scholar]
- Schmidt, A.; Löllmann, H.W.; Kellermann, W. Acoustic self-awareness of autonomous systems in a world of sounds. Proc. IEEE 2020, 108, 1127–1149. [Google Scholar]
- Martinson, E.; Schultz, A. Discovery of sound sources by an autonomous mobile robot. Auton. Robot. 2009, 27, 221–237. [Google Scholar]
- Vincent, E.; Sini, A.; Charpillet, F. Audio source localization by optimal control of a mobile robot. In Proceedings of the 2015 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), South Brisbane, QLD, Australia, 19–24 April 2015; pp. 5630–5634. [Google Scholar]
- Nguyen, Q.V.; Colas, F.; Vincent, E.; Charpillet, F. Long-term robot motion planning for active sound source localization with Monte Carlo tree search. In Proceedings of the 2017 Hands-free Speech Communications and Microphone Arrays (HSCMA), San Francisco, CA, USA, 1–3 March 2017; pp. 61–65. [Google Scholar]
- Nguyen, Q.V.; Colas, F.; Vincent, E.; Charpillet, F. Motion planning for robot audition. Auton. Robot. 2019, 43, 2293–2317. [Google Scholar]
- Asano, F.; Goto, M.; Itou, K.; Asoh, H. Real-time sound source localization and separation system and its application to automatic speech recognition. In Proceedings of the Seventh European Conference on Speech Communication and Technology, Aalborg, Denmark, 3–7 September 2001. [Google Scholar]
- Hioka, Y.; Kingan, M.; Schmid, G.; Stol, K.A. Speech enhancement using a microphone array mounted on an unmanned aerial vehicle. In Proceedings of the 2016 IEEE International Workshop on Acoustic Signal Enhancement (IWAENC), Xi’an, China, 13–16 September 2016; pp. 1–5. [Google Scholar]
- Nakadai, K.; Takahashi, T.; Okuno, H.G.; Nakajima, H.; Hasegawa, Y.; Tsujino, H. Design and Implementation of Robot Audition System’HARK’—Open Source Software for Listening to Three Simultaneous Speakers. Adv. Robot. 2010, 24, 739–761. [Google Scholar]
- Wang, L.; Sanchez-Matilla, R.; Cavallaro, A. Tracking a moving sound source from a multi-rotor drone. In Proceedings of the 2018 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Madrid, Spain, 1–5 October 2018; pp. 2511–2516. [Google Scholar]
- Sekiguchi, K.; Bando, Y.; Itoyama, K.; Yoshii, K. Layout optimization of cooperative distributed microphone arrays based on estimation of source separation performance. J. Robot. Mechatronics 2017, 29, 83–93. [Google Scholar]
Robot Audition | Drone Audition | |
---|---|---|
Not utilizing activity | ||
Utilizing activity |
|
|
TripointAvr | Potamitis04 | Lauzon17 | Yamada20 | Yamada21 | |
---|---|---|---|---|---|
Proposed | 0.37 | 3.78 | 9.13 | 0.55 | 1.64 |
Inner | 7.08 | 4.52 | 38.05 | 9.29 | 20.22 |
Rand1 | 1.52 | 3.76 | 13.14 | 4.67 | 4.88 |
Rand2 | 1.5 | 3.76 | 3.04 | 1.38 | 2.47 |
Surround | 2.53 | 3.46 | 3.49 | 1.62 | 2.36 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yamada, T.; Itoyama, K.; Nishida, K.; Nakadai, K. Placement Planning for Sound Source Tracking in Active Drone Audition. Drones 2023, 7, 405. https://doi.org/10.3390/drones7070405
Yamada T, Itoyama K, Nishida K, Nakadai K. Placement Planning for Sound Source Tracking in Active Drone Audition. Drones. 2023; 7(7):405. https://doi.org/10.3390/drones7070405
Chicago/Turabian StyleYamada, Taiki, Katsutoshi Itoyama, Kenji Nishida, and Kazuhiro Nakadai. 2023. "Placement Planning for Sound Source Tracking in Active Drone Audition" Drones 7, no. 7: 405. https://doi.org/10.3390/drones7070405