Next Article in Journal
A Proposed Algorithm Based on Variance to Effectively Estimate Crack Source Localization in Solids
Previous Article in Journal
Modeling and Analyzing the Availability of Technical Professional Profiles for the Success of Smart Cities Projects in Europe
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Joint Spatio-Temporal-Frequency Representation Learning for Improved Sound Event Localization and Detection

by
Baoqing Chen
1,
Mei Wang
2,* and
Yu Gu
1
1
School of Information and Communication, Guilin University of Electronic Technology, Guilin 541004, China
2
College of Physics and Electronic Information Engineering, Guilin University of Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(18), 6090; https://doi.org/10.3390/s24186090
Submission received: 7 August 2024 / Revised: 13 September 2024 / Accepted: 18 September 2024 / Published: 20 September 2024
(This article belongs to the Section Physical Sensors)

Abstract

:
Sound event localization and detection (SELD) is a crucial component of machine listening that aims to simultaneously identify and localize sound events in multichannel audio recordings. This task demands an integrated analysis of spatial, temporal, and frequency domains to accurately characterize sound events. The spatial domain pertains to the varying acoustic signals captured by multichannel microphones, which are essential for determining the location of sound sources. However, the majority of recent studies have focused on time-frequency correlations and spatio-temporal correlations separately, leading to inadequate performance in real-life scenarios. In this paper, we propose a novel SELD method that utilizes the newly developed Spatio-Temporal-Frequency Fusion Network (STFF-Net) to jointly learn comprehensive features across spatial, temporal, and frequency domains of sound events. The backbone of our STFF-Net is the Enhanced-3D (E3D) residual block, which combines 3D convolutions with a parameter-free attention mechanism to capture and refine the intricate correlations among these domains. Furthermore, our method incorporates the multi-ACCDOA format to effectively handle homogeneous overlaps between sound events. During the evaluation, we conduct extensive experiments on three de facto benchmark datasets, and our results demonstrate that the proposed SELD method significantly outperforms current state-of-the-art approaches.

1. Introduction

Sound event localization and detection (SELD) has played a key role in advancing technologies for environmental monitoring [1], healthcare [2], bioacoustic monitoring [3], multimedia content search and retrieval [4], and autonomous driving [5]. In particular, sound events usually occur in unpredictable and unstructured contexts, characterized by ambient noise, reverberation, and interference. Moreover, such events are frequently non-stationary and exhibit high intra-class variability. These complexities present significant challenges for those interested in utilizing SELD in real-life scenarios.
To address these challenges, many studies have explored 1D and 2D convolution-based approaches for the detection and localization of sound events in spatial audio recordings. For example, a 1D convolutional neural network is constructed within an end-to-end training framework to directly recognize the classes and locations of sound events from raw waveform files [6]. The 2D convolution-based methods are effective at extracting time-frequency patterns within each feature map and learning certain inter-map patterns from the combined features after pre-processing operations [7,8,9], making them more powerful than methods based on 1D convolutional neural networks. The effectiveness of these methods originates from maintaining proper time-frequency (TF) alignment between feature maps, which are created using different frequency distributions from various sound sources over time. This alignment ensures that the network learns meaningful patterns across frequencies, which can help accurately associate target classes with the location of multiple concurrent sound events.
Although these 1D or 2D convolution methods have seen some progress, none of them are fully compatible with the spatio-temporal-frequency volumes intrinsic to multichannel audio data. This incompatibility makes it difficult to explicitly model the interrelationships among audio channels, limiting the effective use of spatial information and resulting in incomplete sound event representations. The issue is exacerbated in cases of concurrent sound events, where spatial locations may be swapped or misinterpreted, thereby compromising the overall performance of SELD. Furthermore, a “reality gap” often emerges when algorithms trained on synthetic spatial audio datasets, such as the TAU NIGENS Spatial Sound Event (TNSSE) 2021 dataset [10], are applied to real-world audio data. These algorithms typically experience a decline in performance due to the acoustic differences and variability present in real-world environments, which differ markedly from those in synthetic datasets.
There has also been a lot of interest in enhancing convolutional neural networks with forms of an attention module for the SELD task [9,11,12,13]. These modules complement backbone or head networks by encoding categorical and directional information of sound events. In practice, the refinement of sound event features is frequently conducted along the feature channel [14] or the time-frequency dimension [15] and then progressively through a cascade of both dimensions [16,17]. When multidimensional dependencies are not considered concurrently, the model’s ability to extract more discriminative information from volumetric data may be constrained. Apart from that, most attention modules are parametric, introducing extra parameters that heighten the risk of overfitting, particularly in small or highly variable datasets.
In response to these concerns, this paper introduces a novel framework for polyphonic SELD that utilizes the newly developed Spatio-Temporal-Frequency Fusion Network (STFF-Net) to jointly learn comprehensive features of sound events across spatial, temporal, and frequency domains. Figure 1 presents an overview of the proposed SELD framework. We start by constructing the input volume for MIC and FOA formats from the recorded multichannel audio signals. The feature volume, after augmentation, is then fed into our STFF-Net, which uses a sequence of Enhanced-3D (E3D) residual blocks to extract and refine spatio-temporal-frequency features of sound events through the integration of 3D convolution with a non-parametric attention mechanism. Furthermore, we leverage the multi-ACCDOA format to enhance the network’s ability to handle overlapping sounds from the same class, making it more robust in diverse acoustic scenarios. During the evaluation, we conduct extensive experiments on both synthetic and real spatial audio datasets to assess the performance of our proposed framework. The experimental results demonstrate that our framework outperforms current state-of-the-art (SOTA) works by a large margin.
In a nutshell, the main contributions of this paper are as follows:
  • We, for the first time, conceptualize sound events as spatio-temporal-frequency entities, similar to actions in videos. This charts a new research trajectory for the SELD task.
  • We propose a novel STFF-Net architecture that cohesively integrates the detection and localization of sound events, simplifies the polyphonic SELD optimization process, and provides a spatio-temporal analysis of sound events.
  • We conduct comprehensive experiments on three public synthetic and real spatial audio datasets to validate the effectiveness of our proposed method.
The remainder of this paper is organized as follows. Section 2 reviews existing works, particularly those related to our approach. Section 3 introduces the details of our newly developed STFF-Net for polyphonic SELD. Section 4 describes the experimental setup and presents the results. Finally, the conclusions related to this work are given in Section 5.

2. Background and Related Works

2.1. Sound Event Localization and Detection

In the past decade, deep neural networks (DNNs) have emerged as the leading technique for sound event detection (SED) and sound source localization (SSL) [18,19]. Researchers have developed DNN-based approaches to jointly model the detection and localization of sound events, known as SELD, to gain a more complete understanding of environmental sounds [20,21]. In particular, polyphonic SELD has garnered substantial research attention due to its close resemblance to real-life scenarios. For example, Adavanne et al. [20] introduced SELDnet, which uses a shared backbone network to perform SED and direction of arrival (DOA) estimation. Cao et al. [22] proposed a two-stage strategy that utilizes separate networks for SED and DOA estimation and filters the DOA output based on the SED results. Despite the success, the fundamental interaction between SED and DOA estimation is not fully exploited. To address this issue, Cao et al. [11] later designed Event-Independent Network V2 (EINV2) with soft-parameter sharing between SED and DOA estimation to improve the interaction and the track-wise output format to handle homogeneous overlapping events. Hu et al. [9] further enhanced the performance of their model by incorporating dense blocks and replacing multi-head self-attention with conformer. As another direction for improvement, Shimada et al. [23] proposed a unified format called Activity-Coupled Cartesian Direction of Arrival (ACCDOA) that fuses SED and DOA estimation into a 3D Cartesian regression task to perform SELD.
In addition, effective feature engineering has been studied recently for the SELD task. For example, Nguyen et al. [24] developed the Spatial Cue-Augmented Log-Spectrogram (SALSA) to enhance polyphonic SELD by accurately mapping signal power with source directional cues for MIC and FOA formats. They then proposed SALSA-Lite [25], which is a computationally optimized variant for the MIC format that employs normalized cross-channel phase differences to enhance efficiency and preserve performance. Rosero et al. [26] improved SELD feature extraction with bio-inspired Gammatone filters and utilized a temporal convolutional network to better capture long-term dependencies compared to previous works. Huang et al. [27] transformed audio signals into graph representations and developed a graph convolutional network to simultaneously learn the spatio-temporal features of sound events.
Distinct from these existing works, our spatio-temporal-frequency fusion network employs a novel model architecture that accurately characterizes sound events by jointly capturing comprehensive features across spatial, temporal, and frequency domains in multichannel audio recordings. Consequently, our model significantly outperforms SOTA methods on several benchmark datasets.

2.2. Attention Mechanisms

The attention mechanism can be viewed as a dynamic selection process that adaptively weights features according to their importance. This approach has proven beneficial in numerous computer vision tasks [28]. One successful example is the Squeeze-and-Excitation (SE) attention module, which establishes interdependencies between different feature channels [14]. The following works further enhance their performance by incorporating long-range dependencies and by extracting detailed contextual information through more advanced modules, such as the Global Context (GC) module and the Convolutional Block Attention Module (CBAM) [16,17]. However, these methods generate 1D or 2D weights that uniformly treat neurons across spatial or channel dimensions of the image features, as shown in Figure 2a,b. This uniformity restricts their ability to extract more discriminative information from volumetric data that contain multiple dimensions.
In subsequent research endeavors, multidimensional dependencies are simultaneously captured to facilitate feature learning. For example, the residual attention network [29] learns 3D weights by integrating sub-networks into its different layers. However, the network’s scalability is limited by heightened computational complexity and resource demands. In contrast, SimAM [30], as depicted in Figure 2c, is a simple, parameter-free attention mechanism that directly generates 3D weights without requiring extensive structural tuning of the original networks. These 3D attention maps align with the temporal, frequency, and spatial dimensions of sound sources, enabling SimAM to enhance the 3D feature volume of sound events effectively while maintaining a low computational cost. Motivated by these existing works, our approach leverages SimAM to refine the spatio-temporal-frequency features extracted by the 3D convolutional network, allowing us to acquire precise and detailed characterizations of sound events recorded in multichannel audio signals.

3. Methodology

In this section, we first introduce our newly developed SELD network and then present its related components, such as the spatio-temporal-frequency volume, the Enhanced-3D residual block, and the class- and track-wise output format.

3.1. Spatio-Temporal-Frequency Fusion Network

The spatial, temporal, and frequency properties of sound sources are usually interdependent, necessitating a comprehensive analysis across these domains to accurately characterize sound events in multichannel audio recordings. This analysis enables the establishment of correlations between sound event classes and their corresponding locations, which is crucial for resolving overlapping events. To this end, we propose a novel STFF-Net model to improve the performance of polyphonic SELD, as illustrated in Figure 3. This model integrates detection and localization of sound events within a unified framework, which consists of a customized backbone network, basic bi-directional gated recurrent units (BiGRUs), and a multi-ACCDOA output format.
Specifically, the STFF-Net model begins by receiving a spatio-temporal-frequency volume with dimension C × D × T × F , where D is the depth defined by the number of feature channels, T is the number of time frames, and F is the number of frequency bins. Because each audio feature map is represented as a grayscale image with a single-channel intensity, we initialize the channel dimension C to 1. The volume is then passed through two 3D convolution layers and a 3D average pooling layer to decrease the resolution of the feature maps while preserving their key characteristics. Next, we deploy a sequence of Enhanced-3D residual blocks in the network to extract specialized representations, as shown on the right side of Figure 3. In particular, each Enhanced-3D residual block consists of two 3D convolutional layers enhanced with SimAM, allowing the network to focus on significant sound events. We set the stride size of the first convolutional layer to 1 or 2. Skip connections are utilized for convolutional layers to ensure that the gradient can flow directly within the entire network. This design helps the model converge fast while simultaneously reducing the probability of overfitting during the training process [31]. Later, we pool the extracted high-level features along the frequency axis, but the other dimensions are kept fixed. To capture temporal context information in both forward and backward directions, we reshape the outputs and then feed them into two BiGRU layers with 256 units [32]. Our network ends with fully connected (FC) layers, in which the first layer includes 256 units and the second layer includes n tracks × n classes × 3 . After further upsampling to match the target annotation’s resolution, we obtain the final shape of T / 6 , n tracks × n classes × 3 , which indicates a 6× reduction of the temporal dimension and provides ACCDOA vector coordinates for each target class per track.

3.2. Spatio-Temporal-Frequency Volume

The spatio-temporal-frequency volume is a 3D representation that encompasses the spatial, temporal, and frequency properties of sound sources in multichannel audio signals. This 3D tensor consists of depth, height, and width dimensions, where depth represents spatial characteristics determined by differences between audio channels, height reflects the evolution of sound sources over time, and width represents the spectral content of the audio signal. Typically, SED relies on time-frequency patterns to differentiate sound classes, while SSL uses amplitude and/or phase differences between microphones to estimate spatial locations. To carry out SELD tasks, we construct a volume that combines SED and SSL features. In detail, we extract multichannel log-linear spectrograms and frequency-normalized inter-channel phase differences (NIPDs) [25] for the MIC format and linear-scale intensity vectors (IVs) [33] for the FOA format. These features for each format are then sequentially stacked in an audio channel order to form a 3D feature volume aligned in all three dimensions, as illustrated in Figure 4a,b. The resulting volumes, referred to as SALSA-Lite for MIC and LinSpecIV for FOA, serve as initial inputs for our STFF-Net model.
Mathematically, the log-linear spectrogram of a spatial audio recording with M channels is defined as follows:
LinSpec ( t , f ) = log X 1 : M ( t , f ) 2 R M × T × F ,
where X i ( t , f ) is the short-time Fourier transform (STFT) of the audio signal from the i-th channel, t is the time frame index, f is the frequency index, and · denotes the element-wise complex modulus. To offer directional cues to the sound source, we compute the NIPD in SALSA-Lite as follows:
Λ ( t , f ) = v ( 2 π f ) 1 arg X 1 X 2 : M ( t , f ) R M 1 ,
where v is the sound velocity, and · is the conjugate transpose. Similarly, IV in LinSpecIV is obtained by:
I ( t , f ) = I X ( t , f ) I Y ( t , f ) I Z ( t , f ) = 1 ρ 0 v X W ( t , f ) X X ( t , f ) X Y ( t , f ) X Z ( t , f ) , I norm = I ( t , f ) I ( t , f ) 2 R 3 ,
where ρ 0 is the sound density, W is the omnidirectional component of the FOA format, and X , Y , and Z correspond to its directional components in the front–back, left–right, and up–down axes, respectively [33,34]. Additionally, · is the real part, and · 2 is the 2 norm of a vector. Table 1 outlines the SELD features used in this study to assess the performance of our chosen initial feature volumes.

3.3. Enhanced-3D Residual Block

Given that standard 1D or 2D convolution methods are unable to adequately capture the features of sound events across spatial, temporal, and frequency domains, we developed a new building module called the Enhanced-3D (E3D) residual block to replace 2D residual units within the residual learning framework. Formally, we define an E3D residual block as follows:
y = Concat E 1 F 1 ( x ) , E 2 F 2 ( x ) , , E C F C ( x ) + H ( x ) .
Here, x and y are the input and output tensors of the building block, respectively; F c ( x ) is the portion of the feature volumes generated after the 3D convolutional layers corresponding to the c-th channel; E c indicates the 3D attention weights for the feature volume F c ( x ) ; ⊙ denotes element-wise multiplication; Concat · denotes the concatenation operation along the feature channel dimension; H ( x ) is the shortcut connection that preserves the input information; and H ( x ) = x is an identity mapping.
The E3D residual block capitalizes on two key technologies. The first is the encoding of spatio-temporal-frequency information of sound events through 3D convolution and 3D pooling operations. 3D convolution [35] jointly captures event-level patterns across spatial, temporal, and frequency domains by convolving a 3D kernel with the feature volumes. Formally, the value of a unit at position ( d , t , f ) in the j-th feature volume in the l-th layer, denoted as z l j d t f , is given by:
z l j d t f = ReLU m α = 0 A 1 β = 0 B 1 γ = 0 Γ 1 w l j m p q r z ( l 1 ) m ( d + α ) ( t + β ) ( f + γ ) + b l j ,
where A , B , and Γ are the sizes of the kernel along the depth, height, and width dimension, b l j is the bias for the given feature volume, m indexes over the set of feature volumes in the ( l 1 ) -th layer connected to the current feature volume, and w l j m p q r is the value at position ( p , q , r ) of the kernel connected to the m-th feature volume in the preceding layer. As the kernel slides across the entire input volume, a complete 3D feature volume is produced.
The second key technology is the integration of a non-parametric attention module using SimAM [30] to generate 3D attention weights, which emphasize the informative and discriminative features of sound events simultaneously in three dimensions. These attention weights for feature volumes are determined based on the energy of the corresponding neuron. The energy function for each neuron n o is given by:
e o = 4 σ 2 + λ n o μ 2 + 2 σ 2 + 2 λ ,
where μ = 1 Δ o = 0 Δ 1 n o and σ 2 = 1 Δ o = 0 Δ 1 n o μ 2 are the mean and variance at a certain depth in a feature volume, Δ = T × F is the number of neurons at that depth, and λ is a regularization parameter. In fact, a neuron n o with lower energy e o is more distinct from the surrounding neurons, which makes it more crucial for feature processing. Therefore, the importance of each neuron is quantified as 1 / e o . For each feature volume, we can obtain 3D attention weights by aggregating 1 / e o over depth, height, and width dimensions and then applying a sigmoid function to constrain overly large values without changing the relative relevance of individual neurons. The attention module is straightforward and intuitive, with all computations being element-wise except for the mean μ and variance σ 2 calculations along the depth dimension of the feature volumes, as shown in Algorithm 1. In essence, the whole refinement phase is analogous to beamforming for each feature volume, but it does not increase network parameters.
Algorithm 1 PyTorch-like Implementation of a Non-parametric Attention Module
Input:
a batch of extracted feature volumes Z R B × C × D × T × F , coefficient λ as defined in Equation (6)
Output:
the refined feature volumes Z ^ R B × C × D × T × F
1:
def forward (Z, lambda):
2:
    # time-frequency size
3:
    Δ = Z.shape[3] ∗ Z.shape[4] − 1
4:
    # square of (n − u)
5:
    d = (Z − Z.mean(dim=[3, 4])).pow(2)
6:
    # d.sum() / n is depth variance of feature volumes
7:
    v = d.sum(dim=[3, 4]) / Δ
8:
    # E_inv groups all importance of feature volumes
9:
    E_inv = d / (4 ∗ (v + lambda)) + 0.5
10:
  # return the attention-weighted feature volumes
11:
  Return Z ∗ sigmoid(E_inv)

3.4. Class- and Track-Wise Output Format

We adopt the multi-ACCDOA format for the model outputs, which merges the activity and DOA of each sound event into a single vector and solves cases with homogeneous overlap in polyphonic SELD. In addition, class-wise auxiliary duplicating permutation-invariant training (ADPIT) is utilized to provide consistent training targets and avoid interference from zero vectors [36]. Overall, the loss function of our STFF-Net model can be defined as follows:
L PIT = 1 S T s S t T min α Perm ( s t ) l α ACCDOA , l α ACCDOA = 1 N n N MSE P α , n , P ^ n ,
where N and S denote the number of tracks and sound classes, respectively. Given class s and frame t, α Perm ( s t ) denotes one possible class-wise frame-level permutation, P α , n denotes an ACCDOA target of a permutation α at track n, and P ^ n denotes the corresponding ACCDOA prediction. We set the parameter N to 3, as it is fairly common for up to three events to occur concurrently. During inference, if the magnitude of event vectors exceeds the preset threshold, we view them as active events and then choose the Cartesian DOAs accordingly.

4. Evaluation

In this section, we conduct extensive experiments on both synthetic and real spatial audio datasets to validate the effectiveness of our newly developed STFF-Net. For fair comparisons, we modify the STFF-Net by replacing the 3D convolution and pooling layers with 2D counterparts and by removing the attention module, creating a simpler baseline called CRNN. We also use consistent experimental settings to rigorously test performance.

4.1. Experimental Setup

4.1.1. Datasets

The models in this work are trained and evaluated using three benchmark datasets: TNSSE 2021 [10], STARSS 2022 [37], and L3DAS22 Task 2 [38]. Table 2 summarizes the main characteristics of these datasets.
TNSSE 2021 dataset: This synthetic spatial audio dataset was released for Task 3 of the DCASE 2021 Challenge. It includes a public set of 400, 100, and 100 one-minute audio recordings for training, validation, and testing, respectively, as well as a private set of 200 one-minute recordings for evaluation. These recordings are crafted by convolving dry audio signals of sound events with multichannel room impulse responses (RIRs) with DOA labels and then adding spatial ambient noise. To replicate real-world conditions, the RIR and noise samples are collected from 15 distinct interior locations, with a balanced distribution of static and moving sound events across the range of azimuth 180 , 180 and elevation 45 , 45 . Additional directional interference from static or moving non-target sound sources is also included in the dataset. Due to its accessibility, well-defined partitions, and relative complexity, this dataset is used in the majority of our experiments.
STARSS 2022 dataset: This dataset marks a significant advancement over the above dataset by utilizing real-world audio recordings with spatiotemporal labels, which are acquired in real time from motion trackers. The dataset consists of audio recordings ranging from 30 s to 6 min, divided into sections for development and evaluation. The development set comprises 67 training clips and 54 testing clips, with a total duration of approximately 4 h 52 min. Audio data collection was conducted in 11 different rooms located in Tokyo and Tampere, where sound events were recorded within azimuth and elevation ranges of 180 , 180 and 90 , 90 , respectively. The diversity and density of these events vary greatly, providing a wide range of natural sound distributions. Despite the presence of four or five overlapping events, they only account for 1.8% of the entire dataset.
L3DAS22 Task 2 dataset: This synthetic dataset was provided in the L3DAS22 Challenge for 3D SELD. Unlike the above challenges, this one requires the estimation of exact locations of sound events. The dataset comprises 7.5 h of multiple-source and multiple-perspective (MSMP) Ambisonics recordings, each of which lasts 30 s . These recordings are divided into 600 recordings for training, 150 recordings for validation, and 150 recordings for testing. Each subset includes an equal number of files that correspond to one, two, and three overlapping sound events. The RIRs are collected from 252 fixed spatial points in an office room that measures about 6 meters in length, 5 m in width, and 3 m in height. Two FOA microphones are positioned centrally within the room.

4.1.2. Data Augmentation

Due to the limited number of samples in both synthetic and real spatial audio datasets, we utilize three data augmentation strategies to enhance data diversity and improve model generalization, including frequency shifting [24], hybrid cutout [39,40], and spatial augmentation [41,42]. Note that only spatial augmentation alters the ground truth, while frequency shifting and hybrid cutout do not.
Frequency shifting is used to mitigate the impact of noise and simulate different microphone distances by changing the frequency content of the audio signal. To do this, we randomly shift the input feature volumes upwards or downwards along the frequency axis by up to 10 bands. Hybrid cutout consists of two operations for preventing overfitting, namely, random cutout and TF masking. Random cutout produces a rectangular mask on the input feature volumes, while TF masking produces a cross-shaped mask. Each operation has an equal probability of occurring, with only one being executed at a time. In addition, spatial augmentation is performed by swapping channels in the input feature volume for MIC and FOA formats, which support 16 and 8 audio channel swapping methods, respectively. To ensure consistency, target labels are adjusted accordingly. This technique increases the variety of spatial locations of sound events within the spatial audio dataset.

4.1.3. Evaluation Metrics

To evaluate SELD performance, we adopt the official metrics introduced in the DCASE and L3DAS22 Challenges [21,43]. These metrics provide a joint measure of localization and detection, including location-dependent F1-score F τ , error rate E R τ , class-dependent localization recall L R CD , and localization error L E CD . Here, τ denotes the spatial threshold, which is set at 20 for DCASE and 2 m for L3DAS22, respectively, while F τ and E R τ consider true positive predictions that are less than a spatial threshold τ from the ground truth. Additionally, L R CD and L E CD only calculate localization predictions when the detected sound class is correct. We also compute an aggregated SELD error metric as a unified measure for the overall evaluation, defined as follows:
E SELD = 1 4 E R τ + 1 F τ + L E CD 180 + 1 L R CD ,
where E SELD is used to select the model and its hyperparameter. An optimal SELD system is characterized by low E R τ , high F τ , low L E CD , high L R CD , and a low overall E SELD metric.
Moreover, we utilize both micro- and macro-averaged evaluation metrics, unlike most previous studies that have used either one or the other. This dual approach allows for a more comprehensive and balanced evaluation by considering both the influence of classes with larger sample sizes and the performance of smaller classes.

4.1.4. Parameter Configuration

We follow a standard training pipeline for all models. Specifically, we use a sampling rate of 24 kHz, Hann window of 1024 samples, hop length of 400 samples, and 1024 FFT points, resulting in the input frame rate of 60 fps. The model temporally downsamples the input by a factor of 12, and the final outputs are temporally upsampled by a factor of 2 to align with the label frame rate of 10 fps. Given that the array response is almost frequency-independent up to about 9 kHz in the DCASE datasets, frequency bands above 9 kHz, corresponding to frequency bin indexes 384 and above, are linearly compressed by a factor of 8. As such, the frequency dimension of all linear-scale features is F = 400 . Unless stated otherwise, initial feature volumes are normalized globally to zero mean and unit standard deviation per channel. We use audio chunks of 8 s in length for model training. The hyperparameter λ in Equation (6) for the attention module is set to 1 × 10 4 , selected through hyperparameter search. A detailed analysis of λ is presented in the following subsection. We employ the Adam optimizer [44] for training. The learning rate is initially set at 3 × 10 4 and is linearly reduced to 1 × 10 5 over the final 35 epochs of the total 70 training epochs. A threshold of 0.4 is used to binarize active class predictions per track in the model outputs.

4.2. Experimental Results

4.2.1. Choice of SELD Features

For a rational comparison of the performance of different input features, we implement the SELD network architecture described in [24] with its default settings. Note that we utilize the multi-ACCDOA format rather than the SELDnet format to output overlapping sound events from the same class. Table 3 shows the SELD performance of all considered features with data augmentation. The comprehensive metric E SELD simplifies the evaluation of overall performance. The directional arrows in the table indicate whether an increase or decrease in a metric value signifies improvement.
As expected, features for FOA generally surpass those for MIC in most evaluation metrics. This suggests that the FOA format, with its built-in spatial audio encoding, is more appropriate for the SELD task. In detail, SALSA, SALSA-Lite, and IV-based features consistently perform better than GCC-based ones, underscoring the importance of maintaining TF alignment between feature maps. This alignment enables the network to effectively learn and discern multichannel event-level patterns in image-like input features. LinSpecIV exhibits superior performance compared to MelSpecIV, indicating that the use of a linear-scale may provide additional performance benefits. The performance of SALSA is relatively subpar for each format, most likely due to its focus on single-source TF bins, which results in the exclusion of important information from multisource TF bins. Among the features, SALSA-Lite achieves the lowest E SELD value with micro- and macro-averages of 0.332 and 0.343 in the MIC format, as highlighted in Table 3. Likewise, LinSpecIV achieves the lowest E SELD value with micro- and macro-averages of 0.331 and 0.335 in the FOA format. Both features take advantage of TF alignment and linear scale, making them ideal input features for subsequent experiments in their respective formats.

4.2.2. Effect of Attention Modules

We evaluate SimAM on the SELD task against three representative attention modules (SE [14], CBAM [16], and GC [17]) and compare the performance of its enhanced CRNN with SOTA SELD systems using the L3DAS22 Task 2 dataset [38]. To adhere to the criteria of the L3DAS22 Challenge, we adapt the output format of CRNN to the track-wise format [11], which is necessary for estimating the exact locations of sound events. The performance is assessed using official evaluation metrics, and the results are shown in Table 4. The majority of the attention modules are capable of improving the performance of CRNN, in which SimAM stands out by achieving the highest F1-score without adding network parameters. In addition, SimAM-enhanced CRNN significantly outperforms the official baseline [38] across all evaluation metrics and even surpasses the second-ranked ensemble model [45] in the Challenge. Although it slightly underperformed compared to the top-ranked ensemble model [9], which benefited from training on both the L3DAS21 and L3DAS22 datasets with two FOA microphone arrays, the overall performance is still impressive. The results of these studies indicate that SimAM generates true 3D attention weights that effectively and efficiently enhance the representation of sound events in all three dimensions.

4.2.3. Analysis of λ in the Attention Module

Figure 5 shows our search results for λ using CRNN on LinSpecIV features from the TNSSE 2021 development dataset. Given the impracticality of verifying all possible real values of λ , we restrict our search to a predetermined set ranging from 10 6 to 10 1 , as shown on the x-axis of the figure. We repeat the process three times for each candidate λ value and report the means and standard deviations of the micro-averaged and macro-averaged E SELD values. The results clearly indicate that SimAM significantly improves the performance of CRNN across the evaluated λ values, with λ = 10 4 emerging as the optimal value for superior SELD performance. Therefore, we select λ = 10 4 for implementation in our method.

4.2.4. Experiments on a Synthetic Spatial Audio Dataset

Table 5 illustrates the performance of our newly developed STFF-Net and its control models for both MIC and FOA formats, as well as SOTA methods, on the test split of the TNSSE 2021 dataset [10]. The evaluations use both micro-average and macro-average metrics, the former for fair comparison with prior works and the latter for ease of comparison with future works. In terms of micro-average metrics, our STFF-Net outperforms the DCASE baseline SELDnet [10] by a substantial margin and surpasses several SOTA systems for the FOA format. In particular, STFF-Net exhibits a significant increase in F 20 and L R CD by 23.8% and 18.1%, respectively, alongside a decrease in E R 20 by 23.5% and in L E CD by 10.55 points compared to G-SELD [26]. Against HAAC-enhanced EINV2 [46], STFF-Net improves F 20 and L R CD by 7.7% and 4%, respectively, and also reduces E R 20 by 7.5% and L E CD by 3.31 points. Although marginally higher than the top-ranking ensemble model [7] from the DCASE 2021 Challenge in micro-averaged E SELD , STFF-Net achieves a higher L R CD and lower E R 20 with fewer than half the parameters.
Our ablation studies, presented in the last two major rows of Table 5, quantify the effect of each module on SELD performance. In general, FOA models consistently outperform MIC models. The integration of SimAM significantly enhances the performance of CRNN over CBAM for both formats, and SimAM-enhanced CRNN achieves the best SELD results. This demonstrates that SimAM directly generates 3D attention weights to capture subtleties that may be neglected by CBAM’s two-step procedure, enabling more precise attention adjustment for target events without increasing network complexity. Furthermore, the SimAM-enhanced CRNN transforms into STFF-Net by substituting 2D convolutional modules with 3D ones. STFF-Net obtains the best macro-averaged F 20 and L E CD among control models for the FOA format and the best L R CD among control models for the MIC format, but the overall performance is slightly inferior to SimAM-enhanced CRNN. One potential explanation is that 3D convolutions incorporate an additional dimension to capture the evolving spatial distribution of sound events, thereby learning comprehensive features of sound events. However, the risk of overfitting is exacerbated by the complexity introduced by the additional dimension, particularly when the event categories are varied or the data are limited. Therefore, it is advisable to conduct additional assessments of STFF-Net with larger datasets.
Additionally, the corresponding class-wise performance analysis for the FOA format, shown in Figure 6, offers insight into the strengths and weaknesses of the systems with respect to each sound class. The F 20 and L E CD metrics for half of the target classes, including Alarm, Crying Baby, and Female Speech, are significantly improved by the proposed STFF-Net compared to CRNN. These improvements collectively elevate the overall system performance. In contrast, the classes Footsteps and Knocking on Door present more of a challenge, as demonstrated by their lower F 20 and higher L E CD .

4.2.5. Experiments on a Real Spatial Audio Dataset

Table 6 illustrates the performance of our STFF-Net and its control models for both MIC and FOA formats, as well as SOTA methods, on the test split of the STARSS 2022 dataset [37]. The models are trained using the training set of the STARSS 2022 dataset and augmented with synthesized simulation data from DCASE 2021 [10] to compensate for the limited duration of the real dataset. Despite the presence of challenging real-world conditions, STFF-Net still considerably outperforms the official baseline SELDnet across all metrics for both MIC and FOA formats. When compared to SOTA systems such as SwG-former [27], SwG-EINV2 [27], and HAAC-enhanced EINV2 [46], STFF-Net shows marked improvements in performance, as indicated by the lowest macro-averaged E SELD and the best location-dependent SED metrics E R 20 and F 20 , although L E CD is slightly higher than HAAC-enhanced EINV2, and L R CD is slightly lower than SwG-EINV2. In addition, STFF-Net is more compact in terms of model complexity than these competing models. Notably, STFF-Net for the FOA format improves macro-averaged E SELD by 2.5% over the second-ranked ensemble model from the DCASE 2022 Challenge, while requiring fewer than one-third of the parameters. These results demonstrate that our proposed STFF-Net effectively learns comprehensive features of sound events to perform polyphonic SELD.
In the ablation study, there remains a large performance gap between the FOA and MIC models when applied in real-world audio contexts. SimAM enhances the performance of CRNN for both formats with a distinct advantage over CBAM. The SimAM-enhanced CRNN is further improved by incorporating 3D convolution modules instead of the original 2D ones, resulting in the STFF-Net. For the FOA format, STFF-Net performs better than the control models, as evidenced by a micro-averaged E SELD of 0.309 and a macro-averaged E SELD of 0.356. These improvements underscore the importance of comprehensive feature extraction of sound events in more varied and complex real-world environments. However, for the MIC format, STFF-Net shows a slight decline in performance. This decline is likely due to the necessity of a lower cutoff frequency in the MIC format to avoid spatial aliasing, which results in the loss of specific sound event information. Overall, our STFF-Net integrates all the proposed improvements.
Figure 7 shows the class-wise performance analysis of STFF-Net and control models for the FOA format on real spatial data. For most target classes, STFF-Net is superior to CRNN in both F 20 and L E CD metrics, leading to better overall performance. Interestingly, the Footsteps and Knock classes show significant improvements, which differ from the findings obtained using the TNSEE 2021 dataset. However, the Faucet and Door classes exhibit a decline in performance, as evidenced by their respective E SELD of 0.536 and 0.507, which are notably greater than the mean E SELD . One possible explanation for the Faucet class’s poor performance is the occurrence of false negatives caused by directional interference from other classes, such as Dishes, Pots, and Pans. Another consideration is that the Door class is rather sparse, comprising just 0.6% of all frames, which might affect the model’s ability to effectively learn this class.
To observe the transformation of features through the various stages of our STFF-Net, we utilize t-SNE [48] to visualize 3 s audio clips sampled every 3 min from the test set of the STARSS 2022 dataset. As shown in Figure 8, the scatter plot of the input features displays somewhat dispersed clustering of the same sound classes, suggesting that intrinsic patterns may exist within the spatio-temporal-frequency volume. As data flow through the backbone network, there is a discernible improvement in the segregation of different classes with reduced overlap, indicating that the backbone effectively extracts more pertinent features. Further processing through the RNN layers, which capture the temporal dependencies, results in sound classes becoming clearly distinguishable and tightly clustered. In the final FC output, the target classes are clearly separated with minimal overlap, reflecting the model’s ability to capture subtle differences between complex sound events. This separation underscores the effectiveness of the refined features, which are highly discriminative and well-suited for the SELD task.

5. Conclusions

In conclusion, this paper introduces a novel framework for polyphonic SELD that utilizes the newly developed Spatio-Temporal-Frequency Fusion Network (STFF-Net) to jointly learn comprehensive features of sound events from multichannel audio recordings. The Enhanced-3D (E3D) building block in our STFF-Net captures and refines intricate correlations across spatial, temporal, and frequency domains by combining 3D convolutions with non-parametric attention mechanisms. Additionally, we incorporate the multi-ACCDOA format to handle overlapping events from the same class. Our rigorous experimental evaluations on a range of benchmark datasets, both synthetic and real-world, demonstrate the effectiveness and generalization of the proposed STFF-Net. Significant enhancements in performance are readily apparent when compared to other SOTA methods. We hope this innovative approach to jointly learning comprehensive features of sound events will drive future advancements in acoustic scene analysis.

Author Contributions

Conceptualization, B.C. and M.W.; methodology, B.C. and Y.G.; software, B.C.; validation, B.C. and Y.G.; formal analysis, B.C. and Y.G.; investigation, B.C.; data curation, B.C.; writing—original draft preparation, B.C. and Y.G.; writing—review and editing, B.C. and Y.G.; visualization, B.C.; supervision, M.W.; project administration, M.W.; funding acquisition, M.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported by the National Natural Science Foundation of China under Grant 62071135.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Data are contained within the article.

Acknowledgments

The authors would like to thank the creators of the papers [10,37,38] for publicly providing spatial audio datasets. The authors also would like to thank the anonymous reviewers for their valuable comments and suggestions in revising this paper.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Foggia, P.; Petkov, N.; Saggese, A.; Strisciuglio, N.; Vento, M. Audio surveillance of roads: A system for detecting anomalous sounds. IEEE Trans. Intell. Transp. Syst. 2015, 17, 279–288. [Google Scholar] [CrossRef]
  2. Despotovic, V.; Pocta, P.; Zgank, A. Audio-based Active and Assisted Living: A review of selected applications and future trends. Comput. Biol. Med. 2022, 149, 106027. [Google Scholar] [CrossRef] [PubMed]
  3. Stowell, D.; Wood, M.D.; Pamuła, H.; Stylianou, Y.; Glotin, H. Automatic acoustic detection of birds through deep learning: The first bird audio detection challenge. Methods Ecol. Evol. 2019, 10, 368–380. [Google Scholar] [CrossRef]
  4. Elizalde, B.; Zarar, S.; Raj, B. Cross modal audio search and retrieval with joint embeddings based on text and audio. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 4095–4099. [Google Scholar]
  5. Zürn, J.; Burgard, W. Self-supervised moving vehicle detection from audio-visual cues. IEEE Robot. Autom. Lett. 2022, 7, 7415–7422. [Google Scholar] [CrossRef]
  6. He, Y.; Trigoni, N.; Markham, A. SoundDet: Polyphonic moving sound event detection and localization from raw waveform. In Proceedings of the International Conference on Machine Learning (ICML), Virtual Event, 18–24 July 2021; pp. 4160–4170. [Google Scholar]
  7. Shimada, K.; Takahashi, N.; Koyama, Y.; Takahashi, S.; Tsunoo, E.; Takahashi, M.; Mitsufuji, Y. Ensemble of ACCDOA-and EINV2-Based Systems with D3Nets and Impulse Response Simulation for Sound Event Localization and Detection. Technical Report. 2021. Available online: https://dcase.community/documents/challenge2021/technical_reports/DCASE2021_Shimada_117_t3.pdf (accessed on 17 September 2024).
  8. Wang, Q.; Chai, L.; Wu, H.; Nian, Z.; Niu, S.; Zheng, S.; Wang, Y.; Sun, L.; Fang, Y.; Pan, J.; et al. The NERC-SLIP System for Sound Event Localization and Detection of DCASE2022 Challenge; Technical Report. 2022. Available online: https://dcase.community/documents/challenge2022/technical_reports/DCASE2022_Du_122_t3.pdf (accessed on 17 September 2024).
  9. Hu, J.; Cao, Y.; Wu, M.; Kong, Q.; Yang, F.; Plumbley, M.D.; Yang, J. A track-wise ensemble event independent network for polyphonic sound event localization and detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 9196–9200. [Google Scholar]
  10. Politis, A.; Adavanne, S.; Krause, D.; Deleforge, A.; Srivastava, P.; Virtanen, T. A dataset of dynamic reverberant sound scenes with directional interferers for sound event localization and detection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2021 Workshop (DCASE2021), Barcelona, Spain, 15–19 November 2021; pp. 125–129. [Google Scholar]
  11. Cao, Y.; Iqbal, T.; Kong, Q.; An, F.; Wang, W.; Plumbley, M.D. An improved event-independent network for polyphonic sound event localization and detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 885–889. [Google Scholar]
  12. Xie, R.; Shi, C.; Le Zhang, Y.L.; Li, H. Ensemble of Attention Based CRNN for Sound Event Detection and Localization. Technical Report. 2022. Available online: https://dcase.community/documents/challenge2022/technical_reports/DCASE2022_Xie_18_t3.pdf (accessed on 17 September 2024).
  13. Kim, J.S.; Park, H.J.; Shin, W.; Han, S.W. A Robust Framework for Sound Event Localization and Detection on Real Recordings. Technical Report. 2022. Available online: https://dcase.community/documents/challenge2022/technical_reports/DCASE2022_Han_54_t3.pdf (accessed on 17 September 2024).
  14. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  15. Hou, Q.; Zhou, D.; Feng, J. Coordinate attention for efficient mobile network design. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 13713–13722. [Google Scholar]
  16. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European conference on computer vision (ECCV), Munich, Germany, 8–14 September 2018; pp. 3–19. [Google Scholar]
  17. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. Global context networks. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 45, 6881–6895. [Google Scholar] [CrossRef] [PubMed]
  18. Mesaros, A.; Heittola, T.; Virtanen, T.; Plumbley, M.D. Sound event detection: A tutorial. IEEE Signal Process. Mag. 2021, 38, 67–83. [Google Scholar] [CrossRef]
  19. Grumiaux, P.A.; Kitić, S.; Girin, L.; Guérin, A. A survey of sound source localization with deep learning methods. J. Acoust. Soc. Am. 2022, 152, 107–151. [Google Scholar] [CrossRef] [PubMed]
  20. Adavanne, S.; Politis, A.; Nikunen, J.; Virtanen, T. Sound event localization and detection of overlapping sources using convolutional recurrent neural networks. IEEE J. Sel. Top. Signal Process. 2019, 13, 34–48. [Google Scholar] [CrossRef]
  21. Politis, A.; Mesaros, A.; Adavanne, S.; Heittola, T.; Virtanen, T. Overview and evaluation of sound event localization and detection in DCASE 2019. IEEE/ACM Trans. Audio, Speech, Lang. Process. 2020, 29, 684–698. [Google Scholar] [CrossRef]
  22. Cao, Y.; Kong, Q.; Iqbal, T.; An, F.; Wang, W.; Plumbley, M.D. Polyphonic sound event detection and localization using a two-stage strategy. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019), New York, NY, USA, 25–26 October 2019; pp. 30–34. [Google Scholar]
  23. Shimada, K.; Koyama, Y.; Takahashi, N.; Takahashi, S.; Mitsufuji, Y. ACCDOA: Activity-coupled cartesian direction of arrival representation for sound event localization and detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 915–919. [Google Scholar]
  24. Nguyen, T.N.T.; Watcharasupat, K.N.; Nguyen, N.K.; Jones, D.L.; Gan, W.S. Salsa: Spatial cue-augmented log-spectrogram features for polyphonic sound event localization and detection. IEEE/ACM Trans. Audio, Speech, Lang. Process. 2022, 30, 1749–1762. [Google Scholar] [CrossRef]
  25. Nguyen, T.N.T.; Jones, D.L.; Watcharasupat, K.N.; Phan, H.; Gan, W.S. SALSA-Lite: A fast and effective feature for polyphonic sound event localization and detection with microphone arrays. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 716–720. [Google Scholar]
  26. Rosero, K.; Grijalva, F.; Masiero, B. Sound events localization and detection using bio-inspired gammatone filters and temporal convolutional neural networks. IEEE/ACM Trans. Audio, Speech, Lang. Process. 2023, 31, 2314–2324. [Google Scholar]
  27. Huang, W.; Huang, Q.; Ma, L.; Chen, Z.; Wang, C. SwG-former: Sliding-window Graph Convolutional Network Integrated with Conformer for Sound Event Localization and Detection. arXiv 2023, arXiv:2310.14016. [Google Scholar]
  28. Guo, M.; Xu, T.; Liu, J.; Liu, Z.; Jiang, P.; Mu, T.; Zhang, S.; Martin, R.; Cheng, M.; Hu, S. Attention mechanisms in computer vision: A survey. Comput. Vis. Media. 2022, 8, 331–368. [Google Scholar] [CrossRef]
  29. Wang, F.; Jiang, M.; Qian, C.; Yang, S.; Li, C.; Zhang, H.; Wang, X.; Tang, X. Residual attention network for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 3156–3164. [Google Scholar]
  30. Yang, L.; Zhang, R.Y.; Li, L.; Xie, X. Simam: A simple, parameter-free attention module for convolutional neural networks. In Proceedings of the International Conference on Machine Learning (ICML), Virtual Event, 18–24 July 2021; pp. 11863–11874. [Google Scholar]
  31. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  32. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. In Proceedings of the Advances in Neural Information Processing Systems (NIPS), Montreal, QC, Canada, 8–13 December 2014; pp. 1–9. [Google Scholar]
  33. Cao, Y.; Iqbal, T.; Kong, Q.; Zhong, Y.; Wang, W.; Plumbley, M.D. Event-independent network for polyphonic sound event localization and detection. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2020 Workshop (DCASE2020), Tokyo, Japan, 2–4 November 2020; pp. 11–15. [Google Scholar]
  34. Zhao, S.; Saluev, T.; Jones, D.L. Underdetermined direction of arrival estimation using acoustic vector sensor. Signal Process. 2014, 100, 160–168. [Google Scholar] [CrossRef]
  35. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [PubMed]
  36. Shimada, K.; Koyama, Y.; Takahashi, S.; Takahashi, N.; Tsunoo, E.; Mitsufuji, Y. Multi-accdoa: Localizing and detecting overlapping sounds from the same class with auxiliary duplicating permutation invariant training. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 316–320. [Google Scholar]
  37. Politis, A.; Shimada, K.; Sudarsanam, P.; Adavanne, S.; Krause, D.; Koyama, Y.; Takahashi, N.; Takahashi, S.; Mitsufuji, Y.; Virtanen, T. STARSS22: A dataset of spatial recordings of real scenes with spatiotemporal annotations of sound events. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2022 Workshop (DCASE2022), Nancy, France, 3–4 November 2022; pp. 125–129. [Google Scholar]
  38. Guizzo, E.; Marinoni, C.; Pennese, M.; Ren, X.; Zheng, X.; Zhang, C.; Masiero, B.; Uncini, A.; Comminiello, D. L3DAS22 challenge: Learning 3D audio sources in a real office environment. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 9186–9190. [Google Scholar]
  39. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; pp. 13001–13008. [Google Scholar]
  40. Park, D.S.; Chan, W.; Zhang, Y.; Chiu, C.C.; Zoph, B.; Cubuk, E.D.; Le, Q.V. SpecAugment: A simple data augmentation method for automatic speech recognition. In Proceedings of the International Speech Communication Association (ISCA), Graz, Austria, 15–19 September 2019; pp. 2613–2617. [Google Scholar]
  41. Mazzon, L.; Koizumi, Y.; Yasuda, M.; Harada, N. First order ambisonics domain spatial augmentation for DNN-based direction of arrival estimation. In Proceedings of the Detection and Classification of Acoustic Scenes and Events 2019 Workshop (DCASE2019), New York, NY, USA, 25–26 October 2019; pp. 154–158. [Google Scholar]
  42. Wang, Q.; Du, J.; Wu, H.X.; Pan, J.; Ma, F.; Lee, C.H. A four-stage data augmentation approach to ResNet-Conformer based acoustic modeling for sound event localization and detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2023, 31, 1251–1264. [Google Scholar] [CrossRef]
  43. Mesaros, A.; Adavanne, S.; Politis, A.; Heittola, T.; Virtanen, T. Joint measurement of localization and detection of sound events. In Proceedings of the IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (WASPAA), New Paltz, NY, USA, 20–23 October 2019; pp. 333–337. [Google Scholar]
  44. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  45. Mao, Y.; Zeng, Y.; Liu, H.; Zhu, W.; Zhou, Y. ICASSP 2022 L3DAS22 Challenge: Ensemble of Resnet-Conformers with Ambisonics Data Augmentation for Sound Event Localization and Detection. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 22–27 May 2022; pp. 9191–9195. [Google Scholar]
  46. Wu, S.; Wang, Y.; Hu, Z.; Liu, J. HAAC: Hierarchical audio augmentation chain for ACCDOA described sound event localization and detection. Appl. Acoust. 2023, 211, 109541. [Google Scholar] [CrossRef]
  47. Hu, J.; Cao, Y.; Wu, M.; Kong, Q.; Yang, F.; Plumbley, M.D.; Yang, J. Sound Event Localization and Detection for Real Spatial Sound Scenes: Event-Independent Network and Data Augmentation Chains. Technical Report. 2022. Available online: https://dcase.community/documents/workshop2022/proceedings/DCASE2022Workshop_Hu_61.pdf (accessed on 17 September 2024).
  48. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. 2008, 9, 2579–2605. [Google Scholar]
Figure 1. SELD task illustration. Our objective is to detect and localize the sources of interest given a set of target event classes and a multichannel audio signal recorded in conditions that include multiple concurrently active target sources, ambient noises, and directional interferences.
Figure 1. SELD task illustration. Our objective is to detect and localize the sources of interest given a set of target event classes and a multichannel audio signal recorded in conditions that include multiple concurrently active target sources, ambient noises, and directional interferences.
Sensors 24 06090 g001
Figure 2. Comparative diagrams of different attention mechanisms. The 1D (a) and 2D (b) attention mechanisms generate weights that adjust channels or regions of the feature maps, respectively. In contrast, the 3D attention mechanism (c) generates weights capable of adjusting each individual point within the features. These weights are visually represented by colored bricks.
Figure 2. Comparative diagrams of different attention mechanisms. The 1D (a) and 2D (b) attention mechanisms generate weights that adjust channels or regions of the feature maps, respectively. In contrast, the 3D attention mechanism (c) generates weights capable of adjusting each individual point within the features. These weights are visually represented by colored bricks.
Sensors 24 06090 g002
Figure 3. Architecture of our STFF-Net model for polyphonic SELD. The network adapts to different input feature volumes in FOA and MIC format.
Figure 3. Architecture of our STFF-Net model for polyphonic SELD. The network adapts to different input feature volumes in FOA and MIC format.
Sensors 24 06090 g003
Figure 4. Spatio-temporal-frequency volumes for different audio formats. (a) Log-linear spectrograms and NIPDs for a four-channel MIC format audio clip, combined and aligned to create a 3D feature volume. (b) Log-linear spectrograms and IVs of an FOA format audio clip, combined and aligned to create a 3D feature volume.
Figure 4. Spatio-temporal-frequency volumes for different audio formats. (a) Log-linear spectrograms and NIPDs for a four-channel MIC format audio clip, combined and aligned to create a 3D feature volume. (b) Log-linear spectrograms and IVs of an FOA format audio clip, combined and aligned to create a 3D feature volume.
Sensors 24 06090 g004
Figure 5. Visualization of the λ search for SELD performance. The analysis is conducted using CRNN on the testing portion of the development dataset. Means and standard deviations are calculated based on three separate trials.
Figure 5. Visualization of the λ search for SELD performance. The analysis is conducted using CRNN on the testing portion of the development dataset. Means and standard deviations are calculated based on three separate trials.
Sensors 24 06090 g005
Figure 6. Class-wise performance comparisons for STFF-Net and control models on the test split of the TNSSE 2021 dataset. F1-score and localization error for both networks are presented on the left and right, respectively.
Figure 6. Class-wise performance comparisons for STFF-Net and control models on the test split of the TNSSE 2021 dataset. F1-score and localization error for both networks are presented on the left and right, respectively.
Sensors 24 06090 g006
Figure 7. Class-wise performance comparisons for STFF-Net and control models on the test split of the STARSS 2022 dataset. F1-score and localization errors for both networks are presented on the left and right, respectively.
Figure 7. Class-wise performance comparisons for STFF-Net and control models on the test split of the STARSS 2022 dataset. F1-score and localization errors for both networks are presented on the left and right, respectively.
Sensors 24 06090 g007
Figure 8. t-SNE visualization of input features, backbone outputs, RNN outputs, and FC outputs for selected test samples from the STARSS 2022 dataset. The colors denote different sound classes. Best viewed in color.
Figure 8. t-SNE visualization of input features, backbone outputs, RNN outputs, and FC outputs for selected test samples from the STARSS 2022 dataset. The colors denote different sound classes. Best viewed in color.
Sensors 24 06090 g008
Table 1. Comparative analysis of different SELD features.
Table 1. Comparative analysis of different SELD features.
Feature ComponentsFormatAbbreviation# Feature ChannelsScaleTF Alignment
Log-mel spectrogram, mel-scale GCC-PHATMICMelSpecGCC 4 + 6 mel
Log-linear spectrogram, linear-scale GCC-PHATMICLinSpecGCC 4 + 6 linear
Log-linear spectrogram, NIPDMICSALSA-Lite 4 + 3 linear
Log-linear spectrogram, normalized principal eigenvectorMIC/FOASALSA 4 + 3 linear
Log-mel spectrogram, mel-scale IVFOAMelSpecIV 4 + 3 mel
Log-linear spectrogram, linear-scale IVFOALinSpecIV 4 + 3 linear
The total number of feature channels is derived from four-channel audio recordings. GCC-PHAT stands for Generalized Cross-Correlation with Phase Transform.
Table 2. Characteristics of the TNSSE 2021, STARSS 2022, and L3DAS22 Task 2 datasets.
Table 2. Characteristics of the TNSSE 2021, STARSS 2022, and L3DAS22 Task 2 datasets.
CharacteristicsTNSSESTARSSL3DAS22
Spatial audio formatMIC, FOAMIC, FOAFOA
Sampling rate24 kHz24 kHz32 kHz
Type of spatial recordingssyntheticrealsynthetic
Moving sources
Ambient noise
Reverberation
Non-target interfering events
Maximum degree of polyphony353
Overlapping of same class events (%)highlowlow
Number of target sound classes121314
Table 3. SELD performances of different features with data augmentation.
Table 3. SELD performances of different features with data augmentation.
FeatureCutoff Freq.Micro-Average MetricsMacro-Average Metrics
ER 20 F 20 LE CD LR CD E SELD ER 20 F 20 LE CD LR CD E SELD
MIC format
MelSpecGCC12 kHz0.6160.46622.2°0.7200.3880.6160.43921.3°0.6850.402
LinSpecGCC12 kHz0.6300.44121.7°0.6810.4070.6300.41120.9°0.6350.425
MIC SALSA4 kHz0.5290.53217.3°0.6900.3510.5290.47318.5°0.6970.365
SALSA-Lite2 kHz0.5000.56715.5°0.6910.3320.5000.51516.6°0.7030.343
FOA format
MelSpecIV12 kHz0.4980.56415.8°0.6850.3340.4980.52216.2°0.7160.338
LinSpecIV12 kHz0.4880.57413.9°0.6670.3310.4880.52515.2°0.7080.335
FOA SALSA9 kHz0.5030.56015.4°0.6640.3410.5030.51316.4°0.6950.347
Table 4. Performance comparison of baseline model and its enhancements with attention modules versus SOTA SELD methods on the test split of the L3DAS22 Task 2 dataset.
Table 4. Performance comparison of baseline model and its enhancements with attention modules versus SOTA SELD methods on the test split of the L3DAS22 Task 2 dataset.
ApproachParams.FOA Num.PrecisionRecallF1-Score
(’22) revised SELDnet [38]7.0 Msingle0.4230.2890.343
(’22 #1) Hu et al. [9] †-double0.7060.6910.699
(’22 #2) Mao et al. [45] †37.8 Msingle0.6000.5840.592
CRNN13.6 Msingle0.7000.5860.638
CRNN + SE13.7 Msingle0.7100.5880.643
CRNN + CBAM13.7 Msingle0.7050.5850.640
CRNN + GC13.7 Msingle0.7020.5840.638
CRNN + SimAM13.6 Msingle0.7010.5950.644
The number following the ‘#’ symbol indicates the model’s rank in the challenge. The ‘†’ symbol denotes an ensemble model.
Table 5. Performance evaluation of SELD methods on the test split of the TNSSE 2021 dataset, including approach comparisons and module ablation studies.
Table 5. Performance evaluation of SELD methods on the test split of the TNSSE 2021 dataset, including approach comparisons and module ablation studies.
ApproachParametersFormatMicro-Average MetricsMacro-Average Metrics
ER 20 F 20 LE CD LR CD E SELD ER 20 F 20 LE CD LR CD E SELD
(’21) SELDnet [10]0.5 MMIC0.750.23430.6°0.3780.5770.75----
FOA0.730.30724.5°0.4050.5390.73----
(’21 #1) Shimada et al. [7] †42 MFOA0.430.69911.1°0.7320.2650.43----
G-SELD [26]-FOA0.650.43922.5°0.5590.4440.65----
HAAC-enhanced EINV2 [46]85 MFOA0.490.6015.26°0.700.3210.49----
CRNN13.7 MMIC0.4530.62414.24°0.7210.2970.4530.62713.70°0.7380.291
CRNN + CBAM13.8 MMIC0.4760.60515.04°0.7230.3080.4760.60214.49°0.7370.304
CRNN + SimAM13.7 MMIC0.4460.63313.96°0.7100.2950.4460.63313.11°0.7310.289
STFF-Net35.8 MMIC0.4940.59316.16°0.7530.3100.4940.60915.39°0.7550.304
CRNN13.7 MFOA0.4110.66012.67°0.7190.2750.4110.66012.03°0.7400.270
CRNN + CBAM13.8 MFOA0.4070.66312.93°0.6990.2790.4070.66412.34°0.7350.269
CRNN + SimAM13.7 MFOA0.3940.68212.12°0.7380.2600.3940.67511.52°0.7580.256
STFF-Net35.8 MFOA0.4150.67711.95°0.7400.2660.4150.68011.33°0.7560.261
The number following the ‘#’ symbol indicates the model’s rank in the challenge. The ‘†’ symbol denotes an ensemble model.
Table 6. Performance evaluation of SELD methods on the test split of the STARSS 2022 dataset, including approach comparisons and module ablation studies.
Table 6. Performance evaluation of SELD methods on the test split of the STARSS 2022 dataset, including approach comparisons and module ablation studies.
ApproachParametersFormatMicro-Average MetricsMacro-Average Metrics
ER 20 F 20 LE CD LR CD E SELD ER 20 F 20 LE CD LR CD E SELD
(’22) SELDnet [37]604 kMIC0.710.36---0.710.1832.2°0.470.560
FOA0.710.36---0.710.2129.3°0.460.551
(’22 #2) Hu et al. [47] †85 MFOA0.53----0.530.48117.8°0.6260.381
HAAC-enhanced EINV2 [46]85 MFOA0.54----0.540.4517.20°0.620.391
SwG-former [27]110.8 MFOA0.64----0.640.45224.5°0.6570.416
SwG-EINV2 [27]288.5 MFOA0.63----0.630.48920.9°0.7180.385
CRNN13.7 MMIC0.5660.48321.62°0.7560.3620.5660.41721.38°0.6210.412
CRNN + CBAM13.8 MMIC0.5850.47121.46°0.7440.3720.5850.41320.52°0.6320.413
CRNN + SimAM13.7 MMIC0.5450.51519.27°0.7370.3500.5450.43918.77°0.5980.403
STFF-Net35.8 MMIC0.6130.44723.31°0.7930.3760.6130.39321.78°0.6310.428
CRNN13.7 MFOA0.5470.52119.69°0.7920.3360.5470.41718.50°0.6580.380
CRNN + CBAM13.8 MFOA0.5260.53020.54°0.7630.3370.5260.44519.14°0.6300.389
CRNN + SimAM13.7 MFOA0.5310.53219.33°0.7710.3340.5310.48117.31°0.6540.373
STFF-Net35.8 MFOA0.4990.57018.00°0.7940.3090.4990.51117.96°0.6630.356
The number following the ‘#’ symbol indicates the model’s rank in the challenge. The ‘†’ symbol denotes an ensemble model.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, B.; Wang, M.; Gu, Y. Joint Spatio-Temporal-Frequency Representation Learning for Improved Sound Event Localization and Detection. Sensors 2024, 24, 6090. https://doi.org/10.3390/s24186090

AMA Style

Chen B, Wang M, Gu Y. Joint Spatio-Temporal-Frequency Representation Learning for Improved Sound Event Localization and Detection. Sensors. 2024; 24(18):6090. https://doi.org/10.3390/s24186090

Chicago/Turabian Style

Chen, Baoqing, Mei Wang, and Yu Gu. 2024. "Joint Spatio-Temporal-Frequency Representation Learning for Improved Sound Event Localization and Detection" Sensors 24, no. 18: 6090. https://doi.org/10.3390/s24186090

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop