Next Article in Journal
Multiband Spectrum Sensing and Power Allocation for aCognitive Radio-Enabled Smart Grid
Previous Article in Journal
Deep Auto-Encoder and Deep Forest-Assisted Failure Prognosis for Dynamic Predictive Maintenance Scheduling
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sound Event Detection by Pseudo-Labeling in Weakly Labeled Dataset

Department of Electronics and Electrical Engineering, Korea University Seoul, Seoul 136-713, Korea
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(24), 8375; https://doi.org/10.3390/s21248375
Submission received: 11 November 2021 / Revised: 9 December 2021 / Accepted: 13 December 2021 / Published: 15 December 2021

Abstract

:
Weakly labeled sound event detection (WSED) is an important task as it can facilitate the data collection efforts before constructing a strongly labeled sound event dataset. Recent high performance in deep learning-based WSED’s exploited using a segmentation mask for detecting the target feature map. However, achieving accurate detection performance was limited in real streaming audio due to the following reasons. First, the convolutional neural networks (CNN) employed in the segmentation mask extraction process do not appropriately highlight the importance of feature as the feature is extracted without pooling operations, and, concurrently, a small size kernel forces the receptive field small, making it difficult to learn various patterns. Second, as feature maps are obtained in an end-to-end fashion, the WSED model would be weak to unknown contents in the wild. These limitations would lead to generating undesired feature maps, such as noise in the unseen environment. This paper addresses these issues by constructing a more efficient model by employing a gated linear unit (GLU) and dilated convolution to improve the problems of de-emphasizing importance and lack of receptive field. In addition, this paper proposes pseudo-label-based learning for classifying target contents and unknown contents by adding ’noise label’ and ’noise loss’ so that unknown contents can be separated as much as possible through the noise label. The experiment is performed by mixing DCASE 2018 task1 acoustic scene data and task2 sound event data. The experimental results show that the proposed SED model achieves the best F1 performance with 59.7% at 0 SNR, 64.5% at 10 SNR, and 65.9% at 20 SNR. These results represent an improvement of 17.7%, 16.9%, and 16.5%, respectively, over the baseline.

1. Introduction

The recent development in deep learning field shows advances in the event detection field, such as earthquake detection [1] and sound event detection [2,3]. The sound event detection (SED) field has been a research focus due to as it is used in various real-life applications to recognize a target event and detect the onset and offset times in an audio clip. We encounter a rich variety of sound events in our daily lives, such as “baby cries”, “dog barks”, “phone rings”, “sirens”, and “boiling water”. SED can be applied to varied fields by utilizing such events. For example, SED can not only be used for public security surveillance [4] and, by monitoring animal sounds, the maintenance and preservation of ecosystem environments [5,6] but also for equipment failure monitoring, oil and gas pipeline anomaly detection, and seismic wave acoustic detection [7]. Although video or image-based event detection systems are also applicable to the above situations, systems based on audio-only have several advantages. First, they can be used even in dark environments because they do not require illumination. Second, while sound can penetrate or pass around obstacles, photos and videos are directly disturbed. Third, events, such as fire alarms sounding, are only detectable by sound.
In the deep learning framework, SED was developed by using frame-level annotated data (strongly labeled) which contains the starting and ending point of the target event [8,9]. However, since such deep learning models require frame-by-frame event information, constructing a large amount of data for training proved difficult. To alleviate the above limitation, weakly supervised learning models using weakly labeled data have recently been proposed. Weakly labeled data is learned by multi-label classification using multi-instance learning (MIL) [10] rather than frame-level classification and uses only the presence or absence of an event in the audio clip. The structure of the Weakly labeled data-based SED (WSED) models consists of a classifier and detector that detects the on-off sets of an event in an audio clip. The classifier is a multi-label classifier that has multiple labels for a single audio clip, not a one-hot label as used in general classifiers. One of the WSED studies tried to extend a low-dimensional feature map through transpose convolution neural network to a higher level for frame-by-frame event prediction, such as auto-encoder [11]. However, information and characteristics on the time and frequency axis are lost in the feature extraction process using a convolutional neural network (CNN), and it is difficult to extend from low-dimensional features to high-dimensional ones. To avoid this problem, a method of fitting the frame size to the low-dimensional size without expanding the features was proposed [12]. In addition, good performance was achieved with a structure in which a gated linear Unit (GLU) was applied to the CNN feature extraction process as the learning the activation function [13]. However, the frequency axis dimension reduction method is sensitive to noise and shows performance degradation in noisy environments. To solve this problem, a learning method that maintains the time-frequency dimension was devised [14]. This learning method finds a segmentation mask that separates the target event from the input audio clip. This model calculates the probability that the target event will be present in the audio through global pooling on the last layer without reducing the dimensions of the feature map. However, this model has some limitations.
First, the segmentation mask is extracted through CNN without a pooling operation. In general, CNN consists of a convolution operation using the kernel and a pooling operation that can find important features while reducing dimensions. However, since this model learns while maintaining the dimension of the feature map without a pooling operation, importance is not incorporated into the extracted features. Second, there is a limited receptive field due to the fixed kernel size. The receptive field represents the amount of information the CNN model can contain when training. In general CNN models, the size of the feature map decreases as it passes through the layers using a pooling operation so that the features of various patterns can be learned using only a small kernel. In contrast, this model can only learn a simplified pattern using a 3 x 3 size kernel at all layers without a pooling operation to maintain the same size as the input. Third, in the segmentation mask-based method, as the proposed architecture estimates target segmentation masks from an input feature without any spectral and temporal feature reduction, the noise contents in an input T-F feature would continuously influence the CNN feature extraction process, and they degrade the performance of SED system.
This paper improved the model problem by applying GLU [15], dilated convolution [16], noise label, which is pseudo label, and noise loss. To solve the first problem of being unable to judge the importance of the extracted features, the proposed model adopts GLU. The GLU can determine whether or not the extracted feature is valid information by training the gate through the CNN’s kernel, and is used by replacing the activation function. The gate has a value between 0 and 1 according to a sigmoid function, and the flow of information can be controlled by saving important information and discarding other information through element-wise multiplication with the extracted features. To solve the second problem of over-simplified pattern learning due to the limited receptive area, the proposed model uses dilated convolution. Dilated convolution can widen the receptive field by extending the spacing between weights without increasing the parameters of the model and still maintaining the size of the feature map. Lastly, to reduce the effect of noise, noise labels and noise loss are added. This idea was inspired by the blank label of Connectionist Temporal Classification (CTC) [17], which is used in speech recognizers and applied to SED [18,19]. The CTC improved performance by adding a blank label so that frames corresponding to unnecessary information belong to blank labels to reduce frames corresponding to silence or blank belonging to other classes in the speech recognizer. By doing so, this model adds a noise label, which is pseudo label [20], that can play the same role as the blank label of the CTC, so that background sound and noise do not affect the mask extraction of the target class when the segmentation mask is output. The noise information is extracted by additionally outputting the segmentation mask for the noise label. Furthermore, as the segmentation mask from which the noise component is to be extracted brings unnecessary information to the target event, noise loss is added so that each event mask can contain only important information. An overall summary of the main contribution of this paper is that Constructing a more noise-robust model by adding a noise label that can extract noise segmentation mask and applying a noise loss that allows maximum noise contents extraction. Additionally, to supplement the problems of the previous model, GLU and dilated convolution were applied to improve the performance.
The experiments are performed by using DCASE 2018 dataset [21] task1 and task2 [22]. Task 2 dataset which is for the audio tagging is used for the audio tagging and the frame-level detecting evaluation. For noise data, Task 1 dataset which is the scene classification is augmented with three types of SNR levels (0, 10, and 20 dB). The results show that our proposed pesudo label trick method enhances the robustness of the system in the noisy audio stream over state-of-the-art methods.
Section 2 covers the task approach to WSED and techniques applied in this paper. Section 3 introduces the proposed method and model. Section 4 details the changes to the experimental results resulting from adding the techniques used in this paper one by one. Finally, Section 5 contains the conclusions based on these experimental results.

2. Related Work

2.1. Multi Instance Learning (MIL)

Weakly supervised learning mainly approaches MIL problems [10,23]. It is usually used not only in sound research but also with medical images [24,25] and for semantic segmentation [26,27]. If the concept of MIL is applied to SED, as shown in Figure 1, one audio clip has as many bags as the number of classes, and each bag contains per frame information on whether an event occurs. If one or more frames in the bag have a positive value, then the bag also takes a positive value and is mapped to 1. On the contrary, if there is no event information in any frame in the bag, the bag has a negative value and is mapped to 0. If SED is approached in this way, as a MIL problem through weakly supervised learning rather than through Acoustic Scene Classification (ASC) [28,29] in which audio clips are mapped to one class, the model can be trained through a multiple-instance binary classification that can be mapped to multiple event classes.

2.2. Dilated Convolution

Kernels of different sizes are required to learn the features of varied patterns through CNN. However, when the size of the kernel is increased, the learning speed of the model slows as the number of parameters increases. This problem can be improved by a pooling operation between the CNN layers. When the pooling operation is performed, the size of the feature map is reduced, so the receptive area can be expanded without increasing the size of the kernel. However, as the size of the feature map decreases after the pooling operation, the values excluding specific information disappear, resulting in information loss. This loss of information is fatal for issues, such as image segmentation and image separation, which require the classification of classes by pixel. In this paper, since we also need to learn the segmentation mask of each class in pixel units, we applied dilated convolution [15,30], a method that can learn various patterns while maintaining the information of the feature map. Dilated convolution is already widely used in fields that require classification by pixel [31,32]. As shown in Figure 2, dilated convolution maintains the size of the feature map and expands the distance between the parameters without increasing the number of parameters, meaning that various patterns can be learned as the receptive fields of different size are formed. Red dots indicate parameters that can be learned, and the other areas have a value of 0. The dilation rate refers to the distance between the parameters, and, as is apparent from the figure, the receptive field increases as the dilation rate increases.

2.3. Gated Linear Unit (GLU)

Dilation convolution can be used to learn varied patterns with a small number of parameters without loss of information. However, meaningless features can be extracted because they are learned without the benefit of a pooling operation to determine the importance of pixel information. In this model, information can be controlled according to importance by using a GLU, a learnable activation function [13,16]. A GLU is used as an alternative to the activation function, and, as shown in Figure 3, the front half of each feature extracted through CNN is used as information, and the back half as a gate to determine the importance of the information in the front half. The difference from ReLU [33], which is a commonly used activation function, is that the range of output values is the same, but information flow can be controlled by using a gate. As the gate has a value between 0 and 1 through the sigmoid function, the gate of a pixel with meaningful information has a value close to 1, and the gate of a pixel with meaningless information has a value of 0. The obtained information and the gate control the flow of information through element-wise multiplication so that features can be extracted according to their importance.

3. Proposed Method

3.1. Extraction Segmentation Mask

To learn the segmentation mask for each event in pixel units from the audio clip, the dilated convolution and GLU described above are used. As shown in Figure 4, the input data is a log mel spectrogram, and features for obtaining a segmentation mask are extracted by repeating a dilated convolution block four times with two dilated convolutions in which the activation function is replaced by a GLU. The proposed model architecture is shown in Table 1. When performing the dilated convolution, the dilation rate was set to 1, 2, 4, and 8, which are the optimal dilation rates obtained by the comparative experiments in paper [34], so the learning is carried out by gradually increasing the distance between the weights. The features obtained through the dilated convolution blocks are reduced dimensions through 1 × 1 convolution using as many kernels as the number of classes, so that S, the segmentation mask for each class, can be extracted. Finally, the extracted S is mapped to a value between 0 and 1 for each pixel through the sigmoid function.

3.2. Global Pooling to Predict the Presence or Absence of Target Events

The segmentation mask (S) corresponding to each extracted class predicts whether an event occurs in the audio clip through a global pooling operation. The global pooling used global weighted rank pooling (GWRP) [35], which showed the best performance in the baseline model [14]. GWRP can alleviate the problem of underestimating and overestimating through global max pooling, which transfers only the largest value, and global average pooling, which calculates and transfers the average value [35]. This method assigns different weights according to the values of each pixel of S. By increasing the weight of a pixel with a large value and decreasing the weight of a pixel with a small value, the information on all pixels is optimally reflected in the prediction value that tries to determine whether an event occurrence is an output. It is as shown in Equation (1) below.
p k = G W R P ( S k ) = 1 N ( r ) i = 1 M r i 1 ( S k ) i .
S is the segmentation mask, and k is the index of each class. i is an index in which each pixel value of S is sorted in descending order, and M is the number of pixels in time × frequency. r is a hyper parameter, set to the value of 0 r 1 , and N ( r ) = i = 1 M r i 1 is a normalization term. In this way, when training the entire model, the value predicted through G W R P calculates the binary cross entropy loss with the weak label. The binary cross-entropy loss is calculated as shown in Equation (3) below.
B C E l o s s ( p k , t k ) = k = 1 K t k log G W R P ( S k ) ,
= k = 1 K t k log p k .
In Equation (2), p denotes the target event probability extracted through G W R P , and t denotes the weak label of the input data. k is the index of the class.

3.3. Noise Label and Noise Loss to Extract Noise Contents

In this paper, a model that could reduce the effect of noise is constructed by adding a pseudo label, which is for classifying unknown contests which is naturally faced in the real stream environment. Furthermore, a loss function is added to extract and separate unknown contents as much as possible through the added pseudo label to build a noise-robust model. Here, the pseudo label is represented by the noise label, and the added loss function is represented by noise loss. Before explaining the noise label and noise loss, let us first think about unknown contents. When recording audio in the real environment, it is inevitably mixed with unknown contents, which is noise, according to the environment. In addition, that noise can be distributed across varied frequency bands depending on the environment. Moreover, it is not known when and what noise will be present. In other words, noise is distributed over the time and frequency axes of the audio clip. Considering these characteristics of noise, noise label and noise loss are applied to this model. The noise label was inspired by the blank label of CTC [17]. The blank label of CTC is a label added to prevent confusion in model training due to the prediction of blank and silent frames that are not included in any class of the speech recognizer model. Using this approach, our method predicts and labels noise that is not included in any target events with a noise label. The noise label maps to 1 according to the characteristics of the noise that can be included anywhere in the aforementioned audio clip. That is, the model is trained by adding a noise label mapped to 1 to the weak labels of the event classes constituting the audio clip. To calculate the noise loss that helps to extract the noise as much as possible, first, a noise segmentation mask corresponding to the noise label is additionally extracted during the segmentation mask extraction process. The noise segmentation mask extracted calculates the KL divergence loss for each pixel so that information about noise can be maximally contained and reflect the characteristics of noise that may be distributed over all time-frequency axes. KL divergence is used since it allows a particular probability distribution to be converged to the desired distribution, and the output value of the noise segmentation mask to be given a value between 0 and 1 through the sigmoid. If the mean square error (MSE) function is used, the model does not train well due to the small gradient, so the learning speed of the model is optimized through KL divergence with a large gradient value. Then, the noise loss Equation (4) is as follows.
N l o s s ( ρ ^ M , ρ ) = 1 M m = 1 M ρ log ρ ρ ^ m + ( 1 ρ ) log 1 ρ 1 ρ ^ m .
In the formula, M is a time × frame, which means the total number of pixels, ρ , is a hyperparameter that is designated to converge the probability distribution to the desired value. In this model, ρ is 0.9999 . ρ ^ is the value of each pixel in the output noise segmentation mask. That is, the noise loss plays the role of converging the output value to close to 1 so that the noise segmentation mask can extract noise information that can be maximally distributed to all pixels. As a result, during the segmentation mask extraction process, the event segmentation mask can be obtained with minimal noise influence since noise is separately extracted. Therefore, the losses used when training the model proposed in this paper are the B C E l o s s and the N l o s s , also proposed in this paper. To define Loss, it is as shown in Equation (5) below.
L t o t a l = B C E l o s s + N l o s s .

4. Experiment

4.1. Database

The database used in the experiment is the public data used in DCASE 2018 task1 [21] and task2 [22]. Task 1 is a scene classification that classifies the environment to which the audio clip corresponds, and 8640 recordings consisting of 10 classes recorded in 10 cities were used. Task 2 is audio tagging that determines whether there is an event in an audio clip. The data is composed of 41 event classes, which vary in length from less than 1 s to longer than 30 s. To proceed with the SED experiment, data was generated using task 1 data as background sounds and task 2 data as event sounds. Before generating the data, several pre-processing tasks were performed. The task 1 data to be used as the background sound was divided so that no background sound was included in both the training and test sets, and task 2 data longer than 4 s to be used as event data was randomly cut to a length between 2 and 4 s. We use only the manually verified audio clips from Task 2 as sound events because the remaining audio clips are unverified and may contain noisy labels. In addition, since the volume of data differs for each event class, the classes comprised of only a small amount of data were made to have at least 600 data items through time-shifting, and it was ensured that there was no event overlap between the training and test sets. The data is generated by randomly selecting the background sound, which is noise, and also picking an arbitrary number of events to be mixed with the background sound and adding to the background sound. For the noise-dependent experiment, data corresponding to SNRs of 0, 10, and 20 are generated by calculating the SNR between the average amplitude of the randomly selected events and the amplitude of the background sound. The generated audio clips contain at least 3 to 8 events, and the training set includes 8000 samples divided into 4 cross-validation folds. Thus, 2000 model evaluation test sets are made for performance measurement.

4.2. Feature Extraction

The sampling rate of the audio clip was 32 Khz, and the spectrogram was extracted using short-time Fourier transform(STFT) with a window size of 2048 and a hop size of 1024 since this configuration is known to have a good resolution in the time-frequency domain [36]. As for the generated spectrogram, the mel spectrogram was extracted using 64 mel filter banks, and the log-mel spectrogram was obtained through log operation and used as input. Log-mel spectrograms are widely used in acoustic studies [32,37].

4.3. Evaluation and Metric

As evaluation indicators, F1 score [38], Area Under the Curve (AUC) [39], and mean Average Precision (mAP) [40] were used to verify the performance of Audio Tagging (AT) and SED. The F1 score is calculated based on the precision representing the accuracy from the system perspective and the recall representing the accuracy from the data perspective. The F1 score derives a value close to 0 when either the precision or recall has a low value, and a value close to 1 when both have high values.
The AUC is the area under the Receiver Operating Characteristic (ROC) curve, which plots the true positive rate and the false positive rate, and is expressed as a single value. When using the AUC, there is no need to manually designate a threshold, and, when the system outputs a random value, such as when the system is not trained, it has a value of 0.5.
As with AUC, AP refers to the area under the graph for precision and recall. As with the F1 score, the higher the precision and reproducibility, the closer to 1, and, in the case of multiple event detection instead of single-event detection, the performance is evaluated according to the AP average (mAP) of each class.

4.4. Post Processing for Performance Evaluation

Post-processing for performance is the same as in the baseline paper extracting the segmentation mask [14]. The performance for AT is measured by determining that there is an event if the probability that each event derived through G W R P is in the audio clip is 0.2 or more, and, if it is less than 0.2, there is no event. It is expressed as (7) in the formula.
e k = 1 p k > 0.2 0 o t h e r w i s e .
p k is the probability for the event class calculated by Equation (1), and k is the index indicating event class. The SED predicts the onset of an event when the average sum of information on the frequency axis included in time t is 0.1 or more in the segmentation mask and the event offset when it falls below. In addition, the SED result was measured by considering only the case of the class determined to have an event in the AT result. The equations for determining whether an event occurs for time t are as in (8) and (9), respectively.
q k ( t ) = 1 N i = 1 N S k ( t , f i ) ,
d k t = 1 q k > 0.1 0 o t h e r w i s e .
N is the size of the frequency axis, t is the time, S is the segmentation mask, and k is the number of classes. That is, q k ( t ) is the probability that an event occurs at time t for the k t h class, and d k t indicates whether an event occurs at time t of the k t h class.

4.5. Model

This section describes in detail the architecture of the model proposed in this paper and the baseline model. As shown in Table 1, the input data is log mel spectrogram, and the proposed model used dilated convolution while the baseline used convolution. The kernel used is a constant size of 3 × 3 in all layers of both models, and the number of kernels is 64, 128, 256, and 256. The dilation rates of the proposed model are set to 1, 2, 4, and 8. The activation function uses ReLU in the baseline and GLU in the proposed model. The convolution operation is repeated twice in each kernel, that is, a total of 8 times. The features extracted through the convolution operation reduce the number of channels by the number of event classes through 1 × 1 convolution. In the proposed model, the segmentation mask corresponding to the added noise label is additionally extracted, making one more output channel than the number of classes. The segmentation mask extracted through 1 × 1 convolution passes through the sigmoid and then calculates the probability of each event in the audio clip through GWRP. When training the model, the batch size is 8, the learning rate is 0.001, and the Adam optimizer [41] is used. In addition, batch normalization [42] is applied to all convolution layers except the 1 × 1 convolution to stabilize the model and improve learning speed. The GWRP’s hyperparameter r is set to 0.9998.

4.6. Ablation Analysis

In this part, we analyze how the performance changes when dilated convolution, GLU, noise label, and noise loss, the techniques used in this paper, are added to the baseline model. The experimental results analyzed are the F1, AUC, and mAP scores of Audio Tagging (AT) and Sound Event Detection (SED) at differing SNRs. In addition, it shows the F1 score for each event class when the SNR is 0 and the output segmentation mask.

4.6.1. Effect of Applying GLU

The reason for applying the GLU is that the CNN, which extracts the segmentation mask from the baseline model, proceeds without pooling so that the importance of the feature is not considered. The experiment was carried out by applying GLU to all convolution layers except the 1 × 1 convolution. Looking at the experimental results in Table 2 and Table 3, AT performance was slightly higher than baseline across all performance indicators and all SNRs, while SED performance differed only slightly at SNRs of 0 or 10, but, at 20, the performance indicators excluding AUC were higher than baseline.

4.6.2. Effect of Applying Dilated Convolution

The reason for using dilated convolution is to expand the receptive field without losing information about the input data while retaining the number of parameters in baseline. The convolution layers of the baseline model except the 1 × 1 convolution were replaced by dilated convolution. Table 2 and Table 3 show that at all SNRs AT and SED show significantly higher performance than baseline. This indicates that for SED the receptive field plays a very important role in the segmentation mask extraction-based method.

4.6.3. Effects of Simultaneous Application of GLU and Dilated Convolution (DCGLU)

This experiment applies Dilated convolution and GLU together. As in the above experiments, the convolution layer of the baseline model was replaced with a dilated convolution layer, and the activation function was changed to GLU. For the proposed model, as shown in Table 2 and Table 3, all indicators except for AUC at SNR 10 show better performance than when only dilated convolution is applied. Since GLU, which is a learnable activation function, is learned through CNN in the same way as the general features, the degree of the effect differs according to the size of the receptive field.

4.6.4. Change When Noise Label Is Added

This experiment added a noise label to the DCGLU model. It was conducted to determine the effect of extracting the noise segmentation mask by adding only the noise label without noise loss. Looking at the difference from the DCGLU model in the experimental results presented in Table 2 and Table 3 indicates that the results do not differ significantly with SNRs. There is almost no difference in the average F1 score performance of the entire class in Table 4 and Table 5. As for events with a slight difference, the DCGLU model showed somewhat higher AT and SED performance in events distributed in low frequency bands, such as bass drum, chime, and double base. However, with the model to which the noise label was added, AT and SED performance were better for events occupying a rather wide frequency band, such as cough and shatter, or a high frequency band, such as snare drum. In some cases, in Figure 5, events not included in the audio clip are extracted from the segmentation mask output when the noise label is added, but compared to other models the microwave has the cleanest output. That is, although, when the noise label is added, the effect is insignificant, event and noise are nevertheless separated. Performance is different for each event class because noise is heavily distributed in the low frequency band, and, so, for events occupying a low frequency band, the separation from noise causes more performance degradation.

4.6.5. Validity of Noise Label and Noise Loss

This experiment adds noise label and noise loss to the DCGLU model. Noise loss is added to extract as much noise information as possible into the noise segmentation mask. Figure 6 shows the extracted noise. Figure 6a is the noise information obtained by adding only a noise label to the DCGLU model, and Figure 6b is the noise information obtained by applying both noise label and noise loss. In Figure 6, yellow represents a value closer to 1, and the darker color a value closer to 0. Comparing the two figures, it seems that the model to which only the noise label was added extracts information about noise from all pixels. On the other hand, for the model to which both the noise label and noise loss are applied, less noise is extracted than the model to which only the noise label is added. However, looking at Figure 5, you can see that the model using noise label and noise loss together extracts each event segmentation mask better than other models, and produces the clearest output for Cello, Clarinet, and Fireworks events, which are absent from the audio clip. Therefore, the phenomenon shown in Figure 6 seems to reflect the event and the noise being rather effectively separated and maximally extracting noise through noise loss.

5. Discussion

This is the experimental result of the model that added both noise label and noise loss to the DCGLU proposed in this paper and the weakly labeled data-based model, including the baseline proposed previously. As with the previously proposed models, we experimented with FrameCNN, which is an autoencoder model [11], a WLDCNN model in which all layers are convolutional [12], and an attention model that combines CNN and biRNN, and applies GLU [13]. As Table 2 shows, the attention model demonstrated the highest F1 performance, that is, 0.648 when SNR was 0, 0.695 when it was 10, and 0.698 when it was 20. However, its F1 performance of SED, as shown in Table 3, was lower than for other models, including the baseline. On the other hand, the proposed model, showed better AUC and mAP AT performance, than the Attention model, as Table 2 shows. Furthermore, the proposed model demonstrated the best SED performance across all indicators and all SNRs. AT performance is high for the attention model because global max pooling is used in its last layer to determine the presence of audio events. However, its SED performance is lower than that of the other models because it does not perform backpropagation well due to the effect of global max pooling, which delivers only the largest value. In the case of SED, the proposed model shows the F1 best performance, 0.543 at 0 SNR, 0.595 at 10 SNR, and 0.610 at 20 SNR. This represents increases compared to the baseline of 0.177, 0.169, or 0.165, respectively, with the largest difference at 0 SNR where the most noise was mixed. As for performance by class, in terms of AT, the proposed model outperformed the attention model in 12 event classes, including acoustic guitar and double base, out of a total of 41. However, in the case of SED, the proposed model performed the best across all event classes, as is shown in Table 4 and Table 5. Therefore, considering the above experimental results, it can be seen that other proposed papers are vulnerable to the noise environment and it can be proved that the proposed noise label and noise loss work effectively to distinguish between event contents and noise. Although the performance has been improved, it needs improvement because the performance is not yet usable in a real system. As a future research direction, we plan to apply it to a speaker separation model that extracts a mask, such as a segmentation mask.

6. Conclusions

In this paper, a more efficient model was constructed by applying dilated convolution and GLU to improve the lack of receptive field and non-interference of feature importance, which was the problem with the SED models extracting segmentation masks from time-frequency domains. Moreover, a noise-robust model was developed by proposing a noise label that could separate the noise contents in the segmentation mask extraction process and a noise loss that could extract the noise contents to the maximum to solve the performance degradation problem due to noise inevitably mixed with the input data. In the experiment, we proved the performance of the proposed model by showing how the model changed and improved through the performance when the techniques used were applied step by step and the segmentation mask output.

Author Contributions

Conceptualization, C.P.; methodology, C.P.; software, C.P.; validation, C.P.; formal analysis, C.P.; investigation, C.P.; resources, C.P.; data curation, C.P. and D.K.; writing—original draft preparation, C.P.; writing—review and editing, C.P., D.K., and H.K.; visualization, C.P.; supervision, H.K.; project administration, C.P. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Korea Environment Industry & Technology Institute (KEITI) through Exotic Invasive Species Management Program, funded by Korea Ministry of Environment (MOE) (2021002280004).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

[dataset] Heittola, Toni; Mesaros, Annamaria; Virtanen, Tuomas. 2018. TUT Urban Acoustic Scenes 2018, Development dataset; Zenodo; Version 1.0; 10.5281/zenodo.1228142, [dataset] Eduardo Fonseca; Xavier Favory; Jordi Pons; Frederic Font; Manoj Plakal; Daniel P. W. Ellis; Xavier Serra. 2019. FSDKaggle2018; Zenodo; Version 1.0; 10.5281/zenodo.2552860.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dong, L.; Tang, Z.; Li, X.; Chen, Y.; Xue, J. Discrimination of mining microseismic events and blasts using convolutional neural networks and original waveform. J. Cent. South Univ. 2020, 27, 2078–3089. [Google Scholar] [CrossRef]
  2. Mesaros, A.; Heittola, T.; Virtanen, T.; Plumbley, M.D. Sound event detection: A tutorial. IEEE Signal Process. Mag. 2021, 38, 67–83. [Google Scholar] [CrossRef]
  3. Xia, X.; Togneri, R.; Sohel, F.; Zhao, Y.; Huang, D. A survey: Neural network-based deep learning for acoustic event detection. Circuits Syst. Signal Process. 2019, 38, 3433–3453. [Google Scholar] [CrossRef]
  4. Crocco, M.; Cristani, M.; Trucco, A.; Murino, V. Audio surveillance: A systematic review. ACM Comput. Surv. 2016, 48, 1–46. [Google Scholar] [CrossRef]
  5. Park, S.; Elhilali, M.; Han, D.; Ko, H. Amphibian Sounds Generating Network Based on Adversarial Learning. IEEE Signal Process. Lett. 2020, 27, 640–644. [Google Scholar] [CrossRef]
  6. Ko, K.; Park, J.; Han, D.; Ko, H. Channel and Frequency Attention Module for Diverse Animal Sound Classification. IEICE Trans. Inf. Syst. 2019, 102, 2615–2618. [Google Scholar] [CrossRef] [Green Version]
  7. Schröder, J.; Goetze, S.; Grützmacher, V.; Anemüller, J. Automatic acoustic siren detection in traffic noise by part-based models. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Vancouver, BC, Canada, 26–31 May 2013; pp. 493–497. [Google Scholar]
  8. Cakır, E.; Parascandolo, G.; Heittola, T.; Huttunen, H.; Virtanen, T. Convolutional recurrent neural networks for polyphonic sound event detection. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 1291–1303. [Google Scholar] [CrossRef] [Green Version]
  9. Parascandolo, G.; Huttunen, H.; Virtanen, T. Recurrent neural networks for polyphonic sound event detection in real life recordings. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Shanghai, China, 20–25 March 2016; pp. 6440–6444. [Google Scholar]
  10. Dietterich, T.G.; Lathrop, R.H.; Lozano-Pérez, T. Solving the multiple instance problem with axis-parallel rectangles. Artif. Intell. 1997, 89, 31–71. [Google Scholar] [CrossRef] [Green Version]
  11. Chou, S.Y.; Jang, J.S.R.; Yang, Y.H. FrameCNN: A weakly-supervised learning framework for frame-wise acoustic event detection and classification. Recall 2017, 14, 55–64. [Google Scholar]
  12. Kumar, A.; Raj, B. Deep cnn framework for audio event recognition using weakly labeled web data. arXiv 2017, arXiv:1707.02530. [Google Scholar]
  13. Xu, Y.; Kong, Q.; Wang, W.; Plumbley, M.D. Large-scale weakly supervised audio classification using gated convolutional neural network. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 121–125. [Google Scholar]
  14. Kong, Q.; Xu, Y.; Sobieraj, I.; Wang, W.; Plumbley, M.D. Sound event detection and time-frequency segmentation from weakly labelled data. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 777–787. [Google Scholar] [CrossRef]
  15. Dauphin, Y.N.; Fan, A.; Auli, M.; Grangier, D. Language modeling with gated convolutional networks. In Proceedings of the International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; pp. 933–941. [Google Scholar]
  16. Yu, F.; Koltun, V. Multi-scale context aggregation by dilated convolutions. arXiv 2015, arXiv:1511.07122. [Google Scholar]
  17. Graves, A.; Fernández, S.; Gomez, F.; Schmidhuber, J. Connectionist temporal classification: Labelling unsegmented sequence data with recurrent neural networks. In Proceedings of the 23rd International Conference on Machine Learning, Pittsburgh, PA, USA, 25–29 June 2006; pp. 369–376. [Google Scholar]
  18. Wang, Y.; Metze, F. A first attempt at polyphonic sound event detection using connectionist temporal classification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2017; pp. 2986–2990. [Google Scholar]
  19. Hou, Y.; Kong, Q.; Li, S.; Plumbley, M.D. Sound event detection with sequentially labelled data based on connectionist temporal classification and unsupervised clustering. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 46–50. [Google Scholar]
  20. Lee, D.H. Pseudo-label: The simple and efficient semi-supervised learning method for deep neural networks. In Proceedings of the ICML 2013 Workshop: Challenges in Representation Learning (WREPL), Atlanta, GA, USA, 16–21 June 2013; Volume 3, p. 896. [Google Scholar]
  21. Mesaros, A.; Heittola, T.; Virtanen, T. A multi-device dataset for urban acoustic scene classification. arXiv 2018, arXiv:1807.09840. [Google Scholar]
  22. Fonseca, E.; Plakal, M.; Font, F.; Ellis, D.P.; Favory, X.; Pons, J.; Serra, X. General-purpose tagging of freesound audio with audioset labels: Task description, dataset, and baseline. arXiv 2018, arXiv:1807.09902. [Google Scholar]
  23. Maron, O.; Lozano-Pérez, T. A framework for multiple-instance learning. Adv. Neural Inf. Process. Syst. 1998, 570–576. Available online: https://citeseerx.ist.psu.edu/viewdoc/download?doi=10.1.1.51.7638&rep=rep1&type=pdf (accessed on 10 November 2021).
  24. Quellec, G.; Cazuguel, G.; Cochener, B.; Lamard, M. Multiple-instance learning for medical image and video analysis. IEEE Rev. Biomed. Eng. 2017, 10, 213–234. [Google Scholar] [CrossRef]
  25. Xu, Y.; Mo, T.; Feng, Q.; Zhong, P.; Lai, M.; Eric, I.; Chang, C. Deep learning of feature representation with multiple instance learning for medical image analysis. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; pp. 1626–1630. [Google Scholar]
  26. Papandreou, G.; Chen, L.C.; Murphy, K.P.; Yuille, A.L. Weakly-and semi-supervised learning of a deep convolutional network for semantic image segmentation. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; pp. 1742–1750. [Google Scholar]
  27. Wu, J.; Zhao, Y.; Zhu, J.Y.; Luo, S.; Tu, Z. Milcut: A sweeping line multiple instance learning paradigm for interactive image segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 256–263. [Google Scholar]
  28. Naranjo-Alcazar, J.; Perez-Castanos, S.; Zuccarello, P.; Cobos, M. Acoustic scene classification with squeeze-excitation residual networks. ACM Comput. Surv. 2016, 48, 1–46. [Google Scholar] [CrossRef]
  29. McDonnell, M.D.; Gao, W. Acoustic scene classification using deep residual networks with late fusion of separated high and low frequency paths. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 141–145. [Google Scholar]
  30. Chen, Y.; Guo, Q.; Liang, X.; Wang, J.; Qian, Y. Environmental sound classification with dilated convolutions. Appl. Acoust. 2019, 148, 123–132. [Google Scholar] [CrossRef]
  31. Wei, Y.; Xiao, H.; Shi, H.; Jie, Z.; Feng, J.; Huang, T.S. Revisiting dilated convolution: A simple approach for weakly-and semi-supervised semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7268–7277. [Google Scholar]
  32. Wang, Y.; Hu, S.; Wang, G.; Chen, C.; Pan, Z. Multi-scale dilated convolution of convolutional neural network for crowd counting. Multimed. Tools Appl. 2020, 79, 1057–1073. [Google Scholar] [CrossRef]
  33. Agarap, A.F. Deep learning using rectified linear units (relu). arXiv 2018, arXiv:1803.08375. [Google Scholar]
  34. Li, Y.; Liu, M.; Drossos, K.; Virtanen, T. Sound event detection via dilated convolutional recurrent neural networks. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 286–290. [Google Scholar]
  35. Kolesnikov, A.; Lampert, C.H. Seed, expand and constrain: Three principles for weakly-supervised image segmentation. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 11–14 October 2016; pp. 695–711. [Google Scholar]
  36. Kong, Q.; Iqbal, T.; Xu, Y.; Wang, W.; Plumbley, M.D. DCASE 2018 challenge surrey cross-task convolutional neural network baseline. arXiv 2018, arXiv:1808.00773. [Google Scholar]
  37. Miyazaki, K.; Komatsu, T.; Hayashi, T.; Watanabe, S.; Toda, T.; Takeda, K. Weakly-Supervised Sound Event Detection with Self-Attention. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 66–70. [Google Scholar]
  38. Mesaros, A.; Heittola, T.; Virtanen, T. Metrics for polyphonic sound event detection. Appl. Sci. 2016, 6, 162. [Google Scholar] [CrossRef]
  39. Hanley, J.A.; McNeil, B.J. The meaning and use of the area under a receiver operating characteristic (ROC) curve. Radiology 1982, 142, 29–36. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Girshick, R.; Donahue, J.; Darrell, T.; Malik, J. Rich feature hierarchies for accurate object detection and semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Columbus, OH, USA, 23–28 June 2014; pp. 580–587. [Google Scholar]
  41. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  42. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. arXiv 2015, arXiv:1502.03167. [Google Scholar]
Figure 1. In order to apply WSE to MIL, it is necessary to map the strong label to the weak label through the bag concept. As shown in the picture above, if there is more than one positive value in the strong label, the weak label is mapped to 1; otherwise, it is mapped to 0.
Figure 1. In order to apply WSE to MIL, it is necessary to map the strong label to the weak label through the bag concept. As shown in the picture above, if there is more than one positive value in the strong label, the weak label is mapped to 1; otherwise, it is mapped to 0.
Sensors 21 08375 g001
Figure 2. This figure shows how the kernel and receptive field change according to the dilation rate change in dilated convolution. The green area indicates the size of the kernel, and the red dot indicates the learnable parameters. Except for the red dot, the kernel has zero values.
Figure 2. This figure shows how the kernel and receptive field change according to the dilation rate change in dilated convolution. The green area indicates the size of the kernel, and the red dot indicates the learnable parameters. Except for the red dot, the kernel has zero values.
Sensors 21 08375 g002
Figure 3. The GLU controls the flow of information by extracting features through the kernel of the convolutional neural network and using one half as information, and the other half as a gate through a sigmoid.
Figure 3. The GLU controls the flow of information by extracting features through the kernel of the convolutional neural network and using one half as information, and the other half as a gate through a sigmoid.
Sensors 21 08375 g003
Figure 4. The figure above represents the architecture of the model proposed in this paper. The model receives a log mel spectrogram as input and outputs a segmentation mask through 4 dilated conv blocks and 1 × 1 convolution. The output segmentation mask calculates the probability that an event is included in the audio clip through global pooling. The model is trained using two losses, one of which is binary cross-entropy between predicted probability and weak label, and the other is KL divergence between noise segmentation mask and user hyper-parameter.
Figure 4. The figure above represents the architecture of the model proposed in this paper. The model receives a log mel spectrogram as input and outputs a segmentation mask through 4 dilated conv blocks and 1 × 1 convolution. The output segmentation mask calculates the probability that an event is included in the audio clip through global pooling. The model is trained using two losses, one of which is binary cross-entropy between predicted probability and weak label, and the other is KL divergence between noise segmentation mask and user hyper-parameter.
Sensors 21 08375 g004
Figure 5. Changes in the output of the segmentation mask as the techniques proposed in this model are applied to the baseline. Ground truth is the log mel spectrogram extracted from each event.
Figure 5. Changes in the output of the segmentation mask as the techniques proposed in this model are applied to the baseline. Ground truth is the log mel spectrogram extracted from each event.
Sensors 21 08375 g005
Figure 6. The noise segmentation mask of the model to which only the noise label is applied, and the model to which both the noise label and noise loss are applied. The yellow areas are judged to have noise and represent values closer to 1, and the darker areas values closer to 0.
Figure 6. The noise segmentation mask of the model to which only the noise label is applied, and the model to which both the noise label and noise loss are applied. The yellow areas are judged to have noise and represent values closer to 1, and the darker areas values closer to 0.
Sensors 21 08375 g006
Table 1. Comparison of proposed model and baseline model.
Table 1. Comparison of proposed model and baseline model.
Proposed ModelBaseline Model
Layers, Activation Function
{Kernel Size, Dilation Rate, Repeat}
Number
of Kernel
Output Size
{Channel × Time × Frequency}
Layers, Activation Function
{Kernel Size, Repeat}
Number
of Kernel
Output Size
{Channel × Time × Frequency}
Input log mel spectrogram-1 × 431 × 64Input log mel spectrogram-1 × 431 × 64
Dilated CNN, GLU {3 × 3, 1, 2}6432 × 431 × 64CNN, ReLU {3 × 3, 2}6464 × 431 × 64
Dilated CNN, GLU {3 × 3, 2, 2}12864 × 431 × 64CNN, ReLU {3 × 3, 2}128128 × 431 × 64
Dilated CNN, GLU {3 × 3, 4, 2}256128 × 431 × 64CNN, ReLU {3 × 3, 2}256256 × 431 × 64
Dilated CNN, GLU {3 × 3, 8, 2}256128 × 431 × 64CNN, ReLU {3 × 3, 2}256256 × 431 × 64
CNN, Sigmoid {1 × 1, -, -}4242 × 431 × 64CNN, Sigmoid {1 × 1, -}4141 × 431 × 64
Global Weighted Rank Pooling-42Global Weighted Rank Pooling-41
Table 2. Audio tagging performance according to SNR.
Table 2. Audio tagging performance according to SNR.
0 dB10 dB20 dB
ModelsF1AUCmAPF1AUCmAPF1AUCmAP
FrameCNN [11]0.3010.6750.3190.3180.6960.3520.3200.7080.367
WLDCNN [12]0.2890.5990.2680.3120.6210.3090.3000.6170.295
Attention [13]0.6480.8540.6960.6950.8740.7380.6980.8750.743
Baseline [14]0.4230.8440.4920.4630.8740.5680.4710.8810.591
Baseline + GLU0.4270.8460.4990.4650.8740.5710.4820.8860.605
Baseline + Dilated0.5590.9020.6870.6160.9280.7580.6160.9310.764
Baseline + GLU + Dilated (DCGLU)0.5650.9000.6870.6130.9260.7520.6270.9330.776
DCGLU + Noise label0.5650.9000.6870.6140.9250.7540.6280.9330.776
DCGLU + Noise label + Noise loss0.5970.9100.7210.6450.9330.7850.6590.9400.802
Table 3. Sound event detection performance according to SNR.
Table 3. Sound event detection performance according to SNR.
0 dB10 dB20 dB
ModelsF1AUCmAPF1AUCmAPF1AUCmAP
FrameCNN [11]0.1430.6210.0640.1550.6380.0690.1580.6480.071
WLDCNN [12]0.1050.5760.1260.1460.5980.1550.1210.5890.135
Attention [13]0.1170.7810.2180.1290.7990.2350.1370.8020.242
Baseline [14]0.3660.7150.2840.4260.7490.3440.4450.7600.374
Baseline + GLU0.3660.7130.2860.4250.7450.3510.4520.7580.383
Baseline + Dilated0.5140.8230.4730.5600.8490.5260.5750.8570.545
Baseline + GLU + Dilated (DCGLU)0.5170.8190.4710.5680.8510.5330.5880.8610.558
DCGLU + Noise label0.5160.8190.4710.5700.8510.5360.5870.8610.559
DCGLU + Noise label + Noise loss0.5430.8320.5020.5950.8610.5640.6100.8700.583
Table 4. F1 score of audio tagging for each class at SNR 0.
Table 4. F1 score of audio tagging for each class at SNR 0.
ModelsAcous.
Guitar
ApplauseBarkBase
Drum
BurpingBusCelloChimeClarinetKeyboardCoughCowbellDouble
Bass
DrawerElec.
Piano
FartFinger
Snap
Fire
Works
FluteGlock.Gong
FrameCNN [11]0.2750.5770.2490.2880.2250.4990.3450.2960.4160.2170.2020.1790.2180.1950.3680.2330.2060.1880.4230.2880.270
WLDCNN [12]0.1940.8550.1950.1900.1910.5890.2300.1900.4680.1920.1960.1900.1860.1930.2520.1960.1990.1960.4490.3490.194
Attention [13]0.3070.8660.8600.5290.7870.6690.5260.6930.7500.6400.7830.8410.3530.3370.5360.5990.7110.4300.7490.5340.442
Baseline [14]0.3720.5690.5180.2690.3840.4410.3890.5120.4460.4800.4020.3450.2730.2620.4220.3770.3790.2730.4400.4390.348
Baseline + GLU0.3820.5570.5140.2660.4060.4320.3990.5220.4530.4860.4060.3460.2640.2520.4200.3750.3820.2700.4490.4380.355
Baseline + Dilated0.4370.7630.7060.3970.5580.5420.4840.6040.5740.5720.5550.5650.3670.3140.5060.5340.4940.3430.5720.5670.406
Baseline + GLU + Dilated (DCGLU)0.4290.7460.7340.3990.5600.5170.4860.6460.6090.5740.5660.5890.3680.3120.5100.5330.5050.3360.6020.5820.405
DCGLU + Noise label0.4250.7520.7370.3750.5660.5130.4880.6320.5990.5770.5810.6020.3490.3200.4920.5260.4970.3280.6020.5820.404
DCGLU + Noise label + Noise loss0.4660.7830.7630.4310.6070.5560.4940.6880.6290.5860.6050.6350.3800.3310.5280.5990.5230.3680.6270.6030.461
ModelsGunshotHarmonicaHihatKeysKnockLaughterMeowMicrowaveOboeSexophoneScissorsShatterSnare
drum
SqueakTambourineTearingTelephoneTrumpetViolinWritingAvg.
FrameCNN [11]0.2200.5590.2790.3020.1750.2080.2230.2850.4890.5480.2310.2080.3830.2020.3390.2260.3120.3810.4670.2130.301
WLDCNN [12]0.2450.3620.4720.1970.1980.1910.2000.2020.4820.7020.1910.2370.2960.1890.1920.1970.2330.4860.4950.1930.289
Attention [13]0.5720.8790.9240.8510.5460.6510.5860.6520.6780.7820.6030.5970.8480.5000.7530.5260.5480.8450.7870.5050.648
Baseline [14]0.3980.5320.5150.5060.3270.4240.4630.3600.4870.5110.4040.4290.4970.3620.6490.4440.3880.5020.4380.3780.423
Baseline + GLU0.4010.5550.5150.5020.3200.4170.4740.3710.4940.5210.4090.4100.5170.3740.6890.4500.3870.5250.4480.3730.427
Baseline + Dilated0.5140.7630.7640.7310.3850.5370.6110.4930.6740.6360.5000.5990.6550.4970.8670.5590.5100.7200.6380.4460.559
Baseline + GLU + Dilated (DCGLU)0.5050.7420.7210.7400.3890.5590.6460.4780.6950.6600.5050.5870.6390.4920.8890.5750.5280.7290.6430.4490.565
DCGLU + Noise label0.5040.7460.7340.7300.3820.5540.6520.4840.7000.6640.4950.6060.6520.5090.8920.5770.5250.7360.6270.4450.565
DCGLU + Noise label + Noise loss0.5460.7900.7650.7520.4070.5880.6750.5280.7300.6900.5240.6420.6950.5440.8920.6100.5730.7460.6570.4770.597
Table 5. F1 score of sound event detection for each class at SNR 0.
Table 5. F1 score of sound event detection for each class at SNR 0.
ModelsAcous.
Guitar
ApplauseBarkBase
Drum
BurpingBusCelloChimeClarinetKeyboardCoughCowbellDouble
Bass
DrawerElec.
Piano
FartFinger
Snap
Fire
Works
FluteGlock.Gong
FrameCNN [11]0.1360.3430.1110.0460.0710.2620.2060.1570.2270.0710.0640.0490.0720.0610.1910.0770.0550.0610.2280.1320.142
WLDCNN [12]0.0000.5440.0000.0000.0000.4170.1130.0000.3580.0000.0000.0000.0020.0000.0750.0600.0000.0000.2830.1140.000
Attention [13]0.0690.2330.0760.0650.1540.1890.1700.1630.2310.0360.0510.0700.0260.0080.1270.0770.0570.0230.2170.0720.110
Baseline [14]0.2900.6070.4030.1490.3410.3880.3460.5620.4010.4570.2620.3190.0620.0480.3810.3050.2780.1330.4010.3720.332
Baseline + GLU0.2960.5960.3920.1510.3370.3650.3470.5450.4150.4530.2650.3150.0540.0460.3520.3050.2720.1250.4230.3580.317
Baseline + Dilated0.4180.7400.6370.2650.4930.5200.4630.6290.5890.5520.4760.5070.2660.1860.5050.5070.4620.2410.5390.4620.388
Baseline + GLU + Dilated (DCGLU)0.3960.7250.6570.2640.5010.4710.4430.6520.5960.5580.4950.5350.2510.1620.4980.5060.4720.2160.5710.4650.375
DCGLU + Noise label0.3880.7310.6620.2450.5000.4690.4480.6390.6020.5540.5000.5150.2320.1640.4720.4990.4690.2230.5750.4690.384
DCGLU + Noise label + Noise loss0.4320.7530.6720.3100.5380.5200.4740.6720.6210.5660.5320.5530.2810.1770.5050.5450.5050.2500.5900.4770.425
ModelsGunshotHarmonicaHihatKeysKnockLaughterMeowMicrowaveOboeSexophoneScissorsShatterSnare
drum
SqueakTambourineTearingTelephoneTrumpetViolinWritingAvg.
FrameCNN [11]0.0800.3110.1290.1390.0480.0810.0820.1690.3100.2860.0870.0760.2230.0740.2080.0770.1570.2140.2560.0790.143
WLDCNN [12]0.0310.2640.1190.0000.0000.000.0240.1000.2870.6180.0030.0390.2150.0000.0000.0000.0000.2530.3820.0000.105
Attention [13]0.0560.3090.1100.0980.0350.0140.0540.1660.2760.2940.0330.0320.1490.0610.1140.0310.1390.2680.2690.0400.117
Baseline [14]0.2370.5930.4410.4480.1590.2950.3130.3410.5850.5030.3490.3100.5740.2870.6240.3530.3630.5680.4690.3460.366
Baseline + GLU0.2480.6220.4330.4600.1530.2910.3090.3190.6050.5290.3500.3040.6000.2980.6380.3600.3610.5790.4750.3600.366
Baseline + Dilated0.4310.7530.6350.6820.2630.4870.5170.4890.6690.6270.4820.5480.6710.4790.7130.5150.5000.6950.6370.4380.514
Baseline + GLU + Dilated (DCGLU)0.4240.7430.6120.6950.2610.5090.5680.4800.6930.6460.5030.5310.6770.5000.7240.5200.5010.7160.6410.4270.517
DCGLU + Noise label0.4250.7520.6100.7000.2540.5000.5540.4850.7010.6590.5060.5460.6850.4920.7220.5220.5150.7170.6390.4350.516
DCGLU + Noise label + Noise loss0.4570.7620.6440.7090.2850.5340.5950.5140.7060.6750.5110.5620.7030.5260.7250.5500.5390.7280.6520.4680.543
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Park, C.; Kim, D.; Ko, H. Sound Event Detection by Pseudo-Labeling in Weakly Labeled Dataset. Sensors 2021, 21, 8375. https://doi.org/10.3390/s21248375

AMA Style

Park C, Kim D, Ko H. Sound Event Detection by Pseudo-Labeling in Weakly Labeled Dataset. Sensors. 2021; 21(24):8375. https://doi.org/10.3390/s21248375

Chicago/Turabian Style

Park, Chungho, Donghyeon Kim, and Hanseok Ko. 2021. "Sound Event Detection by Pseudo-Labeling in Weakly Labeled Dataset" Sensors 21, no. 24: 8375. https://doi.org/10.3390/s21248375

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop