Next Article in Journal
Prediction of Three-Directional Ground Reaction Forces during Walking Using a Shoe Sole Sensor System and Machine Learning
Previous Article in Journal
Multi-Domain Rapid Enhancement Networks for Underwater Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Acoustic-Sensing-Based Attribute-Driven Imbalanced Compensation for Anomalous Sound Detection without Machine Identity

1
Key Innovation Group of Digital Humanities Resource and Research, Shanghai Normal University, Shanghai 200234, China
2
Unisound AI Technology Co., Ltd., Beijing 100089, China
3
Department of ECE, University of Texas at Dallas, Richardson, TX 75080, USA
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(21), 8984; https://doi.org/10.3390/s23218984
Submission received: 25 September 2023 / Revised: 2 November 2023 / Accepted: 3 November 2023 / Published: 5 November 2023
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Acoustic sensing provides crucial data for anomalous sound detection (ASD) in condition monitoring. However, building a robust acoustic-sensing-based ASD system is challenging due to the unsupervised nature of training data, which only contain normal sound samples. Recent discriminative models based on machine identity (ID) classification have shown excellent ASD performance by leveraging strong prior knowledge like machine ID. However, such strong priors are often unavailable in real-world applications, limiting these models. To address this, we propose utilizing the imbalanced and inconsistent attribute labels from acoustic sensors, such as machine running speed and microphone model, as weak priors to train an attribute classifier. We also introduce an imbalanced compensation strategy to handle extremely imbalanced categories and ensure model trainability. Furthermore, we propose a score fusion method to enhance anomaly detection robustness. The proposed algorithm was applied in our DCASE2023 Challenge Task 2 submission, ranking sixth internationally. By exploiting acoustic sensor data attributes as weak prior knowledge, our approach provides an effective framework for robust ASD when strong priors are absent.

1. Introduction

Acoustic-sensing-based anomalous sound detection (ASD) has become an increasingly important technique for predictive maintenance and condition monitoring in industrial environments, especially with the emergence of Industry 4.0. ASD aims to detect anomalous noises in acoustic signals that may indicate a fault or deterioration in mechanical equipment. When machinery begins to degrade, the vibration and sounds emitted often change subtly before failure occurs. By identifying these anomalous acoustic patterns, ASD systems can provide early warning of impending faults, enabling proactive maintenance to avoid catastrophic breakdowns. Traditional manual acoustic monitoring is labor-intensive and prone to human variability. The emergence of automated ASD systems addresses these limitations, reducing personnel costs and providing more consistent machine health assessment.
In most real-world scenarios, abnormal samples cannot be obtained by damaging the machine, and the complex engineering environment introduces much noise into the sound samples. The operating settings of different machines are also diverse. Therefore, the main challenge of the acoustic-sensing-based ASD task is to detect anomalous sounds when only normal sound samples are provided as training data [1,2,3].
In addition to acoustic sensing data, anomaly detection and fault diagnosis methods designed for other data types are also worth considering as references. In [4], the authors innovatively utilized event-based cameras for anomaly detection, collecting vibration signals of machines in a contactless manner and providing a new perspective for machine condition monitoring. Facing similar challenges of imbalanced data and dynamic operations, the authors in [5] combined a self-supervised anomaly detector based on a local outlier factor (LOF) and a deep Q-network (DQN) supervised reinforcement learner to classify interturn short-circuit, local demagnetization, and mixed faults. Additionally, in the context of small datasets in industrial settings, the authors in [6] optimized the friction-drilling process through model ensembling in order to work with no complete information. Feature engineering is also crucial for rotary machine monitoring. The authors in [7] proposed a novel feature extraction method called weighted multi-scale fluctuation-based dispersion entropy for detecting faults in planetary gearboxes. In [8], permutation entropy was integrated with a flexible analytical wavelet transform for bearing defect detection. These real-world practices in industrial scenarios provide valuable references for ASD work.
To drive the development of acoustic-sensing-based ASD technology, a sub-challenge (Task 2) of ‘Unsupervised Detection of Anomalous Sounds for Machine Condition Monitoring’ has been launched in the IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events (DCASE) since 2020 [9]. In the previous work of the DCASE Challenge Task 2, the anomaly detection system AutoEncoder (AE) [9] based on generative model was widely used due to its simple design and efficient inference. In addition to acoustic sensing tasks, AEs have also been extensively utilized as unsupervised anomaly detectors in many other application domains [10,11]. However, AE-based anomaly detection relies on the assumption that anomalies are difficult to reconstruct. Given the inherent denoising characteristics of AE [12], enhancing the representation capacity of AE may inadvertently treat anomalies as noise, thus constraining the anomaly detection performance reliant on AE. In addition to AE, other generative models, such as IDNN [13], Efficient GAN [14], and Glow_Aff [15], detect anomalies by modeling the distribution of normal sounds and determining whether the sound under test is within the range of the normal sound distribution. However, due to the complexity of anomalous sounds, it is difficult to model a stable distribution for anomaly detection [16], which limits the generative model.
Therefore, in order to better model the characteristics of normal sounds, systems based on discriminative models [17] were designed and achieved excellent performance. These models utilize powerful deep learning feature extraction networks, such as ResNet [18], MobileNetV2 [19], and STgram [20], for self-supervised classification tasks related to machine ID. During inference, abnormal samples are exposed due to the difficulty in classification, allowing for effective anomaly detection. Undoubtedly, in previous challenges regarding DCASE Task 2 evaluations, the competitive anomaly detection systems were also based on utilization of machine ID classification. The algorithm’s success is attributed to the high-quality classification boundaries established by leveraging strong prior knowledge of machine ID. However, in practical applications, obtaining such high-quality prior knowledge of machine ID is often unfeasible. This raises an important question: how can we adapt anomaly detection algorithms based on discriminative models to operate effectively under limited prior knowledge conditions?
Instead, we need to design anomaly detection algorithms under weak prior knowledge conditions. According to the task setting of DCASE2023 Task 2, we cannot obtain high-quality prior knowledge such as machine ID, but we can obtain the attributes information of each audio clip, such as microphone number, machine running speed, machine load status, etc. We define the attributes information of the audio clips as weak prior knowledge. Unfortunately, there is no free lunch in the world. Intuitively, these attribute labels are extremely imbalanced, complex categories, and cannot form clear classification boundaries, but they are more accessible in the real world.
In this paper, we propose the attribute-driven imbalanced compensation (AIC) method, aiming to overcome the disadvantages of weak prior knowledge and use attribute labels to build discriminative models for anomaly detection. Our main contributions are as follows: (1) we propose an attribute classifier using the weak prior knowledge, making the application of discriminative models possible when machine ID labels are limited. (2) We propose the imbalanced compensation strategy to solve the common problem of extreme sample imbalance in attribute labels. (3) We propose a score fusion method based on AIC to enhance the robustness of the model.

2. Proposed Method

The AIC framework we propose contains an imbalanced compensation module, M attribute classifiers, and an ensemble attribute anomaly detector. The overview of the overall framework is shown in Figure 1. The overall framework consists of training and testing stages. First, the raw data are augmented via imbalanced compensation separately for each attribute. In the training stage, a classifier is trained for the augmented data of each attribute with cross-entropy loss. In the testing stage, the augmented data are fed into the trained classifiers to obtain embeddings, with each attribute corresponding to one embedding space. Then, the embedding of the test samples is extracted through the trained classifier, and KNN is used to calculate the score in the ensemble attribute anomaly detector.

2.1. Attribute Classifier

Although the previous work using machine ID for classification achieved good results [1,2], it is often more limited in practical applications, such as only one machine is working. In this case, strong prior knowledge such as machine ID cannot be used. Nevertheless, machines still have weak prior knowledge that is easy to obtain, such as attributes in Table 1; both the ‘ToyCar’ and ‘ToyTrain’ have three types of weak attributes. Therefore, in this study, we propose to train the attribute classifier using such weak prior knowledge as attributes.
In real-world applications, different attributes of a machine may work under different operating status. These statuses can be easily collected and labeled. For example, as illustrated in Table 2, in the ToyCar dataset provided by the DCASE2023 Challenge, the attribute ‘Mic’ has two types of operating status: ‘1’ and ‘2’. These status labels can be used to train an anomaly classifier, allowing it to distinguish between different operating status for each attribute. This approach helps the anomaly classifier learn the details of the training dataset more comprehensively and deeply, similar to observing the same object from different perspectives. The operating status information provided by DCASE [21,22] naturally accompanies machine operation and is readily accessible.
Similar to previous machine ID classifiers [1], in our attribute classifier, we adopt the cross-entropy loss to classify each operating status of each attribute of each machine. As shown in Figure 1 training classifier step, and  both of Table 1 and Table 2, this can be described as training M classifiers for M attributes, where each classifier performs an K m -class classification task and K m represents the total distinct types of operating status for the m-th attribute. We use ResNet18 [23] as the backbone encoder of the classifier to obtain the embedding of each attribute. The proposed m-th attribute classifier (AC) is trained with the cross-entropy loss function as
L AC m = 1 N K m i = 1 N k m = 1 K m y i k m log p θ E x i k m
where N means the total input samples of the m-th attribute, x i k m , y i k m represent the i-th input sample and its k-th operating status in attribute m. p θ E ( · ) is the softmax output of the encoder with parameters θ E . As depicted in Figure 1 training detector step, following the independent training of the M attribute classifiers, we correspondingly learn M anomaly detectors, with each detector associated with one of the M-trained attribute classifiers. We utilize KNN as the anomaly detector, with the embeddings extracted by the trained classifiers used as the training data. During testing, the test sample is passed through each of the M-trained classifiers to obtain M embeddings, which are then fed into their corresponding M anomaly detectors to produce anomaly scores. The final aggregated anomaly score is obtained by taking the harmonic mean of the individual scores from each detector.
However, classifiers trained solely on operating status information for machine attributes often struggle to converge. Taking ToyCar dataset as an example, in Table 2, the information of training samples is expressed in the form of ‘Category: #samples’, and we can see that different attributes have varying numbers of operating status categories. For example, ‘Car model’ has 10 categories from A1 to E2, ‘Speed’ has 5 categories, controlled by voltage levels from 2.8 V to 4.0 V, and ‘Mic’ includes 2 categories, 1 and 2. Additionally, the number of samples for each operating status category is highly unbalanced. These factors make machine attributes weaker prior knowledge compared to machine IDs and pose several challenges with using operating status alone for training attribute classifiers: (1) the inconsistency in the number of attributes and operating status categories across different types of machines makes it difficult to establish consistent classification boundaries; (2) the severe sample imbalance within the same attribute and operating status category affects the classifier’s ability to accurately characterize normal samples; (3) during testing, the machine attributes and their corresponding operating status are unknown, further complicating the classification process. As a result, classifiers trained only on operating status are prone to misclassifying normal unseen samples as anomalies. To address this issue, we need to propose a method that strengthens the weak attribute knowledge as prior information in anomaly detection models.

2.2. Imbalanced Compensation

To solve the problem of severe unbalanced samples among different operating status shown in Table 2, in this section, we propose an imbalanced compensation module to enhance the proposed attribute classifier training. The module mainly includes two parts: (1) maximum expansion uniform sampling, and (2) robust data transformation. Figure 2 illustrates the schematic diagram of the effects of imbalanced compensation on acoustic-sensing-based ASD training data.
In Figure 2, different symbol shapes denote different categories with originally imbalanced number of samples. After maximum expansion and uniform sampling, the categories are balanced in terms of sample counts. Boundary shape changes following robust data transformation signify altered data distributions. In the first step, we identify the category with the maximum number of samples and expand all other categories to this level via oversampling. While this balances the quantities, the original data distributions remain unchanged, impeding effective training of the attribute classifier. Therefore, we subsequently apply robust data transformations, randomly augmenting the balanced data with 4 different techniques to alter the data distribution and simulate varied recording conditions. This enables successful training of the attribute classifier and enhances model robustness. In summary, our proposed pipeline tackles data imbalance through expansion and synthesizes robustness via data transformation, enabling learning from skewed real-world data.
The detail algorithm of our proposed imbalanced compensation module is presented in Algorithm 1. Given an unbalanced dataset with all N samples for one attribute of one machine, { x i k m } i = 1 , k m = 1 N , K m , which has N k m samples, it is classified into the k m -th category of operating status for each attribute, satisfying N = k m = 1 K m N k m . With the k m varying, N k m takes different values, resulting in an imbalance of data in each attribute.
Algorithm 1 Proposed imbalanced compensation method in m-th attribute
Input: An unbalanced dataset of all N samples in m-th attribute, { x i k m } i = 1 , k M = 1 N , K m
Output: A balanced dataset of R samples after imbalanced compensation (IC) in m-th attribute, { x i k m I C } i = 1 , k = 1 R , K m
  1. Find the maximum count of operating status T = max N k m
  2: Calculate the sample increment Δ k m = T N k m for each operating status
  3: Expand sample numbers to N * = K m × T
  4: Get a balanced dataset { x i k m M E U S } i = 1 , k m = 1 N * , K m after maximum expansion uniform sampling (MEUS)
  5: Sample R times in { x i k m M E U S }
  6: for  i , k , m in R , K m , M  do
  7:     x i k m I C = T 4 ( T 3 ( T 2 ( T 1 ( x i k m M E U S ) ) ) )
  8: end for
  9: Obtain the final dataset { x i k m I C } i = 1 , k m = 1 R , K m after imbalanced compensation
Maximum Expansion Uniform Sampling: As shown in Figure 2, we first introduce the maximum expansion uniform sampling to expand the original dataset to a balanced one. When we apply maximum expansion uniform sampling, we take the maximum value of N k m , set it to T = max N k m . Then, for the N k m samples in each operating status, we copy the data to increase the number of samples defined as Δ k m = T N k m . At this time, the total number of samples is expanded to N * = K m × T , and the number of samples among each operating status reaches balance. The original samples x i k m can be represented as x i k m M E U S after applying maximum expansion uniform sampling. Therefore, maximum expansion uniform sampling solves the problem of severe imbalance of training samples, allowing the classifier training to converge.
For example, the machine ToyCar has three attributes: ‘Car model’, ‘Speed’, and ‘Mic’. We train three separate classifiers for this machine. For the ‘Car model’ attribute, there are 10 categories ‘C1’–‘E1’ with extremely imbalanced quantities as shown in Table 2. With imbalanced compensation, we first apply maximum expansion uniform sampling. Specifically, we take the number of samples in the largest category ‘C1’, which is 215. Then, we resample each category to have 215 samples, making the number of samples balanced across categories. Similarly, we apply the same procedure to the other two attributes of ToyCar, balancing the number of samples for each category within every attribute.
Robust Data Transformation: Based on the enhanced maximum expansion uniform sampling, we then perform robust data transformation to augment the environment robustness of training data for improving the generalization ability of the resulting attribute classifier model. When we apply robust data transformation, we first sample R times after maximum expansion uniform sampling according to the law of large numbers. When R is large enough, the distribution of samples after R sampling is the same as the balanced sample distribution after maximum expansion uniform sampling. At the same time, each sampling is accompanied by 4 data augmentations, which are
  • AddGaussianNoise: Directly adding a noise signal obeying a zero-mean Gaussian distribution to the original audio signal in the time domain. In practical environments, many background noises can be regarded as additive noise. After such noise addition to the audio signal, it can capture various and complicate acoustic characteristics of real environments.
  • TimeStretch: Changing the speed of audio without altering its pitch by a pre-defined rate. Here, we randomly applied rates in the range of [0.8, 1.25].
  • PitchShift: Randomly increased or decreased the original pitch. Here, we vary the pitch by pre-defined semitones in the range of [ 4 , 4].
  • TimeShift: Shifts the entire audio signal forward or backward. Here, the shift range was [ 0.5 , 0.5] of the total signal length.
The above augmentations are represented by T 1 , T 2 , T 3 , and T 4 , respectively. Specifically, these four data augmentations are applied to each sampled sample with a 50% probability, distorting the distribution after sampling. To some extent, robust data transformation simulates unknown samples and enhances the robustness of the classifier, making it less prone to errors when classifying completely unknown samples in the test set. In summary, the samples after robust data transformation can be expressed as x i k m I C = T 4 ( T 3 ( T 2 ( T 1 ( x i k m M E U S ) ) ) ) .
Still taking ToyCar as an example, after applying maximum expansion uniform sampling, the number of samples is balanced across categories within each attribute. However, merely having a balanced quantity does not mean the data distribution is suitable for training classifiers. Therefore, we apply robust data transformation to transform the data. For each sample, there is a 50% chance to be applied with transformations of ‘AddGaussianNoise’, ‘TimeStretch’, ‘PitchShift’, and ‘TimeShift’, which can be combined. After applying robust data transformation to every sample, we discard the original samples. This completes imbalanced compensation.
In summary, the application of maximum expansion uniform sampling balanced the extremely imbalanced data between operating status. After maximum expansion uniform sampling, robust data transformation was applied, and each sampling was accompanied by 4 types of audio time domain conversion, which simulated various noises in real situations to some extent and improved the robustness of the model at the data level. In addition, the oversampling technique increased the sample size and achieved class balance, which solved the problem that the classifier was difficult to train.

2.3. Ensemble Attribute Anomaly Detector

Currently, in the field of anomaly detection, probability-based confidence methods [24,25] have been widely used. Specifically, this kind of method trains a classifier on normal samples for classification. During testing, normal samples will be classified into known categories by the classifier, while abnormal samples are difficult to distinguish. Intuitively, abnormal samples will receive a lower confidence score. However, incorrectly classifying samples during testing has a catastrophic impact on the performance of anomaly detection [26]. In addition, this method relies on the quality of model training, but, in practical use, the model is often overfitting, which will also affect the performance of anomaly detection.
Instead, we propose ensemble attribute anomaly detector. The key is to combine the traditional machine learning algorithm KNN with the classifier obtained from deep learning to improve the fault tolerance and robustness of the model for anomaly detection. The detailed algorithm of our proposed module is presented in Algorithm 2. Utilizing the data after imbalanced compensation, which are also used to train the classifiers, we train M separate KNN models. Specifically, the embeddings extracted from the trained M classifiers are leveraged as quality training data for each KNN. After training a KNN search tree for each model, test embeddings are extracted by passing the test sample through the corresponding classifier. The trained m-th KNN search tree is then utilized to find the t o p K nearest neighbors of the test embedding, constructing the set T k ( e test ) . Subsequently, the Euclidean distance d ( e t e s t ) between e t e s t and the samples in T k ( e t e s t ) is computed to obtain the distance matrix D t e s t . The anomaly score is calculated as the maximum value in the distance matrix
S m = max ( D t e s t )
Algorithm 2 KNN for anomaly detection from the perspective of the m-th attribute
Input: Train data: { x i k m I C } i = 1 , k = 1 R , K m , Test data: x t e s t , Trained m-th classifier
Output: Anomaly score S m for test sample x t e s t
  1: Extract embeddings { e i k m I C } i = 1 , k = 1 R , K m by the m-th classifier from { x i k m I C } i = 1 , k = 1 R , K m
  2: Extract embedding e t e s t by the m-th classifier from x t e s t
  3: Construct KNN search T R E E from { e i k m I C } i = 1 , k = 1 R , K m
  4: Find t o p K nearest neighbors of e t e s t using T R E E
  5: Let T k ( e t e s t ) be the set comprising the t o p K nearest neighbors of e t e s t
  6: Compute distance d ( e t e s t ) between x t e s t and samples in T k ( e t e s t )
  7: Obtain score set D t e s t of test distance
  8: return Anomaly score S m = max ( D t e s t )
Finally, we take the mean of the results obtained for each attribute in the score domain to obtain the final ensemble anomaly value
S = m = 1 M S m M
Although KNN is a classic machine learning method, it is prone to curse of dimensionality when dealing with high-dimensional data, such as audio in this work. Naturally, we thought of using the outstanding feature extraction capability of deep neural networks to reduce the dimension of audio data to a low-dimensional space that KNN can characterize. Therefore, we use the attribute classifier mentioned above as a proxy task for the anomaly detection task and obtain supervision by distinguishing different operating status. After the training is completed, in the latent space, the samples of each operating status will gather together, while abnormal samples will be exposed because they are difficult to distinguish. Notably, our work trains multiple classifiers for each attribute, which enables each classifier to distinguish abnormal samples from different attribute perspectives. Such an operation improves the fault tolerance of anomaly detection. Even if one classifier makes a mistake, the results of other classifiers can compensate the errors.
Taking ToyCar as an example, after applying the imbalanced compensation module, we pretrain three separate classifiers for the three attributes, respectively. Meanwhile, using the training data after imbalanced compensation, three different embeddings are extracted via the three classifiers, which we term as embedding spaces. During testing, a test sample is fed into the three pretrained classifiers to obtain three test embeddings, each corresponding to one embedding space. Then, we apply a KNN algorithm to retrieve the t o p K nearest neighbors for each test embedding in its embedding space. The Euclidean distances between the test embedding and its t o p K neighbors are calculated. After obtaining the three sets of Euclidean distances, we take their average as the final anomaly score.
Therefore, for anomaly detection, the three different attributes provide three distinct detection perspectives for the same test sample. Fusing their scores allows the three perspectives to complement each other. Meanwhile, the imbalanced compensation technique enables classifier training and enhances classifier robustness through data augmentation. The resultant high-quality embeddings together with the proposed ensemble attribute anomaly detector boost anomaly detection performance.

3. Experimental Setup

3.1. Datasets

We evaluate our proposed approach on the development dataset provided for the DCASE2023 Challenge Task 2 [3], which contains two subsets: ToyADMOS2 [21] and MIMII DG [22]. The development dataset includes normal and anomalous operating sounds of seven types of machines recorded in single-channel, including Fan, Gearbox, Bearing, Slide rail (slider), Valve, ToyCar, and ToyTrain. For each of the 7 machine types, the dataset provides (1) 990 normal sound clips of 10 s length downsampled to 16 kHz for training in the source domain, (2) 10 normal sound clips for training in the target domain, and (3) 100 clips each of normal and anomalous sounds for testing. The source/target domain labels and attribute labels in the train dataset of each sample are provided, while the test dataset is not. The overview of datasets is shown in Figure 3.
This work focuses on attribute labels, which are mentioned in Section 2. Unlike machine IDs that may be unavailable, attribute labels are labels extracted from metadata. Attributes such as operating speed, operating voltage, etc., necessarily accompany the operation of the machine, so obtaining such labels is actually feasible. Unfortunately, there is no free lunch in the world, and such labels are extremely imbalanced and inconsistent, posing challenges to the design of our anomaly detection system. The various labels in the dataset are shown in Figure 4.

3.2. Evaluation Metrics

To evaluate the performance of the proposed model, we adopt the area under the curve (AUC) of the receiver operating characteristic (ROC) as the evaluation metric [9]. The ROC curve is created by plotting the true positive rate (TPR) against the false positive rate (FPR) at various threshold settings. The AUC measures the entire two-dimensional area underneath the ROC curve, which represents the degree or measure of separability between normal and anomalous instances. An AUC of 1 represents a perfect classifier, while an AUC of 0.5 represents a worthless classifier. Compared with metrics such as accuracy, AUC provides a more comprehensive evaluation of the model’s performance in the imbalance scenario. Therefore, AUC is more suitable for evaluating anomaly detection methods where negative samples dominate. Furthermore, the pAUC is also used in this work, which is calculated as the AUC over a low FPR range [0, p]. The AUC and pAUC are defined as
AUC = 1 N N + i = 1 N j = 1 N + H ( S ( x j + ) S ( x i ) )
p AUC = 1 p N N + i = 1 p N j = 1 N + H ( S ( x j + ) S ( x i ) )
where · is the flooring function, S ( x i ) and S ( x j + ) mean the score of normal and anomalous test clips, respectively. N and N + represent the number of them, respectively. Otherwise, the function H ( x ) is
H ( x ) = 1 , x > 0 0 , x 0
In practical acoustic-sensing-based ASD scenarios, a lower FPR is required. Therefore, we set p = 0.1 .

3.3. Implementation Details

For data preprocessing, we first converted the raw audio signals to log-mel-spectrogram using a short-time Fourier transform (STFT) with a window size of 1024 and a hop length of 512. A mel filterbank with 128 filters was applied and the magnitude of the STFT was converted to decibels. In this study, we use 128-dimensional log-mel-spectrogram as input features for the classifier. We adopt ResNet18 [23] as the backbone to design our classifier. The model is optimized using the Adam [27] optimizer with a learning rate of 0.0001 and trained for 15 epochs with a batch size of 128. In addition, the data augmentation used in the IC module uses the audiomendation [28] toolkit, and the KNN training uses the Pyod toolkit [29], the t o p K = 5. In this work, the number of samples R for the imbalanced compensation module is set to 4096.

4. Experimental Results

In this section, we present a comprehensive analysis of the experimental results. We conducted experiments on the development dataset released in DCASE2023 Challenge Task 2 and compared the experimental results with the official baseline AE. In addition, to verify the effectiveness of the imbalanced compensation module, we applied the module on the AE baseline, called AEIC, and obtained competitive performance. Finally, we ensemble AEIC with our AIC model, whose performance ranked sixth in DCASE2023 Challenge Task 2.

4.1. Signal Analysis

Considering the abstract nature of audio signals, we are unable to analyze them through direct observation of the waveforms. Furthermore, the audio in acoustic-sensing-based ASD tasks cannot be distinguished as normal or anomalous through the human ear. Therefore, we transform the signals into the frequency domain through STFT and observe them in the Mel scale, namely the mel-spectrogram.
As shown in the Figure 5, we plot the mel-spectrograms of the audio signals from seven machines. Based on the mel-spectrograms, we can simply classify the audio signals of these seven machines into stationary and non-stationary signals. It is noteworthy that the signal of the Slider machine appears to be non-stationary but is actually a periodic stationary signal [30]. The specific classification is as follows:
  • Stationary signals: ToyCar, Bearing, Fan, Slider
  • Non-stationary signals: ToyTrain, Gearbox, Valve

4.2. Results

The proposed AIC model, as illustrated in Figure 1, utilizes the weak attribute labels to train a classifier without machine ID and embeds the data with the classifier to expose anomalies. The experimental results are shown in Table 3, where AE-MSE and AE-MAHA are two official baselines. Both utilize autoencoders trained solely on normal samples with an MSE loss function. The difference lies in the testing phase, where AE-MSE uses the MSE as the anomaly score while AE-MAHA employs the Mahalanobis distance. AC is the result of direct attribute classification without the IC module.
For evaluation metrics, AUCs and AUCt denote the AUC of the model on the source domain data and target domain data, respectively. pAUC represents AUC at low FPR, as mentioned in Section 3. To measure the model performance under these three metrics of AUCs, AUCt, and pAUC, we take the harmonic mean of them, denoted by ‘hmean’. Similarly, to measure the model performance across the seven machines of data, we take the harmonic mean of each metric across the seven machines. This is also denoted by ‘hmean’ for consistency. By taking the harmonic mean of metrics on each subset, we summarize the performance across machines into a single representative value.
The experimental results show that the proposed AIC model is competitive compared with the two AE-based baselines, demonstrating the potential of discriminative models based on weak attribute labels for anomaly detection tasks. Notably, attribute classifier performing direct classification of the raw attribute labels achieved poor performance, indicating that the extremely imbalanced sample and other issues of the attribute such as weak labels make the classifier untrainable. However, the AIC model with the imbalanced compensation module, compared with attribute classifier, gained significant performance improvements. This shows that the proposed imbalanced compensation module effectively alleviated the weak label problem of the attributes. Moreover, as observed in the baseline results and other previous findings in the literature [32], it is interesting to find that the experimental results of the seven machines show inconsistent patterns. For example, the best model performance on ToyCar is not the same one as with ToyTrain. This is mainly because the seven machines have different acoustic characteristics, which makes it difficult to design a universally applicable model.

4.3. Imbalanced Compensation Module Analysis

As analyzed above, the proposed imbalanced compensation module plays a crucial role in the entire anomaly detection model. Therefore, this section further explores the effectiveness of the imbalanced compensation module and how to determine the number of samples S in the imbalanced compensation module.
First, the imbalanced compensation module is applied to two official baselines, AE-MSE and AE-MAHA. Although AE itself does not utilize attribute labels, it has source domain and target domain labels, with imbalanced scenarios similar to those of attribute labels. Therefore, the imbalanced compensation module is applied according to domain labels. As shown in Table 4, AEIC-MSE and AEIC-MAHA are two baselines with the applied imbalanced compensation module, which achieved significant improvements over the original baselines. This demonstrates the universal effectiveness of the proposed imbalanced compensation module for both generative and discriminative models.
Furthermore, the performance changes in the AIC model with different numbers of imbalanced compensation samples R 1024, 2048, 4096, 8192 are explored. The harmonic mean of AUC on seven machines is used as the evaluation metric. We finally chose 4096 samples. As shown in Figure 6, surprisingly, the model performance of four stable signal machines, ToyCar, Bearing, Fan, and Slider, increases with the increasing number of samples. In contrast, the performance of three non-stable signal machines, ToyTrain, Gearbox, and Valve, cannot be improved with more samples. This could be because the time-domain data transformation of non-stable signals causes distortion, while that of stable signals helps to improve the model’s robustness.

4.4. Visualization

To better demonstrate the performance of AIC, t-SNE [33] is used to visualize the training set and test set. As shown in Figure 7, dots of different colors represent normal samples of the source domain in the training set, normal samples of the target domain in the training set, normal samples of the source domain in the test set, normal samples of the target domain in the test set, abnormal samples of the source domain in the test set, and abnormal samples of the target domain in the test set, respectively. The embeddings of attribute classifier and AIC are extracted separately for visualization. Since the model training objective is attribute classification, abnormal samples are hard to be classified into any category, thereby exposing the abnormal samples. This is manifested as abnormal samples are far from normal samples, forming lower-density areas in the visualization.
Comparing Figure 7a,b, taking Fan as an example, it is found that, for attribute classifier without the imbalanced compensation module, a small portion of normal samples are misclassified into areas close to abnormal samples, which will damage the anomaly detection performance [26]. However, for AIC with the applied imbalanced compensation, the misclassified normal samples disappear. This shows that the proposed AIC model alleviates the problem of normal sample misclassification.
Comparing Figure 7c,d, taking Slider as an example, it is found that, compared with attribute classifier, AIC with the applied imbalanced compensation forms a more compact data distribution. In anomaly detection tasks, the more compact the distribution of normal samples, the lower the density of abnormal samples, which is conducive to abnormal sample detection [32]. This shows that the proposed AIC model helps normal samples form a more compact distribution.

4.5. Ensemble

Finally, anomaly detection is performed by fusing the scores of AEIC-MAHA and AIC models through model ensemble. As shown in Table 5, this fusion of generative and discriminative models significantly improves the anomaly detection performance, ranking sixth internationally in the DCASE2023 Challenge Task 2. The score of model fusion can be expressed as
S e n s e m b l e = S A E I C + λ S A I C
By changing the value of lambda multiple times, we obtained the optimal performance of the ensemble model. In this work, we chose λ = 0.3. Figure 8 shows the relationship between the value of λ and the AUC of the seven machines. Here, AUC refers to the hmean of AUCs, AUCt, and pAUC. Experiments show that the AIC model has high complementarity with the AE model.

5. Limitations and Conclusions

Acoustic sensing provides crucial data for ASD in machine condition monitoring. In the MIMII DG [22] dataset, the machines Fan, Valve, Gearbox, Bearing, and Slider were recorded by a TAMAGO-03 microphone. The Fan and Valve were recorded in a sound-proof room, while the Gearbox, Bearing, and Slider were recorded in an anechoic chamber. In the TOYADMOS2 [21] dataset, the ToyCar and ToyTrain were recorded using either a SURE SM11-CN or TOMOCA EM-700 microphone, which introduced some domain shift. Notably, machines other than these seven are unknown, so transferring trained models to unseen machines and enabling anomaly detection is an important area for future research.
While we have utilized all the available datasets from the DCASE Challenge, the experiments are still limited in scope. Regarding audio signals, relatively comprehensive experiments have been conducted to demonstrate the efficacy of the proposed method. However, as the number of samples provided in the datasets is small, it provides challenges to the statistical significance of the results and generalizability of the conclusions. It is noteworthy that simple data augmentation techniques did not lead to significant performance gains for this task. In addition, for other types of signals such as vibration that lack officially provided data, they are not covered in our current work.
Despite the stated limitations, this work makes several valuable contributions. While the datasets are limited, our extensive experiments nonetheless validate the effectiveness of the proposed method for acoustic signals. Our proposed AIC method enables the use of discriminative models for DCASE2023 Task 2, complementing the reconstruction-based AE approach. The AIC model outperforms the baseline AE, presenting an alternative solution direction for this task. Additionally, our proposed IC module is applicable to AE models as well. Incorporating the imbalanced compensation module improves AE-MSE and AE-MAHA to 57.36% and 59.32%, respectively, demonstrating the versatility of our design. Finally, ensembling the AEIC-MAHA and AIC models yields a 9.7% performance gain over the baseline AE-MSE system. This results in a ranking of sixth place internationally in DCASE2023 Challenge Task 2, proving highly competitive. Our contributions not only advance the performance of the algorithm through the novel AIC method but also have a broader impact by enhancing traditional AE models. The consistent improvements across architectures highlight the significance of the ideas introduced in this work.
In conclusion, this work innovatively proposes the AIC framework for acoustic-sensing-based ASD, making it possible to perform anomaly detection with discriminative models without machine ID. The proposed imbalanced compensation module improves both the AIC framework and traditional AE methods significantly. Through integrating AIC and AE using the imbalanced compensation module via model ensemble, we have achieved highly competitive performance. Furthermore, the efficacy of the proposed imbalnce compensation module is further corroborated through t-SNE visualization, which exhibits clear separation between normal and anomalous data. Future work will validate the generalizability of the proposed method on different data types, such as vibration data. Additionally, we will examine the generalization capabilities of our algorithm on other genres of acoustic data.

Author Contributions

Conceptualization, Y.Z. and Y.L.; methodology, Y.Z.; software, Y.Z.; validation, Y.Z., Y.L. and H.W.; writing—original draft preparation, Y.Z.; writing—review and editing, Y.L.; visualization, Y.Z.; supervision, Y.L.; project administration, Y.Z.; funding acquisition, Y.L. and H.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data used in this work were derived from the DCASE2023 Challenge Task 2. All data is available in https://zenodo.org/records/7882613 (accessed on 2 November 2023).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
IDIdentity
ASDAnomalous Sound Detection
ACAttribute Classifier
ICImbalanced Compensation
AICAttribute-driven Imbalanced Compensation
AEAutoencoder
DCASEDetection and Classification of Acoustic Scenes and Events
MEUSMaximum Expansion Uniform Sampling
RDTRobust Data Transformation
STFTShort-Time Fourier Transform

References

  1. Kawaguchi, Y.; Imoto, K.; Koizumi, Y.; Harada, N.; Niizumi, D.; Dohi, K.; Tanabe, R.; Purohit, H.; Endo, T. Description and discussion on DCASE 2021 challenge task 2: Unsupervised anomalous sound detection for machine condition monitoring under domain shifted conditions. In Proceedings of the Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, Online, 15–19 November 2021. [Google Scholar]
  2. Dohi, K.; Imoto, K.; Harada, N.; Niizumi, D.; Koizumi, Y.; Nishida, T.; Purohit, H.; Endo, T.; Yamamoto, M.; Kawaguchi, Y. Description and discussion on DCASE 2022 challenge task 2: Unsupervised anomalous sound detection for machine condition monitoring applying domain generalization techniques. In Proceedings of the Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, Nancy, France, 3–4 November 2022. [Google Scholar]
  3. Dohi, K.; Imoto, K.; Harada, N.; Niizumi, D.; Koizumi, Y.; Nishida, T.; Purohit, H.; Tanabe, R.; Endo, T.; Kawaguchi, Y. Description and Discussion on DCASE 2023 Challenge Task 2: First-Shot Unsupervised Anomalous Sound Detection for Machine Condition Monitoring. arXiv 2023, arXiv:2305.07828. [Google Scholar]
  4. Li, X.; Yu, S.; Lei, Y.; Li, N.; Yang, B. Intelligent Machinery Fault Diagnosis with Event-Based Camera. IEEE Trans. Ind. Inform. 2023. [Google Scholar] [CrossRef]
  5. Attestog, S.; Senanayaka, J.S.L.; Van Khang, H.; Robbersmyr, K.G. Robust active learning multiple fault diagnosis of PMSM drives with sensorless control under dynamic operations and imbalanced datasets. IEEE Trans. Ind. Inform. 2023, 19, 9291–9301. [Google Scholar] [CrossRef]
  6. Bustillo, A.; Urbikain, G.; Perez, J.M.; Pereira, O.M.; de Lacalle, L.N.L. Smart optimization of a friction-drilling process based on boosting ensembles. J. Manuf. Syst. (JMS) 2018, 48, 108–121. [Google Scholar] [CrossRef]
  7. Sharma, S.; Tiwari, S. A novel feature extraction method based on weighted multi-scale fluctuation based dispersion entropy and its application to the condition monitoring of rotary machines. Mech. Syst. Signal Process. (MSSPS) 2022, 171, 108909. [Google Scholar] [CrossRef]
  8. Sharma, S.; Tiwari, S.; Singh, S. Integrated approach based on flexible analytical wavelet transform and permutation entropy for fault detection in rotary machines. Measurement 2021, 169, 108389. [Google Scholar] [CrossRef]
  9. Koizumi, Y.; Kawaguchi, Y.; Imoto, K.; Nakamura, T.; Nikaido, Y.; Tanabe, R.; Purohit, H.; Suefusa, K.; Endo, T.; Yasuda, M.; et al. Description and discussion on DCASE2020 challenge task2: Unsupervised anomalous sound detection for machine condition monitoring. In Proceedings of the Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, Virtual, 2–4 November 2020. [Google Scholar]
  10. Karapalidou, E.; Alexandris, N.; Antoniou, E.; Vologiannidis, S.; Kalomiros, J.; Varsamis, D. Implementation of a Sequence-to-Sequence Stacked Sparse Long Short-Term Memory Autoencoder for Anomaly Detection on Multivariate Timeseries Data of Industrial Blower Ball Bearing Units. Sensors 2023, 23, 6502. [Google Scholar] [CrossRef] [PubMed]
  11. Abbasi, S.; Famouri, M.; Shafiee, M.J.; Wong, A. OutlierNets: Highly compact deep autoencoder network architectures for on-device acoustic anomaly detection. Sensors 2021, 21, 4805. [Google Scholar] [CrossRef] [PubMed]
  12. Jiang, A.; Zhang, W.Q.; Deng, Y.; Fan, P.; Liu, J. Unsupervised Anomaly Detection and Localization of Machine Audio: A Gan-Based Approach. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  13. Suefusa, K.; Nishida, T.; Purohit, H.; Tanabe, R.; Endo, T.; Kawaguchi, Y. Anomalous sound detection based on interpolation deep neural network. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020; pp. 271–275. [Google Scholar]
  14. Hatanaka, S.; Nishi, H. Efficient GAN-based unsupervised anomaly sound detection for refrigeration units. In Proceedings of the IEEE International Symposium on Industrial Electronics (ISIE), Kyoto, Japan, 20–23 June 2021; pp. 1–7. [Google Scholar]
  15. Dohi, K.; Endo, T.; Purohit, H.; Tanabe, R.; Kawaguchi, Y. Flow-based self-supervised density estimation for anomalous sound detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Toronto, ON, Canada, 6–11 June 2021; pp. 336–340. [Google Scholar]
  16. Chen, H.; Ran, L.; Sun, X.; Cai, C. SW-WAVENET: Learning Representation from Spectrogram and Wavegram Using Wavenet for Anomalous Sound Detection. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  17. Wang, Y.; Zheng, Y.; Zhang, Y.; Xie, Y.; Xu, S.; Hu, Y.; He, L. Unsupervised Anomalous Sound Detection for Machine Condition Monitoring Using Classification-Based Methods. Appl. Sci. 2021, 11, 11128. [Google Scholar] [CrossRef]
  18. Hojjati, H.; Armanfard, N. Self-supervised acoustic anomaly detection via contrastive learning. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 3253–3257. [Google Scholar]
  19. Giri, R.; Tenneti, S.V.; Cheng, F.; Helwani, K.; Isik, U.; Krishnaswamy, A. Self-supervised classification for detecting anomalous sounds. In Proceedings of the Detection and Classification of Acoustic Scenes and Events Workshop (DCASE), Virtual, 2–4 November 2020. [Google Scholar]
  20. Liu, Y.; Guan, J.; Zhu, Q.; Wang, W. Anomalous sound detection using spectral-temporal information fusion. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Singapore, 23–27 May 2022; pp. 816–820. [Google Scholar]
  21. Harada, N.; Niizumi, D.; Takeuchi, D.; Ohishi, Y.; Yasuda, M.; Saito, S. ToyADMOS2: Another Dataset of Miniature-Machine Operating Sounds for Anomalous Sound Detection under Domain Shift Conditions. In Proceedings of the Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, Barcelona, Spain, 15–19November 2021; pp. 1–5. [Google Scholar]
  22. Dohi, K.; Nishida, T.; Purohit, H.; Tanabe, R.; Endo, T.; Yamamoto, M.; Nikaido, Y.; Kawaguchi, Y. MIMII DG: Sound Dataset for Malfunctioning Industrial Machine Investigation and Inspection for Domain Generalization Task. In Proceedings of the Detection and Classification of Acoustic Scenes and Events (DCASE) Workshop, Nancy, France, 3–4 November 2022. [Google Scholar]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  24. Hendrycks, D.; Gimpel, K. A Baseline for Detecting Misclassified and Out-of-Distribution Examples in Neural Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
  25. Hendrycks, D.; Mazeika, M.; Dietterich, T. Deep Anomaly Detection with Outlier Exposure. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  26. CEN, J.; Luan, D.; Zhang, S.; Pei, Y.; Zhang, Y.; Zhao, D.; Shen, S.; Chen, Q. The Devil is in the Wrongly-classified Samples: Towards Unified Open-set Recognition. In Proceedings of the International Conference on Learning Representations (ICLR), Kigali, Rwanda, 1–5 May 2023. [Google Scholar]
  27. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  28. Asteroid-Team. Torch-Audiomentations. 2022. Available online: https://github.com/asteroid-team/torch-audiomentations (accessed on 1 May 2023).
  29. Zhao, Y.; Nasrullah, Z.; Li, Z.; Han, S.; Hu, X.; Huang, H.; Jiang, M.; Zhao, Y. PyOD: A Python Toolbox for Scalable Outlier Detection. J. Mach. Learn. Res. (JMLR) 2020, 20, 1–7. [Google Scholar]
  30. Jiang, A.; Hou, Q.; Liu, J.; Fan, P.; Ma, J.; Lu, C.; Zhai, Y.; Deng, Y.; Zhang, W.Q. THUEE System for First-Shot Unsupervised Anomalous Sound Detection for Machine Condition Monitoring. In Proceedings of the IEEE AASP Challenge on Detection and Classification of Acoustic Scenes and Events, Tampere, Finland, 20–22 September 2023. Technical report. [Google Scholar]
  31. Harada, N.; Niizumi, D.; Ohishi, Y.; Takeuchi, D.; Yasuda, M. First-shot anomaly sound detection for machine condition monitoring: A domain generalization baseline. arXiv 2023, arXiv:2303.00455. [Google Scholar]
  32. Fang, Z.; Li, Y.; Lu, J.; Dong, J.; Han, B.; Liu, F. Is Out-of-Distribution Detection Learnable? In Proceedings of the Advances in Neural Information Processing Systems (NIPS), New Orleans, LA, USA, 28 November 2022.
  33. Van der Maaten, L.; Hinton, G. Visualizing data using t-SNE. J. Mach. Learn. Res. (JMLR) 2008, 9, 2579–2605. [Google Scholar]
Figure 1. The framework of proposed AIC.
Figure 1. The framework of proposed AIC.
Sensors 23 08984 g001
Figure 2. A schematic diagram of the effects of imbalanced compensation on data. ‘Catagory’ means the catagory of operating status.
Figure 2. A schematic diagram of the effects of imbalanced compensation on data. ‘Catagory’ means the catagory of operating status.
Sensors 23 08984 g002
Figure 3. An overview of datasets.
Figure 3. An overview of datasets.
Sensors 23 08984 g003
Figure 4. The taxonomy of various labels in the dataset.
Figure 4. The taxonomy of various labels in the dataset.
Sensors 23 08984 g004
Figure 5. Mel-spectrograms of the 7 machines. (a) ToyCar; (b) ToyTrain; (c) Bearing; (d) Fan; (e) Gearbox; (f) Slider; (g) Valve.
Figure 5. Mel-spectrograms of the 7 machines. (a) ToyCar; (b) ToyTrain; (c) Bearing; (d) Fan; (e) Gearbox; (f) Slider; (g) Valve.
Sensors 23 08984 g005aSensors 23 08984 g005b
Figure 6. The relationship between the number of samples R in the imbalanced compensation module and model performance.
Figure 6. The relationship between the number of samples R in the imbalanced compensation module and model performance.
Sensors 23 08984 g006
Figure 7. t-SNE visualization comparison between attribute classifier and AIC. (a) AC_fan; (b) AIC_fan; (c) AC_slider; (d) AIC_slider.
Figure 7. t-SNE visualization comparison between attribute classifier and AIC. (a) AC_fan; (b) AIC_fan; (c) AC_slider; (d) AIC_slider.
Sensors 23 08984 g007
Figure 8. System ensemble performance with the varying of score fusion weight λ .
Figure 8. System ensemble performance with the varying of score fusion weight λ .
Sensors 23 08984 g008
Table 1. Attributes of different machines.
Table 1. Attributes of different machines.
Attribute 1Attribute 2Attribute 3
ToyCarCar modelSpeedMic
ToyTrainTrain modelSpeedMic
FanMixing of machineN/AN/A
GearboxVoltageWeightN/A
BearingVelocityMicN/A
SliderVelocityAccelerationN/A
ValveOpen/closeN/AN/A
Table 2. Different operating status labels (‘Category: #samples’) of each attribute in ToyCar dataset.
Table 2. Different operating status labels (‘Category: #samples’) of each attribute in ToyCar dataset.
Car ModelSpeedMic
Label 1C1: 2153.1 V: 3501: 990
Label 2D1: 2144.0 V: 3502: 10
Label 3B1: 1663.4 V: 290N/A
Label 4B2: 1642.8 V: 5N/A
Label 5D2: 1163.7 V: 5N/A
Label 6C2: 115N/AN/A
Label 7A1: 3N/AN/A
Label 8E2: 3N/AN/A
Label 9A2: 2N/AN/A
Label 10E1: 2N/AN/A
Table 3. Performance of AUC and pAUC (p = 0.1) comparison on 7 different machines. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
Table 3. Performance of AUC and pAUC (p = 0.1) comparison on 7 different machines. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
AE-MSE [31]hmeanAE-MAHA [31]hmeanAChmeanAIChmean
AUCsAUCt p AUCAUCsAUCt p AUCAUCsAUCt p AUCAUCsAUCt p AUC
ToyCar70.1046.8952.4754.8974.5343.4249.1852.8362.3737.6250.8448.1772.2350.0553.8957.27
ToyTrain57.9357.0248.5754.1655.9842.4548.1348.2361.5861.8253.7858.8163.4048.6851.4253.80
Bearing65.9255.7550.4256.6765.1655.2851.3756.7160.9848.9352.3153.6279.8654.6457.4762.21
Fan80.1936.1859.0452.5987.1045.9859.3359.9073.0228.1651.2643.6665.1879.6263.7368.82
Gearbox60.3160.6953.2257.8671.8870.7854.3464.6060.5551.4754.0052.4052.4060.5654.1555.49
Slider70.3148.7756.3757.1884.0273.2954.7268.4683.6858.4650.7861.5482.6456.4454.3162.20
Valve55.3550.6951.1852.3356.3151.4051.0852.8369.4316.3447.7331.0771.3834.8549.7847.78
hmean64.7949.5952.8455.0268.8452.3752.3656.9166.5235.6351.4547.6768.1852.1254.6657.53
Table 4. Performance comparisons after applying imbalanced compensation to AE baselines. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
Table 4. Performance comparisons after applying imbalanced compensation to AE baselines. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
 AE-MSE [31]hmeanAE-MAHA [31]hmeanAEIC-MSEhmeanAEIC-MAHAhmean
AUCsAUCt p AUCAUCsAUCt p AUCAUCsAUCt p AUCAUCsAUCt p AUC
ToyCar70.1046.8952.4754.8974.5343.4249.1852.8359.6464.1651.0557.7670.2459.6449.2158.45
ToyTrain57.9357.0248.5754.1655.9842.4548.1348.2355.4858.8848.3653.8750.2248.8747.7348.92
Bearing65.9255.7550.4256.6765.1655.2851.3756.7163.9862.8051.2658.7560.3863.5452.2158.31
Fan80.1936.1859.0452.5987.1045.9859.3359.9085.8662.4063.7869.2081.8085.4469.6378.35
Gearbox60.3160.6953.2257.8671.8870.7854.3464.6065.6464.8054.7861.3273.9269.3652.1063.64
Slider70.3148.7756.3757.1884.0273.2954.7268.4663.2247.0454.9454.2779.7070.6151.7865.19
Valve55.3550.6951.1852.3356.3151.4051.0852.8351.0247.9851.4750.1154.4048.6651.1551.30
hmean64.7949.5952.8455.0268.8452.3752.3656.9162.1057.3553.3057.3665.1861.5152.6959.32
Table 5. Performance of the system ensemble. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
Table 5. Performance of the system ensemble. ‘hmean’ represents the harmonic mean of AUCs, AUCt, and pAUC.
 AEIC-MAHAhmeanAIChmeanEnsemblehmean
AUCsAUCt p AUCAUCsAUCt p AUCAUCsAUCt p AUC
ToyCar70.2459.6449.2158.4572.2350.0553.8957.2772.8757.7449.7858.67
ToyTrain50.2248.8747.7348.9263.4048.6851.4253.8059.2247.2548.0550.97
Bearing60.3863.5452.2158.3179.8654.6457.4762.2176.6858.3953.1061.22
Fan81.8085.4469.6378.3565.1879.6263.7368.8280.8987.1670.1578.76
Gearbox73.9269.3652.1063.6452.4060.5654.1555.4971.1669.5053.9463.86
Slider79.7070.6151.7865.1982.6456.4454.3162.2085.5369.3651.9466.13
Valve54.4048.6651.1551.3071.3834.8549.7847.7864.1242.7351.4251.33
hmean65.1861.5152.6959.3268.1852.1254.6657.5371.9058.6853.3460.37
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhou, Y.; Long, Y.; Wei, H. Acoustic-Sensing-Based Attribute-Driven Imbalanced Compensation for Anomalous Sound Detection without Machine Identity. Sensors 2023, 23, 8984. https://doi.org/10.3390/s23218984

AMA Style

Zhou Y, Long Y, Wei H. Acoustic-Sensing-Based Attribute-Driven Imbalanced Compensation for Anomalous Sound Detection without Machine Identity. Sensors. 2023; 23(21):8984. https://doi.org/10.3390/s23218984

Chicago/Turabian Style

Zhou, Yifan, Yanhua Long, and Haoran Wei. 2023. "Acoustic-Sensing-Based Attribute-Driven Imbalanced Compensation for Anomalous Sound Detection without Machine Identity" Sensors 23, no. 21: 8984. https://doi.org/10.3390/s23218984

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop