Next Article in Journal
Sensitivity Analysis of Numerical Coherency Model for Rock Sites
Previous Article in Journal
Study of Synergetic Optimization Operation for Distribution Network Considering Multiple Reactive Power Output Modes of Photovoltaics and Different Port Numbers of Flexible Interconnection Devices
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Self-Supervised Method for Speaker Recognition in Real Sound Fields with Low SNR and Strong Reverberation

1
School of Instrument and Electronics, North University of China, Taiyuan 030051, China
2
School of Information and Communication Engineering, North University of China, Taiyuan 030051, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2025, 15(6), 2924; https://doi.org/10.3390/app15062924
Submission received: 7 February 2025 / Revised: 2 March 2025 / Accepted: 4 March 2025 / Published: 7 March 2025
(This article belongs to the Section Computing and Artificial Intelligence)

Abstract

:
Speaker recognition is essential in smart voice applications for personal identification. Current state-of-the-art techniques primarily focus on ideal acoustic conditions. However, the traditional spectrogram struggles to differentiate between noise, reverberation, and speech. To overcome this challenge, MFCC can be replaced with the output from a self-supervised learning model. This study introduces a TDNN enhanced with a pre-trained model for robust performance in noisy and reverberant environments, referred to as PNR-TDNN. The PNR-TDNN employs HuBERT as its backbone, while the TDNN is an improved ECAPA-TDNN. The pre-trained model employs the Canopy/Mini Batch k-means++ strategy. In the TDNN architecture, several enhancements are implemented, including a cross-channel fusion mechanism based on Res2Net. Additionally, a non-average attention mechanism is applied to the pooling operation, focusing on the weight information of each channel within the Squeeze-and-Excitation Net. Furthermore, the contribution of individual channels to the pooling of time-domain frames is enhanced by substituting attentive statistics with multi-head attention statistics. Validated by zhvoice in noisy conditions, the minimized PNR-TDNN demonstrates a 5.19% improvement in EER compared to CAM++. In more challenging environments with noise and reverberation, the minimized PNR-TDNN further improves EER by 3.71% and 9.6%, respectively, and MinDCF by 3.14% and 3.77%, respectively. The proposed method has also been validated on the VoxCeleb1 and cn-celeb_v2 datasets, representing a significant breakthrough in the field of speaker recognition under challenging conditions. This advancement is particularly crucial for enhancing safety and protecting personal identification in voice-enabled microphone applications.

1. Introduction

Speaker recognition encompasses tasks such as verification, identification, and diarization [1,2,3]. The process involves the analysis of voiceprint features extracted from speech, which are shaped by individual variations in vocal tract structure, pronunciation, and accent [4]. Comparable to fingerprint, facial, and iris recognition, voiceprint recognition offers the benefits of contactless operation and high accuracy. It finds extensive application in access control, user authentication, and forensic evidence collection, and has been integrated into various smart microphone systems, including cellphones, smart homes, and smart vehicles [3,5,6,7,8].
Recently, deep neural networks have achieved significant advancements in speaker recognition [9]. Many studies have focused on enhancing feature extraction by using advanced DNN architectures, such as Time-Delay Neural Networks (TDNNs) [10] and Residual Networks (ResNets) [11]. The X-Vector model, proposed by [10], was introduced as an improvement over the D-Vector model [12]. Unlike the frame-by-frame modeling approach employed in the D-Vector model, the X-Vector model utilizes a pooling layer to capture speech embedding units. It also incorporates data augmentation techniques to enhance the performance of TDNN-based speaker recognition systems. In further research on speaker recognition tasks, feature extraction models based on Deep Neural Networks (DNNs) incorporate various architectures [5,13,14,15,16,17]. For instance, ref.  [18] utilizes Convolutional Neural Networks (CNNs) to locally extract features from speech, while [19] employs Recurrent Neural Networks (RNNs) to capture temporal dependencies and patterns within the speech. Specifically, ref. [20] innovated the X-Vector approach by incorporating techniques from computer vision, namely the Squeeze Excitation network, Res2Net, and the channel propagation aggregation mechanism. This integration significantly improved speaker recognition performance, marking it as one of the most effective systems of its time.
In environments with an ideal sound field, which are free from ambient noise, sensing noise, interference from other speakers, and echoes, numerous deep neural network approaches for speaker recognition have demonstrated high performance in critical metrics such as EER and MinDCF. However, in actual sound field conditions, it is typical to experience sudden occurrences of high-energy noise and significant speech reverberation. For instance, in a confined conference room, sounds such as doors opening and closing, coughing, and footsteps are commonly present. Furthermore, in scenarios where the sound source is distant, the SNR between speech and the noise generated by acoustic sensor equipment is also low. This results in a marked reduction in the effectiveness of speaker recognition techniques, as reported in [21]. Preprocessing of the original speech, which includes the extraction of mel-spectrogram, spectrogram, and MFCC (mel frequency cepstrum coefficient), is essential prior to input into the back-end model. The occurrence of noise and reverberation leads to alterations in the speech’s spectral characteristics. To solve this problem, scholars focus on speech preprocessing, feature extraction, loss function design, and the integration of a speech separation frontend [22,23,24,25,26]. For example, ref. [26] introduced the integration of FiLM into a Voice-Filter frontend to enhance the robustness of automatic speech recognition (ASR) and text-independent speaker verification (TI-SV) in noisy and reverberant conditions. These strategies primarily address specific noisy scenarios.
In recent years, self-supervised learning (SSL) has achieved significant success in fields such as computer vision and natural language processing [27,28,29,30,31]. SSL can extract generalized information from unlabeled data without relying on labels, annotations, or structured text [32]. Refs. [32,33,34,35,36,37] successively demonstrated the impressive translational results continuously achieved in ASR tasks using SSL. SSL is not only employed for ASR and phoneme recognition but also extends beyond acoustic character recognition to include tasks such as speaker recognition [36] and emotion recognition [37].
To overcome the challenges posed by noisy speech during preprocessing, ref. [38] introduced WavLM, which integrates a SSL model with a robust speaker recognition model to create a novel cascaded architecture. However, the model exhibits certain drawbacks: (1) Despite incorporating strategies such as mask learning and data augmentations involving speed, volume, and noise, the model’s performance under conditions of strong interference consistently falls short compared to robust baseline models like ECAPA-TDNN [20] and CAM++ [39]; (2) In [32], the base HuBERT required 32 V100 GPUs to train on a dataset of 960 h of audio from the LibriSpeech. This level of resource usage is highly detrimental to the goals of miniaturization and low power consumption in smart devices.
This study introduces a novel TDNN-based architecture incorporating a pre-trained model for robust speaker recognition in challenging noisy and reverberant conditions (PNR-TDNN). The architecture leverages the strengths of both SSL and TDNN. Firstly, the SSL model, inspired by HuBERT, is utilized in this study. Given that the number of pseudo-labels is typically a fixed hyperparameter for SSL models, the Canopy and Mini Batch K-Means++ algorithm is proposed to more accurately ascertain the appropriate number of pseudo-labels, thereby serving as dynamic adaptive filtering and extraction units for subsequent speaker recognition techniques. Secondly, inspired by the methods for extracting neighboring information in visual detection and semantic segmentation tasks [40,41,42,43,44,45,46,47], this work recognizes the importance of processing inter-contextual channel information in speech features. To address this, the ECAPA-TDNN framework is modified by incorporating several enhancements, particularly in the TDNN and pooling layers, to more effectively address the challenges of noise and reverberation.
The article’s contributions are as follows:
(1)
The development of a novel training strategy for the SSL model in the presence of strong interference, as well as the introduction of a miniaturized method for the integration of SSL model and speaker recognition model.
(2)
The pseudo-label generation mechanism of a typical SSL model is improved through the application of the Canopy and Mini Batch K-Means++ algorithms.
(3)
Enhancements are made to the pooling layer and Res2SE module of the ECAPA-TDNN by incorporating multi-head attention-based statistic pooling, cross-channel fusion for residual learning, and a non-average full-channel processing mechanism.
The rest of this paper is structured as follows: Section 2 introduces the typical SSL and speaker recognition model, as well as the proposed PNR-TDNN. Section 3 describes the details from dataset preparation to model training. Section 4 analyzes and compares the results of multiple sets of experiments. The conclusions are summarized in Section 5.

2. Background and Proposed Method

2.1. Problem Description

The speaker verification task is characterized by the use of a labeled dataset, which is partitioned into training and test subsets. The subsets are created by matching speech segments with their corresponding labels. Following the training phase, the speeches from the test dataset are inputted into the model to obtain prediction vectors. The similarity between these vectors is assessed using an appropriate metric to ascertain the likelihood of their origin from the same speaker. Most traditional speaker recognition systems employ MFCCs as their primary features. As illustrated in Figure 1, however, the presence of strong noise and reverberation significantly interferes with the speech signal, posing challenges to the effectiveness of MFCC-based recognition. The complexity of noise sources makes their elimination challenging, and the differentiation of reverberation from the clean signal is difficult [48]. Consequently, bypassing the spectrogram and employing an alternative initial feature extractor may be considered as a potential solution. This paper proposes employing a small amount of labeled data in conjunction with a large amount of unlabeled data, which is obviously more applicable. This results in a model with greater generalization and migration capabilities.

2.2. Speech Model of SSL

HuBERT is derived from the integration of BERT [29] and wav2vec [35]. Its architecture is characterized by a multilayer CNN encoder, Transformer [49] encoders, and a projection layer. The parameter details for the base HuBERT are presented in Table 1.
The network architecture is shown in Figure 2. An input speech of length L is processed by the CNN encoder, which consists of seven convolutional layers, resulting in T (T = L/320) feature vectors with 512 channels. The T feature vectors are categorized into two types: masked feature vectors x ˜ , where the masking process involves randomly selecting the starting point for the mask and setting 10 consecutive features to a value of 0, and unmasked feature vectors, x. The masked and unmasked features are processed by the Transformer encoders, resulting in the output feature sequences [ o 1 m , …, o T 1 m ] and [ o 1 u m , …, o T 2 u m ], respectively. The sum of T 1 and T 2 equals the total length T, with t representing the feature index. The Transformer encoders utilize both global and local information, maintaining strict consistency in data dimensions between input and output.
The subsequent step involves the calculation of the loss function. Initially, we need to establish the correspondence between the output ( o t ) and the pseudo-labels (z), which are associated through a probability density function. In this context, each pseudo-label (z) (where the total number of classifications is C) is assigned to a specific codebook. Pseudo-labels with identical values are mapped to the same codebook. The probability density functions for the variables ( o t m and o t u m ) with respect to the pseudo-label (z) are given by Equations (1) and (2), respectively.
p m z , c x ˜ , t = exp s i m P · o t m , e c / τ c = 1 C exp s i m P · o t m , e c / τ
p u m z , c x , t = exp s i m P · o t u m , e c / τ c = 1 C exp s i m P · o t u m , e c / τ
The projection matrix P is back-propagable; e c denotes the codebook corresponding to label z, and  e c represents the standard codebook. The variable c indicates a specific classification. The scale parameter τ is utilized for adjustment. The function sim(·,·) calculates the cosine similarity between two tensors. p m represents the probability density function derived from the masked x ˜ , while p u m denotes the probability density function from the unmasked x. Cross-entropy loss function is employed. The loss functions L m and L u m , derived from x with the corresponding pseudo-label, are defined by Equation (3) and Equation (4), respectively. The final loss function, as shown in Equation (5), comprises L m and L u m , where λ represents an adjustment parameter.
L m = t = 1 T 1 log p m z , c x ˜ , t
L u m = t = 1 T 2 log p u m z , c x , t
L = λ · L m + 1 λ · L u m
Notably, HuBERT employs an ensemble of K clustering methods, leading to a multinomial adjustment of the loss function, as demonstrated in Equations (6) and (7).
L m = k = 1 K t = 1 T 1 log p m k z , c x ˜ , t
L u m = k = 1 K t = 1 T 2 log p u m k z , c x , t
During the subsequent fine-tuning phase of the ASR task, the projection matrix P is not updated, and the parameters of the CNN encoder are frozen. Only the Transformer encoders are subjected to parameterization, utilizing connectionist temporal classification (CTC) as the loss function. In the HuBERT strategy, MFCC is initially employed for clustering to generate pseudo-labels, which serve as the target for the initial training epoch. Subsequently, the output from the 6th layer of the Transformer encoders in HuBERT is utilized for clustering purposes, yielding pseudo-labels that facilitate the second phase of training.
In contrast to HuBERT, WavLM enhances the training process by initially introducing additive noise and additional speaker speech to pure speech. It incorporates a gate-relative position bias into the self-attention mechanism of the Transformer architecture. The pre-training strategy is adapted to include noise-contaminated and overlapping speech as input, while still utilizing pure speech for the computation of MFCC-based clustering pseudo-labels.

2.3. Speaker Verification Model

Speech features, including MFCC or Mel-Spectrograms, are extracted through the processing of audio by a characterizer. Subsequently, these extracted features are utilized as input for various neural networks to derive deep embedding features. In the final step, the likelihood of the speaker’s identity is determined through the application of an activation function. Among the commonly employed baseline models in the field are TDNN, ECAPA-TDNN, Res2net, Resnet-SE, Eres2net, and CAM++. The ECAPA-TDNN employs several advanced approaches, including channel-dependent and context-dependent statistics pooling, 1-dimensional Squeeze Excitation Res2Blocks, and a multi-layer feature aggregation and summation mechanism. CAM++ employs the D-TDNN [50] as its backbone architecture, incorporating strategies like context-aware masking.

2.4. Proposed Method Overview

This study proposes a speaker verification method for real-work scenarios: (1) It employs a self-supervised learning-based pseudo-labeling approach for speech corrupted by noise and reverberation. Contrary to HuBERT, pure speech is corrupted with a strong additive noise component to produce corrupted speech for subsequent input. Additionally, HuBERT incorporates a random masking operation into the input, which represents the second phase of interference addition. This is done to mimic the actual acoustic characteristics of a noisy environment. The ultimate objective is to predict pseudo-labels for pure speech. Additionally, the prediction of pseudo-labels is extended to scenarios involving strong reverberation, targeting reverberant speech. (2) To integrate the pre-trained model with the robust speaker verification architecture, the extracted audio features from the last Transformer encoder are adapted through a 1-dimensional convolution layer to a compatible format for the verification network. (3) ECAPA-TDNN is enhanced in this study through the integration of a convolutional aggregation mechanism inspired by U-Net, which facilitates the fusion of contextual features within the TDNN sub-network. The integration of attentive average pooling into the Squeeze Excitation module enables the TDNN subnet to adaptively and discriminately learn channel-wise weight information. This modification enhances the network’s capability to extract speech features even under strong noise interference across varying time intervals. Lastly, we incorporate a multi-heads attentive statistics mechanism into the pooling factor α , enhancing the network’s resilience to interference from low and medium-frequency noise after pooling.

2.4.1. Network Architecture

The proposed model architecture comprises two components. The first component, as illustrated in Figure 3, integrates a CNN encoder with Transformer encoders, incorporating a projection matrix. This projection matrix is utilized exclusively during the pre-training phase. The Transformer encoders consist of 12 layers, each performing self-attention operations. Each layer has an output channel dimension of 768, with 12 attention heads, and a projection dimension of 256 for the loss function. The second part shown in Figure 4 depicts details of the improved ECAPA-TDNN. Among them, Res2net-U, SE-attentive mean, and multi-heads attentive statistics pooling are shown in Figure 4b, Figure 4c and Figure 4d, respectively. The PNR-TDNN and ECAPA-TDNN share similarities in their capability to process speech segments of varying lengths. Signals with a duration exceeding 0.5 s are used for training and testing. When the signal length exceeds 3 s, it is truncated to 3 s.

2.4.2. Canopy/Mini Batch K-Means++

In the application of K-means [51] clustering to HuBERT/WavLM, the determination of the number of clusters and the identification of initial cluster centers can present challenges, particularly when large-scale datasets are involved. Figure 5 and Algorithm 1 illustrate the approach adopted in this study, which employs Canopy [52] and Mini Batch k-means++ [53] to address these issues. The approach is as follows: (1) Set initial distance thresholds T 1 and T 2 ( T 1 > T 2 ). Calculate the Euclidean distance between all data and thresholds. Determine whether new clusters are formed and find the appropriate number of clusters K. (2) Assuming that n initial clustering centers have been selected ( 0 < n < K ), points that are further away from current centers will have a higher probability of being selected as the ( n + 1 )th cluster center until selection of n clustering centroids is completed. (3) A portion of the dataset is randomly selected and allocated to clusters based on proximity to their centroids. The centroids are iteratively updated until stability is achieved or the maximum number of iterations is reached.
Algorithm 1: Canopy/Mini Batch k-means++ algorithm
Applsci 15 02924 i001

2.4.3. Fusion Res2net and SE-Attentive Mean

In Figure 6a, the input is divided into n feature groups by Res2net. The i-th feature is fused with the ( i + 1 )th feature through a 3 × 3 convolutional residual connection, except for the first feature, which is preserved in an unchanged state. Inspired by multi-rate tracking in navigation system [54], this study utilizes the multi-scale fusion capability of the U-net architecture. Figure 4b illustrates the convolutional fusion process within the U-net architecture. In the downsampling phase, features are combined by applying a 3 × 3 convolution, where the ith-level feature is integrated with the ( i + 1 )th level feature. Conversely, during the upsampling phase, the ith-level feature is concatenated with the ( i 1 )th-level feature through a similar 3 × 3 convolutional process. Furthermore, at the deepest level (nth feature), convolution is applied, and the result is directly propagated to the subsequent layer. This mechanism facilitates the integration of contextual information and the fusion of neighboring channels.
Equations (8)–(10) illustrate the three steps executed by the SE (Squeeze Excitation) module within the ECAPA-TDNN framework. (1) x ( B , F , T ) obtains x m e a n ( B , F , 1 ) in the T dimension mean, as shown in Figure 4c, “Mean Parameter”; (2) x m e a n is squeezed and excited to obtain weight coefficients for all F channels x f ( B , F , 1 ) . The linear operation corresponding to W 1 and b 1 represents squeeze, whereas W 2 and b 2 represent excitation, as shown in Figure 4c, “Fully Connected”; (3) x f is extended to x f t (in T dimension), and subsequently, x f is re-weighted with x to obtain x ˜ ( B , F , T ) , as shown in Figure 6c, “Re-weight”.
x m e a n = 1 T t = 1 T x t
x f = s i g m o i d W 2 · r e l u W 1 · x m e a n + b 1 + b 2
x ˜ = x f t · x
This study introduces an attentive channel mean mechanism to replace the traditional mean operation, acknowledging the varying contributions of different frames to the speaker’s message, particularly in the differentiation between vocalized and silent states. Concretely, x ( B , F , T ) is sequentially processed through compression, recovery, and probability estimation to yield the attentive channel mean score α .
In Equation (11), x undergoes squeeze and recovery operations along the F channel. The linear transformations corresponding to weight matrix W 3 and bias vector b 3 denote the squeeze operation, while W 4 and b 4 denote the recovery operation, as illustrated in Figure 4(c1), “Fully Connected”. The hyperbolic tangent function, t a n h , is used as the activation function. After the above operation, we further obtain the linear relationship between all F channels. Subsequently, the Softmax activation function is applied to the T channel, resulting in the fractional distribution of different time frames. Equation (12) illustrates that α is element-wise multiplied with x, followed by the aggregation of the resulting values across the T channel. Conclusively, the optimized mean feature x ˜ m e a n is derived, as shown in Figure 4(c1), “Hadamard Product+Sum”.
α = S o f t m a x W 4 · tanh W 3 · x + b 3 + b 4
x ˜ m e a n = t = 1 T α t · x t

2.4.4. Multi-Head Attentive Statistics Pooling

Analysis of MFCC features extracted from the original speech reveals that noise is predominantly distributed in the low-frequency and mid-frequency bands, while speech is predominantly found in the low-frequency band. Upon substituting the MFCC extractor with a pre-trained model, the noise in the newly extracted features is also characterized by a full-channel distribution. Attentive statistics pooling in ECAPA-TDNN is designed to capture global channel information, which, however, may lead to heightened noise interference in the pooling outcomes. To address this issue, this work introduces a modification involving the replacement of the attentive statistics pooling with a multi-head attention-based pooling mechanism [55]. This approach divides the entire channel space into several autonomous sub-channel spaces, aiming to reduce mutual interference.
Specifically, the mean and standard deviation of the output features from the final TDNN layer are computed initially. As shown in Figure 4d “Catenate”, the resulting feature, after aggregating the above three, is defined as o ( d , T ) . Here, d can be understood as the frequency band and T as the time. o is separated into k non-overlapping subspaces ( x ( d / k , T ) ) along the channel dimension d, as shown in Figure 4(d1), “Chunk”. Among these subspaces, x t for a T channel dimension in x is defined as [ x t 1 , x t 2 , , x t k ] ( 1 t T , 1 i k , x t i R d k ) . Similarly, x i for a d channel dimension in x is defined as [ x 1 i , x 2 i , , x T i ] . Next, unlike the channel-independent self-attention scoring mechanism in [20], the channel-isolated self-attention scoring mechanism is used here. The corresponding scoring factor e for each sub-channel space (in d channel) is calculated by using Equation (13).
e = W 6 · tanh W 5 · x + b 5 + b 6
W 5 R F ( d k ) and W 6 R F × F are linear unit operations, b 5 and b 6 are their corresponding bias variables, as shown in Figure 4(d1) “Fully Connected”. Then, the Softmax function (Equation (14)) is utilized to compute the importance percentage scores across various T channels. For each dimension of the d channel, the score is denoted as β i , which represents the importance distribution of data frames across different T channels within each d channel dimension.
β i = exp e j i j = 1 T exp e j i
β is concatenated along the channel dimension d from the β i components. Subsequently, γ is constructed by concatenating the β across the k subspaces, as depicted in Figure 4(d1) “Catenate”. γ t , representing the T channel dimension within γ , is defined as the concatenation of [ γ t 1 , γ t 2 , , γ t k ] , where each γ t i is a component of the respective subspace. Due to the specificity of different d channel dimensions, the mean μ ˜ and variance σ ˜ , obtained through multi-head attentive statistics pooling, are formulated in Equations (15) and (16), also as shown in Figure 4d, “Hadamard Product”, “mean”, and “std”. The pooling output is derived from the concatenation of μ ˜ and σ ˜ .
μ ˜ = t T γ t · x t
σ ˜ = t T γ t · x t 2 μ ˜

3. Experimental Setup

3.1. Dataset Generation

Voice data for this study were obtained from the open-source Chinese speech corpus, zhvoice, which comprises eight datasets. The corpus, following noise reduction and silence removal, includes more than 3200 speakers and approximately 900 h of audio, all recorded at a 16 kHz sampling rate. Two versions of the zhvoice corpus are adapted for this paper, tailored to conference room conditions. One version simulates scenarios with both noise and reverberation (herein referred to as NR-zhvoice), while the other version represents conditions with noise only (referred to as N-zhvoice). For NR-zhvoice, the simulation of reverberant speech is a preliminary step. The room impulse response is simulated using the wham_room library [56], where the configuration of parameters for reverb is shown in Table 2. In addition, wham_room utilizes a dual-source and dual-microphone configuration by default, in contrast to the minimalist speaker verification approach, which operates with monaural audio. Consequently, identical sets of sound sources and microphone parameters are input into wham_room, with only the left channel of the first microphone being employed for reverb output.
In the second processing step, reverberant speech is corrupted by adding a randomly chosen noise at a randomly determined SNR within the range of [0, 0.5] dB. In cases where the noise duration is shorter than that of the speech, the noise is looped or extended by concatenating its beginning to ensure matching lengths. The noises are derived from two primary collections. The first is the soundsnap website, which encompasses a diverse array of sounds, including air conditioners, fans, cell phone vibration, coughing, sneezing, walking, flipping through documents, keyboards, and more. The dataset includes 1950 records across twelve categories. Additionally, as shown in Figure 7 and Figure 8, actual recordings (A total of 480 records are sampled at 16 kHz, all lengths are cropped to 3 s) were made in a working conference room using a U-PHORIA UMC202HD sound sensor manufactured by Behringer. N-zhvoice is generated by a process that follows the second step of NR-zhvoice, involving the addition of noise while excluding reverberation to the zhvoice.

3.2. Hyperparameter Settings

The speech from N-zhvoice and NR-zhvoice are used as input for the pre-trained model. Meanwhile, the k-means clustering value of MFCC from zhvoice/reverberation zhvoice dataset serves as the iterative target. A sub-dataset of the N-zhvoice/NR-zhvoice is used as the training data for the speaker verification model, with the speaker classification labels designated as the objective for learning. The pre-trained model is trained for 15 epochs, after which the speaker verification network is further trained for an additional 60 epochs. The proposed method, implemented in Python 3.8 using PyTorch 2.0.0, was executed on an Ubuntu 20.04 system equipped with two pieces of RTX 4090 GPUs and one piece of Intel(R) Xeon(R) Gold 5218R CPU (Intel, Santa Clara, CA, USA).
(1)
Training details and settings of the pre-trained model: For MFCC, the frame-length is 25 ms, the frame-shift is 10 ms, the sampling frequency is 16 kHz, and the MFCCs are 39. The CNN encoder utilized a downsampling rate of 1/320. The values for both mask-prob and mask-channel-prob are both set to 0.9, with an approximately 1:1 ratio of masked to non-masked features. The mask length is set to 10 and the learning rate is set to 5 × 10−4, weight decay to 1 × 10−2, and the optimizer employed is Adam with beta parameters randomly sampled from the range (0.9, 1.8). The dropout rate is fixed at 0.1.
(2)
Training details and network settings of TDNN: The model is trained with a learning rate of 1 × 10−3 and weight decay of 1 × 10−6, using Adam as the optimization algorithm, and 512 feature channels. The batch size is set to 12, and the maximum duration of audio clips is limited to 3 s. For the loss function, the Angular Additive Margin (AAM) [57] is employed with a margin of 0.2 and a scale factor of 30.

3.3. Evaluation Metrics

The pre-trained model utilizes two datasets, N-zhvoice and NR-zhvoice, each comprising 1,090,000 unlabeled samples. Approximately 5% of these data are randomly selected and labeled for use in subsequent speaker verification tasks. Of the labeled data, 1∼1.5% is designated as the test set, with the remainder allocated for training purposes. The evaluation criteria for the speaker verification tasks encompass the Equal Error Rate (EER) and the Minimum Detection Cost Function (MinDCF). Associated metrics include the False Acceptance Rate ( F A R ), the False Rejection Rate ( F R R ), and the threshold settings. The EER is defined as the point at which the FAR and FRR are equal, reflecting a balance between the two types of errors at different threshold settings. The formula for computing MinDCF is provided in Equation (17).
D C F = C F R R × F R R × P T + C F A R × F A R × P I
The costs associated with false rejection ( C F R R ) and false acceptance ( C F A R ) are both set as 1, while the prior probabilities for the real speaker ( P T ) and the impostor speaker ( P I ) are assigned as 0.01 and 0.99, respectively.
The metrics are computed through the following steps: (1) acoustic features, extracted either by a MFCC extractor or a pre-trained model, are input into a speaker recognition model to derive classification features; (2) the cosine similarity between the classification features for each audio is calculated and designated as the target score; (3) the target label and the corresponding target score are compiled into a score list; (4) the threshold is incremented from 0 to 0.99 in steps of 0.01. For each threshold, the F R R and F A R are determined. Subsequently, the EER and MinDCF values are computed based on these rates.

4. Results and Comparison

This section discusses the convergence of the pre-trained model, the performance of speaker verification models utilizing MFCC-based input features, and the performance of PNR-TDNN with a self-supervised learning-based feature extractor.

4.1. Pre-Trained Model Experiments

As shown in Figure 9, there were 100 clusters when T 2 took 165. In general, an increase in T 2 is associated with a decrease in the size of the clusters, whereas the impact of T 1 on cluster formation is minimal. Figure 10 illustrates the convergence of pre-trained model associated with varying values of MFCC label clusters: (1) Convergence of the pre-trained model becomes particularly challenging when the clusters value is increased. It is evident that WavLM achieves convergence more readily compared to HuBERT. (2) Both HuBERT and WavLM face increased difficulty in fitting the data as the number of clusters grows. From the perspective of model refinement, a larger number of clusters generally corresponds to improved back-end modeling capabilities. (3) It is crucial to select appropriate values for T 2 and T 1 to accurately determine the number of clusters. In the subsequent experiment, the optimal number of clusters was found to range between 60 and 90 when T 2 is set to approximately 180. This finding serves as a useful guideline for selecting the number of clusters in similar scenarios.

4.2. PNR-TDNN Experiments

Unlike models that utilize MFCCs for speaker verification, the PNR-TDNN employs the output features of a pre-trained network for the speaker verification task. These features, with their higher dimensionality, provide increased flexibility and enable adaptability across diverse application scenarios. This results in slower convergence rates in recognition networks, so training epoch is set to 60.

4.2.1. Discussion of Baselines

To facilitate a fair comparison, the experiments in this study utilize a subset of the data referred to in Section 3.1. These data are further annotated and include instances of noise interference as well as combined noise and reverberation interference. Specifically, the sub-datasets from N-zhvoice/NR-zhvoice for speaker verification are presented in Table 3. The baseline model comprises TDNN, Resnet-SE, Res2net, Eres2net, ECAPA- TDNN (512 channels), CAM++ (32 channels), along with other state-of-the-art methods within the last two years (as shown in Table 4, such as MFA-Conformer, Wespeaker, RedimNet, Gemini). All baseline models exhibit a satisfactory fit within 30 epochs, which is why the training is limited to this number of epochs.
For dataset A, as presented in Table 4, CAM++ demonstrates superior performance compared to other methods with respect to both EER and MinDCF. RedimNet achieves the second lowest EER. In the case of datasets B and C, CAM++ achieves a notable lead in EER, whereas ECAPA-TDNN and Resnet-SE yield the optimal MinDCF results. Additionally, the results following the addition of reverberation interference are worse than those with noise alone, indicating the detrimental effect of reverberation on the speaker recognition task. This study argues that the trailingphenomenon in the frequency domain impacts the network’s capability to process information effectively. The computational complexity of the model is measured in terms of floating-point operations (FLOPs) and is assessed using a CPU (Intel(R) Xeon(R) Gold 5218R CPU@2.1GHz, the batch size for these evaluations is fixed at 12). The real-time factor (RTF) is a metric used to gauge the efficiency of the model’s processing time, defined as the ratio of the time taken by the model to process a single audio to the duration of the audio itself.

4.2.2. Discussion of Clusters and Compressed Models

Table 5 presents the experimental results of various PNR-TDNNs. We found that an increased number of MFCC clusters corresponding to pseudo-labels within the pre-trained model lead to enhanced effectiveness in the back-end recognition model. When the number of clusters exceeds 100, HuBERT experiences gradient overflow within the first three epochs. In this case, the corresponding PNR-TDNN exhibits significantly weak recognition performance. Therefore, the optimal strategy is to select a maximum number of clusters that ensures the convergence of the pre-trained model, balancing the model’s capacity with its ability to learn effectively. In addition, recognition performance is at its peak when the cluster number is set to 80. Based on the methodology described in Section 2.4.2, which involves Canopy and Mini Batch k-means++, it is determined that for the specific application scenario, when the values of T 1 and T 2 are set at 400 and 180, respectively, the optimal number of clusters is obtained.
In the case of noise-only interference, the EER difference between Hu + EC(80) and Wa + EC(80) is negligible, whereas the difference in MinDCF is notable. Hu + EC(80) exhibits superior performance in both metrics under conditions of combined noise and reverberation. Our study suggests that the ambiguity introduced by full-channel reverberation makes use of gated relative position bias that is inappropriate for PNR-TDNN. Additionally, the results for Hu + CA(80) consistently demonstrate lower performance compared to Hu + EC(80) across all evaluated metrics and computational factors. Consequently, Hu + EC(80) is designated as the reference implementation for PNR-TDNNs in subsequent experiments.
Another key issue is that Hu + EC occupies 400 M, which is a very large amount that is unacceptable and cannot be deployed for smart microphone devices. As shown in Table 6, we reduced the parameters for HuBERT’s Transformer encoder and feed forward net. The minimized HuBERT + EC (T-HuBERT + EC) is only 90 M, which is in the same order of magnitude as ECAPA-TDNN and CAM++. In subsequent experiments, small HuBERT + EC and tiny HuBERT + EC outperform or almost equalize base HuBERT + EC in both EER and minDCF. This indicates that base HuBERT has not been fully utilized in pre-training.
The pseudo-labels for pre-trained model carry a certain degree of uncertainty, and their generation and application methods are far from comparable to real labels. Of course, this is also advantageous, as they can better learn fuzzy information and adapt to various downstream tasks. However, the biggest disadvantage of fuzzy information is that it cannot refine learning, and although base HuBERT contains 12 layers of Transformer encoders, the goal is too vague, and the multi-layered and repeated cascading network structure cannot obtain more useful information. Therefore, it is feasible to simplify the number of encoder layers and reduce the feature dimension appropriately within the narrow range of SNR.
Finally, the computational complexity analysis of the proposed method can be seen in Table 5, and a faster speed of 75 speech features per second can be achieved in one part of Intel(R) Xeon(R) Gold 5218R CPU.

4.2.3. Discussion of PNR-TDNN Ablation Experiments

Table 7 presents the performance of the proposed method, which is discussed in detail. The experiments are numbered for ease of reference. Experiments 1–3 modify the residual module of the Hu + EC(80), with the various residual networks depicted in Figure 6a, Figure 6c and Figure 6b, respectively. Experiment 4 introduces the SE-attentive mean mechanism. Experiment 5 employs the multi-head attentive statistics pooling mechanism. In experiment 6, we combine the SE-attentive mean with the multi-head attentive statistics pooling mechanism, integrating these enhancements into the Hu + EC(80). Experiments 7–9 are the enhanced versions achieved after applying fusion Res2net, SE-attentive mean, and multi-head attentive statistics pooling mechanism.
In comparison to the joint training approach where the pre-trained model and back-end model are trained together, as described in references [32,38], our findings indicate that the entire model tends to converge more effectively when the pre-trained model is entirely frozen during the training process.
For experiments under noise-only interference conditions: (1) Employing pseudo-labels based on MFCC clusters of clean speech significantly enhances the performance of the PNR-TDNN architecture, surpassing that of the established baselines. (2) Figure 11a reveals that the EER for each PNR-TDNN experiences a marked decline beyond 30 epochs. Conversely, the MinDCF continues to refine. Research indicates that the points of optimal performance for EER and MinDCF are not reached concurrently.
For experiments conducted under noise and reverberation interference conditions, our findings were as follows: (1) The PNR-TDNN consistently outperforms the baselines only when MFCC clusters with reverberation are utilized as pseudo-labels. (2) When the extended version of dataset A, comprising 66,000 utterances with noise and reverberation interference, is utilized as the training set, an improvement in the EER of the test set is challenging to achieve. However, when dataset A is reduced to create dataset B (51,000 utterances) and dataset C (30,000 utterances), a notable reduction in EER is observed. The study suggests that the limitation of pre-trained models in capturing concrete information, which is less specific than MFCC features, is magnified by the presence of an excessive number of labels during PNR-TDNN training. (3) Figure 11b,c illustrate that the EERs for datasets B and C remain consistent. The slower convergence of the loss function indicates that the pre-trained model exhibits robust resistance to over-fitting. The resilience enables the PNR-TDNN to progressively enhance its efficacy in the latter phase of the training process, a trend that shares similarities with the performance characteristics observed in CAM++.
In comparison to the best model (CAM++) discussed in Section 4.2.1, experiment 8 yields an EER improvement of 9.54% while almost maintaining a similar MinDCF on dataset A. On dataset B, an 11.13% EER improvement is observed, accompanied by a slight improvement in the MinDCF. For experiment 7 on dataset C, a 23.77% EER improvement and a 1.88% enhancement in the MinDCF are achieved. These results suggest that the enhancement mechanisms proposed in this paper demonstrate superior performance in speaker verification tasks under conditions of strong noise and reverberation.
In addition, the EER of experiment 8 is slightly better than experiment 9, indicating that larger pre-trained models still have a few advantages in backend applications, but the improvement effect is not significant. For example, for dataset A, experiment 8’s EER improves by 4.5% compared to that of experiment 9. Considering the miniaturization requirements of smart microphone devices, choosing smaller pre-trained model is still the preferred choice.
Finally, we performed significance tests on the model. Specifically, five cross-training sessions were conducted using dataset A, and then the t-test was cross-validated using the mean and variance of the calculated EER. As shown in the Figure 12, this reflects that the five models are statistically significantly different, and also shows that our proposed model has a very uplifting effect with respect to the other four baseline models, and improvements are not due to random variations.To better present the results, as shown in Figure 13, we present the best EER comparison plots of the four baseline models with the proposed T-Hu+EC+FAM on dataset A, dataset B, and dataset C.

4.2.4. Discussion of Other Noisy Situations

To assess the generalization capability of the PNR-TDNN, ESC50 (primarily environmental sounds) and a subset of cn-celeb_v2 (mainly Chinese speech) are selected to generate noise interference for the zhvoice. The ESC50 encompasses animal sounds, natural phenomena, water noises, human non-verbal cues, as well as indoor and domestic sounds. These closely resemble the noise sources used in Section 4.2.1 (soundsnap and actual conference room recordings), enhancing the realism of the testing conditions. Furthermore, the impact of human speech interference is significant and cannot be overlooked. To simulate this, a random selection of speech segments from the cn-celeb_v2 dataset is used as a second noise source. The speech in zhvoice and cn-celeb_v2 is sourced from different individuals, reflecting realistic conditions in which multiple speakers may be present.
Table 8 reveals that PNR-TDNNs outperformed all baselines in terms of recognition accuracy when exposed to ESC50 noise. Despite the presence of vocal interference, the S-Hu + EC + FAM yields robust recognition outcomes. In addition, S-Hu + EC + FAM substantially outperforms Hu + EC + FAM on all three datasets, which corroborates the conclusion in Section 4.2.2 that the small pre-trained model is more effective at capturing speech feature information in a strongly noisy environment with a narrow range of SNR. This demonstrates the strong generalization capability of the proposed method in diverse scenarios, thereby showcasing the model’s adaptability to different conditions.

4.2.5. Discussion of Other SNR’s Situations

The SNR used in previous sections to characterize the noise level ranges from 0 to 0.5 dB, which is a very narrow range. This means that through repeated training, the network may cleverly avoid the influence of noise (in Section 2.1, we found that noise energy is more easily dispersed than speech energy). Therefore, we supplemented the comparative experiments with a larger range of SNR. In the production of noisy speech, the SNR is in narrow (0 to 0.5 dB) and wide (−5 to 3 dB) ranges. The SNR was chosen at random, i.e., in the form of a uniform distribution. More than 60% of these speeches were in the negative SNR condition, and less than 40% were in the lower positive SNR. These conditions are closer to the complex and challenging real sound field, and we list the experimental results in Table 9.
On dataset A, the larger Hu + EC + FAM has the best EER, with an improvement of 12.6% with respect to CAM++, while both S-Hu + EC + FAM and T-Hu + EC + FAM also present a stronger performance than CAM++. This phenomenon suggests that the larger the pre-trained model is, the better the adaptation to the back-end task. As for dataset B and C, the EER of Hu + EC + FAM improves 5.2%/6.12% relative to RedimNet, while the performance of CAM++ is not stable enough to be used as a baseline. Furthermore, the EER of S-Hu + EC + FAM and T-Hu + EC + FAM are both slightly weaker than RedimNet.
In narrow-range SNR, due to the weakening of the noise by fusion Res2net and SE-attentive mean, it is very likely that the noise effect will be ignored by implicit noise suppression. However, smaller Hu + EC + FAM can easily outperform the baseline for milder general acoustic field environments, but the large pre-trained model still has a significant advantage when faced with the challenge of noise intensity greater than speech intensity and a very wide range of SNR.
When exposed to stronger noise energy and a wide range of SNR, the large pre-trained model is still able to extract more feature information about a particular speaker through its deeper network parameters; this finding is significantly different from the experimental results obtained for the narrow-range SNR.

4.2.6. Discussion of Other Public Datasets

The PNR-TDNNs are also evaluated in the well-known dataset of VoxCeleb1 [62] and cn-celeb_v2 [63]. Notably, VoxCeleb1 and cn-celeb_v2 contain a large amount of noise. To match the initial conditions with the zhvoice, a total of 300 noise-free and reverberation-free speakers are meticulously selected from VoxCeleb1, amounting to 100,482 utterances for pre-training (denoted as VoxCeleb1-S). Similarly, 122 noise-free and reverberation-free speakers are chosen from cn-celeb_v2, resulting in a total of 54,272 utterances for pre-training (denoted as cn-celeb_v2-S).
To assess the performance of the PNR-TDNNs, two subsets of VoxCeleb1-S are randomly selected and labeled. As described in Table 10, these two subsets are named VoxCeleb1-A and VoxCeleb1-B, respectively. The same processing steps described above are also applied to the cn-celeb_v2-S. Table 11 shows the evaluation results in VoxCeleb1-A/B and cn-celeb_v2-A/B. As shown in Figure 14, we also present the best EER comparison plots of the four baseline models with the proposed T-Hu+EC+FAM on dataset VoxCeleb1-B and dataset cn-celeb_v2-B.The performance of CAM++’s EER on VoxCeleb1 and cn-celeb_v2 is erratic, which may be attributed to the more complex acoustic environments encountered in these datasets. Several new approaches, including MFA-Conformer, yield suboptimal EER compared to the ECAPA-TDNN.
In evaluation experiments with VoxCeleb1 with only strong noise interference, small PNR-TDNNs presented optimal EER, e.g., T-Hu + EC + FAM achieved a 32% improvement in EER relative to ECAPA-TDNN. However, in the strong reverberation and noise situation, reverberation seriously interfered with speaker identification accuracy, although the PNR-TDNNs still had the lowest EER relative to numerous baselines. The effectiveness of the proposed method in the English dataset is indicated by these results. As for cn-celeb_v2 s2, the proposed method continued to yield the best recognition performance, albeit with a relatively smaller improvement. This can be attributed to the inherently stronger noise present in the original cn-celeb_v2, which negatively impacts the feature extraction capabilities of the backend network.

5. Conclusions

This paper proposes PNR-TDNN, a speaker verification method designed to handle challenging conditions such as strong noise and reverberation interference. This study improves upon the current model by applying a new training strategy to the SSL model, introducing multi-head attentive statistics pooling, cross-channel fusion of residuals, and non-average full-channel mechanisms into the embedded extraction units, such as the pooling layer and SE-Res2 block of ECAPA-TDNN, respectively. A miniaturization method for PNR-TDNN is also proposed and tested. The proposed method has been tested on several datasets characterized by strong noise and reverberation interference. The experimental results demonstrate varying levels of reduction in EER, thereby confirming the effectiveness of the PNR-TDNN approach. Further research is anticipated to extend the application of PNR-TDNN to the domain of speaker identification.

Author Contributions

Conceptualization, X.Z.; methodology, X.Z. and C.S.; software, X.Z.; validation, X.Z. and C.S.; formal analysis, X.Z. and C.S.; investigation, X.Z. and C.S.; resources, J.T. and J.L.; data curation, X.Z. and C.S.; writing—original draft preparation, X.Z.; writing—review and editing, X.Z., C.S., H.C. and C.W.; visualization, X.Z.; supervision, J.T. and J.L.; project administration, C.S.; funding acquisition, J.L., J.T., C.S., C.W. and H.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was partially supported by Innovative Research Group Project of National Natural Science Foundation of China (Nos. 51821003), the Excellent Youth foundation of Shanxi Province (Nos. 202103021222011), the Key research and development project of Shanxi Province (Nos. 202202020101002), the fundamental research program of Shanxi Province (Nos. 202303021211150), the Aviation Science Foundation (Nos. 2022Z0220U0002), the Shanxi province key laboratory of quantum sensing and precision measurement (Nos. 201905D121001), 173 Foundation (Nos. 2021XXXX0668), ZBZDJCYJKT (Nos. 51405XX01).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The raw data can be obtained via the following website: zhvoice URL: https://github.com/fighting41love/zhvoice; VoxCeleb1 URL: https://mm.kaist.ac.kr/datasets/voxceleb; cn-celeb_v2 URL: http://openslr.org/82/; soundsnap URL: https://www.soundsnap.com/ (accessed on 6 February 2025).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, Q.; Lin, X.D. Efficient and Privacy-Preserving Speaker Recognition for Cybertwin-Driven 6G. IEEE Internet Things J. 2021, 8, 16195–16206. [Google Scholar] [CrossRef]
  2. Lee, S.W.; Lee, D.H. From Attack to Identification: MEMS Sensor Fingerprinting Using Acoustic Signals. IEEE Internet Things J. 2023, 10, 5447–5460. [Google Scholar] [CrossRef]
  3. Zhang, L.; Tan, S.; Chen, Y.; Yang, J. A Continuous Articulatory-Gesture-Based Liveness Detection for Voice Authentication on Smart Devices. IEEE Internet Things J. 2022, 9, 23320–23331. [Google Scholar] [CrossRef]
  4. Kinnunen, T.; Li, H. An overview of text-independent speaker recog-nition: From features to supervectors. Speech Commun. 2010, 52, 12–40. [Google Scholar] [CrossRef]
  5. Li, X.; Ze, J. Enrollment-Stage Backdoor Attacks on Speaker Recognition Systems via Adversarial Ultrasound. IEEE Internet Things J. 2024, 11, 13108–13124. [Google Scholar] [CrossRef]
  6. Dong, Y.; Yao, Y.D. Secure mmWave-Radar-Based Speaker Verification for IoT Smart Home. IEEE Internet Things J. 2021, 8, 3500–3511. [Google Scholar] [CrossRef]
  7. Huang, W.; Tang, W.; Jiang, H.; Luo, J.; Zhang, Y. Stop Deceiving! An Effective Defense Scheme Against Voice Impersonation Attacks on Smart Devices. IEEE Internet Things J. 2022, 9, 5304–5314. [Google Scholar] [CrossRef]
  8. Bian, J.; Al Arafat, A. Machine Learning in Real-Time Internet of Things (IoT) Systems: A Survey. IEEE Internet Things J. 2022, 9, 8364–8386. [Google Scholar] [CrossRef]
  9. Hanifa, R.M.; Isa, K.; Mohamad, S. A review on speaker recognition: Technology and challenges. Comput. Electr. Eng. 2021, 90, 107005. [Google Scholar] [CrossRef]
  10. Snyder, D.; Garcia-Romero, D.; Sell, G.; Povey, D.; Khudanpur, S. X-vectors: Robust DNN embeddings for speaker recognition. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5329–5333. [Google Scholar]
  11. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning forimage recognition. In Proceedings of the Computer Vision and Pattern Recognition (CVPR) 2016, IEEE Computer Society, Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  12. Variani, E.; Lei, X.; McDermott, E.; Moreno, I.L.; Dominguez, J.G. Deep neural networks for small footprint text-dependent speaker verification. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP, Florence, Italy, 4–9 May 2014; pp. 4052–4056. [Google Scholar]
  13. Snyder, D.; Garcia-Romero, D.; Sell, G.; McCree, A.; Povey, D.; Khudanpur, S. Speaker recognition for multi-speaker conversations using x-vectors. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP, Brighton, UK, 12–17 May 2019; pp. 5796–5800. [Google Scholar]
  14. Garcia-Romero, D.; McCree, A.; Snyder, D.; Sell, G. JHUHLTCOE system for the VoxSRC speaker recognition challenge. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing ICASSP, Virtual, 4–9 May 2020; pp. 7559–7563. [Google Scholar]
  15. Shi, Y.; Huang, Q.; Hain, T. H-VECTORS: Improving the robustness in utterance-level speaker embeddings using a hierarchical attention model. Neural Netw. 2021, 142, 329–339. [Google Scholar] [CrossRef]
  16. Zhou, T.; Zhao, Y.; Wu, J. Resnext and res2net structures for speaker verification. In Proceedings of the IEEE Spoken Language Technology Workshop 2021, Shenzhen, China, 19–22 January 2021; pp. 301–307. [Google Scholar]
  17. Liu, B.; Chen, Z.; Wang, S.; Wang, H.; Han, B.; Qian, Y. Dfresnet: Boosting speaker verification performance with depth-first design. In Proceedings of the Interspeech 2022, Incheon, Republic of Korea, 18–22 September 2022; Ko, H., Hansen, J.H.L., Eds.; ISCA: Singapore, 2022; pp. 296–300. [Google Scholar]
  18. Gao, Z.; Song, Y.; McLoughlin, I.V.; Guo, W.; Dai, L. An improved deep embedding learning method for short duration speaker verification. In Proceedings of the Interspeech, Hyderabad, India, 2–6 September 2018; pp. 3578–3582. [Google Scholar]
  19. Chen, C.P.; Zhang, S.Y.; Yeh, C.T.; Wang, J.C.; Wang, T.; Huang, C.L. Speaker characterization using TDNN-LSTM based speaker embedding. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 6211–6215. [Google Scholar]
  20. Desplanques, B.; Thienpondt, J.; Demuynck, K. ECAPATDNN: Emphasized channel attention, propagation and aggregation in tdnn based speaker verification. In Proceedings of the 21st Annual conference of the International Speech Communication Association INTER SPEECH, Shanghai, China, 25–29 October 2020; pp. 3830–3834. [Google Scholar]
  21. Togneri, R.; Pullella, D. An overview of speaker identification: Accuracy and robustness issues. IEEE Circuits Syst. Mag. 2011, 11, 23–61. [Google Scholar] [CrossRef]
  22. Chen, Z. On the Detection of Adaptive Adversarial Attacks in Speaker Verification Systems. IEEE Internet Things J. 2023, 10, 16271–16283. [Google Scholar] [CrossRef]
  23. Li, X.; Zheng, Z.; Yan, C.; Li, C.; Ji, X.; Xu, W. Toward Pitch-Insensitive Speaker Verification via Soundfield. IEEE Internet Things J. 2024, 11, 1175–1189. [Google Scholar] [CrossRef]
  24. Kheder, W.B.; Matrouf, D. Additive noise compensation in the i-vector space for speaker recognition. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing ICASSP, Brisbane, Australia, 19–24 April 2015; pp. 4190–4194. [Google Scholar]
  25. Juneja, K. Two level Noise Robust and Block Featured PNN Model for Speaker Recognition in Real Environment. Wirel. Pers. Commun. 2022, 125, 3741–3771. [Google Scholar] [CrossRef]
  26. O’Malley, T.; Ding, S.; Narayanan, A.; Wang, Q.; Rikhye, K.; Liang, Q.; He, Y.; McGraw, I. Conditional Conformer: Improving Speaker Modulation For Single And Multi-User Speech Enhancement. In Proceedings of the 2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Rhodes Island, Greece, 4–10 June 2023; pp. 1–5. [Google Scholar]
  27. Liu, X.; Zhang, T.; Liu, M. Joint estimation of pose, depth, and optical flow with a competition–cooperation transformer network. Neural Netw. 2024, 171, 263–275. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, X.; Zhang, T.; Liu, M. UDF-GAN: Unsupervised dense optical-flow estimation using cycle Generative Adversarial Networks. Knowl.-Based Syst. 2023, 271, 110568. [Google Scholar] [CrossRef]
  29. Devlin, J.; Chang, M.-W.; Lee, K.; Toutanova, K. Bert: Pre-training of deep bidirectional transformers for language understanding. In Proceedings of the North American Chapter of the Association for Computational Linguistics (NAACL), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar]
  30. Yang, Z.; Dai, Z.; Yang, Y.; Carbonell, J.; Salakhutdinov, R.R.; Le, Q.V. Xlnet: Generalized autoregressive pretraining for language understanding. In Advances in Neural Information Processing Systems (NeurIPS); Wallach, H., Larochelle, H., Beygelzimer, A., d’Alche-Buc, F., Fox, E., Garnett, R., Eds.; Curran Associates, Inc.: New York, NY, USA, 2019; Volume 32. [Google Scholar]
  31. Raffel, C.; Shazeer, N.; Roberts, A.; Lee, K.; Narang, S.; Matena, M.; Zhou, Y.; Li, W.; Liu, P.J. Exploring the limits of transfer learning with a unified text-to-text transformer. J. Mach. Learn. Res. JMLR 2020, 21, 1–67. [Google Scholar]
  32. Hsu, W.-N.; Bolte, B.; Tsai, Y.-H.H.; Lakhotia, K.; Salakhutdinov, R.; Mohamed, A. Hubert: Self-supervised speech representation learning by masked prediction of hidden units. arXiv 2021, arXiv:2106.07447. [Google Scholar] [CrossRef]
  33. Chung, Y.-A.; Hsu, W.-N.; Tang, H.; Glass, J. An unsupervised autoregressive model for speech representation learning. arXiv 2019, arXiv:1904.03240. [Google Scholar]
  34. Baevski, A.; Schneider, S.; Auli, M. vq-wav2vec: Selfsupervised learning of discrete speech representations. arXiv 2019, arXiv:1910.05453. [Google Scholar]
  35. Baevski, A.; Zhou, H.; Mohamed, A.; Auli, M. wav2vec 2.0: A framework for self-supervised learning of speech representations. arXiv 2020, arXiv:2006.11477. [Google Scholar]
  36. Chen, Z.; Chen, S. Large-scale self-supervised speech representation learning for automatic speaker verification. In Proceedings of the ICASSP 2022-2022 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Virtual, 7–13 May 2022. [Google Scholar]
  37. Yan, X.C.; Lin, Z.H. A Novel Exploitative and Explorative GWO-SVM Algorithm for Smart Emotion Recognition. IEEE Internet Things J. 2023, 10, 9999–10011. [Google Scholar] [CrossRef]
  38. Chen, S.; Wang, C. WavLM: Large-Scale Self-Supervised Pre-Training for Full Stack Speech Processing. arXiv 2022, arXiv:2110.13900v5. [Google Scholar] [CrossRef]
  39. Wang, H.; Zheng, S.; Chen, Y.; Cheng, L.; Chen, Q. CAM++: A Fast and Efficient Network for Speaker Verification Using Context-Aware Masking. arXiv 2023, arXiv:2303.00332. [Google Scholar]
  40. Slade, S.; Zhang, L.; Huang, H.; Asadi, H.; Lim, C.P.; Yu, Y.; Zhao, D.; Lin, H.; Gao, R. Neural Inference Search for Multiloss Segmentation Models. IEEE Trans. Neural Netw. Learn. Syst. 2023, 35, 15113–15127. [Google Scholar] [CrossRef]
  41. Zhang, L.; Slade, S.; Lim, C.P.; Asadi, H.; Nahavandi, S.; Huang, H.; Ruan, H. Semantic segmentation using Firefly Algorithm-based evolving ensemble deep neural networks. Knowl.-Based Syst. 2023, 277, 110828. [Google Scholar] [CrossRef]
  42. Wang, A.; Xu, Y.; Wei, X.; Cui, B.B. Semantic Segmentation of Crop and Weed using an Encoder-Decoder Network and Image Enhancement Method under Uncontrolled Outdoor Illumination. IEEE Access 2020, 8, 81724–81734. [Google Scholar] [CrossRef]
  43. Wang, A.; Pen, T.; Cao, H.; Xu, Y.; Wei, X.; Cui, B.B. TIA-YOLOv5: An improved YOLOv5 network for real-time detection of crop and weed in the field. Front. Plant Sci. 2022, 13, 1091655. [Google Scholar] [CrossRef]
  44. Shen, C.; Wu, Y.; Qian, G.; Wu, X.; Cao, H.; Wang, C.; Tang, J.; Liu, J. Intelligent Bionic Polarization Orientation Method Using Biological Neuron Model for Harsh Conditions. IEEE Trans. Pattern Anal. Mach. Intell. 2024, 47, 789–806. [Google Scholar] [CrossRef]
  45. Wu, X.; Liu, J.; Shen, C.; Cao, H.; Wang, C.; Tang, J. Vehicle-Mounted Polarization Orientation Strategy Considering Inclined Pavement and Occlusion Conditions. IEEE Trans. Intell. Veh. 2024. early access. [Google Scholar] [CrossRef]
  46. Shen, C.; Zhao, X.; Wu, X.; Cao, H.; Wang, C.; Tang, J.; Liu, J. Multiaperture Visual Velocity Measurement Method Based on Biomimetic Compound-Eye for UAVs. IEEE Internet Things J. 2024, 11, 11165–11174. [Google Scholar] [CrossRef]
  47. Liu, X.; Tang, J.; Shen, C.; Wang, C.; Zhao, D.; Guo, X.; Li, J.; Liu, J. Brain-like position measurement method based on improved optical flow algorithm. ISA Trans. 2023, 143, 221–230. [Google Scholar] [CrossRef] [PubMed]
  48. Zhang, X.; Tang, J.; Cao, H.; Wang, C.; Shen, C.; Liu, J. Cascaded Speech Separation Denoising and Dereverberation Using Attention and TCN-WPE Networks for Speech Devices. IEEE Internet Things J. 2024, 11, 18047–18058. [Google Scholar] [CrossRef]
  49. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. In Advances in Neural Information Processing Systems (NeurIPS); MIT Press: Cambridge, MA, USA, 2017; pp. 5998–6008. [Google Scholar]
  50. Yu, Y.-Q.; Li, W.-J. Densely connected time delay neural network for speaker verification. In Proceedings of the Annual Conference of the International Speech Communication Association (INTER SPEECH), Shanghai, China, 25–29 October 2020; pp. 921–925. [Google Scholar]
  51. Lloyd, S. Least squares quantization in pcm. IEEE Trans. Inf. Theory 1982, 28, 129–137. [Google Scholar] [CrossRef]
  52. Zhang, G.; Zhang, C.; Zhang, H. Improved K-means Algorithm Based on Density Canopy. Knowl.-Based Syst. 2018, 145, 289–297. [Google Scholar] [CrossRef]
  53. Arthur, D.; Vassilvitskii, S. K-Means++: The Advantages of Careful Seeding. In Proceedings of the Eighteenth Annual ACM-SIAM Symposium on Discrete Algorithms, SODA 2007, New Orleans, LA, USA, 7–9 January 2007. [Google Scholar]
  54. Shen, C.; Xiong, Y.; Zhao, D.; Wang, C.; Cao, H.; Song, X.; Tang, J.; Liu, J. Multi-rate strong tracking square-root cubature Kalman filter for MEMS-INS/GPS/polarization compass integrated navigation system. Mech. Syst. Signal Process. 2022, 163, 108146. [Google Scholar] [CrossRef]
  55. India, M.; Safari, P.; Hernando, J. Self Multi-Head Attention for Speaker Recognition. arXiv 2019, arXiv:1906.09890. [Google Scholar]
  56. Scheibler, R.; Bezzam, E.; Dokmanic, I. Pyroomacoustics: A python package for audio room simulation and array processing algorithms. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 351–355. [Google Scholar]
  57. Deng, J.; Guo, J.; Xue, N.; Zafeiriou, S. Arc-face: Additive angular margin loss for deep face recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4690–4699. [Google Scholar]
  58. Zhang, Y.; Lv, Z.; Wu, H.; Zhang, S.; Hu, P.; Wu, Z.; Lee, H.; Meng, H. Mfa-conformer: Multi-scale feature aggregation conformer for automatic speaker verification. arXiv 2022, arXiv:2203.15249. [Google Scholar]
  59. Wang, H.; Liang, C.; Wang, S.; Chen, Z.; Zhang, B.; Xiang, X.; Deng, Y.; Qian, Y. Wespeaker: A Research and Production Oriented Speaker Embedding Learning Toolkit. In Proceedings of the ICASSP 2023 IEEE, Rhodes Island, Greece, 4–10 June 2023. [Google Scholar]
  60. Yakovlev, I.; Makarov, R.; Balykin, A.; Malov, P.; Okhotnikov, A.; Torgashov, N. Reshape Dimensions Network for Speaker Recognition. arXiv 2024, arXiv:2407.18223. [Google Scholar]
  61. Liu, T.; Lee, K.A.; Wang, Q.; Li, H. Golden Gemini is All You Need: Finding the Sweet Spots for Speaker Verification. IEEE/ACM Trans. Audio Speech Lang. Process. 2024, 32, 2324–2337. [Google Scholar] [CrossRef]
  62. Nagrani, A.; Chung, J.S.; Zisserman, A. VoxCeleb: A largescale speaker identification dataset. In Proceedings of the Interspeech, Stockholm, Sweden, 20–24 August 2017; pp. 2616–2620. [Google Scholar]
  63. Fan, Y.; Kang, J.W.; Li, L.T.; Li, K.C.; Chen, H.L.; Cheng, S.T.; Zhang, P.Y.; Zhou, Z.Y.; Cai, Y.Q.; Wang, D. CN-Celeb: A Challenging Chinese Speaker Recognition Dataset. In Proceedings of the International Conference on Acoustics, Speech, and Signal Processing IEEE, Barcelona, Spain, 4–8 May 2020. [Google Scholar]
Figure 1. Various spectrograms for actual sound field situations: (a) clean spectrum; (b) with weak noise; (c) with strong noise; (d) with reverberation; (e) with weak noise and reverberation; (f) with strong noise and reverberation; (g) noise spectrum at low frequency; (h) speaker-like noise spectrum; (i) noise spectrum across frequencies.
Figure 1. Various spectrograms for actual sound field situations: (a) clean spectrum; (b) with weak noise; (c) with strong noise; (d) with reverberation; (e) with weak noise and reverberation; (f) with strong noise and reverberation; (g) noise spectrum at low frequency; (h) speaker-like noise spectrum; (i) noise spectrum across frequencies.
Applsci 15 02924 g001
Figure 2. The architecture of HuBERT and WavLM.
Figure 2. The architecture of HuBERT and WavLM.
Applsci 15 02924 g002
Figure 3. The pipeline of PNR. The PNR processes contaminated speech, which is encoded sequentially through CNN and HuBERT encoders. The representation is then transformed into output by a projection layer.
Figure 3. The pipeline of PNR. The PNR processes contaminated speech, which is encoded sequentially through CNN and HuBERT encoders. The representation is then transformed into output by a projection layer.
Applsci 15 02924 g003
Figure 4. The PNR-TDNN pipeline and the specifics of the proposed TDNN architecture are as follows: (a) depicts the conversion of speech to loss; (b) outlines the structure of the fusion Res2Net within the TDNN module; (c,c1) detail the dimensionality of the features in the SE-attentive mean module; (d) illustrates the dimension transformation of features in the multi-head attentive statistics pooling module, while (d1) provides the computation details of the pooling factor α .
Figure 4. The PNR-TDNN pipeline and the specifics of the proposed TDNN architecture are as follows: (a) depicts the conversion of speech to loss; (b) outlines the structure of the fusion Res2Net within the TDNN module; (c,c1) detail the dimensionality of the features in the SE-attentive mean module; (d) illustrates the dimension transformation of features in the multi-head attentive statistics pooling module, while (d1) provides the computation details of the pooling factor α .
Applsci 15 02924 g004
Figure 5. The algorithm of Canopy/Mini Batch k-means++.
Figure 5. The algorithm of Canopy/Mini Batch k-means++.
Applsci 15 02924 g005
Figure 6. Different Res2nets comparisons.
Figure 6. Different Res2nets comparisons.
Applsci 15 02924 g006
Figure 7. Conference room for noise acquisition.
Figure 7. Conference room for noise acquisition.
Applsci 15 02924 g007
Figure 8. Noise acquisition equipment.
Figure 8. Noise acquisition equipment.
Applsci 15 02924 g008
Figure 9. Number of clusters corresponding to T 1 and T 2 .
Figure 9. Number of clusters corresponding to T 1 and T 2 .
Applsci 15 02924 g009
Figure 10. Convergence of HuBERT/WavLM.
Figure 10. Convergence of HuBERT/WavLM.
Applsci 15 02924 g010
Figure 11. Metrics of different PNR-TDNNs in dataset A/B/C: (a) EER in dataset A; (b) EER in dataset B; (c) EER in dataset C; (d) MinDCF in dataset A; (e) MinDCF in dataset B; (f) MinDCF in dataset C.
Figure 11. Metrics of different PNR-TDNNs in dataset A/B/C: (a) EER in dataset A; (b) EER in dataset B; (c) EER in dataset C; (d) MinDCF in dataset A; (e) MinDCF in dataset B; (f) MinDCF in dataset C.
Applsci 15 02924 g011
Figure 12. Histogram of significant differences between the four baseline models and the THu+ EC+FAM model on dataset A (a, b, c, d, and e represent the comparison notation between the groups corresponding to the proposed T-Hu+EC+FAM, Gemini, RedimNet, CAM++, and ECAPATDNN, respectively).
Figure 12. Histogram of significant differences between the four baseline models and the THu+ EC+FAM model on dataset A (a, b, c, d, and e represent the comparison notation between the groups corresponding to the proposed T-Hu+EC+FAM, Gemini, RedimNet, CAM++, and ECAPATDNN, respectively).
Applsci 15 02924 g012
Figure 13. Best EERs of the four baseline models and the T-Hu + EC + FAM model (blue for dataset A, red for dataset B, and gray for dataset C).
Figure 13. Best EERs of the four baseline models and the T-Hu + EC + FAM model (blue for dataset A, red for dataset B, and gray for dataset C).
Applsci 15 02924 g013
Figure 14. Best EERs of the four baseline models and the T-Hu+EC+FAM model (blue for cnceleb_v2-B, red for VoxCeleb1-B).
Figure 14. Best EERs of the four baseline models and the T-Hu+EC+FAM model (blue for cnceleb_v2-B, red for VoxCeleb1-B).
Applsci 15 02924 g014
Table 1. Parameters of base HuBERT.
Table 1. Parameters of base HuBERT.
CNN Encoder
(Seven Convolutional Layers)
Transformer EncodersProjection Layer
strides [5,2,2,2,2,2,2,2]layers [12]dimension [256]
kernels [10,3,3,3,3,3,2,2]attention heads [12]
Table 2. Reverb configuration (u: uniform distribution).
Table 2. Reverb configuration (u: uniform distribution).
Room/mL (length)u(5, 10)Source/mLL/2+u(−0.2, 0.2)
W (width)u(5, 10)WW/2 + u(−0.2, 0.2)
H (height)u(3, 4)Hu(0.9, 1.8)
T 60 lowu(0.1, 0.5)Microphone/mLL/2 + u(−1.6, −0.8) or u(0.8, 1.6)
middleu(0.5, 1)WW/2 + u(−1.6, −0.8) or u(0.8, 1.6)
highu(1, 1.5)Hu(5, 10)
Table 3. Dataset for speaker verification (percentage: refers to the proportion of selected data to the original unlabeled speech, “✓” denotes containing, noise source: soundsnap and actual record).
Table 3. Dataset for speaker verification (percentage: refers to the proportion of selected data to the original unlabeled speech, “✓” denotes containing, noise source: soundsnap and actual record).
Percentage (%)Noise (SNR: 0 1 )ReverbSpeakers NumberTraining Set AmountTest Set Amount
dataset A6.0587766,000855
dataset B4.6769651,000642
dataset C2.7542230,000359
Table 4. Baseline model’s best EER and MinDCF.
Table 4. Baseline model’s best EER and MinDCF.
MethodologyDataset ADataset BDataset CParamsComputational Efficiency
EER (%)MinDCFEER (%)MinDCFEER (%)MinDCFModel (M)Flops (G)RTF (1 × 10−3)
TDNN7.520.2019.470.1927.350.15511.30.3180.47
Resnet-SE8.400.1848.720.1889.730.15956.13.71911.56
Res2net10.260.20410.770.1959.780.169111.61.0022.97
Eres2net7.610.1967.890.1896.650.16027.42.4287.18
ECAPA-TDNN [20]7.180.1888.440.1906.770.15433.90.9731.23
CAM++ [39]5.970.1787.810.1916.350.15928.70.8134.63
MFA-Conformer [58]7.980.1988.780.1897.400.16077.00.9942.48
Wespeaker [59]10.880.20612.120.19710.220.1740.960.0060.19
RedimNet [60]6.590.1978.190.1927.450.16621.11.2908.79
Gemini [61]7.780.2018.430.1908.120.16326.73.83426.7
Table 5. PNR-TDNN’s best EER and MinDCF in different clustering numbers (Note: Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, EC = ECAPA-TDNN, Wa = WavLM, CA = CAM++).
Table 5. PNR-TDNN’s best EER and MinDCF in different clustering numbers (Note: Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, EC = ECAPA-TDNN, Wa = WavLM, CA = CAM++).
MethodologyDataset ADataset BDataset CParamsComputational Efficiency
EER (%)MinDCFEER (%)MinDCFEER (%)MinDCFModel (M)Flops (G)RTF (1 × 10−3)
Hu + EC(60)6.410.1898.510.1967.560.161405.116.1123.50
Hu + EC(70)6.390.1908.100.1956.840.161405.116.1123.50
Hu + EC(80)6.350.1867.840.1946.010.158405.116.1123.50
Hu + EC(90)6.580.1958.370.1977.350.164405.116.1123.50
Hu + EC(100)12.690.2089.870.2018.810.172405.116.1123.50
Hu + EC(110)25.780.83229.310.97824.690.728405.116.1123.50
Hu + EC(120)32.661.23537.571.42433.711.354405.116.1123.50
Wa + EC(80)6.530.1939.210.1947.560.162405.116.1125.20
Hu + CA(80)6.330.1927.990.1997.160.166404.616.2426.80
S-Hu + EC(80)5.780.1967.340.1915.940.172140.69.5515.67
T-Hu + EC(80)5.700.1868.050.1856.060.15690.28.4613.32
Table 6. Parameters of compressed HuBERT.
Table 6. Parameters of compressed HuBERT.
Transformer EncodersFeed Forward Net
base HuBERTLayers: 12 heads: 12dim: 3072
small HuBERTLayers: 4 heads: 12dim: 1536
tiny HuBERTLayers: 2 heads: 12dim: 512
Table 7. PNR-TDNN’s best EER and MinDCF in multiple enhancement methods (Note: Hu = HuBERT, EC = ECAPA-TDNN, S-Hu = small HuBERT, T-Hu = tiny HuBERT, F = fusion Res2net, A = SE-attentive mean, M = multi-heads attentive statistics pooling).
Table 7. PNR-TDNN’s best EER and MinDCF in multiple enhancement methods (Note: Hu = HuBERT, EC = ECAPA-TDNN, S-Hu = small HuBERT, T-Hu = tiny HuBERT, F = fusion Res2net, A = SE-attentive mean, M = multi-heads attentive statistics pooling).
MethodologyDataset ADataset BDataset C
EER (%)MinDCFEER (%)MinDCFEER (%)MinDCF
1Hu + EC (original Res2net)6.380.1888.370.1936.170.157
2Hu + EC (reverse Res2net)6.510.1888.110.1905.970.163
3Hu + EC + F5.820.1818.090.1925.930.161
4Hu + EC + A6.140.1807.760.1915.380.165
5Hu + EC + M5.820.1837.670.1915.710.158
6Hu + EC + AM5.880.1827.580.1924.870.155
7Hu + EC + FAM5.650.1787.470.1904.840.156
8S-Hu + EC + FAM5.400.1806.940.1905.650.161
9T-Hu + EC + FAM5.660.1827.520.1855.740.153
Table 8. PNR-TDNN’s best EER and MinDCF in ESC50 and Cnceleb noise situations (Note:EC = ECAPA-TDNN, Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, FAM = fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
Table 8. PNR-TDNN’s best EER and MinDCF in ESC50 and Cnceleb noise situations (Note:EC = ECAPA-TDNN, Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, FAM = fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
NoiseMethodologyDataset ADataset BDataset C
SourcesEER (%)MinDCFEER (%)MinDCFEER (%)MinDCF
ESC50/CnCelebEC9.42/8.890.200/0.2039.41/9.450.193/0.1957.31/7.080.170/0.170
CAM++8.21/7.680.197/0.1988.83/10.430.193/0.1978.58/8.300.174/0.177
MFA-Conformer8.95/10.520.204/0.2039.91/9.810.194/0.1926.96/7.030.171/0.167
Wespeaker13.3/13.30.209/0.20713.2/13.60.197/0.19711.0/11.80.176/0.174
RedimNet9.11/8.390.209/0.2088.67/8.890.194/0.19510.63/9.150.176/0.171
Gemini8.76/9.140.207/0.2059.37/9.700.194/0.1948.69/9.360.175/0.175
Hu + EC7.61/7.430.199/0.1958.35/8.950.191/0.1947.11/6.840.160/0.160
Hu + EC + FAM7.41/7.270.196/0.1928.12/8.710.190/0.1906.80/6.630.162/0.156
S-Hu + EC + FAM5.87/5.790.191/0.1887.27/6.980.187/0.1855.25/4.190.155/0.146
T-Hu + EC + FAM6.26/6.440.200/0.1908.35/7.630.188/0.1856.51/5.390.168/0.145
Table 9. PNR-TDNN’s best EER and MinDCF in a wide SNR range (−5∼3 dB) (Note:EC = ECAPA-TDNN, Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, FAM = fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
Table 9. PNR-TDNN’s best EER and MinDCF in a wide SNR range (−5∼3 dB) (Note:EC = ECAPA-TDNN, Hu = HuBERT, S-Hu = small HuBERT, T-Hu = tiny HuBERT, FAM = fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
MethodologyDataset ADataset BDataset C
EER (%)MinDCFEER (%)MinDCFEER (%)MinDCF
EC8.720.2019.590.1886.440.157
CAM++7.430.1968.620.19217.220.334
MFA-Conformer9.250.2039.320.1886.990.161
Wespeaker12.430.20812.040.19510.260.174
RedimNet7.970.2068.390.1876.530.160
Gemini7.770.2028.510.1918.860.166
Hu + EC6.650.1978.570.1916.880.159
Hu + EC + FAM6.490.1967.950.1856.130.154
S-Hu + EC + FAM6.730.1958.480.1896.360.162
T-Hu + EC + FAM7.200.2028.620.1917.020.163
Table 10. VoxCeleb1-A/B and cn-celeb_v2-A/B setup (percentage: refers to the proportion to the original unlabeled speech(VoxCeleb1-S or cn-celeb_v2-S)).
Table 10. VoxCeleb1-A/B and cn-celeb_v2-A/B setup (percentage: refers to the proportion to the original unlabeled speech(VoxCeleb1-S or cn-celeb_v2-S)).
Percentage (%)Training Set AmountTest Set Amount
VoxCeleb1-A88098142
VoxCeleb1-B40.240,398478
cn-celeb_v2-A14.58035155
cn-celeb_v2-B4926,618570
Table 11. PNR-TDNN’s best EER and MinDCF in other public datasets (Note:EC=ECAPA-TDNN, S-Hu = small HuBERT, T-Hu = tiny HuBERT, Hu = HuBERT, FAM= fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
Table 11. PNR-TDNN’s best EER and MinDCF in other public datasets (Note:EC=ECAPA-TDNN, S-Hu = small HuBERT, T-Hu = tiny HuBERT, Hu = HuBERT, FAM= fusion Res2net + SE-attentive mean + multi-head attentive statistics pooling).
MethodologyVoxCeleb1-A/B
(Noise)
VoxCeleb1-A/B
(Noise and Reverb)
cnceleb_v2-A/B
(Noise)
cnceleb_v2-A/B
(Noise and Reverb)
EER (%)MinDCFEER (%)MinDCFEER (%)MinDCFEER (%)MinDCF
EC6.87/13.50.250/0.33118.0/23.30.308/0.50912.6/8.680.201/0.33414.1/14.00.241/0.438
CAM++15.9/15.50.306/0.40934.7/59.70.403/0.93013.8/7.520.235/0.31016.2/13.60.241/0.426
MFA-Conformer17.1/15.50.329/0.42119.3/24.20.356/0.57313.4/11.90.241/0.50415.8/18.70.255/0.560
Wespeaker13.9/25.70.319/0.58222.2/34.70.351/0.62416.6/20.30.246/0.76818.0/27.30.270/0.818
RedimNet13.9/15.10.306/0.41216.4/25.60.335/0.58014.8/10.30.236/0.36917.2/16.50.241/0.579
Gemini17.0/18.90.351/0.46820.5/22.40.329/0.54516.7/12.10.255/0.49319.9/16.90.266/0.620
Hu + EC4.62/9.960.247/0.21217.1/21.70.304/0.53612.2/7.150.203/0.30914.2/13.20.244/0.445
Hu + EC + FAM3.30/9.270.208/0.20515.8/20.90.296/0.52311.9/6.730.197/0.28113.8/12.80.240/0.414
S-Hu + EC + FAM3.06/9.780.195/0.27515.0/19.20.307/0.50510.9/4.870.240/0.25812.9/11.90.233/0.408
T-Hu + EC + FAM4.39/9.190.219/0.24416.5/20.50.313/0.5429.69/5.260.221/0.26614.5/13.30.241/0.449
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, X.; Tang, J.; Cao, H.; Wang, C.; Shen, C.; Liu, J. A Self-Supervised Method for Speaker Recognition in Real Sound Fields with Low SNR and Strong Reverberation. Appl. Sci. 2025, 15, 2924. https://doi.org/10.3390/app15062924

AMA Style

Zhang X, Tang J, Cao H, Wang C, Shen C, Liu J. A Self-Supervised Method for Speaker Recognition in Real Sound Fields with Low SNR and Strong Reverberation. Applied Sciences. 2025; 15(6):2924. https://doi.org/10.3390/app15062924

Chicago/Turabian Style

Zhang, Xuan, Jun Tang, Huiliang Cao, Chenguang Wang, Chong Shen, and Jun Liu. 2025. "A Self-Supervised Method for Speaker Recognition in Real Sound Fields with Low SNR and Strong Reverberation" Applied Sciences 15, no. 6: 2924. https://doi.org/10.3390/app15062924

APA Style

Zhang, X., Tang, J., Cao, H., Wang, C., Shen, C., & Liu, J. (2025). A Self-Supervised Method for Speaker Recognition in Real Sound Fields with Low SNR and Strong Reverberation. Applied Sciences, 15(6), 2924. https://doi.org/10.3390/app15062924

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop