Next Article in Journal
Resolving Multi-Path Interference in Compressive Time-of-Flight Depth Imaging with a Multi-Tap Macro-Pixel Computational CMOS Image Sensor
Previous Article in Journal
Experimental Verification of Dielectric Models with a Capacitive Wheatstone Bridge Biosensor for Living Cells: E. coli
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Speech Enhancement by Multiple Propagation through the Same Neural Network

Institute of Automatic Control and Robotics, Poznan University of Technology, 60-965 Poznan, Poland
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(7), 2440; https://doi.org/10.3390/s22072440
Submission received: 17 February 2022 / Revised: 11 March 2022 / Accepted: 18 March 2022 / Published: 22 March 2022
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Monaural speech enhancement aims to remove background noise from an audio recording containing speech in order to improve its clarity and intelligibility. Currently, the most successful solutions for speech enhancement use deep neural networks. In a typical setting, such neural networks process the noisy input signal once and produces a single enhanced signal. However, it was recently shown that a U-Net-based network can be trained in such a way that allows it to process the same input signal multiple times in order to enhance the speech even further. Unfortunately, this was tested only for two-iteration enhancement. In the current research, we extend previous efforts and demonstrate how the multi-forward-pass speech enhancement can be successfully applied to other architectures, namely the ResBLSTM and Transformer-Net. Moreover, we test the three architectures with up to five iterations, thus identifying the method’s limit in terms of performance gain. In our experiments, we used the audio samples from the WSJ0, Noisex-92, and DCASE datasets and measured speech enhancement quality using SI-SDR, STOI, and PESQ. The results show that performing speech enhancement up to five times still brings improvements to speech intelligibility, but the gain becomes smaller with each iteration. Nevertheless, performing five iterations instead of two gives additional a 0.6 dB SI-SDR and four-percentage-point STOI gain. However, these increments are not equal between different architectures, and the U-Net and Transformer-Net benefit more from multi-forward pass compared to ResBLSTM.

1. Introduction

Speech enhancement allows for the extraction of the desired speech signal from a mixture of speech with interfering sounds coming from different sources. These methods can be used in hearing aids [1], smartphones [2], or, as a pre-processing step, in automatic speech or speaker recognition [3].
In recent years, great progress has been made with deep neural networks (DNNs) applied to speech enhancement. In these methods, a degraded signal is provided with the input of the neural network, which is trained to provide an estimate of the clean, undistorted signal. The representation of a signal at the input of the DNN can be a spectrogram, MFCCs (mel-frequency cepstral coefficients), or raw (time-domain) signal. The network can be trained to obtain a mask (e.g., an IRM (ideal ratio mask, [4]), a cIRM (a complex ideal ratio mask, [5])), or a representation from which the signal can be reconstructed (for example, the STFT (short-term Fourier transform) of the clean speech signal).
There are many neural network architectures for speech enhancement that can be found in the literature. These include DNNs with fully-connected layers [6,7], fully-convolutional networks [8], recurrent neural networks (RNNs) [9], and combinations of convolutional and recurrent neural networks [10]. Convolutional layers used in speech-enhancement neural networks often use a dilation greater than one, which helps in utilizing long-range dependencies without increasing the number of models’ parameters [11,12,13].
An interesting example of RNNs for speech enhancement is ResBLSTM [14,15], which uses a skip connection between the input to LSTM (long short-term memory) and its output. Such a connection results in a performance boost in comparison to pure stacked LSTM layers. The presence of such a connection forces the layer to learn residual mapping, and it can potentially take long context into account.
Recently, a multi-head attention [16] has been used in speech enhancement [17,18,19]. This technique provides an alternative way to take into account possibly large contexts of the input signal. Neural models that use self-attention are, by convention, named transformer networks. Currently, a lot of research is devoted to optimizing ways of encoding positional information in sequences on the input to the transformer networks. Except for the absolute positional encoding used in the original research paper [16], relative [20] and continuous dynamic [21] positional encodings have also been proposed, but their performance in speech enhancement has not yet been tested.
One of the neural network architectures used in this work comes from our earlier work on U-Net-based speech enhancement [22,23]. The results presented in these works suggest that when the time-frequency resolution of hidden activations (feature maps) is not reduced by max-pooling or strided convolutions, a better performance can be achieved in terms of SDR (signal-to-distortion ratio), STOI (spectro-temporal objective intelligibility), and PESQ (perceptual evaluation of speech quality). However, to achieve this performance, the receptive field needs to be enlarged using dilated convolutions or recurrent layers.
Although there has been progress in DNN-based speech enhancement, the enhanced speech is still not perfect and may not be intelligible. When this occurs, it is mainly due to the existence of the residual noise and distortions of the speech. In the case of residual noise, one can imagine that processing the noisy signal many times with the same network can bring the desired effect. However, the accumulation of distortions and loss of the speech information can make this task difficult.
A recent technique to improve performance of neural networks is progressive learning. In the context of speech enhancement, progressive learning was applied by sequentially connecting many sub-networks with shared weights [24]. The subsequent networks had different SNRs of target representations. The target SNRs in the subsequent sub-networks were from the lowest (slightly enhanced speech) to the highest (clean speech). As such, each sub-network removed a portion of the noise present in the noisy input signal, hence the name “progressive”.
In this paper, multi-pass speech enhancement is proposed. Using this method, in the original neural network architecture (which can be any one of the most popular neural networks used for speech enhancement), three main blocks are identified: input, base, and output. The output from the input block is processed by the same base block many times. Finally, the outputs are computed using the output block. We test this scheme for the following neural network architectures: (1) U-Net with dilated convolutions, (2) RNN, and (3) Transformer. The improvement in speech enhancement after subsequent passes through the base block is achieved thanks to a special skip connection from the previous pass, as in [25].
The main difference between the method proposed in this paper and the progressive learning described in [24,26,27] is that the original noisy signal is provided at the input of each step, while the target at the output of each step is the clean signal. In this regard, instead of removing a certain portion of the input noise, our proposed solution attempts to remove all the noise at each processing step, but with each additional step, this process becomes more refined.
The paper is structured as follows. In Section 2, the undertaken problem is formulated, the proposed multi-pass approach is introduced, and the datasets and experimental setup are described. In Section 3, the results are presented and analyzed. Next, the discussion of the results in the context of the literature is presented in Section 4. Finally, the conclusions are enumerated in Section 5.

2. Materials and Methods

2.1. Problem Formulation

We consider the speech enhancement problem in which a speech signal is corrupted by an additive noise, i.e.,
y ( k ) = x ( k ) + d ( k ) ,
where y ( k ) , x ( k ) , and d ( k ) are the k′th samples of the mixture, clean speech, and noise, respectively.
In order to attenuate the noise, a neural network can be employed. In this article, we consider networks that accept on input a tensor Y R N × B × T × 2 that contains STFT of the noisy speech, where the first dimension denotes an example in the batch with N examples, B and T denote the number of frequency bands and the number of frames, respectively, and the last dimension refers to real and imaginary part of the STFT. The neural network maps this input tensor to the output tensor O R N × B × T × 2 , which should predict M R N × B × T × 2 containing cIRM (processed by a nonlinear function to limit the range of values; see Section 2.6 for more details), which is further used to compute the STFT of the estimated clean speech signal.
In this article, the speech enhancement neural network is built from three main subnetworks: input layers, base layers, and output layers. The input subnetwork extracts a representation of the input mixture. Next, this representation is processed many times by the main subnetwork’s base layers, which refine the representation in such a way that the output subnetwork can project these representations to cIRM masks, which will provide a high-quality speech signal.

2.2. Multi-Pass

The general processing scheme used in multi-pass speech enhancement is shown in Figure 1. The input tensor is processed by input layers to extract a representation of the input signal. Next, this representation is processed by the base subnetwork L times. There are skip connections that add the feature representation of the input signal to the output of the base subnetwork before each iteration. The output of the base subnetwork after each iteration is processed by the output layers that project the representation to the cIRM mask, O l for l = 1 , , L , which is further used to reconstruct the estimated signal (see Section 2.5 for more information).
During the training for each output, its loss is computed as
L l = 1 2 N B T n = 1 N b = 1 B t = 1 T p = 1 2 ( ( O l ) n b t p M n b t p ) 2 .
The total loss used to compute gradient is the sum of the losses from all outputs divided by the number of outputs
L = 1 L l = 1 L L l .
Such a combined loss function allows the trained network to process the input signal multiple times. During the second, third, and each additional iteration, the network can make use of the previously-denoised signal representation as well as the original signal representation, which allows it to enhance the signal further. The skip connections that add the feature representation of the input signal to the output from base layers before each consecutive iteration are a critical part of the proposed solution, as they prevent the accumulation of the distortion added by base layers at each pass.
In the proposed scheme, unlike in the progressive learning, each enhancement step is performed by the same set of layers and weights, which are trained to provide the best possible output signal after each pass. During inference, the multi-pass network can be used to perform any number of enhancement iterations between 1 and L. In that case, only the final output should be used, as it contains the most refined cIRM mask. In progressive learning, each processing stage is realized by a separate network that removes a certain, pre-defined amount of noise.

2.3. Base Sub-Networks

The multi-pass speech enhancement can be applied on top of any existing neural network architecture. The only requirement is the distinction of the input layers, base layers, and output layers within the neural network. The best choice for input and output layers is the usage of single convolutional layers in each case. The output feature tensor from input layers needs to have the same shape as the input tensor to the output layers; otherwise, the skip connection between enhancement iteration will not be possible. If no suitable layers can be identified within the existing architecture, they can be added. In this work, we used this approach with three popular neural network designs.

2.3.1. Dilated U-Net

This is a U-Net-based architecture similar to the one used in [25]. The Dilated U-Net uses two-dimensional convolutional operations to map the spectrogram representation of the noisy signal into the output cIRM mask. The base subnetwork includes three levels of convolutional layers, each separated by a skip connection that connects the encoder to the decoder. Each convolutional layer consists of 54 convolutional filters of size 3 × 3.
The feature maps at the input to each aforementioned convolutional layer are extended with features produced by processing blocks called CEMs (context extension modules). The role of CEM is to provide wide context in terms of either time or frequency without reducing the time-frequency resolution between the subsequent levels of the U-Net. This is achieved with dilated convolutions. Each CEM consists of two parallel colvolutional layers, each using eight filters and a different dilation rate. The outputs of the two layers are concatenated at the output of each CEM. The variant of CEM that extends the context in time dimension uses kernels of size (1, 7) and dilation rates of (1, 3) and (1, 4). The frequency-dilated CEMs use (7, 1) kernels and (2, 1) and (3, 1) dilation rates.
The number of filters used by the “main” convolutional filters of the U-Net (54 filters) and the number of filters inside the CEMs ( 2 × 8 filters) have been established using the same procedure as described in [25] given the context of current experiments, which essentially means that the highest possible values were used that allowed it to perform the training on a GPU with 11 GB of RAM.
In the case of Dilated U-Net, the input layers (as defined in Figure 1) include a single convolutional layer with 54 filters with 3 × 3 kernels and stride equal to two in both time and frequency dimensions. Analogically, the output layers include a single transposed convolution with two filters (one for the real and one for the imaginary part), kernels of size 6 × 6, and stride equal to two in both dimensions. For an overview of this and other architectures, please refer to Figure 2. Please note that the first number in each block denotes the number of filters or recurrent units in case or recurrent layers.

2.3.2. ResBLSTM

The ResBLSTM differs significantly from the aforementioned U-Net in terms of feature tensor composition. This base subnetwork does not distinguish between the channel dimension of the feature tensor and the frequency dimension (i.e., features at the output of any given layer in the ResBLSTM are not localized in frequency). The features at the input of this base subnetwork are organized in a three-dimensional array F R N × C × T ; i.e., for each spectrogram time frame, there are C features. Each of these features potentially depends on all signal’s frequencies. The recurrent layers from the ResBLSTM network iterate over time frames accepting the C features on the input at each step. The output layers transform this signal representation into an array, which can be reshaped and transposed into the output tensor that contains the cIRM mask.
A detailed structure of the multi-pass-enabled ResBLSTM is depicted in Figure 2. In order to fit the original noisy speech spectrogram into the three-dimensional tensor, the ResBLSTM input layers start with transposition and reshape from N × B × T × 2 into N × 2 B × T . This tensor is then processed by a single one-dimensional convolutional layer with 512 filters of kernel size one and stride equal to one. This produces a feature tensor of shape N × C × T with C = 512 that is further processed by the base subnetwork. The output layers use convolution with a kernel size of one and the number of filters is double the number of frequency bands (one set for the real and one for the imaginary part) to produce output tensor of shape N × 2 B × T . This is followed by the reverse reshape and transposition that restores the original tensor shape ( N × B × T × 2 ).
The ResBLSTM base subnetwork is composed of three bidirectional LSTM [28] layers with 256 units in each direction, concatenated on output and encircled with the residual connection. This module concludes with a single convolutional layer with 512 filters with kernel of size one followed by batch normalization and ELU (exponential linear unit) nonlinearity [29] defined as
φ ( x ) = x x > 0 e x 1 x 0 .
Using this configuration, the ResBLSTM can be trained to transform features into a representation from which cIRM mask can be obtained. Using the capacity of a LSTM layer to store information for many recurrent steps, the ResBLSTM network can compute elements of the cIRM mask based on potentially large context, which makes it a very effective model for speech enhancement.

2.3.3. Transformer-Net

Transformer networks are another powerful tools that can be employed to perform speech enhancement. Unlike the recurrent layers, they use the self-attention monarchism that can model relation between any two locations in the input signal independent on their distance. This allows the model to make use of potentially large context, which, as was previously shown in [22], is crucial for high quality speech enhancement. When working with spectrogram representations of signals, two types of distances should be considered: distance in time and frequency. To properly model each of them, two attention modules are used - one for each dimension.
Because the Transformer-Net distinguishes frequency dimension as being separate from the channel dimension, the input and output layers are configured identically as in the case of Dilated U-Net. The Transformer-Net base subnetwork starts with a sequence of three convolutional layers, each with K = 60 filters with 3 × 3 kernels. Each layer ends with batch normalization and ELU nonlinearity and additionally the second layer uses stride equal two in frequency dimension (analogically as on the second level of the Dilated U-Net). This is coupled with transposed convolution with K filters of size 6 × 3 and stride equal two in frequency dimension, batch normalization and ELU on the output on this base subnetwork.
The central part of this base subnetwork is composed of a sequence of two transformer encoders, the first modeling features in frequency dimension and the second one in time dimension. Each of the two encoders uses a multi-head attention block [16] with four heads and d m o d e l = K (i.e., 15 per head). Each multi-head attention block is followed by a point-wise feed-forward network (i.e., features for each time-frequency location are transformed using the feed-forward network independently), each such feed-forward network is composed of two fully-connected layers with K neurons and ELU nonlinearities. For performance reasons the transformers lack the normalization layers after each multi-head attention and point-wise feed-forward networks. The value of K, similarly as in the case of Dilated U-Net, was set experimentally to fill the available GPU memory.

2.4. Dataset

The effectiveness of multi-pass speech enhancement was tested on the popular Wall Street Journal (WSJ0) dataset [30], from which only the utterances included on the SI-84 list were used. From the 83 speakers included on the list, 3 random male and 3 random female speakers were held-out for testing. This test set included 599 speech utterances. The remaining 77 speakers contributing 6637 utterances were used for training and validation (i.e., control of the model over-fitting). As can be seen, there was no overlap between the test and the training datasets.
The WSJ0 training speech utterances were mixed with a diverse set of noises coming from the FreeField72k dataset, which was also used in [25]. Motivated by the popularity of the freefield1010 dataset [31], which was also included in the DNS challenge [32], the FreeField72k extends its predecessor by taking a much bigger sample of the free field recordings from the source database. In the process of creation of the FreeField72k, the Freesound online database [33] was queried, and all of the recordings tagged with the “field-recording” tag were downloaded. For the sake of effectiveness, all recordings shorter than 10 s and those whose download time exceeded 30 s were skipped. A resulting set of 24,237 recordings was resampled to 8k samples per second and segmented into 10 s non-overlapping fragments with the restriction that no more than five fragments could be taken from one recording. This process yielded 93,301 10 s noise fragments, from which 23% were removed due to low signal level. Finally, the FreeField72k included 72,028 10 s audio excerpts coming from 19,990 original Freesound database recordings and totaled to 200 h of diverse, real-life noise.
In the test phase, the trained models were fed with WSJ0 test speech utterances mixed with babble noise (Noisex-92, [34]) and shopping mall noise (DCASE, [35]).

2.5. Features

In all experiments, the neural network input consisted of a scaled STFT of the speech–noise mixture sampled at 8k samples per second. The STFT was obtained by dividing the original signals into 25 ms frames with hop-time of 10 ms and applying the Hanning window. For each frame, a 512-point fast Fourier transform was derived. The resulting STFT was eventually divided by a constant (40.0) in order to reduce the range on values on the input to the neural network to roughly fit the range −1.0 to 1.0, which is known to speed up the training process.
During the inference, The cIRM mask predicted by the neural network was applied to the STFT of the mixture (complex multiplication). Finally, after multiplication by 40.0 (inverse scaling), the inverse STFT was performed in order to obtain the time-domain-enhanced signal.

2.6. Targets

All networks were trained to predict the cIRM mask. In order to limit the range of values in cIRM masks, a compression using hyperbolic tangent was used. The slope parameter of this compressing function was set to 0.1, which effectively limited the targets’ values to the range of (−10, 10). During the inference, the predicted mask was uncompressed using the inverse of this compressing function.

2.7. Hyperparameters and Training

Each neural network was trained for 400 epochs, and during each epoch, 745 model updates were performed. First, the dataset of 6637 training speech utterances was split into actual training samples (90%—5974 samples) and validation samples (10%—663 samples). During one training epoch, each of the 5974 training speech utterances was used once (but in random order). From each training speech utterances, a random 5 s excerpt was extracted and mixed with a random 5 s excerpt from the FreeField72k dataset at SNR randomly chosen from [−5, −4, −3, −2, −1, 0] dB. If a training speech utterance was shorter than 5 s, it was padded on both sides with zeros, and this padded area was masked-out from the loss computation. For each speech–noise mixture, a training target based on a cIRM mask was computed. Training examples comprising the speech–noise mixtures and a target compressed cIRMs were grouped in batches of eight and used for training. The random number generators used in the experiments were initialized with fixed seeds so that the training data sequence was identical in all experiments.
The training used the Adam optimizer [36] with an initial learning rate of 0.002, which was further reduced by 1% after each epoch.
After each training epoch, a validation was carried out that used all of the 663 validation speech samples mixed with a random (but repeatable) sequence of FreeField72k samples at 0 dB. If a new best value of SI-SDR was observed on these validation data, a new model snapshot was taken. In the case of multi-forward experiments, only the final output from the network ( O L ) was used to compute the SI-SDR. The last model snapshot taken during each training was evaluated on the 599 test speech utterances mixed with different noises as described in next section.

2.8. Experiments

To evaluate the performance of the multi-forward speech enhancement, six models were trained. Each of the three architectures described in Section 2.3 was trained two times: first with L = 1 , which is the baseline approach that performs the speech enhancement once, and with L = 5 , which is when the network is trained to perform the speech enhancement up to five times.
Models trained with L = 1 were tested with only one speech enhancement iteration (using its only output). The model trained with L = 5 was tested on all five outputs, i.e., when performing 1, 2, 3, 4, and 5 enhancement passes ( O 1 , O 2 , , O 5 ). In each case, the models were tested for removal of babble noise from Noisex-92 and shopping mall noise from DCASE at −5, 0, 5, and 10 dB SNR. The effectiveness of the enhancement was evaluated using SI-SDR (scale-independent SDR), STOI, and PESQ metrics.

3. Results

The performance of the proposed method (in terms of SI-SDR and STOI) for babble noise at 0dB SNR is shown in Figure 3 and Figure 4. The horizontal dashed lines represent the performance of the models trained with L = 1 and therefore serve as baselines for the three base architectures (Dilated U-Net, ResBLSTM, and Transformer-Net) trained in the multi-pass manner. In the case of the SI-SDR metric, it can be noticed that the multi-pass approach outperformed baselines by 1.36 dB for Transformer-Net, 1.2 dB for Dilated U-Net, and 0.98 dB for ResBLSTM. In the case of STOI, the improvements are: 0.041 for ResBLSTM, 0.063 for Transformer-Net, and finally 0.082 for Dilated U-Net.
When the Dilated U-Net and Transformer-Net architectures are trained in a multi-pass manner ( L = 5 ), their performance for one-pass denoising during the test (for output O 1 ) is worse than for the same network trained using single-pass ( L = 1 ). This is not always true for ResBLSTM, where the SI-SDR for ( L = 5 and output O 1 ) is higher than for the baseline model ( L = 1 , O 1 ).
The highest difference in performance of the multi-pass approach is between the first and the second pass. With additional passes, the benefit decreases. The highest difference between four and five passes is achieved with Transformer-Net (0.05 dB SI-SDR and 0.006 STOI), while the lowest is achieved for ResBLSTM (0.01 dB SI-SDR and 0.001 STOI).
In terms of SI-SDR, the best-performing system is Transformer-Net, with five passes (SDR = 8.63 dB). However, this is not the case for the STOI metric, where the best performance was achieved using ResBLSTM (STOI = 0.871).
The results are presented in Table 1 and Table 2. They contain evaluation metrics obtained for babble noise and shopping mall noise, respectively, for all tested conditions (SNR −5, 0, 5, and 10 dB) and all metrics (SI-SDR, STOI, and PESQ).
The multi-pass approach gave substantial performance gain for all tested conditions. In the case of babble noise, the best results in terms of SI-SDR were achieved by Transformer-Net, while in terms of STOI, the best results were achieved by ResBLSTM (at lower SNR values) and Transformer-Net (at higher SNR values). In the case of shopping mall noise, Transformer-Net was the best in terms of SI-SDR (at all SNR values) and PESQ (at higher SNR values), while in terms of STOI, Dilated U-Net gave the best outcome.
It can be also noticed that ResBLSTM can bring some drop in performance between four and five passes for the PESQ metric. This occurred for both babble and shopping mall noise.
An example of multi-forward speech enhancement performed by Transformer-Net is provided in Figure 5. On the top left, the magnitude spectrogram of a noisy speech utterance is presented. Below it, there are the spectrograms of the denoised signals obtained after subsequent passes through the Transformer-Net ( O = 1 , 2, 3, 4, and 5). Furthermore, finally, at the bottom, the clean (target) utterance is presented. At the right-hand side of each spectrogram of reconstructed clean speech signal, there is an image showing the difference between this spectrogram and the spectrogram located directly above it. As such, these right-hand-side images show the additional amount of noise removed at each pass. It can be noticed that with subsequent passes, residual noise patterns (concentrated in time and frequency) are removed.

4. Discussion

The presented results show a very clear advantage of adding the multi-pass functionality to the existing well-established neural network solutions for speech enhancement. Even if the starting architecture does not contain suitable candidate layers for input and output sub-networks as described in Figure 1, it is enough to add one or two convolutional layers at the beginning and/or end of the network to serve this purpose.
The only true disadvantage of the proposed solution that we were able to identify is the additional requirement for GPU memory during training. When using the loss function proposed in this work, and under the assumption that the base layers from Figure 1 contain the vast majority of the network’s layers, the requirement to train a multi-pass network for L passes in almost L times as high as for the baseline network that is trained to do a single enhancement pass. Some possible solutions were presented in [25], but they do not seem to scale well with values of L higher than three. However, the memory consumption during inference is identical as for the baseline model, and obviously the processing time scales linearly with the number of passes.
Amongst the three tested neural network architectures, this was a considerable factor for the Dilated U-Net and the Transformer-Net, which use feature tensors with two spatial dimensions: time and frequency. Because these networks preserve the feature’s location in the frequency dimension, their memory consumption is significantly higher, so the additional memory required to train these models in multi-pass manner should be taken into consideration. This is a much smaller issue in the case of the ResBLSTM, where features are located only in time, and therefore they are much more compact.
Finally, we would like to highlight two areas where the proposed solution might be especially useful. The first application is related to mobile devices, where the network sizes are relatively small. In this case, the training is usually performed on separate machines with abundant resources, so increasing the amount of memory used during training should not be an issue. However, the big advantage of the proposed solution is that it yields a potentially small, flexible model that is able to perform speech enhancement of varying quality, and the quality scales logarithmically with the number of passes. We can therefore deliver the same model to the smaller/older-generation devices, which will perform one-pass enhancement and provide a well-enhanced signal, and to the more powerful devices which, can perform multiple passes in the same time window to benefit from the increased speech denoising quality.
The second application that might significantly benefit from the proposed solution is the high-throughput network services that usually deal with varying load. Because the multi-pass networks are flexible, the number of speech enhancement iterations can be dynamically adjusted depending on the current utilization of server resources. For example, if the load is low, all five iterations might be performed. However, when the load increases, the number of iterations might be dynamically reduced so that the service is able to process all requests on time. Such a solution should result in a more effective resource utilization and improve the end-user experience.

5. Conclusions

A deepened research of the multi-pass enhancement methods has revealed that this technique is applicable to a wide range of machine learning models, and in each case it brings significant improvement in the speech enhancement performance as measured by SI-SDR, STOI, and PESQ. As expected, with each additional iteration, the increments in performance become smaller, but it is still reasonable to perform five passes, and additional ones should improve the signal quality even further. The proposed solution is quite simple in theory and implementation and should be especially useful with smaller models that have limited hardware resources available during inference or to optimize resource utilization by network services that perform online speech enhancement.

Author Contributions

Conceptualization, T.G.; methodology, S.D. and T.G.; software, T.G.; validation, T.G.; formal analysis, T.G. and S.D.; investigation, T.G.; resources, S.D. and T.G.; data curation, T.G. and S.D.; writing—original draft preparation, T.G. and S.D.; writing—review and editing, T.G. and S.D.; visualization, T.G. and S.D.; supervision, T.G. and S.D.; project administration, T.G.; funding acquisition, S.D. All authors have read and agreed to the published version of the manuscript.

Funding

This paper was prepared with the subvention financial means (Project number: 0211/SBAD/0222).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

All datasets used in this study are available (free or commercially). The scripts which we used to generate noise samples used for model training are available at https://drive.google.com/drive/folders/1KgOjTTcFfglBeSJSBM72WJY_cdek2Qmh?usp%3Dsharing&sa=D&source=docs&ust=1646776164896590&usg=AOvVaw2nK4omJ5lIZJ90m_9Ufd9H, accessed on 17 February 2022.

Acknowledgments

We would like to thank StethoMe sp. z o.o. for providing computational resources and support that were needed to conduct experiments described in this work.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Wang, D. Deep learning reinvents the hearing aid. IEEE Spectr. 2017, 54, 32–37. [Google Scholar] [CrossRef] [PubMed]
  2. Ananthakrishnan, K.S.; Dogancay, K. Recent trends and challenges in speech-separation systems research—A tutorial review. In Proceedings of the TENCON 2009—2009 IEEE Region 10 Conference, Singapore, 23–26 November 2009; IEEE: Piscataway, NJ, USA, 2009; pp. 1–6. [Google Scholar]
  3. Mowlaee, P.; Saeidi, R.; Christensen, M.G.; Martin, R. Subjective and objective quality assessment of single-channel speech separation algorithms. In Proceedings of the 2012 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Kyoto, Japan, 25–30 March 2012; IEEE: Piscataway, NJ, USA, 2012; pp. 69–72. [Google Scholar]
  4. Wang, Y.; Narayanan, A.; Wang, D. On training targets for supervised speech separation. IEEE/ACM Trans. Audio Speech Lang. Process. (TASLP) 2014, 22, 1849–1858. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Williamson, D.S.; Wang, D. Time-frequency masking in the complex domain for speech dereverberation and denoising. IEEE/ACM Trans. Audio Speech Lang. Process. 2017, 25, 1492–1501. [Google Scholar] [CrossRef] [PubMed]
  6. Huang, P.S.; Kim, M.; Hasegawa-Johnson, M.; Smaragdis, P. Deep learning for monaural speech separation. In Proceedings of the 2014 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Florence, Italy, 4–9 May 2014; IEEE: Piscataway, NJ, USA, 2014; pp. 1562–1566. [Google Scholar]
  7. Huang, P.S.; Kim, M.; Hasegawa-Johnson, M.; Smaragdis, P. Joint optimization of masks and deep recurrent neural networks for monaural source separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 2136–2147. [Google Scholar] [CrossRef] [Green Version]
  8. Park, S.R.; Lee, J. A fully convolutional neural network for speech enhancement. arXiv 2016, arXiv:1609.07132. [Google Scholar]
  9. Chen, J.; Wang, D. Long short-term memory for speaker generalization in supervised speech separation. J. Acoust. Soc. Am. 2017, 141, 4705–4714. [Google Scholar] [CrossRef]
  10. Zhao, H.; Zarar, S.; Tashev, I.; Lee, C.H. Convolutional-Recurrent Neural Networks for Speech Enhancement. arXiv 2018, arXiv:1805.00579. [Google Scholar]
  11. Tan, K.; Chen, J.; Wang, D. Gated residual networks with dilated convolutions for monaural speech enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2019, 27, 189–198. [Google Scholar] [CrossRef]
  12. Pirhosseinloo, S.; Brumberg, J.S. Monaural Speech Enhancement with Dilated Convolutions. Proc. Interspeech 2019, 2019, 3143–3147. [Google Scholar] [CrossRef] [Green Version]
  13. Pandey, A.; Wang, D. TCNN: Temporal Convolutional Neural Network for Real-Time Speech Enhancement in The Time Domain. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 6875–6879. [Google Scholar]
  14. Nicolson, A.; Paliwal, K.K. Deep learning for minimum mean-square error approaches to speech enhancement. Speech Commun. 2019, 111, 44–55. [Google Scholar] [CrossRef]
  15. Kim, J.; El-Khamy, M.; Lee, J. Residual LSTM: Design of a deep recurrent architecture for distant speech recognition. arXiv 2017, arXiv:1701.03360. [Google Scholar]
  16. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention Is All You Need. arXiv 2017, arXiv:1706.03762. Available online: http://xxx.lanl.gov/abs/1706.03762 (accessed on 8 January 2022).
  17. Wang, K.; He, B.; Zhu, W.P. TSTNN: Two-stage Transformer based Neural Network for Speech Enhancement in the Time Domain. arXiv 2021, arXiv:2103.09963. Available online: http://xxx.lanl.gov/abs/2103.09963 (accessed on 8 January 2022).
  18. Xu, Z.; Jiang, T.; Li, C.; Yu, J. An Attention-augmented Fully Convolutional Neural Network for Monaural Speech Enhancement. In Proceedings of the 2021 12th International Symposium on Chinese Spoken Language Processing (ISCSLP), Hong Kong, China, 24–26 January 2021; pp. 1–5. [Google Scholar] [CrossRef]
  19. Zhang, Q.; Song, Q.; Ni, Z.; Nicolson, A.; Li, H. Time-Frequency Attention for Monaural Speech Enhancement. arXiv 2021, arXiv:2111.07518. Available online: http://xxx.lanl.gov/abs/2111.07518 (accessed on 8 January 2022).
  20. Shaw, P.; Uszkoreit, J.; Vaswani, A. Self-Attention with Relative Position Representations. In Proceedings of the 2018 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 2 (Short Papers), New Orleans, LA, USA, 1–6 June 2018; Association for Computational Linguistics: New Orleans, LA, USA, 2018; pp. 464–468. [Google Scholar] [CrossRef]
  21. Liu, X.; Yu, H.F.; Dhillon, I.; Hsieh, C.J. Learning to Encode Position for Transformer with Continuous Dynamical Model. In Proceedings of the 37th International Conference on Machine Learning, Vienna, Austria, 13–18 July 2020; Volume 119, pp. 6327–6335. [Google Scholar]
  22. Grzywalski, T.; Drgas, S. Application of recurrent U-net architecture to speech enhancement. In Proceedings of the Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), Poznan, Poland, 19–21 September 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 82–87. [Google Scholar]
  23. Grzywalski, T.; Drgas, S. Using recurrences in time and frequency within U-net architecture for speech enhancement. In Proceedings of the ICASSP 2019—2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; IEEE: Piscataway, NJ, USA, 2019; pp. 6970–6974. [Google Scholar]
  24. Li, A.; Yuan, M.; Zheng, C.; Li, X. Speech enhancement using progressive learning-based convolutional recurrent neural network. Appl. Acoust. 2020, 166, 107347. [Google Scholar] [CrossRef]
  25. Grzywalski, T.; Drgas, S. Speech enhancement by iterating forward pass through U-net. In Proceedings of the 2020 Signal Processing: Algorithms, Architectures, Arrangements, and Applications (SPA), Poznan, Poland, 23–25 September 2020; IEEE: Piscataway, NJ, USA, 2020. [Google Scholar] [CrossRef]
  26. Gao, T.; Du, J.; Dai, L.R.; Lee, C.H. SNR-Based Progressive Learning of Deep Neural Network for Speech Enhancement. Interspeech 2016, 2016, 3713–3717. [Google Scholar]
  27. Gao, T.; Du, J.; Dai, L.R.; Lee, C.H. Densely connected progressive learning for lstm-based speech enhancement. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; IEEE: Piscataway, NJ, USA, 2018; pp. 5054–5058. [Google Scholar]
  28. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. Available online: http://xxx.lanl.gov/abs/https://direct.mit.edu/neco/article-pdf/9/8/1735/813796/neco.1997.9.8.1735.pdf (accessed on 19 January 2022). [CrossRef]
  29. Clevert, D.A.; Unterthiner, T.; Hochreiter, S. Fast and accurate deep network learning by exponential linear units (elus). arXiv 2015, arXiv:1511.07289. [Google Scholar]
  30. Garofalo, J.; Graff, D.; Paul, D.; Pallett, D. Csr-i (wsj0) Complete. Linguistic Data Consortium, Philadelphia. 2007. Available online: https://catalog.ldc.upenn.edu/LDC93S6A (accessed on 17 February 2022).
  31. Stowell, D.; Plumbley, M.D. An Open Dataset for Research on Audio Field Recording Archives: Freefield1010. arXiv 2013, arXiv:1309.5275. Available online: http://xxx.lanl.gov/abs/1309.5275 (accessed on 28 May 2021).
  32. Reddy, C.K.; Dubey, H.; Koishida, K.; Nair, A.; Gopal, V.; Cutler, R.; Braun, S.; Gamper, H.; Aichner, R.; Srinivasan, S. INTERSPEECH 2021 Deep Noise Suppression Challenge. In Proceedings of the INTERSPEECH 2021, Brno, Czech Republic, 30 August–3 September 2021. [Google Scholar]
  33. Freesound. Available online: https://freesound.org/ (accessed on 28 September 2019).
  34. Varga, A.; Steeneken, H.J. Assessment for automatic speech recognition: II. NOISEX-92: A database and an experiment to study the effect of additive noise on speech recognition systems. Speech Commun. 1993, 12, 247–251. [Google Scholar] [CrossRef]
  35. Mesaros, A.; Diment, A.; Elizalde, B.; Heittola, T.; Vincent, E.; Raj, B.; Virtanen, T. Sound event detection in the DCASE 2017 Challenge. IEEE/ACM Trans. Audio Speech Lang. Process. (TASLP) 2019, 27, 992–1006. [Google Scholar] [CrossRef] [Green Version]
  36. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. The general architecture of multi-pass speech enhancement. The ⊕ symbol represents element-wise addition operation.
Figure 1. The general architecture of multi-pass speech enhancement. The ⊕ symbol represents element-wise addition operation.
Sensors 22 02440 g001
Figure 2. Input, Base, and Output layers in the three neural network architectures tested with multi-pass speech enhancement.
Figure 2. Input, Base, and Output layers in the three neural network architectures tested with multi-pass speech enhancement.
Sensors 22 02440 g002
Figure 3. Dependence of SI-SDR on number of passes.
Figure 3. Dependence of SI-SDR on number of passes.
Sensors 22 02440 g003
Figure 4. Dependence of STOI on number of passes.
Figure 4. Dependence of STOI on number of passes.
Sensors 22 02440 g004
Figure 5. Example of multi-forward speech enhancement (Transformer-Net, babble noise at 0 dB SNR). Warmer colors indicate higher energy in given time frame and frequency band.
Figure 5. Example of multi-forward speech enhancement (Transformer-Net, babble noise at 0 dB SNR). Warmer colors indicate higher energy in given time frame and frequency band.
Sensors 22 02440 g005
Table 1. Babble noise (TIMIT). Best performances in each category are marked with bold font.
Table 1. Babble noise (TIMIT). Best performances in each category are marked with bold font.
MetricLOSI-SDR (dB)STOIPESQ
SNR −5 dB0 dB5 dB10 dB−5 dB0 dB5 dB10 dB−5 dB0 dB5 dB10 dB
original--−5.04−0.044.969.970.3750.5550.7160.8331.671.922.232.55
Dilated111.827.2011.2113.700.5750.7750.8940.9481.902.412.803.10
U-Net511.416.8510.8412.890.5590.7590.8820.9401.852.352.753.05
22.517.8211.7514.440.6120.8120.9170.9591.972.482.863.17
33.018.2111.9914.800.6440.8400.9300.9642.032.542.923.21
43.268.3512.0614.910.6580.8520.9340.9662.062.572.943.22
53.378.4012.0614.950.6670.8570.9350.9662.072.582.953.22
Res112.166.8510.5313.300.6340.8300.9190.9561.982.502.903.21
BLSTM512.127.0610.8113.600.6230.8180.9190.9591.932.452.883.21
22.927.6511.2814.090.6680.8580.9340.9632.052.582.983.28
33.157.7811.3714.140.6860.8670.9360.9642.072.603.003.29
43.227.8211.3914.130.6920.8700.9370.9642.082.613.003.29
53.257.8311.3914.120.6960.8710.9370.9642.082.603.003.28
Trans-112.067.2711.3514.370.5980.7980.9090.9571.872.382.793.12
former511.637.0211.1313.870.5670.7680.8880.9461.872.372.783.10
-Net 22.828.0912.0815.240.6250.8240.9220.9622.012.522.923.23
33.498.4512.3415.550.6560.8440.9320.9662.092.602.993.28
43.718.5812.4115.620.6660.8550.9370.9682.132.633.023.30
53.848.6312.4315.650.6760.8610.9390.9692.132.633.023.30
Table 2. Shopping mall noise (DCASE). Best performances in each category are marked with bold font.
Table 2. Shopping mall noise (DCASE). Best performances in each category are marked with bold font.
MetricLOSI-SDR (dB)STOIPESQ
SNR −5 dB0 dB5 dB10 dB−5 dB0 dB5 dB10 dB−5 dB0 dB5 dB10 dB
original--−5.31−0.314.689.530.4710.6380.7760.8791.702.022.362.74
Dilated115.659.6213.0015.940.7490.8820.9440.9702.322.703.013.28
U-Net515.319.3412.7215.520.7350.8710.9390.9682.202.582.913.21
26.129.9513.2816.320.7880.8990.9500.9742.292.662.983.26
36.4110.1313.4416.510.8120.9080.9530.9752.342.703.013.28
46.5310.1913.4816.550.8200.9100.9540.9762.362.713.013.28
56.5710.2113.4816.560.8220.9110.9540.9762.362.713.013.28
Res115.368.9912.1815.020.7780.8900.9420.9682.272.632.953.23
BLSTM515.329.0812.3115.110.7610.8850.9420.9692.292.622.923.20
25.869.4812.6515.490.8020.9020.9480.9712.392.712.993.24
36.019.5712.7115.560.8120.9050.9490.9712.402.722.993.24
46.039.5912.7215.560.8150.9060.9490.9712.402.722.993.24
56.039.5912.7215.570.8160.9060.9490.9712.402.712.983.23
Trans-115.429.4612.9416.050.7500.8810.9430.9712.162.562.923.24
former514.969.2412.7815.750.7140.8610.9360.9692.162.552.923.25
-Net 26.069.9813.4216.590.7770.8950.9500.9752.302.673.013.31
36.4210.1913.5516.720.7980.9040.9530.9762.362.723.043.32
46.5510.2513.5816.740.8040.9070.9530.9762.372.733.053.33
56.5910.2613.6016.750.8060.9070.9540.9762.382.733.053.33
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Grzywalski, T.; Drgas, S. Speech Enhancement by Multiple Propagation through the Same Neural Network. Sensors 2022, 22, 2440. https://doi.org/10.3390/s22072440

AMA Style

Grzywalski T, Drgas S. Speech Enhancement by Multiple Propagation through the Same Neural Network. Sensors. 2022; 22(7):2440. https://doi.org/10.3390/s22072440

Chicago/Turabian Style

Grzywalski, Tomasz, and Szymon Drgas. 2022. "Speech Enhancement by Multiple Propagation through the Same Neural Network" Sensors 22, no. 7: 2440. https://doi.org/10.3390/s22072440

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop