Next Article in Journal
In Vitro and In Silico Screening of 2,4,5-Trisubstituted Imidazole Derivatives as Potential Xanthine Oxidase and Acetylcholinesterase Inhibitors, Antioxidant, and Antiproliferative Agents
Next Article in Special Issue
A Basic Study on Sound Control System for Ultra-Compact Electric Vehicle by Using Masking
Previous Article in Journal
Performance of the OncomineTM Lung cfDNA Assay for Liquid Biopsy by NGS of NSCLC Patients in Routine Laboratory Practice
Previous Article in Special Issue
A Unified Speech Enhancement System Based on Neural Beamforming With Parabolic Reflector
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Supervised Speech Enhancement Approach with Residual Noise Control for Voice Communication

1
Key Laboratory of Noise and Vibration Research, Institute of Acoustics, Chinese Academy of Sciences, Beijing 100190, China
2
University of Chinese Academy of Sciences, Beijing 100049, China
*
Author to whom correspondence should be addressed.
Appl. Sci. 2020, 10(8), 2894; https://doi.org/10.3390/app10082894
Submission received: 20 March 2020 / Revised: 16 April 2020 / Accepted: 17 April 2020 / Published: 22 April 2020
(This article belongs to the Collection Recent Applications of Active and Passive Noise Control)

Abstract

:
For voice communication, it is important to extract the speech from its noisy version without introducing unnaturally artificial noise. By studying the subband mean-squared error (MSE) of the speech for unsupervised speech enhancement approaches and revealing its relationship with the existing loss function for supervised approaches, this paper derives a generalized loss function that takes residual noise control into account with a supervised approach. Our generalized loss function contains the well-known MSE loss function and many other often-used loss functions as special cases. Compared with traditional loss functions, our generalized loss function is more flexible to make a good trade-off between speech distortion and noise reduction. This is because a group of well-studied noise shaping schemes can be introduced to control residual noise for practical applications. Objective and subjective test results verify the importance of residual noise control for the supervised speech enhancement approach.

1. Introduction

Speech enhancement plays an important role in noisy environments for many applications, such as speech communication, speech interaction and speech translation. Numerous researchers have spent much effort on separating the speech from its noisy version and various approaches have already been proposed in the last five decades. Conventional approaches include spectral subtraction [1], statistical method [2,3] and subspace-based method [4], which has proved to be valid when the additive noise is stationary or quasi-stationary. However, their performance often suffers from heavy degradation under non-stationary and low signal-to-noise ratio (SNR) conditions.
Recently, supervised deep learning approaches have shown their powerful capability on suppressing both stationary and highly non-stationary noise signals, which is mainly because of the highly nonlinear mapping ability of deep neural networks (DNNs) [5,6,7,8,9]. In DNN-based algorithms, minimum mean-squared error (MMSE) is often adopted as a loss criterion to update the weights of the network. Nevertheless, the usage of this criterion directly may suffer from some problems. First, although MSE is the most popular and well-known criterion, it is based on the assumption, i.e., each time-frequency (T-F) bin bears the same importance during the training phase. However, the auditory system has different sensitivity towards different frequency regions and the perception knowledge should be also taken into account [10,11]. Second, the speech spectrogram has unbalanced distribution, i.e., the formants often exist in the low- and middle-frequency regions while they are sparse in the high-frequency regions. Therefore, global MSE optimization usually obtains an over-smoothing estimation which omits some important detailed information. To solve these problems, many new criteria, that consider speech perception, have been proposed in recent years [12,13,14,15]. The first one is to use perceptually weighted MSE functions, which are proposed to weight the loss in different T-F regions [13,16]. Despite the weighted coefficients mitigate the over-smoothing issue of MSE to some extent, most of them are based on the heuristic principles and cannot solve the inherent over-smoothing problem of MSE completely. The second one is to use objective metrics as loss functions. For examples, perceptual evaluation speech quality (PESQ) [17], short-time objective intelligibility (STOI) [18] and scale-invariant speech distortion ratio (SI-SDR) [19] have been adopted as loss functions. Although the metric-based methods facilitate the better optimization of the specific objective metric, other metrics are often non-optimized. Moreover, the calculation of some metrics is often too complicated and non-continuous [20]. In [21], speech distortion and residual noise are considered separately in the loss function, known as the components loss (CL), which obtains relatively better metric scores than MSE when suitable loss-weighted coefficients are selected.
Note that all the above mentioned loss functions aim at suppressing noise as much as possible at noise-only segments. In other words, at noise-only segments, the amount of noise reduction is expected to be a positive infinite value. However, this aim could not be achieved in most cases for many reasons. First, the noise is often stochastic, and thus it is inevitable that the estimation accuracy is often constrained by a limited number of available observations [22,23]. Second, there are a great variety of noise signals, so that a DNN model cannot be expected to distinguish all of them correctly from the speech in each T-F unit. Therefore, when the noise cannot be suppressed totally as expected, some unnatural residual noise may severely degrade speech quality [24], which needs to be considered carefully. In this paper, we derive a generalized loss function by introducing multiple manual parameters to flexibly make a balance between speech distortion and noise attenuation. More specifically, we use the residual noise control introduced for voice communication [25,26]. By theoretical derivations, MSE and some other often-used loss functions can be included in the proposed generalized loss function.
The remainder of the paper is structured as follows. Section 2 formulates the problem. Section 3 derives the generalized loss in detail and introduces used network architecture. Section 4 is the experimental settings. Results and analysis are given in Section 5. Section 6 presents some conclusions.

2. Problem Formulation

In the time domain, the noisy signal can be modelled as
x n = s n + d n ,
where s n is the clean speech and d ( n ) is the additive noise. In the frequency domain, (1) can be written as
X l k = S l k + D l k ,
where X l k , S l k , and D l k are, respectively, discrete Fourier transforms (DFT) of x ( n ) , s ( n ) , and d ( n ) with the frame index l and the frequency bin k.
For practical applications, we only have the time-domain noisy signal x ( n ) or its frequency-domain version X l ( k ) , the problem becomes how to estimate s ( n ) or S l ( k ) from its noisy signal. It is common to use MMSE as a criterion in unsupervised speech enhancement approaches. Before introducing MMSE, we first define the square error as
J x M l k = f S l k g S l k , D l k , M l k 2 ,
where M l k is a nonlinear spectral gain function, f ( a ) and g ( a , b , c ) are the transform functions. When f ( a ) = | a | and g ( a , b , c ) = | ( a + b ) c | , min M l k E J x M l k results in MMSE spectral amplitude estimator in [2], where E is the expectation operator. When f ( a ) = log ( | a | ) and g ( a , b , c ) = log ( | ( a + b ) c | ) , min M l k E J x M l k leads to MMSE log-spectral amplitude estimator in [3]. More complicated forms of f ( a ) and g ( a , b , c ) can be chosen; for example, many perceptually-weighted error criteria can be included, which can be referred to [10].
For supervised approaches, the square error in the subband is often defined as the loss function in the fullband, which is
J x = k K l L J x M l k .
One can get that, when f ( a ) = log ( | a | ) and g ( a , b , c ) = log ( | ( a + b ) c | ) , min M l k J x is to minimize the MSE of log-spectral amplitude between the clean speech and the estimated speech, which is the training target in [6].
Note that (3) and (4) are quite similar and the most obvious difference between them is that J x M l k is the subband square error, while J x is the fullband square error. The other difference is that the nonlinear spectral gain can be derived theoretically by minimizing E J x M l k when the probability density function (p.d.f.) of the speech and that of the noise are both given, while it is difficult to derive the nonlinear spectral gain by minimizing J x , where this gain can often be mapped from the input noisy features after training the supervised machine learning model. In all, it seems that all subband square error functions can be generalized to the fullband ones as supervised training targets.
In the above formulation, MMSE is utilized to optimize the speech spectrum recovery in the T-F domain. However, due to the stochastic characteristic of noise components and the performance limit of the network, the residual noise tends to be unnatural and may severely degrade the speech quality. Moreover, noise suppression and speech distortion are not separately considered in an explicit way, which motivates us to reformulate the optimization towards the trade-off between noise suppression and speech distortion and derive a new type of generalized loss function in Section 3.

3. Proposed Algorithm

Only using MMSE as a criterion, it is difficult to make a balance between speech distortion and noise reduction. This section derives a more generalized fullband loss function.

3.1. Trade-Off Criterion in Subband

In traditional speech enhancement approaches, speech distortion and noise reduction in the subband can be considered separately. The subband square error of the speech and the subband residual noise can be, respectively, given by
J s M l k = f S l k g S l k , D l k , M l k 2 ,
and
J d M l k = h S l k , D l k , M l k 2 ,
where h ( a , b , c ) is a transform function. When f ( a ) = | a | , g ( a , b , c ) = | a c | , and h ( a , b , c ) = | b c | , E J s M l k and E J d M l k become the MSE of the speech magnitude and the residual noise power in the subband, respectively, which are identical with [27] (8.31) and (8.32).
By minimizing the subband MSE of the speech with a residual noise control, an optimization problem can be given to derive the nonlinear spectral gain, which is given by
min M l k E J s M l k , s . t . E J d M l k = ƛ β l k , D l k 2 ,
where ƛ β , b is a transform function. β l k [ 0 1 ] could be both a frequency and frame-dependent factor that can be introduced to control the residual noise flexibly.
The optimal spectral gain in (7) can be solved theoretically by the Lagrange multiplier method, which is
min M l k E J s M l k + μ E J d M l k μ ƛ β l k , D l k 2 ,
where μ 0 is a Lagrange multiplier. When f ( a ) = | a | , g ( a , b , c ) = | a c | , h ( a , b , c ) = | b c | , and ƛ β , b = β E { | b | 2 } , the optimal spectral gain can be derived from (8) and the constraint in (7), which can be given by
M l k = ξ l k ξ l k ξ l k + μ l k ξ l k + μ l k ,
where ξ l k = E { | S l k | 2 } / E { | D l k | 2 } is the a priori SNR. It is not always possible to derive M l k mathematically, especially when f ( a ) , g ( a , b , c ) , h ( a , b , c ) , and ƛ β , b have very complicated expressions. Moreover, it is difficult to accurately estimate the noise power spectral density in non-stationary noise environments [28,29,30]. However, it seems that this optimization can be easily solved by supervised approaches. To transfer this problem, we need to define the fullband square error of the speech and the fullband residual noise power to derive the loss function for supervised approaches.

3.2. Trade-Off Criterion in Fullband

The fullband MSE of the speech and the fullband residual noise can be, respectively, given by
J s = k = K l = L J s M l k ,
and
J d = k = K l = L J d M l k .
The loss function without any constraints can be given by
J x = J s + μ J d ,
where (12) is the same as the newly proposed components loss function as given in [21].
The loss function with residual noise control is
J x = J s + μ J d con ,
where
J d con = k = K l = L J d M l k ƛ β l k , D l k 2 .
It is obvious that (13) is a generalization of (12), where (13) reduces to (12) when ƛ β l k , D l k 2 0 . One can observe that β l k is both frequency and frame-dependent, so it can control the residual noise in each time-frequency bin.

3.3. A Generalized Loss Function

We further generalize the subband square error in (5) and (6), the square is substituted by a variable γ 0 and an additional variable α is also introduced on the spectra, then (5) and (6) can be, respectively, given by
J s γ , α M l k = f S l α k g S l α k , X l α k , M l α k γ ,
and
J d γ , α M l k = h S l α k , D l α k , M l α k γ .
Analogously, with the residual noise control, the optimization problem in the subband becomes
min M l k E J s γ , α M l k , s . t . E J d γ , α M l k = ƛ β l α k , D l α k γ .
By setting f ( a ) = | a | , g ( a , b , c ) = | a c | , h ( a , b , c ) = | b c | , and ƛ β , b = ( β | b | ) , one can derive a generalized gain function with the Lagrange multiplier method, which is
M l k = ξ l k c 1 μ l k 2 c 1 c 2 1 + ξ l k c 1 c 2 ,
where c 1 = α γ / ( 2 γ 2 ) and c 2 = 1 / α , where (17) is identical to [31] (6). Note that [31] (6) is given intuitively without theoretical derivation. When γ = 2 and α = 1 , (17) reduces to (9). When γ = 2 , one can get M l k = ξ l k α ξ l k α μ l k + ξ l k α μ l k + ξ l k α 1 / α , which has already been derived and presented in ([31] (22)).
Similarly, the generalized loss function for supervised approaches can be given by
J x γ , α = J s γ , α + μ J d γ , α , con ,
where the first item J s γ , α = k = K l = L J s γ , α M l k relates to the fullband speech distortion and the second item J d γ , α , con = k = K l = L J d γ , α M l k ƛ β l α k , D l α k γ is introduced to control the residual noise.
Equation (18) is a generalized loss function that includes (12) and (13). This is because (18) reduces to (13) when γ = 2 , α = 1 and it can further reduce to (12) by setting ƛ β l α k , D l α k γ 0 . It is interesting to see that (3) also can be separated into two components, where one is the MSE of the speech and the other is related to the residual noise. When f ( a ) = a and g ( a , b , c ) = ( a + b ) c , we have
E J x M l k = E J s M l k + E J d M l k ,
where E J s M l k = 1 M l k 2 E S l k 2 relates to the power of speech distortion and E J d M l k = M l k 2 E D l k 2 relates to the power of residual noise. E J x M l k is a combination of speech distortion and residual noise, so the fullband MSE loss function of a complex spectrum is also a special case of the generalized loss function in (18). If f ( a ) = | a | and g ( a , b , c ) = | ( a + b ) c | are chosen, the decomposition of E J x M l k is more complicated than (19), which will not be further discussed for limited space.
In this paper, we emphasize the importance of introducing the residual noise control. f ( a ) = | a | , g ( a , b , c ) = | a c | , h ( a , b , c ) = | b c | , and ƛ β , b = ( β | b | ) are applied, although more complicated expressions can be chosen when taking the perceptual quality into account. Accordingly, we have
J s γ , α = l = L k = K 1 M l α k S l α k γ ,
and
J d γ , α , c o n = l = L k = K M l k D l k α γ β l k D l k α γ ,
where both α and β l k are constant values over frequency for simplicity, that is to say, β l k β 0 and α = α 0 are used in the following. α 0 is set to 1 in the major experiments and we will also separately analyze the role of α . We study the impact of β 0 , μ , γ and α 0 on supervised approaches.

4. Experimental Setup

4.1. Dataset

Experiments are conducted with TIMIT corpus, where 1000 and 200 clean utterances are randomly chosen as the training and the validation datasets, respectively. In total, 125 types of environment noises [6,32] are used for generating noisy utterances under different SNR levels ranging from −5 dB to 15 dB with the interval 5 dB. During each mixing process, a clean utterance is mixed with two types of noises and 5 SNR levels. As a consequence, 10,000 ( 1000 × 5 × 2 ) , 2000 ( 200 × 5 × 2 ) noisy-clean pairs are established for training and validation, respectively. For model test, additional 10 male and 10 female utterances are chosen to mix with 5 types of unseen noises taken from the NOISEX92 [33] (babble, factory1, hfchannel, pink, and white) with SNR ranging from −5 dB to 10 dB with the interval 5 dB.

4.2. Experimental Settings

We sample all the utterances at 16 kHz, which are subsequently enframed by a 20-ms Hamming window and 10-ms overlap between adjacent frames. A 320-point short-time Fourier transform (STFT) is applied to transform the frames into the T-F domain, leading to 161-point spectral feature vectors. The magnitude of the spectrum is deployed as the input feature. The models are trained with stochastic gradient descent (SGD) optimized by Adam [34]. The learning rate is initialized at 0.0005. We halve the learning rate only when three consecutive validation loss increment arises and the training process is early-stopped unless ten consecutive validation loss increment happens. Totally 100 epochs are trained to guarantee the network convergence. Within each epoch, the minibatch is set to 16 at the utterance level, where all the utterances are zero-padded to have the same timestep with the longest utterance.

4.3. Network Architecture

U-Net is chosen as the network in this paper, which has been widely adopted for the speech separation task [35]. As shown in Figure 1, the network consists of the convolutional encoder and decoder, both of which are comprised of five convolutional blocks where the 2-D convolution layer is adopted, followed by batch normalization (BN) [36] and exponential linear unit (ELU) [37]. Within each convolutional block, the kernel size is set to ( 2 , 3 ) along the temporal and frequency axis. Skip connections are introduced to compensate for the information loss during the features compression process. Note that the mapping target is the gain function and the sigmoid function is adopted to make sure that the output ranges from 0 to 1. A causal mechanism is used to achieve real-time processing, where only the past frames are involved in the convolution calculation. The tensor output size of each layer is given with ( C h a n n e l s , T i m e S t e p , F e a t ) format, which is shown in Figure 1. A more detailed description of the network can refer to Table 1. The total number of trainable parameters of the network is 0.59 M.

4.4. Loss Functions and Training Models

This paper chooses three loss functions including MSE in (4), Time-MSE-based loss (TMSE) [35] and recently proposed SI-SDR-based loss [35] as baselines. As a T-F domain-based network is used, an additional iSTFT layer is needed to transform the estimated T-F spectrum back into time domain for TMSE- and SI-SDR-based loss [38]. The iSTFT layer is a type of specific deconvolutional layer, whose basis function corresponds to the iSTFT coefficient matrix. The baselines are compare with the proposed generalized loss function given in (18) with (20) and (21).

5. Results and Analysis

This paper uses four objective measurements to analyze the performance of proposed generalized loss, including noise attenuation (NA) [25], speech attenuation (SA) [25], PESQ [17], and SDR [39].

5.1. The Impact of γ , β 0 and μ

The testing results w.r.t. γ , β 0 and μ are shown in Figure 2, where γ = 1 , 2 , 3 , β 0 = 10 dB , 20 dB , 30 dB and μ = 0.5 , 1 , 2 , 3 , 4 are considered. Here α 0 is set to 1 for all the conditions. The test results of three baselines are also presented as the comparison. From this figure, one can observe the following phenomena. First, the increase of β 0 will decrease NA. This is because the residual noise control mechanism is introduced for optimization, which means, during the training process, the residual noise in the estimated spectra will gradually get close to the preset residual noise threshold. As a consequence, the characteristic of the residual noise is expected to be effectively preserved, which will be further confirmed by subjective listening tests in the following. Second, the increase of μ is beneficial to noise suppression and meanwhile introducing more speech distortion. As generalized loss can be viewed as the joint optimization of both speech distortion and noise reduction, a larger μ shows that more emphasis is given on noise suppression and it leads to smaller gain values, as (17) states, where on the one hand more interference is suppressed and on the other hand, more speech components are inevitably abandoned. Third, the increase of γ has a negative influence on NA and SA. In addition, when γ is set to 2, it shows a better objective speech quality than γ = 1 , 3 .
According to the above results, we have some general guidelines to choose the three parameters including γ , β 0 and μ . First, NA is expected to be as large as possible while SA needs to be as small as possible, indicating more noise components can be suppressed with less speech distortion. Second, both PESQ and SDR need to be as large as possible, indicating better speech quality is achieved. Considering the effects of different parameters illustrated in the last paragraph, among various parameter configurations with γ , β 0 , μ format, (2, −30 dB, 0.5), (2, −30 dB, 1) and (2, −20 dB, 1) can be chosen for practical applications. This is because relatively better performance can be achieved for four objective metrics. One can observe that the three competing loss functions can get better performance in some objective metrics, while they may suffer much worse performance in others. For example, SI-SDR has the largest value of SDR, while its PESQ score is even lower than the MSE, which is consistent with the study in [19].

5.2. The Impact of α

To analyze the impact of α , we select one type of parameter configuration, with γ = 2 , β 0 = 20 dB, and μ = 1 , which has shown the best performance among various configurations. The value of α ranges from 1 to 2 with the interval 0.1. The reason for choosing α 1 is to avoid the gradient value infinite problem during the back-propagation process for α < 1 . The metric results w.r.t. α are given in Figure 3. One can observe the following phenomena. First, the increase of α will decrease both NA and SA values. This is because the estimated gain in each T-F bin ranges from 0 to 1, and the increase of α will lead to a larger value, cf. (17). As a consequence, the network tends to attenuate less background interference and preserve more speech components. Second, both PESQ and SDR tend to decline with the increase of α , which can be explained as more residual noise components are preserved, and they will heavily degrade the speech quality although the speech distortion is reduced.

5.3. Subjective Evaluation

To evaluate speech quality of the proposed generalized loss (GL) function, a subjective evaluation test is conducted among GL and baselines, where we follow the subjective testing procedures of [40]. In this comparison, we choose the parameter configuration (2, −20 dB, 1) and α 0 = 1 for the proposed GL function. The experiment is conducted in a standard listening room with the size (5 m × 4 m × 3 m), where 10 listeners participate. The listening material consists of 20 utterances, each of which includes one male and female utterance selected from TIMIT corpus and is mixed with one of five noises including aircraft, babble, bus, cafeteria, and car. Four SNR conditions are selected for mixing, i.e., −5 dB, 0 dB, 5 dB, 10 dB. A speech pause of 3s duration is specifically inserted before each utterance. Then, the duration of each listening utterance is about 13s. Each listener needs to write down the utterance index that they prefer considering both noise naturalness and speech quality. The same as [40], “Equal” option is also provided if no subjective preference can be given. To overcome inertia, the utterance index in each pair is shuffled. The averaged subjective results are presented in Table 2. From this table, one can observe that the proposed GL function with residual noise control achieves better performance in subjective testing, which can be explained as the proposed GL method can effectively recover speech components while preserving the characteristic of background noise to some extent compared with all the baselines.

6. Conclusions

This paper derives a generalized loss function which can easily make a balance between noise attenuation and speech distortion with multiple manual parameters. In addition, MSE and other typical loss functions are revealed to be special cases. Both objective and subjective tests are conducted to show that it is important to control the residual noise for supervised speech enhancement approaches, where the residual noise becomes much more natural than before. Moreover, compared with other competitive loss functions, the proposed loss function obtains comparable performance in objective metrics and much better subjective evaluation results when suitable parameter configurations are selected. Further work could concentrate on studying a combination of the residual noise control scheme with objective metrics-based loss functions to improve the naturalness of the residual noise.

Author Contributions

Conceptualization and methodology, A.L. and C.Z.; software and validation, A.L. and R.P.; writing—original draft preparation: A.L.; writing—review and editing, C.Z. and X.L.; supervision: X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Science Foundation of China under Grant No. 61571435, No. 61801468, and No. 11974086.

Conflicts of Interest

The authors declare no conflict of interest.

Acronyms and Abbreviations

SNRsignal-to-noise ratio
MMSEminimum mean-squared error
MSEmean-squared error
DNNdeep neural network
PESQperceptual evaluation speech quality
STOIshort-time objective intelligibility
CLcomponent loss
GLgeneralized loss
SI-SDRscale-invariant speech distortion ratio
T-Ftime-frequency
DFTdiscrete Fourier transform
SGDstochastic gradient descent
STFTshort-time Fourier transform
BNbatch normalization
ELUexponential linear unit
TMSEtime mean-squared error
iSTFTinverse short-time Fourier transform
NAnoise attenuation
SAspeech attenuation
SDRspeech distortion ratio

References

  1. Boll, S. Suppression of acoustic noise in speech using spectral subtraction. IEEE Trans. Acoust. Speech Signal Process. 1979, 27, 113–120. [Google Scholar] [CrossRef] [Green Version]
  2. Ephraim, Y.; Malah, D. Speech enhancement using a minimum-mean square error short-time spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1984, 32, 1109–1121. [Google Scholar] [CrossRef] [Green Version]
  3. Ephraim, Y.; Malah, D. Speech enhancement using a minimum mean-square error log-spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1985, 33, 443–445. [Google Scholar] [CrossRef]
  4. Ephraim, Y.; Van Trees, H.L. A signal subspace approach for speech enhancement. IEEE Trans. Acoust. Speech Signal Process. 1995, 3, 251–266. [Google Scholar] [CrossRef]
  5. Wang, Y.; Narayanan, A.; Wang, D. On training targets for supervised speech separation. IEEE Trans. Audio Speech Lang. Process. 2014, 22, 1849–1858. [Google Scholar] [CrossRef] [Green Version]
  6. Xu, Y.; Du, J.; Dai, L.-R.; Lee, C.-H. An experimental study on speech enhancement based on deep neural networks. IEEE Signal Process. Lett. 2013, 21, 65–68. [Google Scholar] [CrossRef]
  7. Li, A.; Yuan, M.; Zheng, C.; Li, X. Speech enhancement using progressive learning-based convolutional recurrent neural network. Appl. Acoust. 2020, 166, 107347. [Google Scholar] [CrossRef]
  8. Zhang, L.; Wang, M.; Zhang, Q.; Liu, M. Environmental Attention-Guided Branchy Neural Network for Speech Enhancement. Appl. Sci. 2020, 10, 1167. [Google Scholar] [CrossRef] [Green Version]
  9. Wu, J.; Hua, Y.; Yang, S.; Qin, H.; Qin, H. Speech Enhancement Using Generative Adversarial Network by Distilling Knowledge from Statistical Method. Appl. Sci. 2019, 9, 3396. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, Y.; Loizou, P.C. A perceptually motivated approach for speech enhancement. IEEE Trans. Audio Speech Lang. Process. 2003, 11, 457–465. [Google Scholar] [CrossRef]
  11. Loizou, P.C.; Kim, G. Reasons why current speech-enhancement algorithms do not improve speech intelligibility and suggested solutions. IEEE Trans. Audio Speech Lang. Process. 2010, 19, 47–56. [Google Scholar] [CrossRef] [Green Version]
  12. Martín-Doñas, J.M.; Gomez, A.M.; Gonzalez, J.A.; Peinado, A.M. A deep learning loss function based on the perceptual evaluation of the speech quality. IEEE Signal Process. Lett. 2018, 25, 1680–1684. [Google Scholar] [CrossRef]
  13. Liu, Q.; Wang, W.; Jackson, P.J.; Tang, Y. A perceptually-weighted deep neural network for monaural speech enhancement in various background noise conditions. In Proceedings of the European Signal Processing Conference (EUSIPCO), Kos, Greece, 28 August–2 September 2017; pp. 1270–1274. [Google Scholar]
  14. Kolbæk, M.; Tan, Z.-H.; Jensen, J. Monaural speech enhancement using deep neural networks by maximizing a short-time objective intelligibility measure. In Proceedings of the IEEE International Conference on Acoustics, Speech, Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5059–5063. [Google Scholar]
  15. Venkataramani, S.; Casebeer, J.; Smaragdis, P. Adaptive front-ends for end-to-end source separation. In Proceedings of the Neural Information Processing Systems (NIPS), Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
  16. Shivakumar, P.G.; Georgiou, P.G. Perception optimized deep denoising autoencoders for speech enhancement. In Proceedings of the Conference of the International Speech Communication Association (INTERSPEECH), San Francisco, CA, USA, 8–12 September 2016; pp. 3743–3747. [Google Scholar]
  17. Rim, A.W.; Beerends, J.G.; Hollier, M.P.; Hekstra, A.P. Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Salt Lake City, UT, USA, 7–11 May 2001; pp. 749–752. [Google Scholar]
  18. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. A short-time objective intelligibility measure for time-frequency weighted noisy speech. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Dallas, TX, USA, 15–19 March 2010; pp. 4214–4217. [Google Scholar]
  19. Le Roux, J.; Wisdom, S.; Erdogan, H.; Hershey, J.R. SDR–half-baked or well done? In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 626–630.
  20. Fu, S.; Wang, T.; Tsao, Y.; Lu, X.; Kawai, H. End-to-end waveform utterance enhancement for direct evaluation metrics optimization by fully convolutional neural networks. IEEE Trans. Audio Speech Lang. Process. 2018, 26, 1570–1584. [Google Scholar] [CrossRef] [Green Version]
  21. Xu, Z.; Elshamy, S.; Zhao, Z.; Fingscheidt, T. Components loss for neural networks in mask-based speech enhancement. arXiv 2019, arXiv:1908.05087. [Google Scholar]
  22. Zheng, C.; Zhou, Y.; Hu, X.; Li, X. Two-channel post-filtering based on adaptive smoothing and noise properties. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 1745–1748. [Google Scholar]
  23. Zheng, C.; Liu, H.; Peng, R.; Li, X. A statistical analysis of two-channel post-filter estimators in isotropic noise fields. IEEE Trans. Audio Speech Lang. Process. 2012, 21, 336–342. [Google Scholar] [CrossRef]
  24. Gelderblom, F.B.; Tronstad, T.V.; Viggen, E.M. Subjective evaluation of a noise-reduced training target for deep neural network-based speech enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 27, 583–594. [Google Scholar] [CrossRef]
  25. Gustafsson, S.; Jax, P.; Vary, P. A novel psychoacoustically motivated audio enhancement algorithm preserving background noise characteristics. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seattle, WA, USA, 12–15 May 1998; pp. 397–400. [Google Scholar]
  26. Braun, S.; Kowalczyk, K.; Habets, E.A. Residual noise control using a parametric multichannel wiener filter. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), New Orleans, LA, USA, 5–9 March 2015; pp. 360–364. [Google Scholar]
  27. Benesty, J.; Chen, J.; Huang, Y.; Cohen, I. Noise Reduction in Speech Processing; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  28. Martin, R. Noise power spectral density estimation based on optimal smoothing and minimum statistics. IEEE Trans. Speech Audio Process. 2001, 9, 504–512. [Google Scholar] [CrossRef] [Green Version]
  29. Cohen, I. Noise spectrum estimation in adverse environments: Improved minima controlled recursive averaging. IEEE Trans. Speech Audio Process. 2003, 11, 466–475. [Google Scholar] [CrossRef] [Green Version]
  30. Gerkmann, T.; Hendriks, R.C. Unbiased mmse-based noise power estimation with low complexity and low tracking delay. IEEE Trans. Audio Speech Lang. Process. 2011, 20, 1383–1393. [Google Scholar] [CrossRef]
  31. Inoue, T.; Saruwatari, H.; Shikano, K.; Kondo, K. Theoretical analysis of musical noise in wiener filtering family via higher-order statistics. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing (ICASSP), Prague, Czech Republic, 22–27 May 2011; pp. 5076–5079. [Google Scholar]
  32. Duan, Z.; Mysore, G.J.; Smaragdis, P. Speech enhancement by online non-negative spectrogram decomposition in nonstationary noise environments. In Proceedings of the Conference of the International Speech Communication Association (INTERSPEECH), Portland, OR, USA, 9–13 September 2012; pp. 1–4. [Google Scholar]
  33. Varga, A.; Steeneken, H.J. Assessment for automatic speech recognition: Ii. noisex-92: A database and an experiment to study the effect of additive noise on speech recognition systems. Speech Commun. 1993, 12, 247–251. [Google Scholar] [CrossRef]
  34. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  35. Kolbæk, M.; Tan, Z.-H.; Jensen, S.H.; Jensen, J. On loss functions for supervised monaural time-domain speech enhancement. arXiv 2019, arXiv:1909.01019. [Google Scholar] [CrossRef] [Green Version]
  36. Ioffe, S.; Szegedy, C. Batch normalization: Accelerating deep network training by reducing internal covariate shift. In Proceedings of the International Conference on Machine Learning (ICML), Lille, France, 6–11 July 2015; pp. 448–456. [Google Scholar]
  37. Clevert, D.-A.; Unterthiner, T.; Hochreiter, S. Fast and accurate deep network learning by exponential linear units (elus). arXiv 2015, arXiv:1511.07289. [Google Scholar]
  38. Wichern, G.; Roux, J.L. Phase reconstruction with learned time-frequency representations for single-channel speech separation. In Proceedings of the International Workshop on Acoustic Signal Enhancement (IWAENC), Tokyo, Japan, 17–20 September 2018; pp. 396–400. [Google Scholar]
  39. Vincent, E.; Gribonval, R.; Févotte, C. Performance measurement in blind audio source separation. IEEE/ACM Trans. Audio Speech Lang. Process. 2006, 14, 1462–1469. [Google Scholar] [CrossRef] [Green Version]
  40. Breithaupt, C.; Gerkmann, T.; Martin, R. Cepstral smoothing of spectral filter gains for speech enhancement without musical noise. IEEE Signal Process. Lett. 2007, 14, 1036–1039. [Google Scholar] [CrossRef]
Figure 1. The network architecture adopted in this study. Input is the noisy magnitude spectra and output is the estimated gain functions. T refers to the timestep length of the utterances within a minibatch.
Figure 1. The network architecture adopted in this study. Input is the noisy magnitude spectra and output is the estimated gain functions. T refers to the timestep length of the utterances within a minibatch.
Applsci 10 02894 g001
Figure 2. Test results in terms of NA, SA, PESQ and SDR. The averaged PESQ score of the noisy signals is 1.80 and its averaged SDR is 2.51 dB. Here α 0 is fixed at 1 for all conditions.
Figure 2. Test results in terms of NA, SA, PESQ and SDR. The averaged PESQ score of the noisy signals is 1.80 and its averaged SDR is 2.51 dB. Here α 0 is fixed at 1 for all conditions.
Applsci 10 02894 g002
Figure 3. Metric scores with the increase of α 0 . (a) NA scores with the increase of α 0 . (b) SA scores with the increase of α 0 . (c) PESQ scores with the increase of α 0 . (d) SDR scores with the increase of α 0 .
Figure 3. Metric scores with the increase of α 0 . (a) NA scores with the increase of α 0 . (b) SA scores with the increase of α 0 . (c) PESQ scores with the increase of α 0 . (d) SDR scores with the increase of α 0 .
Applsci 10 02894 g003
Table 1. Detailed description of the netowrk used in the manuscript. The input size and output size of 3-D representation are given in C h a n n e l s , T i m e S t e p , F e a t format. The hyperparameters are specified with K e r n e l , S t r i d e , C h a n n e l s format.
Table 1. Detailed description of the netowrk used in the manuscript. The input size and output size of 3-D representation are given in C h a n n e l s , T i m e S t e p , F e a t format. The hyperparameters are specified with K e r n e l , S t r i d e , C h a n n e l s format.
Layer NameInput SizeHyperparametersOutput Size
reshape_size_1T × 161-1 × T × 161
conv2d_11 × T × 1612 × 3, (1, 2), 1616 × T × 80
conv2d_216 × T × 802 × 3, (1, 2), 3232 × T × 39
conv2d_332 × T × 392 × 3, (1, 2), 6464 × T × 19
conv2d_464 × T × 192 × 3, (1, 2), 128128 × T × 9
conv2d_5128 × T × 92 × 3, (1, 2), 256256 × T × 4
deconv2d_1256 × T × 42 × 3, (1, 2), 128128 × T × 9
skip_1128 × T × 9-256 × T × 9
deconv2d_2256 × T × 92 × 3, (1, 2), 6464 × T × 19
skip_264 × T × 19-128 × T × 19
deconv2d_3128 × T × 192 × 3, (1, 2), 3232 × T × 39
skip_332 × T × 39-64 × T × 39
deconv2d_464 × T × 392 × 3, (1, 2), 1616 × T × 80
skip_416 × T × 80-32 × T × 80
deconv2d_532 × T × 802 × 3,(1, 2), 11 × T × 161
reshape_size_21 × T × 161-T × 161
Table 2. Results of subjective listening test. The numbers indicate the percentage of votes in favor of one approach. The choice “Equal” means no subjective difference.
Table 2. Results of subjective listening test. The numbers indicate the percentage of votes in favor of one approach. The choice “Equal” means no subjective difference.
MethodsGLMSEEqual
Preference70.0%22.0%8.0%
MethodsGLTMSEEqual
Preference66.5%22.0%12.5%
MethodsGLSI-SDREqual
Preference70.5%23.5%6.0%

Share and Cite

MDPI and ACS Style

Li, A.; Peng, R.; Zheng, C.; Li, X. A Supervised Speech Enhancement Approach with Residual Noise Control for Voice Communication. Appl. Sci. 2020, 10, 2894. https://doi.org/10.3390/app10082894

AMA Style

Li A, Peng R, Zheng C, Li X. A Supervised Speech Enhancement Approach with Residual Noise Control for Voice Communication. Applied Sciences. 2020; 10(8):2894. https://doi.org/10.3390/app10082894

Chicago/Turabian Style

Li, Andong, Renhua Peng, Chengshi Zheng, and Xiaodong Li. 2020. "A Supervised Speech Enhancement Approach with Residual Noise Control for Voice Communication" Applied Sciences 10, no. 8: 2894. https://doi.org/10.3390/app10082894

APA Style

Li, A., Peng, R., Zheng, C., & Li, X. (2020). A Supervised Speech Enhancement Approach with Residual Noise Control for Voice Communication. Applied Sciences, 10(8), 2894. https://doi.org/10.3390/app10082894

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop