Next Article in Journal
Deep Learning Based Vehicle Detection and Classification Methodology Using Strain Sensors under Bridge Deck
Next Article in Special Issue
Low-Rank Matrix Recovery from Noise via an MDL Framework-Based Atomic Norm
Previous Article in Journal
In-Vehicle Situation Monitoring for Potential Threats Detection Based on Smartphone Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Real-Time Dual-Microphone Speech Enhancement Algorithm Assisted by Bone Conduction Sensor

1
School of Communication and Information Engineering, Chongqing University of Posts and Telecommunications, Chongqing 400065, China
2
Suresense Technology, Chongqing 400065, China
*
Author to whom correspondence should be addressed.
Sensors 2020, 20(18), 5050; https://doi.org/10.3390/s20185050
Submission received: 17 August 2020 / Revised: 2 September 2020 / Accepted: 3 September 2020 / Published: 5 September 2020
(This article belongs to the Special Issue Signal Processing and Machine Learning for Smart Sensing Applications)

Abstract

:
The quality and intelligibility of the speech are usually impaired by the interference of background noise when using internet voice calls. To solve this problem in the context of wearable smart devices, this paper introduces a dual-microphone, bone-conduction (BC) sensor assisted beamformer and a simple recurrent unit (SRU)-based neural network postfilter for real-time speech enhancement. Assisted by the BC sensor, which is insensitive to the environmental noise compared to the regular air-conduction (AC) microphone, the accurate voice activity detection (VAD) can be obtained from the BC signal and incorporated into the adaptive noise canceller (ANC) and adaptive block matrix (ABM). The SRU-based postfilter consists of a recurrent neural network with a small number of parameters, which improves the computational efficiency. The sub-band signal processing is designed to compress the input features of the neural network, and the scale-invariant signal-to-distortion ratio (SI-SDR) is developed as the loss function to minimize the distortion of the desired speech signal. Experimental results demonstrate that the proposed real-time speech enhancement system provides significant speech sound quality and intelligibility improvements for all noise types and levels when compared with the AC-only beamformer with a postfiltering algorithm.

1. Introduction

In recent years, the signal transmission bandwidth and network technology have been significantly improved, and therefore, communication system can real-time transmit speech signals with a higher sampling rate and deeper sampling bit depth. In speech communication systems, the sound field noise at the mobile communication terminal is the dominant aspect that degrades the communication quality [1]. Speech enhancement is a technology for improving the sound quality and intelligibility of speech signals in the acoustic front end. It uses various speech signal processing approaches in stationary and non-stationary noises, reverberant, point source interference, and diffuse field interference scenarios. Speech enhancement technology can be broadly divided into two categories according to the number of microphones used, i.e., single-channel and multi-channel [2].
Single-channel speech enhancement usually assumes that the noise is additive, and the statistical characteristics of the noise change much more smoothly compared to the speech. The noise is estimated by statistics or minima-controlled methods, and finally the gain function is calculated to complete the noise reduction, for example, spectral subtraction [3], Weiner filter [4], and optimally modified log-spectral amplitude (OMLSA) estimator [5], to name a few. Single-channel speech enhancement algorithms are difficult to effectively remove non-stationary noise and usually introduce additional musical noise [1]. In recent studies, the single-channel speech enhancement technology based on a deep neural network (DNN) has been proven to be effective to suppress transient non-stationary noise [6]. However, DNN-based methods often have a large number of parameters and high computational complexity, making it difficult to be deployed in real-time systems with low power consumption requirements. Recently, Valin [7] proposed a low computational complexity method, recurrent neural network (RNNoise), which combines deep learning and digital signal processing and can process 48 kHz sampling rate audio on a low-power central processing unit (CPU) in real time. However, this lightweight network lacks a powerful for front-end processing method like microphone array to improve the signal-to-noise ratio (SNR) to ensure acceptable performance.
The fundamental difference between single-channel and multi-channel speech enhancement algorithms is that the multi-channel method utilizes the spatial information with multiple microphone sensors. The interferences from undesired directions will be attenuated by the spatial filter composed of the microphone array. Adaptive beamforming is a widely used microphone array speech enhancement algorithm [8], which updates the weight coefficients according to the relationship between the interference and the desired signal in a frame-by-frame manner. Many studies have been conducted to improve the speech quality and intelligibility using microphone arrays and beamforming technology. The minimum variance distortionless response (MVDR) beamformer minimizes the variance of the noise while maintaining the desired signal distortion free. Habets [9] implemented an MVDR beamformer in room acoustics for speech denoising and dereverberation. The least constraint minimum variance (LCMV) [10] method was developed with additional linear constraints, and a variant of LCMV, the generalized sidelobe canceller (GSC), was proposed by Griffiths and Jim [11]. The GSC is composed of three blocks: a fixed beamformer (FBF), a signal-blocking matrix (BM), and an adaptive noise canceller (ANC). FBF is a spatial filter with fixed weight coefficients which retains the desired signal and suppresses the undesired signals. The BM blocks the desired speech components in reference noise signals and ANC eliminates the noise components leaking in the sidelobes of FBF. The work of this paper extends the GSC beamforming structure.
The above discussed methods are based on air-conduction (AC) microphone arrays and showed limited performance in low SNR and highly diffuse noise environments [12]. Recently, the internet of things (IOT) network and devices have developed rapidly. Those technologies have been applied in industry and consumer electronics [13,14,15,16]. As a micro-electro-mechanical system (MEMS) sensor, the bone-conduction (BC) sensor collects vibration signals transmitted from the vocal cords to the bones. Unlike the AC microphone, the BC sensor is relatively insensitive to the ambient acoustic noises and interferences that are conducted by air [17]. However, there is a non-trivial issue that the high-frequency components of the BC signal are seriously attenuated, and the low-frequency components are not exactly the same as those of the AC signal, due to the transmission loss and the sensitivity of the sensor. This problem is particularly prominent for low-cost mass production sensors. Therefore, an additional enhancement of the BC signal is necessary. Several algorithms have been proposed to use the BC signal to aid speech enhancement [17]. In [18], a method was proposed to find a non-linear mapping mechanism between BC and AC signals using a DNN-based autoencoder to enhance the speech signal. The offline training process requires a high sensor quality with less noise and higher sensitivity. In [19], an approach using a bandwidth extension of the BC signal via an adaptive filter for speech enhancement was proposed. Shin [20] proposed an approach using AC and BC signals to estimate a prior SNR. The BC sensor was deployed to assist the noise estimation in [21]. However, using BC signal characteristics to assist microphone array beamforming based on a spatial adaptive filter is rare. It is known that the microphone array beamformer yields a significant improvement in SNR and speech quality [8]. However, the noise suppression is insufficient when the spatial discrimination between noise and desired speech signal is low or the noise field is diffuse [22]. To reduce this kind of residual noise, an additional postfiltering is usually required. The research on the postfiltering algorithm of the microphone array can be divided into two groups. The first is to process the output of the adaptive beamformer with a single channel speech enhancement algorithm [1,23,24]. The second multichannel postfilters utilize the spatial information in the GSC structure and the relationship between the beam and the interference to distinguish between speech and non-stationary noise and thereby obtains a better postfiltering noise suppression capability [25,26,27]. In the non-stationary noise sound field, when the noise is a non-point source or the noise source is in the same direction with the desired signal, the array postfiltering algorithm has difficulty obtaining benefit from the spatial information provided by the beamformer. To combine a real-time simple recurrent unit (SRU)-based neural network noise reduction algorithm with a small number of parameters, this paper treats the array postfiltering as a single-channel noise reduction task.
In this paper, a GSC-based and BC signal-assisted dual-microphone adaptive beamforming algorithm with an SRU-based postfilter was proposed. An improved voice activity detection (VAD) was extracted from the BC, which effectively assists the weight coefficient update of the ANC and the adaptive block matrix (ABM). We designed a filter that was named a compensation filter (CF) in this paper to learn the delay and amplitude difference between the BC and the AC signals. The BC signal is filtered to approximate the clean AC signal and is fused with the output of the ANC in an appropriate ratio to further suppress the noise. To combat the wind noise, a dual-microphone wind noise detector is deployed and the wind noise portion will be replaced by the output of the CF. With the assistance of the BC signal, the noise suppression capability of the dual-microphone GSC system under low SNR or harsh conditions can be improved effectively. Finally, a lightweight real-time single-channel noise reduction system based on SRU was used as a postfilter to eliminate the residual noise in the microphone array output.
The rest of this paper was organized as follows: the proposed dual-microphone array assisted by a BC signal approach was introduced in Section 2. The integration of a BC-assisted microphone array and SRU-based postfilter is presented in Section 3. Experimental details are described in Section 4. Results and discussion are shown in Section 4. Finally, a conclusion of this paper is given in Section 5.

2. Dual-Microphone Array Assisted by BC Signal

2.1. Signal Model

In this work, we considered a frame-by-frame processing of an AC dual-microphone array and BC signals in the frequency domain. The dual-microphone array can be treated as a uniform linear array (ULA), depicted in Figure 1. It is important to estimate the position of the sound source for the microphone array signal processing. The direction of arrival (DOA) was used to obtain the steering vector. There are many methods to locate the direction of the source indoors [28,29,30]. Since the algorithm application scenario proposed in this paper is a wearable device, we assume that the interference source was in the broadside direction and that the desired source was fixed in the end fire direction, along the axis of the array. The observed signal in the time domain at the m-th sensor ( m = 1 , 2 ) is:
y m ( t ) = x ( t ) s m ( t ) + v m ( t )
where denotes convolution. x ( t ) is the desired speech signal, v m ( t ) contains the directional and non-directional noises, and s m ( t ) is the system function that represents the acoustic wave propagation delay and the room impulse response (RIR).
Applying short-time Fourier transform (STFT) to (1) yields its complex spectral time–frequency domain expression as
y ( k , l ) = X ( k , l ) s ( k , l ) + v ( k , l ) = X 1 ( k , l ) d ( k , l ) + v ( k , l )
where:
d ( k , l ) = [ 1 , S 2 ( k , l ) S 1 ( k , l ) ] T
and   v ( k , l ) = [ V 1 ( k , l ) , V 2 ( k , l ) ] T
l and k are the time-frame and frequency-bin indices, respectively, and X 1 ( k , l ) is the desired AC signal collected by the reference microphone in the frequency domain. Since the desired signal impinges on the ULA from the end fire direction, considering the delay-only model, the steering vector d ( k , l ) is:
d ( k ) = [ 1 , e j 2 π k τ ] T
where τ = d / c and d is the distance between two microphones, c is the speed of sound.

2.2. Reliable VAD from BC Signal

2.2.1. BC Signal

Unlike the AC microphone, the BC sensor is relatively insensitive to the ambient acoustic noise and interferences which are conducted by air. Due to the transmission loss and the sensitivity of the sensor, the high-frequency components of the BC signal are seriously attenuated, and the low-frequency components are not exactly the same as those of the AC signal. This problem is particularly prominent for low-cost mass production sensors. In Figure 2, it can be observed that the BC signal suffers from a significant attenuation around 1.5 kHz and above, and the sensor thermal noise is also present.

2.2.2. BC VAD Estimator

In order to obtain an accurate and robust VAD from the noisy BC signal, we used a simple noise estimator. The BC speech that is corrupted by the uncorrelated additive thermal noise is given by
Y BC ( k , l ) = S ( k , l ) + N ( k , l )
To derive the VAD estimator, it is assumed that the thermal noise and the BC speech spectral coefficients have a complex Gaussian distribution [31]. Two hypotheses for VAD are:
H 0 :   speech   absent :   Y = N H 1 :   speech   present :   Y = N + S
The probability density functions (PDF) of the two hypotheses H 0 and H 1 are written as
p ( Y | H 0 ) = k = 0 L 1 1 π λ N ( k ) exp ( | Y ( k ) | 2 λ N ( k ) ) ,
and   p ( Y | H 1 ) = k = 0 L 1 1 π ( λ N ( k ) + λ s ( k ) ) exp ( | Y ( k ) | 2 λ N ( k ) + λ s ( k ) )
where L , λ N ( k ) , and λ s ( k ) , respectively, denote the coefficients’ dimension of the discrete Fourier transform (DFT), the variance of N ( k , l ) and the variance of S ( k , l ) . The likelihood ratio is given by
Λ = p ( Y | H 0 ) p ( Y | H 1 )
For mathematical convenience, the logarithm operation is applied which yields:
log Λ = log p ( Y | H 0 ) p ( Y | H 1 ) = k = 0 L 1 ( log ( ( λ N ( k ) + λ s ( k ) ) exp ( | Y ( k ) | 2 λ N ( k ) ) λ N ( k ) exp ( | Y ( k ) | 2 λ N ( k ) + λ s ( k ) ) ) = k = 0 L 1 log ( 1 1 + ξ ( k ) exp ( γ ( k ) ξ ( k ) 1 + ξ ( k ) ) )
where ξ ( k ) = λ s ( k ) / λ N ( k ) and γ ( k ) = | Y ( k ) | 2 / λ N ( k ) are called a priori and a posteriori SNRs. For a real-time implementation, γ ( k ) can be estimated as γ ( k ) = | Y ( k ) | 2 / λ N ( k ) and the estimate of ξ ( k ) can be treated as ξ ( k ) = max ( 0 , ξ ml ( k ) ) = max ( 0 , γ 1 ) by a limited maximum likelihood (ML) [32]. Therefore, the frame VAD decision rule is given by
f VAD ( l ) = { 1 ,        log Λ η , 0 ,       log Λ < η .
To obtain the noise variance λ N ( k ) , it can be estimated by a simple recursive function as
λ N ( k , l ) = { α 0 λ N ( k , l 1 ) + ( 1 α 0 ) | Y ( k , l ) | 2 ,        f VAD ( l ) = 1 , λ N ( k , l 1 ) ,           f VAD ( l ) = 0 .

2.3. BC VAD-Assisted Adaptive Beamforming

2.3.1. Robust Generalized Sidelobe Canceller

The robust GSC structure [33] is shown in Figure 3. The dual-microphone inputs are received by the fixed beamforming (FBF) that directs the beam towards the end fire desired signal direction, i.e., at 0°. Y FBF ( k , l ) is used as the reference signal of the adaptive block matrix (ABM) to eliminate the desired signal components in BM. In order to suppress only those desired signals and maximize the noise reduction, the adaptive filter coefficients were constrained within DOA-based boundaries. Then, the output of ABM U ABM ( k , l ) was passed to the adaptive noise canceller (ANC) as the reference signal to suppress the components that are correlated to the interference signal in the FBF output Y FBF ( k , l ) .

2.3.2. Adaptive Block Matrix Assisted by BC Signal

The ABM utilizes an adaptive algorithm to suppress the speech portion of the reference microphone. Its output is
U ( k , l )   = X 1 ( k , l ) G ABM H ( k , l ) Y FBF ( k , l )
Unlike the method in [33], where the weight coefficients are constrained by DOA boundaries, the reliable VAD obtained from the BC signal is used to control the weight coefficients update of the normalized least mean square (NLMS) adaptive filter of ABM, shown in Figure 4, which yields:
G ABM ( k , l + 1 ) = G ABM ( k , l ) + f VAD ( l ) μ 0 Y FBF ( k , l ) U ( k , l ) p ¯ FBF ( k , l )
where μ 0 = 0.3 is a step size constant when the speech is present. U ( k , l ) represents the conjugate of U ( k , l ) and p ¯ FBF ( k , l ) is the smoothed power of Y FBF ( k , l ) by a first order recursive smoothing as
p ¯ FBF ( k , l ) = α 0 p ¯ FBF ( k , l 1 ) + ( 1 α 0 ) | Y FBF ( k , l 1 ) | 2
The adaptive filter update controller based on BC VAD does not need to estimate the DOA, and in the harsh sound field environment, it is also difficult to effectively estimate the accurate DOA.

2.3.3. Adaptive Noise Canceler Assisted by BC signal

The ANC adaptively updates the filter G ANC ( k , l ) with the minimum output power criterion and uses the ABM output U ( k , l ) as a reference signal to suppress the residual noise in the output of FBF, that is:
Y GSC ( k , l )   = Y FBF ( k , l ) G ANC H ( k , l ) U ( k , l )
This adaptive problem can be solved by the following NLMS algorithm:
G ANC ( k , l + 1 ) = G ANC ( k , l ) + μ ANC ( k , l ) U ( k , l ) Y GSC ( k , l ) p ¯ U ( k , l )
Similar to (15), p ¯ U ( k , l ) is:
p ¯ U ( k , l ) = α 0 p ¯ U ( k , l 1 ) + ( 1 α 0 ) | U ( k , l 1 ) | 2
where μ ctrl ( k , l ) controls the update of the step size. Considering the signal to interference ratio (SIR) and the VAD of BC signal, μ ctrl ( k , l ) can be updated as
μ ctrl ( k , l ) = μ 0 S I R ( k , l ) f VAD ( l ) + 1
where:
S I R ( k , l ) = | Y FBF ( k , l ) | 2 | U ( k , l ) | 2 .
The update mechanism is as follows. When f VAD ( l ) is equal to 0, the algorithm tends to update with a larger step size, whereas when f VAD ( l ) is equal to 1, the step size update depends on the value of S I R ( k , l ) .

2.4. Robust Compensation Filter for Low Frequencies

After the processing of BC-assisted GSC, an algorithm is developed by using low-frequency BC speech components to suppress the low-frequency noise in the final output. Another NLMS adaptive filter is employed, given by
N ( k , l ) = Y GSC ( k , l ) G CF ( k , l ) Y BC ( k , l )
where Y GSC ( k , l ) is the output of GSC and G CF ( k , l ) represents the CF weight coefficient which can be updated by NLMS as
G CF ( k , l + 1 ) = G CF ( k , l ) + μ CF ( k , l ) Y BC ( k , l ) N ( k , l ) p u ( k , l ) ,
and   p ¯ u ( k , l ) = α 1 p ¯ u ( k , l 1 ) + ( 1 α 1 ) | Y BC ( k , l ) | 2
where μ CF ( k , l ) is used to control the step size update of the CF, and its solution is:
μ CF = μ 1 f VAD ( l )
where μ 1 is a constant controlling the overall step size of NLMS. This solution combines the BC and VAD, so that CF will be updated when speech is present and not be updated when speech is absent. Therefore, the output of CF is:
Y CF ( k , l ) = G CF ( k , l ) Y BC ( k , l )
Based on the above derivations, an optimizer is designed to fuse the low-frequency part of Y CF ( k , l ) with the GSC output so as to provide a reasonable compromise between the noise suppression and voice quality, given by
Y LF ( k , l ) = p 1 ( k , l ) Y GSC ( k , l ) + ( 1 p 1 ( k , l ) ) Y CF ( k , l ) ,
and   p 1 ( k , l ) = t a n h ( S I R ( k , l ) ) = e S I R ( k , l ) e S I R ( k , l ) e S I R ( k , l ) + e S I R ( k , l )
where p 1 ( k , l ) is a value that uses the t a n h ( x ) function to map the SIR to [0.5, 1]. It increases (decreases) the proportion of Y CF ( k , l ) when the SNR is low (high). This ensures the sound quality and intelligibility at high a SNR with good noise suppression capability. Considering the cut-off frequency bin of the BC sensor k cutoff , the output of the system can be written as
Y Fused ( k , l ) = { Y GSC ( k , l ) , k k cutoff , min ( Y LF ( k , l ) , Y GSC ( k , l ) ) , k < k cutoff .
where min ( Y LF ( k , l ) , Y GSC ( k , l ) ) protects the output signal from the thermal noise of the BC sensor when the output SNR of the GSC is high.

2.5. Wind Noise Suppression

In order to suppress the wind noise, we used the output of (21) Y LF ( k , l ) to compensate the low-frequency speech components corrupted by wind noise. The processed BC and clean AC signal are very similar in the frequency spectrum, but simply replacing all the low-frequency parts of the AC signal by BC signals will impair the intelligibility of the GSC output at high SNR. To alleviate this problem, a wind noise detector is deployed here, and only the speech components corrupted by wind noise will be replaced by BC signals.
A wind noise detector based on the ratio of the cross power spectral density (PSD) Φ x 1 x 2 ( k , l ) and the square root of the product of two auto-PSD Φ x 1 x 1 ( k , l ) and Φ x 2 x 2 ( k , l ) is utilized [34], which is given by
Γ ( k , l ) = Φ x 1 x 2 ( k , l ) Φ x 1 x 1 ( k , l ) Φ x 2 x 2 ( k , l )
where:
Φ x 1 x 2 ( k , l ) = α sm Φ x 1 x 2 ( k , l 1 ) + ( 1 α sm ) Y 1 ( k , l ) Y 2 ( k , l ) ,
Φ x 1 x 1 ( k , l ) = α sm Φ x 1 x 1 ( k , l 1 ) + ( 1 α sm ) Y 1 ( k , l ) Y 1 ( k , l ) ,
Φ x 2 x 2 ( k , l ) = α sm Φ x 2 x 2 ( k , l 1 ) + ( 1 α sm ) Y 2 ( k , l ) Y 2 ( k , l )
where α sm = 0.3 is a factor for smoothing the PSD. The wind noise is directly produced by turbulences on a boundary layer close to the microphones. In [35], it is shown that the wind noise between the different microphones is uncorrelated. Therefore, a magnitude squared coherence (MSC) coefficient Γ ( k , l ) was used to describe the correlation of the two microphone signals. With MSC, an indicator of the existence of wind noise is obtained by comparing the ratio with a fixed threshold as follows:
r ( k , l ) = { 0 , | Γ ( k , l ) | σ 1 , 1 . | Γ ( k , l ) | < σ 1 .
where σ 1 = 0.35 is a constant threshold. Finally, the output now is:
Y output ( k , l ) = r ( k , l ) Y GSC ( k , l ) + ( 1 r ( k , l ) ) Y LF ( k , l )

3. Recurrent Neural Network-Based Real-Time Postfiltering with a Small Number of Parameters

3.1. Feature Compression

For training a neural network, it is important to find appropriate features, which affect the training efficiency and deduction performance of neural networks [36]. However, in many approaches summarized by the overview [6], neural networks are used to directly estimate the frequency bin ideal ratio mask (IRM) or magnitude and require a million weights to process 16 kHz speech. Obviously, these types of networks are difficult to deploy in low-power real-time systems. Inspired by the sub-band signal processing method [37], we divide the frequency bin spectrum into 40-dimensional band amplitudes of the Mel-scale [38] as the input features of the neural network, which is shown in Figure 5. The transfer function of the m-th filter is:
H m ( k ) = { 0 , k < f ( m 1 ) , k f ( m 1 ) f ( m ) f ( m 1 ) , f ( m 1 ) k f ( m ) f ( m + 1 ) k f ( m + 1 ) f ( m ) , f ( m ) < k f ( m + 1 ) 0 , k > f ( m + 1 )
where:
f ( m ) = ( N f s ) F mel 1 ( F mel ( f 1 ) + m F mel ( f h ) F mel ( f l ) M + 1 )
F mel ( f ) = 1125 log ( 1 + f / 700 )
M = 40 is the number of filters. For processing audio with 16 kHz sampling rate, f h = 8000 , f l = 0 , f s = 16,000 were used and F mel 1 ( f ) is the inverse function of F mel ( f ) . Then, 161-dimensional features are compressed to 40-dimensions.

3.2. Learning Machine and Training Setup

Recurrent neural network (RNN) presents an outstanding performance in real-time speech-enhancement tasks [36]. Many recurrent architectures, including the long short-term memory (LSTM) [39] and the gated recurrent unit (GRU) [40] rely on gate to control the flow of sequence information to avoid the problem of exponential weight decay or explosion. Simple recurrent unit (SRU) has a better performance and higher computational efficiency than the GRU and LSTM, and it has strong parallel computing capabilities (in both forward propagation and back propagation) [41].
We stacked three layers of SRU and a fully connected layer with a sigmoid activation function as the output to estimate the band IRM as shown in Figure 6. The training label of the band IRM is:
g birm ( b ) = E s ( b ) E x ( b )
where b is the band index, E s ( b ) and E x ( b ) are the clean and noisy band energies, respectively. As shown in Figure 6, we concatenated the features of the t-th frame, the (t − 1)-th frame, and the (t − 2)-th to the neural network, and estimated the band IRM gain of the t − 2th frame to better eliminate non-stationary noise. It is worth pointing out that this does not violate the real-time system’s rule of one frame in and one frame out, but it will cause a two-frame (20 ms) delay. The constant matrix (untrainable) interpolate matrix (IM) is deployed afterwards to interpolate the band IRM from 40 to 161 dimensions and the noisy signal in the complex frequency domain is multiplied with the estimated IRM, and finally the inverse fast Fourier transform (IFFT) is conducted to obtain the time-domain waveform estimation.
For the predicted band IRM, the mean squared error (MSE) is usually used as the loss function in the frequency domain, given by
L 1 = ( g ^ birm ( b ) g birm ( b ) ) 2
In this work, we also propose to utilize the scale-invariant signal-to-distortion ratio (SI-SDR) [42] in time domain loss function to better minimize the distortion of model prediction, which:
L 2 = 10 log 10 ( || s ^ T s | | s | | 2 s || 2 || s ^ T s | | s | | 2 s s ^ || 2 )
where s ^ represents the prediction of the time domain speech wave.
Finally, to integrate two losses, the loss function of the network is weighted by L 1 and L 2 , which can be written as
L = α L 1 + ( 1 α ) L 2
where α is the factor that controls the weights of two losses. In practice, through training experiments, we found that the value α = 0.3 provides a good trade-off between noise suppression and speech distortion.

4. Experimentation Details

4.1. System Processing Pipeline

The system pipeline is depicted in Figure 7, where the BC-assisted real-time speech-enhancement system proposed in this paper consists of two parts: the BC-assisted array signal processing and the SRU-based postfiltering. The former is used to suppress the directional point source interference, and the latter uses a computational efficiency neural network structure as a postfilter algorithm to further suppress residual noise. In Figure 7, x 0 ( t ) , x 1 ( t ) and x BC ( t ) are captured simultaneously. The processing of the sub-modules is carried out frame-by-frame synchronously, where STFT indicates short time Fourier transform and VAD means voice activity detection.

4.2. Performance Metrics

We used two performance metrics, the perceptual evaluation of speech quality (PESQ) [43] and the short-time objective intelligibility (STOI) [44] to evaluate the speech quality and intelligibility, respectively. For these evaluation indicators, higher scores indicate the better results. The PESQ and STOI are all objective, and the evaluations require clean reference signals. Therefore, objective scores can only be performed on the synthetic test set. Research in [45] pointed out that there is still a gap between the objective scores and human’s subjective hearing, so we use data recorded in real scenes and subjective scores are also provided as a supplement to performance measures [45]. The listener is given a score of 1 (very poor sound quality) to 5 (very good sound quality), and each listener is required to go through scoring training and pass the qualification test with a scored (labeled) dataset before scoring. We averaged the scores to obtain the mean opinion score (MOS).

4.3. Comparisons

The method proposed in this paper integrates an array-based adaptive beamformer and a DNN-based postfilter. In order to fully verify the superiority of the proposed algorithm and show the effectiveness of the BC signal assistance in the algorithm performance, we split the proposed algorithm into the BC–GSC module: BC-assisted array signal processing, and the SRU-P module: SRU-based postfiltering to compare with some relevant algorithms. The whole system is named BC–SE. In details, we compare proposed BC–GSC module with two baselines, i.e., the robust (R)-GSC [33], and the transfer-function (TF)–GSC [46]. For the evaluations of the module SRU-P, we selected OMLSA [5] as the conventional method baseline, and RNNoise [7] as the deep learning method baseline. OMLSA is a speech-enhancement algorithm based on digital signal processing (DSP) which is widely used in engineering because of its robustness and low computational complexity. RNNoise is the state-of-the-art deep learning-based single-channel real-time algorithm with a number of parameters less than 100 K. For comparison with whole BC–SE system, the postfilter transient beam-to-reference ratio (TBRR) [27] and RNNoise are combined with basic GSC and R-GSC to form complete array processing postfilter noise reduction system baselines.

4.4. Experimental Setting

In the experiments, the LIS25BA from STMicroelectronics was selected as the BC sensor which is a high-bandwidth three-axis digital accelerometer chip. The price of LIS25BA is about 2 USD. The LIS25BA is available in a small thin plastic land grid array (LGA) package. The development board with dual-microphone and LIS25BA can be seen in Figure 8. LIS25BA is attached on the ear back side to collect the BC signals. The microphone array end-fire direction, along the axis of the array is pointed to the speaker’s mouth with a distance of 10 cm to record clean speech in a reverb-free environment. The STM32 development board is employed to synchronize the capture of the BC and AC signals. In order to objectively assess the performance of the proposed algorithm, the noise and clean speech were first separately recorded and then mixed at different SNRs (−5 dB, 0 dB, 5 dB, 15 dB) with room reverberation. The babble noise generated by a number of talking speakers and car cabin noise are recorded as the diffuse non-directional interference. The music noise and speech-shaped noise are recorded as the directional interferences (60°, 90°, 180°). The microphone array end-fire direction is pointed to the speaker’s mouth with a distance of 10 cm to record the clean speech in a reverb-free environment. LIS25BA is attached to the ear back side to collect the BC signals. We recorded 20 clips of each type of noise, and the average length of each clip was 5 s, which was used to synthesize the noisy data. In addition, we recorded 15 clips of each type of noise in real scenes for subjective evaluation testing. We used the deep noise suppression challenge (DNS)dataset [47] to train the proposed SRU-based network and RNNoise network, and used the Adam [48] optimizer to train with a learning rate of 0.001 and a batch size of 128 and 10 epochs. The original RNNoise was also modified to process a 16 kHz sampling rate audio. We used DNS data collection to form a 500 h training set (single channel, 16 kHz sampling rate, SNR ranging from −5 dB to 30 dB). The DNS test set (synthetic) was used to evaluate the performance of the SRU-based postfilter. In this work, the frame size was 20 ms and the shift was 10 ms (50% overlap), and the FFT size was equal to the frame size.

5. Results and Discussions

5.1. Performance Evaluation of BC–GSC

Thanks to the wind noise suppression module based on the BC signal, the wind noise suppression performance is excellent, which can be determined from the spectrogram in Figure 9 The GSC performance for the AC-only signal is good against point source interference and can significantly improve the voice quality, but the performance for non-point source noise is limited. GSC assisted by BC signal effectively enhances the suppression of non-point source noise at a low SNR.
A clean female speech was mixed with SNR = 0 dB music noise. Three enhancement approaches were applied to obtain results as shown by the spectrograms in Figure 10. It can be observed the R-GSC effectively suppressed the directional noise above 2 kHz, and the BC–GSC better suppressed the directional noise between 1 and 2 kHz. The noise in the non-existent segment of the speech was effectively eliminated. The evaluation metric for each algorithm is obtained on the test set described above, shown by the histograms in Figure 11. For non-point source noise, the performance of AC-only GSC is unsatisfactory. The BC–GSC in the case of car noise and wind noise is significantly better than that of the R-GSC, and TF–GSC. This is because in a diffuse noise scenario, the latter two algorithms have difficulty distinguishing the interference and the desired signal, resulting in the failure of an adaptive filter update. Due to the BC signal that is not affected by sound field noise, the advantage of the proposed approach at low SNR is more significant.

5.2. Performance Evaluation of Postfiltering

Table 1 shows the evaluation performance of OMLSA, RNNoise and our proposed method and the best performing scores are marked in boldface. It can be seen that the DNN-based method is significantly better than the traditional OMLSA method, because the conventional method depends on the assumption that noise changes much more slowly than speech. However, the DNN-based methods do not require this assumption. The proposed method is slightly better than RNNoise at a high SNR, but significantly better than it at a low SNR. This is due to two reasons. One is that the proposed sub-band density is nearly twice that of RNNoise (Proposed: 40, RNNoise: 22), so the spectral resolution is higher. The second is that the usage of two additional frames, as a cost of 20 ms delay, increases the ability to estimate noise and protects the speech component. Since the number of parameters for the proposed method and for RNNoise is almost the same, both can be processed in real time by low-power devices.

5.3. Performance Evaluation of the Proposed Speech Enhancement Algorithm

We averaged the PESQ, STOI and MOS of various noises in the test set. It can be seen from Table 2 that the proposed system is superior to the baseline in terms of subjective and objective evaluation. This is consistent with the results for the separate modules as shown in the previous sections of the performance evaluations. It demonstrates that the proposed speech enhancement algorithm is effective.

6. Conclusions

This paper proposes a new algorithm using BC signals to assist dual-microphone GSC adaptive beamforming for speech enhancement. First, the BC signals were used to conduct highly reliable VAD-assisted ANC and ABM weight coefficient updates in GSC. Second, an adaptive filter CF is designed to compensate the amplitude and phase difference between AC and BC signals. Third, wind noise is detected and replaced with the output of CF to recover low-frequency speech components from the wind noise. Finally, a real-time neural network-based postfilter is designed and trained to effectively remove the residual noise. Experimental results show that the proposed algorithm effectively improves SNR and the speech quality in different scenarios, and the assistance of BC signals can effectively improve the noise reduction performance of beamforming.

Author Contributions

Conceptualization, Y.Z.; methodology, Y.C. and Y.M.; software, Y.C.; validation, Y.C. and Y.M.; formal analysis, Y.C.; investigation, Y.C. and Y.Z.; resources, Y.Z. and H.L.; data curation, Y.M.; writing—original draft preparation, Y.C.; writing—review and editing, Y.C. and H.L.; visualization, Y.M., H.L. and Y.C.; supervision, Y.Z. and H.L.; project administration, Y.Z.; funding acquisition, H.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant 61801066.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Loizou, P.C. Speech Enhancement: Theory and Practice; CRC Press: Boca Raton, FL, USA, 2013. [Google Scholar]
  2. Upadhyay, N.; Karmakar, A. An improved multi-band spectral subtraction algorithm for enhancing speech in various noise environments. Procedia Eng. 2013, 64, 312–321. [Google Scholar] [CrossRef] [Green Version]
  3. Martin, R. Spectral subtraction based on minimum statistics. Power 1994, 6, 1182–1185. [Google Scholar]
  4. Abd El-Fattah, M.A.; Dessouky, M.I.; Diab, S.M.; Abd El-Samie, F.E.S. Speech enhancement using an adaptive wiener filtering approach. Prog. Electromagn. Res. 2008, 4, 167–184. [Google Scholar] [CrossRef] [Green Version]
  5. Cohen, I. Optimal speech enhancement under signal presence uncertainty using log-spectral amplitude estimator. IEEE Signal Process. Lett. 2002, 9, 113–116. [Google Scholar] [CrossRef]
  6. Wang, D.; Chen, J. Supervised speech separation based on deep learning: An overview. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1702–1726. [Google Scholar] [CrossRef]
  7. Valin, J.M. A hybrid DSP/deep learning approach to real-time full-band speech enhancement. In Proceedings of the 2018 IEEE 20th International Workshop on Multimedia Signal Processing (MMSP), Vancouver, BC, Canada, 29–31 August 2018; pp. 1–5. [Google Scholar]
  8. Benesty, J.; Chen, J.; Huang, Y. Microphone Array Signal Processing; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2008. [Google Scholar]
  9. Habets, E.A.P.; Benesty, J.; Cohen, I.; Gannot, S.; Dmochowski, J. New insights into the MVDR beamformer in room acoustics. IEEE Trans. Audio Speech Lang. Process. 2009, 18, 158–170. [Google Scholar] [CrossRef]
  10. Frost, O.L. An algorithm for linearly constrained adaptive array processing. Proc. IEEE 1972, 60, 926–935. [Google Scholar] [CrossRef]
  11. Griffiths, L.; Jim, C.W. An alternative approach to linearly constrained adaptive beamforming. IEEE Trans. Antennas Propag. 1982, 30, 27–34. [Google Scholar] [CrossRef] [Green Version]
  12. Bitzer, J.; Simmer, K.U.; Kammeyer, K.D. Theoretical noise reduction limits of the generalized sidelobe canceller (GSC) for speech enhancement. In Proceedings of the 1999 IEEE International Conference on Acoustics, Speech, and Signal Processing, ICASSP99 (Cat. No. 99CH36258), Phoenix, AZ, USA, 15–19 March 1999; Volume 5, pp. 2965–2968. [Google Scholar]
  13. Su, J.; Xu, R.; Yu, S.; Wang, B.; Wang, J. Redundant rule Detection for Software-Defined Networking. KSII Trans. Internet Inf. Syst. 2020, 14, 2735–2751. [Google Scholar]
  14. Su, J.; Xu, R.; Yu, S.; Wang, B.; Wang, J. Idle Slots Skipped Mechanism based Tag Identification Algorithm with Enhanced Collision Detection. KSII Trans. Internet Inf. Syst. TIIS 2020, 14, 2294–2309. [Google Scholar]
  15. Su, J.; Chen, Y.; Sheng, Z.; Huang, Z.; Liu, A.X. From M-Ary query to bit query: A new strategy for efficient large-scale RFID identification. IEEE Trans. Commun. 2020, 68, 2381–2393. [Google Scholar] [CrossRef]
  16. Su, J.; Sheng, Z.; Liu, A.X.; Han, Y.; Chen, Y. A group-based binary splitting algorithm for UHF RFID anti-collision systems. IEEE Trans. Commun. 2019, 68, 998–1012. [Google Scholar] [CrossRef] [Green Version]
  17. Shin, H.S.; Kang, H.G.; Fingscheidt, T. Survey of speech enhancement supported by a bone conduction microphone. In Proceedings of the 10th ITG Symposium on Speech Communication, Braunschweig, Germany, 26–28 September 2012; pp. 1–4. [Google Scholar]
  18. Liu, H.P.; Tsao, Y.; Fuh, C.S. Bone-conducted speech enhancement using deep denoising autoencoder. Speech Commun. 2018, 104, 106–112. [Google Scholar] [CrossRef]
  19. Bouserhal, R.E.; Falk, T.H.; Voix, J. In-ear microphone speech quality enhancement via adaptive filtering and artificial bandwidth extension. J. Acoust. Soc. Am. 2017, 141, 1321–1331. [Google Scholar] [CrossRef] [PubMed]
  20. Shin, H.S.; Fingscheidt, T.; Kang, H.G. A priori SNR estimation using air-and bone-conduction microphones. IEEE/ACM Trans. Audio Speech Lang. Process. 2015, 23, 2015–2025. [Google Scholar] [CrossRef]
  21. Lee, C.H.; Rao, B.D.; Garudadri, H. Bone-Conduction Sensor Assisted Noise Estimation for Improved Speech Enhancement. In Proceedings of the INTERSPEECH, Hyderabad, India, 2–6 September 2018; pp. 1180–1184. [Google Scholar]
  22. McCowan, I.A.; Bourlard, H. Microphone array post-filter for diffuse noise field. In Proceedings of the 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, Orlando, FL, USA, 13–17 May 2002; Volume 1, pp. 905–908. [Google Scholar]
  23. Soni, M.H.; Shah, N.; Patil, H.A. Time-frequency masking-based speech enhancement using generative adversarial network. In Proceedings of the 2018 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Calgary, AB, Canada, 15–20 April 2018; pp. 5039–5043. [Google Scholar]
  24. Zezario, R.E.; Huang, J.W.; Lu, X.; Tsao, Y.; Hwang, H.T.; Wang, H.M. Deep denoising autoencoder based post filtering for speech enhancement. In Proceedings of the 2018 Asia-Pacific Signal and Information Processing Association Annual Summit and Conference (APSIPA ASC), Honolulu, HI, USA, 12–15 November 2018; pp. 373–377. [Google Scholar]
  25. Li, J.; Akagi, M. A hybrid microphone array post-filter in a diffuse noise field. Appl. Acoust. 2008, 69, 546–557. [Google Scholar] [CrossRef] [Green Version]
  26. Kumatani, K.; Raj, B.; Singh, R.; McDonough, J. Microphone array post-filter based on spatially-correlated noise measurements for distant speech recognition. In Proceedings of the Thirteenth Annual Conference of the International Speech Communication Association, Portland, OR, USA, 9–13 September 2012; pp. 298–301. [Google Scholar]
  27. Cohen, I.; Berdugo, B. Microphone array post-filtering for non-stationary noise suppression. In Proceedings of the 2002 IEEE International Conference on Acoustics, Speech, and Signal Processing, Orlando, FL, USA, 13–17 May 2002; Volume 1, pp. 901–904. [Google Scholar]
  28. Zhou, M.; Wang, Y.; Liu, Y.; Tian, Z. An information-theoretic view of WLAN localization error bound in GPS-denied environment. IEEE Trans. Veh. Technol. 2019, 68, 4089–4093. [Google Scholar] [CrossRef]
  29. Zhou, M.; Wang, Y.; Tian, Z.; Lian, Y.; Wang, Y.; Wang, B. Calibrated data simplification for energy-efficient location sensing in internet of things. IEEE Internet Things J. 2018, 6, 6125–6133. [Google Scholar] [CrossRef]
  30. Lleida, E.; Fernandez, J.; Masgrau, E. Robust continuous speech recognition system based on a microphone array. In Proceedings of the 1998 IEEE International Conference on Acoustics, Speech and Signal Processing, ICASSP’98 (Cat. No. 98CH36181), Seattle, WA, USA, 15 May 1998; Volume 1, pp. 241–244. [Google Scholar]
  31. Ephraim, Y.; Malah, D. Speech enhancement using a minimum mean-square error log-spectral amplitude estimator. IEEE Trans. Acoust. Speech Signal Process. 1985, 33, 443–445. [Google Scholar] [CrossRef]
  32. Gerkmann, T.; Hendriks, R.C. Unbiased MMSE-based noise power estimation with low complexity and low tracking delay. IEEE Trans. Audio Speech Lang. Process. 2011, 20, 1383–1393. [Google Scholar] [CrossRef]
  33. Yoon, B.J.; Tashev, I.; Acero, A. Robust adaptive beamforming algorithm using instantaneous direction of arrival with enhanced noise suppression capability. In Proceedings of the 2007 IEEE International Conference on Acoustics, Speech and Signal Processing-ICASSP’07, Honolulu, HI, USA, 15–20 April 2007; Volume 1, pp. 133–136. [Google Scholar]
  34. Nelke, C.M.; Vary, P. Dual microphoneWind Noise Reduction by Exploiting the Complex Coherence. In Proceedings of the 11th ITG Symposium on Speech Communication, Erlangen, Germany, 24–26 September 2014; pp. 1–4. [Google Scholar]
  35. Müller, G.; Möser, M. (Eds.) Handbook of Engineering Acoustics; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012. [Google Scholar]
  36. Chen, J.; Wang, Y.; Wang, D. A feature study for classification-based speech separation at low signal-to-noise ratios. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 1993–2002. [Google Scholar] [CrossRef]
  37. Doblinger, G. Computationally efficient speech enhancement by spectral minima tracking in subbands. In Proceedings of the Fourth European Conference on speech Communication and Technology, Madrid, Spain, 18–21 September 1995; Volume 2, pp. 1513–1516. [Google Scholar]
  38. Davis, S.; Mermelstein, P. Comparison of parametric representations for monosyllabic word recognition in continuously spoken sentences. IEEE Trans. Acoust. Speech Signal Process. 1980, 28, 357–366. [Google Scholar] [CrossRef] [Green Version]
  39. Greff, K.; Srivastava, R.K.; Koutník, J.; Steunebrink, B.R.; Schmidhuber, J. LSTM: A search space odyssey. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2222–2232. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  40. Chung, J.; Gulcehre, C.; Cho, K.; Bengio, Y. Empirical evaluation of gated recurrent neural networks on sequence modeling. arXiv 2014, arXiv:1412.3555. [Google Scholar]
  41. Lei, T.; Zhang, Y.; Artzi, Y. Training RNNs as Fast as CNNs. arXiv 2017, arXiv:1709.02755. [Google Scholar]
  42. Le Roux, J.; Wisdom, S.; Erdogan, H.; Hershey, J.R. SDR–half-baked or well done? In Proceedings of the ICASSP 2019–2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Brighton, UK, 12–17 May 2019; pp. 626–630. [Google Scholar]
  43. Rix, A.W.; Beerends, J.G.; Hollier, M.P.; Hekstra, A.P. Perceptual evaluation of speech quality (PESQ)-a new method for speech quality assessment of telephone networks and codecs. In Proceedings of the 2001 IEEE International Conference on Acoustics, Speech, and Signal Processing, (Cat. No. 01CH37221), Salt Lake City, UT, USA, 7–11 May 2001; Volume 2, pp. 749–752. [Google Scholar]
  44. Taal, C.H.; Hendriks, R.C.; Heusdens, R.; Jensen, J. An algorithm for intelligibility prediction of time–frequency weighted noisy speech. IEEE Trans. Audio Speech Lang. Process. 2011, 19, 2125–2136. [Google Scholar] [CrossRef]
  45. Reddy, C.K.; Beyrami, E.; Pool, J.; Cutler, R.; Srinivasan, S.; Gehrke, J. A scalable noisy speech dataset and online subjective test framework. arXiv 2019, arXiv:1909.08050. [Google Scholar]
  46. Gannot, S.; Cohen, I. Speech enhancement based on the general transfer function GSC and postfiltering. IEEE Trans. Speech Audio Process. 2004, 12, 561–571. [Google Scholar] [CrossRef]
  47. Reddy, C.K.; Beyrami, E.; Dubey, H.; Gopal, V.; Cheng, R.; Cutler, R.; Matusevych, S.; Aichner, R.; Aazami, A.; Braun, S.; et al. The interspeech 2020 deep noise suppression challenge: Datasets, subjective speech quality and testing framework. arXiv 2020, arXiv:2001.08662. [Google Scholar]
  48. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Dual-microphone array with end fire desired signal.
Figure 1. Dual-microphone array with end fire desired signal.
Sensors 20 05050 g001
Figure 2. Clean air-conduction (AC) signal and bone-conduction (BC) signal.
Figure 2. Clean air-conduction (AC) signal and bone-conduction (BC) signal.
Sensors 20 05050 g002
Figure 3. The robust generalized sidelobe canceller (GSC) with coefficients constrained ABM structure.
Figure 3. The robust generalized sidelobe canceller (GSC) with coefficients constrained ABM structure.
Sensors 20 05050 g003
Figure 4. The framework of the proposed BC signal-assisted adaptive beamforming algorithm for speech enhancement.
Figure 4. The framework of the proposed BC signal-assisted adaptive beamforming algorithm for speech enhancement.
Sensors 20 05050 g004
Figure 5. The scheme for extracting Mel-scale sub-band features.
Figure 5. The scheme for extracting Mel-scale sub-band features.
Sensors 20 05050 g005
Figure 6. The diagram of the proposed Mel-scale sub-band simple recurrent unit (SRU)-based neural network. FC means the fully connection. IM indicates interpolate matrix. IFFT means Inverse fast Fourier transform and IRM is ideal ratio mask.
Figure 6. The diagram of the proposed Mel-scale sub-band simple recurrent unit (SRU)-based neural network. FC means the fully connection. IM indicates interpolate matrix. IFFT means Inverse fast Fourier transform and IRM is ideal ratio mask.
Sensors 20 05050 g006
Figure 7. The system pipeline is divided into two parts: BC-assisted array signal processing and SRU-based postfiltering.
Figure 7. The system pipeline is divided into two parts: BC-assisted array signal processing and SRU-based postfiltering.
Sensors 20 05050 g007
Figure 8. The STM32 development board with the dual-microphone array and the bone conduction sensor LIS25BA.
Figure 8. The STM32 development board with the dual-microphone array and the bone conduction sensor LIS25BA.
Sensors 20 05050 g008
Figure 9. Wind noise processed by the proposed dual-microphone BC sensor-assisted method.
Figure 9. Wind noise processed by the proposed dual-microphone BC sensor-assisted method.
Sensors 20 05050 g009
Figure 10. Spectrograms of an utterance corrupted by 0 dB music interference at 90° and the processing results by the robust (R)-GSC, the transfer-function (TF)–GSC and the proposed BC–GSC methods, respectively.
Figure 10. Spectrograms of an utterance corrupted by 0 dB music interference at 90° and the processing results by the robust (R)-GSC, the transfer-function (TF)–GSC and the proposed BC–GSC methods, respectively.
Sensors 20 05050 g010
Figure 11. The perceptual evaluation of speech quality (PESQ) of the R-GSC, the TF–GSC and the proposed BC–GSC methods in different noises.
Figure 11. The perceptual evaluation of speech quality (PESQ) of the R-GSC, the TF–GSC and the proposed BC–GSC methods in different noises.
Sensors 20 05050 g011
Table 1. The postfilter (single-channel) speech enhancement results obtained with optimally modified log-spectral amplitude (OMLSA), Recurrent neural network (RNN) noise and the proposed method, where noisy means the unprocessed original signal.
Table 1. The postfilter (single-channel) speech enhancement results obtained with optimally modified log-spectral amplitude (OMLSA), Recurrent neural network (RNN) noise and the proposed method, where noisy means the unprocessed original signal.
SNR 0–5 dBSNR 5–20 dB
NoisyOMLSARNNoiseSRU-PNoisyOMLSARNNoiseSRU-P
PESQ1.751.772.432.512.402.642.862.94
STOI (%)76.576.076.777.988.288.388.588.7
ParametersNoneOracle88 k89 kNoneOracle88 k89 k
Table 2. The objective and subjective measurements comparing the PESQ and MOS performance among TF–GSC+TBRR, TF–GSC + RNNoise, R-GSC + RNNoise and BC–SE.
Table 2. The objective and subjective measurements comparing the PESQ and MOS performance among TF–GSC+TBRR, TF–GSC + RNNoise, R-GSC + RNNoise and BC–SE.
0–5 dB
NoisyTF–GSC + TBRRTF–GSC+RNNoiseR-GSC+RNNoiseBC–SE
PESQ1.181.291.521.551.64
STOI (%)71.371.470.67074.3
MOS2.542.713.563.453.64
5–15 dB
NoisyTF–GSC+TBRRTF–GSC+RNNoiseR-GSC+RNNoiseBC–SE
PESQ1.871.952.152.192.24
STOI (%)76.476.976.176.977.1
MOS3.753.843.873.823.86

Share and Cite

MDPI and ACS Style

Zhou, Y.; Chen, Y.; Ma, Y.; Liu, H. A Real-Time Dual-Microphone Speech Enhancement Algorithm Assisted by Bone Conduction Sensor. Sensors 2020, 20, 5050. https://doi.org/10.3390/s20185050

AMA Style

Zhou Y, Chen Y, Ma Y, Liu H. A Real-Time Dual-Microphone Speech Enhancement Algorithm Assisted by Bone Conduction Sensor. Sensors. 2020; 20(18):5050. https://doi.org/10.3390/s20185050

Chicago/Turabian Style

Zhou, Yi, Yufan Chen, Yongbao Ma, and Hongqing Liu. 2020. "A Real-Time Dual-Microphone Speech Enhancement Algorithm Assisted by Bone Conduction Sensor" Sensors 20, no. 18: 5050. https://doi.org/10.3390/s20185050

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop