Next Article in Journal
A Multilevel Switched Capacitor Inverter with Reduced Components and Self-Balance
Next Article in Special Issue
Environment-Aware Knowledge Distillation for Improved Resource-Constrained Edge Speech Recognition
Previous Article in Journal
Multi-Task Deep Evidential Sequence Learning for Trustworthy Alzheimer’s Disease Progression Prediction
Previous Article in Special Issue
End-to-End Mispronunciation Detection and Diagnosis Using Transfer Learning
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Orthogonalization of the Sensing Matrix Through Dominant Columns in Compressive Sensing for Speech Enhancement

Department of Electronics and Communication Engineering, University Institute of Technology RGPV, Bhopal 462033, India
*
Author to whom correspondence should be addressed.
Appl. Sci. 2023, 13(15), 8954; https://doi.org/10.3390/app13158954
Submission received: 21 April 2023 / Revised: 20 June 2023 / Accepted: 22 July 2023 / Published: 4 August 2023
(This article belongs to the Special Issue Advances in Speech and Language Processing)

Abstract

:
This paper introduces a novel speech enhancement approach called dominant columns group orthogonalization of the sensing matrix (DCGOSM) in compressive sensing (CS). DCGOSM optimizes the sensing matrix using particle swarm optimization (PSO), ensuring separate basis vectors for speech and noise signals. By utilizing an orthogonal matching pursuit (OMP) based CS signal reconstruction with this optimized matrix, noise components are effectively avoided, resulting in lower noise in the reconstructed signal. The reconstruction process is accelerated by iterating only through the known speech-contributing columns. DCGOSM is evaluated against various noise types using speech quality measures such as SNR, SSNR, STOI, and PESQ. Compared to other OMP-based CS algorithms and deep neural network (DNN)-based speech enhancement techniques, DCGOSM demonstrates significant improvements, with maximum enhancements of 42.54%, 62.97%, 27.48%, and 8.72% for SNR, SSNR, PESQ, and STOI, respectively. Additionally, DCGOSM outperforms DNN-based techniques by 20.32% for PESQ and 8.29% for STOI. Furthermore, it reduces recovery time by at least 13.2% compared to other OMP-based CS algorithms.

1. Introduction

In recent years, the domains of signal processing and information theory have shown significant interest in compressive sensing (CS) [1,2]. Compressive sensing, alternatively referred to as compressed sensing or compressive sampling, has emerged as a potent technique for signal processing, with extensive applications across multiple fields. By leveraging the innate sparsity or compressibility of signals, compressive sensing enables the acquisition and reconstruction of signals using considerably fewer measurements than conventional methods demand. This pioneering concept has transformed data acquisition and processing across diverse domains, including image and video processing, audio processing, medical imaging, wireless communication, etc. [3,4,5,6,7].
Conventional approaches often depend on uniformly sampling signals at the Nyquist rate, which can be computationally burdensome and require significant storage capacity. In contrast, compressive sensing enables sub-Nyquist sampling, thereby lowering the demands for data acquisition and storage while preserving the fidelity of the signal.
The CS suggests that, compared to the number of samples required by conventional Nyquist-based approaches, a signal can be reconstructed with fewer samples (observations) [8]. However, for CS to work properly, the input signal must be highly compressible, or more specifically, sparse. A sparse signal has few active (nonzero) components in relation to its length. The signals can exhibit this property either in their sampling domain or any other underlying transform domain such as Fourier, wavelet, curvelet, etc. In CS, there are two major components: encoding (sampling) and decoding (recovery). Let vector x R N × 1 denote N successive samples coming from a transducer. The vector x can be transformed into a sparse vector X = ψ x through a N × N sparse transforming matrix ψ . If X contains k non-zero elements, it is said to be k sparse. The reconstruction of the original data x from y R M × 1 measurements, where y = ϕ X = ϕ ψ x , requires k < < M < < N . Generating a suitable sensing matrix to improve the reconstruction process is one of the most challenging tasks in CS. The restricted isometry property (RIP) states that the performance of compressed sensing algorithms improves when the mutual coherence between the sensing matrix ϕ and the transforming matrix ψ decreases [9,10].
The rest of the paper is organized as follows: Section 2 briefly reviews the related work. Section 3 provides a detailed explanation of the proposed technique along with the theoretical background of the different processes used in this work. In Section 4, the evaluation results for the proposed technique with detailed analysis are presented. This section also includes a comparison of various methods. Finally, Section 5 concludes the paper.

2. Related Work

Finding an optimal CS matrix for compression and reconstruction is one of the most difficult problems in CS. Randomly constructing CS matrix elements that meet the RIP requirement does not produce the optimal solution. Because of this, random sensing matrices have been abandoned in favor of deterministic matrices in a number of research studies. The Ternary matrix [11], Walsh–Hadamard [12], and low-density parity checks [13] are examples of deterministic matrices that can be used in computations.
Another challenge is finding suitable algorithms for compressed speech enhancement and a domain of transformation that is suitable for sparse signals. Many basis functions are available for speech signals [14,15], like the Fourier Transform, Cosine Transform, and Wavelet Transforms [16].
There are several different types of sensing matrices; for example, the Gaussian random matrix, Bernoulli, Fourier, wavelets, etc. [14]. Several optimization techniques, such as l 1 minimization [17], orthogonal matching pursuit (OMP) [18], and compressive sampling matching pursuit (CoSaMP) [19], were also utilized to recover the compressively sampled speech signals. The study presented by Pilastri et al. [20] analyzed the performance of several CS reconstruction algorithms applied to image reconstruction and suggested that the l 1 minimization-based basis pursuit algorithm outperforms the other OMP methods.
As presented in [21], the basis pursuit algorithm also provides a slight edge over the CoSaMP for speech enhancement. Several CS-based methods for speech enhancement have also been proposed in [22,23,24]. Wu et al. [22,24] employed a random sampling matrix in their sensing process.
At present, the most accurate approach for determining the quality of speech is to conduct subjective listening tests. However, subjective evaluations of speech enhancement algorithms are costly and time-consuming, despite the fact that they are accurate, reliable, and performed under rigorous conditions [25,26,27,28]. Therefore, many efforts have been made to develop objective measures that are highly correlated with speech quality.
Several attempts have been made over the years to predict subjective speech quality using objective speech quality metrics [25]. Objective measures based on LPC, such as cepstrum distance measures (CEP) [29] and frequency-weighted segmental SNR (fwsegSNR) [25], are among the objective speech quality measures examined in [28], along with others, such as segmental SNR (segSNR) [30], weighted-slope spectral distance (WSS) [31], and PESQ [32,33].
The preceding discussion has highlighted that the basic compressive sensing idea has been modified in different ways to achieve better speech enhancement performance. Many ideas have been proposed for efficient reconstruction of the compressed signal [34], as shown in Figure 1. However, this work is primarily concerned with the efficient reconstruction of compressed signals while achieving better enhancement using greedy algorithms. Greedy algorithms recover the signal iteratively by choosing a local optimal solution at each iteration with the intention of eventually discovering the global optimal solution. In general, greedy algorithms take the steps depicted in Figure 2. In each repetition, the quantity of the column selected varies slightly. Only one column is selected for each iteration of the matching pursuit (MP) [35], orthogonal matching pursuit (OMP), and matching pursuit based on least squares (MPLS) [36] algorithms.
In contrast, the subspace pursuit (SP) [37], stagewise-OMP (StOMP) [38], compressive sampling matching pursuit (CoSaMP) [19], regularized-OMP (ROMP) [39], generalized-OMP (GOMP) [40], generalized orthogonal adaptive matching pursuit (GOAMP) [41], constrained backtracking matching pursuit (CBMP) [42], generalized backtracking regularized adaptive matching pursuit (GBRAMP) [43], OMP algorithm based on singular value decomposition [44], enhanced block-based OMP pursuit [45], gradient pursuit (GP) [46], and multipath matching pursuit (MMP) [47], involve columns with projected values that exceed the thresholds selected by the algorithm. Other differences between the algorithms include the way the residual vector γ i t e r is calculated and how the non-zero values of x e s t are estimated. For instance, the MPLS and the subspace pursuit (SP) algorithms do not estimate x e s t until the last step of their respective processes.
As the aforementioned discussion has highlighted, all the existing greedy algorithms use either single or groups of column selection. For every frame of the signal, iterations are required to find the dominating columns set ( ϕ i t e r d o m i n a n t ), as described in Figure 2. This makes the algorithm complex and time-consuming. To overcome this problem, we proposed an optimized measurement matrix generation in such a way that all dominating columns consisting of the original signal components are known in advance. Furthermore, the optimization minimizes the presence of noise components in these columns to achieve better signal enhancement.
Based on these investigations, this paper proposes a CS-based speech enhancement technique called the dominant columns group orthogonalization of the sensing matrix (DCGOSM).

3. Theoretical Background

3.1. Preprocessing

In signal processing, preprocessing plays an important role. Once experimental results are obtained, they are modeled to extract useful information. Generally, the data output is either too large, too small, or fragmented. As a part of preprocessing, the data are classified and processed accordingly [48]. In this work, we use the following normalization procedure to translate the signal into the range of 1 ,   1 [49]:
S n o r m i = 2 × S o r g i S m i n S m a x S m i n 1
where S o r g i , S n o r m i are the i t h samples of the original and normalized (scaled) signal, respectively, and S m a x and S m i n are the maximum and minimum values achieved by the signal S o r g , respectively.

3.2. Discrete Cosine Transform (DCT) and Inverse DCT (IDCT)

The DCT transforms a signal into the frequency domain, similar to the discrete Fourier transform (DFT). However, unlike to the DFT, the DCT relies only on real-valued cosine functions; therefore, it is simpler to compute [50]. Another property that makes the DCT superior to the DFT is its high spectral compaction. Therefore, a signal’s DCT transformation includes more energy in fewer coefficients than other transformations, such as the DFT.
For sparse encoding of a signal, where a small number of nonzero samples are required, this is preferable. Due to the small number of DCT coefficients that hold substantial energy in a speech signal, all remaining coefficients can be reduced to zero without losing any significant information [51]. Therefore, the voice signal will be sparsely represented in this way.
Figure 3 illustrates the process of generating a sparse signal representation using the discrete cosine transform (DCT). In Figure 3a, a speech signal frame consisting of 32 samples is depicted. The corresponding DCT transform with 32 coefficients is shown in Figure 3b. The DCT coefficients are then arranged in descending order and presented in Figure 3c. Furthermore, Figure 3d shows the plot of energy in the largest K coefficients, where K takes values of 5, 10, 15, 20, 25, and 30. The plot in Figure 3d demonstrates that the top five DCT coefficients encompass 75.65% of the total energy of the 32 samples in the speech signal frame, while the top ten coefficients capture 95.36% of the total energy. These findings provide empirical evidence that only a small number (five to ten) of coefficients are sufficient to extract the most pertinent information from the complete frame. Therefore, zeroing the non-significant coefficients will produce a sparse signal with minimal information loss.
There are several different iterations of DCT available, typically referred to as DCT-I to DCT-IV. They all have very subtle distinctions amongst one another [52]. Among these four variants, DCT-II is by far the most common and has thus been chosen for this work. The following equation can be used as a general formula for computing the DCT of a one-dimensional signal f n with N number of samples:
g k = Λ k 2 N n = 0 N 1 f n cos π k 2 n + 1 2 N ,   0 k < N 0 ,          O t h e r w i s e
where g [ k ] denotes the DCT coefficients vector of length N .
The IDCT is calculated as:
f n = 2 N k = 0 N 1 Λ k g k cos π k 2 n + 1 2 N ,   0 k < N 0 ,          O t h e r w i s e
where Λ k is defined as:
Λ k = 1 2 ,      k = 0 1 ,   1 k N

3.3. Framing and Deframing

In signal processing, framing refers to the process of splitting samples into small groups (blocks) known as “frames”. Real-time block-based signal processing systems often use framing because it minimizes the processing and memory requirements by dividing one large chunk of samples into several small frames [53]. It also helps in extracting features from non-stationary signals (i.e., speech) by dividing them into small blocks during which they are assumed to be stationary. Figure 4 demonstrates how the frames are designed to overlap with one another to prevent data loss between successive frames [54]. In this figure F a , F b , and F c represent the frames a , b , and c , respectively, and O L a b and O L b c are the overlapped frames.
To avoid spectral distortions at the frame edges and to achieve seamless continuity among consecutive frames of speech, the Hamming window ( w n ) is multiplied with each frame [55]:
w n = 0.54 0.46 cos 2 π n N 1 ,   0 n N
where N denotes the samples in the speech frame. The windowed signal y n can be calculated as:
y n = x n × w n ,   0 n N
Windowing acts as a low-pass filter, enhancing the signal at the center and smoothing it at the edges.

3.4. Voice Activity Detector

Voice activity detection (VAD) is a process that detects the presence of a voice in a noisy signal. It is an essential component in speech processing applications like voice and speech detection, speech recognition, speech enhancement, and speaker recognition [56].
In this work, a VAD detector based on the work presented in [56] is utilized. Figure 5 shows an overview of the VAD. According to this model, the four different feature streams are used as follows:
  • Spectral Shape: The short-term power spectrum envelope of the speech signal.
  • Spectro–temporal modulations: The spectral scales and temporal rates of speech.
  • Voicing: The vibrations of the vocal cords (folds) that appear in speech as fundamental and harmonic components of the pitch (frequency).
  • Long-term variability: The variations in speech caused by successive phone generation.
These feature streams are extracted from the input audio stream. These streams are divided into small frames, then the DCT of each feature stream is taken. The DCT contains a large number of coefficients. However, the first five DCT components are sufficient to extract the most relevant context information from those frames [56].
The resulting five-dimensional vectors from each of the four feature streams are normalized and combined into a single twenty-dimensional frame feature vector.
Finally, these frame feature vectors are used to train a multilayer perceptron (MLP) classifier to detect the presence of speech. Speech segments are detected by thresholding the ratio of speech and non-speech outputs of the MLP.

3.5. Particle Swarm Optimization (PSO)

Particle swarm optimization is a meta-heuristic searching algorithm that is used to obtain the optimal solution to numerical problems. PSO was originally proposed by Kennedy and Elberhartin in 1995, [57] influenced by the social behavior of bird flocking or fish schooling. Currently, the algorithm has been well-tested on many complex mathematical problems and found to be efficient and fast. The pseudocode for PSO is presented in Algorithm 1. PSO initially populates particles (points in the solution space), and these particles are considered a probable solution to the optimization problem, which is given by a fitness function. In every iteration, a new position for the particles is calculated using several parameters such as the velocity vector, global best solution, local best solution, etc.
Figure 6 shows the velocity and position update procedure for i t h particle at time step t using Equations (7) and (8), respectively. p b e s t i represents the particle’s information gained from its past movements. Similarly, g b e s t represents the collective information gained by all the particles.
The velocity update of each particle is performed using the following equation:
v i t + 1 = w v i t + c 1 p b e s t i x i t + c 2 g b e s t x i t
where c 1 , and c 2 are priority factors used to configure the dominance of p b e s t i and g b e s t . The new position x i t + 1 of the i t h particle is calculated as:
x i t + 1 = x i t + v i t + 1
In Equation (7), the first term ( w v i t ) contains the variable w , which represents the inertia weight. Therefore, the whole term is known as the inertia component, which is responsible for maintaining the particle moment it initially possessed.
The value of w is usually maintained between 0.8 and 1.2 . Setting up a smaller value for w reduces the algorithm’s search space and converges quickly to the optima (sometimes trapping in local optima), while higher values force the algorithm to search over a wider space (which sometimes skips the global optima).
The second term in Equation (7), c 1 p b e s t i x i t , serves as the particle’s memory and is known as cognitive component. It forces the process to search in the area where the best solution can be found based on the particle’s perspective. This coefficient ( c 1 ) is usually set to 2.
The third term in Equation (7), c 2 g b e s t x i t , is known as the social component. It forces the particle movements towards the area belonging to the best of each particle. This component is also set to 2.
Algorithm 1: PSO Algorithm
Initialize:
 Set initial positions x i ( 0 ) |   i { 1,2 , 3 , , N } and velocities v i ( 0 ) |   i { 1,2 , 3 , N } of each of the N particles randomly within a specified range.
 Set intertia weight w .
 Set constant c 1 and c 2 .
 Set p b e s t i for each particle as the initial position.
 Set g b e s t as the best position among all particles.
Main:
 While termination criterion is not met do:
  For each particle do:
   Evaluate fitness of particle through objective function presented in eq. (18)
    If current position is better than p b e s t i :
     Update p b e s t i with current position
    If current position is better than g b e s t :
     Update g b e s t with current position
   For each particle do:
 Update particle velocity using:
   v i ( t + 1 )   =   w v i t +   c 1 r 1 p b e s t     x i t +   c 2 r 2 ( g b e s t     x i ( t ) )
 Update positions of particles using:
   x i t + 1 = x i t + v i ( t + 1 )
 End while
 Return g b e s t as the best solution found

3.6. Compressive Sensing (CS)

Compressive sensing (CS) is also referred as compressive sampling, sparse recovery, or compressed sensing. Conventional Shannon theorem-based sampled signal recovery algorithms have the following basic constraints [58]:
  • The Shannon theorem states that the sampling rate needs to be more than or equal to twice the signal’s bandwidth [59].
  • Linear algebra stipulates that the length of the observed samples must be equal to or greater than the length of original signal [38].
Compressive sensing aims to overcome these limitations by stating that sparse signals can be reconstructed from incomplete samples. It enables the reconstruction of a compressed signal with only a few linear and non-adaptive measurements [1]. Additionally, unlike conventional signal compression methods, which use a sample-then-compress procedure, the CS achieves compression while sampling. The fundamental equation for recovering the CS signal is as follows:
y M × 1 = ϕ M × N x N × 1 + v M × 1
where x , y , and v denote the sparse signal, observation vector, and noise, respectively, while ϕ denotes the sensing (recovery) matrix. To achieve proper compression and recovery through CS, the following conditions must be satisfied:
  • M N , to ensure the compression [60,61].
  • x is required to be a k-sparse vector that satisfies k < < N , where N denotes the length of x .
  • ϕ required to be a matrix having full rank.
The reconstruction of x from y using Equation (9) requires the estimation of ϕ -inverse:
x ϕ 1 y ,   neglecting   the   ν
Since ϕ is a non-square matrix ( M N ), instead of inverse, the pseudo-inverse is used. Pseudo-inverses are defined for matrices with real or complex entries, and it is unique for these matrices [62]. The estimation of pseudo-inverse requires the decomposition of the phi matrix into three matrices using singular value decomposition (SVD) factorization as follows:
ϕ = U S V T
where ϕ and U are an orthogonal matrix and S is a diagonal matrix containing singular values. Therefore, the pseudo-inverse of matrix ϕ through singular values is calculated as:
ϕ + = V S + U T
where S + is obtained by replacing each non-zero singular value in S with its reciprocal. Using the above solution, the reconstruction of x can be achieved using following equation:
x = ( V S + U T ) y
Equation (13) may have an infinite number of solutions, but given the sparsity in x, the best solution could be obtained through l 0 norm minimization, as shown below:
min x 0 ,   s u b j e c t   t o   y = ϕ x
where 0 denotes the l 0 norm operation. In the case of partial reconstruction, the above equation can be rewritten as:
min x 1 ,   s u b j e c t   t o   y = ϕ x 2 γ t h
where 1 , 2 denotes the l 1 , l 2 norm operations, respectively, and γ t h denotes the termination residue threshold.

3.7. Orthogonal Matching Pursuit (OMP)

Solving the Equations (14) and (15) using the l 0 norm is an NP-hard problem; therefore, an alternative approach known as the orthogonal matching pursuit (OMP) algorithm [18] could be used. However, OMP has the following prerequisites:
  • The signal x needs to be a k-sparse, where k < < N .
  • The matrix ϕ is required to fulfill the condition M ϕ < 1 2 k 1 , where M ϕ denotes the mutual coherence of column vectors of matrix ϕ and is calculated as follows:
    M ϕ = max i j ϕ i ϕ j x i 2 x j 2
    where ϕ i and ϕ j denote the i t h and j t h columns of the ϕ matrix, respectively.
OMP is a greedy method that seeks to discover the solution for elements of x in one-by-one fashion through an iterative process [40]. The pseudocode for the OMP algorithm is presented in Algorithm 2.
Algorithm 2: OMP Algorithm
Inputs:
 Input signal vector x R N must be k-sparse.
/* sparse representation of speech frame using DCT */
 Sensing matrix ϕ R M × N , where M ϕ < 1 2 k 1 .
 Termination residue threshold γ t h . /* unwanted or noise components */
 Maximum number of iterations i t e r m a x .
Initialize:
i t e r   = 0 /* iteration counts */
y   = ϕ x /* measurement vector creation */
γ i t e r = y /* current residue. */
C i t e r = /*indices of dominant columns.*/
λ i t e r = /*contributions of dominant columns.*/
ϕ i t e r d o m i n a n t = /*dominent columns matrix. */
x e s t = 0 1 × N /* estimated filtered x . */
Main:
ϕ i n o r m = ϕ i ϕ i , where ϕ i c o l i ϕ /* columns normalization of sensing matrix */
While ( ( γ i t e r > γ t h )   o r   ( i t e r < i t e r m a x ) ) do:
   λ c t m p = argmax c γ i t e r ϕ i n o r m /* finds the column ( c ) with highest contribution ( λ c t m p ). */
  If ( c C i t e r ) then:/* verifying that the column indices in not the dominant column list */
    C i t e r = C i t e r 1 ,   c /* adding column indices in the list */
    λ i t e r = λ i t e r 1 ,   λ c t m p /* adding the contribution of the column */
    ϕ i t e r d o m i n a n t = ϕ i t e r 1 d o m i n a n t ,   ϕ c n o r m   /* update dominating column matrix */
  Else
    λ i t e r , c = λ i t e r 1 , c + λ c t m p
/* update dominating columns contributions only as column indices already exists */
  End If
   z i t e r = ϕ i t e r d o m i n a n t ϕ i t e r d o m i n a n t T ϕ i t e r d o m i n a n t + ϕ i t e r d o m i n a n t T y
/* finding the projection of the   y onto the dominant basis */
   γ i t e r   =   y     z i t e r /* remaining residue. */
   i t e r = i t e r + 1
End While
For i = 1 ; i c a r d C i t e r ;   i = i + 1 do:
  For j = 1 ; j c a r d C i t e r ;   j = j + 1 do:
    ϕ r e d u c e d i ,   j = ϕ i t e r d o m i n e n t : ,   i T ϕ i t e r d o m i n e n t : ,   j
/* generating reduced dimension sensing matrix through dominent columns */
  End For
End For
y r e d u c e d = ϕ d o m i n a n t T y T /* generating reduced dimension y */
x e s t r e d u c e d = ϕ r e d u c e d + y r e d u c e d /* estimation of reduced dimension x from y r e d u c e d */
For ( i = 1 ; i < = N ; i = i + 1 ) do:
   x e s t C i t e r i =   x e s t r e d u c e d i /* filling up the full dimension x e s t from x e s t r e d u c e d */
End For
Outputs:
C i t e r
λ i t e r
ϕ i t e r d o m i n a n t
x e s t

3.8. Dominant Columns Group Orthogonalization of Sensing Matrix (DCGOSM)

As described in Algorithm 2, the OMP algorithm collects the coefficients of the sensing matrix’s dominant columns while excluding the columns containing residue components.
Considering these characteristics of OMP, the orthogonalization of the dominant columns group of the sensing matrix for speech and noise signals can be achieved as follows:
  • Let x s and x n be the sparse representation of speech and noise signals, respectively.
  • If the sensing matrix is denoted by ϕ , then the compressed speech and noise signals can be obtained by y s = ϕ x s and y n = ϕ x n , respectively.
  • The OMP algorithm can be used to recover the enhanced speech signal x s from y s . The OMP finds the dominant columns, their contributions, and the reduced sensing matrix with only dominant columns (denoted by C i t e r , λ i t e r , and ϕ i t e r d o m i n e n t , respectively, in Algorithm 2) to estimate the enhanced speech signal while leaving the residue noise component columns intact.
  • If V s and V n are the reduced sensing matrix columns’ contributions to the speech and noise signals obtained through OMP, respectively, then the orthogonalization of dominant columns group of the sensing matrix can be achieved by optimizing the columns of the sensing matrix, such that:
    f o b j = m i n i m i z e i = 1 N s j = 1 N n V s : , i V n : , j  
    where N s and N n are the number of columns in matrix V s and V n , respectively. The optimization is performed to minimize the f o b j value to its maximum extent or up to a certain termination threshold.

3.9. Proposed DCGOSM-Based Speech Enhancement Technique

Figure 7 shows an illustrative block diagram of the proposed approach. Referring to this diagram, the proposed method can be divided into two parts. The first part is used to obtain the required sensing matrix, and the second part is used to enhance noisy speech.

3.9.1. Process of Obtaining the Required Sensing Matrix

The process of obtaining the required sensing matrix involves obtaining an optimized sensing matrix with orthogonalized dominant columns for clean speech and noise signals. The speech enhancement performed using greedy algorithms assumes that speech signals exhibit a structured temporal pattern and a relatively narrowband spectral content, whereas the noise signals lack these properties. Therefore, when reconstruction is performed, the noise components equally spread over all columns of the sensing matrix, whereas the speech components become concentrated over a small number of columns (dominant columns, see Section 3.8). Therefore, when reconstruction is performed through dominant columns, the amount of the noise significantly reduces in the reconstructed speech. However, such approaches have following limitations:
  • They assume that the noise components spread equally over all columns or noise must be white noise. However, for sounds other than white noise, this distribution of components will not be possible. Therefore, for sounds other than white noise, the performance of these approaches will naturally decrease.
  • Since the dominant columns are not known initially, they need to be searched every time from all columns, which increases processing time.
  • The algorithms have no awareness of speech and noise signals, and the selection of dominant columns is done solely on the basis of contribution. The algorithms always assume that higher contribution is due to speech components. Therefore, during higher noise conditions, the algorithms may select columns with dominating noise components. The condition may worsen in non-white noise cases.
To overcome these limitations, the proposed DCGOSM approach obtains the sensing matrix based on the type of noise present in the noisy speech. The obtained sensing matrix guarantees separate dominating columns for clean speech and noise components. During the generation of the required sensing matrix, the dominant columns for the speech components are also stored; hence, during the enhancement process, the iterations will be performed only through these columns. The algorithm identifies speech and non-speech frames and obtains the dominant columns separately for speech and non-speech frames. Therefore, the dominant columns selected for the speech components are guaranteed to contain only speech components. The effect of this can also be seen in the results presented in Table 1, Table 2 and Table 3, where the proposed DCGOSM gives better relative improvement at higher noise conditions (lower SNRs).
The process of obtaining the required sensing matrix involves the following steps:
  • In the first step, the clean speech sample is mixed with the noise using the mixer block, which adjusts the amplitude of the noise signal according to the given SNR and then adds it to the clean speech to generate the noisy speech sample.
  • The noisy speech sample is passed through the preprocessing block, which normalizes the signal. Normalization helps to reduce the impact of variations in recording conditions, microphone characteristics, and speaker differences. By normalizing the speech signals, the relative differences in amplitude caused by these factors are minimized, making the subsequent processing algorithms more robust and less sensitive to such variations.
  • Subsequently, the signal is divided into frames using overlapping Hamming windows. Considering the block-based processing nature of the CS, these windows are required to divide the speech signal into frames. This process also reduces the computational complexity of subsequent processing steps, and the analysis can be performed on smaller segments, reducing the amount of data to be processed and improving efficiency. The framing of the signal causes spectral leakage at the edges of the frame. The overlapping Hamming windows are used to mitigate spectral leakage by smoothly transitioning between adjacent frames, reducing the abrupt changes at the frame boundaries.
  • Signal sparsity is the fundamental condition for CS. Therefore, to make the frame sparse, each frame is converted into the frequency domain using the discrete cosine transform (DCT), then the insignificant DCT components within the transformed frame are set to zero. The term “significance” refers to the components that encompass the most energy, specifically those exceeding 90%. This objective can be accomplished by strategically selecting the highest 25% of the coefficients.
  • The non-transformed speech frames are sent to the VAD black for identification of the speech and non-speech frames. In addition, the preliminary noise variance ( v n ) estimation is performed using the initial frame, which is considered to carry only noise.
  • The noisy speech and non-speech frames are then grouped into separate databases for D B s p e e c h and D B n o n s p e e c h   . This helps in obtaining dominant columns separately for noisy speech and non-speech frames.
  • To minimize the optimization time, only a small number of frames are chosen from these databases. The selected noise frames are further used for the noise variance estimation ( v n f i n a l ), which is used during the final enhancement process. Estimating the noise variance from the multiple selected frames reduces the error possibilities, especially in the cases where the noise is not uniformly distributed throughout the entire signal duration.
  • PSO is initialized according to the number of particles and maximum generations. Each particle p i j (here, the superscript j represents the generation, and the subscript i represents the particle) is defined as an M × N dimension row vector, where each element is bounded within the range of [ 1 ,   1 ] .
    p i j = [ e 1 i ,   e 2 i ,   .   .   .   ,   e M × N i ] ,   w h e r e   i ,   j : { 1 e j i 1 }
  • Each of these particles ( p i j ) represents a sensing matrix ( ϕ i j ) when converted into an M rows and N columns matrix:
    ϕ i = e 1 i e 2 i e N i e N + 1 i e N + 2 i e 2 × N i e M 1 N + 1 e M 1 N + 2 e M × N i ,
    where
    i , j : { 1 e j i 1 }
  • However, before using this matrix as a sensing matrix, it is required to make ϕ i j a sample from the uniform spherical ensemble (USE). This is done by dividing each element of each column of the matrix by the column’s l 2 norm:
    ϕ i = e 1 i r 1 e 2 i r 2 e N i r N e N + 1 i r 1 e N + 2 i r 2 e 2 × N i r N e M 1 N + 1 r 1 e M 1 N + 2 r 2 e M × N i r N ,
    where
    r i = e 0 × N + j i ,   e 1 × N + j i , , e M 1 × N + j i 2 = k = 0 M 1 e k × N + j i 2
  • Using the sensing matrix ( ϕ i , U S E j ) formed by i t h particle p i j , the observation vector y s i = ϕ i , U S E j x s for speech, and y n i = ϕ i , U S E j x n for noise, frames are generated.
  • The OMP is applied to y s i and y n i independently to reconstruct x s and x n , respectively. However, our aim is not to find x s and x n , but to find the dominating basis vectors (columns) V s i and V n i in ϕ U S E i for y s i and y n i . The OMP uses v n as the termination threshold.
  • Our aim is to make ϕ i , U S E j a dominant columns group orthogonal sensing matrix where the columns used to represent the speech signals and noise signals are different. Therefore, speech signals can be extracted from the noisy samples using only the speech dominant columns group.
  • To match the above-mentioned condition, the ϕ i , U S E j matrix must satisfy the condition f o b j = 0 (Equation (17)).
  • However, the condition f o b j = 0 is only possible for the ideal cases. Therefore, the optimization is performed to minimize the f o b j value to its maximum extent or up to a certain termination threshold ( γ t h ) (Equation (18)).
  • f i , o b j j is calculated through ϕ j i , U S E for each particle p i j .
  • Once the f i , o b j j is calculated for each particle in the current generation, based on the f i , o b j j values obtained from all the generations ( 1 to j ), the p i , b e s t   (particle’s best),\ and g b e s t (global best) values are estimated.
  • Based on these estimations, the new positions of particles p i j + 1 are calculated as described in Equations (7) and (8).
  • PSO repeats the process until the proper ϕ i , U S E j matrix is generated.
  • After completion of this process, we obtain the sensing matrix ( ϕ i , U S E j ), final noise variance ( v n f i n a l ), dominating columns for speech signal components ( V s i ), and noisy-speech frames database ( D B s p e e c h ).

3.9.2. Process of Speech Enhancement Using the Obtained Sensing Matrix

  • Speech enhancement is performed only on the noisy speech frames separated by VAD during the generation of sensing matrix. Additionally, the enhancement is performed on a frame-by-frame basis using the obtained ϕ i , U S E j matrix, the final noise variance ( v n f i n a l ), and the OMP algorithm. However, because the dominating columns for speech signal components ( V s i ) are already known, the OMP is only iterated through these columns instead of all the columns of ϕ i n o r m .
  • The enhanced speech frames are then converted into the time domain by taking IDCT and finally converted into a single enhanced speech stream through the de-framing operation.

4. Performance Evaluation and Result Analysis

4.1. Sensing Matrix Optimization

The idea behind the proposed approach is to isolate the components of speech and noise signals into different columns of a sensing matrix. The degree of isolation is reflected by the fitness function of the optimization algorithm, where a lower value reflects a higher level of isolation. To validate this concept, we have shown the column contributions of the sensing matrix for speech and noise frames before (Figure 8) and after (Figure 9) optimization. The convergence plot of the optimization process is also shown (Figure 10) to link the fitness function and the level of isolation.
The convergence plot shows that, when the optimization begins, the average and best fitness values were 0.64 and 0.52, respectively, which resulted in highly overlapping column contributions, whereas at the end of the optimization, the average and best fitness values were 0.27 and 0.25, respectively, which resulted in well-separated column contributions. Figure 9 also shows that after optimization, 95% of the total signal energy was concentrated in half the number of columns.

4.2. Speech Quality Analysis

The proposed DCGOSM-based speech enhancement technique was compared with four baseline methods: OMP [18], CoSaMP [19], StOMP [38], and the K-SVD-based CS technique (K-SVDCS) [63] (note: this technique was only compared for babble noise, as the paper only provided the results for this noise). Additionally, it was compared with some other methods based on human perception-related objective functions, such as DNN-MSE [64], DNN-PMSEQ [65], BLSTM-MSE [64], BLSTM-PMSQE [65], and DNN-Quality-Net [64].
To evaluate the performance of the proposed algorithm, the NOISEX-92 [66] and NOIZEUS [67] datasets were used. The clean speech samples were taken from the NOIZEUS dataset, which contains 30 clean speech files (sp01.wav to sp30.wav) with a single channel (mono), 16-bit PCM, sampled at 8 kHz, and encoded in the Windows “wav” format. The noise samples were taken from the NOISEX-92, which contains white noise, babble, and f-16 noise files with a single channel sampled using a 16-bit analog-to-digital converter (A/D) at a sampling rate of 19.98 kHz. These samples were encoded in the MATLAB “mat” format. However, for the experimental analysis, we downsampled all signals to 8 kHz.
The clean speech signal was mixed with one of the noise signals at a time to evaluate the performance of the proposed technique. The amplitude of the noise was modified before mixing according to the required SNR. The noisy speech was then processed through the enhancement algorithm to obtain the enhanced speech. This process is depicted in Figure 11, which shows the time domain plots of the “sp05.wav” clean speech sample, the “babble” noise, and the noisy speech signal generated by corrupting the clean speech signal with babble noise, and their respective spectrogram plots (Figure 11b,d,f). The blue bars in the spectrogram of the enhanced speech (Figure 11f) depict the events where VAD detected non-speech components. These bars were mostly synchronized with the silence events in the clean speech (Figure 11a), which demonstrate that the VAD is capable of detecting non-speech signals even in the presence of noise (Figure 11c).
Finally, to compare the performance of each algorithm, four objective measures, the signal-to-noise ratio (SNR), segmental SNR (SSNR), short-time objective intelligibility (STOI), and perceptual evaluation of speech quality (PESQ), were utilized.
The performance comparison of the proposed technique with different variants of the OMP algorithm for white, babble, and f-16 noises is presented in Table 1, Table 2 and Table 3, respectively. These results demonstrate the superior performance of the proposed technique over competitors for SNR, SSNR, and PESQ, irrespective of the noise type and magnitude. However, for STOI, the proposed algorithm performed better under low SNR conditions.
A further in-depth analysis of these tables showed that the proposed technique achieved the highest SNR improvement of 42.54% for f-16 noise, the second-highest improvement of 21.56% for babble noise, and the third-highest improvement of 11.03% for white noise at 0 dB of SNR. The minimum improvement of 0.41% was achieved for babble noise at 20 dB. Similar behavior was observed for SSNR; however, at 0 dB of SNR, the higher improvements were 62.97%, 28.80%, and 26.84% for f-16, babble, and white noise, respectively. White noise achieved the lowest improvement of 2.77% at 20 dB.
In comparison to SNR and SSNR, the proposed algorithm showed different behavior for PESQ and achieved the three highest improvements for white noise at 0 dB, 5 dB, and 10 dB SNR, which were 27.48%, 20.53%, and 20.31%, respectively. The fourth-highest improvement of 18.64% was achieved for f-16 noise, followed by 17.01% for babble noise, both at 0 dB SNR.
For STOI, the only significant improvement of 8.72% was achieved for white noise at 0 dB SNR, followed by 2.73% for f-16 at 0 dB for SNR.
It is also worth mentioning that for every noise, the proposed algorithm always achieved the best SNR, SSNR, and PESQ, except for the babble noise at 0 dB SNR, where the K-SVDCS achieved the best SSNR.
To further validate the performance of the proposed algorithm, a comparison with non-OMP (non-CS) algorithms for PESQ and STOI measures is presented in Table 4 and Table 5, respectively. These results show that the proposed technique performed better at all SNRs except 18 dB, where DNN-Quality-Net improved by 3.96%. However, the proposed algorithm achieved the highest PESQ improvement of 20.32% at −6 dB. For STOI, the BLSTM (MSE) performed better for 18 dB, 12 dB, and 6 dB, while for 0 dB and −6 dB, the proposed technique performed better and achieved a maximum improvement of 8.29% at −6 dB SNR.

4.3. Recovery Time Analysis

A reconstruction time comparison of the proposed algorithm with the different OMP-based reconstruction algorithms is shown in Table 6. The results show that the proposed algorithm consistently outperformed the other three algorithms in terms of execution time across all the configurations. In Configuration ( 8,32,64 ), the proposed algorithm achieved an execution time of 0.3261, which was lower than the execution times of OMP (0.7956), CoSaMP (0.8071), and StOMP (0.3439). Similarly, in Configuration ( 16,64,128 ), the proposed algorithm (0.3527) demonstrated a lower reconstruction time compared to OMP (0.9434), CoSaMP (0.7652), and StOMP (0.3694). This trend continued in Configurations ( 32,128,256 ) and ( 64,256,512 ), where the proposed algorithm consistently exhibited the lowest reconstruction time among the four algorithms, followed by the StOMP as the second best, while the OMP took the longest time. In terms of percentage improvements, the proposed algorithm achieved a maximum reduction of 70% in reconstruction time for the single column selection algorithm (OMP), while the maximum improvements over the multiple column selection algorithms were 59% and 13% for CoSaMP and STOMP, respectively. The results also show that the reconstruction time increased with the sensing matrix size.

5. Conclusions

This paper proposes a compressive sensing-based speech enhancement technique that uses the dominant columns group orthogonalization of the sensing matrix (DCGOSM). The proposed method attempts to find the sensing matrix in such a way that each column can either contain the noise component or the clean speech component, which is obtained by calculating the dot product of column contributions for the noise and speech samples. The proposed technique greatly reduces the speech recovery time by iterating only through the speech-dominating columns of the sensing matrix, as opposed to other OMP-based techniques, which require iterations through all columns of the sensing matrix. Another advantage of the proposed technique is that it provides uniform improvement for different quality measures, unlike the other sensing matrix optimization techniques that only improve the measures defined in the objective function. In this work, the PSO is adopted for the optimization of the sensing matrix. However, in the future, other optimization algorithms can also be tested. Furthermore, the OMP can be replaced with multiple column selection approaches such as CoSaMP and STOMP.

Author Contributions

Methodology, V.S.; Software, V.S.; Validation, V.S.; Formal analysis, P.D.S.; Supervision, P.D.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The datasets used are publicly available.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Donoho, D.L. Compressed Sensing. IEEE Trans. Inf. Theory 2006, 52, 1289–1306. [Google Scholar] [CrossRef]
  2. Candes, E.J.; Romberg, J.; Tao, T. Robust Uncertainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information. IEEE Trans. Inf. Theory 2006, 52, 489–509. [Google Scholar] [CrossRef] [Green Version]
  3. Rani, M.; Dhok, S.B.; Deshmukh, R.B. A Systematic Review of Compressive Sensing: Concepts, Implementations and Applications. IEEE Access 2018, 6, 4875–4894. [Google Scholar] [CrossRef]
  4. Xia, K.; Pan, Z.; Mao, P. Video Compressive Sensing Reconstruction Using Unfolded LSTM. Sensors 2022, 22, 7172. [Google Scholar] [CrossRef]
  5. Vanjari, H.B.; Kolte, M.T. Comparative Analysis of Speech Enhancement Techniques in Perceptive of Hearing Aid Design. In Proceedings of the Third International Conference on Information Management and Machine Intelligence, Jaipur, India, 23–24 December 2021; Springer Nature: Singapore, 2023; pp. 117–125. [Google Scholar]
  6. Calisesi, G.; Ghezzi, A.; Ancora, D.; D’Andrea, C.; Valentini, G.; Farina, A.; Bassi, A. Compressed Sensing in Fluorescence Microscopy. Prog. Biophys. Mol. Biol. 2022, 168, 66–80. [Google Scholar] [CrossRef]
  7. Kwon, H.-M.; Hong, S.-P.; Kang, M.; Seo, J. Data Traffic Reduction with Compressed Sensing in an AIoT System. Comput. Mater. Contin. 2022, 70, 1769–1780. [Google Scholar] [CrossRef]
  8. Shannon, C.E. Communication in the Presence of Noise. Proc. IRE 1949, 37, 10–21. [Google Scholar] [CrossRef]
  9. Donoho, D.L.; Stark, P.B. Uncertainty Principles and Signal Recovery. SIAM J. Appl. Math. 1989, 49, 906–931. [Google Scholar] [CrossRef] [Green Version]
  10. Candès, E.J.; Romberg, J.K.; Tao, T. Stable Signal Recovery from Incomplete and Inaccurate Measurements. Commun. Pure Appl. Math. 2006, 59, 1207–1223. [Google Scholar] [CrossRef] [Green Version]
  11. Amini, A.; Marvasti, F. Deterministic Construction of Binary, Bipolar, and Ternary Compressed Sensing Matrices. IEEE Trans. Inf. Theory 2011, 57, 2360–2370. [Google Scholar] [CrossRef] [Green Version]
  12. Ben-Haim, Z.; Eldar, Y.C.; Elad, M. Coherence-Based Performance Guarantees for Estimating a Sparse Vector Under Random Noise. IEEE Trans. Signal Process. 2010, 58, 5030–5043. [Google Scholar] [CrossRef] [Green Version]
  13. Baraniuk, R.; Davenport, M.; DeVore, R.; Wakin, M. A Simple Proof of the Restricted Isometry Property for Random Matrices. Constr. Approx. 2008, 28, 253–263. [Google Scholar] [CrossRef] [Green Version]
  14. Abrol, V.; Sharma, P.; Budhiraja, S. Evaluating Performance of Compressed Sensing for Speech Signals. In Proceedings of the 2013 3rd IEEE International Advance Computing Conference (IACC), Ghaziabad, India, 22–23 February 2013; pp. 1159–1164. [Google Scholar]
  15. Xu, S.-F.; Chen, X.-B. Speech Signal Acquisition Methods Based on Compressive Sensing. In Systems and Computer Technology; CRC Press: Boca Raton, FL, USA, 2015; ISBN 978-0-429-22578-9. [Google Scholar]
  16. Swami, P.D.; Sharma, R.; Jain, A.; Swami, D.K. Speech Enhancement by Noise Driven Adaptation of Perceptual Scales and Thresholds of Continuous Wavelet Transform Coefficients. Speech Commun. 2015, 70, 1–12. [Google Scholar] [CrossRef]
  17. Donoho, D.L. For Most Large Underdetermined Systems of Linear Equations the Minimal ℓ1-Norm Solution Is Also the Sparsest Solution. Commun. Pure Appl. Math. 2006, 59, 797–829. [Google Scholar] [CrossRef]
  18. Yang, H.; Hao, D.; Sun, H.; Liu, Y. Speech Enhancement Using Orthogonal Matching Pursuit Algorithm. In Proceedings of the 2014 International Conference on Orange Technologies, Xi’an, China, 20–23 September 2014; pp. 101–104. [Google Scholar]
  19. Needell, D.; Tropp, J.A. CoSaMP: Iterative Signal Recovery from Incomplete and Inaccurate Samples. Appl. Comput. Harmon. Anal. 2009, 26, 301–321. [Google Scholar] [CrossRef] [Green Version]
  20. Pilastri, A.L.; Tavares, J.M.R. Reconstruction Algorithms in Compressive Sensing: An Overview. In Proceedings of the 11th edition of the Doctoral Symposium in Informatics Engineering (DSIE-16), Porto, Portugal, 3 February 2016. [Google Scholar]
  21. Firouzeh, F.F.; Ghorshi, S.; Salsabili, S. Compressed Sensing Based Speech Enhancement. In Proceedings of the 2014 8th International Conference on Signal Processing and Communication Systems (ICSPCS), Gold Coast, Australia, 15–17 December 2014; pp. 1–6. [Google Scholar]
  22. Wu, D.; Zhu, W.-P.; Swamy, M.N.S. Compressive Sensing-Based Speech Enhancement in Non-Sparse Noisy Environments. IET Signal Process. 2013, 7, 450–457. [Google Scholar] [CrossRef]
  23. Gemmeke, J.F.; Van Hamme, H.; Cranen, B.; Boves, L. Compressive Sensing for Missing Data Imputation in Noise Robust Speech Recognition. IEEE J. Sel. Top. Signal Process. 2010, 4, 272–287. [Google Scholar] [CrossRef] [Green Version]
  24. Wu, D.; Zhu, W.-P.; Swamy, M.N.S. The Theory of Compressive Sensing Matching Pursuit Considering Time-Domain Noise with Application to Speech Enhancement. IEEE/ACM Trans. Audio Speech Lang. Process. 2014, 22, 682–696. [Google Scholar] [CrossRef]
  25. Quackenbush, S.R.; Barnwell, T.P.; Clements, M.A. Objective Measures of Speech Quality; Prentice Hall: Upper Saddle River, NJ, USA, 1988; ISBN 978-0-13-629056-8. [Google Scholar]
  26. ITU-T Recommendation Database. Available online: https://www.itu.int/ITU-T/recommendations/rec.aspx?rec=14949&lang=en (accessed on 7 June 2023).
  27. Speech Coding and Synthesis; Kleijn, W.B.; Paliwal, K.K. (Eds.) Elsevier: Amsterdam, The Netherlands; New York, NY, USA, 1995; ISBN 978-0-444-82169-0. [Google Scholar]
  28. Hu, Y.; Loizou, P.C. Evaluation of Objective Quality Measures for Speech Enhancement. IEEE Trans. Audio Speech Lang. Process. 2008, 16, 229–238. [Google Scholar] [CrossRef]
  29. Tribolet, J.; Noll, P.; McDermott, B.; Crochiere, R. A Study of Complexity and Quality of Speech Waveform Coders. In Proceedings of the ICASSP’78. IEEE International Conference on Acoustics, Speech, and Signal Processing, Tulsa, OK, USA, 10–12 April 1978; Volume 3, pp. 586–590. [Google Scholar]
  30. Hansen, J.H.L.; Pellom, B.L. An Effective Quality Evaluation Protocol for Speech Enhancement Algorithms. In Proceedings of the 5th International Conference on Spoken Language Processing (ICSLP 1998), Sydney, Australia, 30 November 1998; ISCA; p. 0917-0. [Google Scholar]
  31. Klatt, D. Prediction of Perceived Phonetic Distance from Critical-Band Spectra: A First Step. In Proceedings of the ICASSP’82. IEEE International Conference on Acoustics, Speech, and Signal Processing, Paris, France, 3–5 May 1982; Volume 7, pp. 1278–1281. [Google Scholar]
  32. P.862: Perceptual Evaluation of Speech Quality (PESQ): An Objective Method for End-to-End Speech Quality Assessment of Narrow-Band Telephone Networks and Speech Codecs. Available online: https://www.itu.int/rec/T-REC-P.862 (accessed on 7 June 2023).
  33. P.862.3: Application Guide for Objective Quality Measurement Based on Recommendations P.862, P.862.1 and P.862.2. Available online: https://www.itu.int/rec/T-REC-P.862.3/_page.print (accessed on 7 June 2023).
  34. Crespo Marques, E.; Maciel, N.; Naviner, L.; Cai, H.; Yang, J. A Review of Sparse Recovery Algorithms. IEEE Access 2019, 7, 1300–1322. [Google Scholar] [CrossRef]
  35. Mallat, S.G.; Zhang, Z. Matching Pursuits with Time-Frequency Dictionaries. IEEE Trans. Signal Process. 1993, 41, 3397–3415. [Google Scholar] [CrossRef] [Green Version]
  36. de Paiva, N.M.; Marques, E.C.; de Barros Naviner, L.A. Sparsity Analysis Using a Mixed Approach with Greedy and LS Algorithms on Channel Estimation. In Proceedings of the 2017 3rd International Conference on Frontiers of Signal Processing (ICFSP), Paris, France, 6–8 September 2017; pp. 91–95. [Google Scholar]
  37. Dai, W.; Milenkovic, O. Subspace Pursuit for Compressive Sensing Signal Reconstruction. IEEE Trans. Inf. Theory 2009, 55, 2230–2249. [Google Scholar] [CrossRef] [Green Version]
  38. Donoho, D.L.; Tsaig, Y.; Drori, I.; Starck, J.-L. Sparse Solution of Underdetermined Systems of Linear Equations by Stagewise Orthogonal Matching Pursuit. IEEE Trans. Inf. Theory 2012, 58, 1094–1121. [Google Scholar] [CrossRef]
  39. Needell, D.; Vershynin, R. Uniform Uncertainty Principle and Signal Recovery via Regularized Orthogonal Matching Pursuit. Found. Comput. Math. 2009, 9, 317–334. [Google Scholar] [CrossRef]
  40. Wang, J.; Kwon, S.; Shim, B. Generalized Orthogonal Matching Pursuit. IEEE Trans. Signal Process. 2012, 60, 6202–6216. [Google Scholar] [CrossRef] [Green Version]
  41. Sun, H.; Ni, L. Compressed Sensing Data Reconstruction Using Adaptive Generalized Orthogonal Matching Pursuit Algorithm. In Proceedings of the 2013 3rd International Conference on Computer Science and Network Technology, Dalian, China, 12–13 October 2013; pp. 1102–1106. [Google Scholar]
  42. Bi, X.; Leng, L.; Kim, C.; Liu, X.; Du, Y.; Liu, F. Constrained Backtracking Matching Pursuit Algorithm for Image Reconstruction in Compressed Sensing. Appl. Sci. 2021, 11, 1435. [Google Scholar] [CrossRef]
  43. GBRAMP: A Generalized Backtracking Regularized Adaptive Matching Pursuit Algorithm for Signal Reconstruction—ScienceDirect. Available online: https://www.sciencedirect.com/science/article/abs/pii/S0045790621001907 (accessed on 7 June 2023).
  44. Zhang, C. An Orthogonal Matching Pursuit Algorithm Based on Singular Value Decomposition. Circuits Syst. Signal Process. 2020, 39, 492–501. [Google Scholar] [CrossRef]
  45. Das, S.; Mandal, J.K. An Enhanced Block-Based Compressed Sensing Technique Using Orthogonal Matching Pursuit. Signal Image Video Process. 2021, 15, 563–570. [Google Scholar] [CrossRef]
  46. Blumensath, T.; Davies, M.E. Gradient Pursuits. IEEE Trans. Signal Process. 2008, 56, 2370–2382. [Google Scholar] [CrossRef]
  47. Kwon, S.; Wang, J.; Shim, B. Multipath Matching Pursuit. IEEE Trans. Inf. Theory 2014, 60, 2986–3001. [Google Scholar] [CrossRef] [Green Version]
  48. Elad, M. Optimized Projections for Compressed Sensing. IEEE Trans. Signal Process. 2007, 55, 5695–5702. [Google Scholar] [CrossRef]
  49. Singh, R.; Bhattacharjee, U.; Singh, A.K. Performance Evaluation of Normalization Techniques in Adverse Conditions. Procedia Comput. Sci. 2020, 171, 1581–1590. [Google Scholar] [CrossRef]
  50. Blinn, J.F. What’s That Deal with the DCT? IEEE Comput. Graph. Appl. 1993, 13, 78–83. [Google Scholar] [CrossRef]
  51. Stanković, L.; Brajović, M. Analysis of the Reconstruction of Sparse Signals in the DCT Domain Applied to Audio Signals. IEEE/ACM Trans. Audio Speech Lang. Process. 2018, 26, 1220–1235. [Google Scholar] [CrossRef]
  52. Reznik, Y.A. Relationship between DCT-II, DCT-VI, and DST-VII Transforms. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 5642–5646. [Google Scholar]
  53. Prabhu, K.M.M. Window Functions and Their Applications in Signal Processing; Taylor & Francis: Oxford, UK, 2014; ISBN 978-1-4665-1584-0. [Google Scholar]
  54. Shukla, V.; Swami, P.D. Speech Enhancement Using VAD for Noise Estimation in Compressive Sensing. In Proceedings of the Data, Engineering and Applications; Sharma, S., Peng, S.-L., Agrawal, J., Shukla, R.K., Le, D.-N., Eds.; Springer Nature: Singapore, 2022; pp. 357–369. [Google Scholar]
  55. Lokesh, S.; Devi, M.R. Speech Recognition System Using Enhanced Mel Frequency Cepstral Coefficient with Windowing and Framing Method. Clust. Comput 2019, 22, 11669–11679. [Google Scholar] [CrossRef]
  56. Van Segbroeck, M.; Tsiartas, A.; Narayanan, S.S. A Robust Frontend for VAD: Exploiting Contextual, Discriminative and Spectral Cues of Human Voice. In Proceedings of the INTERSPEECH, Lyon, France, 25–29 August 2013; pp. 704–708. [Google Scholar]
  57. Kennedy, J.; Eberhart, R. Particle Swarm Optimization. In Proceedings of the ICNN’95—International Conference on Neural Networks, Perth, Australia, 27 November–1 December 1995; Volume 4, pp. 1942–1948. [Google Scholar]
  58. Candes, E.J.; Wakin, M.B. An Introduction To Compressive Sampling. IEEE Signal Process. Mag. 2008, 25, 21–30. [Google Scholar] [CrossRef]
  59. Jerri, A.J. The Shannon Sampling Theorem—Its Various Extensions and Applications: A Tutorial Review. Proc. IEEE 1977, 65, 1565–1596. [Google Scholar] [CrossRef]
  60. Yuan, X.; Haimi-Cohen, R. Image Compression Based on Compressive Sensing: End-to-End Comparison with JPEG 2020. IEEE Trans. Multimed. 2020, 22, 2889–2904. [Google Scholar] [CrossRef] [Green Version]
  61. Fira, M.; Costin, H.-N.; Goraș, L. A Study on Dictionary Selection in Compressive Sensing for ECG Signals Compression and Classification. Biosensors 2022, 12, 146. [Google Scholar] [CrossRef]
  62. Golub, G.; Kahan, W. Calculating the Singular Values and Pseudo-Inverse of a Matrix. J. Soc. Ind. Appl. Math. Ser. B Numer. Anal. 1965, 2, 205–224. [Google Scholar] [CrossRef]
  63. Haneche, H.; Boudraa, B.; Ouahabi, A. A New Way to Enhance Speech Signal Based on Compressed Sensing. Measurement 2020, 151, 107117. [Google Scholar] [CrossRef]
  64. Fu, S.-W.; Liao, C.-F.; Tsao, Y. Learning With Learned Loss Function: Speech Enhancement With Quality-Net to Improve Perceptual Evaluation of Speech Quality. IEEE Signal Process. Lett. 2020, 27, 26–30. [Google Scholar] [CrossRef] [Green Version]
  65. Martin-Doñas, J.M.; Gomez, A.M.; Gonzalez, J.A.; Peinado, A.M. A Deep Learning Loss Function Based on the Perceptual Evaluation of the Speech Quality. IEEE Signal Process. Lett. 2018, 25, 1680–1684. [Google Scholar] [CrossRef]
  66. Varga, A.; Steeneken, H.J.M. Assessment for Automatic Speech Recognition: II. NOISEX-92: A Database and an Experiment to Study the Effect of Additive Noise on Speech Recognition Systems. Speech Commun. 1993, 12, 247–251. [Google Scholar] [CrossRef]
  67. Hu, Y.; Loizou, P.C. Subjective Comparison and Evaluation of Speech Enhancement Algorithms. Speech Commun. 2007, 49, 588–601. [Google Scholar] [CrossRef] [Green Version]
Figure 1. Sparse recovery algorithms.
Figure 1. Sparse recovery algorithms.
Applsci 13 08954 g001
Figure 2. Flow chart for greedy algorithms.
Figure 2. Flow chart for greedy algorithms.
Applsci 13 08954 g002
Figure 3. Illustration of the process of generating a sparse signal representation using the DCT. Where plot (a) shows the sampled speech signal, plot (b) shows the amplitudes of its DCT coefficients, plot (c) shows DCT coefficients arranged in descending order, and plot (d) shows the percentage of total energy contained by the largest K coefficients, where K takes values of 5, 10, 15, 20, 25, and 30.
Figure 3. Illustration of the process of generating a sparse signal representation using the DCT. Where plot (a) shows the sampled speech signal, plot (b) shows the amplitudes of its DCT coefficients, plot (c) shows DCT coefficients arranged in descending order, and plot (d) shows the percentage of total energy contained by the largest K coefficients, where K takes values of 5, 10, 15, 20, 25, and 30.
Applsci 13 08954 g003
Figure 4. Framing process and frame overlapping, where blue and red dashed lines represent the starting and ending of the frames, respectively.
Figure 4. Framing process and frame overlapping, where blue and red dashed lines represent the starting and ending of the frames, respectively.
Applsci 13 08954 g004
Figure 5. Overview of preprocessing steps used in VAD [24].
Figure 5. Overview of preprocessing steps used in VAD [24].
Applsci 13 08954 g005
Figure 6. Particle velocity update procedure in PSO.
Figure 6. Particle velocity update procedure in PSO.
Applsci 13 08954 g006
Figure 7. Block diagram of the proposed algorithm.
Figure 7. Block diagram of the proposed algorithm.
Applsci 13 08954 g007
Figure 8. Column contributions for speech components (blue bars) and noise components (red bars) in the non-optimized sensing matrix.
Figure 8. Column contributions for speech components (blue bars) and noise components (red bars) in the non-optimized sensing matrix.
Applsci 13 08954 g008
Figure 9. Column contributions for speech components (blue bars) and noise components (red bars) in the optimized sensing matrix.
Figure 9. Column contributions for speech components (blue bars) and noise components (red bars) in the optimized sensing matrix.
Applsci 13 08954 g009
Figure 10. Sensing matrix optimization convergence plot showing global best (red) and average (blue) fitness value for each iteration.
Figure 10. Sensing matrix optimization convergence plot showing global best (red) and average (blue) fitness value for each iteration.
Applsci 13 08954 g010
Figure 11. Plots for (a) clean speech, (c) noisy speech with 5 dB SNR corrupted by babble noise, (e) speech samples enhanced by the proposed algorithm with their respective spectrogram plots in (b,d,f).
Figure 11. Plots for (a) clean speech, (c) noisy speech with 5 dB SNR corrupted by babble noise, (e) speech samples enhanced by the proposed algorithm with their respective spectrogram plots in (b,d,f).
Applsci 13 08954 g011
Table 1. Performance comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Table 1. Performance comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Techniques
OMPCoSaMPStOMPProposed
SNR (dB)SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
05.744.6313.1132.8144.4173.5816.3805.874
59.1248.0497.4816.3288.3927.1649.6279.157
1012.33911.38511.54210.22111.91610.77312.63912.298
1514.86214.10114.65813.52714.69613.80714.99314.745
2016.26615.94116.26715.76516.29915.94416.39016.386
SNR (dB)PESQSTOIPESQSTOIPESQSTOIPESQSTOI
01.8010.6881.7590.6341.7960.6082.2960.748
52.3190.7642.0690.6892.1730.6992.7950.731
102.5700.7932.4030.7342.5090.7683.0920.809
152.7590.8372.6110.7762.7080.8173.2010.825
202.9570.8772.7600.8242.8740.8693.2200.862
The bold font indicates the best obtained result for the selected measure in that row.
Table 2. Performance comparison for babble noise using the NOIZEUS dataset sample sp05.wav.
Table 2. Performance comparison for babble noise using the NOIZEUS dataset sample sp05.wav.
Techniques
OMPCoSaMPStOMPK-SVDCSProposed
SNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
02.5602.3850.7581.6381.3881.853--3.803.1123.072
56.2015.8355.1934.6025.6814.965--3.506.6736.592
1010.1199.3459.3498.4719.5328.708--3.0610.1769.885
1513.11812.43813.11912.19513.07412.168--1.6013.36113.056
2015.38314.87315.45314.89015.15614.685---0.9415.51615.325
SNR (dB)PESQSTOIPESQSTOIPESQSTOIPESQSTOIPESQSTOI
01.7800.652l.9170.6831.9690.6441.960.662.3040.615
52.2160.7362.2510.7262.3080.7152.280.722.3840.689
102.4340.8112.5130.7592.5130.7712.520.792.6740.748
152.6060.8272.6460.7742.7090.8052.690.812.9030.785
202.6670.8592.7720.8162.8530.8472.850.833.1940.824
The bold font indicates the best obtained result for the selected measure in that row.
Table 3. Performance comparison for f-16 noise using the NOIZEUS dataset sample sp05.wav.
Table 3. Performance comparison for f-16 noise using the NOIZEUS dataset sample sp05.wav.
Techniques
OMPCoSaMPStOMPProposed
SNR (dB)SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
SNR
(dB)
SSNR
(dB)
03.2882.490l.292l.7442.0562.0364.6874.058
57.2206.1015.6394.6566.2665.1297.9267.454
1010.6899.6319.8448.44510.2648.90111.20610.748
1513.77212.80313.59912.35813.51712.36614.03713.602
2015.68615.09515.71114.93315.62914.94915.82115.599
SNR (dB)PESQSTOIPESQSTOIPESQSTOIPESQSTOI
01.6840.6591.8870.6291.9420.5912.3040.677
52.1010.7512.1830.6822.2680.6802.3840.685
102.5090.7872.4960.7452.5770.7562.6740.772
152.6540.8362.6320.7562.7470.8082.9030.791
202.8380.8622.7470.7962.8780.8393.1940.808
The bold font indicates the best obtained result for the selected measure in that row.
Table 4. PESQ comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Table 4. PESQ comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Techniques
SNR
(dB)
DNN
(MSE) [64]
DNN-
PMSQE [65]
BLSTM
(MSE) [64]
BLSTM
(PMSQE) [65]
DNN-
Quality-Net [64]
Proposed
182.8103.0823.2872.8993.3773.243
122.5762.8192.9082.7773.0103.024
62.2752.4972.5042.5782.6142.777
01.9122.1112.0652.2612.1712.337
−61.5301.7111.5691.8651.6712.244
The bold font indicates the best obtained result for the selected measure in that row.
Table 5. STOI comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Table 5. STOI comparison for white noise using the NOIZEUS dataset sample sp05.wav.
Techniques
SNR
(dB)
DNN (MSE) [64]DNN PMSQE [65]BLSTM (MSE) [64]BLSTM (PMSQE) [65]DNN-Quality-Net [64]Proposed
180.8550.8860.9720.8820.9660.842
120.8310.8650.9420.8670.9370.821
60.7880.8220.8850.8360.8820.815
00.7150.7410.7960.7730.7940.807
−60.6040.6150.6630.6670.6630.718
The bold font indicates the best obtained result for the selected measure in that row.
Table 6. Reconstruction time comparison (in seconds).
Table 6. Reconstruction time comparison (in seconds).
Techniques
Configuration   ( k , m , n ) OMPCoSaMPStOMPProposed
(8,32,64)0.79560.80710.34390.3261 *
(16,64,128)0.94340.76520.36940.3527
(32,128,256)1.3928l.20970.46370.4165
(64,256,512)2.80482.04870.98450.8542
The bold font indicates the best obtained result for the selected measure in that row. * k = sparsity, m = rows in sensing matrix, n   = columns in sensing matrix.
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Shukla, V.; Swami, P.D. Orthogonalization of the Sensing Matrix Through Dominant Columns in Compressive Sensing for Speech Enhancement. Appl. Sci. 2023, 13, 8954. https://doi.org/10.3390/app13158954

AMA Style

Shukla V, Swami PD. Orthogonalization of the Sensing Matrix Through Dominant Columns in Compressive Sensing for Speech Enhancement. Applied Sciences. 2023; 13(15):8954. https://doi.org/10.3390/app13158954

Chicago/Turabian Style

Shukla, Vasundhara, and Preety D. Swami. 2023. "Orthogonalization of the Sensing Matrix Through Dominant Columns in Compressive Sensing for Speech Enhancement" Applied Sciences 13, no. 15: 8954. https://doi.org/10.3390/app13158954

APA Style

Shukla, V., & Swami, P. D. (2023). Orthogonalization of the Sensing Matrix Through Dominant Columns in Compressive Sensing for Speech Enhancement. Applied Sciences, 13(15), 8954. https://doi.org/10.3390/app13158954

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop