Next Article in Journal
Initial Solution Generation and Diversified Variable Picking in Local Search for (Weighted) Partial MaxSAT
Previous Article in Journal
A Finite Element Approximation for Nematic Liquid Crystal Flow with Stretching Effect Based on Nonincremental Pressure-Correction Method
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Newton Recursion Based Random Data-Reusing Generalized Maximum Correntropy Criterion Adaptive Filtering Algorithm

1
School of Information Engineering, Southwest University of Science and Technology, Mianyang 621010, China
2
School of Power and Energy, Northwestern Polytechnical University, Xi’an 710072, China
3
Science and Technology on Altitude Simulation Laboratory, Mianyang 621700, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(12), 1845; https://doi.org/10.3390/e24121845
Submission received: 2 October 2022 / Revised: 30 November 2022 / Accepted: 13 December 2022 / Published: 18 December 2022
(This article belongs to the Section Signal and Data Analysis)

Abstract

:
For system identification under impulsive-noise environments, the gradient-based generalized maximum correntropy criterion (GB-GMCC) algorithm can achieve a desirable filtering performance. However, the gradient method only uses the information of the first-order derivative, and the corresponding stagnation point of the method can be a maximum point, a minimum point or a saddle point, and thus the gradient method may not always be a good selection. Furthermore, GB-GMCC merely uses the current input signal to update the weight vector; facing the highly correlated input signal, the convergence rate of GB-GMCC will be dramatically damaged. To overcome these problems, based on the Newton recursion method and the data-reusing method, this paper proposes a robust adaptive filtering algorithm, which is called the Newton recursion-based data-reusing GMCC (NR-DR-GMCC). On the one hand, based on the Newton recursion method, NR-DR-GMCC can use the information of the second-order derivative to update the weight vector. On the other hand, by using the data-reusing method, our proposal uses the information of the latest M input vectors to improve the convergence performance of GB-GMCC. In addition, to further enhance the filtering performance of NR-DR-GMCC, a random strategy can be used to extract more information from the past M input vectors, and thus we obtain an enhanced NR-DR-GMCC algorithm, which is called the Newton recursion-based random data-reusing GMCC (NR-RDR-GMCC) algorithm. Compared with existing algorithms, simulation results under system identification and acoustic echo cancellation are conducted and validate that NR-RDR-GMCC can provide a better filtering performance in terms of filtering accuracy and convergence rate.

1. Introduction

Adaptive filters have been widely used in different engineering fields. Typical applications include system identification, active noise control, acoustic echo cancellation and channel equalization [1,2,3,4]. The adaptive filtering algorithms based on a mean square error (MSE) criterion mainly include the family of least mean square (LMS) algorithm, the tribe of affine projection (AP) algorithm and the class of least square algorithm [5,6,7]. Among them, the LMS algorithm is widely used because of its simple structure, low computational complexity, and good convergence in smooth environments. However, the above-mentioned MSE-based algorithms will have obvious filtering performance degradation in a non-Gaussian environment, especially in a noise environment with heavy-tailed distribution in which the typical heavy-tailed distribution noises include Laplace noise, Cauchy noise, mixed Gaussian noise, and alpha-stable distribution ( α -SD) noise [8].
In order to enhance the ability of an adaptive filtering algorithm to suppress impulsive noise and improve the robustness of algorithms, from the perspective of similarity measurement, a series of nonlinear optimization criteria and/or cost functions have been proposed and applied to signal processing. Typical examples include the l p -norm with p [ 1 , 2 ) [9], the M-estimation theory [10], and the information theoretic learning (ITL) family [11]. More details about robust adaptive signal processing schemes can be seen in a review article [12]. In the family of ITL criteria, thanks to all the even-order moment information of the error signal contained in the minimization of error entropy (MEE) [13] and the maximum correntropy criterion (MCC) [14,15], they are widely used in robust signal processing and machine learning. Generally speaking, the MCC criterion has a smaller computational burden than that of the MEE criterion. Although correntropy can provide a generalized similarity measure between two random variables, the Gaussian kernel function used in the MCC criterion may not always be the best choice. Therefore, as an extension of the MCC criterion, the correntropy criterion based on the generalized Gaussian density function, usually called the generalized maximum correntropy criterion (GMCC), has been proposed and widely used in the field of robust adaptive filtering [16,17,18,19,20,21]. Similar to the LMS algorithm family, when the correlation degree of the input signal gradually increases, the gradient-based generalized maximum correntropy criterion algorithm (GB-GMCC) will have obvious convergence performance degradation. In addition, the gradient method can only provide relevant first-order derivative information, which leads to certain deficiencies in the filtering performance of corresponding algorithms.
The GB-GMCC algorithm has the disadvantage of a slow convergence rate in highly colored input situations. In other words, the gradient method is not suitable for the correlated inputs. There are many methods to overcome this issue. Examples include affine projection, the recursive method [10], the subband method [22] and the data-reusing method, etc. However, relatively speaking, the complexity of the data-reusing method is usually lower. Therefore, we can consider it as a candidate method for solving the problem introduced by the highly colored inputs. On the other hand, the first-order derivative information may damage the convergence rate of GB-GMCC, therefore we can consider the second-order derivative information to remedy such issues, that is to say, the Newton recursion method can be a good choice to replace the gradient method.
A new robust adaptive filtering algorithm is proposed by the methods of data-reusing, Newton recursion and GMCC. The new algorithm is called the Newton recursion-based data-reusing generalized maximum correntropy criterion algorithm (NR-DR-GMCC). In addition, to further improve the convergence rate and filtering accuracy of the NR-DR-GMCC algorithm, based on the random data-reusing strategy, an enhanced version of NR-DR-GMCC is derived and is named the Newton recursion-based random data-reusing generalized maximum correntropy criterion algorithm (NR-RDR-GMCC). Some simulation results under system identification and acoustic echo cancellation are conducted and demonstrate that NR-RDR-GMCC can achieve a better filtering accuracy and faster convergence rate than existing related algorithms in the α -SD noise environment. The main contributions of this work are as follows:
  • Based on the data-reusing and Newton recursion method, we propose the NR-DR-GMCC algorithm, which is derived by maximizing the generalized correntropy. Based on the data-reusing method, the information contained in the latest M input vectors can be obtained to combat the negative influence of highly colored inputs. Using the Newton recursion method, the second-order derivative information, i.e., the Hessian matrix, can be explored. Thus, the Hessian matrix can be used to update the weight vector leading to a faster convergence rate, thereby overcoming the disadvantage of the gradient method that only considers the first-order derivative information.
  • Inspired by the random strategy method, NR-RDR-GMCC can be derived by introducing a cache window with length C M into the NR-RDR-GMCC algorithm. Compared with the NR-DR-GMCC algorithm, NR-RDR-GMCC has very similar computational complexity except for requiring extra memory to contain C input vectors. However, NR-RDR-GMCC can explore more historical information and avoid the effects of consecutive large outliers through the cache window and use it to update the weight, thereby improving the filtering performance of the NR-DR-GMCC algorithm.
The rest of this paper is organized as follows. In Section 2, some preliminaries about α -SD Noise, the GMCC criterion, and the Newton recursion method are reviewed. Section 3 presents our proposed algorithms, namely, NR-DR-GMCC and NR-RDR-GMCC. Simulation results are presented in Section 4. Finally, the conclusion is given in Section 5.

2. Preliminaries

2.1. Alpha-Stable Distribution Noise

In this paper, the impulsive noise can be simulated by α -SD, and the noise model can construct the ubiquitous non-Gaussian noise with strong impulsive characteristics well [23]. The characteristic function of this model is defined as:
Φ ( θ ) = exp j δ θ λ | θ | α 1 + j ς sgn ( θ ) tan α π 2 , α 1 exp j δ θ λ | θ | 1 + j ς sgn ( θ ) 2 π log | t | , α = 1 ,
where j 2 = 1 ; α ( 0 , 2 ] denotes a characteristic exponent measuring the thickness of the tail of a distribution. The smaller the value of α , the thicker the tail of the corresponding distribution, and the more significant the impulsive characteristics. When the value of α is close to 2, the corresponding distribution is close to the Gaussian distribution; λ > 0 denotes a dispersion parameter, which is similar to the variance of a Gaussian distribution; ς [ 1 , 1 ] means the symmetry parameter, which can define the inclination of the distribution; and δ ( , + ) denotes the location parameter. In general, when α = 2 , ς = 0 , α -SD is equivalent to a Gaussian distribution. In this paper, for the sake of simplicity, the parameter vector of α -SD is denoted as d α T = [ α , ς , δ , λ ] .

2.2. Generalized Maximum Correntropy Criterion

Correntropy measures the local similarity between two random variables X and Y [11],
C σ ( X , Y ) = E κ σ ( X , Y ) = κ σ ( x , y ) d F X , Y ( x , y ) ,
where E [ · ] stands for the expectation operator; F X , Y ( x , y ) denotes the joint distribution function of X and Y; and κ δ ( · , · ) denotes a Gaussian kernel defined as:
κ σ x , y = exp σ | x y | 2 ,
where σ > 0 means a kernel factor. In practice, the joint distribution F X , Y ( x , y ) is usually unknown, and thus only a finite number of data x n , y n n = 1 N are used to derive the sample estimator of (3) as
C ^ σ X , Y = 1 N n = 1 N κ σ x n , y n .
From a probabilistic statistical point of view, the maximization of correntropy can lead to the error producing the maximum probability density at the origin, and thus the Equation (4) can be regarded as a nonlinear cost function, and it has been widely adopted in the field of robust adaptive signal processing. However, the Gaussian kernel function used in the estimation (4) may not always be the best choice. Therefore, other types of kernel functions can be used to replace the Gaussian kernel, such as a kernel function based on the generalized Gaussian density (GGD) function, i.e.,
κ s , τ x , y = γ s , t exp τ | x y | s ,
where τ = t s is the kernel parameter; s and t are positive numbers and present the shape parameter and the scale factor, respectively; and γ s , t = s / 2 t Γ s 1 denotes the normalization constant with Γ ( · ) being the Gamma function. Injecting GGD into (4), a new sample estimator can be obtained as
C ^ s , τ X , Y = 1 N n = 1 N κ s , τ x n , y n .
In the field of robust adaptive signal processing, (6) is often referred to as generalized correntropy. In applications such as regression analysis and classification studies, the correntropy loss (C-loss) metric can be used in place of correntropy [24]. Inspired by this, the generalized C-loss (GC-loss) function between the random variables X and Y can be defined as
J X , Y = C ^ s , τ 0 1 N n = 1 N κ s , τ e n ,
where e n = x n y n means the error information. Equation (7) shows that minimizing GC-loss is equivalent to maximizing generalized correntropy, which is usually denoted as GMCC and is widely used in the field of adaptive filtering [18,20,25].

2.3. Newton Recursion

Using the second-order Taylor formula to expand a continuous function f ( x ) at x 0 ,
f ( x ) f ( x 0 ) + ( x x 0 ) g ( x 0 ) + 1 2 ( x x 0 ) H ( x 0 ) ,
where g ( x 0 ) represents the gradient of f ( x ) at x 0 and H ( x 0 ) is the gradient of g ( x 0 ) at x 0 . Therefore, based on (8), we can obtain the gradient of f ( x ) with respect to x as
g ( x ) g ( x 0 ) + H ( x 0 ) ( x x 0 ) .
Setting g ( x ) = 0 , we obtain
x x 0 H ( x 0 ) 1 g ( x 0 ) ,
which leads to the general form of the Newton recursion method [26],
x k + 1 = x k H ( x k ) 1 g ( x k ) .
This recursive method is often used to obtain approximate solutions of nonlinear equations g ( x ) = 0 . If g ( x ) is a real-valued function, then x k + 1 represents the point where the tangent y g ( x k ) = H ( x k ) ( x x k ) of the function g ( x ) at point x k intersects the x-axis [27]. It is worth noting that when g ( x ) is a vector function, the H ( x ) becomes a Hessian matrix and, in this case, the use of Newton recursion must ensure that the Hessian matrix is positive definite.

3. The Proposed Algorithms

3.1. GB-GMCC

The following linear model is considered to reconstruct an unknown system such that the output y ( n ) of a adaptive filter matches the desired signal d ( n )
d ( n ) = u ( n ) T w o + v ( n ) ,
where u ( n ) = [ u ( n ) , u ( n 1 ) , , u ( n L + 1 ) ] T denotes the input vector with length L; w o = [ w 1 o , w 2 o , , w L o ] T means the intrinsic weight vector of the unknown system; and v ( n ) is the α -SD noise. For simplicity, the normalized constant γ s , t in the GGD function is ignored, and the sample estimator in (6) is replaced by an instantaneous estimator to establish the following cost function,
J w = exp τ | e n | s ,
where e n = d n u n T w denotes the estimation error and w = [ w 1 , w 2 , , w L ] T means the estimated weight vector. Calculate the gradient of (13), and use the gradient ascent method to obtain the weight vector update formula of the GB-GMCC algorithm as
w n + 1 = w n + μ f e n u n ,
where
f e n = exp τ | e n | s | e n | s 2 e n
denotes a nonlinear function of the error information; and μ > 0 stands for the learning step size.

3.2. NR-DR-GMCC

Observing from (14), one can find that the GB-GMCC algorithm only uses the current input vector u ( n ) to update the weight vector. Although this method has a lower computational complexity, it results in the slower convergence rate of the GB-GMCC algorithm when facing the increment of the correlation of the input signal. This problem can be overcome by using the data-reusing method, and thus the cost function in (13) can be changed as
J w = exp τ e n s s ,
where
e n = e n , e n 1 , , e n M + 1 T
represents an error vector consisting of M recent error signals and · s s means the s power of s -norm. In addition, the GB-GMCC algorithm updates the weight vector by using the gradient ascent strategy, which only considers the first-order derivative information. When all elements in the gradient are zero values, the stationary point of the gradient ascent method may be a local maximum and/or a saddle point. In other words, the gradient ascent method is an inefficient or even ineffective optimization method in some situations. To this end, inspired by the Newton recursion strategy, we can derive a new generalized maximum correntropy criterion algorithm as follows.
Let g w = J w w , and take the gradient of (16) with respect to w as
g w = J w w = τ s exp τ e n s s i = n n M + 1 e i s 2 e i u i ,
where sgn · is a symbolic function. Further denoting g w w as H w , we can obtain the Hessian matrix as
H w = τ s exp τ e n s s τ s i = n n M + 1 e i s 2 e i u i 2 s 1 i = n n M + 1 u i u i T .
Based on the following definitions, the g w and H w can be represented as more compact formations as
g ( w ) = τ s exp τ e n s s U n φ n H w = τ s exp τ e n s s τ s U n φ n φ n T U n T s 1 U n U n T U n = u n , u n 1 , , u n M + 1 φ n = b n sgn e n b n = e n s 1 , e n 1 s 1 , , e n M + 1 s 1 T sgn e n = sgn e n , sgn e n 1 , , sgn e n M + 1 T ,
where ⊙ is the Hartman product. Therefore, based on Newton recursion in (11), a weight vector update formation of the Newton recursion-based data-reusing generalized maximum correntropy criterion (NR-DR-GMCC) algorithm is derived as
w n + 1 = w n η τ s R u , φ n s 1 R u n + ϵ I 1 U n φ n ,
where R u , φ n = U n φ n φ n T U n T ; R u = U n U n T is a covariance matrix of the input matrix U n ; η > 0 is step-size; and ϵ > 0 denotes a small-valued smoothing factor.

3.3. NR-RDR-GMCC

Although the filtering performance of NR-DR-GMCC can be improved by digging much information contained in the latest M input vectors, the computational complexity of the algorithm will be increased when the data-reusing order M is large. Therefore, to balance the computational burden and the good filtering performance, we further inject a random strategy of data-reusing into the NR-DR-GMCC algorithm, and thus an enhanced version of NR-DR-GMCC is obtained as follows.
Based on this, we can define the error vector as
e r n = e r n 1 , e r n 2 , , e r n M T ,
which is composed of randomly selected error signals in which i 1 , M , r n i denotes the index of an interval [ n C + 1 , n ] and satisfies a b r n a r n b , where C > M is the length of a cache window. Then, (16) can be modified as
J w = exp τ e r n s s .
Based on some similar operations to (18) and (19), we can obtain a weight vector update formation of the Newton recursion-based random data-reusing generalized maximum correntropy criterion (NR-RDR-GMCC) algorithm as
w n + 1 = w n η τ s R u , φ r n s 1 R u r n + ϵ I 1 U r n φ r n ,
in which the following symbols are used
U r n = u r n 1 , u r n 2 , , u r n M φ r n = b r n sgn e r n b r n = e r n 1 s 1 , e r n 2 s 1 , , e r n M s 1 T sgn e r n = sgn e r n 1 , sgn e r n 2 , , sgn e r n M T R u , φ r n = U r n φ r n φ r n T U r n T R u r n = U r n U r n T .
Remarks: (1) Based on the data-reusing method, the NR-DR-GMCC algorithm can use the latest M input information to update the weight equation, remitting the convergence rate degradation problem of GB-GMCC due to the increased correlation of the input signal; (2) In comparison with GB-GMCC only using the information of the first-order derivative, the NR-DR-GMCC algorithm can use the second-order information provided by the Newton recursion method, and thus NR-DR-GMCC achieves faster convergence behavior; (3) As an improved version of NR-DR-GMCC, the NR-RDR-GMCC algorithm uses a cache window with length C > M to store the latest C input vectors. Then, to update the weight vector, NR-RDR-GMCC randomly selects M input vectors from the cache window. According to the random data-reusing method, NR-RDR-GMCC obtains abundant error information from the cache window and avoids the convergence loss caused by continuous outliers. Furthermore, when the data-reusing order M is fixed, NR-RDR-GMCC does not have a similar computational complexity to that of NR-DR-GMCC with data-reusing order M, but also achieves a better filtering performance than NR-DR-GMCC. Moreover, observing from (12) and (16), one can find that, when C = M , the NR-RDR-GMCC algorithm becomes the NR-DR-GMCC algorithm. That is to say, NR-RDR-GMCC can be regarded as an effective extension of NR-DR-GMCC. Algorithm 1 summarizes the pseudocode of the proposed algorithms. Moreover, Figure 1 plots the schematic diagram of the data-reusing and the random data-reusing methods, and shows the difference between these two methods, namely, the data-reusing method only utilizes the latest M inputs, whereas the random data-reusing method saves the latest C inputs and randomly selects M entries from these C inputs.
Algorithm 1. Pseudocode of Proposed Algorithms
1
Initialization: 0 < M < C , ϵ > 0 , η > 0 , s > 0 , τ > 0 , w 0 = 0
2
Computation:
3
for n > 0 do
4
  if n M
5
    i = n , u r n i = u i
6
  else if M < n C
7
     r n i 1 , n u r n i
8
  else n > C
9
     r n i n + C 1 , n u r n i
10
  end if
11
   U r n = u r n 1 , u r n 2 , , u r n M
12
   e r n = d r n U r n T w n
13
   b r n = e r n 1 s 1 , e r n 2 s 1 , , e r n M s 1 T
14
   sgn e r n = sgn e r n 1 , sgn e r n 2 , , sgn e r n M T
15
   φ r n = b r n sgn e r n
16
   H r w n = τ s U r n φ r n φ r n T U r n T s 1 U r n U r n T
17
   g r w n = U r n φ r n
18
   w n + 1 = w n η H r w n + ϵ I 1 g r w n
19
End for

3.4. Computation Complexity Analysis

Different from the traditional GB-GMCC algorithm, the random strategy is applied in NR-RDR-GMCC for efficiently reusing the past M input vectors, and thus our proposal can extract more historical information supporting the weight vector update inspired by Newton recursion. Although, compared with GB-GMCC algorithm, our proposed NR-RDR-GMCC has a higher computational complexity, it is worthwhile since NR-RDR-GMCC outperforms GB-GMCC in terms of filtering accuracy as will be shown in the simulation results. It is worth noting that the matrix inverse involved in NR-RDR-GMCC can be achieved with the help of some iterative optimization methods, thereby reducing the computational complexity of the NR-RDR-GMCC algorithm [28]. In addition, Table 1 lists the computational complexity of the proposed NR-RDR-GMCC and other related algorithms per iteration in terms of multiplications and additions in which M represents the order of data-reusing and/or projection, L stands for the tap length, N exp τ , s indicates the computational complexity associated with the nonlinear function exp τ e i s e i s 1 [20], and O ( L 3 ) is the computational burden required for the direct inverse of the square matrix in the L × L dimension.

4. Simulation Results

4.1. System Identification

In order to evaluate the filtering performance of the NR-RDR-GMCC algorithm, we consider the system identification problem represented in (12), where the unknown weight vector w o is randomly generated and the tap length L = 32 . Impulsive noise is modeled by α -SD with the parameter vector d α T = 1.5 , 0 , 0 , 0.1 . The input signal u ( n ) is generated by a zero-mean Gaussian noise with unit variance filtered by the following second-order system:
H z = 1 + 0.5 z 1 + 0.81 z 2 1 0.59 z 1 + 0.4 z 2 .
To compare the filtering performance of competing algorithms, the normalized mean square deviation (NMSD) is used, i.e.,
NMSD = 20 lg w w o w o ;
in addition, all NMSD simulation results are averaged over 100 independent experiments.
Firstly, in order to study the effect of the order M of data-reusing on the convergence performance of the NR-RDR-GMCC algorithm, we set different M from 2 , 4 , 6 , 8 , 16 , 32 , and the other parameters are fixed as C = 512 , s = 1.5 , τ = 0.0001 , η = 0.005 , and ϵ = 0.0001 . Figure 2 plots the corresponding NMSD curves. From Figure 2, we can observe that: (1) When M is large, such as M = 32 , the NR-RDR-GMCC algorithm does not converge; (2) Under the premise of the convergence of NR-RDR-GMCC, by appropriately increasing the M value, such as from 2 to 16, the convergence rate of NR-RDR-GMCC will be significantly improved; (3) With some moderate values, such as M = 4 , NR-RDR-GMCC can achieve an acceptable convergence rate and the best filtering accuracy. Therefore, for NR-RDR-GMCC, in practical applications, it is necessary to select a proper value of M to make the trade-off among the computational complexity, filtering accuracy and convergence rate.
Secondly, we investigate the influence of the length of cache window C on the performance of NR-RDR-GMCC. To this end, some values of C from 16 , 32 , 64 , 128 , 256 , 512 are set and other parameters are fixed as M = 16 , s = 1.5 , τ = 0.0001 , η = 0.005 and ϵ = 0.0001 . The corresponding NMSD curves are plotted in Figure 3, which clearly shows that: (1) The filtering accuracy of NR-RDR-GMCC can be enhanced by increasing the values of C. In addition, with some large C values, such as C { 256 , 512 } , the jitter behavior of NR-RDR-GMCC can be restrained, thereby improving the stability of our proposed algorithm; (2) When C = M = 16 , NR-RDR-GMCC reduces to the NR-DR-GMCC algorithm. However, compared to NR-RDR-GMCC with other large C values, the NR-DR-GMCC algorithm realizes the fastest convergence rate at the cost of the worst accuracy. That is to say, with some moderate C values, NR-RDR-GMCC can achieve a good balance between a smaller misadjustment and a good convergence rate.
Thirdly, we study the effect of the shape parameter s and kernel parameter τ on the filtering accuracy of NR-RDR-GMCC. To this end, different s and τ values are considered and other parameters are set as C = 512 , M = 16 , η = 0.005 and ϵ = 0.0001 . Figure 4 shows the steady-state NMSD (SS-NMSD) estimated by the last 100 NMSD for NR-RDR-GMCC with respect to various values of s ( 0 , 3 ) and τ ( 0.00001 , 0.0002 ) . From this figure we can observe that: (1) With the same shape parameter s, different τ values lead to NR-RDR-GMCC obtaining very similar filtering accuracy; (2) With the same kernel parameter τ , different s values result in obvious steady-state behaviors of NR-RDR-GMCC. That is to say, when s ( 1.0 , 3.0 ) , the filtering accuracy will gradually increase with the decrease of s; when s = 1.0 , the algorithm mutates into a divergent state; when s ( 0.0 , 1.0 ) , the algorithm does not converge. In addition, to clearly show the filtering behavior, some NMSD curves of NR-RDR-GMCC with τ 0.00001 , 0.002 and s 1.5 , 2.0 , 2.5 , 3.0 are plotted in Figure 5. From this figure, we have the following observations: (1) When the s parameter is same, in most cases, the NR-RDR-GMCC algorithms with τ = 0.00001 or τ = 0.002 realize similar convergence behavior; (2) With the same τ parameter, NR-RDR-GMCC can realize an enhanced convergence rate and filtering accuracy by decreasing the values of s from 3.0 to 1.5 .
Finally, in order to verify the effectiveness of the proposed NR-RDR-GMCC algorithm, some related algorithms, such as NR-DR-GMCC, GB-GMCC [19], affine projection GMCC (AP-GMCC) [20], AP sign algorithm (APSA) [29], AP Versoria (APV) [30] and MCC-APA [31], are considered in this part. For all algorithms, the smooth parameter or regularization parameter ϵ = 0.001 ; for all GMCC based algorithms, s = 1.5 , τ = 0.0001 ; other parameter settings are experimentally tested so that the algorithms have similar initial convergence rates. Table 2 lists the parameter setting. In addition, to further verify the advantages of the GMCC criterion, this experiment also considers the special case of s = 2.0 in the NR-RDR-GMCC algorithm; in other words, the NR-RDR-GMCC algorithm reduces to an MCC-based NR-RDR algorithm. Furthermore, we change the unknown weight vector from w o to w o at the middle iterations to simulate system mutation and explore the tracking performance of various algorithms. Figure 6 plots the corresponding NMSD curves and reveals that: (1) Compared to the GB-GMCC algorithm, the NR-RDR-GMCC ( s = 1.5 ) algorithm has a very obvious advantage in terms of filtering accuracy and convergence rate. In addition, irrespective of the mutation of system, in comparison with other algorithms, the NR-RDR-GMCC ( s = 1.5 ) algorithm achieves the comparable convergence rate and the best filtering accuracy; (2) In this simulation, the NR-RDR-GMCC ( s = 2.0 ) algorithm is inferior to the other algorithms in terms of convergence rate and steady-state misadjustment as shown in Table 2. That is to say, in some situations, to derive some robust adaptive filtering algorithms, GMCC is better than MCC.

4.2. Acoustic Echo Cancellation

To further compare the filtering behaviors of the mentioned algorithms, a typical application, i.e., acoustic echo cancellation (AEC), is also considered in this work. Under the condition of double-ended voice, in this part, two kinds of environmental noises are considered, namely: (1) the noise v ( n ) being impulsive and (2) the noise v ( n ) not being impulsive. In all trials, a Gaussian background noise with a signal-to-noise ratio of 60 dB was considered. The echo paths w 1 o and w 2 o [32] involved in this experiment and the voice input signal are shown in Figure 7 and Figure 8, respectively, where the signal sample rate was set to 8 kHz. To simulate the system mutation, the echo path was abruptly changed from w 1 o to w 2 o at the 7000-th iteration. Except where otherwise mentioned, for all algorithms, the data-reusing order or projection order is M = 16 , the regularization parameter or smooth parameter is ϵ = 0.0001 , and for GMCC-based algorithms the kernel factor is τ = 0.001 .

4.2.1. Without Impulsive Noise

In order to compare the acoustic echo cancellation filtering performance of the NR-RDR-GMCC algorithm and other related algorithms in an environment without impulsive noise, based on various experimental tests, Table 3 lists the remaining parameters’ setting to realize a similar initial convergence rate. Figure 9 plots the corresponding NMSD curves. From this figure, we can see that: (1) Due to the interference of near-end voice, the NR-RDR-GMCC algorithm and other related algorithms have a slightly jitter phenomenon after convergence; (2) Without interference of impulsive noise, NR-RDR-GMCC is superior to other algorithms in terms of convergence rate and filtering accuracy before or after system mutation.

4.2.2. With Impulsive Noise

In this part, we use the Bernoulli–Gaussian distribution model to obtain the impulsive noise, i.e., v n = b n v g n , where b ( n ) is the Bernoulli process with the probability density model P b ( n ) = 1 = P r = 0.001 . Before the system mutation, v g ( n ) is a zero-mean Gaussian noise with a variance of 151. After system mutation, the variance of v g ( n ) becomes 110. For all algorithms, Table 4 lists the remaining parameters’ setting to realize a similar initial convergence rate. The corresponding NMSD curves are plotted in Figure 10, which shows that, although some impulsive noises may have a negative effect on the filtering performance of the considered algorithms, our proposed NR-RDR-GMCC still achieves the smallest steady-state misadjustment before or after the mutation of the system.

5. Conclusions

The traditional GB-GMCC algorithm has a degradation of convergence performance when facing highly colored input signals. In this work, to overcome this issue, the NR-DR-GMCC algorithm was derived by using the data-reusing method and the Newton recursion method. In addition, to further avoid the influence of continuous large outliers in the output signal and extract more historical data, a cache window is considered to collect the latest C input vectors, and then M input vectors in the cache window are randomly selected to derive an enhanced version of NR-DR-GMCC, namely, NR-RDR-GMCC. From a mathematical point of view, when the cache window length C is equal to the data-reusing order M, NR-RDR-GMCC can be converted to the NR-DR-GMCC algorithm, and thus the former can be regarded as a good extension of the latter. The system identification simulation results show that the NR-RDR-GMCC algorithm has a better filtering accuracy and a faster convergence rate than the corresponding AP-type algorithm and GB-GMCC. Furthermore, in acoustic echo cancellation applications, the NR-RDR-GMCC algorithm has obvious advantages over other related algorithms in terms of steady-state misalignment.

Author Contributions

Methodology, J.Z.; Software, Y.M.; Investigation, Y.M.; Writing—original draft, Y.M.; Writing—review & editing, J.Z.; Supervision, J.Z.; Project administration, J.Z., Y.Q. and Q.L.; Funding acquisition, J.Z., Y.Q. and Q.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the Southwest University of Science and Technology Doctor Fund (Grant no. 20zx7119), in part by the National Natural Science Foundation of China (Grant no. 62201478), in part by the Central Financial Stability Support Special Project (Grant no. GJCZ-0032-19), in part by the Sichuan Science and Technology Program (Grant no. 2022YFG0148), and in part by the National Defense Basic Research Program (Grant no. JCKY2020404C001).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

We are grateful to the associate editor and all reviewers for reviewing this manuscript and providing valuable suggestions.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Lee, K.A.; Gan, W.S.; Kuo, S.M. Subband Adaptive Filtering: Theory and Implementation; John Wiley & Sons: New York, NY, USA, 2009. [Google Scholar]
  2. Sayed, A.H. Adaptive Filters; John Wiley & Sons: New York, NY, USA, 2011. [Google Scholar]
  3. Bhattacharjee, S.S.; Shaikh, M.A.; Kumar, K.; George, N.V. Robust constrained generalized correntropy and maximum versoria criterion adaptive filters. IEEE Trans. Circuits Syst. II Express Briefs 2021, 68, 3002–3006. [Google Scholar] [CrossRef]
  4. Rusu, A.G.; Paleologu, C.; Benesty, J.; Ciochină, S. A variable step size normalized least-mean-square algorithm based on data reuse. Algorithms 2022, 15, 111. [Google Scholar] [CrossRef]
  5. Ozeki, K. Theory of Affine Projection Algorithms for Adaptive Filtering; Springer: New York, NY, USA, 2016. [Google Scholar]
  6. Yu, T.; Li, W.; de Lamare, R.C.; Yu, Y. M-estimate affine projection spline adaptive filtering algorithm: Analysis and implementation. Digit. Signal Process. 2022, 123, 103452. [Google Scholar] [CrossRef]
  7. Sun, X.; Ji, J.; Ren, B.; Xie, C.; Yan, D. Adaptive forgetting factor recursive least square algorithm for online identification of equivalent circuit model parameters of a lithiumion battery. Energies 2019, 12, 2242. [Google Scholar] [CrossRef] [Green Version]
  8. Zhao, J.; Zhang, J.A.; Li, Q.; Zhang, H.; Wang, X. Recursive constrained generalized maximum correntropy algorithms for adaptive filtering. Signal Process. 2022, 199, 108611. [Google Scholar] [CrossRef]
  9. Zhao, J.; Zhang, H.; Wang, G.; Zhang, J.A. Projected kernel least mean p-power algorithm: Convergence analyses and modifications. IEEE Trans. Circuits Syst. I Regul. Pap. 2020, 67, 3498–3511. [Google Scholar] [CrossRef]
  10. Xu, W.; Zhao, H. Robust constrained recursive least M-estimate adaptive filtering algorithm. Signal Process. 2022, 194, 108433. [Google Scholar] [CrossRef]
  11. Principe, J.C. Information Theoretic Learning: Renyi’s Entropy and Kernel Perspectives; Springer Science & Business Media: New York, NY, USA, 2010. [Google Scholar]
  12. Kumar, K.; Pandey, R.; Karthik, M.; Bhattacharjee, S.S.; George, N.V. Robust and sparsity-aware adaptive filters: A review. Signal Process. 2021, 189, 108276. [Google Scholar] [CrossRef]
  13. Chen, B.; Dang, L.; Gu, Y.; Zheng, N.; Príncipe, J.C. Minimum error entropy Kalman filter. IEEE Trans. Syst. Man Cybern. Syst. 2019, 51, 5819–5829. [Google Scholar] [CrossRef] [Green Version]
  14. Jiang, Z.; Li, Y.; Huang, X. A correntropy-based proportionate affine projection algorithm for estimating sparse channels with impulsive noise. Entropy 2019, 21, 555. [Google Scholar] [CrossRef]
  15. Yue, P.; Qu, H.; Zhao, J.; Wang, M. Newtonian-type adaptive filtering based on the maximum correntropy criterion. Entropy 2020, 22, 922. [Google Scholar] [CrossRef] [PubMed]
  16. Qu, H.; Shi, Y.; Zhao, J. A smoothed algorithm with convergence analysis under generalized maximum correntropy criteria in impulsive interference. Entropy 2019, 21, 1099. [Google Scholar] [CrossRef] [Green Version]
  17. Liu, D.; Zhao, H.; He, X.; Zhou, L. Polynomial constraint generalized maximum correntropy normalized subband adaptive filter algorithm. Circuits Syst. Signal Process. 2022, 41, 2379–2396. [Google Scholar] [CrossRef]
  18. Zhao, J.; Zhang, H.; Wang, G. Fixed-point generalized maximum correntropy: Convergence analysis and convex combination algorithms. Signal Process. 2019, 154, 64–73. [Google Scholar] [CrossRef]
  19. Chen, B.; Xing, L.; Zhao, H.; Zheng, N.; Príncipe, J.C. Generalized correntropy for robust adaptive filtering. IEEE Trans. Signal Process. 2016, 64, 3376–3387. [Google Scholar] [CrossRef] [Green Version]
  20. Zhao, J.; Zhang, H.; Zhang, J.A. Generalized maximum correntropy algorithm with affine projection for robust filtering under impulsive-noise environments. Signal Process. 2020, 172, 107524. [Google Scholar] [CrossRef]
  21. Zheng, H.; Qian, G. Generalized Maximum Complex Correntropy Augmented Adaptive IIR Filtering. Entropy 2022, 24, 1008. [Google Scholar] [CrossRef]
  22. Yu, Y.; He, H.; de Lamare, R.C.; Chen, B. Study of General Robust Subband Adaptive Filtering. arXiv 2022, arXiv:2208.08856. [Google Scholar]
  23. Nikias, C.L.; Shao, M. Signal Processing with Alpha-Stable Distributions and Applications; Wiley-Interscience: New York, NY, USA, 1995. [Google Scholar]
  24. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  25. Zhu, Y.; Zhao, H.; Zeng, X.; Chen, B. Robust generalized maximum correntropy criterion algorithms for active noise control. IEEE/ACM Trans. Audio, Speech Lang. Process. 2020, 28, 1282–1292. [Google Scholar] [CrossRef]
  26. Qi, L.; Sun, J. A nonsmooth version of Newton’s method. Math. Program. 1993, 58, 353–367. [Google Scholar] [CrossRef]
  27. Yau, L.; Ben-Israel, A. The Newton and Halley methods for complex roots. Am. Math. Mon. 1998, 105, 806–818. [Google Scholar] [CrossRef]
  28. Yang, F.; Yang, J. A comparative survey of fast affine projection algorithms. Digit. Signal Process. 2018, 83, 297–322. [Google Scholar] [CrossRef]
  29. Shao, T.; Zheng, Y.R.; Benesty, J. An affine projection sign algorithm robust against impulsive interferences. IEEE Signal Process. Lett. 2010, 17, 327–330. [Google Scholar] [CrossRef]
  30. Huang, F.; Zhang, J.; Zhang, S. Affine projection versoria algorithm for robust adaptive echo cancellation in hands-free voice communications. IEEE Trans. Veh. Technol. 2018, 67, 11924–11935. [Google Scholar] [CrossRef]
  31. Liu, C.; Qiu, T.; Li, J.; Li, R. Affine project algorithm based on maximum correntropy criterion for impulsive noise environment. J. Dalian Univ. Technol. 2018, 58, 86–91. [Google Scholar]
  32. Sector, T.S. Digital network echo cancellers. In Series G: Transmission Systems and Meaid, Digital Systems and Netwroks, Recommendation G.168; International Telecommunication Union (ITU-T): Geneva, Switzerland, 2015. [Google Scholar]
Figure 1. Schematic diagram of the data-reusing and the random data-reusing methods.
Figure 1. Schematic diagram of the data-reusing and the random data-reusing methods.
Entropy 24 01845 g001
Figure 2. The NMSD curves of NR-RDR-GMCC with different values of M 2 , 4 , 6 , 8 , 16 , 32 .
Figure 2. The NMSD curves of NR-RDR-GMCC with different values of M 2 , 4 , 6 , 8 , 16 , 32 .
Entropy 24 01845 g002
Figure 3. The NMSD curves of NR-RDR-GMCC with different values of C 16 , 32 , 64 , 128 , 256 , 512 .
Figure 3. The NMSD curves of NR-RDR-GMCC with different values of C 16 , 32 , 64 , 128 , 256 , 512 .
Entropy 24 01845 g003
Figure 4. The steady-state NMSD results of NR-RDR-GMCC with different values of s ( 0 , 3 ) and τ ( 0.00001 , 0.0002 ) .
Figure 4. The steady-state NMSD results of NR-RDR-GMCC with different values of s ( 0 , 3 ) and τ ( 0.00001 , 0.0002 ) .
Entropy 24 01845 g004
Figure 5. The NMSD curves of NR-RDR-GMCC with different values of τ 0.00001 , 0.002 and s 1.5 , 2.0 , 2.5 , 3.0 .
Figure 5. The NMSD curves of NR-RDR-GMCC with different values of τ 0.00001 , 0.002 and s 1.5 , 2.0 , 2.5 , 3.0 .
Entropy 24 01845 g005
Figure 6. The NMSD curves of competing algorithms with d α T = 1.5 , 0 , 0 , 0.1 .
Figure 6. The NMSD curves of competing algorithms with d α T = 1.5 , 0 , 0 , 0.1 .
Entropy 24 01845 g006
Figure 7. Two acoustic finite impulse responses with L = 128 : (a) w 1 o ; (b) w 2 o .
Figure 7. Two acoustic finite impulse responses with L = 128 : (a) w 1 o ; (b) w 2 o .
Entropy 24 01845 g007
Figure 8. Speech signals: (a) Far-end; (b) Near-end.
Figure 8. Speech signals: (a) Far-end; (b) Near-end.
Entropy 24 01845 g008
Figure 9. The NMSD curves of various algorithms in ACE without impulsive noise interference.
Figure 9. The NMSD curves of various algorithms in ACE without impulsive noise interference.
Entropy 24 01845 g009
Figure 10. The NMSD curves of various algorithms in ACE under an impulsive noise environment.
Figure 10. The NMSD curves of various algorithms in ACE under an impulsive noise environment.
Entropy 24 01845 g010
Table 1. Computational Complexity of Algorithms Per Iteration.
Table 1. Computational Complexity of Algorithms Per Iteration.
AlgorithmsMultiplicationsAdditions
GB-GMCC M N e x p τ , s + 2 L + 1 2 M L + L
AP-GMCC M N e x p τ , s + 2 M L + 2 L + 1 2 M L + L
APSA M L + 2 L + K 2 M L + L
APV M ( 2 L + M + 7 ) + 1 2 M L + L
MCC-APA O ( L 3 ) + ( M + 2 ) L 2 + N e x p τ , s + 2 ( M + 1 ) L + 2 ( M + L ) L 2 + ( 2 M 1 ) L + M 1
NR-RDR-GMCC O ( L 3 ) + ( M + 4 ) L 2 + N e x p τ , s + 3 ( M + 1 ) L + M ( M + 1 ) L 2 + ( 2 M 1 ) L + M
Table 2. Parameter Setting And Steady-State NMSD Results In System Identification.
Table 2. Parameter Setting And Steady-State NMSD Results In System Identification.
Algorithms η MCRadiusKernel WidthSS-NMSD (dB)
r σ w o w o
GB-GMCC0.004----−20.9597−20.8324
AP-GMCC0.00516---−20.7933−20.6320
APSA0.00616---−22.1333−22.0104
APV0.00516-10-−20.4574−20.2859
MCC-APA0.00516--10−19.1079−18.7148
NR-DR-GMCC0.0021616--−21.2425−20.7102
NR-RDR-GMCC0.00516512--−23.8190−23.5575
Table 3. Parameter Setting and Steady-State NMSD Results in AEC without Impulsive Noise.
Table 3. Parameter Setting and Steady-State NMSD Results in AEC without Impulsive Noise.
AlgorithmsParametersSS-NMSD (dB)
w 1 o w 2 o
GB-GMCC η = 0.05 , s = 1.7 −18.4309−14.3984
AP-GMCC η = 0.10 , s = 1.7 −22.5124−22.3412
APSA η = 0.06 −21.1401−20.5691
APV η = 0.05 −28.0652−27.7924
MCC-APA η = 0.50 −37.9557−38.9585
NR-RDR-GMCC η = 0.10 , C = 256 , s = 1.7 −40.8235−41.9047
Table 4. Parameter Setting and Steady-State NMSD Results in AEC under Impulsive Noise.
Table 4. Parameter Setting and Steady-State NMSD Results in AEC under Impulsive Noise.
AlgorithmsParametersSS-NMSD (dB)
w 1 o w 2 o
GB-GMCC η = 0.15 , s = 1.6 −3.6680−3.5960
AP-GMCC η = 0.20 , s = 1.6 −6.2958−7.1930
APSA η = 0.30 −8.3213−8.0550
APV η = 0.10 −11.7331−11.6527
MCC-APA η = 0.10 −2.4777−3.6380
NR-RDR-GMCC η = 0.10 , C = 256 , s = 1.6 −15.6412−17.2785
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhao, J.; Mu, Y.; Qiao, Y.; Li, Q. Newton Recursion Based Random Data-Reusing Generalized Maximum Correntropy Criterion Adaptive Filtering Algorithm. Entropy 2022, 24, 1845. https://doi.org/10.3390/e24121845

AMA Style

Zhao J, Mu Y, Qiao Y, Li Q. Newton Recursion Based Random Data-Reusing Generalized Maximum Correntropy Criterion Adaptive Filtering Algorithm. Entropy. 2022; 24(12):1845. https://doi.org/10.3390/e24121845

Chicago/Turabian Style

Zhao, Ji, Yuzong Mu, Yanping Qiao, and Qiang Li. 2022. "Newton Recursion Based Random Data-Reusing Generalized Maximum Correntropy Criterion Adaptive Filtering Algorithm" Entropy 24, no. 12: 1845. https://doi.org/10.3390/e24121845

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop