Next Article in Journal
An Experimental Investigation into the Thermal Characteristics of Bump Foil Journal Bearings
Next Article in Special Issue
Lagrangian Regularized Twin Extreme Learning Machine for Supervised and Semi-Supervised Classification
Previous Article in Journal
Structural Brain Asymmetries for Language: A Comparative Approach across Primates
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Low-Complexity Constrained Recursive Kernel Risk-Sensitive Loss Algorithm

School of Computer Science and Engineering, Chongqing Three Gorges University, Wanzhou, Chongqing 404120, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(5), 877; https://doi.org/10.3390/sym14050877
Submission received: 30 March 2022 / Revised: 16 April 2022 / Accepted: 19 April 2022 / Published: 25 April 2022
(This article belongs to the Special Issue Adaptive Filtering and Machine Learning)

Abstract

:
The constrained recursive maximum correntropy criterion (CRMCC) combats the non-Gaussian noise effectively. However, the performance surface of maximum correntropy criterion (MCC) is highly non-convex, resulting in low accuracy. Inspired by the smooth kernel risk-sensitive loss (KRSL), a novel constrained recursive KRSL (CRKRSL) algorithm is proposed, which shows higher filtering accuracy and lower computational complexity than CRMCC. Meanwhile, a modified update strategy is developed to avoid the instability of CRKRSL in the early iterations. By using Isserlis’s theorem to separate the complex symmetric matrix with fourth-moment variables, the mean square stability condition of CRKRSL is derived, and the simulation results validate its advantages.

1. Introduction

The constrained adaptive filters (CAFs) [1], where the weight is subject to linear constraints, have been widely studied in the field of adaptive signal processing. The original research of CAFs was derived from the antenna array processing, which employed the linearly-constrained minimum-variance (LCMV) criterion to estimation the direction of the antenna array [2]. CAFs have since been successfully applied to adaptive beamforming [3], system identification [4], channel equalization [5], and blind multiuser detection [6].
The simplest linearly-constrained adaptive filter, named constrained least mean-square (CLMS) [2], is developed from the LCMV criterion, and its mean square performance is analyzed based on a decomposable symmetric matrix in [7]. Due to stochastic gradient optimization, the CLMS has a simple structure with low computational complexity. However, its performance is highly influenced by the step size and correlated input. To improve the convergence speed, the constrained fast least-squares (CFLS) algorithm [8], linear-equality-constrained recursive least-squares (CRLS) algorithm [9] and its relaxed version are proposed at the expense of high computational complexity. Furthermore, the reduced-complexity constrained recursive least-squares algorithm based on the dichotomous coordinate descent (CRLS-DCD) iterations [10] and low-complexity constrained affine-projection (CAP) algorithm [11] with the data-selection method are proposed for reducing the computational complexity effectively. Other types of improved constrained filters [12,13,14] have been widely used, which make a trade-off between the computational complexity and filtering performance, i.e., convergence speed and filtering accuracy. All the mentioned constrained algorithms above are developed based on the mean square error (MSE) criterion [15] with good performances under Gaussian assumptions. However, with non-Gaussian cases, the filtering accuracy of these algorithms will decline sharply.
Therefore, the maximum correntropy criterion (MCC) [16,17], generalized MCC (GMCC) [18] and minimum error entropy (MEE) [19] criteria from information theoretic learning (ITL) [20] become alternative criteria, showing strong robustness to non-Gaussian signals. Adding the linear constraints on the weights of MCC and GMCC, the constrained MCC (CMCC) [21] and constrained GMCC (CGMCC) [22] are developed by using stochastic gradient optimization. CMCC and CGMCC display good filtering performances in the presence of single-peak heavy-tailed noise. However, when coping with multi-peak noise, their performances will decline. Due to the symmetry of the errors, MEE counteracts the influence of multi-peak noise effectively. Adding the linear constraints into MEE, a gradient-based constrained MEE (CMEE) algorithm [23] with a sliding window is proposed at the expense of higher complexity but with better accuracy than CMCC and GMCC in the presence of multi-peak noise. Except for the ITL-based constrained filters, there also exist other criteria [24,25] to show good performance under a non-Gaussian environment. A constrained least mean M-estimation (CLMM) algorithm based on an improved M-estimation loss function has been proposed in [24]. Inspired by the boundedness of the gradient of the lncosh function, a constrained least lncosh adaptive filtering algorithm (CLLAF) has been developed in [25]. These constrained algorithms show good performances under different non-Gaussian noise. However, the gradient-based constrained adaptive filters need to make a trade-off between accuracy and convergence speed through an adjustable step size. Hence, the constrained recursive MCC (CRMCC) algorithm [26] is developed, which not only improves the accuracy but also accelerates the convergence.
Recently, an advanced ITL-based criterion, named kernel risk-sensitive loss (KRSL) [27], has been proposed by introducing a beneficial risk-sensitive parameter to regulate the shape of its performance surface. Compared with MCC, the KRSL is more “convex”, which conduces better accuracy and faster convergence. Based on the KRSL, a gradient-based constrained mixture KRSL algorithm [28] is proposed, which indicates a higher filtering accuracy than CMCC.
In this paper, thanks to the advantage of KRSL criterion, a novel constrained recursive KRSL (CRKRSL) algorithm is proposed by using an average approximation method [29]. Thus, the proposed CRKRSL algorithm is converted to a variable step size gradient-based algorithm, which has a lower complexity than the traditional constrained recursive algorithms. Meanwhile, due to the instability of CRKRSL at the initial update stage, we use a gradient algorithm with a fixed step size to replace its initial update. Moreover, the convergence condition of mean square stability regarding iteration times is derived by decomposing a symmetric matrix with fourth-order variables. Simulation results indicate the advantages of CRKRSL on filtering accuracy and computational complexity.
The rest of the paper is organized as follows. The constrained KRSL loss and the CRKRSL algorithm are presented in Section 2. Stability analysis of CRKRSL is given in Section 3. Simulation results and discussion of CRKRSL are shown in Section 4. Finally, the conclusion is given in Section 5.

2. CRKRSL Algorithm

2.1. Notations

Throughout this paper, R denotes the real field, R m denotes an m-dimensional real-valued vector space, R n × m denotes n × m matrix where its entries belong to R ; ( · ) T represents the transpose operation; E [ · ] is the expectation operation; · denotes the Euclidean norm; and O ( · ) represents the computational complexity of algorithm.

2.2. KRSL Loss

As a nonlinear similarity measure of variable X , Y R , the KRSL [27] is defined as
S ( X , Y ) = 1 γ E [ exp ( γ 2 φ ( X ) φ ( Y ) 2 ) ] = 1 γ E [ exp ( γ ( 1 κ ( X Y ) ) ) ] ,
where γ [ 0 , + ) is a risk-sensitive parameter. φ ( · ) is the corresponding mapping variable in reproducing kernel Hilbert spaces (RKHS) [30], satisfying φ T ( X ) φ ( Y ) = κ ( X Y ) with Gaussian kernel
κ ( X Y ) = exp X Y 2 2 σ 2 .
Generally, in terms of the unknown joint distribution between X and Y , we adopt a finite sampled data pair { x ˜ l , y ˜ l } l = 1 N to approximate the expected loss, which is given by
S ^ ( X , Y ) = 1 γ N l = 1 N exp ( γ ( 1 κ ( x ˜ l y ˜ l ) ) ) .

2.3. Constrained KRSL Loss

When applied to constrained adaptive filtering, the optimization problem with KRLS loss becomes
min w 1 γ l = 1 n exp ( γ ( 1 κ ( d l w T u l ) ) ) s . t . C T w = f ,
where { u l , d l } l = 1 N R m × R are the input–output training data pair; w R m is the weight vector and e l = d l w T u l is the corresponding error; C R m × q and f R q are the constrained matrix and vector, respectively. By constructing the Lagrange function, the constrained problem is transformed to minimize the following constrained KRSL loss:
J c ( n ) = σ 2 γ l = 1 n exp ( γ ( 1 κ ( d l w T u l ) ) ) + θ n T ( f C T w )
with θ n R q being the Lagrange multiplier.

2.4. Proposed CRKRSL Algorithm

Taking the gradient of J c ( n ) to zero at instant n, one has
J c ( n ) w n = l = 1 n ϕ ( e l ) ( d l w n T u l ) u l + C θ n = 0
with ϕ ( e l ) = exp ( γ ( 1 κ ( e l ) ) ) κ ( e l ) .
Define
U n 1 = l = 1 n 1 ϕ ( e l ) u l u l T + ϕ ( e n ) u n u n T
d n = l = 1 n 1 ϕ ( e l ) u l d l + ϕ ( e n ) u n d n .
Then, the constrained solution is derived as
w n = U n d n + U n C θ n
with θ n = ( C T U n C ) 1 ( f C T U n d n ) . By the matrix inversion lemma [31], Equation (7) is further rewritten by
U n = U n 1 g n u n T U n 1 ,
where the gain is given by
g n = U n 1 u n ϕ 1 ( e n ) + u n T U n 1 u n .
Reorganizing Equation (11), we get another form of g n by
g n = ϕ ( e n ) ( U n 1 u n g n u n T U n 1 u n ) = ϕ ( e n ) U n u n .
To obtain a recursive solution, we expand Equation (9) as follow:
w n = U n ( d n 1 + ϕ ( e n ) u n d n ) + U n C θ n = ( U n 1 g n u n T U n 1 ) d n 1 + ϕ ( e n ) U n u n d n + U n C θ n = w n 1 + ϕ ( e n ) U n u n d n g n u n T w n 1 + U n C θ n = w n 1 + ϕ ( e n ) e n U n u n + U n C θ n ,
where the corrected error e n updated by the a priori weight w n 1 is obtained by
e n = d n u n T w n 1 .
Substituting Equation (12) into Equation (13), we get
w n = w n 1 + g n e n + U n C θ n
with
θ n = ( C T U n C ) 1 ( f C T ( w n 1 + g n e n ) ) .
Therefore, combing Equations (10), (11), (14) and (15), we obtain the constrained recursive CRKRSL algorithm. The main drawback of Equation (15) is that the inverse matrix ( C T U n C ) 1 needs to be updated iteratively with complexity O ( q 3 ) . To reduce the computational complexity of the CRKRSL algorithm, we consider the following linear model and give some assumptions. The linear model is described by
d n = w * T u n + v n ,
where w * is the model parameter and v n is the noise at instant n. The assumptions are given as follows:
  • A1 { u n } is independent identically distributed (i.i.d), generated from a multivariate Gaussian distribution with covariance matrix R = E [ u n u n T ] ;
  • A2 { v n } is zero mean, i.i.d, and independent with { u n } , satisfying δ 2 = E [ v n 2 ] ;
  • A3 the error e n is uncorrelated with u n u n T .
Inspired by the average approximation [29] and based on A1–A3, the correlation matrix U n is approximated by
U n = 1 n ( 1 n l = 1 n ϕ ( e l ) u l u l T ) 1 1 n E [ ϕ ( e n ) ] 1 E [ u n u n T ] 1 = η n Z ,
where Z = R 1 = E [ u n u n T ] 1 and η n = ( n E [ ϕ ( v n ) ] ) 1 is approximated by the first-order Taylor expansion around the noise.
Furthermore, Equation (13) can be simplified as
w n = w n 1 + ϕ ( e n ) e n U n u n + U n C θ n = w n 1 + η n ϕ ( e n ) e n Z u n + Z C θ ^ ,
where θ ^ = Θ ( f C T ( w n 1 + η n ϕ ( e n ) e n Z u n ) ) and the constrained inverse matrix is defined as Θ = ( C T Z C ) 1 .
Therefore, w n is further expressed as
w n = Q ( w n 1 + η n ϕ ( e n ) e n Z u n ) + p
with Q = I Z C Θ C T and p = Z C Θ f .
Based on the approximation, the recursive CRKRSL algorithm has been changed to a gradient one with variable step size η n and transformed input Z u n .
Remark 1.
The term ϕ ( e n ) in Equation (20) has a significant impact on the stability of the CRKRSL algorithm under non-Gaussian noise since CRKRSL can suppress large outliers ( e n ) with a small ϕ ( e n ) . Figure 1 shows the relation between ϕ ( e n ) and the error e n . It is clear to see that ϕ ( e n ) in CRKRSL is bigger than κ ( e n ) in CRMCC, when the error is small. (Note that ϕ ( e n ) = κ ( e n ) if γ = 0 ). Moreover, when γ = 0 , the CRKRSL degenerates to an efficient CRMCC [26] algorithm and ϕ ( e n ) reaches the maximum at e n = 0 . When γ > 0 , ϕ ( e n ) reaches the maximum at local points around e n = 0 with a larger increment, resulting in a faster convergence speed and higher accuracy than CRMCC.
However, Equation (20) is not stable at the initial update phase, since the variable step size η n is large with a small instant n. Especially, when the kernel width of KRSL is small, the step size η n even exceeds the convergence range, leading to the attenuation of filtering performance. To overcome this unfavorable factor, we introduce a gradient strategy with a fixed step size μ to replace the update of Equation (20) in the initial L iterations, which is described by
w n = Q ^ ( w n 1 + μ ϕ ( e n ) e n u n ) + p ^
with Q ^ = I C ( C T C ) 1 C T and p ^ = C ( C T C ) 1 f .
Finally, the CRKRSL is summarized in Algorithm 1.
Algorithm 1: The CRKRSL Algorithm.
        Input:
        Data pair { u n , d n } R m × R , n = 1 , 2 , .
 
        Initialization:
        Choose step-size μ ; kernel width σ ; risk-sensitive γ ; initial iterative length L;
        training size N t r ;
        initial weight w 0 = 0 .
             e n = d n w n 1 T u n
            for n = 1 : L
                w n = Q ^ ( w n 1 + μ ϕ ( e n ) e n u n ) + p ^
            end
            for n = ( L + 1 ) : N t r
                w n = Q ( w n 1 + η n ϕ ( e n ) e n Z u n ) + p
            end
Remark 2.
In Equation (20), the constant inverse matrix Θ = ( C T Z C ) 1 needs to be calculated only once before the update. On the contrary, ( C T U n C ) 1 in Equation (15) needs to be updated iteratively. Meanwhile, the update of matrix U n by Equation (10) is avoided by using Equation (20). Therefore, the proposed CRKRSL with Equation (20) has a lower computational complexity than one with Equation (15).

3. Stability Analysis

To obtain the stability condition of mean square weight error of CRKRSL, we define the weight error and model error as
w ˜ n = w n w o
w Δ = w * w o
with the optimal weight
w o = lim n ( U n d n + U n C θ n ) = lim n U n d n + lim n U n C ( C T U n C ) 1 ( f C T U n d n ) = lim n 1 n l = 1 n ϕ ( e l ) u l u l T 1 1 n l = 1 n ϕ ( e l ) u l d l       + lim n 1 n l = 1 n ϕ ( e l ) u l u l T 1 C C T 1 n l = 1 n ϕ ( e l ) u l u l T 1 C 1 ( f C T U n d n ) = U d + U C ( C T U C ) 1 ( f C T U d ) ,
where the robust correlation matrix and vector are defined as U = lim n [ 1 n l = 1 n ϕ ( e l ) u l u l T ] 1 = E [ ϕ ( e n ) u n u n T ] 1 and d = lim n [ 1 n l = 1 n ϕ ( e l ) u l d l ] = E [ ϕ ( e n ) u n d n ] .
Subtracting w o from both sides of Equation (20), we obtain
w ˜ n = Q ( w n 1 + η n ϕ ( e n ) Z u n d n η n ϕ ( e n ) Z u n u n T w n 1 ) + p w o = Q ( w n 1 + η n ϕ ( e n ) Z u n ( u n T w * + v n ) η n ϕ ( e n ) Z u n u n T w n 1 ) + p w o = Q ( w n 1 + η n ϕ ( e n ) Z u n ( u n T ( w Δ + w o ) + v n ) η n ϕ ( e n ) Z u n u n T w n 1 ) + p w o = Q ( w n 1 η n ϕ ( e n ) Z u n u n T w ˜ n 1 ) + η n ϕ ( e n ) Q Z u n ( u n T w Δ + v n ) + p w o = Q ( I η n ϕ ( e n ) Z u n u n T ) w ˜ n 1 + η n ϕ ( e n ) Q Z u n ( u n T w Δ + v n ) + Q w o w o + p .
Since Q is an idempotent matrix, we have Q w o w o + p = 0 and Q w ˜ n = w ˜ n . Then, we get
w ˜ n = Q ( I η n ϕ ( e n ) Z u n u n T ) w ˜ n 1 + η n ϕ ( e n ) Q Z u n ( u n T w Δ + v n ) = ( I η n ϕ ( e n ) Q Z u n u n T ) w ˜ n 1 + η n ϕ ( e n ) Q Z u n ( u n T w Δ + v n ) .
We take the expectation operator on both sides of the square norm, since the noise v n is independent with u n and the input sequence { u n } is i.i.d under assumptions A1–A2. We further get the useful results that the a priori error weight w ˜ n 1 is independent with u n and v n based on the independence assumptions [32]. Hence, the cross terms are equal to zeros. Then, we obtain
E [ w ˜ n 2 ] = E [ w ˜ n 1 F n 2 ] + η n 2 δ 2 E [ ϕ 2 ( e n ) ] E [ u n T Z Q Z u n ] + η n 2 E [ ϕ 2 ( e n ) ] w Δ T E [ u n u n T Z Q Z u n u n T ] w Δ
with
F n = E [ ( I η n ϕ ( e n ) Q Z u n u n T ) T ( I η n ϕ ( e n ) Q Z u n u n T ) ] = I 2 η n E [ ϕ ( e n ) ] E [ Q Z u n u n T ] + η n 2 E [ ϕ 2 ( e n ) ] E [ u n u n T Z Q Z u n u n T ] .
According to Isserlis’s theorem [33], the symmetric matrix with fourth-moment Gaussian variables can be separated by
V = E [ u n u n T Z Q Z u n u n T ] = E [ u n u n T Z ] E [ Q Z u n u n T ] + E [ u n u n T Z Q ] E [ Z u n u n T ] + E [ u n u n T ] E [ u n T Z Q Z u n ] = 2 Q + tr { Q Z } R ,
where tr { · } denotes the trace operator and R = Z 1 = E [ u n u n T ] is a positive define correlation matrix.
Therefore, Equation (27) can be simplified as
E [ w ˜ n 2 ] = E [ w ˜ n 1 F n 2 ] + η n 2 δ 2 E [ ϕ 2 ( e n ) ] tr { Q Z } + η n 2 E [ ϕ 2 ( e n ) ] w Δ T ( 2 Q + tr { Q Z } R ) w Δ = E [ w ˜ n 1 F n 2 ] + η n 2 E [ ϕ 2 ( e n ) ] tr { Q Z } ( w Δ T R w Δ + δ 2 )
with a simplified
F n = I 2 η n E [ ϕ ( e n ) ] Q + η n 2 E [ ϕ 2 ( e n ) ] V = I 2 η n E [ ϕ ( e n ) ] Q + η n 2 E [ ϕ 2 ( e n ) ] ( 2 Q + tr { Q Z } R ) .
Define q k and r k , k { 1 , 2 , , m } , being the kth eigenvalue of matrix Q and R . To ensure the square stability, the eigenvalues of F n should satisfy the following condition:
| 1 2 η n E [ ϕ ( e n ) ] q k + η n 2 E [ ϕ 2 ( e n ) ] ( 2 q k + tr { Q Z } r k ) | < 1 , for k = 1 , 2 , , m .
Then, the convergence condition about the step size can be expressed as
η n < min i ( 2 E [ ϕ ( e n ) ] q k E [ ϕ 2 ( e n ) ] ( 2 q k + tr { Q Z } r k ) ) .
Since the step size η n = ( n E [ ϕ ( v n ) ] ) 1 is related to the iteration n, finally, we obtain the weight square error stability condition about the iteration n by
n > max i ( E [ ϕ 2 ( e n ) ] ( 2 q k + tr { Q Z } r k ) 2 E [ ϕ ( v n ) ] E [ ϕ ( e n ) ] q k ) ,
where the nonlinear terms E [ ϕ ( e n ) ] and E [ ϕ 2 ( e n ) ] can be approximated by the Taylor expansion.
Remark 3.
Inequality (34) implies that the iteration n should be sufficiently large to guarantee the convergence. When n is small, CRKRSL with Equation (20) cannot satisfy (34), resulting in fluctuations at the initial update stage. Therefore, it is reasonable to use Equation (21) to replace Equation (20) to improve the convergence speed and filtering accuracy.

4. Results and Discussion

In this section, we will show the advantages of the CRKRSL algorithm on the filtering accuracy and computational complexity for both low-dimensional and high-dimensional inputs. The noise model, data selection and algorithms comparison are given as follows, respectively.
Noise model: We first consider a pure Gaussian noise to test the filtering accuracy, i.e., Gaussian noise with v n N ( 0 , δ 2 ) , where N ( μ ¯ , δ ¯ 2 ) denotes the Gaussian distribution with mean μ ¯ and variance δ ¯ 2 . Then, a mixed noise is considered to test the robustness of CRKRSL. The mixed noise model under a probability process is denoted as
v n = b ( n ) v 1 ( n ) + ( 1 b ( n ) ) v 2 ( n ) ,
where b ( n ) { 0 , 1 } satisfies the binary distribution with probability P { b ( n ) = 0 } = 0.1 and P { b ( n ) = 1 } = 0.9 . v 1 ( n ) with high probability generates the ordinary noise by a Gaussian distribution N ( 0 , δ 1 2 ) and v 2 ( n ) with low probability generates a few impulsive noise. Two types of v 2 ( n ) are considered: (a) Gaussian noise with large variance, i.e., v 2 ( n ) N ( 0 , 100 ) ; (b) α -stable noise [34] with parameter function F ( 0.8 , 0 , 0.1 , 0 ) .
Data selection: The training inputs sampled from a Gaussian distribution with zero mean and variance matrix R and 5000 samples are chosen for the simulation. The parameters C , f , R are configured as the same as in [7]. Note that there exist two sets of data with different input dimensions in [7]. The underlying dimension is either m = 7 or m = 31 . The simulated mean square deviation (MSD) is define as MSD ( dB ) = 10 log 10 ( w ˜ n 2 ) and the steady-state MSD is defined as the mean of last 1000 samples. The obtained results are averaged over 500 Monte Carlo trials. The results were run on MATLAB version R2020b on a Windows 10 operating system, configured with an Intel(R) Core(TM) i7-8700 CPU 3.20 GHz and RAM 16 GB.
Compared algorithms: The constrained algorithms, including the CLMS [7], CMCC [21], CLLAF [25] CRLS [9], and CRMCC [26], are chosen to be compared with the proposed CRKRSL. For fair comparison, the kernel widths σ of CMCC, CRMCC, and CRKRSL are set by the same value; the regularization terms of CRLS and CRMCC are set as 0.001; the initial iterative length of CRKRSL is set as the same the input dimension, i.e., L = m .

4.1. Low-Dimensional Input

In this part, the input dimension m and constrained dimension q are set as m = 7 and q = 3 , respectively. The noise considered here satisfies the following conditions, i.e., Gaussian noise v ( n ) N ( 0 , 0.1 ) and mixed noise v 1 ( n ) N ( 0 , 0.1 ) . To reflect the influence of risk-sensitive parameter γ on the MSD, the relations between γ and steady-state MSD are shown in Figure 2 and Figure 3 under different noise models. From Figure 2, one can see that CRKRSL has a stable steady-state MSD under Gaussian noise. Therefore, γ has little influence on the performance of CRKRSL. From Figure 3, it is observed that the CRKRSL is sensitive to γ under mixed noise with model (a) and achieves the lowest steady-state MSD around γ = 2.5 . Therefore, we choose γ = 2.5 for the following algorithm comparison in Figure 4 and Figure 5. Note that ρ denotes the shape parameter in the CLLAF algorithm. All necessary parameters are given in the figures.
From Figure 4, one can see that CRKRSL, CRMCC and CRLS coincide and have almost the same MSDs under Gaussian noise. This implies that the CRKRSL deals with the Gaussian noise well by choosing a large kernel width. From Figure 5, it is observed that CRKRSL has the best performance among all constrained algorithms since the risk-sensitive parameter γ can avoid fluctuations caused by a small kernel width. Moreover, CRKRSL is more stable than CRMCC at the initial stage, potentially leading to a better filtering performance. In Table 1, we further compare the consumed time at each iteration and steady-state MSDs of each algorithm under mixed noise with model (a). The consumed time of CRKRSL is far less than that of CRMCC and CRLS.
To show the advantage of CRKRSL on computational complexity, Table 2 lists the compared results of all mentioned algorithms at each iteration. One can see that CRKRSL has a lower computational complexity than CRLS and CRMCC by avoiding calculating the inverse matrix. Although the inverse matrix U n of CRMCC is not required to be calculated, the inverse matrix ( C T U n C ) 1 is still needed to be calculated at each iteration, resulting in a high computational complexity with a large q.
To test the performance of CRKRSL under mixed noise with model (b), Figure 6 gives the MSD results of all mentioned algorithms. One can see from Figure 6 that CRKRSL, CRMCC, CMCC and CLLAF show strong robustness to outliers, and CRKRSL has the lowest MSD, whereas the CLMS and CRLS are not stable due to being sensitive to α -stable noise.

4.2. High-Dimensional Input

In this part, the input dimension m and constrained dimension q are set as m = 31 and q = 1 , respectively. We only consider the mixed noise, since the Gaussian noise influences the performance of CRKRSL little by selecting a large kernel width. The mixed noise satisfies v 1 ( n ) N ( 0 , 1 ) .
Figure 7 and Figure 8 show the MSDs of different algorithms under mixed noise with model (a) and model (b), respectively. It is clear to see that CRKRSL shows the best performance out of all the compared algorithms both under mixed noise with model (a) and model (b). The initial iterative length L influences the convergence speed, significantly. Therefore, the initial iterative length should be not smaller than the input dimension. In Table 3, we further compare the consumed time at each iteration and steady-state MSDs of each algorithm under mixed noise with model (a). One can see that the consumed time of CRKRSL is far less than that of CRMCC. Moreover, CRKRSL has the lowest steady-state MSD value.

5. Conclusions

By introducing the linear constraints to the kernel risk-sensitive loss (KRSL), a low-complexity constrained recursive KRSL (CRKRSL) algorithm is presented with the help of average approximation. Since the risk-sensitive parameter is able to control the smoothness of the performance surface, CRKRSL achieves higher accuracy than some existing constrained recursive algorithms. Due to the inaccuracy of the average approximation with a few inputs, a fixed step-size gradient method is adopted to avoid the instability of CRKRSL at the initial update stage. Moreover, mean square analysis indicates that iteration times influence the stability of CRKRSL significantly and simulation results confirm the advantages of CRKRSL. The effectiveness of CRKRSL highly relies on the average approximation method, which has limits when coping with nonstationary signals. In the future, we focus on finding a novel approximation method to process both stationary and nonstationary signals and further improve the computational efficiency and accuracy of constrained recursive algorithms.

Author Contributions

Conceptualization, S.X. and C.Z.; methodology, S.X.; software, S.X. and Z.G.; validation, S.X.; formal analysis, S.X.; investigation, S.X. and D.Y.; resources, C.Z.; data curation, S.X. and D.Y.; writing—original draft preparation, S.X.; writing—review and editing, C.Z. and Z.G.; visualization, S.X.; supervision, C.Z.; project administration, C.Z. and D.Y.; funding acquisition, C.Z. and Z.G. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by Chongqing Social Science Planning Project (2021BS038) and Natural Science Foundation of Chongqing, China (cstc2021jcyj-bshX0035, cstc2018jcyjA2453).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. De Campos, M.L.R.; Werner, S.; Apolinário, J.A. Constrained adaptive filters. In Adaptive Antenna Arrays: Trends and Applications; Springer: New York, NY, USA, 2004; pp. 46–64. [Google Scholar]
  2. Frost, O. An algorithm for linearly constrained adaptive array processing. Proc. IEEE 1972, 60, 926–935. [Google Scholar] [CrossRef]
  3. Li, J.; Stoica, P. Robust Adaptive Beamforming; John Wiley & Sons: New York, NY, USA, 2005. [Google Scholar]
  4. Diniz, P.S. Adaptive Filtering: Algorithms and Practical Implementation; Springer: Cham, Switzerland, 1997. [Google Scholar]
  5. Wu, Q.; Li, Y.; Xue, W. A kernel recursive maximum versoria-like criterion algorithm for nonlinear channel equalization. Symmetry 2019, 11, 1067. [Google Scholar] [CrossRef] [Green Version]
  6. Verdu, S. Multiuser Detection; Cambridge University Press: Cambridge, UK, 1998. [Google Scholar]
  7. Arablouei, R.; Doğançay, K.; Werner, S. On the mean-square performance of the constrained LMS algorithm. Signal Process. 2015, 117, 192–197. [Google Scholar] [CrossRef] [Green Version]
  8. Resende, L.S.; Romano, J.M.T.; Bellanger, M.G. A fast least-squares algorithm for linearly constrained adaptive filtering. IEEE Trans. Signal Process. 1996, 44, 1168–1174. [Google Scholar] [CrossRef]
  9. Arablouei, R.; Doğançay, K. Performance analysis of linear-equality-constrained least-squares estimation. IEEE Trans. Signal Process. 2015, 63, 3762–3769. [Google Scholar] [CrossRef] [Green Version]
  10. Arablouei, R.; Doğançay, K. Reduced-complexity constrained recursive least-squares adaptive filtering algorithm. IEEE Trans. Signal Process. 2012, 60, 6687–6692. [Google Scholar] [CrossRef]
  11. Werner, S.; Apolinário, J.A.; de Campos, M.L.R.; Diniz, P. Low-complexity constrained affine-projection algorithms. IEEE Trans. Signal Process. 2005, 53, 4545–4555. [Google Scholar] [CrossRef]
  12. Apolinário, J.A., Jr.; de Campos, M.L.R.; Bernal, O.C.P. The constrained conjugate gradient algorithm. IEEE Signal Process. Lett. 2000, 12, 351–354. [Google Scholar] [CrossRef]
  13. De Campos, M.L.R.; Werner, S.; Apolinário, J.A. Constrained adaptation algorithms employing Householder transformation. IEEE Trans. Signal Process. 2002, 9, 2187–2195. [Google Scholar] [CrossRef]
  14. Arablouei, R.; Doğançay, K. Linearly-constrained recursive total least-squares algorithm. IEEE Signal Process. Lett. 2012, 12, 821–824. [Google Scholar] [CrossRef]
  15. Haykin, S. Adaptive Filter Theory; Prentice-Hall: Upper Saddle River, NJ, USA, 2002. [Google Scholar]
  16. Liu, W.; Pokharel, P.P.; Príncipe, J.C. Correntropy: Properties and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  17. Li, Y.; Wang, Y.; Sun, L. A proportionate normalized maximum correntropy criterion algorithm with correntropy induced metric constraint for identifying sparse systems. Symmetry 2018, 10, 683. [Google Scholar] [CrossRef] [Green Version]
  18. Chen, B.; Xing, L.; Zhao, H.; Zheng, N.; Príncipe, J.C. Generalized correntropy for robust adaptive filtering. IEEE Trans. Signal Process. 2016, 64, 3376–3387. [Google Scholar] [CrossRef] [Green Version]
  19. Erdogmus, D.; Príncipe, J.C. An error-entropy minimization algorithm for supervised training of nonlinear adaptive systems. IEEE Trans. Signal Process. 2002, 50, 1780–1786. [Google Scholar] [CrossRef] [Green Version]
  20. Príncipe, J.C. Information Theoretic Learning: Renyi’s Entropy and Kernel Perspectives; Springer: New York, NY, USA, 2010. [Google Scholar]
  21. Peng, S.; Chen, B.; Sun, L.; Ser, W.; Lin, Z. Constrained maximum correntropy adaptive filtering. Signal Process. 2017, 140, 116–126. [Google Scholar] [CrossRef] [Green Version]
  22. Bhattacharjee, S.S.; Shaikh, M.A.; Kumar, K.; George, N.V. Robust constrained generalized correntropy and maximum versoria criterion adaptive filters. IEEE Trans. Circuits Syst. II Express Briefs 2021, 68, 3002–3006. [Google Scholar] [CrossRef]
  23. Peng, S.; Ser, W.; Chen, B.; Sun, L.; Lin, Z. Robust constrained adaptive filtering under minimum error entropy criterion. IEEE Trans. Circuits Syst. II Express Briefs 2018, 65, 1119–1123. [Google Scholar] [CrossRef]
  24. Wang, Z.; Zhao, H.; Zeng, X. Constrained least mean M-estimation adaptive filtering algorithm. IEEE Trans. Circuits Syst. II Express Briefs 2021, 68, 1507–1511. [Google Scholar] [CrossRef]
  25. Liang, T.; Li, Y.; Zakharov, Y.V.; Xue, W.; Qi, J. Constrained least lncosh adaptive filtering algorithm. Signal Process. 2021, 183, 108044. [Google Scholar] [CrossRef]
  26. Qian, G.; Ning, X.; Wang, S. Recursive constrained maximum correntropy criterion algorithm for adaptive filtering. IEEE Trans. Circuits Syst. II Express Briefs 2020, 67, 2229–2233. [Google Scholar] [CrossRef]
  27. Chen, B.; Xing, L.; Xu, B.; Zhao, H.; Zheng, N.; Príncipe, J.C. Kernel risk-sensitive loss: Definition, properties and application to robust adaptive filtering. IEEE Trans. Signal Process. 2017, 65, 2888–2901. [Google Scholar] [CrossRef] [Green Version]
  28. Qian, G.; Dong, F.; Wang, S. Robust constrained minimum mixture kernel risk-sensitive loss algorithm for adaptive filtering. Digit. Signal Process. 2020, 107, 102859. [Google Scholar] [CrossRef]
  29. Qian, G.; Wang, S.; Wang, L.; Duan, S. Convergence analysis of a fixed point algorithm under maximum complex correntropy criterion. IEEE Signal Process. Lett. 2018, 25, 1830–1834. [Google Scholar] [CrossRef]
  30. Liu, W.; Príncipe, J.C.; Haykin, S. Kernel Adaptive Filtering: A Comprehensive Introduction; Wiley: Hoboken, NJ, USA, 2011. [Google Scholar]
  31. Zhang, X.-D. Matrix Analysis and Applications; Cambridge University Press: Cambridge, UK, 2017. [Google Scholar]
  32. Sayed, A.H. Adaptive Filters; John Wiley & Sons: Hoboken, NJ, USA, 2008. [Google Scholar]
  33. Isserlis, L. On a formula for the product-moment coefficient of any order of a normal frequency distribution in any number of variables. Biometrika 1918, 12, 134–139. [Google Scholar] [CrossRef] [Green Version]
  34. Zhang, Q.; Feng, W.; Iu, H.H.C.; Wang, S. Adaptive filters with robust augmented space linear model: A weighted k-NN method. IEEE Trans. Signal Process. 2021, 69, 6448–6461. [Google Scholar] [CrossRef]
Figure 1. The relation between ϕ ( e n ) and the error e n ( σ = 2 ).
Figure 1. The relation between ϕ ( e n ) and the error e n ( σ = 2 ).
Symmetry 14 00877 g001
Figure 2. Steady-state MSD versus risk-sensitive parameter γ under Gaussian noise ( m = 7 , q = 3 ).
Figure 2. Steady-state MSD versus risk-sensitive parameter γ under Gaussian noise ( m = 7 , q = 3 ).
Symmetry 14 00877 g002
Figure 3. Steady-state MSD versus risk-sensitive parameter γ under mixed noise with model (a) ( m = 7 , q = 3 ).
Figure 3. Steady-state MSD versus risk-sensitive parameter γ under mixed noise with model (a) ( m = 7 , q = 3 ).
Symmetry 14 00877 g003
Figure 4. MSDs of compared algorithms under Gaussian noise ( m = 7 , q = 3 ).
Figure 4. MSDs of compared algorithms under Gaussian noise ( m = 7 , q = 3 ).
Symmetry 14 00877 g004
Figure 5. MSDs of compared algorithms under mixed noise with model (a) ( m = 7 , q = 3 ).
Figure 5. MSDs of compared algorithms under mixed noise with model (a) ( m = 7 , q = 3 ).
Symmetry 14 00877 g005
Figure 6. MSDs of compared algorithms under mixed noise with model (b) ( m = 7 , q = 3 ).
Figure 6. MSDs of compared algorithms under mixed noise with model (b) ( m = 7 , q = 3 ).
Symmetry 14 00877 g006
Figure 7. MSDs of compared algorithms under mixed noise with model (a) ( m = 31 , q = 1 ).
Figure 7. MSDs of compared algorithms under mixed noise with model (a) ( m = 31 , q = 1 ).
Symmetry 14 00877 g007
Figure 8. MSDs of compared algorithms under mixed noise with model (b) ( m = 31 , q = 1 ).
Figure 8. MSDs of compared algorithms under mixed noise with model (b) ( m = 31 , q = 1 ).
Symmetry 14 00877 g008
Table 1. Consumed time and steady-state MSD of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL under mixed noise with model (a) based on low-dimensional input.
Table 1. Consumed time and steady-state MSD of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL under mixed noise with model (a) based on low-dimensional input.
AlgorithmConsumed Time (s)Steady-State MSD (dB)
CLMS 0.0024 6.70
CMCC 0.0028 10.86
CLLAF 0.0025 18.58
CRLS 0.0324 21.23
CRMCC 0.0327 27.75
CRKRSL 0.0038 34.34
Table 2. Computational complexity of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL at each iteration.
Table 2. Computational complexity of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL at each iteration.
AlgorithmsComputational Complexity
CLMS O ( m 2 )
CMCC O ( m 2 )
CLLAF O ( m 2 )
CRLS O ( m 3 + q 3 )
CRMCC O ( m 2 + q 3 )
CRKRSL O ( m 2 )
Table 3. Consumed time and steady-state MSD of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL under mixed noise with model (a) based on high-dimensional input.
Table 3. Consumed time and steady-state MSD of CLMS, CMCC, CLLAF, CRLS, CRMCC and CRKRSL under mixed noise with model (a) based on high-dimensional input.
AlgorithmConsumed Time (s)Steady-State MSD (dB)
CLMS 0.0035 2.98
CMCC 0.0034 2.21
CLLAF 0.0034 3.13
CRLS 0.0112 11.02
CRMCC 0.0114 15.45
CRKRSL 0.0063 19.72
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Xiang, S.; Zhao, C.; Gao, Z.; Yan, D. Low-Complexity Constrained Recursive Kernel Risk-Sensitive Loss Algorithm. Symmetry 2022, 14, 877. https://doi.org/10.3390/sym14050877

AMA Style

Xiang S, Zhao C, Gao Z, Yan D. Low-Complexity Constrained Recursive Kernel Risk-Sensitive Loss Algorithm. Symmetry. 2022; 14(5):877. https://doi.org/10.3390/sym14050877

Chicago/Turabian Style

Xiang, Shunling, Chunzhe Zhao, Zilin Gao, and Dongfang Yan. 2022. "Low-Complexity Constrained Recursive Kernel Risk-Sensitive Loss Algorithm" Symmetry 14, no. 5: 877. https://doi.org/10.3390/sym14050877

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop