Next Article in Journal
MFAN: Multi-Level Features Attention Network for Fake Certificate Image Detection
Previous Article in Journal
On the Depth of Decision Trees with Hypotheses
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variational Bayesian-Based Improved Maximum Mixture Correntropy Kalman Filter for Non-Gaussian Noise

Department of Intelligent Systems Science and Engineering, Harbin Engineering University, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Entropy 2022, 24(1), 117; https://doi.org/10.3390/e24010117
Submission received: 29 November 2021 / Revised: 7 January 2022 / Accepted: 10 January 2022 / Published: 12 January 2022

Abstract

:
The maximum correntropy Kalman filter (MCKF) is an effective algorithm that was proposed to solve the non-Gaussian filtering problem for linear systems. Compared with the original Kalman filter (KF), the MCKF is a sub-optimal filter with Gaussian correntropy objective function, which has been demonstrated to have excellent robustness to non-Gaussian noise. However, the performance of MCKF is affected by its kernel bandwidth parameter, and a constant kernel bandwidth may lead to severe accuracy degradation in non-stationary noises. In order to solve this problem, the mixture correntropy method is further explored in this work, and an improved maximum mixture correntropy KF (IMMCKF) is proposed. By derivation, the random variables that obey Beta-Bernoulli distribution are taken as intermediate parameters, and a new hierarchical Gaussian state-space model was established. Finally, the unknown mixing probability and state estimation vector at each moment are inferred via a variational Bayesian approach, which provides an effective solution to improve the applicability of MCKFs in non-stationary noises. Performance evaluations demonstrate that the proposed filter significantly improves the existing MCKFs in non-stationary noises.

1. Introduction

The state estimation problem in dynamic systems is an important research topic in engineering applications and scientific research. As an excellent optimal state-space estimator, the Kalman filter (KF) is commonly applied in various fields like control systems and signal processing. Unfortunately, the optimality of KF requires exact system models and ideal noise conditions as summarized in [1].
The widely used KF, which usually refers to the Kalman filter based on the Hidden Markov Models (HMM), has rigorous requirements for the noise models. Both process and measurement noise are assumed as ideal independent Gaussian noise sequences. However, in practical applications, ideal noise conditions are not likely, and model uncertainties such as system structure changes and environmental disturbances are generally inevitable. Moreover, unexpected noise interference, such as colored noise and non-Gaussian noise, widely exist, and the performance of the KF is likely to worsen when applied to such situations. For example, when the independent noise assumption no longer holds, and the colored noise needs to be considered in the system model, as one of the efficient solutions, the pairwise Markov models (PMM), which can be deemed a general form of HMM, can be taken as an efficient improvement scheme for KF. In [2], the framework of the KF based on PMM is derived, which allows the cross-dependence between observations conditionally on hidden variables. As an extension, the KF based on triplet Markov chains is provided in [3]. These methods provided efficient solutions for the colored noise filtering from the perspective of the state-space models. On this basis, to solve the model parameters’ uncertain problems in practical applications. The robust parameter estimation problem for the PMM model is further explored in [4,5], which provided several representative schemes for the state estimation problems in colored noise or model parameter uncertain conditions.
On the other hand, the research on filtering algorithms under non-Gaussian noise has also attracted extensive attention. Generally, non-Gaussian distribution noises are often caused by impulsive noise, which gives the overall noise distributions obvious heavy-tailed features (such as some special Gaussian-mixture distributions). Extensive research has been carried out regarding the filtering problems in non-Gaussian noises. H-Infinity filtering and m-estimators are typical robust methods used for this purpose [6]. The Huber Kalman filter (HKF) is one of the most representative m-estimators, which retains a consistent form with KF and provides reliable robustness to external impulsive noise interference. In the existing relevant studies, various improvements based on the HKF have also been proposed, such as a batch mode [7] and several typical nonlinear schemes [8,9]. Besides m-estimators, student-t distribution filters have also been designed to deal with state estimate problems under heavy-tailed noise conditions. The filters attenuate the interference of outliers by their heavy-tailed distribution statistical characteristics [10,11]. However, student-t filters suffer from high-order statistics’ loss due to moment matching and inaccurate parameter selection, such as the degrees of freedom limiting their applicability. To solve the problem, another novel student-t filter based on the hierarchical Gaussian state-space model and variational Bayes (VB) inference approach is proposed, which is compactly called the robust student-t Kalman filter (RSTKF) [12]. Compared with the student-t filter, several unknown parameters, such as the degrees of freedom of the student-t distribution and the scale coefficient, are updated accordingly in the RSTKF [13,14,15].
Recently, a new robust maximum correntropy Kalman filter (MCKF) was proposed for the filtering problem in non-Gaussian noises [16,17]. The MCKF is derived on the maximum correntropy criterion (MCC), which introduces the Gaussian correntropy as a novel local similarity measure, and a novel object function based on correntropy is utilized to overcome the non-Gaussian noise interference [18]. Compared with other existing robust filters, the Gaussian correntropy-based MCKF has better statistical characteristics [19]. Meanwhile, the MCKF can be formulated in the same form as the KF. Similarly, the states constrained [20] and nonlinear forms [21,22] have been proposed and validated in various applications [23]. Moreover, an adaptive MCKF was proposed to improve the filtering performance by updating the measurement noise covariance via VB approach [24].
Due to the excellent robustness to outliers, in many cases, the filtering problem of a linear system corrupted by impulsive noise can be solved effectively by the MCKF. However, just like other robust filters, the filtering performance of the MCKF is closely related to its initial parameters, which are usually obtained by experience or trial and error method. Similar to many existing robust filtering algorithms with fixed parameters, it might result in performance degradation in non-stationary noises. Several studies have focused on the parameter problem of correntropy function. In prior works, heuristic solutions were proposed to adjust the kernel bandwidth during filtering [25,26]. As it is difficult to directly find an optimal kernel bandwidth during filtering, the mixture correntropy concept was proposed in [27,28], the method takes mixture Gaussian correntropy with different kernel parameters instead of the Gaussian correntropy, and a new maximum mixture correntropy criterion (MMCC) is derived from replacing MCC. However, the mixing probability of the mixture correntropy needs to be configured manually and fixed, which results in performance degradation similar to that of the MCKF.
To improve the filtering performance when applied to non-stationary noise conditions, it may be an efficient improved solution that considers the model’s switching scenarios. For instance, in [29], the optimal recursive filtering method is studied for non-Gaussian Markov switching models, in which a semi-supervised parameter estimation method is used. In addition, for the non-stationary noise conditions concerned in this paper, the variable Bayesian approach is an implementation scheme worthy of consideration. Based on the triplet Markov chain model, the general form of variable Bayesian inference is deduced in [30], and a structured variable Bayesian inference frame with regime switching is obtained. For typical linear filtering applications, the RSTKF proposed in [12] also adopts a similar operation scheme, where VB approximation is operated by rescaling the covariance and inferring parameters to deal with the non-stationary non-Gaussian noises. In [31], a similar problem is further studied, and a VB-based robust student’s t KF is applied to the linear PMM systems, which extends the filter’s applicability to more general conditions, and independent noise assumption is no longer needed. Therefore, inspired by the related references, the model switch concept is adopted in this work, and the variational Bayesian approach is also operated to infer estimation results. In view of the filtering accuracy degradation problem of MCKF in the non-stationary and non-Gaussian noise, a series of studies are carried out.
In this work, an improved maximum mixture Kalman filter (IMMCKF) is therefore proposed, the intermediate random variables are used to represent the mixing probability of mixture correntropy. The state variables and mixing probability are approximated by the derived variational Bayesian approach. Compared with the existing MCKF algorithms, the numerical test results demonstrate that the proposed filter could deal with the filtering problem well in a non-stationary non-Gaussian noises environment. The contributions of this paper are listed as follows:
(a)
The accuracy degradation problem of existing fixed-parameter robust filtering algorithms in non-stationary noise environment is considered. By analyzing, it is inferred that the mixture correntropy function can be improved as the breakthrough point and applied to the filtering problem in such noise conditions. On this basis, a novel improved robust filtering algorithm is then further derived.
(b)
By employing Beta-Bernoulli distributed intermediate random variables, a new hierarchical Gaussian state-space model is therefore derived, the system state vector, and the unknown variables are simultaneously estimated by utilizing the variational Bayesian technique.
(c)
Through analyses and derivations, the necessary selection strategy of initial parameters is derived, and the numerical test results show that the filtering performance is enhanced obviously after several iterations. On the one hand, it achieved desired robust filtering performance when applied to non-stationary noise conditions. On the other hand, the proposed algorithm effectively avoids the possible filtering divergence difficulties of MCKF in practical application.
This paper is organized as follows: In Section 2, we review the concept of correntropy and existing MCKF. In Section 3, a variational Bayesian-based improved maximum mixture correntropy KF is derived to solve the filtering problem in non-stationary non-Gaussian noises, in which the variational Bayesian approach is applied in the proposed filter. Section 4 provides performance evaluations and analysis, demonstrating the advantages of the proposed filter in different noise conditions. Conclusions are given in Section 5.

2. Materials and Methods

2.1. Definition of Correntropy and Properties

Correntropy is a useful local similarity measure tool for state estimation in heavy-tailed noise environments. Given two random variables X , Y R with a joint distribution function F X Y ( x , y ) , correntropy can be defined as
V ( X , Y ) = E [ κ σ ( X , Y ) ] = κ σ ( x , y ) d F X Y ( x , y )
where κ σ ( · ) is a shift-invariant Mercer kernel. In this work, the Gaussian kernel function is given by
κ σ ( X , Y ) = G σ ( e ) = exp ( e 2 2 σ 2 )
where e = X Y , and σ > 0 denotes the kernel bandwidth. In order to make correntropy applicable to complex noise conditions, the default Gaussian kernel function can be extended into a mixture correntropy form as follows:
M ( X , Y ) = E [ ρ G σ 1 ( e ) + ( 1 ρ ) G σ 2 ( e ) ]
where G σ 1 , G σ 2 represent the Gaussian correntropy with two different kernel bandwidth parameters, and ρ [ 0 , 1 ] represents the mixing probability. For convenience of application, (3) can be approximately expressed as
M ( X , Y ) = 1 N i = 1 N ρ exp ( e i 2 2 σ 1 2 ) + ( 1 ρ ) exp >( e i 2 2 σ 2 2 )
where N represents the number of samples. The mixture correntropy can be taken as a generalization of the original correntropy, if ρ = 1 or ρ = 0 , it reduces to the Gaussian correntropy with single kernel parameter and M ( X , Y ) = 1 if X = Y .

2.2. Robust Kalman Filter Based on Maximum Correntropy Criterion

Consider the linear state-space system based on HMM, the equations to be
x k = F k x k 1 + w k
y k = H k x k + v k
where k is a discrete time index, x k R n is the system state vector at discrete-time index k, F k is the state transition matrix, y k R m is the measurement vector, H k is the measurement matrix; w k and v k are zero-mean process and measurement noise vector with nominal covariance Q k and R k , respectively. It is assumed that both process and measurement noise are statistically independent and time uncorrelated. When the process and measurement noises are assumed ideal Gaussian distributions, and initial state x 0 is random Gaussian variables, the state can be inferred by the KF, which is an optimal filter in the minimum mean square error (MMSE) criterion. Additionally, the quadratic objective function can be formulated as follows:
x ^ k | k = arg min x k J ( x k ) = arg min x k 1 2 x k x ^ k | k 1 P k | k 1 1 2 + 1 2 y k H k x k R k 1 2
where x A 2 = x T Ax . To minimize Equation (7), the KF is implemented in two steps as below, the prior estimation is
x ^ k | k 1 = F k x ^ k 1 | k 1
P k | k 1 = F k P k 1 | k 1 F k T + Q k 1
and the posterior measurement update is
K k = P k | k 1 H k T ( H k P k | k 1 H k T + R k ) 1
x ^ k = x ^ k | k 1 + K k ( y k H k x ^ k | k 1 )
P k | k = ( I n K k H k ) P k | k 1
where P k | k 1 and P k | k represents the prior and posterior covariance matrix, respectively. The KF is an optimal estimator in an ideal white Gaussian noise environment. In Equation (7), only the second-order statistics are used during the state update, and the KF is susceptible to non-Gaussian noise interference.
To solve the filtering problem in non-Gaussian noise conditions, recently, MCC is often considered an efficient potential solution. As described in Section 2, correntropy has several necessary properties that make it capable of dealing with the non-Gaussian estimation problem. Different from the global similarity measure-mean square error (MSE), which only contains the second-order statistics, Gaussian correntropy incorporates all even order moments [17,18]. In geometric meaning, MSE gives the L2 norm distance while correntropy offers a hybrid norm distance, where the correntropy behaves like the L2→L0 norm with the increasing difference between two points.
As proved by the research in [17], maximizing the correntropy of two different random variables can be used as a criterion in dealing with the non-Gaussian noise problem, especially the filtering problem in heavy-tailed noises, which leads to MCC. Therefore, to enhance the robustness of the Kalman filter, the objective function based on MCC is introduced to replace the original quadratic cost function, and the new objective function therefore can be formulated as:
x ^ k = argmax x k J MCC ( x k ) = arg max x k Σ i = 1 n a i G σ m ( e i ) + Σ j = 1 m b j G σ p ( f j )
where e = y k H k x ^ k | k 1 R k 1 , f = x ^ k F k x ^ k 1 | k 1 P k | k 1 1 the subscript i represents the ith element of the vector; a and b are tuning parameter vectors; and σ m , σ p denotes the kernel bandwidths corresponding to the R k and P k | k 1 , respectively. For simplicity, it is assumed that σ m = σ p = σ , to ensure the filter converges to the KF when the kernel bandwidth goes to infinity, assuming that a i = σ , b j = σ [21]. The prior estimation steps of the MCKF are the same as in Equations (8) and (9) and the posterior state estimation of x k is obtained by below KF like equations with the updated filter gain as
K ¯ k = P ¯ k | k 1 H k T ( H k P ¯ k | k 1 H k T + R ¯ k ) 1
P ¯ k | k 1 = S P k | k 1 Λ p 1 S P k | k 1 T
R ¯ k = S R k Λ m 1 S R k T
where Λ p = diag ( G σ ( f 1 ) G σ ( f 2 ) . . . G σ ( f n ) ) , Λ m = diag ( G σ ( e 1 ) G σ ( e 2 ) . . . G σ ( e m ) ) , and P k | k 1 = S P k | k 1 S P k | k 1 T , R k = S R k S R k T . It is worth noting that the MCC solution cannot be obtained in a closed form and usually solves it using an iterative update algorithm such as the fixed-point iterative algorithm, which involves no step size and may converge fast, the condition that guarantees the convergence of the fixed point MCC was given in [32].

3. Main Results

3.1. Robust Kalman Filter Based on Mixture Correntropy Criterion

The kernel parameters of Gaussian correntropy determine the filtering performance of the MCKF, and an improper kernel bandwidth might lead to filtering performance degradation or even diverge. For this reason, the mixture correntropy is an alternative solution to improve the solution of problem by reducing the filter’s sensitivity to the kernel parameters.
The maximum mixture correntropy Kalman filter (MMCKF) is derived in this section. According to the mixture correntropy Function (4), the objective function of the maximum mixture correntropy criterion can be formulated as
x ^ k = arg max x k J MMCC ( x k ) = arg max x k Σ i = 1 n a i M ( e i ) + Σ j = 1 m b j M ( f j )
where M ( e i ) represents the mixture correntropy function as (3). Similar to Equation (13), it is assumed that a i = b j = λ , and mixture correntropy is the convex combination of the two Gaussian correntropy functions. To maximize this objective function, the solution can be obtained by solving
J MMCC ( x k ) x k = λ Σ i = 1 n Ø i M ( e i ) + Σ j = 1 m Ø j M ( f j ) = 0
where
Ø i = ρ G σ 1 ( e i ) σ 1 2 + ( 1 ρ ) G σ 2 ( e i ) σ 2 2
represents the derivation of the correntropy function.
To maintain consistency to the KF, a tuning factor λ should be properly assigned as
λ = σ 1 2 σ 2 2 ρ σ 2 2 + ( 1 ρ ) σ 1 2
and the MMCKF converges to the optimal KF while the process and measurement noises obey ideal Gaussian distributions. A modified mixture correntropy function C ( e i ) is then formulated as
C ( e i ) = λ Ø i = μ G σ 1 ( e i ) + ( 1 μ ) G σ 2 ( e i )
with μ = ρ σ 2 2 ρ σ 2 2 ( ρ σ 2 2 + ( 1 ρ ) σ 1 2 ) ( ρ σ 2 2 + ( 1 ρ ) σ 1 2 ) , and the optimal estimate of x k is obtained by KF like equations, where Λ p = diag ( C ( f 1 ) C ( f 2 ) . . . C ( f n ) ) , Λ m = diag ( C ( e 1 ) C ( e 2 ) . . . C ( e m ) ) , and μ [ 0 , 1 ] . Here, C ( e i ) can be regarded as a linear transformation of M ( e i ) , they are positive correlated, and C ( e i ) = M ( e i ) if ρ , μ 0 , 1 .

3.2. Improved MMCKF via Variational Bayesian Approximation

Similar to the Gaussian correntropy of the MCKF, in Equation (21), the mixing probability parameters are constant, which inevitably results in the filtering performance degradation in non-stationary noises. To improve the filtering performance in such complex noise environments, an improved algorithm is derived in this section. The mixing probability of the mixture correntropy is reassigned as an unknown random variable that can be further approximated during filtering.
According to the Bayesian theorem, the posterior probability density function (PDF) p ( x k | y 1 : k ) is formulated as
p ( x k | y 1 : k ) p ( y k | x k ) p ( x k | y 1 : k 1 )
in which p ( x k | y 1 : k 1 ) is determined using the Chapman–Kolmogorov equation. Therefore,
p ( x k | y 1 : k 1 ) p ( x k | x k 1 ) p ( x k 1 | y 1 : k 1 ) d x k 1
where p ( x k | x k 1 ) is the posterior density at time k 1 and p ( x k 1 | y 1 : k 1 ) is the transition density.
In this work, we assumed that μ is an unknown random variable, and μ [ 0 , 1 ] . In order to approximate the unknown state and reasonable mixing probability via VB approach, we first need to determine the assumed distribution of unknown variable. Then, it is necessary to introduce a prior probability distribution p ( μ ) . According to the desired scenario, μ is assumed to be an Beta distributed variable. Since the likelihood distribution of the mixing probability μ can be formulated as a Bernoulli distribution by introducing Bernoulli random variables ξ , according to the Bayesian probability theory, we have
p ( μ | ξ ) = p ( ξ | μ ) · p ( μ ) p ( ξ | μ ) · p ( μ ) d ξ
and then, p ( μ | ξ ) p ( ξ | μ ) · p ( μ ) . According to the Bayesian formula, the posterior probability distribution is directly proportional to the product of the prior probability distribution and the likelihood function. The form of the posterior probability distribution will be the same as that of the prior. Therefore, the conjugate prior distribution of μ is selected as a Beta distribution according to Bayesian probability theory, have p μ = B e μ ; a , b , where E ( μ ) = a / ( a + b ) .
Now that we have determined the distribution of the required intermediate variables, we can further derive the following specific formula based on the system model. Using two random variables s k and t k obey Bernoulli distribution to generate the likelihood distribution. In order to ensure that the prior and posterior probability distributions retain the same form, the Beta distribution is taken as the prior conjugation distribution for inference. Therefore, the conditional probability p ( t k | α k ) and p ( s k | β k ) can be formulated as
p ( t k | α k ) = α k t k ( 1 α k ) ( 1 t k ) , s . t . t k { 0.1 }
p ( s k | β k ) = β k s k ( 1 β k ) ( 1 s k ) , s . t . s k { 0.1 }
in which α k and β k obey the Beta distribution as
p ( α k ) = Be ( α k ; a 0 , b 0 )
p ( β k ) = Be ( β k ; c 0 , d 0 )
where a 0 , b 0 , c 0 and d 0 represent the prior Beta parameters for the mixing probabilities. Then, the conditional PDF p ( y k | x k , t k ) and p ( x k | y 1 : k 1 , s k ) can be rewritten as the following hierarchical Gaussian form by
p ( y k | x k , t k ) = N ( y k ; H k x k , R ˜ k ) t k N ( y k ; H k x k , R ¯ k ) ( 1 t k )
p ( x k | y 1 : k 1 , s k ) = N ( x k ; x ^ k | k 1 , P ˜ k | k 1 ) s k N ( x k ; x ^ k | k 1 , P ¯ k | k 1 ) ( 1 s k )
where R ˜ k , P ˜ k | k 1 and R ¯ k , P ¯ k | k 1 represent the updated covariance matrix corresponding to the Gaussian correntropy functions G σ 1 and G σ 2 , respectively. The modified prior error covariance is formulated as P ˜ k | k 1 = S P k | k 1 Λ p , σ 1 1 S P k | k 1 T and P ¯ k | k 1 = S P k | k 1 Λ p , σ 2 1 S P k | k 1 T , and the measurement noise covariance is formulated as R ˜ k = S R k Λ m , σ 1 1 S R k T and R ¯ k = S R k Λ m , σ 2 1 S R k T .
Therefore, the conditional probability density distribution of p ( y k | x k , α k , t k ) and p ( x k | y 1 : k 1 , β k , s k ) can be formulated as
p ( y k | x k , α k , t k ) = N ( y k ; H k x k , R ˜ k ) t k N ( y k ; H k x k , R ¯ k ) ( 1 t k ) p ( t k | α k ) p ( α k ) s . t . t k { 0 , 1 }
p ( x k | y 1 : k 1 , β k , s k ) = N ( x k ; x ^ k | k 1 , P ˜ k | k 1 ) s k N ( x k ; x ^ k | k 1 , P ¯ k | k 1 ) ( 1 s k ) p ( s k | β k ) p ( β k ) s . t . s k { 0 , 1 }
According to the likelihood PDF derived above and the Bayesian theorem, the joint PDF can be given as follows:
p ( Θ k , y 1 : k ) = p ( y 1 : k 1 ) N ( x k ; x ^ k | k 1 , P ˜ k | k 1 ) s k N ( x k ; x ^ k | k 1 , P ¯ k | k 1 ) ( 1 s k ) N ( y k ; H k x k , R ˜ k ) t k N ( y k ; H k x k , R ¯ k ) 1 t k α k t k ( 1 α k ) ( 1 t k ) β k s k ( 1 β k ) ( 1 s k ) Be ( α k ; a 0 , b 0 ) Be ( β k ; c 0 , d 0 )
where Θ k = Δ x k , s k , t k , α k , β k contains the state variable x k that needs to be estimated, and the Beta-Bernoulli variables α k , β k , s k , t k are used for the inference of mixing probability parameters. Therefore, to obtain the estimated value of Θ using the VB inference, according to prior work [12,33], the approximate posterior PDF of the element in Θ needs to satisfy the following equation:
log q ( θ ) = E Θ k θ [ log p ( Θ k , y 1 : k ) ] + C θ
where θ is an element of Θ k , Θ θ means the remain elements of Θ except θ , and C θ is a constant with respect to θ . As (34) cannot be solved analytically, fixed-point iteration is required to achieve an approximate solution. By expanding (34), the following can be obtained:
log p ( Θ k , y 1 : k ) = ( 1 s k ) log N ( x k ; x ^ k | k 1 , P ¯ k | k 1 ) + s k logN ( x k ; x ^ k | k 1 , P ˜ k | k 1 ) + ( 1 t k ) logN ( y k ; H k x k , R ¯ k ) + t k log N ( y k ; H k x k , R ˜ k ) + t k log α k + ( 1 t k ) log ( 1 α k ) + s k log β k + ( 1 s k ) log ( 1 β k ) + logBe ( α k ; a 0 , b 0 ) + log Be ( β k ; c 0 , d 0 ) + C Θ k
Initialize the parameters a 0 , b 0 , c 0 , d 0 and calculate E ( i + 1 ) [ t k ] = a 0 / ( a 0 + b 0 ) , E ( i + 1 ) [ s k ] = c 0 / ( c 0 + d 0 ) , note that E ( 0 ) [ log ( α 0 ) ] = ψ ( a 0 ) ψ ( a 0 + b 0 ) , E ( 0 ) [ log ( β 0 ) ] = ψ ( c 0 ) ψ ( c 0 + d 0 ) , where ψ represents the digamma function.
By exploiting (35), the estimation of the unknown parameters in Θ k is implemented by fixed-point iteration loop in (a)–(c).
( a ) Let θ = x k , utilizing (34) in (35), and log q ( i + 1 ) ( x k ) can be rewritten as
log q ( i + 1 ) ( x k ) = E ( i ) [ s k ] log N ( x k ; H k x ^ k | k ( i ) , P ˜ k | k ( i ) ) + ( 1 E ( i ) [ s k ] ) log N ( x k ; H k x ^ k | k ( i ) , P ¯ k | k ( i ) ) E ( i ) [ t k ] log N ( y k ; H k x k , R ˜ k ( i ) ) + ( 1 E ( i ) [ t k ] ) log N ( y k ; H k x k , R ¯ k ( i ) ) + C x k
where i represents the ith fixed-point iterations. Using E ( i + 1 ) [ s k ] and E ( i + 1 ) [ t k ] replace the mixing probability ρ in (21) and then calculate the conditional PDF p ( x k | y 1 : k 1 ) and p ( y k | x k ) by proposed mixture correntropy as
q ( i + 1 ) ( x k | y 1 : k 1 ) q ( i + 1 ) ( y k | x k ) = N ( x k ; x ^ k 1 , P ^ k | k 1 ) N ( y k ; H k x k , R ^ k )
where the modified prediction error covariance matrix P ^ k | k 1 ( i + 1 ) and the modified measurement noise covariance matrix R ^ k ( i + 1 ) are used. After deformation, the posterior PDF q ( i + 1 ) ( x k ) can be formulated as
q ( i + 1 ) ( x k ) N ( x k ; H k x ^ k | k 1 ( i ) , P ^ k | k 1 ( i ) ) N ( y k ; H k x k , R ^ k ( i ) )
According to (38), q ( i + 1 ) ( x k ) is updated using a nominal Gaussian distribution as q ( i + 1 ) ( x k ) = N ( x k ; H k x ^ k | k ( i ) , P ^ k | k ( i + 1 ) ) . Then x ^ k | k ( i ) and the corresponding estimation error covariance matrix P ^ k | k ( i + 1 ) are obtained by the measurement update of the KF as shown in (10)–(12) with P ^ k | k 1 ( i + 1 ) and R ^ k ( i + 1 ) .
( b ) As for the expectation of the Bernoulli distribution, E ( i + 1 ) [ s k ] and E ( i + 1 ) [ t k ] can be updated as follows:
E ( i + 1 ) [ s k ] = Pr ( i + 1 ) ( s k = 1 ) Pr ( i + 1 ) ( s k = 1 ) + Pr ( i + 1 ) ( s k = 0 )
E ( i + 1 ) [ t k ] = Pr ( i + 1 ) ( t k = 1 ) Pr ( i + 1 ) ( t k = 1 ) + Pr ( i + 1 ) ( t k = 0 )
By setting θ = s k and θ = t k , and then updating q ( i + 1 ) ( s k ) and q ( i + 1 ) ( t k ) using the Bernoulli distribution, the probabilities of s k and t k being 1 are given by:
Pr ( t k = 1 ) = Δ 1 i + 1 exp { E ( i ) [ log α k ] + 0.5 tr ( log Λ m , σ 1 ) 0.5 tr ( A k i + 1 R ¯ k 1 ) }
Pr ( s k = 1 ) = Δ 2 i + 1 exp { E ( i ) [ log β k ] + 0.5 tr ( log Λ p , σ 1 ) 0.5 tr ( B k i + 1 P ¯ k | k 1 1 ) }
where Δ is a normalizing constant, and two auxiliary variables A , B are used to evaluate the prior probability covariance and measurement covariance as
A k ( i + 1 ) = E ( i + 1 ) [ ( y k H k x k ) ( y k H k x k ) T ]
B k ( i + 1 ) = E ( i + 1 ) [ ( x k x ^ k | k 1 ) ( x k x ^ k | k 1 ) T ]
To implement the fixed-point iteration, (43) and (44) is approximated as
A k ( i + 1 ) = ( y k H k x ^ k | k ( i + 1 ) ) ( y k H k x ^ k | k ( i + 1 ) ) T + H k P k | k i + 1 H k T
B k ( i + 1 ) = P k | k ( i + 1 ) + ( x ^ k | k ( i + 1 ) x ^ k | k 1 ) ( x ^ k | k ( i + 1 ) x ^ k | k 1 ) T
( c ) Next, q ( i + 1 ) ( α k ) and q ( i + 1 ) ( β k ) are updated using the Beta distribution as shown in (27) and (28). Take the PDF of the Beta distribution into derivation, and, by setting θ = α k and θ = β k , then
q ( i + 1 ) ( α k ) = ( a 0 E ( i + 1 ) [ t k ] ) log ( α k ) + ( b 0 + E ( i + 1 ) [ t k ] 1 ) log ( 1 α k ) + C α
q ( i + 1 ) ( β k ) = ( c 0 E ( i + 1 ) [ s k ] ) log ( β k ) + ( d 0 + E ( i + 1 ) [ s k ] 1 ) log ( 1 β k ) + C β
where q ( i + 1 ) ( α k ) and q ( i + 1 ) ( β k ) can be updated with new shape parameters, therefore, the shape parameters of the Beta prior distribution can be updated as follows:
a k ( i + 1 ) = a 0 + E ( i + 1 ) [ t k ]
b k ( i + 1 ) = b 0 E ( i + 1 ) [ t k ] + 1
c k ( i + 1 ) = c 0 + E ( i + 1 ) [ s k ]
d k ( i + 1 ) = d 0 E ( i + 1 ) [ s k ] + 1
After completing c), checking the convergence of the iteration according to the preset error threshold, if x ^ k ( i + 1 ) x ^ k ( i ) x ^ k ( i ) < ε then stop the iteration and output the result. Otherwise, return to (a) for the next iteration.
In addition to the previous state and measurement vector, to begin the algorithm, the proposed filter only needs two kernel parameters and the initial prior beta shape parameters for the mixture of Gaussian correntropy. Among them, the kernel parameters of the Gaussian correntropy subfunction are more likely to be chosen by experience. In most cases, they have minimal impact on the filtering performance. This conclusion will be confirmed in the subsequent simulations.
On the other hand, the chosen of prior Beta distribution parameters needs a simple discussion. By combining the characteristics of the Beta distribution, the initial mixing probability ρ is determined by its prior shape parameters. If a 0 = c 0 = 0 or b 0 = d 0 = 0 , the proposed filter converges to the original MCKF. In order to reduce input parameters and simplify the algorithm, it is generally assumed that a 0 = c 0 and b 0 = d 0 in most cases.
In this work, two Gaussian correntropies with different kernel parameters are used to cope with the stable filtering process and the process corrupted by dynamic abnormal errors, such as impulsive noise disturbance. In order to ensure filtering accuracy and stability, the prior parameters are not arbitrarily set, but obey a certain regularity. For example, in an ideal Gaussian noise environment, we have E ( j + 1 ) [ t k ] 1 and then the performance of the mixture correntropy converge to the Gaussian correntropy with a large kernel parameter, so that it has basic robustness and retains convergence to KF as much as possible, expanding (41) and we can find that
E ( i + 1 ) [ t k ] = p ( i + 1 ) ( t k = 1 ) p ( i + 1 ) ( t k = 1 ) + p ( i + 1 ) ( t k = 0 ) = 1 1 + p ( i + 1 ) ( t k = 0 ) / p ( i + 1 ) ( t k = 1 ) = 1 1 + exp ( E ( i + 1 ) ( 1 α 0 ) E ( i + 1 ) ( α 0 ) + C k ) = 1 1 + exp ( ψ ( b 0 ) ψ ( a 0 ) + C k ) 1
where C k represents the terms independent of a 0 and b 0 . According to (41) and (42), if the distributions of measurement and state noise nearly meet the ideal Gaussian assumption is needed.
In contrast, for the filtering process corrupted by severe impulsive noise, the mixing probability is likely to be redistributed appropriately to enhance the robustness to abnormal noise, and then
E ( i + 1 ) [ t k ] = 1 1 + p ( i + 1 ) ( t k = 0 ) / p ( i + 1 ) ( t k = 1 ) 0 exp ( ψ ( b 0 ) ψ ( a 0 ) + C k ) > > 0
As the definition of Gaussian correntropy Equation (2), if σ 1 > σ 2 , for the residual term e i with significant abnormality, there is a great of difference between G σ 1 ( e i ) and G σ 2 ( e i ) , and G σ 1 ( e i ) G σ 2 ( e i ) . Therefore, it is further inferred that C k increases significantly and then ψ ( b 0 ) ψ ( a 0 ) + C k 0 at this time, where ψ ( b 0 ) ψ ( a 0 ) can be taken as a constant term to regulate the transition of E ( i + 1 ) [ t k ] , and a 0 > b 0 still applies to this case.
Therefore, in this work, the parameters can be chosen within the range 0.95 a 0 0.85 and b 0 = 1 a 0 , which can be used as a typical parameter configuration. On this basis, it may sometimes be useful to make a minor adjustment on the initial parameter configurations according to the specified application. This will contributes to further improving the algorithm’s performance in complex environments. The related results are shown in the later simulation.

4. Performance Evaluations and Analysis

4.1. Example I: 2-D Moving Target Tracking Model

Consider a two-dimensional (2D) moving target tracking system as
F = 1 sin ( w T ) w 0 ( 1 cos ( w T ) ) w 0 cos ( w T ) 0 sin ( w T ) 0 ( 1 cos ( w T ) ) w 1 sin ( w T ) w 0 sin ( w T ) 0 cos ( w T ) , Γ = 0.5 T 2 0 T 0 0 0.5 T 2 T 0 , H = 1 0 0 0 0 0 1 0
in which Γ represents the process noise gain matrix, T = 0 . 2 , w = 0 . 2 , and the total simulation steps N = 200 / T . It is assumed that Q k = 0 . 1 I 2 and R k = 10 I 2 . The true initial state x 0 = [ 1 1 1 1 ] T , the initial state estimation x ^ 0 = [ 0 0 0 0 ] T and the error covariance matrix P 0 | 0 = I 4 . In addition to the KF, the HKF with loss function parameter r = 1.345; the MCKF with typical kernel parameters σ = 2, σ = 3, σ = 5 and σ = 9, are used for comparison. For the algorithm proposed in this paper, we chose σ 1 = 9, σ 2 = 3 and a 0 = 0.9 as the default initial parameters. The maximum iteration times N m = 10 for the robust filters. The numerical test was coded with MATLAB and executed on a computer with Intel Core i7-9700 CPU @3.0 GHz.
In order to evaluate the filtering performance, the root mean square errors (RMSE) and averaged RMSE (ARMSE) are chosen as evaluation indicators of position and velocity estimation, which are defined as follows:
RMSE ( k ) = Δ 1 M s = 1 M ( ( x k s x ^ k s ) 2 + ( y k s y ^ k s ) 2 ) ARMSE = Δ 1 M N k = 1 N s = 1 M ( ( x k s x ^ k s ) 2 + ( y k s y ^ k s ) 2 )
where ( x k s , y k s ) and ( x ^ k s , y ^ k s ) are the true and estimated positions(or velocity) at the kth step of the sth Monte Carlo run, and RMS E p o s ( k ) is the RMSE of position at k step. The ARMSEs of position and velocity are denoted as ARMS E p o s and ARMS E v e l , respectively [12,21].
In order to verify whether the proposed filtering algorithm improve the solution of accuracy degradation problem of original MCKF in no-stationary non-Gaussian noise conditions, the simulation was divided into two stages, and the noises with different distributions were used to test the performance of the proposed filter. To simulate the non-Gaussian noise sequence contaminated by impulsive noise interference, the Gaussian mixture distribution model with specific parameters are used to generate the noise.
The number of Monte Carlo runs was M = 1000 and the specific noise parameters were set as follows:
w k 0.95 N ( 0 , Q k ) + 0.05 N ( 0 , 100 Q k ) , stage 1 : 1 < t N / 2 w k 0.90 N ( 0 , Q k ) + 0.10 N ( 0 , 100 Q k ) , stage 2 : N / 2 < t N v k 0.90 N ( 0 , R k ) + 0.10 N ( 0 , 100 R k ) , stage 1 : 1 < t N / 2 v k 0.95 N ( 0 , R k ) + 0.05 N ( 0 , 100 R k ) , stage 2 : N / 2 < t N
The RMSEs of the position and velocity from different filters are shown in Figure 1 and Figure 2. In addition, the ARMSEs from different filters at all stages are listed in Table 1 for comparison. In the intervals 0 , N / 2 and N / 2 , N , the process and measurement vectors were contaminated by non-Gaussian heavy-tailed noise with different distributions. The plots show that the RMSEs from different filters have obvious differences, and the best kernel size for the MCKF at each stage varied accordingly. For example, MCKF2 ( σ = 3) and MCKF3 ( σ = 5) achieved higher accuracy, respectively, at these two stages; nevertheless, a too large or small kernel size also resulted in the filtering performance to degrade or even diverge, as shown by MCKF1 ( σ = 2) and MCKF4 ( σ = 9).
In comparison, the proposed filter had the lowest estimation error at each stage, especially in the case of noise distribution change, the results show that the proposed VB interference method played a positive role in the filtering process, which takes into account both stability performance and robustness to non-Gaussian noise. Therefore, by comparing the filtering performance of all stages, the superiority of the proposed algorithm is preliminarily verified.
For the existing robust filters with fixed parameters, such as the classic MCKF or HKF, the filtering parameters can be obtained by experience or trial and error methods, which are more applicable to stationary noise conditions. However, as mentioned before, the classic MCKF does not always achieve satisfactory estimation accuracy in non-stationary noise environment. In order to show this change more concretely, the process and measurement noise distribution can be expressed as w k ( 1 p 1 ) N ( 0 , Q k ) + p 1 N ( 0 , 100 Q k ) and v k ( 1 p 2 ) N ( 0 , R k ) + p 2 N ( 0 , 100 R k ) , where p 1 and p 2 represent the outlier percentage of noise.
In Figure 3 and Figure 4, the outlier percentage of process noise was fixed as p 1 = 0.05, and the ARMSEs of the different filters varying with p 2 : 0 ∼ 0.15 are shown. The overall performance of the ARMSEs showed an increasing trend, which means that as the proportion of impulsive noise increased, the estimation accuracy of the filters also decreased. In this test, the MCKFs with different kernel parameters obtained significant filtering accuracy differences. This means that, under the interference of different non-Gaussian noise, the MCKFs with a fixed kernel parameter could not ensure reliable estimation results, and the MCKFs lack enough self-adaptive ability.
In addition, the fixed-parameters MMCKF without variational Bayesian iteration also taken for comparison. The comparison shows that the results of MMCKF always converges to the MCKF of a specific fixed parameter. Therefore, although it achieves better accuracy than MCKF in some cases, it does not avoid similar accuracy degradation problem as MCKF. Since the filtering results of MMCKF are similar to the typical MCKF with specific parameters, therefore, the comparison and discussion of the simulation mainly focus on the classic MCKF.
Compared with other filtering results, it can be concluded that the proposed filter has better performance than other existing algorithms. On the one hand, the algorithm does not lose much optimality of estimation caused by changing the objective function of KF, and on the other hand, it retains robustness to the increased impulsive noise probability. Therefore, the proposed filter further demonstrates its performance advantage in various non-Gaussian noise environments.
In addition, the simulation tests were performed to compare the filtering performance with different initial parameter configurations. For the proposed IMMCKF, the Gaussian correntropy G σ 1 ( e ) , G σ 2 ( e ) were mixed to generate the mixture correntropy, where σ 1 and σ 2 represent the specific kernel parameters, and σ 1 > σ 2 . It has been proved that Gaussian correntropy with a smaller kernel bandwidth is more sensitive to impulsive noise. However, this also degrades the filtering stability and might lead to divergence.
Table 2 lists the RMSEs of the proposed filter with different σ 1 and σ 2 . In general, the filtering results of the proposed filter with different σ 1 and σ 2 were not significant. Within a certain range, the filtering accuracy improved with the increase of σ 1 , as it provided better compatibility to Gaussian noises for the filtering algorithm. As too large of a kernel bandwidth may reduce Gaussian correntropy’s robustness to impulsive noise, σ 1 much greater than σ 2 would not be considered as the parameter set. Therefore, generally speaking, it is easy to choose appropriate σ 1 and σ 2 for the proposed algorithm.
In this case, several potential shape parameter options were used for comparison. Table 3 lists the ARMSEs of position and velocity from the proposed filter with different initial shape parameters. As demonstrated, the estimation accuracy can be improved somewhat by fine-tuning the parameter within the given range, but in general, it does not significantly affect the overall filtering performance. In summary, for the state estimation problem in non-stationary noises, the filter proposed in this work has good compatibility with the initial parameter sets.
In order to balance the computational efficiency and the filtering performance, it is necessary to choose a reasonable number of iterations for the proposed filter. Figure 5 shows the ARMSEs of filters with different numbers of iterations. For comparison, several filtering results with similar performance in the simulation are also taken into accounts. It can be concluded that the accuracy of the proposed algorithm can be greatly improved after several iterations, and the filtering result also gradually converged. In practical applications, the increase of iteration times brings more computational burden. As shown from Figure 5, when the number of iterations N m is 3∼5, the filtering algorithm has obtained satisfactory estimation accuracy. Therefore, when considering the accuracy and calculation burden factors, the number of iterations N m set as 3∼5 is reasonable.
The implementation time of the proposed filter and existing filters in a single-step run with N m = 1 is evaluated. It shows that KF (0.0054 ms) is the fastest, MCKF (0.0253 ms), HKF (0.0251 ms) and MMCKF (0.0212 ms) has similar computational burden and the proposed IMMCKF (0.0490 ms). Compared with the MCKF and MMCKF, the computation complexity of the proposed filter increased simultaneously due to the additional variational Bayesian iterations. In view of this, the additional computation cost of the IMMCKF can be compensated by adjusting the number of iterations. Therefore, it is still feasible to apply this algorithm in real-time applications.

4.2. Example II: INS/GPS Integrated Navigation System

To validate the effectiveness and superiority of the proposed algorithm in this paper, the experimental data collected in a vehicle-mounted INS/GPS integrated navigation experiment was used for the test. The experiment was carried out in the campus of Harbin Engineering University, the test trajectory is shown in Figure 6. A low-cost MEMS-IMU based INS/GPS integrated navigation system is used to provide the navigation data.
An INS/GNSS integration navigation system includes a self-made navigation-grade fiber optic strap-down INS and a double-antenna GPS receiver is used for reference. The initial velocity and position of INS/GPS are obtained directly from GPS measurement, the initial level attitude information is acquired from the alignment results of high accuracy SINS. In the experimental test, the car moves along a bumpy road, and the GPS might work abnormally due to the occlusion of trees and buildings.
The sampling frequencies of the low-cost IMU and GPS were 100 Hz and 1 Hz, respectively. The loosely coupled configuration is used in the integration, and a linear closed-loop feedback scheme as [15,23] is used. The state variables vector is defined as x k = [ φ n δ v n δ p n ε b b ] T , where φ n , δ v n , δ p n denote the attitude error, velocity error and position error expressed in the n-frame, respectively. The bias of the gyro and accelerometer are ε b = 10 / h and b = 100 μ g, respectively. The initial estimation was x ^ 0 | 0 = 0 15 × 1 . The process noise vector is expressed as w = [ w g x b w g y b w g z b w a x b w a y b w a z b ] T , where w g b = 1 / h and w a b = 1500 μ g / Hz , respectively, and the measurement noise covariance matrix R = diag ( 0.5 m / s · I 3 1.5 m · I 3 ) . The initial position and velocity errors were set as δ p = [ 1.5 m 1.5 m 1.5 m ] T , δ V n = [ 0.1 m / s 0.1 m / s 0.1 m / s ] T , and attitude errors ϕ = [ 0.1 0.1 0.1 ] T .
The filter only performs time update while there are no GPS outputs. To compare the robustness of different filters to non-Gaussian measurement noise interference, inspired by the scheme in [34,35], the gross errors were added artificially to the measurement data. In the integrated navigation system, the velocity and position state variables are susceptible to external interference. Therefore, the velocity and position errors are used for comparison.
For comparison, several typical filtering results are shown in Figure 7 and Figure 8. It can be seen that most filters obtain similar results in stable periods, as the measurement output is reliable in these periods. For the existing robust filters with fixed-parameter, due to the uncertain interference factors in practical application. In this example, both MCKF1 ( σ = 2) and MCKF2 ( σ = 3) have a divergence trend in the filtering process. In contrast, MCKFs with larger kernel parameters and the proposed filter have more stable performance.
In this example, the measurement sequence is disturbed by impulsive noise with a certain probability. It is evident from the Figures that the estimation accuracy of KF is seriously corrupted during filtering. Generally, a small kernel size can be more effective in attenuating the measurement outliers, but at the same time, it inevitably decreases the filter’s stability, e.g., MCKF1 ( σ = 2) failed to obtain reliable filtering results due to divergence. By comparing the robustness of MCKF3 ( σ = 5), MCKF4 ( σ = 9), and other algorithms to non-Gaussian noise, it can be concluded that, for existing MCKFs, it is difficult to obtain both stable and robustness estimation results.
The results listed in Table 4 confirmed that the filter proposed in this work solves the problem well. NaN represents invalid output due to filter divergence. Compared with other robust filters, the proposed filter achieves better filtering results by taking into account both the robustness and stability of the filter, which demonstrated the conclusions of the previous simulation example.

5. Conclusions

In this work, the performance degradation problem of the existing MCKF in non-stationary noises is explored, and a new improved mixture correntropy filtering algorithm is proposed as an effective solution. To cope with the dynamic process where both Gaussian and non-Gaussian noises may occur, the intermediate random variables is used to construct the mixture correntropy. By derivation, the state variables and intermediate parameters are approximated via a variational Bayesian approach. The theoretical derivation and numerical test results show that the proposed method significantly improves upon the existing MCKF algorithms in different conditions, which offers a promising improvement for the robust filtering problem in complex noise environments.

Author Contributions

Conceptualization, X.L.; methodology, Y.G.; software, Y.G.; data curation, Q.M.; writing—original draft preparation, Y.G.; writing—review and editing, Q.M.; supervision, X.L.; project administration, X.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

No new data were created or analyzed in this study. Data sharing is not applicable to this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Simon, D. Optimal State Estimation: Kalman, H Infinity, and Nonlinear Approaches; John Wiley & Sons: Hoboken, NJ, USA, 2006. [Google Scholar]
  2. Pieczynski, W.; Desbouvries, F. Kalman filtering using pairwise Gaussian models. In Proceedings of the 2003 IEEE International Conference on Acoustics, Speech, and Signal Processing, Proceedings.(ICASSP’03), Hong Kong, China, 6–10 April 2003. [Google Scholar]
  3. Ait-El-Fquih, B.; Desbouvries, F. Kalman Filtering in Triplet Markov Chains. IEEE Trans. Signal Process. 2006, 54, 2957–2963. [Google Scholar] [CrossRef]
  4. Kulikova, M.V. Gradient-Based Parameter Estimation in Pairwise Linear Gaussian System. IEEE Trans. Autom. Control 2017, 62, 1511–1517. [Google Scholar] [CrossRef]
  5. Nemesin, V.; Derrode, S. Robust Blind Pairwise Kalman Algorithms Using QR Decompositions. IEEE Trans. Signal Process. 2013, 61, 5–9. [Google Scholar] [CrossRef]
  6. Huber, P.J. Robust Statistics; John Wiley & Sons: Hoboken, NJ, USA, 2004; Volume 523. [Google Scholar]
  7. Gandhi, M.A.; Mili, L. Robust Kalman filter based on a generalized maximum-likelihood-type estimator. IEEE Trans. Signal Process. 2009, 58, 2509–2520. [Google Scholar] [CrossRef]
  8. Chang, L.; Hu, B.; Chang, G.; Li, A. Robust derivative-free Kalman filter based on Huber’s M-estimation methodology. J. Process. Control 2013, 23, 1555–1561. [Google Scholar] [CrossRef]
  9. Karlgaard, C.D.; Schaub, H. Huber-based divided difference filtering. J. Guid. Control Dyn. 2007, 30, 885–891. [Google Scholar] [CrossRef]
  10. Roth, M.; Özkan, E.; Gustafsson, F. A Student’s t filter for heavy tailed process and measurement noise. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 5770–5774. [Google Scholar]
  11. Huang, Y.; Zhang, Y.; Li, N.; Chambers, J. Robust Student-t based nonlinear filter and smoother. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 2586–2596. [Google Scholar] [CrossRef] [Green Version]
  12. Huang, Y.; Zhang, Y.; Li, N.; Wu, Z.; Chambers, J.A. A novel robust student’st-based kalman filter. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1545–1554. [Google Scholar] [CrossRef] [Green Version]
  13. Agamennoni, G.; Nieto, J.I.; Nebot, E.M. Approximate inference in state-space models with heavy-tailed noise. IEEE Trans. Signal Process. 2012, 60, 5024–5037. [Google Scholar] [CrossRef]
  14. Huang, Y.; Zhang, Y.; Zhao, Y.; Chambers, J.A. A Novel Robust Gaussian–Student’s t Mixture Distribution Based Kalman Filter. IEEE Trans. Signal Process. 2019, 67, 3606–3620. [Google Scholar] [CrossRef]
  15. Huang, Y.; Zhang, Y. A New Process Uncertainty Robust Student’s t Based Kalman Filter for SINS/GPS Integration. IEEE Access 2017, 5, 14391–14404. [Google Scholar] [CrossRef]
  16. Izanloo, R.; Fakoorian, S.A.; Yazdi, H.S.; Simon, D. Kalman filtering based on the maximum correntropy criterion in the presence of non-Gaussian noise. In Proceedings of the 2016 Annual Conference on Information Science and Systems (CISS), Princeton, NJ, USA, 16–18 March 2016; pp. 500–505. [Google Scholar]
  17. Chen, B.; Liu, X.; Zhao, H.; Principe, J.C. Maximum correntropy Kalman filter. Automatica 2017, 76, 70–77. [Google Scholar] [CrossRef] [Green Version]
  18. Liu, W.; Pokharel, P.P.; Principe, J.C. Correntropy: Properties and applications in non-Gaussian signal processing. IEEE Trans. Signal Process. 2007, 55, 5286–5298. [Google Scholar] [CrossRef]
  19. Chen, B.; Príncipe, J.C. Maximum correntropy estimation is a smoothed MAP estimation. IEEE Signal Process. Lett. 2012, 19, 491–494. [Google Scholar] [CrossRef]
  20. Liu, X.; Chen, B.; Zhao, H.; Qin, J.; Cao, J. Maximum correntropy Kalman filter with state constraints. IEEE Access 2017, 5, 25846–25853. [Google Scholar] [CrossRef]
  21. Wang, G.; Li, N.; Zhang, Y. Maximum correntropy unscented Kalman and information filters for non-Gaussian measurement noise. J. Frankl. Inst. 2017, 354, 8659–8677. [Google Scholar] [CrossRef]
  22. Wang, H.; Li, H.; Zhang, W.; Zuo, J.; Wang, H. Maximum correntropy derivative-free robust Kalman filter and smoother. IEEE Access 2018, 6, 70794–70807. [Google Scholar] [CrossRef]
  23. Liu, X.; Qu, H.; Zhao, J.; Yue, P. Maximum correntropy square-root cubature Kalman filter with application to SINS/GPS integrated systems. ISA Trans. 2018, 80, 195–202. [Google Scholar] [CrossRef]
  24. Wang, G.; Gao, Z.; Zhang, Y.; Ma, B. Adaptive maximum correntropy Gaussian filter based on variational Bayes. Sensors 2018, 18, 1960. [Google Scholar] [CrossRef] [Green Version]
  25. Huang, F.; Zhang, J.; Zhang, S. Adaptive filtering under a variable kernel width maximum correntropy criterion. IEEE Trans. Circuits Syst. II Express Briefs 2017, 64, 1247–1251. [Google Scholar] [CrossRef]
  26. Fakoorian, S.; Izanloo, R.; Shamshirgaran, A.; Simon, D. Maximum correntropy criterion Kalman filter with adaptive kernel size. In Proceedings of the 2019 IEEE National Aerospace and Electronics Conference (NAECON), Dayton, OH, USA, 15–19 July 2019; pp. 581–584. [Google Scholar]
  27. Chen, B.; Wang, X.; Lu, N.; Wang, S.; Cao, J.; Qin, J. Mixture correntropy for robust learning. Pattern Recognit. 2018, 79, 318–327. [Google Scholar] [CrossRef]
  28. Wang, H.; Zhang, W.; Zuo, J.; Wang, H. Outlier-robust Kalman filters with mixture correntropy. J. Frankl. Inst. 2020, 357, 5058–5072. [Google Scholar] [CrossRef]
  29. Zheng, F.; Derrode, S.; Pieczynski, W. Semi-supervised optimal recursive filtering and smoothing in non-Gaussian Markov switching models. Signal Process. 2020, 171, 107511. [Google Scholar] [CrossRef]
  30. Petetin, Y.; Janati, Y.; Desbouvries, F. Structured Variational Bayesian Inference for Gaussian State-Space Models with Regime Switching. IEEE Signal Process. Lett. 2021, 28, 1953–1957. [Google Scholar] [CrossRef]
  31. Zhang, G.; Lan, J.; Zhang, L.; He, F.; Li, S. Filtering in Pairwise Markov Model With Student’s t Non-Stationary Noise with Application to Target Tracking. IEEE Trans. Signal Process. 2021, 69, 1627–1641. [Google Scholar] [CrossRef]
  32. Chen, B.; Wang, J.; Zhao, H.; Zheng, N.; Principe, J.C. Convergence of a fixed-point algorithm under maximum correntropy criterion. IEEE Signal Process. Lett. 2015, 22, 1723–1727. [Google Scholar] [CrossRef]
  33. Huang, Y.; Zhang, Y.; Wu, Z.; Li, N.; Chambers, J. A novel adaptive Kalman filter with inaccurate process and measurement noise covariance matrices. IEEE Trans. Autom. Control 2017, 63, 594–601. [Google Scholar] [CrossRef] [Green Version]
  34. Li, S.; Xu, B.; Wang, L.; Razzaqi, A.A. Improved Maximum Correntropy Cubature Kalman Filter for Cooperative Localization. IEEE Sens. J. 2020, 20, 13585–13595. [Google Scholar] [CrossRef]
  35. Yang, C.; Shi, W.; Chen, W. Robust M–M unscented Kalman filtering for GPS/IMU navigation. J. Geod. 2019, 93, 1093–1104. [Google Scholar] [CrossRef]
Figure 1. RMSEs of the position from different filters.
Figure 1. RMSEs of the position from different filters.
Entropy 24 00117 g001
Figure 2. RMSEs of the velocity from different filters.
Figure 2. RMSEs of the velocity from different filters.
Entropy 24 00117 g002
Figure 3. ARMSEs of position from different filters with varying p 2 .
Figure 3. ARMSEs of position from different filters with varying p 2 .
Entropy 24 00117 g003
Figure 4. ARMSEs of velocity from different filters with varying p 2 .
Figure 4. ARMSEs of velocity from different filters with varying p 2 .
Entropy 24 00117 g004
Figure 5. ARMSEs versus iteration number from different filters.
Figure 5. ARMSEs versus iteration number from different filters.
Entropy 24 00117 g005
Figure 6. The test trajectory of the vehicle.
Figure 6. The test trajectory of the vehicle.
Entropy 24 00117 g006
Figure 7. The position errors from different filters in non-Gaussian noises.
Figure 7. The position errors from different filters in non-Gaussian noises.
Entropy 24 00117 g007
Figure 8. The velocity errors from different filters in non-Gaussian noises.
Figure 8. The velocity errors from different filters in non-Gaussian noises.
Entropy 24 00117 g008
Table 1. ARMSEs of different filters at each stage.
Table 1. ARMSEs of different filters at each stage.
FiltersStage 1Stage 2All Stages
Position (m)Velocity (m/s)Position (m)Velocity (m/s)Position (m)Velocity (m/s)
KF4.0141.2073.4201.3023.7291.255
HKF2.3470.9702.9451.2682.6631.129
MCKF1 ( σ = 2 )2.3460.9773.0811.2902.7381.144
MCKF2 ( σ = 3 )2.1670.9532.6591.2422.4261.107
MCKF3 ( σ = 5 )2.2010.9532.5721.2262.3941.098
MCKF4 ( σ = 9 )2.5520.9932.6831.2312.6181.118
IMMCKF1 ( σ 1 = 9 , σ 2 = 3 )2.0840.9392.5031.2182.3031.087
IMMCKF2 ( σ 1 = 9 , σ 2 = 2 )2.0780.9392.5081.2182.3031.088
Table 2. ARMSEs of the proposed filter with different σ 1 and σ 2 .
Table 2. ARMSEs of the proposed filter with different σ 1 and σ 2 .
FiltersStage 1Stage 2All Stages
Position (m)Velocity (m/s)Position (m)Velocity (m/s)Position (m)Velocity (m/s)
KF4.0141.2073.4201.3023.7291.255
IMMCKF1 ( σ 1 = 6 , σ 2 = 3 )2.1170.9432.5301.2212.3321.091
IMMCKF2 ( σ 1 = 6 , σ 2 = 2 )2.1080.9422.5321.2222.3301.091
IMMCKF3 ( σ 1 = 9 , σ 2 = 3 )2.0840.9392.5031.2182.3031.087
IMMCKF4 ( σ 1 = 9 , σ 2 = 2 )2.0780.9392.5081.2182.3031.088
IMMCKF5 ( σ 1 = 12 , σ 2 = 3 )2.0740.9382.4941.2172.2941.086
IMMCKF6 ( σ 1 = 12 , σ 2 = 2 )2.0540.9362.5021.2182.2891.086
Table 3. ARMSEs of the proposed filter with different a 0 .
Table 3. ARMSEs of the proposed filter with different a 0 .
FiltersStage 1Stage 2All Stages
Position(m)Velocity (m/s)Position (m)Velocity (m/s)Position (m)Velocity (m/s)
KF4.0141.2073.4201.3023.7291.255
IMMCKF1 ( a 0 = 0.94)2.1180.9432.5141.2182.3241.089
IMMCKF2 ( a 0 = 0.92)2.0950.9402.5051.2182.3091.088
IMMCKF3 ( a 0 = 0.90)2.0840.9392.5031.2182.3031.087
IMMCKF4 ( a 0 = 0.88)2.0790.9382.5041.2182.3011.087
IMMCKF5 ( a 0 = 0.86)2.0770.9382.5101.2192.3041.088
Table 4. The RMSEs of position, velocity from different filters.
Table 4. The RMSEs of position, velocity from different filters.
Filtering AlgorithmsPosE (m)PosN (m)PosU (m)VelE (m/s)VelN (m/s)VelU (m/s)
KF1.7361.7260.7990.4660.3630.118
HKF1.8171.4280.6810.4880.3200.112
MCKF1 ( σ = 2)NaNNaN0.675NaNNaN0.117
MCKF2 ( σ = 3)NaN1.1870.671NaN0.3330.112
MCKF3 ( σ = 5)1.0691.0610.6710.3790.2900.111
MCKF4 ( σ = 9)1.1211.1110.6740.3850.3000.111
MMCKF1 ( σ 1 = 2, σ 2 = 9, α = 0.8)NaNNaN0.711NaNNaN0.120
MMCKF2 ( σ 1 = 3, σ 2 = 9, α = 0.5)1.2210.9950.6740.4000.2750.112
MMCKF3 ( σ 1 = 3, σ 2 = 9, α = 0.2)1.0261.0070.6730.3740.2780.111
IMMCKF1 ( σ 1 = 9, σ 2 = 3)0.8880.9030.6570.3450.2610.110
IMMCKF2 ( σ 1 = 9, σ 2 = 2)0.7510.7750.6400.3190.2380.110
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, X.; Guo, Y.; Meng, Q. Variational Bayesian-Based Improved Maximum Mixture Correntropy Kalman Filter for Non-Gaussian Noise. Entropy 2022, 24, 117. https://doi.org/10.3390/e24010117

AMA Style

Li X, Guo Y, Meng Q. Variational Bayesian-Based Improved Maximum Mixture Correntropy Kalman Filter for Non-Gaussian Noise. Entropy. 2022; 24(1):117. https://doi.org/10.3390/e24010117

Chicago/Turabian Style

Li, Xuyou, Yanda Guo, and Qingwen Meng. 2022. "Variational Bayesian-Based Improved Maximum Mixture Correntropy Kalman Filter for Non-Gaussian Noise" Entropy 24, no. 1: 117. https://doi.org/10.3390/e24010117

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop