Next Article in Journal
Crack Assessment of Wheel Hubs via an Ultrasonic Transducer and Industrial Robot
Next Article in Special Issue
Multi-Sensor Data Fusion for Real-Time Surface Quality Control in Automated Machining Systems
Previous Article in Journal
A Novel Instantaneous Phase Detection Approach and Its Application in SSVEP-Based Brain-Computer Interfaces
Previous Article in Special Issue
Observable Degree Analysis for Multi-Sensor Fusion System
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Hybrid Adaptive Cubature Kalman Filter with Unknown Variance of Measurement Noise

1
College of Intelligent Manufacturing and Automation, Henan University of Animal Husbandry and Economy, Zhengzhou 450011, China
2
Information Technology Center, Zhejiang University, Hangzhou 310027, China
3
College of Electrical Engineering, Henan University of Technology, Zhengzhou 450001, China
4
School of Information Engineering, Zhengzhou University, Zhengzhou 450000, China
5
School of Business, University of Shanghai for Science and Technology, Shanghai 200093, China
*
Authors to whom correspondence should be addressed.
Sensors 2018, 18(12), 4335; https://doi.org/10.3390/s18124335
Submission received: 23 September 2018 / Revised: 13 November 2018 / Accepted: 27 November 2018 / Published: 7 December 2018
(This article belongs to the Collection Multi-Sensor Information Fusion)

Abstract

:
This paper is concerned with the filtering problem caused by the inaccuracy variance of measurement noise in real nonlinear systems. A novel weighted fusion estimation method of multiple different variance estimators is presented to estimate the variance of the measurement noise. On this basis, a hybrid adaptive cubature Kalman filtering structure is proposed. Furthermore, the information filter of the hybrid adaptive cubature Kalman filter is also studied, and the stability and filtering accuracy of the filter are theoretically discussed. The final simulation examples verify the validity and effectiveness of the hybrid adaptive cubature Kalman filtering methods proposed in this paper.

1. Introduction

In recent decades, nonlinear filtering has been widely used in military and civil fields such as target tracking, navigation, positioning, and intelligent manufacturing [1,2]. The theory and method of nonlinear filtering has became one of the most important research issues in the signal processing field, and has attracted increasing attention from researchers.
There are two main kinds of nonlinear filtering methods. The representative of the first kind of nonlinear filter is the extended Kalman filter (EKF), which linearizes the system model by Taylor expansion, holds the first order term, and ignores the second- and higher-order terms. The second kind of nonlinear filtering approximates the statistics of the system state, with examples being the unscented Kalman filter (UKF) and the cubature Kalman filter (CKF). Due to the model error of the linearization of the nonlinear system, the accuracy of the EKF is slightly lower, even leading to filtering divergence. Based on the unscented transformation to approximate the statistics of the system state, UKF was presented in [3]. Further, the cubature Kalman filter algorithm was proposed by Ienkaran in [4]. The CKF algorithm uses a third-degree spherical-radial cubature rule based on a Gaussian filtering framework. The algorithm has higher numerical stability and a smaller amount of calculation. Its excellent performance has made it widely used in various nonlinear system scenarios. Many advantages of CKF have attracted scholars to conduct in-depth research on it, considering that the traditional nonlinear filters often need to overcome the filtering divergence caused by high-dimensional operational errors. Drawing on the idea of square root filtering in the Kalman filter, Ienkaram and Haykin proposed the square-root cubature Kalman filter (SCKF) [5], which further improved the accuracy and stability of filtering. However, the traditional nonlinear filtering method requires knowledge of the mathematical model and the prior statistical information of noise when in practical application. Additionally, the statistical characteristics of noise in actual systems are usually indeterminate, which leads to a decline in the filtering accuracy.
For the problem of unknown statistical characteristics of measurement noise in real applications [6,7], Sage and Husa [8,9] proposed an excellent Sage–Husa suboptimal unbiased maximum a posteriori (MAP) estimator. Many scholars also adopted adaptive filtering techniques to improve the performance of the estimation algorithm, such as sliding window method [10], fading factor adjustment (FFA) [11,12,13], maximum a posteriori (MAP) estimator [14,15], and the variational Bayesian (VB) method [16,17,18,19], etc. Different methods for estimating the statistical characteristics of system noise are usually designed under different estimation criterions. How to use these methods to estimate the statistical characteristics of system noise is still an open issue.
For a class of nonlinear stochastic systems with inaccurate or unknown measurement noise variance (i.e., the priori measurement noise variance is not a precise value), an adaptive filtering algorithm based on SCKF is designed in this paper. Firstly, a novel fusion approach is proposed to estimate the measurement noise variance on the basis of the MAP and VB methods. Then, we use FFA to adjust the part of the variance matrix of the SCKF algorithm to obtain a hybrid adaptive SCKF algorithm (HASCKF) and the corresponding information filter (recorded as HASCIF). This is beneficial to reduce the effects on the adaptive filtering algorithm performance that may be caused by the estimation deviation of noise fusion. At the same time, the performance of the adaptive filtering algorithm is analyzed from two aspects based on the established HASCKF algorithm: the stability of the adaptive filtering algorithm and the filtering accuracy.
This paper is organized as follows: Section 2 formulates the nonlinear stochastic system and describes the problem of the inaccuracy of the the measurement noise variance. In Section 3, a novel noise variance fusion estimation algorithm HASCKF and the corresponding information filter are proposed based on the idea of weighted fusion. In Section 4, two simulation examples are utilized to display and verify the performance of the proposed algorithms. Section 5 provides the conclusions of this work.

2. Problem Description

Considering a class of discrete nonlinear stochastic systems, the state space model is described as follows [1]:
x ( k + 1 ) = f ( x ( k ) ) + w ( k ) ,
z ( k ) = h ( x ( k ) ) + v ( k ) ,
where x ( k ) R n is the state of the target, z ( k ) R m is the measurement, f : R n R n is the evolution process of the nonlinear state, and h : R n R m is the corresponding nonlinear measurement mapping.
The process noise w ( k ) R n is a Gaussian white noise with zero means and variance Q ( k ) .
The measurement noise v ( k ) R m is a Gaussian white noise with zero means and variance R ( k ) .
Hypothesis 1.
The process noise w ( k ) and measurement noise v ( k ) in the model are mutually statistically independent.
Hypothesis 2.
The initial state of the system is x ( 0 ) , with mean x 0 and variance P 0 , and it is uncorrelated with v ( k ) and w ( k ) .
Hypothesis 3.
The process noise variance Q ( k ) is known, but the measurement noise variance is only with an inaccuracy prior state R 0 .
For the nonlinear system described in (1) and (2), assuming that at the time k, we have the optimal estimation x ^ ( k 1 | k 1 ) and the square root matrix of the error variance matrix S ( k 1 | k 1 ) . Then, the state estimate x ^ ( k | k ) and the square root of the variance matrix S ( k | k ) can be calculated according to the standard SCKF algorithm.
For nonlinear systems with determined noise variance, the SCKF algorithm has better estimation performance. However, when the priori value of the measurement noise variance is inaccurate, the final estimation x ^ ( k | k ) and S ( k | k ) will have large errors.

3. Hybrid Adaptive SCKF Algorithm (HASCKF)

In order to improve the adaptive filtering accuracy for an inaccurate modeling system with unknown measurement noise variance, and to let the estimated noise variance be closer to the true noise variance, in this paper, a novel noise variance fusion estimation algorithm HASCKF is proposed based on the idea of weighted fusion.

3.1. Estimation Method of Measurement Noise

Theorem 1.
Assume that the measurement noise variance estimated by MAP and VB methods at time k are denoted as R ^ 1 ( k ) and R ^ 2 ( k ) , respectively. Then, the weighted fusion estimation of the measurement noise variance R ^ g ( k ) is
R ^ g ( k ) = R ^ 1 ( k ) + [ T 1 ( k ) T 12 ( k ) ] [ T 1 ( k ) + T 2 ( k ) 2 T 12 ( k ) ] 1 [ R ^ 2 ( k ) R ^ 1 ( k ) ] ,
where
T 1 ( k ) = R ^ g ( k 1 ) R ^ 1 ( k ) F 2 , T 12 ( k ) = t r [ ( R ^ g ( k 1 ) R ^ 1 ( k ) ) ( R ^ g ( k 1 ) R ^ 2 ( k ) ) T ] , T 2 ( k ) = R ^ g ( k 1 ) R ^ 2 ( k ) F 2 .
In the above equation, · F represents the Frobenius norm of the matrix, and the initial value of the fusion estimation R ^ g ( 0 ) = R 0 .
Proof. 
The weighted fusion estimation of the measurement noise variance R ^ g ( k ) can be expressed as the following linear combination:
R ^ g ( k ) = a 1 ( k ) R ^ 1 ( k ) + a 2 ( k ) R ^ 2 ( k ) .
Under the condition a 1 ( k ) + a 2 ( k ) = 1 , we minimize the performance criterion:
J ( k ) = t r [ ( R ( k ) R ^ g ( k ) ] [ ( R ( k ) R ^ g ( k ) ] T
From a 1 ( k ) + a 2 ( k ) = 1 , we can get a 1 ( k ) = ( 1 a 2 ( k ) ) , then substituting it into formula (5), we obtain
R ^ g ( k ) = R ^ 1 ( k ) + a 2 ( k ) [ R ^ 2 ( k ) R ^ 1 ( k ) ] .
Then, the overall estimation error is
R ˜ g ( k ) = R ( k ) R ^ g ( k ) = a 1 ( k ) R ˜ 1 ( k ) + a 2 ( k ) R ˜ 2 ( k ) = [ 1 a 2 ( k ) ] R ˜ 1 ( k ) + a 2 ( k ) R ˜ 2 ( k ) .
So, we can get
J ( k ) = t r [ R ˜ g ( k ) ( R ˜ g ( k ) ) T ] = [ 1 a 2 ( k ) ] 2 t r [ R ˜ 1 ( k ) ( R ˜ 1 ( k ) ) T ] + [ 1 a 2 ( k ) ] t r [ R ˜ 1 ( k ) ( R ˜ 2 ( k ) ) T ] a 2 ( k ) + a 2 ( k ) t r [ R ˜ 2 ( k ) ( R ˜ 1 ( k ) ) T ] [ 1 a 2 ( k ) ] + a 2 2 ( k ) t r [ R ˜ 2 ( k ) ( R ˜ 2 ( k ) ) T ] .
Let J ( k ) a 2 ( k ) = 0 , and after simplification, we can get
a 2 ( k ) = [ T 1 ( k ) T 12 ( k ) ] [ T 1 ( k ) + T 2 ( k ) T 12 ( k ) T 21 ( k ) ] 1 ,
where
T 1 ( k ) = t r [ R ˜ 1 ( k ) ( R ˜ 1 ( k ) ) T ] = t r [ ( R ( k ) R ^ 1 ( k ) ) ( R ( k ) R ^ 1 ( k ) ) T ] = R ( k ) R ^ 1 ( k ) F 2 , T 12 ( k ) = t r [ R ˜ 1 ( k ) ( R ˜ 2 ( k ) ) T ] = t r [ ( R ( k ) R ^ 1 ( k ) ) ( R ( k ) R ^ 2 ( k ) ) T ] , T 21 ( k ) = t r [ R ˜ 2 ( k ) ( R ˜ 1 ( k ) ) T ] = t r [ ( R ( k ) R ^ 2 ( k ) ) ( R ( k ) R ^ 1 ( k ) ) T ] = T 12 ( k ) , T 2 ( k ) = t r [ R ˜ 2 ( k ) ( R ˜ 2 ( k ) ) T ] = t r [ ( R ( k ) R ^ 2 ( k ) ) ( R ( k ) R ^ 2 ( k ) ) T ] = R ( k ) R ^ 2 ( k ) F 2 .
Since the measurement noise variance R ( k ) is uncertain, Equation (9) cannot be directly calculated. For this reason, we replace R ( k ) with fusion estimated value R ^ g ( k 1 ) of the measurement noise variance at time ( k 1 ) . Obviously, the initial value satisfies R ^ g ( 0 ) = R 0 , then substituting it to Equations (9), (8), and (6), respectively, we can get Equations (3) and (4). □
Note 1 In the above theorem, since the measurement noise variance R ( k ) is uncertain, we replace R ( k ) with fusion estimate R ^ g ( k 1 ) of the measurement noise variance. This approximate substitution has certain rationality, especially for the case of constant noise variance and slowly varying noise variance. From Theorem 1, the physical meanings of T 1 ( k ) and T 2 ( k ) are equivalent to the estimated error variance of the two noise variance estimation methods, and T 12 ( k ) and T 21 ( k ) are similar to their cross-variance.
Inference 1.
If we do not consider the correlation between the estimation error of noise variance, that is T 12 ( k ) = T 12 ( k ) = 0 , then the weighted fusion estimation of measurement noise variance R ^ g ( k ) is
R ^ g ( k ) = [ T 1 1 ( k ) + T 2 1 ( k ) ] 1 T 1 1 ( k ) R ^ 1 ( k ) + [ T 1 1 ( k ) + T 2 1 ( k ) ] 1 T 2 1 ( k ) R ^ 2 ( k ) , T 1 ( k ) = T 1 1 ( k ) + T 2 1 ( k ) .
It can be directly derived according to the principle of simple convex combination fusion [1].
Note 2 Obviously, the result of inference 1 is easy to generalize to the case where the number of noise variance estimators N e 3 . Assume that the error variance of the ith estimator T i ( k ) = R ^ g ( k 1 ) R ^ i ( k ) F 2 . Assuming that the estimation errors of various estimation methods are not related to each other, the fusion estimation of noise variance is
R ^ g ( k ) = j N e T j 1 ( k ) 1 j N e T j 1 ( k ) R ^ j ( k ) , T g 1 ( k ) = j N e T j 1 ( k ) ,
where T g ( k ) can be regarded as the error variance of noise variance fusion estimation. Since T i ( k ) 0 , T g ( k ) T i ( k ) , which indicates that the estimated noise variance after fusion is superior to that of any single noise variance estimator.

3.2. HASCKF Algorithm

Based on Theorem 1, combined with the fading factor adjustment technique [11], we propose the hybrid adaptive SCKF algorithm (HASCKF). The principle block diagram is shown in Figure 1.
Firstly, the variance of measurement noise is estimated by the MAP estimator and VB method respectively. Then, the weighted fusion technique is introduced to fuse the two noise variance estimators. Finally, the variance matrix of the SCKF measurement update is adjusted by the fading factor adjustment technique to obtain the final state estimation and the root-mean square error (RMSE) variance matrix. The detailed implementation of HASCKF is described in the following Theorem 2.
Theorem 2.
Consider a class of nonlinear system as described in (1) and (2). Under the condition of Hypotheses 1–3, if the optimal estimate x ^ ( k 1 | k 1 ) and the square-root matrix of the estimation error variance S ( k 1 | k 1 ) have been obtained, the state estimate x ^ ( k | k ) and the square-root matrix of the estimation error variance S ( k | k ) can be calculated according to the following steps:
Step 1: Time Update
x i ( k 1 | k 1 ) = S ( k 1 | k 1 ) ξ i + x ^ ( k 1 | k 1 ) ,
x i * ( k | k 1 ) = f x i ( k 1 | k 1 ) ,
x ^ ( k | k 1 ) = 1 N x i = 1 N x x i * ( k | k 1 ) .
In (12)–(14), i = 1 , 2 , , N x , N x = 2 n . The parameters ξ i are given below ( ε i is an n-order unit vector):
ξ i = N x N x 2 2 · ε i , i = 1 , 2 , , n , N x N x 2 2 · ε i n x , i = n + 1 , n + 2 , , N x .
Calculate the square root of the variance matrix:
S ( k | k 1 ) = T r i a ( [ X * ( k | k 1 ) S Q ( k ) ] ) ,
where T r i a ( · ) represents a triangular operation. S Q ( k ) represents the square-root of the new process noise variance Q ( k ) , that is, Q ( k ) = S Q ( k ) S Q T ( k ) , and
X * ( k | k 1 ) = 1 N x x 1 * ( k | k 1 ) x ^ ( k | k 1 ) , x 2 * ( k | k 1 ) x ^ ( k | k 1 ) , , x N x * ( k | k 1 ) x ^ ( k | k 1 ) .
Step 2: Measurement Update
(1) According to Equations (16)–(18), we can calculate the predicted value z ^ ( k | k 1 ) .
x i ( k | k 1 ) = S ( k | k 1 ) ξ i + x ^ ( k | k 1 ) ,
z i ( k | k 1 ) = h ( x i ( k | k 1 ) ) ,
z ^ ( k | k 1 ) = 1 N x i = 1 N x z i ( k | k 1 ) .
(2) The measurement noise variance R ^ 1 ( k ) of the MAP estimator is calculated by using Lemma 1 (19) or Lemma 2 (20).
Lemma 1
([14]). When the measurement noise variance is constant, the suboptimal MAP estimate R ^ ( k ) of noise variance R ( k ) can be obtained by the recursive calculation:
R ^ ( k ) = 1 k [ ( k 1 ) R ^ ( k 1 ) + z ˜ ( k ) z ˜ T ( k ) z ( k | k 1 ) z T ( k | k 1 ) ] ,
where z ˜ ( k ) = z ( k ) z ^ ( k | k 1 ) is the residual vector of measurement. The initial value R ^ ( 0 ) = R 0 .
Lemma 2
([15]). When the measurement noise variance is time-varying, the suboptimal MAP estimate R ^ ( k ) of noise variance R ( k ) can be obtained by the recursive calculation:
R ^ ( k ) = [ 1 d ( k 1 ) ] R ^ ( k 1 ) + d ( k 1 ) [ z ˜ ( k ) z ˜ T ( k ) z ( k | k 1 ) z T ( k | k 1 ) ] ,
where d ( k ) = ( 1 b ) / ( 1 b k + 1 ) . b is the forgetting factor, and its value range is usually between 0.95 and 0.99.
(3) Using the following Equation (21), combined with Equations (16)–(18), (22), (25), (26), and (23)–(32), we can iteratively compute the measurement noise variance estimation R ^ 2 ( k ) of the VB method.
Estimation of prediction parameters of measurement noise variance by VB method:
ζ ( k | k 1 ) = ρ · ζ ( k 1 ) , η ( k | k 1 ) = ρ · η ( k 1 ) ,
where “·” represents the point operation in Matlab. ρ = [ ρ 1 , , ρ m ] T , ζ ( k ) = [ ζ 1 ( k ) , , ζ m ( k ) ] T , η ( k ) = [ η 1 ( k ) , , η m ( k ) ] T . ζ i ( k ) and η i ( k ) are two parameters of the inverse gamma distribution. ρ i ( 0 , 1 ) is the predictive weighting factor. It reflects the degree of correlation between the noise at the last moment and the noise at the current moment. When the difference between the measurement noise variance at the last moment and the measurement noise variance at the current moment is small, a larger ρ i value should be used. Conversely, ρ i should take a smaller value.
The square-root of the variance matrix:
S z z ( k | k 1 ) = T r i a ( [ z ( k | k 1 ) S R ( k ) ] ) ,
where S R ( k ) represents the square root of measurement noise variance R ( k ) (namely, R ( k ) = S R ( k ) S R T ( k ) ) and
z ( k | k 1 ) = 1 N x z 1 ( k | k 1 ) z ^ ( k | k 1 ) z 2 ( k | k 1 ) z ^ ( k | k 1 ) z N x ( k | k 1 ) z ^ ( k | k 1 ) ,
P x z ( k | k 1 ) = x ( k | k 1 ) z T ( k | k 1 ) ,
where
x ( k | k 1 ) = 1 N x x 1 ( k | k 1 ) x ^ ( k | k 1 ) x 2 ( k | k 1 ) x ^ ( k | k 1 ) x N x ( k | k 1 ) x ^ ( k | k 1 ) .
Then, the VB method gets R ^ ( k ) through M iterations:
Iterative initialization: let t = 1 , for a given number of iterations M, we have
ζ ( k ) = 1 / 2 , 1 / 2 , , 1 / 2 T + ζ ( k | k 1 ) , x ^ 1 ( k | k 1 ) = x ^ ( k | k 1 ) .
Calculating the estimate of the measurement noise variance:
R ^ t ( k ) = d i a g ( η t ( k ) · η t ( k ) · ζ ( k ) ζ ( k ) ) ,
where d i a g ( A ) represents a diagonal matrix composed of matrix A diagonal elements.
Use Equations (16) and (32) to calculate the t-th iteration state estimate x ^ t ( k | k ) and the root of its mean square error matrix S t ( k | k ) . If t < M , update the parameter η t ( k ) .
η t ( k ) = η ( k | k 1 ) + ( z ( k ) z ^ t ( k | k 1 ) ) · 2 / 2 + d i a g { P z z t ( k | k ) } / 2 .
Let t = t + 1 , x ^ t ( k | k 1 ) = x ^ t 1 ( k | k ) , return to the beginning of the iterative.
When t = M , end the iteration, we can get
η ( k ) = η M ( k ) R ^ ( k ) = R ^ M ( k ) , x ^ ( k | k ) = x ^ M ( k | k ) S ( k | k ) = S M ( k | k ) .
(4) Calculate the fusion estimation R ^ g ( k ) of the measurement noise variance according to Equation (3).
(5) According to Equations (28) and (29), S z z ( k | k 1 ) and P x z ( k | k 1 ) are adaptively adjusted using the FFA fading factor.
S z z ( k | k 1 ) = T r i a [ z ( k | k 1 ) z ( k | k 1 ) τ ( k ) , τ ( k ) S R ( k ) ]
P x z ( k | k 1 ) = 1 τ ( k ) x ( k | k 1 ) z T ( k | k 1 ) ,
where τ ( k ) is the adaptive factor, 0 < τ ( k ) 1 , and it is calculated by the following Lemma 3.
Lemma 3
([11]). For nonlinear systems with unknown measurement noise variance, the adaptive fading factor is determined by the following equation:
τ ( k ) = 1 , t r [ z ˜ ( k | k 1 ) z ˜ T ( k | k 1 ) ] t r [ S z z ( k | k 1 ) S z z T ( k | k 1 ) ] , t r [ S z z ( k | k 1 ) S T ( k | k 1 ) ] t r [ z ˜ ( k | k 1 ) z ˜ T ( k | k 1 ) ] , t r [ z ˜ ( k | k 1 ) z ˜ T ( k | k 1 ) ] > t r [ S z z ( k | k 1 ) S z z T ( k | k 1 ) ] .
In the above equation, t r denotes the trace of the matrix, z ˜ ( k | k 1 ) = z ( k ) z ^ ( k | k 1 ) is the measurement residual error vector.
(6) The following Equations (30)–(32) are used to obtain the updated estimation x ^ ( k | k ) and the root of its mean square error variance S ( k | k ) .
K ( k ) = P x z ( k | k 1 ) / S z z T ( k | k 1 ) / S z z ( k | k 1 ) ,
where the symbol “/” indicates the matrix right divide operation (e.g., A / B = A B 1 )
x ^ ( k | k ) = x ^ ( k | k 1 ) + K ( k ) [ z ( k ) z ^ ( k | k 1 ) ] ,
S ( k | k ) = T r i a [ x ( k | k 1 ) K ( k ) z ( k | k 1 ) K ( k ) S R ( k ) ] .
Proof: We can directly derive Theorem 1 and Lemma 3 by Lemma 1, omitted here.
Note 3 Theorem 2 only shows the hybrid adaptive filtering algorithm when the number of noise variance estimators N e is 2. Obviously, when N e > 2 , we first use various noise variance estimators to estimate R ^ j ( k ) and then calculate the fused estimate R ^ g ( k ) of the measurement noise variance according to Equation (11). In other words, only steps (2)–(4) in Theorem 2 need to be adjusted.

3.3. HASCIF Information Filter

Compared with traditional filtering, the information filter may not require prior information when it is initialized, and thus has better numerical performance. In addition, the use of an information filter to design a fusion algorithm is also simpler. In the information filter, the state estimate and its estimation error variance matrix are replaced by information vector and information matrix, respectively. Subsequently, we give its corresponding information filtering form (HASCIF) on the basis of HASCKF. According to the literature [20,21], we can obtain the implementation process of HASCIF as follows:
Step 1: Time update [20,21]
y ^ ( k | k 1 ) = Y ( k | k 1 ) x ^ ( k | k 1 ) ,
Y ( k | k 1 ) = P 1 ( k | k 1 ) = [ S ( k | k 1 ) S T ( k | k 1 ) ] 1 ,
where y ^ ( k | k 1 ) and Y ( k | k 1 ) are the predicted information vector and the predicted information matrix, respectively. x ^ ( k | k 1 ) and S ( k | k 1 ) can be calculated according to Equations (14) and (15).
Step 2: Measurement update
y ^ ( k | k ) = Y ( k | k ) x ^ ( k | k ) = y ^ ( k | k 1 ) + Y ( k | k 1 ) P x z ( k | k ) [ R ^ g ( k ) ] 1 { z ˜ ( k | k 1 ) + P x z T ( k | k ) Y T ( k | k 1 ) x ^ ( k | k 1 ) } = Δ y ^ ( k | k 1 ) + θ ( k ) ,
Y ( k | k ) = Y ( k | k 1 ) + Y ( k | k 1 ) P x z ( k | k ) [ R ^ g ( k ) ] 1 ( k ) P x z T ( k | k ) Y T ( k | k 1 ) = Δ Y ( k | k 1 ) + Θ ( k ) .
In Equations (35) and (36), the information state vectors y ^ ( k | k ) and Y ( k | k ) are the information vector and information matrix of the state estimate, respectively. θ ( k ) and Θ ( k ) are the information contribution vector and the information contribution matrix. R ^ g ( k ) is determined by formula (3), and the cross-variance matrix P x z ( k | k ) can be calculated by Formula (29).

3.4. Performance Analysis of HASCKF

References [22,23,24] proposed the bounded convergence theorem of the UKF algorithm, and Reference [14] extended its theorem to the adaptive cubature Kalman filter (ACKF). In this section, the CKF bounded convergence theorem proposed in Reference [14] and the Cramer–Rao lower bound (CRLB) [25] are used to analyze the convergence of the HASCKF algorithm.

3.4.1. Stability Analysis

Assuming that the noise variances Q ( k ) and R ( k ) are known accurately, and considering the nonlinear systems (1), (2), and the standard CKF algorithm [4], the corresponding state error variance matrix can be written as [14,22,24]:
P ( k | k 1 ) = x ( k | k 1 ) x T ( k | k 1 ) + Q ( k 1 ) = A ( k ) P ( k 1 | k 1 ) A T ( k ) + δ P ( k | k 1 ) + Q ( k 1 ) = A ( k ) P ( k 1 | k 1 ) A T ( k ) + Ξ ( k ) ,
P z z ( k | k 1 ) = z ( k | k 1 ) z T ( k | k 1 ) + R ( k ) = G ( k ) P ( k | k 1 ) G T ( k ) + δ P z z ( k | k 1 ) + R ( k ) = G ( k ) P ( k | k 1 ) G T ( k ) + Σ ( k ) ,
K ( k ) = P ( k | k 1 ) G T ( k ) [ G ( k ) P ( k | k 1 ) G T ( k ) + Σ ( k ) ] 1 ,
where
δ P ( k | k 1 ) = x ( k | k 1 ) x T ( k | k 1 ) A ( k ) P ( k 1 | k 1 ) A T ( k ) ,
δ P z z ( k | k 1 ) = z ( k | k 1 ) z T ( k | k 1 ) G ( k ) P ( k | k 1 ) G T ( k ) .
Each correlation matrix is defined as follows:
A ( k ) = β ( k ) F ( k ) , B ( k ) = α ( k ) H ( k ) , C ( k ) = I K ( k ) α ( k ) H ( k ) ,
G ( k ) = α ( k ) H ( k ) γ T ( k ) , n m , γ T ( k ) α ( k ) H ( k ) , n < m ,
Ξ ( k ) = δ P ( k | k 1 ) + Q ( k 1 ) , Σ ( k ) = δ P z z ( k | k 1 ) + R ( k ) ,
where
F ( k ) = f x x = x ^ ( k 1 | k 1 ) , H ( k ) = h x x = x ^ ( k | k 1 ) .
α ( k ) , β ( k ) , and γ ( k ) are auxiliary diagonal matrixes. Refer to Reference [22] for specific meanings.
Lemma 4
([14]). Consider the nonlinear systems (1), (2), and the standard CKF algorithm. If k 0 , both satisfy the following two assumptions:
(1) There are non-zero real numbers a min , a max , b max , c max , g min , and g max existing to let the following formulas be established:
a min 2 I A ( k ) A T ( k ) a max 2 I , B ( k ) B T ( k ) b max 2 I , g min 2 I G ( k ) G T ( k ) g max 2 I , C ( k ) C T ( k ) c max 2 I , [ G ( k ) B ( k ) ] [ G ( k ) B ( k ) ] T ( g max b max ) 2 I .
(2) There are positive real numbers p min , p max , q max , r max , Ξ min , Ξ max , and Σ min existing to let the following forms be established:
p min I P ( k | k ) p max I , Q ( k ) q max I , R ( k ) r max I , Ξ ( k ) Ξ max I , Ξ ( k ) > Ξ min I , Σ ( k ) > Σ min I ,
where
Σ min = max ( Σ 1 , Σ 2 ) , Σ 1 = a max 2 ( g max b max ) 2 ( p max + p max 2 a max 2 Ξ min 1 ) , Σ 2 = b max 2 ( a max 2 p max + Ξ max ) g max 2 ( a min 2 p min Ξ min ) .
Then, the standard CKF state estimation error will be mean-square bounded, that is, the algorithm is stable and convergent.
Note 4 Lemma 4 shows that the statistical characteristics of noise are closely related to the stability of the CKF algorithm. In addition, if Lemma 4 is established, the SCKF with known noise statistics is also stable and convergent. This is because the theoretical framework of SCKF is consistent with that of CKF. The mean square root matrix S ( k | k 1 ) and S ( k | k ) in CKF are only used when transferring the error variance. The root mean square matrix in the SCKF algorithm is obtained through triangulation of the matrix, thus avoiding the filter divergence caused by the non-positive definite variance matrix in the numerical calculation, so it has better stability while ensuring the accuracy of estimation with CKF.
Theorem 3.
If the standard SCKF algorithm is stable and convergent when the statistical characteristics of noise are known accurately, the introduction of the noise variance fusion estimator and the adaptive fade factor can ensure the stable convergence of the HASCKF algorithm.
Proof. 
When the measured noise variance matrix R ( k ) is inaccurate, other sufficient conditions in the bounded convergence theorem can be satisfied, but the condition (46) will be affected.
(a) First only consider the influence of the weighted fusion noise estimator. Let Δ R ( k ) = R ^ g ( k ) R ( k ) , then formula (40) can be rewritten as follows:
Σ ( k ) = δ P z z ( k | k 1 ) + R ^ g ( k ) = δ P z z ( k | k 1 ) + R ( k ) + Δ R ( k ) .
When R ^ g ( k ) R ( k ) , then Δ R ( k ) 0 , and Σ k will become larger. Obviously, condition (46) still holds and the HASCKF algorithm still converges steadily.
When R ^ g ( k ) < R ( k ) , then Δ R ( k ) < 0 , and Σ k will become smaller. Now, condition (46) may not be satisfied. However, the noise variance weighted fusion estimator R ^ g ( k ) estimates and corrects the inaccurate noise variance matrix in real time so that it gradually tracks the real value R ( k ) , thus making Δ R ( k ) 0 , Σ k Σ ( k ) . In this way, Δ R ( k ) 0 is gradually satisfied to Equation (42) to ensure stable convergence of the HASCKF algorithm.
(b) Further consider the effect of fading factors. According to Theorem 2, the variance matrix for measuring residuals is
P z z ( k | k 1 ) = S z z ( k | k 1 ) S z z T ( k | k 1 ) = 1 τ ( k ) z ( k | k 1 ) z T ( k | k 1 ) + R ^ g ( k ) .
Then, Equation (41) is rewritten as:
δ P z z ( k | k 1 ) = 1 τ ( k ) z ( k | k 1 ) z T ( k | k 1 ) G ( k ) P ( k | k 1 ) G T ( k ) = [ 1 τ ( k ) 1 ] z ( k | k 1 ) z T ( k | k 1 ) + z ( k | k 1 ) z T ( k | k 1 ) G ( k ) P ( k | k 1 ) G T ( k ) = [ 1 τ ( k ) 1 ] z ( k | k 1 ) z T ( k | k 1 ) + δ P z z ( k | k 1 ) = Δ Δ P z z ( k | k 1 ) + δ P z z ( k | k 1 ) .
Therefore, Equation (44) can be written as
Σ ( k ) = δ P z z ( k | k 1 ) + R ^ g ( k ) = δ P z z ( k | k 1 ) + R ( k ) + Δ P z z ( k | k 1 ) + Δ R ( k ) = Δ Σ ( k ) + Δ Σ ( k ) .
It is known from the definition of fading factor 0 < τ ( k ) 1 , so we can get
Δ P z z ( k | k 1 ) = [ 1 1 τ ( k ) τ ( k ) 1 ] z ( k | k 1 ) z T ( k | k 1 ) 0 .
Similar to the analysis in (a), when Δ R ( k ) 0 and Δ Σ ( k ) 0 , it is obvious that condition (46) holds and the stability of the HASCKF algorithm remains. When Δ R ( k ) < 0 , if Δ P z z ( k | k 1 ) is large enough, Δ Σ ( k ) 0 can still be satisfied, and condition (46) still holds, so the HASCKF algorithm converges steadily. If Δ P z z ( k | k 1 ) is not enough to guarantee Δ Σ ( k ) 0 , the introduction of the weighted fusion noise variance estimator can also make Δ R ( k ) 0 , so that Δ Σ ( k ) 0 is established stepwise to ensure stable convergence of the HASCKF algorithm.
In summary, the introduction of the noise variance fusion estimator and adaptive fading factor in the HASCKF algorithm can ensure the stable convergence of the algorithm. □
Note 5 This theorem combined with Note 2 shows that the noise variance estimation based on weighted fusion is superior to the estimate of any single noise variance estimator. Therefore, the hybrid adaptive HASCKF estimation algorithm has better stability than the adaptive SCKF using a single noise variance estimation algorithm.

3.4.2. Filtering Accuracy Analysis

There is a lower bound on the minimum variance unbiased estimator of the state of the nonlinear filtering algorithm. It is widely used to evaluate the performance of nonlinear estimation. In practice, the lower limit of Cramer–Rao Lower Bound (CRLB) is commonly used. Denote X 0 : k = { x ( 0 ) , x ( 1 ) , , x ( k ) } , Z 0 : k = { z ( 0 ) , z ( 1 ) , , z ( k ) } , and p ( Z 0 : k , X 0 : k ) is the joint probability density of ( Z 0 : k , X 0 : k ) . x ^ ( k ) is the unbiased estimation of x ( k ) . Then, CRLB is defined as [25]:
P ( k | k ) = Δ E { [ x ( k ) x ^ ( k ) ] [ x ( k ) x ^ ( k ) ] T } J 1 ( k ) ,
where J ( k ) is the Fisher information matrix:
J ( k ) = E 2 ln p ( Z 0 : k , X 0 : k ) 2 x 2 ( k ) .
Theorem 4 Assume that the nonlinear filter is applied to system (1), (2). Then
P ( k | k ) J 1 ( k ) ,
where
J ( k ) = [ Q ( k 1 ) + F ( k ) J 1 ( k 1 ) F T ( k ) ] 1 + H T ( k ) R 1 ( k ) H ( k ) .
Proof. 
We use J ( k ) to denote the information matrix of state x ( k ) . Then, the information matrix J ( k ) can be recursively calculated according to the following formula [26]:
J ( k ) = D 22 ( k ) D 21 ( k ) [ J ( k 1 ) + D 11 ( k ) ] 1 D 12 ( k ) ,
where
D 11 ( k ) = E { Δ x ( k 1 ) x ( k 1 ) ln p [ x ( k ) | x ( k 1 ) ] } , D 12 ( k ) = E { Δ x ( k 1 ) x ( k ) ln p [ x ( k ) | x ( k 1 ) ] } , D 21 ( k ) = E { Δ x ( k ) x ( k 1 ) ln p [ x ( k ) | x ( k 1 ) ] } = D 12 T ( k ) , D 22 ( k ) = E { Δ x ( k ) x ( k ) ln p [ x ( k ) | x ( k 1 ) ] } + E { Δ x ( k ) x ( k ) ln p [ z ( k ) | x ( k ) ] } ,
where Δ x y = x y T means the second-order operator. x = x is the first-order operator. Because process noise and measurement noise are Gaussian white noise, we have
ln p [ x ( k ) | x ( k 1 ) ] = ln e { 1 2 [ x ( k ) f ( x ( k 1 ) ) ] T Q 1 ( k 1 ) [ x ( k ) f ( x ( k 1 ) ) ] } 2 π Q ( k 1 ) = c 1 + 1 2 [ x ( k ) f ( x ( k 1 ) ) ] T Q 1 ( k 1 ) [ x ( k ) f ( x ( k 1 ) ) ]
ln p [ z ( k ) | x ( k ) ] = ln e { 1 2 [ z ( k ) h ( x ( k ) ) ] T R 1 ( k ) [ z ( k ) h ( x ( k ) ) ] } 2 π R ( k ) = c 2 + 1 2 [ z ( k ) h ( x ( k ) ) ] T R 1 ( k ) [ z ( k ) h ( x ( k ) ) ] ,
where constant c 1 = ln [ 2 π Q ( k 1 ) ] , c 2 = ln [ 2 π R ( k ) ] .
Substituting Equations (58) and (59) into Equation (57), then we can get
D 11 ( k ) = E { [ x ( k 1 ) f T ( x ( k 1 ) ) ] Q 1 ( k 1 ) [ x ( k 1 ) f T ( x ( k 1 ) ) ] T } = F T ( k ) Q 1 ( k 1 ) F ( k ) , D 12 ( k ) = E { [ x ( k 1 ) f T ( x ( k 1 ) ) ] Q 1 ( k 1 ) = F T ( k ) Q 1 ( k 1 ) = D 21 T ( k ) , D 22 ( k ) = Q 1 ( k 1 ) + E { [ x ( k ) h T ( x ( k ) ) ] R 1 ( k ) [ x ( k ) h T ( x ( k ) ) ] T } = Q 1 ( k 1 ) + H T ( k ) R 1 ( k ) H ( k ) .
J ( k ) = Q 1 ( k 1 ) + H T ( k ) R 1 ( k ) H ( k ) Q 1 ( k 1 ) F ( k ) × [ J ( k 1 ) + F T ( k ) Q 1 ( k 1 ) F ( k ) ] 1 F T ( k ) Q 1 ( k 1 ) .
Equation (56) can be obtained by applying a matrix inversion lemma to Equation (61). □

4. Simulation

In this section, two numerical simulation examples are provided to display and verify the performance of the SACKF algorithm proposed in this paper, mainly including the following contents:
(1) For a class of inaccurate modeling with unknown measurement noise variance, we compare the performance difference between the weighted fusion estimator (referred to as WF-NE) and the single noise estimator (e.g., MAP estimation or VB estimation, respectively denoted as MAP-NE and VB-NE).
(2) For a class of inaccurate modeling with unknown measurement noise variance, we study the advantages and disadvantages of the HASCKF algorithm and the standard SCKF algorithm, and the equivalence between the HASCKF and HASCIF algorithms.
Example 1.
This example is used to evaluate the performance of three kinds of noise estimators, MAP-NE, VB-NE, and WF-NE. Considering the following first-order nonlinear discrete dynamic system:
x ( k ) = 0.5 x ( k 1 ) + 0.2 x ( k 1 ) 1 + x 2 ( k 1 ) + w ( k ) ,
z ( k ) = x 2 ( k ) 20 + v ( k ) ,
where w ( k ) and v ( k ) are mutually independent Gaussian white noise sequences. Assume that the initial value of the system state and the error variance matrix are
x 0 = 2 , P 0 = 0.01 ,
and the system state initial value x 0 is independent of the two noises. The process noise statistic Q ( k ) = 0.001 . In the following, simulation experiments are performed for two cases where the measurement noise variance R ( k ) is a constant and piecewise continuous function.
(1) Situation 1: If the measurement noise variance is constant and R ( k ) = 0.012 . Assume that the imprecise measurement noise variance of the initial value R ^ 0 = 0.04 . In the simulation, MAP-NE adopted the estimator described in Equation (19). The parameters of VB-NE were selected as follows: ρ = 1 e 5 , ζ ( 0 ) = 1 , η ( 0 ) = 0.04 , M = 1 .
In order to compare the performance of various algorithms, we adopted the absolute error (AE) and the mean absolute error (MAE), which are calculated as follows:
A E ( k ) = y ( k ) y ^ ( k ) , M A E ( k ) = 1 N s k = 1 N s A E ( k ) ,
where y ( k ) and y ^ ( k ) are the value to be estimated and the estimated value respectively. N s is the number of simulation steps. In this case, N s = 1000 .
The estimated noise variance of the three kinds of noise estimators are shown in Figure 2 and Figure 3, and the estimation error is given in Table 1.
(2) Situation 2: If the measurement noise value is time varying, the true statistical characteristics meet the following formula:
R ( k ) = R 0 k 500 2 R 501 k 1000 , a n d R = 0.012 .
Assume that the known inaccurate initial value of the measurement noise variance is R ^ 0 = 0.08 . In the simulation, MAP-NE adopted the estimator described in Equation (20), and the forgetting factor b was 0.98. The parameters of VB-NE were selected as ρ = 1 e 5 , ζ ( 0 ) = 1 , η ( 0 ) = 0.08 , M = 1 . The estimation results of the three kinds of noise estimators on the measured noise variance are shown in Figure 4 and Figure 5, and the mean absolute errors of several algorithms are given in Table 2.
The simulation results of Example 1 show that the fusion estimator WF-NE proposed in Theorem 1 could obtain the best estimated result for either the constant unknown measurement noise variance or for the time-varying one, compared with the MAP-NE and VB-NE. It had the same result with the analysis conclusion in Note 2. In Situation 1, as shown in Figure 2 and Figure 3, compared with the fusion results of MAP-NE and VB-NE, the WF-NE had better estimates in most simulation steps. Due to the randomness of noise, not all the noise variance estimates obtained by WF-NE were better than the ones solved by other methods, especially for the estimation of the time-varying variance in Situation 2. As shown in Figure 3, the MAP-NE method had relatively large estimation errors in the time interval [160–360], and the VB-NE method had relatively large estimation errors in the time interval [390–800]. Nethertheless, the WF-NE could avoid large variation of the noise variance estimation error. Similarly, the general trend of the WF-NE method was best in Figure 4, although the MAP-NE method or the VB-NE method were best in some small time intervals. Although the CPU time cost of WF-NE (0.3725 for 1000 simulation steps) was larger than the other two methods (0.2188 for 1000 simulation steps), it was still acceptable. Meanwhile, the mean absolute errors of WF-NE were smaller than the other two methods, for boththe constant noise variance (in Situation 1) and for the time-varying variance (in Situation 2).
Example 2.
This example is used to verify the pros and cons of the hybrid adaptive SCKF estimation algorithm HASCKF proposed by Theorem 2 and the standard SCKF algorithm described in Reference [4]. Assume that the target is moving in a uniform linear motion on a two-dimensional plane. The system state x ( k ) = [ x ( k ) , x ˙ ( k ) , y ( k ) , y ˙ ( k ) ] T , where x ( k ) and y ( k ) are the position components in the east–north coordinate system. x ˙ ( k ) and y ˙ ( k ) are the corresponding velocity components, respectively. Then, the state equation can be described as:
x ( k ) = 1 T 0 0 0 1 0 0 0 0 1 T 0 0 0 1 · x ( k 1 ) + w ( k ) ,
where the sampling period T = 0.5 s, the process noise w ( k ) is zero mean Gaussian white noise. Its statistical characteristic is Q ( k ) = 0.1 × d i a g ( Q 1 , Q 2 ) , and Q 1 = T 3 T 3 3 3 T 2 T 2 2 2 T 2 T 2 2 2 T .
Consider a radar to track the target. The nonlinear measurement equation can be expressed as
z ( k ) = x 2 ( k ) + y 2 ( k ) arctan y ( k ) x ( k ) + v ( k ) .
The real statistical characteristic of the measuring noise R ( k ) = d i a g { ( 4 m ) 2 , ( 0.1 ) 2 } . During the simulation, the simulation time was set as N s = 100   s . Suppose that the inaccuracy initial value of the measurement noise variance R ^ ( 0 ) = d i a g { ( 81 m ) 2 , ( 0.3 ) 2 } . MAP-NE adopts the estimator described in Equation (19), and the parameters of VB-NE were selected as ρ = [ 1 e 5 , 1 e 5 ] T , ζ ( 0 ) = [ 1 , 1 ] T , η ( 0 ) = [ 100 , 0.02 ] T , M = 1 . The initial state estimate and the estimation error variance matrix are
x 0 = [ 10,000 m , 150 m / s , 15,000 m , 200 m / s ] T P 0 = d i a g { ( 100 m ) 2 , ( 14 m / s ) 2 , ( 100 m ) 2 , ( 15 m / s ) 2 } .
The estimated results of SCKF and HASCKF algorithms are shown in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10. The estimation error is given in Table 3.
From Figure 6 to Figure 10, it can be clearly seen that the standard SCKF estimation showed a large deviation after 40 s. However, HASCKF could still better estimate the state of the target. This is because the standard SCKF adopts an inaccurate prior noise variance R ^ ( 0 ) , while the HASCKF proposed in this paper estimates and corrects the inaccurate measurement noise variance, thus ensuring the accuracy and stability of the algorithm. As analyzed in Example 1, due to the randomness of noise, not all the noise variance estimates obtained by WF-NE were better. Therefore, the estimates of some components of the state obtained by SCKF were better than the one solved by HASCKF at some time in the interval [10 s, 40 s]. However, the general tend of the proposed HASCKF was better than SCKF, as shown in Figure 6, Figure 7, Figure 8, Figure 9 and Figure 10.

5. Conclusions

In this paper, aiming at the filtering problem caused by the inaccurate measurement noise variance in real systems, a weighted fusion estimation method is designed and a class of hybrid adaptive filtering structures is proposed, based on different measurement noise variance estimators. The stability and the filtering accuracy of the hybrid adaptive filter are analyzed and discussed theoretically. The simulation results showed that the algorithm had excellent accuracy and stability.
The work that can continue to be studied in the future includes further study on the research and design of fusion methods based on hybrid adaptive filtering in nonlinear centralized fusion framework and distributed fusion framework, research on functional equivalence among various fusion algorithms, and so on.

Author Contributions

Conceptualization, Y.S.; Methodology, Y.S.; Software, D.B.; Validation, X.Z.; Formal Analysis, X.F. and X.T.; Writing—Original Draft Preparation, Y.S. and X.F.; Writing—Review & Editing, X.Z. and X.T.

Funding

This research was supported in part by the National Natural Science Foundation of China (Grant Nos. 61503174, 61503336, 61503136, and 61602141), Henan higher education school young backbone teacher training program (2017GGJS160), Henan University of Animal Husbandry and Economy key scientific research innovation fund (XKYCXJJ2017009).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Han, C.Z.; Zhu, H.Y.; Duan, Z.S. Multi-source Information Fusion, 2nd ed.; Tsinghua University Press: Beijing, China, 2010. [Google Scholar]
  2. Ge, Q.; Shao, T.; Yang, Q.; Shen, X.; Wen, C. Multisensor Nonlinear Fusion Methods Based on Adaptive Ensemble Fifth-Degree Iterated Cubature Information Filter for Biomechatronics. IEEE Trans. Syst. Man Cybern. Syst. 2017, 46, 912–925. [Google Scholar] [CrossRef]
  3. Julier, S.J.; Uhlmann, J.K. New extension of the Kalman filter to nonlinear systems. In Proceedings of the Signal Processing, Sensor Fusion, and Target Recognition VI, Orlando, FL, USA, 28 July 1997; pp. 182–193. [Google Scholar]
  4. Arasaratnam, I.; Haykin, S. Cubature kalman filter. IEEE Trans. Autom. Control 2009, 54, 1254–1269. [Google Scholar] [CrossRef]
  5. Arasaratnam, I.; Haykin, S.; Thomas, R.H. Cubature kalman filtering forcontinuous-discrete systms: Theory and simulations. IEEE Trans. Signal Process. 2010, 58, 4977–4993. [Google Scholar] [CrossRef]
  6. Hernandez, W.; De Vicente, J.; Sergiyenko, O.Y.; Fernández, E. Improving the Response of Accelerometers for Automotive Applications by Using LMS Adaptive Filters: Part II. Sensors 2010, 10, 952–962. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  7. Yang, Y.; Gao, W. A new learning statistic for adaptive filter based on predicted residuals. Prog. Nat. Sci. Mater. Int. 2006, 16, 833–837. [Google Scholar]
  8. Sage, A.; Husa, G.W. Adaptive filtering with unknown prior statistics. In Proceedings of the Joint Automatic Control Conference, Boulder, CO, USA, 5–7 August 1969; pp. 760–769. [Google Scholar]
  9. Sage, A.P.; Husa, G.W. Algorithms for sequential adaptive estimation of prior statistics. In Proceedings of the 8th Symp on Adaptive Processes Decision and Control, Session 6a, University Park, PA, USA, 17–19 November 1969; pp. 61–70. [Google Scholar]
  10. Gao, S.; Hu, G.; Zhong, Y. Windowing and random weighting-based adaptive unscented Kalman filter. Int. J. Adapt. Control Signal Process. 2015, 29, 201–223. [Google Scholar] [CrossRef]
  11. Gao, W.G.; Yang, Y.X.; Cuix, Q. Application of adaptive kalman filtering algorithm in imu/GPS integrated navigation system. Geo-Spat. Inf. Sci. 2007, 10, 22–26. [Google Scholar] [CrossRef]
  12. Li, W.; Sun, S.; Jia, Y.; Du, J. Robust unscented Kalman filter with adaptation of process and measurement noise variances. Digit. Signal Process. 2016, 48, 93–103. [Google Scholar] [CrossRef]
  13. Zheng, B.; Fu, P.; Li, B.; Yuan, X. A Robust Adaptive Unscented Kalman Filter for Nonlinear Estimation with Uncertain Noise variance. Sensors 2018, 18, 808. [Google Scholar] [CrossRef]
  14. Ding, J.L.; Xiao, J. Design of adaptive cubature Kalman filter based on maximum a posteriori estimation. Control Decis. 2014, 29, 327–334. [Google Scholar]
  15. Zhao, L.; Wang, X.X.; Sun, M.; Ding, J.C.; Yan, C. Adaptive UKF Filtering Algorithm Based on Maximum a Posterior Estimation and Exponential Weighting. Acta Autom. Sin. 2010, 36, 1007–1019. [Google Scholar] [CrossRef]
  16. Sarkka, S.; Nummenmaa, A. Recursive noise adaptive kalman filtering by variational bayesian approximations. IEEE Trans. Autom. Control 2009, 54, 596–600. [Google Scholar] [CrossRef]
  17. Ge, Q.; Wei, Z.; Liu, M.; Yu, J.; Wen, C. Adaptive Quantized Estimation Fusion Using Strong Tracking Filtering and Variational Bayesian. IEEE Trans. Syst. Man Cybern. Syst. 2017, 1–12. [Google Scholar] [CrossRef]
  18. Sun, C.; Zhang, Y.; Wang, G.; Gao, W. A New Variational Bayesian Adaptive Extended Kalman Filter for Cooperative Navigation. Sensors 2018, 18, 2538. [Google Scholar] [CrossRef] [PubMed]
  19. Li, J.; Deng, F.; Chen, J. A Fast Distributed Variational Bayesian Filtering for Multisensor LTV System With Non-Gaussian Noise. IEEE Trans. Cybern. 2018, 1–13. [Google Scholar] [CrossRef] [PubMed]
  20. Chandra, K.P.B.; Gu, D.W.; Postlethwaite, I. Square root cubature information filter. IEEE Sens. J. 2013, 13, 750–758. [Google Scholar] [CrossRef]
  21. Ge, Q.B.; Xu, D.X.; Wen, C.L. Cubature information filters with correlated noises and their applications in decentralized fusion. Signal Process. 2014, 94, 434–444. [Google Scholar] [CrossRef]
  22. Xiong, K.; Zhang, H.Y.; Chan, C.W. Performance evaluation of ukf-based nonlinear fltering. Automatica 2006, 42, 261–270. [Google Scholar] [CrossRef]
  23. Wu, Y.X.; Hu, D.W.; Hu, X.P. Comments on “performance evaluation of ukf-based nonlinear filtering”. Automatica 2007, 43, 567–568. [Google Scholar] [CrossRef]
  24. Xiong, K.; Zhang, H.Y.; Chan, C.W. Author’s reply to “comments on ’performance evaluation of ukf-based nonlinear fltering”’. Automatica 2007, 43, 569–570. [Google Scholar] [CrossRef]
  25. Farina, A.; Ristic, B.; Benvenuti, D. Tracking a ballistic target: Comparison of several nonlinear filters. IEEE Trans. Aerosp. Electron. Syst. 2002, 38, 854–867. [Google Scholar] [CrossRef]
  26. Tichavsky, P.; Muravchik, C.H.; Nehorai, A. Posterior cramér-Rao bounds for discrete-time nonlinear filtering. IEEE Trans. Signal Process. 1998, 46, 1386–1396. [Google Scholar] [CrossRef]
Figure 1. Principle block diagram of the HASCKF algorithm.
Figure 1. Principle block diagram of the HASCKF algorithm.
Sensors 18 04335 g001
Figure 2. Estimation of measurement noise variance.
Figure 2. Estimation of measurement noise variance.
Sensors 18 04335 g002
Figure 3. Absolute estimation error of measurement noise variance.
Figure 3. Absolute estimation error of measurement noise variance.
Sensors 18 04335 g003
Figure 4. Estimation of measurement noise variance.
Figure 4. Estimation of measurement noise variance.
Sensors 18 04335 g004
Figure 5. Absolute estimation error of measurement noise variance.
Figure 5. Absolute estimation error of measurement noise variance.
Sensors 18 04335 g005
Figure 6. Target’s trajectory and tracking result of SCKF and HASCKF.
Figure 6. Target’s trajectory and tracking result of SCKF and HASCKF.
Sensors 18 04335 g006
Figure 7. Absolute error curves of X-displacement.
Figure 7. Absolute error curves of X-displacement.
Sensors 18 04335 g007
Figure 8. Absolute error curves of Y-displacement.
Figure 8. Absolute error curves of Y-displacement.
Sensors 18 04335 g008
Figure 9. Absolute error curves of X-velocity.
Figure 9. Absolute error curves of X-velocity.
Sensors 18 04335 g009
Figure 10. Absolute error curves of Y-velocity.
Figure 10. Absolute error curves of Y-velocity.
Sensors 18 04335 g010
Table 1. The mean absolute error of three noise estimation algorithms in Situation 1.
Table 1. The mean absolute error of three noise estimation algorithms in Situation 1.
AlgorithmMAP-NEVB-NEWF-NE
Mean absolute error0.00100.00150.0005
CPU time cost0.21880.21880.3725
Table 2. The mean absolute error of three noise estimation algorithms in Situation 2.
Table 2. The mean absolute error of three noise estimation algorithms in Situation 2.
AlgorithmMAP-NEVB-NEWF-NE
Mean absolute error0.00240.00220.0019
Table 3. The mean absolute error of two algorithms.
Table 3. The mean absolute error of two algorithms.
Mean Absolute ErrorAlgorithms
SCKFHASCKF
X-Position (m)182.94776.4784
X-Velocity (m/s)5.21130.7898
Y-Position (m)129.522824.6147
Y-Velocity (m/s)3.67661.0834

Share and Cite

MDPI and ACS Style

Shi, Y.; Tang, X.; Feng, X.; Bian, D.; Zhou, X. Hybrid Adaptive Cubature Kalman Filter with Unknown Variance of Measurement Noise. Sensors 2018, 18, 4335. https://doi.org/10.3390/s18124335

AMA Style

Shi Y, Tang X, Feng X, Bian D, Zhou X. Hybrid Adaptive Cubature Kalman Filter with Unknown Variance of Measurement Noise. Sensors. 2018; 18(12):4335. https://doi.org/10.3390/s18124335

Chicago/Turabian Style

Shi, Yuepeng, Xianfeng Tang, Xiaoliang Feng, Dingjun Bian, and Xizhao Zhou. 2018. "Hybrid Adaptive Cubature Kalman Filter with Unknown Variance of Measurement Noise" Sensors 18, no. 12: 4335. https://doi.org/10.3390/s18124335

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop