Next Article in Journal
On the Uniqueness of the Solution to the Inverse Problem of Determining the Diffusion Coefficient of the Magnetic Field Necessary for Constructing Analytical Models of the Magnetic Field of Mercury
Previous Article in Journal
Multilevel Fuzzy Inference System for Estimating Risk of Type 2 Diabetes
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Higher-Order Extended Cubature Kalman Filter Method Using the Statistical Characteristics of the Rounding Error of the System Model

1
School of Information and Control Engineering, Jilin Institute of Chemical Technology, Jilin 132022, China
2
School of Automation, Guangdong University of Petrochemical Technology, Maoming 525000, China
*
Author to whom correspondence should be addressed.
Mathematics 2024, 12(8), 1168; https://doi.org/10.3390/math12081168
Submission received: 10 March 2024 / Revised: 8 April 2024 / Accepted: 9 April 2024 / Published: 12 April 2024

Abstract

:
The cubature Kalman filter (CKF) cannot accurately estimate the nonlinear model, and these errors will have an impact on the accuracy. In order to improve the filtering performance of the CKF, this paper proposes a new CKF method to improve the estimation accuracy by using the statistical characteristics of rounding error, establishes a higher-order extended cubature Kalman filter (RHCKF) for joint estimation of sigma sampling points and random variables of rounding error, and gives a solution method considering the rounding error of multi-level approximation of the original function in the undermeasured dimension. Finally, numerical simulations show that the RHCKF has a better estimation effect than the CKF, and that the filtering accuracy is improved by using the information of the higher-order rounding error, which also proves the effectiveness of the method.

1. Introduction

Nonlinear state estimation is a popular research field, which is widely used in chemical processes, target tracking, signal processing, macroeconomic forecasting, and other engineering fields [1,2]. Nonlinear filtering is an important method to solve the nonlinear state estimation, and there are many research results. Ito et al. proposed a Gaussian nonlinear filtering framework and gave the optimal solution of nonlinear filtering under the Gaussian assumption [3]. Arellano-Valle et al. (2018) proposed the Kalman filter based on the skew-normal distribution as a non-Gaussian assumption of innovations [4]. It is difficult to get the optimal solution in practical processing; thus, only suboptimal approximation methods can be used instead. The commonly used nonlinear filtering algorithms include the extended Kalman filter (EKF) [5], untraceable Kalman filter (UKF) [6], and cubature Kalman filter (CKF) [7]. Among them, the EKF transforms the nonlinear state and the measurement function into a linear problem through the first-order Taylor series expansion [8]; the UKF overcomes the disadvantage of the EKF that the nonlinear function must be continuously differentiable and needs to solve the Jacobi matrix, it approximates the probability density function of the state by using a set of weighted sums of sampling points passed through a nonlinear function, and its filtering effect is significantly improved compared with that of the EKF [9]; however, when the dimensionality is too high, the UKF can be used to filter the state of the nonlinear state, which is a very important factor to improve the filtering effect. When the dimensionality is too high, the UKF filters poorly and even diverges, and its filtering effect is easily affected by the parameter settings [10,11]. Similar to the UKF, the CKF approximates a specific class of nonlinear integrals by obtaining a deterministic set of samples and weights through the spherical–radial Cubature law and moment matching [12]. The CKF has been widely studied and applied due to its advantages of high accuracy, low complexity, and good convergence. Compared with the EKF, the CKF does not need to calculate the Jacobi matrix and can eliminate the linearization error, which solves the theoretical limitations of the EKF; and compared with the UKF, the CKF has more stable performance and higher filtering accuracy when solving the filtering problem of high-dimensional strong nonlinear systems [13,14,15]. Thus, it has been widely used since proposed and has attracted many scholars who have attempted to improve it. Jia et al. [16] utilized the arbitrary-order fully symmetric spherical interpolation criterion and the moment matching method [17,18] to derive a higher-order spherical-phase-diameter cubature rule that can obtain arbitrary-order estimation accuracy. Using this rule, the high-order cubature Kalman filter (HCKF) [19] can be obtained, which improves the filtering accuracy of the CKF. In order to further improve the filtering accuracy, this paper proposes the high-order extended cubature Kalman filter considering rounding error, on the traditional CKF algorithm. This paper considers the rounding error that exists in the nonlinear approximation of the CKF. The algorithmic steps of the rounding error from the first to the general order are derived in detail, and it is proven by the simulation results that the new method improves the filtering accuracy.
The main contributions of this paper as follows: (1) proposing a new CKF method to improve the estimation accuracy by using the statistical characteristics of rounding error; and (2) adopting the idea of multi-level gradual approximation; the rounding error of the multi-level approximation of the original function in the undermeasured dimension is solved.

2. Problem Statement

Consider the following discrete-time nonlinear dynamic system:
x ( k + 1 ) = f ( x ( k ) , k ) + w ( k ) y ( k + 1 ) = h ( x ( k + 1 ) , k + 1 ) + v ( k + 1 )
where k is the discrete time; x ( k ) R n is the state vector; y ( k + 1 ) R m is the observation vector; f ( ) is the state transfer function; h ( ) is the measurement function. System modeling errors w ( k ) and v ( k + 1 ) are uncorrelated white noise sequences.
The system satisfies the following assumptions:
E { w ( k ) } = 0 , E { v ( k ) } = 0
E { w ( k ) w T ( j ) } = Q ( k ) δ k j
E { v ( k + 1 ) v T ( j ) } = R ( k + 1 ) δ { k + 1 , j }
where Q ( k ) is the variance of noise w ( k ) ; R ( k + 1 ) is the variance of noise v ( k + 1 ) ; δ k j is the Kronecker product.
The cubature Kalman filtering algorithm converts nonlinear filtering into an integral problem of solving the product of a nonlinear function and a Gaussian probability density, and then approximates the state a posteriori probability density using a weighted sum of 2 n cubature points. For an arbitrary distribution function ρ ( x ) , the integral problem is solved using the cubature integration criterion [20], the basic cubature points, and the corresponding weights obtained using the cubature criterion, as follows:
R n ρ ( x ) N ( x ; μ , ) d x i = 1 2 n ω i ρ ( μ + ζ i )
ζ i = n e i ,   ω i = 1 2 n ,   i = 1 , 2 , 2 n
where N ( x ; μ , ) denotes that the state variable x satisfying the mean μ is a Gaussian distribution with variance–covariance matrix ; ζ i represents the product points; ω i represents the corresponding weights, is the covariance and satisfies = T , T denotes the transpose of ; 2 n is the number of all cubature points, n is the state dimension of the system; e i denotes the i-th cubature point. When n = 2 , the corresponding set of Cubature points is e = [ 1 0 1 0 0 1 0 1 ] .
Assuming that the state estimates and covariances at the time of k + 1 are x ^ ( k + 1 | k + 1 ) and P ( k + 1 | k + 1 ) , respectively, and the posterior probability density function at this time is P ( x ( k + 1 ) ) = N ( x ^ ( k + 1 | k + 1 ) , P ( k + 1 | k + 1 ) ) , the standard CKF algorithm was originally proposed in [21], and can be divided into temporal updating and measurement updating as follows.
  • Time Update
(1) Initialize the state quantities and estimate the error covariance
x ^ ( 0 | 0 ) = E { x ( 0 ) }
P ( 0 | 0 ) = E { [ x ( 0 ) x ^ ( 0 | 0 ) ] [ x ( 0 ) x ^ ( 0 | 0 ) ] T }
(2) Construct a sigma point set under the nonlinear transformation f ( ) to construct the point set { x i ( k ) } i = 1 2 n , which contains a total of 2 n points, where x i ( k ) are the sampling points obtained by sampling.
(3) Perform the Cholesky decomposition on P ( k | k ) and compute the P ( k | k ) square root matrix:
P ( k | k ) = S ( k | k ) S ( k | k ) T
S ( k | k ) = P ( k | k )
(4) Generate cubature points:
x i ( k | k ) = x ^ ( k | k ) + S ( k | k ) ζ i ,   i = 1 , 2 2 n
where x i ( k | k ) is the cubature point of the state cubature at the time k and ζ i is satisfied:
ζ i = n e i ,   i = 1 , 2 , , 2 n
where e i is the vector of the column of the matrix and n is the dimension of the state vector.
(5) Calculate the state cubature point of the sampling point:
x ^ i ( k + 1 | k ) = f ( x ^ i ( k | k ) )
(6) Calculate the state prediction value:
x ^ ( k + 1 | k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 | k )
(7) Calculate the estimation error covariance:
P ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + Q ( k )  
  • Measurement Updates
(8) Construct the state cubature point set { x ^ i ( k + 1 | k ) } i = 1 2 n + 1 , and calculate the measured cubature point set { y ^ i ( k + 1 | k ) } i = 1 2 n + 1 .
(9) Calculate the P ( k + 1 | k ) square root matrix:
S ( k + 1 | k ) = P ( k + 1 | k )
(10) Calculate the state cubature point:
x ^ i ( k + 1 | k ) = x ^ ( k + 1 | k ) + S ( k + 1 | k ) ζ i
where x ^ i ( k + 1 | k ) is the cubature point of the state quantity of the system at the time k k + 1 .
(11) Calculate the measured cubature points after the nonlinear transformation:
y ^ i ( k + 1 | k ) = h ( x ^ i ( k + 1 | k ) )
(12) Calculate the measured predicted value:
y ^ ( k + 1 | k ) = i = 1 2 n w i c y ^ i ( k + 1 | k )
(13) Calculate the self-covariance:
P y y ( k + 1 | k ) = i = 1 2 n w i c [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] T + R ( k )
(14) Calculate the inter-covariance:
P x y ( k + 1 | k ) = i = 1 2 n w i c [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] T
(15) Calculate the gain array:
K ( k + 1 ) = P x y ( k + 1 | k ) P y y ( k + 1 | k ) 1
(16) Calculate the updated covariance:
P ( k + 1 | k + 1 ) = P ( k + 1 | k ) K ( k + 1 ) P y y K T ( k + 1 )
(17) Calculate the updated state:
x ^ ( k + 1 | k + 1 ) = x ^ ( k + 1 | k ) + K ( k + 1 ) [ y ( k + 1 ) y ^ ( k + 1 | k ) ]

3. Cubature Kalman Filtering Method Considering Rounding Error Information (RHCKF)

In order to utilize the information of the nonlinear system state model and nonlinear observation model to a greater extent, the RHCKF takes the error information discarded in the process of CKF sampling approximation to the real value into account in the state estimation process. Therefore, the RHCKF can obtain higher estimation accuracy. The RHCKF algorithm can be divided into time updating and measurement updating, and the specific steps are as follows:
  • Time Update
(1) Initialize the state quantities and estimate the error covariance:
x ^ ( 0 | 0 ) = E { x ( 0 ) }
P ( 0 | 0 ) = E { [ x ( 0 ) x ^ ( 0 | 0 ) ] [ x ( 0 ) x ^ ( 0 | 0 ) ] T }
(2) Construct a sigma point set under the nonlinear transformation f ( ) to construct the point set { x i ( l ) ( k ) } i = 1 2 n ; the point set contains a total of 2 n points, where x i ( k ) is a sampling point obtained by sampling.
(3) Perform the Cholesky decomposition on P ( k | k ) and compute the P ( k | k ) square root matrix:
P ( k | k ) = S ( k | k ) S ( k | k ) T ,   S ( k | k ) = P ( k | k )
(4) Generate cubature points with equal weights:
x i ( k | k ) = x ^ ( k | k ) + S ( k | k ) ζ i ,   i = 1 , 2 2 n
where x i ( k | k ) is the cubature point of the state cubature at moment k and ξ i is satisfied:
ζ i = n e i ,   i = 1 , 2 , , 2 n
where e i is the vector of the column of the matrix and n is the dimension of the state vector.
(5) Calculate the state Cubature point of the sampling point:
x ^ i ( k + 1 | k ) = f ( x ^ i ( k | k ) )
(6) Weight each prediction in Equation (25) to obtain the weighted fusion prediction estimate of x ( k + 1 ) , respectively:
x ^ ( k + 1 | k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 | k )

3.1. Cubature Kalman Filter Considering First-Order Rounding Error

For the state prediction error, in the standard CKF algorithm uses the weighted average of the state prediction value instead of the true value. In this process, there will inevitably be rounding error; thus, the true value of the rounding error information needs to be taken into account in the following equations to take into account the existence of the first-order error information, to calculate the true value of the first-order RHCKF at the ( k + 1 ) time:
x ( 1 ) ( k + 1 ) = f ( x ( k ) ) + w ( k )   = x ^ ( k + 1 | k ) + f ( x ( k ) ) x ^ ( k + 1 | k ) + w ( k )   = 1 2 n i = 1 2 n x ^ i ( k + 1 | k ) + ξ ( 1 ) ( k ) + w ( k )
where ξ ( 1 ) ( k ) is the first-order rounding error given by ξ ( 1 ) ( k ) = f ( x ( k ) ) x ^ ( k + 1 | k ) , and w ( k ) is zero-mean Gaussian noise.
Calculating the updated state prediction value, we have:
x ^ ( 1 ) ( k + 1 ) = 1 2 n i = 1 2 n x ^ i ( k + 1 | k ) + ξ ^ ( 1 ) ( k )
The first-order rounding error ξ ( 1 ) ( k ) is identified by least squares (LS) by bringing x ( 1 ) ( k + 1 ) into the measurement equation:
y ( 1 ) ( k + 1 ) = h ( x ( 1 ) ( k + 1 ) ) + v ( k + 1 ) H ( k + 1 )   x ( 1 ) ( k + 1 ) + v ( k + 1 ) = H ( k + 1 ) [ 1 2 n i = 1 2 n x ^ i ( k + 1 ) + w ( k ) + ξ ( 1 ) ( k ) ] + v ( k + 1 ) = H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) + H ( k + 1 ) w ( k ) + H ( k + 1 ) ξ ( 1 ) ( k ) + v ( k + 1 )
where H ( k + 1 ) are first-order Jacobian matrices, H ( k + 1 ) = h ( x ( k + 1 ) ) x | x ( k + 1 ) = x ^ ( k + 1 | k ) ; and v ( k + 1 ) is zero-mean Gaussian noise.
y ( 1 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ¯ i ( 0 ) ( k + 1 ) = H ( k + 1 ) ξ ( 1 ) ( k ) + H ( k + 1 ) w ( k )
The left side of the equation is denoted by y ¯ ( k + 1 ) and the right side of the equation is denoted by v ¯ ( k + 1 ) , to get:
y ¯ ( 1 ) ( k + 1 ) = y ( 1 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 )
v ¯ ( 1 ) ( k + 1 ) = H ( k + 1 ) w ( k ) + v ( k + 1 )
Combining Equations (33) and (32), we can write (31) as:
y ¯ ( 1 ) ( k + 1 ) = H ( k + 1 ) ( ξ ( 1 ) ( k ) ) + v ¯ ( 1 ) ( k + 1 ) ξ ( 1 ) ( k ) N [ ξ ^ ( 1 ) ( k ) , P ( 1 , ξ ) ( k | k ) ]
Calculating from the LS formula ξ ^ ( 1 ) ( k ) , we get:
ξ ^ ( 1 ) ( k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 y ¯ ( k + 1 )
Update the state values by rounding error and find the state prediction error by the updated state values P ( 1 ) ( k + 1 | k ) :
P ( 1 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ( 1 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( 1 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + ξ ^ ( k ) ξ ^ ( k ) T + Q ( k )

3.2. Cubature Kalman Filter Considering Second-Order Rounding Error

When considering only the first-order rounding error, there is still unutilized error information such that the estimation accuracy will definitely be inaccurate. Therefore, it is necessary to consider the second-order error information ξ ( 2 ) ( k ) , after identifying the first-order error information, and it is necessary to update the information on the true value of the x ( 2 ) ( k + 1 ) .
x ( 2 ) ( k + 1 ) = f ( x ( k ) ) + w ( k ) = x ^ ( k + 1 | k ) + f ( x ( k ) ) x ^ ( k + 1 | k ) + w ( k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ( 1 ) ( k ) + w ( k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ^ ( 1 ) ( k ) + ξ ˜ ( 1 ) ( k ) + w ( k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ^ ( 1 ) ( k ) + f ( x ( k ) ) x ^ ( k + 1 | k ) ξ ^ ( 1 ) ( k ) + w ( k ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ^ ( 1 ) ( k ) + ξ ( 2 ) ( k ) + w ( k )
where ξ ( 2 ) ( k ) is the second-order rounding error, ξ ( 2 ) ( k ) = f ( x ( k ) ) x ^ ( k + 1 | k ) ξ ^ ( 1 ) ( k ) .
Similarly, x ( 2 ) ( k + 1 ) is brought into the measurement equation to identify the second-order rounding error ξ ( 2 ) ( k ) by LS, and the first-order Jacobian matrix H ( k + 1 ) given by the following:
y ( 2 ) ( k + 1 ) = h ( x ( 2 ) ( k + 1 ) ) + v ( k + 1 ) H ( k + 1 )   x ( 2 ) ( k + 1 ) + v ( k + 1 )   = H ( k + 1 ) [ 1 2 n i = 1 2 n x ^ i ( k + 1 ) + w ( k ) + ξ ^ ( 1 ) ( k ) + ξ ( 2 ) ( k ) ] + v ( k + 1 )   = H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) + H ( k + 1 ) w ( k ) + H ( k + 1 ) ξ ^ ( 1 ) ( k )   + H ( k + 1 ) ξ ( 2 ) ( k ) + v ( k + 1 )
y ( 2 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) H ( k + 1 ) ξ ^ ( 1 ) ( k ) = H ( k + 1 ) ξ ( 2 ) ( k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
The left side of the equation is denoted by y ¯ ( 2 ) ( k + 1 ) and the right side of the equation is denoted by v ¯ ( 2 ) ( k + 1 ) , to get:
y ¯ ( 2 ) ( k + 1 ) = y ( 2 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) H ( k + 1 ) ξ ^ ( 1 ) ( k )
v ¯ ( 2 ) ( k + 1 ) = H ( k + 1 ) w ( k ) + v ( k + 1 )
Combining Equations (41) and (40), one can write Equation (39) as:
y ¯ ( 2 ) ( k + 1 ) = H ( k + 1 ) ξ ( 2 ) ( k ) + v ¯ ( 2 ) ( k + 1 )
Calculating from the undermeasurement LS formula ξ ( 2 ) ( k ) , we have:
ξ ^ ( 2 ) ( k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 y ¯ ( 2 ) ( k + 1 )
Update the state values by rounding error and find the state prediction error by the updated state values P ( 2 ) ( k + 1 | k ) :
P ( 2 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( κ + 1 | k ) x ( 2 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( 2 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + Q ( κ ) + ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) T ( k ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) T ( k )

3.3. Cubature Kalman Filter Considering Rounding of the General Order l − 1

The (l − 1)-order rounding error ξ ( l 1 ) ( k ) calculation steps can be found in the Appendix A:
x ( l 1 ) ( k + 1 ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 2 ) ( k ) + ξ ( l 1 ) ( k ) + w ( k )
where ξ ( l 1 ) ( k ) is the (l − 1)-order rounding error; w ( k ) is zero-mean Gaussian noise.
Update the state values using the rounding error and find the state prediction error covariance matrix by the updated state values P ( l 1 ) ( k + 1 | k ) :
P ( l 1 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ( l 1 ) ( κ + 1 | κ ) ] [ x ^ i ( κ + 1 | κ ) x ( l 1 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + Q ( κ ) + ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) ( k ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) ( k ) + + ξ ^ ( l 1 ) ( k ) ξ ^ ( l 1 ) ( k )
By mathematical induction, we know that when the rounding error is considered to the l-th order, the equation of state for the l-th order rounding that we use instead of considering the rounding error is satisfied:
x ( l ) ( k + 1 ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + w ( k ) + ξ ^ ( 1 ) ( k ) + ξ ˜ ( 1 ) ( k ) + + ξ ( l ) ( k )
Bringing x ( l ) ( k + 1 )   into the observation equation, we have:
y ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) + ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 1 ) ( k ) ] = H ( k + 1 ) ξ ( l ) ( k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
Using LS, we can find ξ ( l ) ( k ) , P ( l , ξ ) ( k | k ) :
ξ ^ ( l ) ( k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 y ¯ ( l ) ( k + 1 )
P ( l , ξ ) ( k | k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 R ¯ ( l ) ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 × H ( k + 1 )
When the rounding error of the system model is considered to the l-th order, is it necessary to continue to consider the rounding error? We could set the threshold λ to determine whether | | ξ ^ ( l ) ( k | k ) | | < λ is valid, where λ is the discriminant parameter decided by the system: if the inequality is valid, then it means that there is no need to consider the higher order; if it is not valid, then it is also necessary to continue to consider the rounding error.
Updating P ( l ) ( k + 1 | k ) is accomplished due to the higher-order unconsidered error information ξ ˜ ( l ) ( k ) in x ( l ) ( k + 1 )   , corresponding to the error covariance matrix P ( l , ξ ) ( k + 1 | k ) given by
P ( l ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ( l 1 ) ( κ + 1 | κ ) ] [ x ^ i ( κ + 1 | κ ) x ( l 1 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + Q ( κ )   + ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) ( k ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) ( k ) + + ξ ^ ( l ) ( k ) ξ ^ ( l ) ( k ) + P ( l , ξ ) ( k + 1 | k )
  • Measurement Updates
(1) Construct the state cubature point set { x ^ i ( l ) ( k + 1 | k ) } i = 1 2 n + 1 and calculate the measured cubature point set y ^ i ( l ) ( k + 1 | k ) } i = 1 2 n + 1
(2) Calculate the P ( k + 1 | k ) square root matrix:
S ( k + 1 | k ) = P ( k + 1 | k )
(3) Calculate the state cubature point:
x ^ i ( k + 1 | k ) = x ^ ( k + 1 | k ) + S ( k + 1 | k ) ζ i
where x ^ i ( k + 1 | k ) is the cubature point of the state quantity of the system at the time k k + 1 .
(4) Calculate the measured cubature points after the nonlinear transformation:
y ^ i ( k + 1 | k ) = h ( x ^ i ( k + 1 | k ) )
(5) Calculate the measurement prediction:
y ^ ( k + 1 | k ) = i = 1 2 n w i c y ^ i ( k + 1 | k )
(7) Calculate the self-covariance:
P y y ( k + 1 | k ) = i = 1 2 n w i c [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] T + R ( k )
(8) Calculate the inter-covariance:
P x y ( k + 1 | k ) = i = 1 2 n w i c [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ y ^ i ( k + 1 | k ) y ^ ( k + 1 | k ) ] T
(9) Calculate the gain array:
K ( k + 1 ) = P x y ( k + 1 | k ) P y y ( k + 1 | k ) 1
(10) Calculate the updated covariance:
P ( k + 1 | k + 1 ) = P ( k + 1 | k ) K ( k + 1 ) P y y K T ( k + 1 )
(11) Calculate the updated state:
x ^ ( l ) ( k + 1 | k + 1 ) = x ^ ( l ) ( k + 1 | k ) + K ( k + 1 ) [ y ( k + 1 ) y ^ ( k + 1 | k ) ]

4. Performance Analysis of RHCKF and CKF

4.1. Performance Analysis in the Prediction Phase

Taking the state value considering the l-order rounding error as the true value:
x ( l ) ( k + 1 ) = 1 2 n i = 1 2 n x ^ i ( k + 1 ) + ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 1 ) ( k ) + ξ ( l ) ( k ) + w ( k + 1 )
The prediction error covariance matrix P ( 0 ) ( k + 1 | k ) P ( l ) ( k + 1 | k ) calculation process can be found in the Appendix B.
Prediction error covariance matrix when considering (l − 1)-order rounding error P ( l 1 ) ( k + 1 | k ) :
P ( l 1 ) ( k + 1 | k ) = P ( l ) ( k + 1 | k ) + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k )
Prediction error covariance matrix when considering second-order rounding error P ( 2 ) ( k + 1 | k ) :
P ( 2 ) ( k + 1 | k ) = P ( 3 ) ( k + 1 | k ) + ξ ^ ( 3 ) ( k ) ξ ^ ( 3 ) T ( k )
Prediction error covariance matrix when considering first-order rounding error P ( 1 ) ( k + 1 | k ) :
P ( 1 ) ( k + 1 | k ) = P ( 2 ) ( k + 1 | k ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) T ( k )
Prediction difference covariance matrix without rounding errors P ( 0 ) ( k + 1 | k ) :
P ( 0 ) ( k + 1 | k ) = P ( 1 ) ( k + 1 | k ) + ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) T ( k )
Compared to the traditional CKF, we have utilized more information in the RHCKF, i.e., the error between the true value and the prediction estimate; therefore, in terms of information utilization, the filter design utilizing more information should have higher accuracy. Also, in terms of the prediction error covariance matrix that represents the performance metrics of the filter prediction stage, combining Equations (63) and (66), we have:
P ( 0 ) ( k + 1 | k ) P ( l ) ( k + 1 | k ) = ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) T ( k ) + + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k ) 0
It can be found that, equivalent to the CKF, the RHCKF utilizes more information in the prediction stage, which reduces the prediction error covariance matrix and increases the model prediction reliability. Therefore, the RHCKF has better prediction performance in the prediction stage compared to the traditional CKF.

4.2. Performance Analysis of the Update Phase

In order to facilitate the comparison of the size of the estimation error covariance matrix of the two filtering methods, first of all, we can get the collapsing transformation based on Equations (60) and (67):
P ( l 1 ) ( k + 1 | k + 1 ) 1 = P ( l 1 ) ( k + 1 | k ) 1 + H T ( k + 1 ) ( R ( l 1 ) ( k + 1 ) ) 1 H ( k + 1 )
P ( l ) ( k + 1 | k + 1 ) 1 = P ( l ) ( k + 1 | k ) 1 + H T ( k + 1 ) ( R ( l ) ( k + 1 ) ) 1 H ( k + 1 )
From the above equation, it can be found that the state estimation performance of the filter is mainly contributed by two aspects. One is the prediction error covariance that contains the prediction information, and the other is the measurement error that contains the measurement prediction information. It is clear that it is possible to obtain
P ( l ) ( k + 1 | k + 1 ) 1 P ( l 1 ) ( k + 1 | k + 1 ) 1 = P ( l ) ( k + 1 | k ) 1 P ( l 1 ) ( k + 1 | k ) 1 + H T ( k + 1 ) [ R ( l ) ( k + 1 ) R ( l 1 ) ( k + 1 ) ] 1 H ( k + 1 ) = P ( l ) ( k + 1 | k ) 1 P ( l 1 ) ( k + 1 | k ) 1 0
P ( l ) ( k + 1 | k + 1 ) P ( l 1 ) ( k + 1 | k + 1 )
It can be shown that, equivalent to the CKF, the RHCKF utilizes more information in the updating phase, and the extra information used improves the accuracy, i.e., more accurate state prediction values and measurement prediction values are obtained. Therefore, the RHCKF has better performance compared to traditional CKF filtering.

5. Numerical Simulation Verification

5.1. Experiment I

Consider the system where the equation of state is a two-dimensional nonlinear equation and the measurement equation is the one-dimensional under-measured nonlinear equation
[ x 1 ( k + 1 ) x 2 ( k + 1 ) ] = [ 0.85 x 1 ( k ) + 0.5 x 2 ( k ) + 0.5 sin ( γ x 1 ( k ) ) 0.5 x 1 ( k ) + 0.5 sin ( μ x 2 ( k ) ) ] + [ w 1 ( k ) w 2 ( k ) ]
z ( k + 1 ) = 2 sin ( x 1 ( k + 1 ) ) + sin ( x 2 ( k + 1 ) ) + v ( k + 1 )
where the white noise is set to w ( k ) N [ 0 , Q ] ; v ( k + 1 ) N [ 0 , R ] ; Q = d i a g { 0.01 , 0.01 } ; R = 0.01 ; and the initial values of the experimental parameters are x ^ 0 = [ 0.1 , 0.1 ] T , P 0 = 2 × d i a g ( 1 , 1 ) , and γ = 1 , respectively.
From Figure 1 and Figure 2, it can be seen that each component of the state estimation can effectively follow the true state component, and Figure 3 and Figure 4 show the results of the root mean square error (RMSE) for 100 steps. Compared with the CKF, the RHCKF proposed in this paper has a better estimation effect.
From Figure 3 and Figure 4 and Table 1, it can be seen that the average estimation errors of the first-order RHCKF for the target state are 0.0854 and 0.1281, respectively; that those of the second-order RHCKF for the target state are 0.0826 and 0.1272, respectively; that the improvements in the first-order RHCKF in comparison with the CKF are 13.23% and 7.33%; and that the improvements in the second-order RHCKF in comparison with the CKF are 17.07% and 8.09%. Moreover, the first-order RHCKF is more effective than the second-order RHCKF in improving the CKF because the first-order rounding error already contains most of the error information, and the second-order rounding error contains very little information, and the RHCKF is able to utilize more information of the model. The RHCKF takes into account the rounding error information in the process of approximating the nonlinear system of the sampled values, and it can utilize more information of the model compared to the CKF, so the filtering performance is also higher.

5.2. Experiment II

Consider a system with a nonlinear equation of state in two dimensions and a nonlinear equation of measurement in two dimensions.
[ x 1 ( k + 1 ) x 2 ( k + 1 ) ] = [ 0.85 x 1 ( k ) + 0.5 x 2 ( k ) + 0.5 sin ( γ x 1 ( k ) ) 0.5 x 1 ( k ) + 0.5 sin ( μ x 2 ( k ) ) ] + [ w 1 ( k ) w 2 ( k ) ]
[ z 1 ( k + 1 ) z 2 ( k + 1 ) ] = [ sin ( x 1 ( k + 1 ) ) sin ( x 2 ( k + 1 ) ) ] + [ v 1 ( k + 1 ) v 2 ( k + 1 ) ]
where the white noise is set to w ( k ) N [ 0 , Q ] ; v ( k + 1 ) N [ 0 , R ] ; Q = d i a g { 0.01 , 0.01 } ; R = d i a g { 0.01 , 0.01 } ; and the initial values of the experimental parameters are x ^ 0 = [ 0.1 , 0.1 ] T , P 0 = 2 × d i a g ( 1 , 1 ) , and γ = 1 , respectively.
For a two-dimensional nonlinear system, Figure 5 and Figure 6 show the filtering effects of several filtering methods, and Figure 7 and Figure 8 show the results of the root mean square error (RMSE) for 100 steps. Combining the data in Table 2, it can be seen that the estimation accuracy of the algorithm in this paper has been significantly improved regardless of the estimation of the state.
The dataa in Table 2 show that the estimated root mean square errors of the RHCKF algorithm in this paper are 0.0821 and 0.0910, respectively, which are significantly improved compared to the estimation accuracy of the EKF, UKF and CKF. Among them, the most improved was the EKF: x 1 = 13.03 % and x 2 = 8.24 % .
There is a lower bound for the unbiased estimator of the minimum variance of the state of the nonlinear filtering algorithm, which is commonly measured by the Cramer–Rao lower bound limit in practice. The simulation results show that, compared with the traditional CKF, the root mean square error of the state of the proposed RHCKF is closer to that of the Cramer–Rao lower bound, and the filtering performance is also better than that of the CKF, so the RHCKF is an effective new method for state estimation.

6. Conclusions

Aiming at discrete-time nonlinear dynamic system, the traditional CKF cannot estimate the nonlinear model accurately, and these errors will have an impact on the estimation accuracy; for this reason, this paper firstly proposes a method to improve the estimation accuracy of the CKF by using the statistical properties of the rounding error. As the number of orders of the rounding error utilized increases, the more the approximation of the nonlinear function is enhanced, which means that there is less information about the rounding error. At the same time, from the point of view of state estimation, the less the rounding error, the more information can be utilized to design the filter, and the better the estimation of the filter will be. Finally, a higher-order extended cubature Kalman filter, based on the statistical characteristics of the system rounding error, was designed. The simulation experiment shows that, compared with the typical nonlinear filtering of the EKF, UKF and CKF, this method has achieved better estimation results, and the filtering accuracy has been significantly improved, providing a new solution to the problem of state estimation.
Although the method in this paper has achieved good filtering results, there are still areas for improvement. We have only analyzed the case where the observation equations of the nonlinear model are linear, and not the case where both the equation of state and the observation equations are nonlinear, both of which need to take advantage of the presence of rounding errors, which needs to be investigated in the future. In addition, a Bayesian approach (Idrovo-Aguirre & Contreras-Reyes, 2022) for an extended cubature Kalman filter can be addressed in a further work [22]. These are the key points and difficulties of the next step of research.

Author Contributions

Conceptualization, H.Z. and C.W.; methodology, C.W.; software, H.Z.; validation, H.Z.; formal analysis, H.Z.; investigation, H.Z.; resources, H.Z.; data curation, H.Z.; writing—original draft preparation, H.Z.; writing—review and editing, C.W.; visualization, C.W.; supervision, C.W.; project administration, C.W.; funding acquisition, C.W. All authors have read and agreed to the published version of the manuscript.

Funding

(1) National Key R&D Program Intelligent Robot Key Special Project “Robot Joint Drive Control Integrated Chip” 2023YFB4704000; (2) National Natural Science Foundation of China 62125307, U22A2046, 61933013.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. The (l − 1)-Order Rounding Error ξ ( l 1 ) ( k ) Calculation Steps

Using LS estimation, the rounding error ξ ( l 1 ) ( k ) is identified, H ( k + 1 ) is the measurement matrix, and x ( l 1 ) ( k + 1 ) is brought into the measurement equation to obtain:
y ( l 1 ) ( k + 1 ) = h (   x ( l 1 ) ( k + 1 ) ) + v ( k + 1 )   H ( k + 1 )   x ( l 1 ) ( k + 1 ) + v ( k + 1 ) = H ( k + 1 ) [ 1 2 n i = 1 2 n x ^ i ( k + 1 ) + w ( k ) + ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 2 ) ( k ) + ξ ( l 1 ) ( k ) ] + v ( k + 1 ) = H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) + H ( k + 1 ) w ( k ) + H ( k + 1 ) [ ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 2 ) ( k ) ] + H ( k + 1 ) ξ ( l 1 ) ( k ) + v ( k + 1 )
y ( l 1 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) H ( k + 1 ) [ ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 2 ) ( k ) ]                                 = H ( k + 1 ) ξ ( l 1 ) ( k ) + H ( k + 1 ) w ( k ) + v ( k + 1 )
The left side of the equation is denoted by y ¯ ( l 1 ) ( k + 1 ) and the right side of the equation is denoted by v ¯ ( l 1 ) ( k + 1 ) , to get:
y ¯ ( l 1 ) ( k + 1 ) = y ( l 1 ) ( k + 1 ) H ( k + 1 ) 1 2 n i = 1 2 n x ^ i ( k + 1 ) H ( k + 1 ) [ ξ ^ ( 1 ) ( k ) + + ξ ^ ( l 2 ) ( k ) ]
v ¯ ( l 1 ) ( k + 1 ) = H ( k + 1 ) w ( k ) + v ( k + 1 )
Combining Equations (A2) and (A3), we can write (A1) as:
y ¯ ( l 1 ) ( k + 1 ) = H ( k + 1 ) ξ ( l 1 ) ( k ) + v ¯ ( l 1 ) ( k + 1 )
Using the under-measurement LS to compute the ξ ( l 1 ) ( k ) , we get:
ξ ^ ( l 1 ) ( k ) = H T ( k + 1 ) [ H ( k + 1 ) H T ( k + 1 ) ] 1 y ¯ ( l 1 ) ( k + 1 )

Appendix B. The Prediction Error Covariance Matrix P ( 0 ) ( k + 1 | k ) P ( l ) ( k + 1 | k ) Calculation Process

Prediction difference covariance matrix without rounding errors P ( 0 ) ( k + 1 | k ) :
P ( 0 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 ) x ( 0 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( 0 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T   + Q ( k ) + ξ ^ ( 1 ) ( k ) ξ ^ ( 1 ) T ( k ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) T ( k ) + + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k ) + P ( l , ξ ) ( k | k )
Prediction error covariance matrix when considering first-order rounding error P ( 1 ) ( k + 1 | k ) :
P ( 1 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ( 1 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( 1 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T   + Q ( κ ) + ξ ^ ( 2 ) ( k ) ξ ^ ( 2 ) T ( k ) + ξ ^ ( 3 ) ( k ) ξ ^ ( 3 ) T ( k ) + + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k ) + P ( l , ξ ) ( k | k )
Prediction error covariance matrix when considering second-order rounding error P ( 2 ) ( k + 1 | k ) :
P ( 2 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ^ i ( k + 1 ) x ( 2 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( 2 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ^ i ( k + 1 ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T   + Q ( κ ) + ξ ^ ( 3 ) ( k ) ξ ^ ( 3 ) T ( k ) + + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k ) + P ( l , ξ ) ( k | k )
Prediction error covariance matrix when considering (l − 1)-order rounding error P ( l 1 ) ( k + 1 | k ) :
P ( l 1 ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ¯ i ( k + 1 | k ) x ( l 1 ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( l 1 ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ¯ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T   + Q ( k ) + ξ ^ ( l ) ( k ) ξ ^ ( l ) T ( k ) + P ( l , ξ ) ( k | k )
Prediction error covariance matrix when considering l-order rounding error P ( l ) ( k + 1 | k ) :
P ( l ) ( k + 1 | k ) = 1 2 n i = 1 2 n [ x ¯ i ( k + 1 | k ) x ( l ) ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ( l ) ( k + 1 | k ) ] T = 1 2 n i = 1 2 n [ x ¯ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] [ x ^ i ( k + 1 | k ) x ^ ( k + 1 | k ) ] T + Q ( k ) + P ( l , ξ ) ( k | k )

References

  1. Li, L.-Q.; Zhao, D.; Luo, C.-D. A Novel Interacting TS Fuzzy Multiple Model by Using UKF for Maneuvering Target Tracking. In Proceedings of the 2019 22th International Conference on Information Fusion (FUSION), Ottawa, ON, Canada, 2–5 July 2019; pp. 1–7. [Google Scholar]
  2. Meles, M.; Rajasekaran, A.; Mela, L.; Ruttik, K.; Jäntti, R. Impact of carrier frequency offset and phase noise on the steering vector for 3D drone localization based on angle of arrival (AOA). In Proceedings of the 2023 16th International Conference on Signal Processing and Communication System (ICSPCS), Bydgoszcz, Poland, 6–8 September 2023; pp. 1–6. [Google Scholar]
  3. Ito, K.; Xiong, K. Gaussian filters for nonlinear filtering problems. IEEE Trans. Autom. Control 2000, 45, 910–927. [Google Scholar] [CrossRef]
  4. Yuan, Y.; Zhou, D.; Li, J.; Lou, C. Time-varying parameters estimation with adaptive neural network EKF for missile-dual control system. J. Syst. Eng. Electron. 2024, 1–12. [Google Scholar] [CrossRef]
  5. Arellano-Valle, R.B.; Contreras-Reyes, J.E.; Quintero, F.O.L.; Valdebenito, A. A skew-normal dynamic linear model and Bayesian forecasting. Comput. Stat. 2019, 34, 1055–1085. [Google Scholar] [CrossRef]
  6. Cheng, C.; Wang, W.; Meng, X.; Shao, H.; Chen, H.J. Sigma-Mixed Unscented Kalman Filter-based Fault Detection for Traction Systems in High-speed Trains. Chin. J. Electron. 2023, 32, 982–991. [Google Scholar] [CrossRef]
  7. Pagoti, S.K.; Vemuri, S.I.D. Development and performance evaluation of Correntropy Kalman Filter for improved accuracy of GPS position estimation. Int. J. Intell. Netw. 2022, 3, 1–8. [Google Scholar] [CrossRef]
  8. Bucy, R.S.; Senne, K.D. Digital synthesis of non-linear filters. J. Autom. 1971, 7, 287–298. [Google Scholar] [CrossRef]
  9. Julier, S.J.; Uhlmann, J.K. Unscented filtering and nonlinear estimation. Proc. IEEE 2004, 92, 401–422. [Google Scholar] [CrossRef]
  10. Zhou, W.; Hou, J. A new adaptive high-order unscented Kalman filter for improving the accuracy and robustness of target tracking. IEEE Access 2019, 7, 118484–118497. [Google Scholar] [CrossRef]
  11. Wang, X.-X.; Pan, Q.; Huang, H.; Gao, A. Overview of deterministic sampling filtering algorithms for nonlinear system. Control Decis. 2012, 27, 801–812. [Google Scholar]
  12. Huber, M.F.; Hanebeck, U.D. Gaussian filter based on deterministic sampling for high quality nonlinear estimation. IFAC Proc. Cuba 2008, 41, 13527–13532. [Google Scholar] [CrossRef]
  13. Tian, Y.; Huang, Z.; Tian, J.; Li, X. State of charge estimation of lithium-ion batteries based on cubature Kalman filters with different matrix decomposition strategies. Energy 2022, 238, 121917. [Google Scholar] [CrossRef]
  14. Li, H.; Sun, H.; Chen, B.; Shen, H.; Yang, T.; Wang, Y.; Jiang, H.; Chen, L. A cubature Kalman filter for online state-of-charge estimation of lithium-ion battery using a gas-liquid dynamic model. J. Energy Storage 2022, 53, 105141. [Google Scholar] [CrossRef]
  15. Peng, J.; Luo, J.; He, H.; Lu, B. An improved state of charge estimation method based on cubature Kalman filter for lithium-ion batteries. Appl. Energy 2019, 253, 113520. [Google Scholar] [CrossRef]
  16. Jia, B.; Xin, M.; Cheng, Y. High-degree cubature Kalman filter. Automatica 2013, 49, 510–518. [Google Scholar] [CrossRef]
  17. Genz, A. Fully symmetric interpolatory rules for multiple integrals over hyper-spherical surfaces. J. Comput. Appl. Math. 2003, 157, 187–195. [Google Scholar] [CrossRef]
  18. He, J.; Sun, C.; Zhang, B.; Wang, P. Maximum correntropy square-root cubature Kalman filter for non-Gaussian measurement noise. IEEE Access 2020, 8, 70162–70170. [Google Scholar] [CrossRef]
  19. Zhang, L.; Cui, N.; Yang, F.; Lu, F.; Lu, B. High-degree cubature Kalman filter and its application in target tracking. J. Harbin Eng. Univ. 2016, 37, 573–578. [Google Scholar]
  20. Yang, H.; Wang, B.; He, J. Adaptive cubature Kalman filter for unknown noise covariance. J. Air Force Eng. Univ. 2021, 22, 42–47. [Google Scholar]
  21. Arasaratnam, I.; Haykin, S. Cubature kalman filters. IEEE Trans. Autom. Control 2009, 54, 1254–1269. [Google Scholar] [CrossRef]
  22. Idrovo-Aguirre, B.J.; Contreras-Reyes, J.E. Bayesian monthly index for building activity based on mixed frequencies: The case of Chile. J. Econ. Stud. 2022, 49, 541–557. [Google Scholar] [CrossRef]
Figure 1. State x 1 estimates.
Figure 1. State x 1 estimates.
Mathematics 12 01168 g001
Figure 2. State x 2 estimates.
Figure 2. State x 2 estimates.
Mathematics 12 01168 g002
Figure 3. RMSE x 1 .
Figure 3. RMSE x 1 .
Mathematics 12 01168 g003
Figure 4. RMSE x 2 .
Figure 4. RMSE x 2 .
Mathematics 12 01168 g004
Figure 5. State x 1 estimates.
Figure 5. State x 1 estimates.
Mathematics 12 01168 g005
Figure 6. State x 2 estimates.
Figure 6. State x 2 estimates.
Mathematics 12 01168 g006
Figure 7. RMSE x 1 .
Figure 7. RMSE x 1 .
Mathematics 12 01168 g007
Figure 8. RMSE x 2 .
Figure 8. RMSE x 2 .
Mathematics 12 01168 g008
Table 1. Comparison of estimation errors.
Table 1. Comparison of estimation errors.
Methodologies x 1 x 2
CKF0.09670.1375
First-order RHCKF0.08540.1281
VS CKF13.23%7.33%
Second-order RHCKF0.08260.1272
VS CKF17.07%8.09%
Table 2. Performance comparison of different algorithms.
Table 2. Performance comparison of different algorithms.
RMSEEKFUKFCKFRHCKFVS EKFVS UKFVSCKF
state x 1 0.09280.08740.08630.082113.03%6.45%5.11%
x 2 0.09850.09420.09350.09108.24%3.51%2.74%
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Zhang, H.; Wen, C. A Higher-Order Extended Cubature Kalman Filter Method Using the Statistical Characteristics of the Rounding Error of the System Model. Mathematics 2024, 12, 1168. https://doi.org/10.3390/math12081168

AMA Style

Zhang H, Wen C. A Higher-Order Extended Cubature Kalman Filter Method Using the Statistical Characteristics of the Rounding Error of the System Model. Mathematics. 2024; 12(8):1168. https://doi.org/10.3390/math12081168

Chicago/Turabian Style

Zhang, Haiyang, and Chenglin Wen. 2024. "A Higher-Order Extended Cubature Kalman Filter Method Using the Statistical Characteristics of the Rounding Error of the System Model" Mathematics 12, no. 8: 1168. https://doi.org/10.3390/math12081168

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop