Next Article in Journal
Kinematic Analysis of a Gear-Driven Rotary Planting Mechanism for a Six-Row Self-Propelled Onion Transplanter
Next Article in Special Issue
Design of an Embedded Energy Management System for Li–Po Batteries Based on a DCC-EKF Approach for Use in Mobile Robots
Previous Article in Journal
RISE-Based Composite Adaptive Control of Electro-Hydrostatic Actuator with Asymptotic Stability
Previous Article in Special Issue
A Simple Soft Computing Structure for Modeling and Control
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improved Invariant Kalman Filter for Lie Groups Attitude Dynamics with Heavy-Tailed Process Noise

1
Key Laboratory of Advanced Process Control for Light Industry, Institute of Automation, Jiangnan University, Wuxi 214111, China
2
School of Electronics and Information Engineering, Harbin Institute of Technology, Shenzhen 518055, China
3
Department of Electronic and Computer Engineering, Hong Kong University of Science and Technology, Hong Kong, China
*
Author to whom correspondence should be addressed.
Machines 2021, 9(9), 182; https://doi.org/10.3390/machines9090182
Submission received: 20 July 2021 / Revised: 21 August 2021 / Accepted: 23 August 2021 / Published: 27 August 2021
(This article belongs to the Special Issue Modeling, Sensor Fusion and Control Techniques in Applied Robotics)

Abstract

:
Attitude estimation is a basic task for most spacecraft missions in aerospace engineering and many Kalman type attitude estimators have been applied to the guidance and navigation of spacecraft systems. By building the attitude dynamics on matrix Lie groups, the invariant Kalman filter (IKF) was developed according to the invariance properties of symmetry groups. However, the Gaussian noise assumption of Kalman theory may be violated when a spacecraft maneuvers severely and the process noise might be heavy-tailed, which is prone to degrade IKF’s performance for attitude estimation. To address the attitude estimation problem with heavy-tailed process noise, this paper proposes a hierarchical Gaussian state-space model for invariant Kalman filtering: The probability density function of state prediction is defined based on student’s t distribution, while the conjugate prior distributions of the scale matrix and degrees of freedom (dofs) parameter are respectively formulated as the inverse Wishart and Gamma distribution. For the constructed hierarchical Gaussian attitude estimation state-space model, the Lie groups rotation matrix of spacecraft attitude is inferred together with the scale matrix and dof parameter using the variational Bayesian iteration. Numerical simulation results illustrate that the proposed approach can significantly improve the filtering robustness of invariant Kalman filter for Lie groups spacecraft attitude estimation problems with heavy-tailed process uncertainty.

1. Introduction

The navigation and control operations in aerospace engineering usually require adequate accurate knowledge of spacecraft orientation in space, and attitude determination from vector observations is usually employed in astronautically applications [1,2,3,4]. The widely used Kalman type filters can infer the state in Euclidean vector space by fusing attitude dynamic and observation sensor according to their probabilities [5,6,7,8]. Recently, the building of attitude dynamics on matrix Lie groups has been investigated actively for the navigation and control of spacecraft targets because it can significantly improve the estimation and control performance to make full use of the geometrical properties of Lie groups models [9,10].
For attitude estimation, there have already been significant research interests in the estimation and control of spacecraft targets such as the invariant observers and filters for attitude dynamics on the special orthogonal group SO(3), or the navigation systems defined on the special Euclidean group SE(3) [11,12]. Accounting for the geometrical invariance of symmetry attitude dynamics, the resulted attitude description with rotation matrix on SO(3) can avoid the trouble of singularities and unwinding encountered when using unit quaternion [13] and, to some extent, also contribute to better estimation performance for attitude determination algorithms [14]. Exploiting the mapping relation between matrix Lie groups and Lie algebra, invariant Kalman filter (IKF) can convert the attitude dynamics on SO(3) to Euclidean vector space and mimic classical Kalman filtering steps [15]. IKF is based on the invariance and log-linearity of symmetry dynamics, which then leads to better convergence on a much bigger set of state trajectories [16]. In these available researches, the process noise of attitude dynamical model was assumed to be a concentrated Gaussian distribution [17,18] and the statistics parameter are also assumed to be accurately given in advance as required in classical Kalman theory [19,20]. However, due to the presence of server maneuvers in space missions, there might be some unpredictable disturbances or outliers that would contaminate the Gaussian distribution and violate the required assumptions for Kalman filtering on matrix Lie groups.
For Euclidean space filtering problems, some adaptive methods are commonly used to deal with the trouble of inaccurate/unknown noise parameters [21,22,23]. Some adaptive techniques have been designed to estimate online unknown/inaccurate noise parameters, mainly including innovation-based adaptive methods and multiple model methods [24,25,26]; nevertheless, the convergence of innovation-based estimation methods cannot be guaranteed while the parallel-running multiple filter banks might increase the computational complexities. The variational Bayesian(VB) iteration-based adaptive methods have also been employed to jointly determine the system state and inaccurate noise statistics [27,28,29]. However, these adaptive methods all assume the noise distribution to be Gaussian and cannot be applied to the filtering cases with heavy-tailed noise. For heavy-tailed system noises, the Student’s t distribution assumes the distribution of the latent state and system noises as a joint Student’s t-distribution, and it has been proved to be more suitable to represent the non-Gaussian probability density function (PDF) [30,31].
To deal with the heavy-tailed process noise caused by severe maneuvering of rigid bodies, a robust Student’s t-based hierarchical Gaussian system model for the SINS/GPS integration is presented by modeling the prior probability density function as a Student’s t distribution and the conjugate prior distributions [21]. To improve the filtering performance of Student’s t-based methods for cases with both heavy-tailed process and observation noises, a robust Student’s t-based Kalman filter is proposed employing the variational Bayesian-based iterative approach in [32]; in the work of [33], by applying the Student’s t approximate distributions to posterior filtering and smoothing probability density functions, the robust nonlinear Student’s t-based filter and smoother are also presented as the generalization and extension of the linear Student’s t filtering methods. Note that the above researches provide a sound theoretical foundation for attitude estimation problems with heavy-tailed noises, but they are originally designed in Euclidean space systems for filtering of vector variable states. To the best of our knowledge, for attitude estimation problem in SO(3), until now there is still no special study on the robust Student’s t-based adaptive methods for invariant Kalman filtering on matrix Lie groups.
Note that although the available invariant Kalman filter also obeys the basic theory of classical Kalman filter, the above researches about robust Student’s t-based methods cannot be directly applied to matrix Lie groups systems. The uncertainty definition of concentrated Gaussian distribution on matrix Lie groups requires the assumption of noisy errors being small enough [34,35], which is the condition for first order approximation to the Baker–Campbell–Hausdorff (BCH) formula [36,37]. For attitude estimation on SO(3) in aerospace engineering, the presence of heavy-tailed noises and disturbances imposes a critical constraint on the application of robust Student’s t-based methods to improve the performance of invariant Kalman filter.
According to the above discussions, it is rather meaningful to investigate the robust Student’s t-based invariant Kalman filtering method especially for the matrix Lie groups attitude estimation problems in the presence of heavy-tailed outliers in system process noise. In this work, for attitude estimation problem with heavy-tailed process noise, this paper proposes a hierarchical Gaussian state-space model for invariant Kalman filtering, and the Lie groups rotation matrix of spacecraft attitude is inferred together with the scale matrix and dof parameter using the variational Bayesian iteration. Numerical simulation results illustrate that the proposed approach can significantly improve the filtering robustness of invariant Kalman filter for Lie groups spacecraft attitude estimation problems with heavy-tailed process uncertainty.

2. Primaries and Problem Definition

2.1. The Attitude Estimation System on the Special Orthogonal Group SO(3)

Attitude determinition from two vector observations is an important task in most aerospace engineering. The attitude dynamic on special orthogonal group S O ( 3 ) 3 × 3 with the associated observation vectors can be formulated as the following system model [1,14]
R k = Exp G w k R k 1 Ω k 1
Y k = y k y k = R k T b + v k R k T b + v k ,
where the special orthogonal group S O ( 3 ) : = R R T = I 3 × 3 , det R = 1   is the group of 3 × 3 rotation matrices and R k S O ( 3 ) is the mapping rotation from the body frame of the spacecraft to the earth-fixed frame at time instant k ; Ω k S O ( 3 ) denotes the control rotation; Y k 6 is a composition of the vector observations y k , y k 3 with their respective observation vector b , b 3 ; v k ~ N c v k ; 0 3 × 1 , Σ v , and v k ~ N c v k ; 0 3 × 1 , Σ v are Gaussian white and mutually independent observation noises with N c denoting the Gaussian distribution. Note that the vector observation based attitude determination is a common application scenario because the Sun sensor, magnetometer, horizon sensor, and the gravitational accelerometer are usually equipped by most spacecrafts in astronautical engineering [1]. A graphical presentation of target application scenario is given in Figure 1.
Note that the additive forms of vector noises in Euclidean space are not supported in Lie Groups space. The vector process noise w k 3 is injected through Lie exponential operator Exp G ( ) and, for any w = w 1   w 2   w 3 T 3 , the projection between the vector noise w = w 1   w 2   w 3 T 3 and the associated groups element R S O 3 are
Exp G : 3 S O 3 , Exp G 1 : S O 3 3
Exp G w : = exp w × = m = 0 w × m m ! = I 3 × 3 + sin w w w × + 2 sin w / 2 2 w 2 w × 2 ,
w × = w 1 w 2 w 3 × = 0 w 3 w 2 w 3 0 w 1 w 2 w 1 0 3 × 3 ,  
where exp ( ) is the matrix exponential operator and w × is the skew-symmetric matrix. The concept of concentrated Gaussian distribution (CGD) was employed to define the probability distribution for random variables on matrix Lie groups [19,20,37]. For attitude estimation, in concentrated Gaussian distribution the groups variable R S O 3 with mean R ¯ S O 3 and covariance Σ 3 × 3 is defined as
R ~ G R ; R ¯ , Σ S O 3   if   ξ Exp G 1 R R ¯ 1 ~ N c ξ ; 0 3 × 1 , Σ 3
where G , denotes the concentrated Gaussian distribution (CGD). Note that, in the definition of CGD, the eigenvalues of Σ are assumed to be small enough, i.e., almost all the mass of the distribution should be concentrated in a small neighborhood around the mean of the rotation variable [20,37]. A graphical illustration of the relation between the concentrated Gaussian distribution on matrix Lie groups SO(3) and the classical Gaussian distribution relation is presented in Figure 2.

2.2. Model Projection Based on the Invariance Property of Attitude Estimation System

For attitude estimation on SO(3), the usual Kalman filtering methods in Euclidean vector space could not be applied directly. However, by properly defining the system variable, the Lie groups dynamic model (1) can be projected into the usual Euclidean space based on the invariance property. Let R ^ k 1 be an estimate for R k 1 at instant k − 1 and R ^ k = R ^ k 1 Ω k 1 be the prediction for R k . Define the invariant error ξ k = Exp G 1 R ^ k R k 1 , then the evolution model of invariant error ξ k obeys
ξ k = Exp G 1 R ^ k R k 1 = Exp G 1 R ^ k 1 Ω k 1 Exp G w k R k 1 Ω k 1 1      = Exp G 1 R ^ k 1 R k 1 1 Exp G w k = Exp G 1 Exp G ξ k 1 Exp G w k      = BCH ξ k 1 , w k = ξ k 1 w k + O ξ k 1 2 , w k 2 , ξ k 1 w k ,
where the BCH , of the last step is a first order form of the Baker–Campbell–Hausdorff (BCH) formula [34,35,36]. Note that if the norm of the error term ξ k and w k are small enough so that the second order term in (7) could be negligible, then the matrix Lie groups attitude dynamic (1) for rotation matrix R k could be approximately converted into the following linear state space model in Euclidean space for the invariant error ξ k
ξ k ξ k 1 w k .
As for observation model (2), the observation sequence Y k can also be reformulated into a function of the invariant error ξ k , i.e., the new innovation sequence z k 6 [14]
z k = R ^ k Y k b b = R ^ k y k b R ^ k y k b = R ^ k R k T b b b b + R ^ k | v k v k : = H ξ ξ k + V k ,
where H = b × b ×   6 × 3 and V k = R ^ k | k 1 v k v k 6 are the projected observation matrix and noise, respectively; note that, since the observation noises v k , v k are isotropic and mutually independent noises and with R ^ k R ^ k T = I 3 × 3 , we have Σ V = Cov V k V k T = d i a g Σ v , Σ v .
Obviously, based on the definition of the invariant error ξ k , the Lie groups dynamics (1) and observation model (2) could be converted as the Euclidean space model (8) and (9) for the invariant error ξ k (as in Figure 3). Note that both the true rotation state R k 1 and the control rotation Ω k 1 do not appear in the converted model (8) and (9); therefore, the evolving motion of ξ k is actually independent of the true trajectory of rotation R k 1 and control rotation Ω k 1 , which is called the invariance property of symmetry model [15,16].

2.3. The Invariant Kalman Filter for Attitude Estimation

As the consequence of invariance property, for the attitude estimation system, the probability distribution of rotation state R k based on the observation Y k is equivalent to that of the invariant error ξ k in 3 , constituting the basis of invariant Kalman filter. Then, if the process noise obeys classical Gaussian distribution, i.e., w k ~ N c w k ; 0 3 × 1 , Σ w 3 , the filtering steps for attitude estimation system on Lie group SO(3) can be implemented by mimicking that of the classical Kalman filter for invariant error ξ k in Euclidean space.
For the attitude dynamic (1) and associated observation (2), invariant Kalman filter (IKF) aims to recursively predict and correct the estimate for the rotation state. Given the initial estimate R ^ 0 | 0 = Exp G ξ ^ 0 | 0 R 0 for true R 0 with error ξ ^ 0 | 0 N c ξ ^ 0 | 0 ; 0 3 × 1 , Σ 0 | 0 , as the prediction for prior error estimate ξ ^ k | k 1 based on (8), the prior error rotation estimate R ^ k | k 1 based on the dynamic model (1) can be implemented as follows
R ^ k | k 1 = R ^ k 1 | k 1 Ω k 1   ,
where R ^ k | k 1 and R ^ k 1 | k 1 are the prior and posterior error estimate for true rotation state R k , respectively; ξ ^ k | k 1 and ξ ^ k 1 | k 1 are associated with the prior and posterior error estimate for true error ξ k . The covariance propagation of rotation group R k is same as that of the classical Kalman filter for invariant error ξ k based on the evolution model (8),
Σ k | k 1 = Σ k 1 | k 1 + Σ w ,
where Σ k 1 | k 1 stands for the covariance of the posterior estimate ξ ^ k | k 1 and R ^ k | k 1 . Therefore, based on the converted innovation (9), the correction steps of IKF for the rotation state R ^ k | k 1 can be implemented by updating the estimate R ^ k | k 1 with innovation sequence z k
R ^ k | k = Exp G K k z k R ^ k | k 1 ,
K k = Σ k | k 1 H T / H Σ k | k 1 H T + Σ V ,
Σ k | k = Σ k | k 1 K k H Σ k | k 1 ,
where K k is the Kalman gain matrix, and the posterior estimate R ^ k | k is then updated with Lie exponential operation of the correction term K k z k as in (12) [16,17]. A graphical presentation of the filtering diagram of invariant Kalman filter is presented in Figure 4.

2.4. The Attitude Estimation Problem with the Trouble of Heavy-Tailed Process Noise

Obviously, the invariant Kalman filter (IKF) is a general extension of the Euclidean space Kalman filter to matrix Lie groups systems based on the invariance property of symmetry dynamics [16]. However, for the attitude estimation in aerospace engineering, there is still some difficulty for practical implementation of the IKF-based attitude estimation because the presence of heavy-tailed process noise in attitude dynamic might introduce great challenge and trouble to the theory of IKF:
Firstly, in the filtering step (10)~(14) of IKF, the process noise of attitude was assumed to be a concentrated Gaussian distribution and the statistics parameter is also required to be accurately known (as in Figure 4). However, due to the presence of severely maneuvering operations in space tasks, there might be some unpredictable disturbances or substantial outliers that would contaminate the Gaussian distribution and violate the assumptions for Kalman filtering on matrix Lie groups. Therefore, the theoretical foundation for optimal Kalman filtering is sure to be destroyed, which will degrade the final estimation.
Secondly, for the invariant projection of the Lie groups dynamics (1) into the linear vector form (8) in Section 2.2, the first order approximation to BCH formula is employed based on the assumption that both ξ k and w k were small enough and the second and higher order residual could be dropped from (7) to (8). However, for aerospace engineering applications, the unpredictable disturbances or substantial outliers caused by severely maneuvering operations might violate this condition and lead to inaccurate approximation to the system model, which is prone to a potential source of error and influences the estimation accuracy of IKF.
For attitude determination in astronautical engineering, it is rather meaningful to further study Lie groups estimation problems with heavy-tailed process noise and investigate new robust adaptive version of invariant Kalman filtering methods to improve the filtering precisions for cases with heavy-tailed process noises.

3. Robust Student’s t Based Invariant Kalman Filter for Attitude Estimation on SO(3)

3.1. Probability View of Attitude Estimation with Heavy-Tailed Process Noise

For attitude estimation, if the process noise w k in dynamic model (1) is heavy-tailed and contaminated by larger outliers, the Gaussian distribution and concentrated Gaussian distribution cannot accurately describe the probability distribution of heavy-tailed process noise. In this work, to account for the larger outliers due to severely maneuvering, the probability of the heavy-tailed noise is described as the student’s t distribution
p w k = S t w k ; 0 3 × 1 , Σ w , τ k = 0 + N c w k ; 0 3 × 1 , Σ w / γ w Γ γ w ; τ w 2 , τ w 2 d γ w
where p w k denotes the probability density of the process noise w k ; S t ; μ , Σ , τ denotes the student’s t distribution with mean vector μ , scale matrix Σ , and degree of parameter τ ; γ w is an auxiliary random variable in (15); and Γ γ ; α , β represents the Gamma probability density function of the scalar γ with α and β as the shape and rate parameter, i.e.,
Γ γ ; α , β = β α ϒ α γ α e β γ , γ > 0 0 , γ 0
where e is the natural constant and ϒ represents the Gamma function. Obviously, the probability density function of the student’s t distribution can be regarded as the infinite mixture of the classical Gaussian probability density function [21,30].
If given the Gaussian approximation of the posterior probability density function p ξ k 1 z 1 : k 1 and concentrated Gaussian approximation of posterior p R k 1 Y 1 : k 1 for instant k−1, the prior probability density function of p ξ k z 1 : k 1 and p R k Y 1 : k 1 for time instant k can be obtained according to the Chapman–Kolmogorov equation [38,39],
p ξ k z 1 : k 1 = p w ξ k ξ k 1 p ξ k 1 z 1 : k 1 d ξ k 1 = p w ξ k ξ k 1 N c ξ k 1 ; ξ ^ k 1 k 1 , Σ k 1 k 1 d ξ k 1
p R k Y 1 : k 1 = p w Exp G 1 R k Ω k 1 R k 1 p R k 1 Y 1 : k 1 d ξ k 1 = p w Exp G 1 R k Ω k 1 R k 1 G R k 1 ; R ^ k 1 k 1 , Σ k 1 k 1 d R k 1
where p ξ k 1 z 1 : k 1 denotes the posterior probability density function of the ξ k 1 based on the innovation sequence z k from time instant 1 to k − 1 and p R k 1 Y 1 : k 1 denotes the posterior probability density function of rotation R k based on observation Y k from instant 1 to k − 1. Note that resorting to the equivalence of the R k and ξ k in probability density function, p w ξ k ξ k 1 = p w Exp G 1 R k Ω k 1 R k 1 is the student’s t distribution p w k for system process noise and so we have p R k Y 1 : k 1 = p ξ k z 1 : k 1 . Further, it is evident that the prior probability p ξ k z 1 : k 1 for instant k is not Gaussian but a student’s t distribution with mean vector parameter ξ ^ k | k 1 , scale matrix Σ k | k 1 and parameter τ k , i.e.,
p R k Y 1 : k 1 = p ξ k z 1 : k 1 S t ξ k ; ξ ^ k | k 1 , Σ k | k 1 , τ k = 0 + N c ξ k ; ξ ^ k | k 1 , Σ k | k 1 / γ k Γ γ k ; τ k 2 , τ k 2 d γ k
where ξ ^ k | k 1 , Σ k | k 1 , and τ k are the mean vector, scale matrix, and degree of freedom parameter of the prior student’s t distribution p R k Y 1 : k 1 ; and γ k is the auxiliary random variable. Employing (15)~(19), the prior student’s t distribution p R k Y 1 : k 1 can be reformulated into a hierarchical composition of the Gaussian distributions
p ξ k γ k , z 1 : k 1 = N c ξ k ; ξ ^ k k 1 , Σ k k 1 / γ k ,
p γ k = Γ γ k ; τ k 2 , τ k 2 .
Attitude estimation aims to propagate the posterior probability density distribution based on the observation Y k . Since the observation noise in (2) and its equivalent form V k in (9) are zero mean Gaussian white sequence with covariance Σ V , the likelihood probability density function of the innovation sequence z k based on the error ξ k is equal to that of observation Y k given the rotation R k , i.e.,
p Y k R k = p z k ξ k = N c z k ; H ξ ξ k , Σ V
where p Y k R k is the probability density function of Y k given rotation R k , and p z k ξ k is the probability density function of z k based on error ξ k .
Note that the accurate propagation of the probability density function requires accurate knowledge of scale matrix Σ k k 1 and degree of freedom τ k . However, for real engineering applications in the presence of unpredictable disturbances or substantial outliers, these two distribution parameters are usually not correct or unavailable in advance. Therefore, from the view of probability density, the main troubles of attitude estimation with heavy-tailed process noise are threefold:
(1)
With the student’s t distribution process noise, the probability density function p ξ k z 1 : k 1 in classical invariant Kalman filter depends on the auxiliary random variable γ k and becomes the hierarchical form p ξ k γ k , z 1 : k 1 in (20) and (21).
(2)
For the unpredictable disturbances or outliers induced by severely maneuvering operations, the accurate scale matrix Σ k k 1 and degree of freedom γ k are usually unavailable but essential to propagate the posterior estimates.
As a consequence, the heavy-tailed attitude estimation problem cannot be solved with the invariant Kalman filter in Section 2 since the distribution parameter Σ k k 1 and τ k should be estimated together with the invariant error ξ k and the rotation state R k .

3.2. Prior Probability Definition for the Parameters of Student’s t Distribution

To solve Lie groups filtering problems with Student’s t distribution process noise, this work aims to estimate the rotation group state R k together with Σ k k 1 and τ k by variational Bayesian iterations as in Figure 5. The prior distributions of parameter Σ k k 1 and τ k based on all historical observation sequences should be conjugate prior distributions so that their posterior distribution would maintain the same form as their prior distribution.
According to (20) and (21), the estimated scale matrix Σ k k 1 is proportional to the covariance matrix of Gaussian distribution. Therefore, the commonly used inverse Wishart distribution and the Gamma distribution of Bayesian statistics are employed as the conjugate prior distributions for parameters Σ k k 1 and γ k , respectively.
The distribution of scale matrix Σ k k 1 is defined as inverse Winshart distribution [29,39,40] which is often used as conjugate prior for the covariance matrix of the Gaussian variable. For attitude estimation problem, the inverse Winshart function of Σ k k 1 can be written as
p Σ k k 1 = IW Σ k k 1 ; λ k , Ψ k = Ψ k λ k / 2 Σ k k 1 ( λ k + 3 + 1 ) / 2 2 3 λ k / 2 ϒ 3 λ k / 2 e 0.5 t r Ψ k Σ k k 1 1 ,
where IW denotes the inverse Winshart distribution; λ k denotes the degrees of freedom; Ψ k is the 3 × 3 inverse scale matrix; is the determinant operator; and ϒ 3 is the 3-variate gamma function [29]. For the scale matrix Σ k k 1 IW Σ k k 1 ; λ k , Ψ k , the expectation of the matrix inverse is E Σ k k 1 1 = λ k 3 1 Ψ k 1   for   λ k > 3 + 1 and so we have
E Σ k k 1 = Ψ k λ k 3 1 ,   if   λ > 3 + 1 .
The distribution parameter τ k can be defined as the Gamma distribution [21],
p τ k = Γ τ k ; a k , b k ,
where a k , b k are respectively the shape and rate parameter of the Gamma distribution τ k , respectively.
In summary, Equations (20)~(25) actually constitute a new student’s t-based hierarchical Gaussian state-space model for the heavy-tailed attitude estimation problem, whose diagram is presented in Figure 6. The initialization of the required distribution parameter is based on the nominal parameter Σ k k 1 , which will be discussed later on. Using the above distribution definition, the rotation state R k and ξ k , the auxiliary random variable γ k , the scale matrix Σ k k 1 , and the parameter τ k will be jointly estimated based on the newly constructed hierarchical Gaussian state-space model by variational Bayesian iterations, which then leads to the new robust student’s t-based invariant Kalman filter for the heavy-tailed attitude estimation problem.

3.3. Variational Beayesian Approximations of Posterior Probability Density Function

To jointly estimate the rotation state R k , auxiliary random variable γ k , scale matrix Σ k k 1 , and parameter τ k requires calculating the posterior probability density function p ξ k , γ k , Σ k | k 1 , τ k z 1 : k . Since the accurate analytical expression of p ξ k , γ k , Σ k | k 1 , τ k z 1 : k is not available, this work tries to obtain its factored approximate based on variational Bayesian iterations [38]. i.e.,
p ξ k , γ k , Σ k | k 1 τ k z 1 : k q ξ k q γ k q Σ k | k 1 q τ k ,
where q ξ k , q γ k , q Σ k | k 1 , and q τ k represent, respectively, the approximate posterior probability density function of ξ k , γ k , Σ k k 1 , τ k ; q ξ k , q γ k , q Σ k | k 1 and q τ k , which can be determined by minimizing the Kullback–Leibler divergence (KLD) [38,41] between the factored approximate q ξ k q γ k q Σ k | k 1 q τ k and the true p ξ k , γ k , Σ k | k 1 , τ k z 1 : k [29,38], i.e.,
q ξ k , q γ k , q Σ k | k 1 , q τ k = arg min KLD q ξ k q γ k q Σ k | k 1 q τ k p ξ k , γ k , Σ k | k 1 , τ k Y 1 : k ,
where K L D q x p x q x log q x p x d x denotes the KLD between q x and p x . According to the variational Bayesian iteration, these approximate posterior q ξ k , q γ k , q Σ k | k 1 , and q τ k to (27) satisfy the following equations [21,38]
log q ξ k = E ξ k log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c ξ k ,
log q γ k = E γ k log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c γ k ,
log q Σ k k 1 = E Σ k k 1 log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c Σ k k 1 ,
log q τ k = E τ k log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c τ k ,
where log denotes the natural logarithm function; E ξ k represents the mathematical expectation operation for all variables except ξ k , while E γ k , E Σ k k 1 , and E τ k are with similar definitions; c ξ k is the constant with respect to ξ k , while c γ k , c Σ k k 1 , and c τ k are similarly defined. Obviously, Equations (28)~(31) are coupled and so no analytical solutions to the above equations exist. Therefore, only the approximate solutions are available and in this work the fixed-point iterations are employed to determine these parameters [38].
According to the conditional independence properties, the joint probability density function p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k can be factored as follows
p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k = p z k ξ k p ξ k z 1 : k 1 , γ k p γ k p Σ k | k 1 p τ k p z 1 : k 1
Then, using (20)~(26), we have
p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k   = N c z k ; H ξ k , Σ V N c ξ k ; ξ ^ k k 1 , Σ k k 1 / γ k   × Γ γ k ; τ k 2 , τ k 2 IW Σ k | k 1 ; λ k , Ψ k Γ τ k ; a k , b k p Y 1 : k 1
Using (16) and (24), log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k can be formulated as
log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k = 1 2 z k H ξ k T Σ V 1 z k H ξ k 1 2 5 + λ k log Σ k | k 1 1 2 tr Ψ k Σ k | k 1 1   1 2 γ k ξ k ξ ^ k k 1 T Σ k | k 1 1 ξ k ξ ^ k k 1 + τ k + 1 2 log γ k τ k γ k 2 + τ k 2 log τ k 2 log ϒ τ k 2   + a k 1 log τ k b k τ k + constant .

3.4. Fixed-Point Iteration of the System State and Distribution Parameters

Based on the conditional independence property of the invariant error ξ k and parameters γ k , Σ k k 1 , and τ k , the Equations (28)~(31) can be rewritten with above Equation (34) and fixed-point methods can be used to obtain the approximate estimates to the invariant error ξ k as well as the parameters γ k , Σ k k 1 , and τ k [21].

3.4.1. Fixed-Point Iteration of the Invariant Error ξ k

With (34) and the independence property of γ k and Σ k k 1 , (28) can be written as
log q ( i + 1 ) ( ξ k ) = E ξ k ( i + 1 ) [ log p ( ξ k , γ k , Σ k k 1 , τ k , z 1 : k ) ] + c ξ k = 1 2 ( z k H ξ k ) T Σ V 1 ( z k H ξ k ) 1 2 E ξ k ( i ) [ γ k ( ξ k ξ ^ k k 1 ) T Σ k k 1 1 ( ξ k ξ ^ k k 1 ) ] + c ξ k = 1 2 ( z k H ξ k ) T Σ V 1 ( z k H ξ k ) 1 2 E γ k ( i ) [ γ k ] ( ξ k ξ ^ k k 1 ) T E Σ k k 1 ( i ) [ Σ k k 1 1 ] ( ξ k ξ ^ k k 1 ) + c ξ k = 1 2 ( z k H ξ k ) T Σ V 1 ( z k H ξ k ) 1 2 ( ξ k ξ ^ k k 1 ) T ( Σ ¯ k k 1 ( i ) ) 1 ( ξ k ξ ^ k k 1 ) + c ξ k
where q i + 1 is the iterative approximation to q at i + 1th iteration; E ξ k i represents the mathematical expectation at ith iteration for variables except the invariant error ξ k while E γ k i and E Σ k | k 1 i denote the mathematical expectation at ith iteration for γ k and Σ k | k 1 , respectively; Σ k | k 1 i is the modified prior error covariance matrix calculated as
Σ k | k 1 i = E Σ k | k 1 i Σ k | k 1 1 1 / E γ k i γ k .
Then, q i + 1 ξ k can also be updated as a Gaussian distribution, i.e.,
q i + 1 ξ k = N c ξ k ; ξ ^ k | k i + 1 , Σ k | k i + 1 ,
where ξ ^ k | k i + 1 and Σ k | k i + 1 can be determined by following the update steps
ξ ^ k | k i + 1 = ξ ^ k | k 1 + K k i + 1 z k H ξ ^ k | k 1 ,
K k i + 1 = Σ k | k 1 i H T H Σ k | k 1 i H T + Σ V 1 ,
Σ k | k i + 1 = Σ k | k 1 i K k | i + 1 H Σ k | k 1 i ,
where K k i + 1 denotes the calibrated Kalman gain matrix at the i + 1th iteration.

3.4.2. Fixed-Point Iteration of the Auxiliary Random Variable γ k

Using (34) and conditional independency, the log q i + 1 γ k in (29) can be rewritten as
log q i + 1 γ k = E γ k i + 1 log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c γ k = E τ k i τ k + 1 2 log γ k 1 2 E γ k i ξ k ξ ^ k | k 1 T Σ k | k 1 1 ξ k ξ ^ k | k 1 + E τ k i τ k γ k + c γ k = E τ k i τ k + 1 2 log γ k 1 2 t r E ξ k i ξ k ξ ^ k | k 1 ξ k ξ ^ k | k 1 T E Σ k | k 1 i Σ k | k 1 1 + E τ k i τ k γ k + c γ k .
where E γ k i represents the mathematical expectation at ith iteration for variables except the γ k while E τ k i denotes the mathematical expectation at ith iteration for the parameter τ k . Note that the approximate ξ ^ k | k i + 1 and Σ k | k i + 1 updated in (35)~(40) can be used to propagate Equation (41) and the term E ξ k i ξ k ξ ^ k | k 1 ξ k ξ ^ k | k 1 T can be reformulated as
E γ k i ξ k ξ ^ k | k 1 ξ k ξ ^ k | k 1 T = E γ k i ξ k ξ ^ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ k ξ ^ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 T      = E γ k i + 1 ξ k ξ ^ k | k i + 1 ξ k ξ ^ k | k i + 1 T + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T = Σ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T
Then (41) could be deduced into
log q i + 1 γ k = E τ k i τ k + 3 2 1 log γ k 1 2 t r Σ k | k i + 1 E Σ k | k 1 i Σ k | k 1 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T E Σ k | k 1 i Σ k | k 1 1 + E τ k i τ k γ k + c γ k .
which is actually also a Gamma distribution as (16), i.e.,
q i + 1 γ k = Γ γ k ; η i + 1 , θ i + 1 ,
where η i + 1 and θ i + 1 are the shape and rate parameter determined by
η i + 1 = E τ k i τ k + 3 2 ,
θ i + 1 = 1 2 t r Σ k | k i + 1 + U k i + ξ ^ k | k i + 1 u ^ k i ξ ^ k | k i + 1 u ^ k i T E Σ k | k 1 i Σ k | k 1 1 + E τ k i τ k .

3.4.3. Fixed-Point Iteration of the Prior Estimate for Covariance Matrix Σ k | k 1

With (34), (37), (44) and the independence of ξ k and γ k , (30) can be rewritten as
log q i + 1 Σ k | k 1 = E Σ k k 1 i + 1 log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c Σ k | k 1 = 1 2 E Σ k k 1 i γ k ξ k ξ ^ k | k 1 T Σ k | k 1 1 ξ k ξ ^ k | k 1 1 2 5 + λ k T log Σ k | k 1 1 2 t r Ψ k Σ k k 1 1 + c Σ k | k 1 = 1 2 5 + λ k T log Σ k | k 1 1 2 t r Ψ k + E ξ k i + 1 ξ k B k i + 1 Σ k k 1 1 + c Σ k | k 1
where the term B k i + 1 is calculated as
B k i + 1 = E Σ k k 1 i + 1 ξ k ξ ^ k | k 1 ξ k ξ ^ k | k 1 T = E Σ k k 1 i + 1 ξ k ξ ^ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ k ξ ^ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 T = Σ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T ,
Therefore, using (24) and (47), q i + 1 Σ k | k 1 is also an inverse Wishart distribution as
q i + 1 Σ k | k 1 = IW Σ k | k 1 ; λ ^ k i + 1 , Ψ ^ k i + 1 ,
where λ ^ k i + 1 , Ψ ^ k i + 1 are given as
λ ^ k i + 1 = λ k + 1 ,
Ψ ^ k i + 1 = Ψ k + E ξ k i + 1 ξ k B k i + 1 .

3.4.4. Fixed-Point Iteration of the Prior Estimate for Parameter τ k

With (34), (37), (44), and (49), (33) can be rewritten as
log q i + 1 τ k = E τ k i + 1 log p ξ k , γ k , Σ k | k 1 , τ k , z 1 : k + c τ k = τ k 2 E γ k i + 1 log γ k 1 2 E γ k i + 1 γ k τ k + τ k 2 log τ k 2 log ϒ τ k 2 + a k 1 log τ k b k τ k + c τ k .
With Stirling’s equation log ϒ τ k 2 τ k 1 2 log τ k 2 τ k 2 , (52) can be approximated as
log q i + 1 τ k = a k + 1 2 1 log τ k b k 1 2 1 2 E γ k i + 1 log γ k + 1 2 E γ k i + 1 γ k τ k + c τ k .
Therefore, using (16) and (53), q i + 1 τ k can be updated as a Gamma distribution, i.e.,
q i + 1 τ k = Γ τ k ; a ^ k i + 1 , b ^ k i + 1 ,
where a ^ k i + 1 , b ^ k i + 1 are given as
a ^ k i + 1 = a k + 1 2 ,
b ^ k i + 1 = b k 1 2 1 2 E γ k i + 1 log γ k + 1 2 E γ k i + 1 γ k .

3.5. The Variational Beayesian Iteration Based Robust Student’s t Invariant Kalman Filter

To propagate the approximate posterior distributions, using (44), (49) and (54), the expectations E γ k i + 1 log γ k , E γ k i + 1 γ k , E Σ k | k 1 i + 1 Σ k | k 1 1 , and E τ k i + 1 τ k can be calculated as follows
E γ k i + 1 γ k = η i + 1 / θ i + 1 ,
E γ k i + 1 log γ k = ψ η i + 1 log θ i + 1 ,
E Σ k | k 1 i + 1 Σ k | k 1 1 = λ ^ k i + 1 4 Ψ ^ k i + 1 1 ,
E τ k i + 1 τ k = a ^ k i + 1 / b ^ k i + 1 ,
where ψ represents a digamma function [42]. After N times of fixed-point iterations, the required posterior probability density function of ξ k could be approximated as
q ξ k N c ξ k ; ξ ^ k | k N , Σ k | k N = N c ξ k ; ξ ^ k | k , Σ k | k ,
where ξ ^ k | k , Σ k | k are respectively the posterior estimate for invariant error ξ k and the posterior estimate error covariance parameter. Then, resorting to the equivalence of probability distribution, the distributing of the propagated rotation state R k can be updated as the concentrated Gaussian distribution with mean being R ^ k | k and covariance being Σ k | k ,i.e.,
q R k = G R k ; R ^ k | k , Σ k | k ,
R ^ k | k = exp K k N z k H ξ ^ k | k 1 R ^ k | k 1 = exp K k N z k H ξ ^ k | k 1 R ^ k 1 | k 1 Ω k 1
where the iterative Kalman gain matrix K k N is calculated in (39). The Equations (62) and (63) are actually the projection of the Euclidean space Equation (38) into Lie group space.
Therefore, the filtering steps of the proposed approach include the steps (36)~(40), (42)~(46), (48)~(51), and (54)~(63). The detailed implementation of the proposed method for attitude estimation is presented in Algorithm 1. For the heavy-tailed attitude estimation, the nominal Σ ˜ k | k 1 could be selected as (11) to initialize the filter parameter Ψ k and λ k ; the invariant error is defined as the Lie exponential error of R ^ k 1 k 1 and R k so it could be assumed to be zero at the initialization to propagate the prior estimate ξ ^ k | k 1 = ξ ^ k 1 | k 1 .
In invariant Kalman filter, the prior error estimate R ^ k | k 1 is actually based on all the historical observations Y 1 : k 1 before time instant k. In this work, the covariance parameter Σ k | k 1 is also inferred using the historical observations Y 1 : k 1 until time instant k. As for the inverse Wishart distribution of Σ k | k 1 , at time instant k the degree of parameter is set as λ k = k + 4 and the inverse scale matrix parameter is set as Ψ k = k Σ ˜ k | k 1 .
In practical numerical integration of the attitude dynamic model, if there exists severe maneuvering, the system process noise will have a heavy-tailed distribution and the degree of freedom parameter τ k depends on the maneuvering intensity. According to our experience, the initial estimates for the shape parameter a k and rate parameter b k could be set as a k / b k 3 , 10 for the iteratively updating a ^ k , b ^ k in Algorithm 1.
Algorithm 1. The filtering steps of one time instant in the proposed approach for attitude estimation.
Inputs: R ^ k 1 | k 1 , Σ k 1 | k 1 , ξ ^ k 1 | k 1 , Ω k 1 , H , Y k , Σ w , Σ V , λ k , Ψ k , a k , b k , N
1. Predict the nominal invariant error ξ ^ k | k 1 and rotation R ^ k | k 1 according to (8) and (10)
    ξ ^ k | k 1 = ξ ^ k 1 | k 1 , R ^ k | k 1 = R ^ k 1 | k 1 Ω k 1
2. Predict the nominal prior error covariance Σ ˜ k | k 1 according to (11)
    Σ ˜ k | k 1 = Σ k 1 | k 1 + Σ w
3. Calculate the converted innovation according to (9)
    z k = R ^ k | k 1 Y k b b
4. Initialize the prior parameters
    λ k 0 = λ k , Ψ ^ k 0 = Ψ k = λ k 4 Σ ˜ k | k 1 , E γ k 0 γ k = 1
5. Calculate initial expectations according to (24) and (25)
    E Σ k | k 1 0 Σ k | k 1 1 = λ ^ k 0 4 Ψ ^ k 0 1 , E τ k 0 τ k = a k / b k
for i = 0:N − 1
6. Update q i + 1 ξ k = N c ξ k ; ξ ^ k | k i + 1 , Σ k | k i + 1 according to (36)~(40)
    Σ k | k 1 i = E Σ k | k 1 i Σ k | k 1 1 1 / E γ k i γ k ,     K k i + 1 = Σ k | k 1 i H T H Σ k | k 1 i H T + Σ V 1 ,
Σ k | k i + 1 = Σ k | k 1 i K k | i + 1 H Σ k | k 1 i , ξ ^ k | k i + 1 = ξ ^ k | k 1 + K k i + 1 z k H ξ ^ k | k 1
7. Update q i + 1 γ k = Γ γ k ; η i + 1 , θ i + 1 according to (44)~(46)
    η i + 1 = E τ k i τ k + 3 2 , θ i + 1 = 1 2 t r Σ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T E Σ k | k 1 i Σ k | k 1 1 + E τ k i τ k
8. Update q i + 1 Σ k | k 1 = IW Σ k | k 1 ; λ ^ k i + 1 , Ψ ^ k i + 1 according to (48)~(51)
    B k i + 1 = Σ k | k i + 1 + ξ ^ k | k i + 1 ξ ^ k | k 1 ξ ^ k | k i + 1 ξ ^ k | k 1 T , λ ^ k i + 1 = λ k + 1 , Ψ ^ k i + 1 = Ψ k + E ξ k i + 1 ξ k B k i + 1
9. Update expectations E γ k i + 1 log γ k , E γ k i + 1 γ k and E Σ k | k 1 i + 1 Σ k | k 1 1 according to (57)~(59)
    E γ k i + 1 γ k = η i + 1 / θ i + 1 ,     E γ k i + 1 log γ k = ψ η i + 1 log θ i + 1 ,
E Σ k | k 1 i + 1 Σ k | k 1 1 = λ ^ k i + 1 4 Ψ ^ k i + 1 1
10. Update q i + 1 τ k = Γ τ k ; a ^ k i + 1 , b ^ k i + 1 according to (55) and (56)
    a ^ k i + 1 = a k + 1 2 , b ^ k i + 1 = b k 1 2 1 2 E γ k i + 1 log γ k + 1 2 E γ k i + 1 γ k
11. Calculate the expectation E τ k i + 1 τ k according to (60)
    E τ k i + 1 τ k = a ^ k i + 1 / b ^ k i + 1
end for
12. Update the posterior rotation group and its covariance according to (63)
    R ^ k | k = exp K k N z k H ξ ^ k | k 1 R ^ k | k 1 , Σ k | k = Σ k | k N
Outputs: R ^ k | k , Σ k | k
The number of N iterations is also a crucial parameter for proposed approach and, generally, it should be set to a value larger than the system dimension to guarantee the local convergence of variational Bayesian iterations; nevertheless, a too larger value is sure to increase the computational cost for algorithm implementation. The balance between precision and cost should be considered according to the detailed applications.

4. Numerical Simulations

To further demonstrate the performance of the variational Bayesian iteration-based invariant Kaman filter, the attitude estimation system (6) (7) is simulated with parameter Σ 0 | 0 = 0.5236 2 I 3 × 3 , Σ   w = 0.01745 2 I 3 × 3 , b = [ 1 , 0 , 0 ] T , b = [ 0 , 1 , 0 ] T , Σ v = 0.0873 2 I 3 × 3 , Σ v = 0.0873 2 I 3 × 3 . The real attitude trajectories are generated for 5000 s and the heavy tailed process noise is generated with heavy tailed outliers according to [42], i.e.,
p   w k = N c   w k ; 0 3 × 1 , Σ w w . p .   1 α % N c   w k ; 0 3 × 1 , β Σ w w . p .   α % ,
which means that α percentage of process noise is drawn from a Gaussian distribution with covariance Σ   w while the covariance of α percentage of process noise is severely increased by scale parameter β due to the presence of outliers.
As a comparison, the filtering methods of the proposed approach (RSIKF), IKF [14,16] the Student’s t filter (STF) [43], the Huber-based Kalman filter (HKF) [44], and the QeIKF [24,25] are conducted together. For HKF, the tuning parameter is set to 1.345 and the number of iterations is selected as 10; in STF the degree of freedom parameter is selected as 3. In the proposed RSIKF, the filter parameters are set to λ k = k + 4 , a k = 2 , b k = 6 , N = 10 . All above filtering methods are implemented with MATLAB codes on a computer with Intel Core i7-7700 CPU at 3.60 GHz. To compare their performance, the error variable in Lie algebra of 5000 random runs is used to calculate the root mean square error R M S E k during the 5000 runs’ filtering time and average root mean square error A R M S E , i.e.,
  R M S E k 1 5000 l = 1 5000 ξ k , l ξ k , l * 2
A R M S E 1 5000 × 5000 k = 1 5000 l = 1 5000 ξ ^ k , l ξ k , l 2
where ξ k , l denotes the true estimate at the k-th time instant of the l-th simulation run and ξ ^ k , l is the corresponding estimate for the true ξ k , l ; denotes the Euclidean vector norm.
In this work, three cases of heavy tailed process noises are simulated with α = 5, 20 and β = 100, 1000 to investigate the filtering performance of different methods. The ARMSE result data of different filtering methods are given in Figure 7 while the R M S E k result data during the first 500 runs’ filtering processes are presented in Figure 8 and Figure 9.
The ARMSE result of IKF in Figure 7 and R M S E k data of IKF in Figure 8, Figure 9, Figure 10 and Figure 11 are always the largest in all methods for all cases, which certifies that the estimation performance of IKF is severely corrupted by heavy tailed outliers.
In STF, the noise distribution is assumed as a given Student’s t distribution, and the ARMSE result in Figure 7 and R M S E k in Figure 8, Figure 9, Figure 10 and Figure 11 is smaller than that of IKF. However, since it still requires accurate prior knowledge of the noise distribution parameter, the unknown or inaccurate parameters still mislead the estimate results.
The QeIKF aims to estimate, online, the unknown noise covariance parameter and also output better results than IKF and STF; however, it assumes the distribution as an Gaussian one so its ARMSE result in Figure 7 and R M S E k in Figure 8 for the case α = 5 become worse than the case of α = 20 in Figure 9; while, as shown in Figure 10 and Figure 11, its filtering precision would be significantly degraded by large outliers with β = 1000.
HKF shows some robustness to heavy tailed outliers IKF, STK, and QeKF for the cases of β = 1000 as in Figure 10 and Figure 11. Note that, the estimate given by HKF is rather conservative and its robustness is at the cost of inferior precision.
In all simulation cases of α and β , the ARMSE and R M S E k result of proposed RSIKF is always the least one in Figure 7, Figure 8, Figure 9, Figure 10 and Figure 11, demonstrating its superiority filtering performance to the other available methods. For RSIKF, different setting of N from 1~15 are conducted with all cases of α with the ARMSE result given in Figure 12; obviously, N = 10 already guarantees the iterative convergence and a larger N would significantly increase the computational cost.

5. Conclusions

For aerospace, satellite, and robotics engineering, the matrix Lie groups attitude estimation problem with heavy-tailed process noise is investigated. An improved robust Student’s t invariant Kalman filtering is proposed based on a hierarchical Gaussian state-space model. The probability density function of state prediction is defined based on student’s t distribution, while the conjugate prior distributions of the scale matrix and degrees of freedom (dofs) parameter are respectively formulated as the inverse Wishart and Gamma distribution. For attitude estimation state-space model, the Lie groups rotation matrix of spacecraft attitude is inferred together with the scale matrix and dof parameter using the variational Bayesian iteration. The new approach can improve the filtering robustness of the invariant Kalman filter for Lie groups spacecraft attitude estimation problems with heavy-tailed process uncertainty. Some further applications of the proposed methods can also be extended to other problems such as state smoother, parameter identification, state observers, and so on.

Author Contributions

Conceptualization, methodology, writing, software, validation, J.W. (Jiaolong Wang) and C.Z.; formal analysis, investigation, resources, J.W. (Jin Wu) and M.L. All authors have read and agreed to the published version of the manuscript.

Funding

This research is funded by the Fundamental Research Funds for Central Universities with number JUSRP121022 and National Natural Science Foundation of China with number 62003112.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data presented in this study are available on request from the corresponding author. The data are not publicly available as the data also forms part of an ongoing study.

Conflicts of Interest

The authors declare no conflict of interest. The funders had no role in the design of the study; in the collection, analyses, or interpretation of data; in the writing of the manuscript, or in the decision to publish the results.

References

  1. Wu, J.; Shan, S. Dot Product Equality Constrained Attitude Determination from Two Vector Observations: Theory and Astronautical Applications. Aerospace 2019, 6, 102. [Google Scholar] [CrossRef] [Green Version]
  2. Phisannupawong, T.; Kamsing, P.; Torteeka, P.; Channumsin, S.; Sawangwit, U.; Hematulin, W.; Jarawan, T.; Somjit, T.; Yooyen, S.; Delahaye, D.; et al. Vision-Based Spacecraft Pose Estimation via a Deep Convolutional Neural Network for Noncooperative Docking Operations. Aerospace 2020, 7, 126. [Google Scholar] [CrossRef]
  3. Soken, H.E.; Sakai, S.-I.; Asamura, K.; Nakamura, Y.; Takashima, T.; Shinohara, I. Filtering-Based Three-Axis Attitude Determination Package for Spinning Spacecraft: Preliminary Results with Arase. Aerospace 2020, 7, 97. [Google Scholar] [CrossRef]
  4. Carletta, S.; Teofilatto, P.; Farissi, M.S. A Magnetometer-Only Attitude Determination Strategy for Small Satellites: Design of the Algorithm and Hardware-in-the-Loop Testing. Aerospace 2020, 7, 3. [Google Scholar] [CrossRef] [Green Version]
  5. Louédec, M.; Jaulin, L. Interval Extended Kalman Filter-Application to Underwater Localization and Control. Algorithms 2021, 14, 142. [Google Scholar] [CrossRef]
  6. Pan, C.; Qian, N.; Li, Z.; Gao, J.; Liu, Z.; Shao, K. A Robust Adaptive Cubature Kalman Filter Based on SVD for Dual-Antenna GNSS/MIMU Tightly Coupled Integration. Remote Sens. 2021, 13, 1943. [Google Scholar] [CrossRef]
  7. Zheng, L.; Zhan, X.; Zhang, X. Nonlinear Complementary Filter for Attitude Estimation by Fusing Inertial Sensors and a Camera. Sensors 2020, 20, 6752. [Google Scholar] [CrossRef]
  8. Deibe, Á.; Antón Nacimiento, J.A.; Cardenal, J.; López Peña, F. A Kalman Filter for Nonlinear Attitude Estimation Using Time Variable Matrices and Quaternions. Sensors 2020, 20, 6731. [Google Scholar] [CrossRef]
  9. Guo, H.; Liu, H.; Hu, X.; Zhou, Y. A Global Interconnected Observer for Attitude and Gyro Bias Estimation with Vector Measurements. Sensors 2020, 20, 6514. [Google Scholar] [CrossRef] [PubMed]
  10. Ayala, V.; Román-Flores, H.; Torreblanca Todco, M.; Zapana, E. Observability and Symmetries of Linear Control Systems. Symmetry 2020, 12, 953. [Google Scholar] [CrossRef]
  11. Bonnabel, S.; Martin, P.; Salaun, E. Invariant Extended Kalman Filter: Theory and application to a velocity-aided attitude estimation problem. In Proceedings of the IEEE Conference on Decision & Control, Shanghai, China, 15–18 December 2009. [Google Scholar] [CrossRef] [Green Version]
  12. Vasconcelos, J.; Cunha, R.; Silvestre, C.; Oliveira, P. A nonlinear position and attitude observer on SE(3) using landmark measurements. Syst. Control Lett. 2010, 59, 155–166. [Google Scholar] [CrossRef]
  13. Chaturvedi, N.; Sanyal, A.; Mcclamroch, A. Rigid-body attitude control using rotation matrices for continuous singularity-free control laws. IEEE Control Syst. Mag. 2011, 31, 30–51. [Google Scholar]
  14. Barrau, A.; Bonnabel, S. Intrinsic filtering on Lie groups with applications to attitude estimation. IEEE Trans. Autom. Contr. 2014, 60, 436–449. [Google Scholar] [CrossRef]
  15. Barrau, A.; Bonnabel, S. The invariant extended Kalman filter as a stable observer. IEEE Trans. Autom. Contr. 2017, 62, 1797–1812. [Google Scholar] [CrossRef] [Green Version]
  16. Barrau, A.; Bonnabel, S. Invariant Kalman filtering. Annu. Rev. Control Robot. Auton. Syst. 2018, 1, 237–257. [Google Scholar] [CrossRef]
  17. Batista, P.; Silvestre, C.; Oliveira, P. A GES attitude observer with single vector observations. Automatica 2012, 49, 388–395. [Google Scholar] [CrossRef]
  18. Chirikjian, G.; Kobilarov, M. Gaussian approximation of non-linear measurement models on lie groups. In Proceedings of the IEEE Conference on Decision and Control, Los Angeles, CA, USA, 15–17 December 2014. [Google Scholar]
  19. Barfoot, T.; Furgale, P. Associating uncertainty with three- dimensional poses for use in estimation problems. IEEE Trans. Robot. 2014, 30, 679–693. [Google Scholar] [CrossRef]
  20. Said, S.; Manton, J. Extrinsic mean of Brownian distributions on compact lie groups. IEEE Trans. Inf. Theory 2012, 58, 3521–3535. [Google Scholar] [CrossRef] [Green Version]
  21. Huang, Y.; Zhang, Y. A New Process Uncertainty Robust Student’s t Based Kalman Filter for SINS/GPS Integration. IEEE Access 2017, 5, 14391–14404. [Google Scholar] [CrossRef]
  22. Karasalo, M.; Hu, X. An optimization approach to adaptive Kalman filtering. Automatica 2011, 47, 1785–1793. [Google Scholar] [CrossRef] [Green Version]
  23. Wang, J.; Wang, J.; Zhang, D.; Shao, X.; Chen, G. Kalman filtering through the feedback adaption of prior error covariance. Signal. Process. 2018, 152, 47–53. [Google Scholar] [CrossRef]
  24. Feng, B.; Fu, M.; Ma, H.; Xia, Y.; Wang, B. Kalman filter with recursive covariance Estimation--sequentially estimating process noise covariance. IEEE Trans. Ind. Electron. 2014, 61, 6253–6263. [Google Scholar] [CrossRef]
  25. Zanni, L.; Le Boudec, J.; Cherkaoui, R.; Paolone, M. A prediction-error covariance estimator for adaptive Kalman filtering in step-varying processes: Application to power-system state estimation. IEEE Trans. Contr. Syst. Technol. 2017, 25, 1683–1697. [Google Scholar] [CrossRef] [Green Version]
  26. Mohamed, A.; Schwarz, K. Adaptive Kalman Filtering for INS/GPS. J. Geod. 1999, 73, 193–203. [Google Scholar] [CrossRef]
  27. Ardeshiri, T.; Özkan, E.; Orguner, U.; Gustafsson, F. Approximate Bayesian smoothing with unknown process and measurement noise covariance. IEEE Signal Process. Lett. 2015, 22, 2450–2454. [Google Scholar] [CrossRef] [Green Version]
  28. Assa, A.; Plataniotinos, K. Adptive Kalman filtering by covariance sampling. IEEE Signal Process. Lett. 2017, 24, 1288–1292. [Google Scholar] [CrossRef]
  29. Huang, Y.; Zhang, Y.; Wu, Z.; Li, N.; Chambers, J. A novel adaptive Kalman filter with inaccurate process and measurement noise covariance matrices. IEEE Trans. Autom. Contr. 2018, 63, 594–601. [Google Scholar] [CrossRef] [Green Version]
  30. Tronarp, F.; Karvonen, T.; Särkkä, S. Student′s t-Filters for Noise Scale Estimation. IEEE Signal Process. Lett. 2019, 26, 352–356. [Google Scholar] [CrossRef]
  31. Dong, P.; Jing, Z.; Leung, H.; Shen, K.; Wang, J. Student-t mixture labeled multi-Bernoulli filter for multi-target tracking with heavy-tailed noise. Signal Process. 2018, 152, 331–339. [Google Scholar] [CrossRef]
  32. Huang, Y.; Zhang, Y.; Li, N.; Wu, Z.; Chambers, J. A novel robust student’s t-based Kalman filter. IEEE Trans. Aerosp. Electron. Syst. 2017, 53, 1545–1554. [Google Scholar] [CrossRef] [Green Version]
  33. Huang, Y.; Zhang, Y.; Li, N.; Chambers, J. Robust student’s t based nonlinear filter and smoother. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 2586–2596. [Google Scholar] [CrossRef] [Green Version]
  34. Ćesić, J.; Markovi, I.; Petrovi, I. Mixture Reduction on Matrix Lie Groups. IEEE Signal Process. Lett. 2017, 24, 1719–1723. [Google Scholar] [CrossRef] [Green Version]
  35. Ćesić, J.; Markovi, I.; Bukal, M.; Petrović, I. Extended Information Filter on Matrix Lie Groups. Automatica 2017, 82, 226–234. [Google Scholar] [CrossRef]
  36. Kang, D.; Jang, C.; Park, F. Unscented Kalman Filtering for Simultaneous Estimation of Attitude and Gyroscope Bias. IEEE/ASME Trans. Mechatron. 2019, 24, 350–360. [Google Scholar] [CrossRef]
  37. Bourmaud, G.; Mégret, R.; Arnaudon, M.; Giremus, A. Continuous-Discrete Extended Kalman Filter on Matrix Lie Groups Using Concentrated Gaussian Distributions. J. Math. Imaging Vis. 2015, 51, 209–228. [Google Scholar] [CrossRef] [Green Version]
  38. Tzikas, D.; Likas, A.; Galatsanos, N. The Variational Approximation for Bayesian Inference. IEEE Signal Process. Mag. 2008, 25, 131–146. [Google Scholar] [CrossRef]
  39. Pavliotis, G.A. Applied Stochastic Processe; Springer: London, UK, 2009. [Google Scholar]
  40. Hagan, T.; Forster, J.J. Kendall’s Advanced Theory of Statistics; Arnold: London, UK, 2004. [Google Scholar]
  41. Bishop, C.M. Pattern Recognition and Machine Learning; Springer: Berlin, Germany, 2007. [Google Scholar]
  42. Huang, Y.; Zhang, Y.; Li, N.; Chambers, J. A Robust Gaussian Approximate Fixed-Interval Smoother for Nonlinear Systems With Heavy-Tailed Process and Measurement Noises. IEEE Signal Process. Lett. 2016, 23, 468–472. [Google Scholar] [CrossRef] [Green Version]
  43. Roth, M.; Ozkan, E.; Gustafsson, F. A Student’s t filter for heavy tailed process and measurement noise. In Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; pp. 5770–5774. [Google Scholar]
  44. Karlgaard, C.D.; Schaub, H. Huber-Based Divided Difference Filtering. J. Guid. Control Dyn. 2012, 30, 885–891. [Google Scholar] [CrossRef]
Figure 1. The vector observations-based attitude determination in astronautical application (permission granted from [1]).
Figure 1. The vector observations-based attitude determination in astronautical application (permission granted from [1]).
Machines 09 00182 g001
Figure 2. The mappings between the probability definition for Lie group rotation matrix R S O 3 and the classical Gaussian distribution vector w 3 .
Figure 2. The mappings between the probability definition for Lie group rotation matrix R S O 3 and the classical Gaussian distribution vector w 3 .
Machines 09 00182 g002
Figure 3. Projecting the Lie groups model into vector space based on the invariant error.
Figure 3. Projecting the Lie groups model into vector space based on the invariant error.
Machines 09 00182 g003
Figure 4. The filtering diagram of the invariant Kalman filter for attitude estimation.
Figure 4. The filtering diagram of the invariant Kalman filter for attitude estimation.
Machines 09 00182 g004
Figure 5. The main idea and key innovation of this work to deal with the heavy-tailed attitude estimation problem on matrix Lie group SO(3).
Figure 5. The main idea and key innovation of this work to deal with the heavy-tailed attitude estimation problem on matrix Lie group SO(3).
Machines 09 00182 g005
Figure 6. Probability view of the proposed filtering method for Student’s t based hierarchical Gaussian state-space model.
Figure 6. Probability view of the proposed filtering method for Student’s t based hierarchical Gaussian state-space model.
Machines 09 00182 g006
Figure 7. The ARMSE result of RSIKF with different N for all cases of α and β .
Figure 7. The ARMSE result of RSIKF with different N for all cases of α and β .
Machines 09 00182 g007
Figure 8. The R M S E k result of different methods for the case of α = 5 and β = 10.
Figure 8. The R M S E k result of different methods for the case of α = 5 and β = 10.
Machines 09 00182 g008
Figure 9. The R M S E k result of different methods for the case of α = 20 and β = 10.
Figure 9. The R M S E k result of different methods for the case of α = 20 and β = 10.
Machines 09 00182 g009
Figure 10. The R M S E k result of different methods for the case of α = 5 and β = 100.
Figure 10. The R M S E k result of different methods for the case of α = 5 and β = 100.
Machines 09 00182 g010
Figure 11. The R M S E k result of different methods for the case of α = 5 and β = 100.
Figure 11. The R M S E k result of different methods for the case of α = 5 and β = 100.
Machines 09 00182 g011
Figure 12. The ARMSE result of RSIKF with different N for the cases of α and β .
Figure 12. The ARMSE result of RSIKF with different N for the cases of α and β .
Machines 09 00182 g012
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Wang, J.; Zhang, C.; Wu, J.; Liu, M. An Improved Invariant Kalman Filter for Lie Groups Attitude Dynamics with Heavy-Tailed Process Noise. Machines 2021, 9, 182. https://doi.org/10.3390/machines9090182

AMA Style

Wang J, Zhang C, Wu J, Liu M. An Improved Invariant Kalman Filter for Lie Groups Attitude Dynamics with Heavy-Tailed Process Noise. Machines. 2021; 9(9):182. https://doi.org/10.3390/machines9090182

Chicago/Turabian Style

Wang, Jiaolong, Chengxi Zhang, Jin Wu, and Ming Liu. 2021. "An Improved Invariant Kalman Filter for Lie Groups Attitude Dynamics with Heavy-Tailed Process Noise" Machines 9, no. 9: 182. https://doi.org/10.3390/machines9090182

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop