Next Article in Journal
Four Constructions of Asymptotically Optimal Codebooks via Additive Characters and Multiplicative Characters
Previous Article in Journal
Q-rung Orthopair Normal Fuzzy Aggregation Operators and Their Application in Multi-Attribute Decision-Making
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Variational Bayesian Iterative Estimation Algorithm for Linear Difference Equation Systems

1
Key Laboratory of Advanced Process Control for Light Industry (Ministry of Education), School of Internet of Things Engineering, Jiangnan University, Wuxi 214122, China
2
Department of Chemical and Materials Engineering, University of Alberta, Edmonton, AB T6G 2G6, Canada
*
Author to whom correspondence should be addressed.
Mathematics 2019, 7(12), 1143; https://doi.org/10.3390/math7121143
Submission received: 30 October 2019 / Revised: 19 November 2019 / Accepted: 20 November 2019 / Published: 22 November 2019
(This article belongs to the Section Engineering Mathematics)

Abstract

:
Many basic laws of physics or chemistry can be written in the form of differential equations. With the development of digital signals and computer technology, the research on discrete models has received more and more attention. The estimates of the unknown coefficients in the discretized difference equation can be obtained by optimizing certain criterion functions. In modern control theory, the state-space model transforms high-order differential equations into first-order differential equations by introducing intermediate state variables. In this paper, the parameter estimation problem for linear difference equation systems with uncertain noise is developed. By transforming system equations into state-space models and on the basis of the considered priors of the noise and parameters, a variational Bayesian iterative estimation algorithm is derived from the observation data to obtain the parameter estimates. The unknown states involved in the variational Bayesian algorithm are updated by the Kalman filter. A numerical simulation example is given to validate the effectiveness of the proposed algorithm.

1. Introduction

Differential equations arise in many instances when using the mathematical models to describe phenomena in economics, engineering, biology, etc. [1,2]. Considering the following nth-order linear differential equation with zero initial conditions
y k ( n ) + α 1 y k ( n 1 ) + α 2 y k ( n 2 ) + + α n 1 y k ( 1 ) + α n y k = β 1 u k ( n 1 ) + β 2 u k ( n 2 ) + + β n 1 u ( 1 ) + β n u k ,
where y k is the system output, u k is the system input, k is the time variable, α i and β i are the unknown real parameters. Each derivative symbol is defined as
y k ( i ) = d i y k d i k ,        y k ( 2 ) = d 2 y k d 2 k = y ¨ k ,        y k ( 1 ) = y k k = y ˙ k ,
u k ( i ) = d i u k d i k ,        u k ( 2 ) = d 2 u k d 2 k = u ¨ k ,        u k ( 1 ) = u k k = u ˙ k .
Although the laws of the systems are continuously evolved over time, the sampled values of the systems are always discretized. In order to obtain the estimates of the unknown parameters in the system mathematical models, there are usually two approaches to further process the differential equation models. One is called the frequency-domain analysis methods, and the other is called the time-domain analysis methods [3,4,5,6].
For linear continuous systems, Laplace transform of Equation (1) under zero initial conditions is given by
( s n 1 + α 1 s n 1 + α 2 s n 2 + + α n 1 s + α n ) Y ( s ) = ( β 1 s n 1 + β 2 s n 2 + + β n 1 s + β n ) U ( s ) ,
where the variable s is the laplace operator, Y ( s ) and U ( s ) are the Laplace transform of y k and u k : Y ( s ) : = L [ y k ] , U ( s ) : = L [ u k ] . Then the transfer function of the system is written as
G ( s ) : = Y ( s ) U ( s ) = β 1 s n 1 + β 2 s n 2 + + β n 1 s + β n s n 1 + α 1 s n 1 + α 2 s n 2 + + α n 1 s + α n .
So the output equation can be further expressed as
Y ( s ) = G ( s ) U ( s ) .
The system will be excited by designing different input signals to obtain different output responses. The parameters of the system are further obtained according to the analysis of the output responses [7,8,9]. By using the Laplace transform, the transfer function is one of the most important mathematical models in control theory. Under the zero initial conditions, the Laplace transform of the system output divided by the Laplace transform of the system input is called the transfer function. However, the transfer function is only suitable for describing linear constant coefficient differential equation systems.
In the time-domain analysis method, the system is represented in discrete time and the sampling interval is assumed to be one time unit. In this case, the discrete-time operator equations or difference equations are necessary [10,11,12,13]. A typical controlled autoregressive difference equation is written as
y k + a 1 y k 1 + a 2 y k 2 + + a n y k n = b 1 u k 1 + b 2 u k 2 + + b n u k n .
Then the output y k can be further written as a form of determining the next value for given previous observations:
y k = a 1 y k 1 a 2 y k 2 a n y k n + b 1 u k 1 + b 2 u k 2 + + b n u k n = ψ k T ϑ ,
where
ψ k : = [ y k 1 , y k 2 , , y k n , u k 1 , u k 2 , , u k n ] T R 2 n ,
ϑ : = [ a 1 , a 2 , , a n , b 1 , b 2 , , b n ] T 2 n .
From Equation (2), it is easy to find that the calculation of y k from past data depends on the parameters in ϑ . Call the calculated value y ^ k | ϑ and write
y ^ k | ϑ = ψ k T ϑ .
The solution of obtaining the estimates of the parameters is to optimize a criterion function. The effective estimates make the prediction error ε ( k , ϑ ) = y k y ^ k | ϑ as small as possible. Given a batch of inputs and outputs over a time interval 1 k N :
Z N = [ u 1 , y 1 , u 2 , y 2 , , u N , y N ] .
The quadratic criterion function is defined as [14]:
V N ( ϑ , Z N ) = 1 N k = 1 N [ y k y ^ k | ϑ ] 2 .
The process of computing the estimate ϑ ^ N is the process of finding the solution of the following optimization problem:
ϑ ^ N = arg min ϑ V N ( ϑ , Z N ) ,
where ‘ arg min ϑ ’ represents the value of the considered parameter vector when the objective function takes the minimum value. The classical least squares-based and gradient descent-based numerical methods have been applied to search the estimates of the parameters by minimizing the criterion function [15,16,17,18]. For example, Xu et al. developed a least squares-based iterative parameter estimation algorithm for dynamic systems based on the parameter decomposition by using the hierarchical identification principle [19].
However, the gradient and least squares-based iterative algorithm only obtain the point estimates of parameters based on the known noise variance. They minimize a cost function on the parameter vector and then obtain the estimate of the parameter vector. Different from the stochastic gradient and least squares algorithm, the variational Bayesian (VB) algorithm needs the priors of parameters and then iteratively updates the posterior expression of the parameters by maximizing the lower bound marginal likelihood function [20,21]. The VB algorithm can obtain the distributions of the unknown parameters rather than the point estimates. For example, Yang and Yin applied the VB inference to estimate the parameters of finite impulse response (FIR) systems with randomly missing output data [22].
In this paper, the difference equation is further transformed to the state-space model. In modern control theory, the state-space model is widely used to describe the dynamics of systems, because it can not only reflect the internal states of systems but also reveals the relationship between the intermediate states and the external input and output variables [23,24,25,26,27,28]. A general linear state-space model
x 1 , k + 1 x 2 , k + 1 x n , k + 1 = a 11 a 12 a 1 n a 21 a 22 a 2 n a n 1 a n 2 a n n x 1 , k x 2 , k x n , k + b 1 b 2 b n u k ,
y k = [ c 1 , c 2 , , c n ] x 1 , k x 2 , k x n , k + v k ,
where y k is the system output, u k is the system input, v k is the random noise, and x i , k is the unmeasurable internal state, which establishes the link between system internal information and external measurements is presented. However, state estimation brings difficulty to the parameter estimation [29]. For linear Gaussian systems, the Kalman filter can achieve optimal estimation performance with a small amount of computational load, which is usually the preferred method of the state estimation for state-space models [30,31,32].
Based on the previous iterative identification schemes [33,34,35,36], this paper concerns a VB-based iterative estimation algorithm for difference equation models with unknown measurement noise. Different from the work studied in [37], which used an open-loop state observer to roughly calculate the system states without considering the effect of measurement noise in the procedure of states estimation, this paper considers the problem of state estimation by using the Kalman filter. The main contributions of this paper are as follows.
  • By defining internal state variables, the difference equation model is rewritten in the form of the state-space model.
  • A Kalman filter based VB (VB-KF) estimation algorithm is provided to estimate both the model states and parameters interactively. Based on the estimated states, the VB algorithm calculates the system parameters and the noise variance. Then the states are updated by means of the Kalman filter with new obtained parameters and noise variance.
  • Different from the existing point parameter estimation methods, the proposed algorithm can give the distributions of the system parameters and noise variance to reflect the uncertainties of their estimates.
The remainder of this paper is organized as follows. Section 2 introduces the state-space model. Section 3 derives a VB iterative algorithm to estimate the unknown parameters of the systems by using the Kalman filter. The simulation is presented in Section 4 to verify the proposed method. Section 5 gives some conclusions.

2. Transformation of the Difference Equation

Reconsidering a general nth-order difference equation in Equation (2), in order to transform the difference equation to the state-space model, n state variables are needed. The choices of the state variables are not unique; they can be measurable or unmeasurable. Usually, the state variables which can help to convert the state equation into a standard form are selected as the intermediate states. Here, we introduce n state variables as follows:
x 1 , k = y k , x 2 , k = y k + 1 b 1 u k , x 3 , k = y k + 2 b 1 u k + 1 b 2 u k , x 4 , k = y k + 3 b 1 u k + 2 b 2 u k + 1 b 3 u k , x n 1 , k = y k + n 2 b 1 u k + n 3 b 2 u k + n 4 b n 2 u k , x n , k = a n y k 1 a n 1 ( y k b 1 u k 1 ) a n 2 ( y k + 1 b 1 u k b 2 u k 1 ) + a 1 ( y k + n 2 b 1 u k + n 3 b 2 u k + n 4 b n 2 u k b n 1 u k 1 ) + b n u k .
Further, replacing the time variable k with k + 1 on both sides of Equation (3) gives the following equations:
x 1 , k + 1 = y k + 1 = x 2 , k + b 1 u k , x 2 , k + 1 = y k + 2 b 1 u k + 1 = x 3 , k + b 2 u k , x j , k + 1 = y k + j b 1 u k + j 1 b j 1 u k + 1 = x j + 1 , k + b j u k , x n 1 , k + 1 = x n , k + b n 1 u k , x n , k + 1 = a n x 1 , k a n 1 x 2 , k a 1 x n , k + b n u k .
Define the state vector x k : = [ x 1 , k , x 2 , k , , x n , k ] T n . Rewrite the state equation for each state variable in Equation (4) by matrix representation; the observability canonical form state-space model of the difference Equation (2) can be described as
x 1 , k + 1 x 2 , k + 1 x n 1 , k + 1 x n , k + 1 = 0 1 0 0 0 0 1 0 0 0 0 1 a n a n 1 a n 2 a 1 x 1 , k x 2 , k x n 1 , k x n , k + b 1 b 2 b n 1 b n u k ,
y k = [ 1    0    0       0 ] x 1 , k x 2 , k x n 1 , k x n , k .
Further, we rewrite (5) and (6) as
x k + 1 = A x k + B u k ,
y k = C x k ,
where
A = 0 1 0 0 0 0 1 0 0 0 0 1 a n a n 1 a n 2 a 1 ,        B = b 1 b 2 b n 1 b n ,
C = [ 1 ,   0 ,   0 ,   ,   0 ] .
The matrix A R n × n is called the system matrix, the vector B R n × 1 is the system input vector, and the vector C R 1 × n is the system output vector. Equation (7) is called the state equation, and Equation (8) is called the output equation. Specially, for the multi-input multi-output systems, the input and output vectors will be extended to matrices. It is easy to find that those unknown parameters are in the system matrix and input vector. In order to estimate the unknown parameters of the system, the output equation needs to be redescribed to include all unknown parameters. Transforming the state equation, the output equation in Equation (8) can be further written as
y k = x 1 , k = x 2 , k 1 + b 1 u k 1 = x 3 , k 2 + b 2 u k 2 + b 1 u k 1 = x 4 , k 3 + b 3 u k 3 + b 2 u k 2 + b 1 u k 1 = a n x 1 , k n a n 1 x 2 , k n a 1 x n , k n + b n u k n + + b 2 u k 2 + b 1 u k 1 .
Define the parameter vector θ , the information vector ψ k as
θ : = [ a n , a n 1 , , a 1 , b n , b n 1 , , b 1 ] T R 2 n , ψ k : = [ x 1 , k n , x 2 , k n , , x n , k n , u k n , u k n + 1 , , u k 1 ] T R 2 n .
Considering that there will be some disturbance in the actual systems, introduce a noise term v k to the output equation. Then the output in Equation (9) can be simplified into a vector form
y k = ψ k T θ + v k ,
where v k is the random disturbance, which can be the white noise or colored noise. For practical systems, the effects of noise should be considered when dealing with the problem of parameter estimation.
Remark 1.
By introducing internal state variables, higher-order difference equations are transformed into first-order difference equations, which simplify the forms of the output equations of the models.
This paper develops a combined parameters, intermediate states, and noise variance estimation algorithm for linear difference equation systems. On the basis of the model shown in Equation (10), the next section will drive a VB-KF iterative algorithm to get the solution for the problem of the parameter estimation.

3. The Kalman Filter-Based Variational Bayesian Iterative Estimation Algorithm

3.1. The Variational Bayesian Parameter Estimation Algorithm

The VB algorithm is formed by introducing the variational approximation theory. It is difficult to deal with the integrations involved in the general posterior probability solution. The VB algorithm introduces a cluster of free factored distribution to approximate the complicated posterior probability, which avoids the direct multiple integrations on marginal distribution [38,39]. In this paper, the parameter vector together with the corresponding estimated error covariance matrix and noise variance are inferred by using the VB approach.
The unknown parameters that need to be estimated include the system model parameter vector θ , the variance of parameter estimates ζ 1 I , and the system noise variance σ 1 . Assuming their priors follow Gaussian, Gamma, and Gamma distributions, respectively,
p ( θ ) = N ( 0 , ζ 1 I ) , p ( ζ ) = Gamma ( ζ ; d 0 , e 0 ) , p ( σ ) = Gamma ( σ ; f 0 , h 0 ) .
The VB algorithm introduces a free distribution q ( · ) as the approximate distribution of parameter to simplify the calculation of logarithmic likelihood function of the measurement data. Then the probability distribution of observation can be expressed as
log   p ( y 1 : k ) = log   p ( θ , ζ , σ , y 1 : k ) log   p ( θ , ζ , σ | y 1 : k ) = q ( θ , ζ , σ ) × log p ( θ , ζ , σ , y 1 : k ) q ( θ , ζ , σ ) d θ d ζ d σ + q ( θ , ζ , σ ) × log q ( θ , ζ , σ ) p ( θ , ζ , σ | y 1 : k ) d θ d ζ d σ = F q ( θ , ζ , σ ) + KL q ( θ , ζ , σ ) | | p ( θ , ζ , σ , y 1 : k ) ,
where F q ( θ , ζ , σ ) is the lower bound of log   p ( y 1 : k ) (referring to the Jensen inequality [38]), that is
log   p ( y 1 : k ) q ( θ , ζ , σ ) × log p ( θ , ζ , σ , y 1 : k ) q ( θ , ζ , σ ) d θ d ζ d σ .
KL ( q ( θ , ζ , σ ) | | p ( θ , ζ , σ | y 1 : k ) is the Kullback–Leibler divergence between the factored approximation q ( θ , ζ , σ ) and the true posterior distribution p ( θ , ζ , σ | y 1 : k ) , which demonstrates the similarity between the two probability distributions.
We assume that the joint probability distribution is decomposable:
q ( θ , ζ , σ ) = q ( θ ) q ( ζ ) q ( σ ) .
We can find that when q ( θ ) q ( ζ ) q ( σ ) and p ( θ , ζ , σ | y 1 : k ) are identically distributed, the KL divergence between them can reach minimum. Therefore, these independent distributions q ( θ ) , q ( ζ ) , q ( σ ) can be solved by minimizing
KL q ( θ ) q ( ζ ) q ( σ ) | | p ( θ , ζ , σ | y 1 : k ) = q ( θ , ζ , σ ) × log p ( θ , ζ , σ | y 1 : k ) q ( θ ) q ( ζ ) q ( σ ) d θ d ζ d σ .
Using the theorems from the calculus of variation [31], the optimal solutions for the parameter distributions satisfy the following equations
log q ( θ ) =   log p ( θ , ζ , σ , y 1 : k ) q ( ζ ) q ( σ ) d ζ d σ + c θ ,
log q ( ζ ) =   log p ( θ , ζ , σ , y 1 : k ) q ( θ ) q ( σ ) d θ d σ + c ζ ,
log q ( σ ) =   log p ( θ , ζ , σ , y 1 : k ) q ( θ ) q ( ζ ) d θ d ζ + c σ ,
where c θ , c ζ , c σ denote the terms which are independent of θ , ζ , σ , respectively.
The initial joint probability p ( θ , ζ , σ , y 1 : k ) can be factored as
p ( θ , ζ , σ , y 1 : k ) = N ( 0 , ζ 1 I ) Gamma ( ζ ; c 0 , d 0 ) Gamma ( σ ; e 0 , f 0 ) k = 1 N N ( y k ; ψ k T θ , σ 1 ) .
Using Equation (16) into Equation (13) and ignoring items which are unrelated to θ gives
log q ( s + 1 ) ( θ ) =   q ( ζ ) q ( σ ) 1 2 θ T ζ θ 1 2 k = 1 N ( y k ψ k T θ ) T σ ( y k ψ k T θ ) d ζ d σ + c θ = 1 2 θ T ζ ( s ) I + k = 1 N ψ k σ ( s ) ψ k T θ + k = 1 N y k 2 k = 1 N θ T ψ k σ ( s ) y k k = 1 N y k σ ( s ) ψ k T θ + c θ ,
where q ( s + 1 ) ( · ) denotes the approximation of q ( · ) at the ( s + 1 ) th iteration. From Equation (17) we can find that q ( s + 1 ) ( θ ) is the probability density function of a Gaussian distribution, and the system parameter θ can be updated as
θ ¯ ( s + 1 ) = var θ ( s + 1 ) k = 1 N ψ k σ ( s ) y k ,
var θ ( s + 1 ) = ζ ( s ) I + k = 1 N ψ k σ ( s ) ψ k T 1 .
Furthermore, we can calculate the mathematical expectation of θ θ T
E ( s + 1 ) ( θ θ T ) = var ( ( s + 1 ) ) + θ ( s + 1 ) ( θ ( s + 1 ) ) T .
According to Equation (14), we have
log q ( s + 1 ) ( ζ ) =   q ( θ ) q ( σ ) 1 2 log ζ 1 2 ( θ ) ζ θ + ( d 0 1 ) log ζ e 0 ζ d θ d σ + c ζ = 1 2 E ( s + 1 ) ( θ T θ ) + e 0 ζ + ( 1 2 + d 0 1 ) log ζ + c ζ = 1 2 tr var ( ( s + 1 ) ) + θ ( s + 1 ) ( θ ( s + 1 ) ) T + e 0 ζ + ( 1 2 + d 0 1 ) log ζ + c ζ ,
which yields
q ( s + 1 ) ( ζ ) = Gamma ( ζ ; d ( s + 1 ) , e ( s + 1 ) ) ,
where d ( s + 1 ) and e ( s + 1 ) are the two parameters of the Gamma distribution and are given by
d ( s + 1 ) = 1 2 + d 0 , e ( s + 1 ) = e 0 + 1 2 tr var ( θ ( s + 1 ) ) + θ ( s + 1 ) ( θ ( s + 1 ) ) T .
Then, update the expectation of ζ at iteration s + 1 by
ζ ¯ ( s + 1 ) = d ( s + 1 ) e ( s + 1 ) .
Similarly, the variational form for q ( s + 1 ) ( σ ) obeys
log q ( s + 1 ) ( σ ) =   q ( θ ) q ( ζ ) ( f 0 1 ) log σ h 0 σ + N 2 log σ 1 2 k = 1 N ( y k ψ k T θ ) T σ ( y k ψ k T θ ) d θ d ζ + c σ = ( f 0 + N 2 1 ) log σ 1 2 h 0 + k = 1 N ( y k 2 2 y k ψ k T θ ( s + 1 ) + ψ k T E ( s + 1 ) ( θ θ T ) ψ k ) σ + c σ ,
which yields
q ( s + 1 ) ( σ ) = Gamma ( σ ; f ( s + 1 ) , h ( s + 1 ) ) ,
where f ( s + 1 ) and h ( s + 1 ) are given by
f ( s + 1 ) = N 2 + f 0 , h ( s + 1 ) = h 0 + k = 1 N y k 2 2 y k ψ k T θ ( s + 1 ) + ψ k T tr [ var ( ( s + 1 ) ) + θ ( s + 1 ) ( θ ( s + 1 ) ) T ψ k } .
Then σ ¯ can be computed as
σ ¯ ( s + 1 ) = f ( s + 1 ) h ( s + 1 ) .

3.2. The State Estimation Algorithm

Above all, we obtain the posterior probability distribution or optimal solution of system parameters and the noise variance. However, the state-space model contains both unknown system parameters and unavailable system states, which brings a difficulty to obtaining the parameter estimates. The Kalman filter is usually used to estimate the intermediate states for linear state-space systems with Gaussian noise. The state observer is a deterministic state estimation algorithm, which is simpler in structure but only obtains an approximate estimation of the system states. To improve the estimation accuracy, this paper adopts a modified Kalman filter to obtain state estimates under the Bayesian framework. According to the considered state-space model, the following Kalman filter algorithm [40] is designed to generate the state estimates
x ^ k + 1 = A ^ x ^ k + b ^ u ¯ k + L k ( y k c x ^ k ) ,      x ^ 1 = 1 n / p 0 ,
L k = A ^ k P k c T ( σ ^ k + c P k c T ) 1 ,
P k + 1 = A ^ P k A ^ T L k c P k A ^ T ,      P 1 = I n .
Since the system parameters are unknown, we have to utilize the estimated parameters to construct the system parameter vector and matrix
A ^ : = 0 1 0 0 0 0 1 0 0 0 0 1 a ^ n a ^ n 1 a ^ n 2 a ^ 1 R n × n ,      b ^ : = b ^ 1 , k b ^ 2 , k b ^ n 1 , k b ^ n , k R n .
The flowchart of executing the proposed VB-KF iterative algorithm is shown in Figure 1.
The iterative steps for calculating parameter and state estimates are summarized as follows:
(1)
Initialization: set s = 1 , x k ( 0 ) = 1 n , θ ¯ ( 0 ) = 1 n 2 , and give a group of random positive values to parameters d 0 , e 0 , f 0 , h 0 and a small positive value to ε .
(2)
Collect input-output data ( u k , y k : k = 1 , 2 , , N ) , and contract information vector ψ k with Equation (10).
(3)
Update θ ¯ ( s + 1 ) , var [ θ ( s + 1 ) ] , E ( s + 1 ) ( θ θ T ) , and ζ ¯ ( s + 1 ) with Equations (18)–(21), respectively.
(4)
Update noise variance σ ¯ ( s + 1 ) with Equation (22).
(5)
Use the new obtained parameter estimates and noise variance to update states with Kalman filter algorithm in Equations (23)–(25).
(6)
Evaluate the relative change between θ ¯ ( s + 1 ) and θ ¯ ( s ) , if | | θ ¯ ( s + 1 ) θ ¯ ( s ) | | ε ( | | · | | means L2-norm), for a very small positive number ε , then stop the iteration procedure and get the parameter estimates θ ¯ ( s + 1 ) ; or else, let s = s + 1 and go to Step 3.
The proposed methods can combine other estimation algorithms [41,42,43,44,45] and the mathematical tools [46,47,48,49] to study the parameter identification problems of other linear and nonlinear systems with colored noise [50,51,52] and can be applied to other literatures [53,54,55,56].

4. Simulation Example

Consider the following 3rd-order difference equation:
y k + 0 . 50 y k 1 + 0 . 32 y k 2 + 0 . 18 y k 3 = 1 . 00 u k 1 + 0 . 55 u k 2 + 0 . 97 u k 3 + v k .
Taking into account the interference of noise, transform the difference equation into the state-space model:
x 1 , k + 1 x 2 , k + 1 x 3 , k + 1 = 0 1 0 0 0 1 0.18 0.32 0.50 x 1 , k x 2 , k x 3 , k + 1.00 0.55 0.97 u k , y k = [ 1    0    0 ] x 1 , k x 2 , k x 3 , k + v k .
The parameter vector to be estimated is
θ = 0.18 ,   0.32 ,   0.50 ,   1.00   , 0.55 ,   0.97 .
In this simulation example, the input is selected as a persistent excitation signal sequence with zero mean and unit variance, v k is Gaussian noise sequence with zero mean and variance σ 2 . Of the collected data length N = 2000 , the first 1000 sets of data ( N 1 ) are used for the training set and the rest of the ( N 2 ) sets are used for the testing set. Design different noise variances to simulate the noise interferences on the output signal. Use the proposed VB-KF iterative estimation algorithm to estimate the parameter vector.
Taking the calculated mean values of the Gaussian distributions for the parameters as the estimates of the unknown parameters, the proposed VB-KF parameter estimates and estimation error δ = | | θ ¯ ( s ) θ | | / | | θ | | with the increasing of the iteration s under different noise variances are presented in Table 1 and Figure 2. The curves of parameter estimates against s are plotted in Figure 3. The estimated noise variances are shown in Table 2.
When the noise variance is σ 2 = 0.10 2 , use the estimated parameters to construct the approximate difference equation:
y k + 0.49918 y k 1 + 0.31520 y k 2 + 0.17246 y k 3 = 0.97571 u k 1 + 0.54883 u k 2 + 0.99432 u k 3 + v k .
The computed prediction y ^ k and the true output y k are compared in Figure 4. From Figure 4, we can see that the predictions are very close to the true outputs.
For comparison, we also use the open-loop state observer to generate the estimates of unknown states [37]. Here, we combine the proposed VB parameter estimation method with state observer to form the state observer-based variational Bayesian (VB-SO) iterative estimation algorithm. In [57], the KF-based least squares iterative algorithm is used to identify the state space system, where the unknown process noise variance is set to one ( R = 1.00 ) in the Kalman filter. Here, we use this KF based-variational Bayesian algorithm (VB-KF2) and the proposed method (VB-KF1), where the noise variance is accurately computed, to estimate system parameters and states under different noises. Furthermore, the parameter estimation errors of the three different algorithms are compared in Figure 5 with σ 2 = 0.50 2 .
In addition, we adopt the other batch of data set for model validation. The parameter estimates with the least estimation error are chosen for the verification test. The root mean square errors (RMSE) of the three estimate models are computed as
RMSE 1 = 1 N 2 i = 1001 2000 [ y i y ^ vbkf 1 ] 2 = 0.2833 ,
RMSE 2 = 1 N 2 i = 1001 2000 [ y i y ^ vbso ] = 0.3007 ,
RMSE 3 = 1 N 2 i = 1001 2000 [ y i y ^ vbkf 2 ] 2 = 0.2930 .
From the calculated results, it can be seen that the proposed VB-KF algorithm has the smallest root mean square error.

5. Conclusions

In this paper, a Kalman filter-based variational Bayesian iterative algorithm is proposed to estimate the unknown parameters for linear difference equation systems. The main advantage of the presented method is that it can give the distribution of the parameter rather than a point estimate. Besides, the proposed algorithm obtains the noise statistics and state estimation. From the simulation results shown in Table 1 and Table 2 and Figure 2, Figure 3, Figure 4 and Figure 5, it is easy to find the following conclusions.
  • The parameter estimation errors are generally becoming smaller with iteration s increasing and with the noise variance decreasing. The estimated parameters are very close to the real values.
  • The estimated noise variances in Table 2 show that variance estimates are relatively large in the previous period because of the inaccurate parameters and states, but with the increasing of the iteration s, the noise variance estimates can converge to their true value.
  • Under the same batch of date and noise variance, the estimation accuracy of the proposed VB algorithm is higher than that of VB-SO and VB-KF2 algorithms.
  • The predicted outputs fit well with true outputs, which indicates the effectiveness of the proposed algorithm.
The proposed variational Bayesian iterative estimation algorithm in this paper for state space models from the observation data can combine some mathematical strategies [58,59,60,61] to study the parameter estimation algorithms of other stochastic systems with colored noise and can be applied to other fields [62,63,64,65,66].

Author Contributions

J.M., Q.F., F.G., and W.X. conceived and worked together to achieve this work; J.M. and F.G. discussed the method and the derivation process of the algorithm; Q.F. and W.X. compiled the simulation program by Matlab and analyzed the data; J.M. and Q.F. wrote the original draft. Finally, all the authors have read and approved the final manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (nos. 61803183, 21706096 and 61773182), the Fundamental Research Funds for the Central Universities (no. JUSRP11923), the 111 Project (no. B12018) and the Natural Science Foundation of Jiangsu Province (nos. BK20180591 and BK20160162).

Acknowledgments

The authors thank the anonymous reviewers for their careful reading and constructive comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Dallon, J.C.; Evans, E.J.; Grant, C.P.; Smith, W.V. Results from a differential equation model for cell motion with random switching show that the model cell velocity is asymptotically independent of force. J. Differ. Equ. 2019, 268, 301–317. [Google Scholar] [CrossRef]
  2. Senthil, R.; Janarthanan, K.; Prakash, J. Nonlinear state estimation using fuzzy Kalman filter. Ind. Eng. Chem. Res. 2006, 45, 8678–8688. [Google Scholar] [CrossRef]
  3. Sekizawa, S.; Inagaki, S.; Suzuki, T.; Hayakawa, S.; Tsuchida, N.; Tsuda, T.; Fujinami, H. Modeling and Recognition of Driving Behavior Based on Stochastic Switched ARX Model. IEEE Trans. Intell. Transp. Syst. 2007, 8, 593–606. [Google Scholar] [CrossRef]
  4. Garus, D.; Gogolla, T.; Krebber, K.; Schliep, F. Brillouin optical-fiber frequency-domain analysis for distributed temperature and strain measurements. J. Light. Technol. 1997, 15, 654–662. [Google Scholar] [CrossRef]
  5. John, V.; Ye, Z.; Kolwalkar, A. Investigation of anti-islanding protection of power converter based distributed generators using frequency domain analysis. IEEE Trans. Power Electron. 2004, 19, 1177–1183. [Google Scholar] [CrossRef]
  6. Worden, K.; Hickey, D.; Haroon, M.; Adams, D.E. Nonlinear system identification of automotive dampers: A time and frequency-domain analysis. Mech. Syst. Signal Process. 2009, 23, 104–126. [Google Scholar] [CrossRef]
  7. Xu, L.; Ding, F. Parameter estimation algorithms for dynamical response signals based on the multi-innovation theory and the hierarchical principle. IET Signal Process. 2017, 11, 228–237. [Google Scholar] [CrossRef]
  8. Xu, L.; Chen, L.; Xiong, W.L. Parameter estimation and controller design for dynamic systems from the step responses based on the Newton iteration. Nonlinear Dyn. 2015, 79, 2155–2163. [Google Scholar] [CrossRef]
  9. Xu, L. The damping iterative parameter identification method for dynamical systems based on the sine signal measurement. Signal Process. 2016, 120, 660–667. [Google Scholar] [CrossRef]
  10. Ding, F.; Liu, X.P.; Liu, G. Gradient based and least-squares based iterative identification methods for OE and OEMA systems. Digit. Signal Process. 2010, 20, 664–677. [Google Scholar] [CrossRef]
  11. Pan, J.; Jiang, X.; Wan, X.K.; Ding, W. A filtering based multi-innovation extended stochastic gradient algorithm for multivariable control systems. Int. J. Control Autom. Syst. 2017, 15, 1189–1197. [Google Scholar] [CrossRef]
  12. Ding, F. Decomposition based fast least squares algorithm for output error systems. Signal Process. 2013, 93, 1235–1242. [Google Scholar] [CrossRef]
  13. Ding, F. Two-stage least squares based iterative estimation algorithm for CARARMA system modeling. Appl. Math. Model. 2013, 37, 4798–4808. [Google Scholar] [CrossRef]
  14. Ljung, L. System Identification: Theory for the User, 2nd ed.; Prentice Hall: Englewood Cliffs, NJ, USA, 1999. [Google Scholar]
  15. Wan, L.J.; Ding, F. Decomposition- and gradient-based iterative identification algorithms for multivariable systems using the multi-innovation theory. Circuits Syst. Signal Process. 2019, 38, 2971–2991. [Google Scholar] [CrossRef]
  16. Ding, F.; Pan, J.; Alsaedi, A.; Hayat, T. Gradient-based iterative parameter estimation algorithms for dynamical systems from observation data. Mathematics 2019, 7, 428. [Google Scholar] [CrossRef]
  17. Liu, S.Y.; Ding, F.; Xu, L.; Hayat, T. Hierarchical principle-based iterative parameter estimation algorithm for dual-frequency signals. Circuits Syst. Signal Process. 2019, 38, 3251–3268. [Google Scholar] [CrossRef]
  18. Ding, F.; Liu, Y.J.; Bao, B. Gradient based and least squares based iterative estimation algorithms for multi-input multi-output systems. Proc. Inst. Mech. Eng. Part I J. Syst. Control Eng. 2012, 226, 43–55. [Google Scholar] [CrossRef]
  19. Xu, L.; Ding, F.; Zhu, Q.M. Hierarchical Newton and least squares iterative estimation algorithm for dynamic systems by transfer functions based on the impulse responses. Int. J. Syst. Sci. 2019, 50, 141–151. [Google Scholar] [CrossRef]
  20. Daunizeau, J.; Friston, K.J.; Kiebel, S.J. Variational Bayesian identification and prediction of stochastic nonlinear dynamic causal models. Phys. D 2009, 238, 2089–2118. [Google Scholar] [CrossRef]
  21. Khatibisepehr, S.; Huang, B. Dealing with irregular data in soft sensors: Bayesian method and comparative study. Ind. Eng. Chem. Res. 2008, 47, 8713–8723. [Google Scholar] [CrossRef]
  22. Yang, X.; Yin, S. Variational Bayesian inference for FIR models with randomly missing measurements. IEEE Trans. Ind. Electron. 2017, 64, 4217–4225. [Google Scholar] [CrossRef]
  23. Ding, F. State filtering and parameter estimation for state space systems with scarce measurements. Signal Process. 2014, 104, 369–380. [Google Scholar] [CrossRef]
  24. Liu, Y.J.; Ding, F.; Shi, Y. An efficient hierarchical identification method for general dual-rate sampled-data systems. Automatica 2014, 50, 962–970. [Google Scholar] [CrossRef]
  25. Ding, F. Combined state and least squares parameter estimation algorithms for dynamic systems. Appl. Math. Modell. 2014, 38, 403–412. [Google Scholar] [CrossRef]
  26. Viberg, M. Subspace-based methods for the identification of linear time-invariant systems. Automatica 1995, 21, 1835–1851. [Google Scholar] [CrossRef]
  27. Mao, Y.; Guo, X.; Zhao, Y. A state space force identification method based on Markov parameters precise computation and regularization technique. J. Sound Vib. 2010, 329, 300–3019. [Google Scholar] [CrossRef]
  28. Ding, F.; Liu, G.; Liu, X.P. Partially coupled stochastic gradient identification methods for non-uniformly sampled systems. IEEE Trans. Autom. Control 2010, 55, 1976–1981. [Google Scholar] [CrossRef]
  29. Zhang, X.; Ding, F.; Yang, E.F. State estimation for bilinear systems through minimizing the covariance matrix of the state estimation errors. Int. J. Adapt. Control Signal Process. 2019, 33, 1157–1173. [Google Scholar] [CrossRef]
  30. Zhou, L.; Li, X.; Shan, L.; Xia, J. Hierarchical recursive least squares parameter estimation of non-unifoimly sampled Hammerstein nonlinear systems based on Kalman filter. J. Frankl. Inst. 2017, 354, 4231–4246. [Google Scholar] [CrossRef]
  31. Sarkka, S.; Nymmenmaa, A. Recursive noise adaptive Kalman filtering by variational Bayesian approximations. IEEE Trans. Automat. Control 2009, 54, 596–600. [Google Scholar] [CrossRef]
  32. Huang, Y.; Zhang, Y.; Li, N.; Chambers, J. A robust Gaussian approximate fixed-interval smoother for nonlinear systems with heavy-tailed process and measurement noises. IEEE Signal Process. Lett. 2016, 23, 468–472. [Google Scholar] [CrossRef]
  33. Li, M.H.; Liu, X.M.; Ding, F. The filtering-based maximum likelihood iterative estimation algorithms for a special class of nonlinear systems with autoregressive moving average noise using the hierarchical identification principle. Int. J. Adapt. Control Signal Process. 2019, 33, 1189–1211. [Google Scholar] [CrossRef]
  34. Xu, L. The parameter estimation algorithms based on the dynamical response measurement data. Adv. Mech. Eng. 2017, 9, 1–12. [Google Scholar] [CrossRef]
  35. Ma, H.; Pan, J.; Ding, F.; Xu, L.; Ding, W. Partially-coupled least squares based iterative parameter estimation for multi-variable output-error-like autoregressive moving average systems. IET Control Theory Appl. 2019, 13, 3040–3051. [Google Scholar] [CrossRef]
  36. Xu, L. Application of the Newton iteration algorithm to the parameter estimation for dynamical systems. J. Comput. Appl. Math. 2015, 288, 33–43. [Google Scholar] [CrossRef]
  37. Xu, L.; Ding, F.; Gu, Y.; Alsaedi, A.; Hayat, T. A multi-innovation state and parameter estimation algorithm for a state space system with d-step state-delay. Signal Process. 2017, 140, 97–103. [Google Scholar] [CrossRef]
  38. Tzikas, D.G.; Likas, A.C.; Galatsanos, N.P. The variational approximation for Bayesian inference. IEEE Signal Process. Mag. 2008, 25, 131–146. [Google Scholar] [CrossRef]
  39. Latouche, P.; Birmele, E.; Ambroise, C. Variational Bayesian inference and complexity control for stochastic block models. Stat. Model. 2012, 12, 93–115. [Google Scholar] [CrossRef] [Green Version]
  40. Ma, J.X.; Xiong, W.L.; Chen, J.; Feng, D. Hierarchical identification for multivariate Hammerstein systems by using the modified Kalman filter. IET Control Theory Appl. 2017, 11, 857–869. [Google Scholar] [CrossRef]
  41. Ding, F.; Liu, G.; Liu, X.P. Parameter estimation with scarce measurements. Automatica 2011, 47, 1646–1655. [Google Scholar] [CrossRef]
  42. Ding, J.; Chen, J.Z.; Lin, J.X.; Wan, L.J. Particle filtering based parameter estimation for systems with output-error type model structures. J. Frankl. Inst. 2019, 356, 5521–5540. [Google Scholar] [CrossRef]
  43. Ding, J.; Chen, J.Z.; Lin, J.X.; Jiang, G.P. Particle filtering-based recursive identification for controlled auto-regressive systems with quantised output. IET Control Theory Appl. 2019, 13, 2181–2187. [Google Scholar] [CrossRef]
  44. Wang, Y.J.; Ding, F.; Wu, M.H. Recursive parameter estimation algorithm for multivariate output- error systems. J. Frankl. Inst. 2018, 355, 5163–5181. [Google Scholar] [CrossRef]
  45. Ding, F.; Wang, F.F.; Xu, L.; Wu, M.H. Decomposition based least squares iterative identification algorithm for multivariate pseudo-linear ARMA systems using the data filtering. J. Frankl. Inst. 2017, 354, 1321–1339. [Google Scholar] [CrossRef]
  46. Yin, C.C.; Wang, C.W. The perturbed compound Poisson risk process with investment and debit interest. Methodol. Comput. Appl. Probab. 2010, 12, 391–413. [Google Scholar] [CrossRef]
  47. Yin, C.C.; Yuen, K.C. Optimality of the threshold dividend strategy for the compound Poisson model. Stat. Probab. Lett. 2011, 81, 1841–1846. [Google Scholar] [CrossRef]
  48. Yin, C.C.; Wen, Y.Z. Optimal dividend problem with a terminal value for spectrally positive Levy processes. Insur. Math. Econom. 2013, 53, 769–773. [Google Scholar] [CrossRef] [Green Version]
  49. Yin, C.C.; Wen, Y.Z. Exit problems for jump processes with applications to dividend problems. J. Comput. Appl. Math. 2013, 245, 30–52. [Google Scholar] [CrossRef]
  50. Pan, J.; Li, W.; Zhang, H.P. Control algorithms of magnetic suspension systems based on the improved double exponential reaching law of sliding mode control. Int. J. Control Autom. Syst. 2018, 16, 2878–2887. [Google Scholar] [CrossRef]
  51. Ma, J.X.; Ding, F. Filtering-based multistage recursive identification algorithm for an input nonlinear output-error autoregressive system by using the key term separation technique. Circuits Syst. Signal Process. 2017, 36, 577–599. [Google Scholar] [CrossRef]
  52. Ma, P.; Ding, F. New gradient based identification methods for multivariate pseudo-linear systems using the multi-innovation and the data filtering. J. Frankl. Inst. 2017, 354, 1568–1583. [Google Scholar] [CrossRef]
  53. Wan, X.K.; Li, Y.; Xia, C.; Wu, M.H.; Liang, J.; Wang, N. A T-wave alternans assessment method based on least squares curve fitting technique. Measurement 2016, 86, 93–100. [Google Scholar] [CrossRef]
  54. Zhao, N. Joint Optimization of cooperative spectrum sensing and resource allocation in multi-channel cognitive radio sensor networks. Circuits Syst. Signal Process. 2016, 35, 2563–2583. [Google Scholar] [CrossRef]
  55. Zhao, X.L.; Liu, F.; Fu, B.; Na, F. Reliability analysis of hybrid multi-carrier energy systems based on entropy-based Markov model. Proc. Inst. Mech. Eng. Part O J. Risk Reliab. 2016, 230, 561–569. [Google Scholar] [CrossRef]
  56. Zhao, N.; Liang, Y.; Pei, Y. Dynamic contract incentive mechanism for cooperative wireless networks. IEEE Trans. Veh. Technol. 2018, 67, 10970–10982. [Google Scholar] [CrossRef]
  57. Ding, F.; Liu, X.; Ma, X. Kalman state filtering based least squares iterative parameter estimation for observer canonical state space systems using decomposition. J. Comput. Appl. Math. 2016, 301, 135–143. [Google Scholar] [CrossRef]
  58. Yin, C.C.; Wen, Y.Z. An extension of Paulsen-Gjessing’s risk model with stochastic return on investments. Insur. Math. Econom. 2013, 52, 469–476. [Google Scholar] [CrossRef] [Green Version]
  59. Yin, C.C.; Wen, Y.Z.; Zhao, Y.X. On the optimal dividend problem for a spectrally positive levy process. Astin Bull. 2014, 44, 635–651. [Google Scholar] [CrossRef] [Green Version]
  60. Yin, C.C.; Yuen, K.C. Exact joint laws associated with spectrally negative Levy processes and applications to insurance risk theory. Front. Math. China 2014, 9, 1453–1471. [Google Scholar] [CrossRef] [Green Version]
  61. Yin, C.C.; Yuen, K.C. Optimal dividend problems for a jump-diffusion model with capital injections and proportional transaction costs. J. Ind. Manag. Optim. 2015, 11, 1247–1262. [Google Scholar] [CrossRef]
  62. Gong, P.C.; Wang, W.Q.; Li, F.C.; Cheung, H. Sparsity-aware transmit beamspace design for FDA-MIMO radar. Signal Process. 2018, 144, 99–103. [Google Scholar] [CrossRef]
  63. Zhao, X.L.; Lin, Z.Y.; Fu, B.; He, L.; Na, F. Research on automatic generation control with wind power participation based on predictive optimal 2-degree-of-freedom PID strategy for multi-area interconnected power system. Energies 2018, 11, 3325. [Google Scholar] [CrossRef] [Green Version]
  64. Zhao, X.L.; Lin, Z.Y.; Fu, B.; He, L.; Li, C.S. Research on the predictive optimal PID plus second order derivative method for AGC of power system with high penetration of photovoltaic and wind power. J. Electr. Eng. Technol. 2019, 14, 1075–1086. [Google Scholar] [CrossRef]
  65. Shi, W.X.; Liu, N.; Zhou, Y.M.; Cao, X.A. Effects of postannealing on the characteristics and reliability of polyfluorene organic light-emitting diodes. IEEE Trans. Electron Devices 2019, 66, 1057–1062. [Google Scholar] [CrossRef]
  66. Wu, T.Z.; Shi, X.; Liao, L.; Zhou, C.J.; Zhou, H.; Su, Y.H. A capacity configuration control strategy to alleviate power fluctuation of hybrid energy storage system based on improved particle swarm optimization. Energies 2019, 12, 642. [Google Scholar] [CrossRef] [Green Version]
Figure 1. The flowchart of the proposed variational Bayesian Kalman filter (VB-KF) iterative algorithm for computing parameters and states.
Figure 1. The flowchart of the proposed variational Bayesian Kalman filter (VB-KF) iterative algorithm for computing parameters and states.
Mathematics 07 01143 g001
Figure 2. The VB parameter estimation error under different noise variance.
Figure 2. The VB parameter estimation error under different noise variance.
Mathematics 07 01143 g002
Figure 3. The VB parameter estimates against s with noise variance σ 2 = 0.50 2 .
Figure 3. The VB parameter estimates against s with noise variance σ 2 = 0.50 2 .
Mathematics 07 01143 g003
Figure 4. The true outputs and predictions with noise variance σ 2 = 0.10 2 .
Figure 4. The true outputs and predictions with noise variance σ 2 = 0.10 2 .
Mathematics 07 01143 g004
Figure 5. Comparison of estimation errors δ versus s under different algorithms ( σ 2 = 0.50 2 ) .
Figure 5. Comparison of estimation errors δ versus s under different algorithms ( σ 2 = 0.50 2 ) .
Mathematics 07 01143 g005
Table 1. The parameter estimates and estimation errors versus s under the proposed VB-KF algorithm.
Table 1. The parameter estimates and estimation errors versus s under the proposed VB-KF algorithm.
σ 2 s a 1 a 2 κ 1 κ 2 κ 3 b 2 δ   ( % )
0.10 2 10.01016−0.02289−0.037001.003970.531940.9778440.69214
2−0.073610.007280.897020.946870.520430.9639235.08236
50.174740.301560.494710.975220.550290.992532.40308
100.172430.315210.499200.975710.548830.994322.19306
200.172460.315200.499180.975710.548830.994322.19278
400.172460.315200.499180.975710.548830.994322.19278
0.50 2 1−0.05182−0.058090.041321.008420.548800.9971139.39494
2−0.115740.049630.900890.938490.526090.9700335.19921
50.176880.319970.493490.967590.560151.000642.85674
100.176740.327780.497670.967740.558741.001892.90030
200.176760.327770.497640.967740.558741.001892.90017
400.176760.327770.497640.967740.558741.001892.90017
1.00 2 1−0.017850.06661−0.156940.983650.518351.0172945.26212
2−0.01462−0.049350.756650.894820.484190.9641931.18481
50.172910.288670.461870.916820.508360.986956.59330
100.170820.295530.465570.917390.507760.988256.41325
200.170830.295530.465560.917390.507760.988256.41343
400.170830.295530.465560.917390.507760.988256.41343
true values0.180000.320000.500001.000000.550000.97000
Table 2. The estimates of noise variance versus s under the proposed VB algorithm.
Table 2. The estimates of noise variance versus s under the proposed VB algorithm.
σ 2 / s 1235101540
0.10 2 0.20000.75090.03560.02800.01080.01080.0108
0.50 2 0.20000.99350.30100.27480.26640.26640.2664
1.00 2 0.20001.62521.02411.00970.99500.99500.9950

Share and Cite

MDPI and ACS Style

Ma, J.; Fei, Q.; Guo, F.; Xiong, W. Variational Bayesian Iterative Estimation Algorithm for Linear Difference Equation Systems. Mathematics 2019, 7, 1143. https://doi.org/10.3390/math7121143

AMA Style

Ma J, Fei Q, Guo F, Xiong W. Variational Bayesian Iterative Estimation Algorithm for Linear Difference Equation Systems. Mathematics. 2019; 7(12):1143. https://doi.org/10.3390/math7121143

Chicago/Turabian Style

Ma, Junxia, Qiuling Fei, Fan Guo, and Weili Xiong. 2019. "Variational Bayesian Iterative Estimation Algorithm for Linear Difference Equation Systems" Mathematics 7, no. 12: 1143. https://doi.org/10.3390/math7121143

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop