Next Article in Journal
Mechanical Impurities Carry-Over from Horizontal Heavy Oil Production Well
Previous Article in Journal
Experimental Simulation on the Stress Disturbance Mechanism Caused by Hydraulic Fracturing on the Mechanical Properties of Shale Formation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Linear Quadratic Optimal Control of Discrete-Time Stochastic Systems Driven by Homogeneous Markov Processes

1
College of Mathematics and Systems Science, Shandong University of Science and Technology, Qingdao 266590, China
2
College of Electronic and Information Engineering, Shandong University of Science and Technology, Qingdao 266590, China
3
College of Electrical Engineering and Automation, Shandong University of Science and Technology, Qingdao 266590, China
*
Author to whom correspondence should be addressed.
Processes 2023, 11(10), 2933; https://doi.org/10.3390/pr11102933
Submission received: 17 June 2023 / Revised: 19 September 2023 / Accepted: 26 September 2023 / Published: 9 October 2023
(This article belongs to the Section Automation Control Systems)

Abstract

:
Random terms in many natural and social science systems have distinct Markovian characteristics, such as Markov jump-taking values in a finite or countable set, and Wiener process-taking values in a continuous set. In general, these systems can be seen as Markov-process-driven systems, which can be used to describe more complex phenomena. In this paper, a discrete-time stochastic linear system driven by a homogeneous Markov process is studied, and the corresponding linear quadratic (LQ) optimal control problem for this system is solved. Firstly, the relations between the well-posedness of LQ problems and some linear matrix inequality (LMI) conditions are established. Then, based on the equivalence between the solvability of the generalized difference Riccati equation (GDRE) and the LMI condition, it is proven that the solvability of the GDRE is sufficient and necessary for the well-posedness of the LQ problem. Moreover, the solvability of the GDRE and the feasibility of the LMI condition are established, and it is proven that the LQ problem is attainable through a certain feedback control when any of the four conditions is satisfied, and the optimal feedback control of the LQ problem is given using the properties of homogeneous Markov processes and the smoothness of the conditional expectation. Finally, a practical example is used to illustrate the validity of the theory.

1. Introduction

Linear quadratic (LQ) optimal control problems play an important role in control theory and practical applications [1,2]. The Riccati equation method is a well-known method for studying the LQ problem for deterministic systems described by ordinary differential equations [3,4,5], and this method can be extended to stochastic cases for Itô-type stochastic systems [6,7,8,9]. With the development of control theories, the LQ problem for discrete-time stochastic systems has also been studied by many scholars, achieving impressive results [10,11,12]. In these studies, the value functions and optimal controls for finite- or infinite-horizon indefinite LQ problems were obtained based on the solutions to some Riccati equations with forms of algebraic, differential, and difference Riccati equations [13,14,15,16,17,18,19]. In system controller design, some external disturbances or parameter uncertainties should be addressed to eliminate or reduce their impact on the system’s performance [20,21,22]. In some studies, the uncertain parts of the systems are treated as unknown disturbances, e.g., Ref. [22] presented the robust non-linear generalized predictive control method to improve the system’s robustness and [23] used the H control design method. Many studies describe these complex systems as stochastic systems in which the random or fluctuating terms are usually modeled by white noise, Markov chains, or Wiener processes [24,25,26,27,28,29,30,31,32].
In practice, random terms in many natural and social science systems exhibit distinct Markovian characteristics. Markovian processes find applications in various fields, such as economy, finance, and engineering systems. For example, Markov jumps, which take values in a finite or countable set, are widely used to describe the jumping phenomenon of various systems [33,34,35]. In some situations, there are processes that possess Markovian properties but take values in a continuous set, such as Wiener processes and fractal Brownian motions, which are used to describe different types of noise in continuous- or discrete-time systems [36,37,38]. These systems are driven by a type of stochastic process with Markovian properties. So, in general, such systems are called Markov-process-driven systems, in which the Markovian processes are seen as an extension of white noise. This paper discusses the LQ problem for discrete-time linear systems driven by Markov processes. In more detail, the following LQ optimal control is considered. Minimize the cost function
J ( u ) = E k = 0 N 1 [ x k T Q ( k ) x k + 2 x k T S ( k ) u k + u k T R ( k ) u k ] + E [ x N T Q ( N ) x N ]
subject to
x k + 1 = A ( k ) x k + B ( k ) u k + A 1 ( k ) x k ω k , x 0 R n , k = 0 , , N 1
where A ( k ) , A 1 ( k ) R n × n , B ( k ) , S ( k ) R n × n v , Q ( k ) S n , and R ( k ) S n v . x k R n is the state variable, u k R n v is the control, and ω k is a Markovian process. Compared to the stochastic systems discussed in [10], the random disturbance ω k in system (2) is no longer an independent process but rather a Markovian process, i.e., the probability distributions of ω k are related to x k 1 . Compared to the system discussed in [39], ω k in (2) is a Markovian process that can take values in a continuous set rather than a countable set. So, the novelties of this paper can be summarized as follows: (1) A discussion of more general systems driven by Markovian processes, extending beyond systems driven by white noise, Markov jumps, Wiener processes, etc. (2) A derivation of more general results with distribution forms to describe the probability distributions of Markov processes.
This paper is organized as follows. In Section 2, the basic theoretical knowledge involved in this paper is introduced. In Section 3, the equivalence of the well-posedness and attainability of the LQ problem in (1) and (2) is discussed, and the optimal controller and minimum value of the cost function are obtained. In Section 4, two examples are used to illustrate the feasibility of the theory.
The following notations are used in this paper. R n : the set of all real n-dimensional vectors; R m × n : the set of m × n real matrices; A T , x T : transpose of a matrix A or vector x; A 1 : inverse of matrix A; A > 0 ( A 0 ) : A is a positive (positive semi-) definite matrix; I D ( x ) : indicator function of set D with I D ( x ) = 1 when x D and I D ( x ) = 0 when x D ; E [ X ] : the expectation of a random variable X; E [ X Y = y ] : the conditional expectation of X under the condition Y = y ; and L 2 ( Ω , F , P ) : the complete space of random variables with E [ | X | 2 ] < , where X L 2 ( Ω , F , P ) .

2. Preliminaries

Let { ω k , k = 0 , 1 , , N } be a homogeneous Markovian process defined on a complete probability space ( Ω , F , P ) with a one-step transition probability density function p ( ξ , η ) , ξ , η R , and the initial distribution is p 0 ( ξ ) . For every given k = 1 , 2 , , N , the joint probability density function of ω 0 , ω 1 , , ω k is
f ( ξ 0 , ξ 1 , , ξ k ) = f 0 ( ξ 0 ) f 1 ( ξ 1 | ξ 0 ) f 2 ( ξ 2 | ξ 0 , ξ 1 ) f k ( ξ k | ξ 0 , ξ 1 , , ξ k 1 )
where f k ( ξ k | ξ 0 , ξ 1 , , ξ k 1 ) denotes the conditional probability density function of ω k under the conditions of ω 0 = ξ 0 , ω 1 = ξ 1 , , ω k 1 = ξ k 1 , ξ i R , 0 i k . Because ω k , k = 0 , 1 , , N is a Markovian process, the conditional probability density function f k ( ξ k | ξ 0 , ξ 1 , , ξ k 1 ) only depends on ω k 1 = ξ k 1 , i.e.,
f k ( ξ k | ξ 0 , ξ 1 , , ξ k 1 ) = f k ( ξ k | ξ k 1 )
and f k ( ξ k | ξ k 1 ) is just the one-step transition probability density function. So,
f k ( ξ k | ξ k 1 ) = p ( ξ k 1 , ξ k ) .
The joint probability density function of ω 0 , ω 1 , , ω k can be represented by
f ( ξ 0 , ξ 1 , , ξ k ) = p 0 ( ξ 0 ) p ( ξ 0 , ξ 1 ) p ( ξ 1 , ξ 2 ) p ( ξ k 1 , ξ k ) .
Suppose that X is a random variable generated by ω 0 , ω 1 , , ω k , i.e., X = X ( ω 0 , ω 1 , , ω k ) . The conditional expectation of X is
E [ X ( ω 0 , ω 1 , , ω k 1 , ω k ) | ω 0 = ξ 0 , ω 1 = ξ 1 , , ω k 1 = ξ k 1 ] = R X ( ξ 0 , ξ 1 , , ξ k 1 , η ) f ( η | ξ 0 , ξ 1 , , ξ k 1 ) d η = R X ( ξ 0 , ξ 1 , , ξ k 1 , η ) p ( ξ k 1 , η ) d η
In the following discussion, the condition expectation E [ X ( ω 0 , ω 1 , , ω k 1 , ω k ) | ω 0 = ξ 0 , ω 1 = ξ 1 , , ω k 1 = ξ k 1 ] is shortened by E [ X | ω 0 , ω 1 , , ω k 1 ]
The following lemma is used in this paper.
Lemma 1.
Given a series of random symmetric matrices P 1 ( ω 0 ) , , P N ( ω N 1 ) , the following results hold
E [ P k + 1 ( ω k ) ω 0 , , ω k 1 ] = E [ P k + 1 ( ω k ) ω k 1 = ξ ] = R P k + 1 ( η ) p ( ξ , η ) d η ,
E [ P k + 1 ( ω k ) ω k ω 0 , , ω k 1 ] = E [ P k + 1 ( ω k ) ω k ω k 1 = ξ ] = R P k + 1 ( η ) η p ( ξ , η ) d η ,
E [ P k + 1 ( ω k ) ω k 2 ω 0 , , ω k 1 ] = E [ P k + 1 ( ω k ) ω k 2 ω k 1 = ξ ] = R P k + 1 ( η ) η 2 p ( ξ , η ) d η .
Proof. 
According to the definition of a Markovian process, the conditional probability density function of ω k satisfies f ( ω k ω 0 , , ω k 1 ) = f ( ω k ω k 1 ) , and according to the definition of the conditional expectation,
E [ φ ( ω k ) ω 0 , , ω k 1 ] = R φ ( ω k ) f ( ω k ω 0 , , ω k 1 ) d η = R φ ( ω k ) f ( ω k ω k 1 ) d η = E [ φ ( ω k ) ω k 1 ] ,
We obtain
E [ P k + 1 ( ω k ) ω k 1 = ξ ] = R P k + 1 ( η ) p ( ξ , η ) d η ,
E [ P k + 1 ( ω k ) ω k ω k 1 = ξ ] = R P k + 1 ( η ) η p ( ξ , η ) d η ,
E [ P k + 1 ( ω k ) ω k 2 ω k 1 = ξ ] = R P k + 1 ( η ) η 2 p ( ξ , η ) d η .
This ends the proof. □
Let F k be the σ field generated by σ 0 , σ 1 , , σ k 1 ( k 1 ) and σ 0 = { , Ω } .
Definition 1.
The LQ problem in (1) and (2) is considered well-posed if
inf u J ( u ) > .
The LQ problem in (1) and (2) is called attainable if there exists an admissible control ( u 0 , , u N 1 ) , such that
inf u J ( u ) = J ( u ) ,
and u is called an optimal control.

3. LQ Problem of the Discrete-Time Linear Stochastic Systems Driven by a Homogeneous Markovian Process

In this section, the well-posedness and attainability of the LQ problem are studied. We suppose that in the optimal control problem in (1) and (2), the admissible control set U = { U k } k = 0 N 1 , U k = L 2 ( Ω , F k ; R n u ) , and { ω k } k = 0 N are the homogeneous Markovian processes given in Section 2. We also assume that u k is the optimal control problem in (1) and (2), and the corresponding optimal trajectory is x k . Under the premise that the optimal cost function is finite, the LQ problem is always attainable through optimal control. Next, we establish the relationship between the well-posedness of the LQ problem and an LMI condition and then prove that the LMI condition is equivalent to both the solvability of the GDRE and the attainability of the LQ problem. In other words, we establish the equivalent relationship between the well-posedness and attainability of the LQ problem, the solvability of the GDRE, and the feasibility of the LMI condition and find the optimal control and optimal value function. In the following discussion, we denote
2 A T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k )
as the corresponding symmetric matrix
A T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A ( k ) .

3.1. Well-Posedness

The following provides a connection between the well-posedness and the feasibility of an LMI involving unknown symmetric matrices [40].
Theorem 1.
The LQ problem in (1) and (2) is well-posed if there exist symmetric matrices P 0 , , P N satisfying the following LMI’s condition
Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + 2 A T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R η 2 P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) P k ( ξ ) S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) 0
for k = 0 , , N 1 , and P N = Q ( N ) 0 , where ∗ denotes the symmetric part.
Proof. 
Let P 1 , , P N satisfy (6). Then, by adding the following trivial equality,
k = 0 N 1 E ( x k + 1 T P k + 1 x k + 1 x k T P k x k ) = E ( x N T P N x N x 0 T P 0 x 0 )
to the cost function,
J ( u ) = E k = 0 N 1 [ x k T Q ( k ) x k + 2 x k T S ( k ) u k + u k T R ( k ) u k ] + E [ x N T Q N x N ] ,
and using the system in Equation (1), we can rewrite the cost function as follows:
J ( u ) = E k = 0 N 1 { x k T [ Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + 2 A T ( k ) A 1 ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η + A 1 T ( k ) A 1 ( k ) R η 2 P k + 1 ( η ) p ( ξ , η ) d η P k ( ξ ) ] x k + 2 u k T [ S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) A 1 ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η ] x k + u k T [ R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) ] u k } + E [ x N T ( Q N P N ) x N ] + E ( x 0 T P 0 x 0 ) = E k = 0 N 1 x k u k T Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + 2 A T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R η 2 P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) P k ( ξ ) S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R η P k + 1 ( η ) p ( ξ , η ) d η A 1 ( k ) Ξ × x k u k + E [ x N T ( Q N P N ) x N ] + E ( x 0 T P 0 x 0 ) ,
where Ξ = R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) . From the above equality, we can see that the cost function J ( u ) is bounded from below by E ( x 0 T P 0 x 0 ) ; hence, the LQ problem in (1) and (2) is well-posed. □
We have shown in the proof of Theorem 1 that the proposed LMI condition (6) is sufficient for the well-posedness of the LQ problem. Next, we show that the LMI condition is equivalent to the solvability of the GDRE, and then provide a connection between the well-posedness of the LQ problem and the solvability of the GDRE. Meanwhile, the necessity of the LMI condition (6) for the well-posedness of the LQ problem is also proven.
Lemma 2
(Extended Schur’s lemma [41]). Let M = M T and C and D = D T be given matrices of appropriate proper sizes. Then, the following conditions are equivalent:
(i) 
M C D 1 C T 0 , D 0 .
(ii) 
M C C T D 0 .
(iii) 
D C T C M 0 .
Lemma 3.
Let F = F T and H and G = G T be given matrices of appropriate sizes. Consider the following quadratic form:
f ( x , u ) = E [ x T F x + 2 x T H u + u T G u ] ,
where x and u are random variables belonging to the space L 2 ( Ω , F , P ) . Then, the following conditions are equivalent:
(i) 
inf u f ( x , u ) > for any random variable x.
(ii) 
There exists a symmetric matrix Z = Z T , such that inf u f ( x , u ) = E [ x T Z x ] for any random variable x.
(iii) 
G 0 and ker ( G ) ker ( H ) .
(iv) 
There exists a symmetric matrix W = W T , such that
F W H H T G 0 .
Moreover, if G > 0 and any of the above conditions holds, then (ii) is satisfied by Z = F H G 1 H T , and for any random variable x, the random variable u = G 1 H T x is optimal with the following optimal value:
f ( x , u ) = E [ x T ( F H G 1 H T ) x ] .
Proof. 
( i ) ( i i i ) : Suppose there is a v such that v T G v < 0 . Then, for any scalar γ > 0 , we have lim γ + f ( x , γ v ) = . This leads to a contradiction through the assumption. So, G must be positive.
Suppose now ker ( G ) ker ( H ) . That is, there exists u, such that G u = 0 and H u 0 . Take any scalar γ > 0 . Then, we have lim γ + f ( H u , γ u ) = , which contradicts (i).
( i i i ) ( i i ) : Through a simple calculation, we can obtain the following:
f ( x , u ) = E [ x T ( F H G 1 H T ) x + ( u T + x T H G 1 ) G ( u + G 1 H T x ) ] .
Let Z = F H G 1 H T . Then, it is immediate that inf u f ( x , u ) = E [ x T Z x ] for any random variable x.
( i i ) ( i v ) : Since f ( x , u ) = E [ x T F x + 2 x T H u + u T G u ] E [ x T Z x ] , that is,
E x u T F Z H H T G x u 0 ,
the condition (iv) holds with W = Z .
( i v ) ( i ) : Since
F H H T G W 0 0 0 = F W H H T G 0 ,
for every x and u, there exists
E x u T F H H T G x u E x u T W 0 0 0 x u 0 ,
i.e.,
f ( x , u ) E [ x T W x ] 0 .
So, for every x , u L 2 ( Ω , F , P )
f ( x , u ) E [ x T W x ] > ,
and
inf u f ( x , u ) > ,
This proves that (i) is true.
Furthermore, if G > 0 , by applying the completing square method to f ( x , u ) with respect to u, we obtain
f ( x , u ) = E x T ( F H G 1 H T ) x + ( u + G 1 H T x ) T G ( u + G 1 H T x ) .
Let u = u = G 1 H T x . We can directly obtain f ( x , u ) = E x T ( F H G 1 H T ) x . This ends the proof. □
The following provides a connection between the well-posedness of the LQ problem and the solvability of the GDRE.
Theorem 2.
The LQ problem in (1) and (2) is well-posed if and only if there exist symmetric matrices P 0 , , P N satisfying the GDRE, where the randomness of P k + 1 ( k = 0 , 1 , , N 1 ) is generated by ω 0 , ω 1 , , ω k . Furthermore, the optimal cost is given by
inf u 0 , , u N 1 J ( u ) = E [ x 0 T P 0 x 0 ] .
Proof. 
We prove that the solvability of the GDRE is necessary for the well-posedness of the LQ problem by induction. This needs to consider the cost function from l to N. Suppose that
V l ( x l ) = inf u l , , u N 1 E [ k = l N 1 ( x k T Q ( k ) x k + 2 x k T S ( k ) u k + u k T R ( k ) u k ) + x N T Q N x N ] .
Then V l 2 ( x l 2 ) is also finite for any l 1 l 2 when V l 1 ( x l 1 ) is finite. This fact is used at each step of the induction: the LQ problem is assumed to be well-posed at the initial time, so the cost function V l ( x l ) is finite at any stage 0 l N 1 .
First, we consider the case of l = N 1 , and let P N = Q N 0 . There exists
V N 1 ( x N 1 ) = inf u N 1 E { x N 1 T [ Q N 1 + A T ( N 1 ) R P N ( η ) p ( ξ , η ) d η A ( N 1 ) + 2 A T ( N 1 ) R P N ( η ) η p ( ξ , η ) d η A 1 ( N 1 ) + A 1 T ( N 1 ) R P N ( η ) η 2 p ( ξ , η ) d η A 1 ( N 1 ) ] x N 1 + 2 x N 1 T [ S N 1 + A T ( N 1 ) R P N ( η ) p ( ξ , η ) d η B ( N 1 ) + A 1 T ( N 1 ) R P N ( η ) η p ( ξ , η ) d η B ( N 1 ) ] u N 1 + u N 1 T [ R N 1 + B T ( N 1 ) R P N ( η ) p ( ξ , η ) d η B ( N 1 ) ] u N 1 } .
So,
V N 1 ( x N 1 ) = E [ x N 1 T P N 1 ( ξ ) x N 1 ] .
According to Lemma 3, we have the following conditions:
P N 1 ( ξ ) = Q N 1 + A T ( N 1 ) R P N ( η ) p ( ξ , η ) d η A ( N 1 ) + A T ( N 1 ) R P N ( η ) η p ( ξ , η ) d η A 1 ( N 1 ) + A 1 T ( N 1 ) R P N ( η ) η p ( ξ , η ) d η A ( N 1 ) + A 1 T ( N 1 ) R P N ( η ) η 2 p ( ξ , η ) d η A 1 ( N 1 ) H N 1 T ( ξ ) G N 1 1 ( ξ ) H N 1 ( ξ ) ,
with G N 1 ( ξ ) = R N 1 + B T ( N 1 ) R P N ( η ) p ( ξ , η ) d η B ( N 1 ) , where
H N 1 ( ξ ) = S N 1 T + B T ( N 1 ) R P N ( η ) p ( ξ , η ) d η A ( N 1 ) + B T ( N 1 ) R P N ( η ) η p ( ξ , η ) d η A 1 ( N 1 ) .
Because p ( ξ , η ) is the transition probability density function, there exists p ( ξ , η ) 0 . So, for every ξ , η , we have P N ( η ) p ( ξ , η ) 0 , and the following results can be obtained:
B T ( N 1 ) R P N ( η ) p ( ξ , η ) d η B ( N 1 ) 0 .
Reminding that R N 1 0 , we have G N 1 ( ξ ) = R N 1 + B T ( N 1 ) R P N ( η ) p ( ξ , η ) d η B ( N 1 ) 0 . So, Equation (8) satisfies the GDRE for k = N 1 .
Next, suppose that we have found a sequence of symmetric matrices P l , , P N 1 , which solves the GDRE for i = l , , N 1 , and satisfies
V l ( x l ) = E [ x l T P l ( ξ ) x l ] .
Then, the following results are derived:
V l 1 ( x l 1 ) = inf u l 1 E [ x l 1 T Q l 1 x l 1 + 2 x l 1 T S l 1 u l 1 + u l 1 T R l 1 u l 1 + V l ( x l ) ] = inf u l 1 E [ x l 1 T Q l 1 x l 1 + 2 x l 1 T S l 1 u l 1 + u l 1 T R l 1 u l 1 + x l T P l x l ] = inf u l 1 E { x l 1 T [ Q l 1 + A T ( l 1 ) R P l ( η ) p ( ξ , η ) d η A ( l 1 ) + 2 A T ( l 1 ) R P l ( η ) η p ( ξ , η ) d η A 1 ( l 1 ) + A 1 T ( l 1 ) R P l ( η ) η 2 p ( ξ , η ) d η A 1 ( l 1 ) ] x l 1 + 2 x l 1 T [ S l 1 + A T ( l 1 ) R P l ( η ) p ( ξ , η ) d η B ( l 1 ) + A 1 T ( l 1 ) R P l ( η ) η p ( ξ , η ) d η B ( l 1 ) ] u l 1 + u l 1 T [ R l 1 + B T ( l 1 ) R P l ( η ) p ( ξ , η ) d η B ( l 1 ) ] u l 1 } .
Since V l 1 ( x l 1 ) is finite, according to Lemma 3, we have
P l 1 ( ξ ) = Q l 1 + A T ( l 1 ) R P l ( η ) p ( ξ , η ) d η A ( l 1 ) + A T ( l 1 ) R P l ( η ) η p ( ξ , η ) d η A 1 ( l 1 ) + A 1 T ( l 1 ) R P l ( η ) η p ( ξ , η ) d η A ( l 1 ) + A 1 T ( l 1 ) R P l ( η ) η 2 p ( ξ , η ) d η A 1 ( l 1 ) H l 1 T ( ξ ) G l 1 1 ( ξ ) H l 1 ( ξ ) , G l 1 ( ξ ) = R l 1 + B T ( l 1 ) R P l ( η ) p ( ξ , η ) d η B ( l 1 ) 0 , H l 1 ( ξ ) = S l 1 T + B T ( l 1 ) R P l ( η ) p ( ξ , η ) d η A ( l 1 ) + B T ( l 1 ) R P l ( η ) η p ( ξ , η ) d η A 1 ( l 1 ) .
In addition,
V l 1 ( x l 1 ) = E [ x l 1 T P l 1 ( ξ ) x l 1 ] .
By the recursion method, the necessity of the GDRE for the well-posedness of the LQ problem has been proven.
According to Lemma 2, we deduce that the solution of the GDRE also satisfies the LMI condition (6), which, according to Theorem 2, implies the well-posedness of the LQ problem. □
Remark 1.
The following constrained difference equation is called a generalized difference Riccati equation (GDRE):
P k ( ξ ) = Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + A T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A ( k ) + A 1 T ( k ) R P k + 1 ( η ) η 2 p ( ξ , η ) d η A 1 ( k ) H k T ( ξ ) G k 1 ( ξ ) H k ( ξ ) , P N = Q N , k = 0 , , N 1 ,
where
H k ( ξ ) = S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) A 1 ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η , G k ( ξ ) = R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) 0 , k = 0 , , N 1 .
Compared to the Riccati equations obtained in [7], the forms of (9) are more complex, and the probability distributions of the Markov processes are necessary. Specifically, if ω k and ω k 1 are independent with identical distributions and with the expectation E [ ω k ] = 0 and the variance E [ ω k 2 ] = 1 , we have p ( ξ , η ) = p ˜ ( η ) for all ξ R , which implies independence between ω k and ω k 1 , and R η p ( ξ , η ) d η = R η p ˜ ( η ) d η = 0 and R η 2 p ( ξ , η ) d η = R η 2 p ˜ ( η ) d η = 1 for all ξ R . In this case, we can take P k as deterministic matrices that are not dependent on ω k 1 or ξ. The forms of the Riccati equations are given more simply:
P k = Q ( k ) + A T ( k ) P k + 1 A ( k ) + A 1 T ( k ) P k + 1 A 1 ( k ) H k T G k 1 H k , P N = Q N , k = 0 , , N 1 ,
where
H k = S ( k ) T + B T ( k ) P k + 1 A ( k ) + B T ( k ) A 1 ( k ) P k + 1 , G k = R ( k ) + B T ( k ) P k + 1 B ( k ) > 0 , k = 0 , , N 1 ,
which is the form of the GDRE provided in [10]. Because the systems discussed in this paper are only for finite-time cases, the system’s stability is not discussed. However, the stability of such Markov-process-driven systems is also a topic worthy of further research, and excellent relevant results can be found in [42,43].
Remark 2.
From Lemma 2 and Theorem 2, it is obvious that the LMI condition (6) is also necessary for the well-posedness of the LQ problem. So, the LMI condition (8) is a sufficient and necessary condition for the well-posedness of the LQ problem.

3.2. Attainability

The following result shows the equivalent relationship between the well-posedness and attainability of the LQ problem, the solvability of the GDRE, and the feasibility of the LMI condition and provides the optimal control by which the LQ problem is attainable, as well as the optimal value function.
Theorem 3.
The following are equivalent:
(i) 
The LQ problem in (1) and (2) is well-posed.
(ii) 
The LQ problem in (1) and (2) is attainable.
(iii) 
The LMI condition (6) is feasible.
(iv) 
The GDRE (9) is solvable.
In addition, when any of the above conditions are satisfied, the LQ problem in (1) and (2) is attainable through
u k = G k 1 ( ξ ) H k ( ξ ) x k , k = 0 , , N 1 ,
and the optimal cost function
J ( u k ) = inf u 0 , , u N 1 J ( u ) = E [ x 0 T P 0 x 0 ] ,
where
G k ( ξ ) = R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) , H k ( ξ ) = S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) A 1 ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η ,
P 0 , , P N are solutions to the GDRE (9).
Proof. 
According to Theorem 1 and Theorem 2, the equivalences ( i ) ( i i i ) ( i v ) are straightforward. Next, we prove that the LQ problem is attainable through the feedback control law (10). To this end, by introducing the symmetric matrices P k , P k + 1 , k = 0 , , N 1 , where the randomness of P k + 1 is generated by ω 0 , ω 1 , , ω k , and the randomness of P k is generated by ω 0 , ω 1 , , ω k 1 , we have
E ( x k + 1 T P k + 1 x k + 1 x k T P k x k ) = E { x k T [ A T ( k ) P k + 1 A ( k ) P k ] x k + 2 u k T B T ( k ) P k + 1 A ( k ) x k + u k T B T ( k ) P k + 1 B ( k ) u k + 2 x k T A T ( k ) P k + 1 A 1 ( k ) x k ω k + 2 u k T B T ( k ) P k + 1 A 1 ( k ) x k ω k + ω k T x k T A 1 T ( k ) P k + 1 A 1 ( k ) x k ω k } .
Through the smoothing property of the conditional expectation, we have
E ( x k + 1 T P k + 1 x k + 1 x k T P k x k ) = E { E [ x k T ( A T ( k ) P k + 1 A ( k ) P k ) x k + 2 u k T B T ( k ) P k + 1 A ( k ) x k + u k T B T ( k ) P k + 1 B ( k ) u k + 2 x k T A T ( k ) P k + 1 A 1 ( k ) x k ω k + 2 u k T B T ( k ) P k + 1 A 1 ( k ) x k ω k + ω k T x k T A 1 T ( k ) P k + 1 A 1 ( k ) x k ω k ω 0 , , ω k 1 ] } ,
since the randomness of x k , u k , and P k is generated by ω 0 , ω 1 , , ω k 1 , but the randomness of P k + 1 is generated by ω 0 , ω 1 , , ω k . According to Lemma 1, we can obtain
E ( x k + 1 T P k + 1 x k + 1 x k T P k x k ) = E { x k T [ A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) P k ( ξ ) + 2 A T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R P k + 1 ( η ) η 2 p ( ξ , η ) d η A 1 ( k ) ] x k + 2 u k T [ B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) ] x k + u k T B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) u k } .
Denote
ϕ = x k T [ A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) P k ( ξ ) + 2 A T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R P k + 1 ( η ) η 2 p ( ξ , η ) d η A 1 ( k ) ] x k + 2 u k T [ B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) ] x k + u k T B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) u k ,
and then
E ( x k + 1 T P k + 1 x k + 1 x k T P k x k ) = E ( ϕ )
Summing k from 0 to N 1 on both sides of Equation (12) at the same time, we can obtain
E ( x N T P N x N x 0 T P 0 x 0 ) = E k = 0 N 1 ( ϕ ) ,
i.e.,
E k = 0 N 1 ( ϕ ) E ( x N T P N x N x 0 T P 0 x 0 ) = 0 .
So,
J ( u ) = E k = 0 N 1 [ x k T Q ( k ) x k + 2 x k T S ( k ) u k + u k T R ( k ) u k ] + E [ x N T Q N x N ] + E k = 0 N 1 ( ϕ ) E ( x N T P N x N x 0 T P 0 x 0 ) = E k = 0 N 1 x k T [ Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) P k ( ξ ) + 2 A T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) + A 1 T ( k ) R P k + 1 ( η ) η 2 p ( ξ , η ) d η A 1 ( k ) ] x k + 2 u k T [ S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) ] x k + u k T [ R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) ] u k + E [ x N T ( Q N P N ) x N ] + E [ x 0 T P 0 x 0 ] .
Completing the square for u k , we have
J ( u ) = E k = 0 N 1 { x k T Φ x k + ( u k u k ) T [ R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B k ] ( u k u k ) } + E [ x N T ( Q N P N ) x N ] + E ( x 0 T P 0 x 0 ) ,
where
u k = G k 1 ( ξ ) H k ( ξ ) x k , Φ = Q ( k ) + A T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) P k ( ξ ) + 2 A T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η × A 1 ( k ) + A 1 T ( k ) R P k + 1 ( η ) η 2 p ( ξ , η ) d η A 1 ( k ) H k T ( ξ ) G k 1 ( ξ ) H k ( ξ ) , H k ( ξ ) = S ( k ) T + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η A ( k ) + B T ( k ) R P k + 1 ( η ) η p ( ξ , η ) d η A 1 ( k ) , G k ( ξ ) = R ( k ) + B T ( k ) R P k + 1 ( η ) p ( ξ , η ) d η B ( k ) .
Let P 0 , , P N solve the GDRE (9), and then when u k = u k , we can obtain the optimal cost function
J ( u k ) = inf u 0 , , u N 1 J ( u ) = E [ x 0 T P 0 x 0 ] .
This shows that the optimal value equals E [ x 0 T P 0 x 0 ] and the LQ problem is attainable through the feedback control u k = G k 1 ( ξ ) H k ( ξ ) x k .

4. Examples

Example 1.
Consider the following one-dimensional system:
x k + 1 = a x k + b u k + a 1 x k ω k x 0 R , k = 0 , 1 , , N 1
with the cost function
J ( u ) = k = 0 N 1 [ q x k 2 + 2 s x k u k + r u k 2 ] + x N 2
where the coefficients of a , b , a 1 , q , s , and r take values in R where q > 0 , r > 0 , x k R , and u k R are the state and control, respectively, ω k is a Markovian process with transition probability density p ( ξ , η ) = 3 4 [ 1 ( η ξ ) 2 ] 1 D with initial probability density p 0 ( ξ ) = 3 4 ( 1 ξ 2 ) 1 D 0 , where the set D = { ( ξ , η ) | | η ξ | < 1 , ξ , η R } and D 0 = { ξ | | ξ | < 1 , ξ R } . According to Theorem 2, The Riccati equation of the LQ problem in (13) and (14) is
P k + 1 ( η ) p ( ξ , η ) d η + 2 a a 1 R P k + 1 ( η ) η p ( ξ , η ) d η + a 1 2 R η 2 p ( ξ , η ) d η s k 2 + a b R P k + 1 ( η ) η p ( ξ , η ) d η + a 1 b R P k + 1 ( η ) η p ( ξ , η ) d η 2 r k + b 2 R P k + 1 ( η ) p ( ξ , η ) d η
and the optimal control is given by
u k = s k 2 + a b R P k + 1 ( η ) η p ( ξ , η ) d η + a 1 b R P k + 1 ( η ) η p ( ξ , η ) d η r k + b 2 R P k + 1 ( η ) p ( ξ , η ) d η x k , k = 0 , 1 , , N 1
Figure 1 shows the profile of the Markovian process ω k and the trajectories of P k which is the solution of the Riccati Equation (15) with coefficients a = 0.97 , b = 0.55 , a 1 = 0.2 , q = 0.2 , s = 0.5 , and r = 0.6 . Figure 2 shows the corresponding trajectories of the optimal control u k and the state x k .
Example 2.
In automobile manufacturing industrial design, automobile suspension is an important piece of equipment. Figure 3 shows the hybrid active suspension studied in [44].
m 1 represents the non-spring-loaded mass; m 2 represents the spring-loaded mass; K 1 is the equivalent stiffness of the tire; K 2 is the suspension stiffness; C 2 is the suspension damping; m, k, c, and u are, respectively, the mass block, spring stiffness, damper damping coefficient, and electromagnetic driving force of the electromagnetic reaction force actuator. When the electromagnetic coil is energized, the electromagnetic driving force u is generated. The force u drives the mass block m to vibrate so that the electromagnetic actuator as a whole generates a reaction force F t . q, X 1 , X 2 , and X 3 represent the displacement of the road surface, wheel, body, and electromagnetic actuator mass block, respectively. The external force F t of the electromagnetic actuator acts only on the non-spring-loaded mass m 1 as an active control force. Therefore, as long as the magnitude and direction of the current of the electromagnetic actuator are controlled, the corresponding active control force can be generated to adjust the vibration of the entire vibration system. According to Newton’s second law, the differential Equation (16) of motion of the hybrid active suspension shown in Figure 1 can be obtained
m 1 X ¨ 1 = K 2 ( X 2 X 1 ) + C 2 ( X ˙ 2 X ˙ 1 ) K 1 ( X 1 q ) F t m 2 X ¨ 2 = K 2 ( X 2 X 1 ) C 2 ( X ˙ 2 X ˙ 1 )
where q = q ( t ) is the displacement in the vertical direction of the road surface, and here, it is assumed that q ( t ) satisfies the following model [45]:
q ˙ ( t ) = 2 π G 0 U 0 x 3 ( t ) ω ( t ) ,
where G 0 is the road roughness coefficient and U 0 is the speed of the vehicle. By selecting x 1 = X 2 X 1 , x 2 = X ˙ 2 , x 4 = X ˙ 1 , and x 3 = X 1 q as the state variables and taking x = [ x 1 x 2 x 3 x 4 ] T as the notation, the system is discretized as
J ( u ) = E t = 1 N ( | x ( t ) | 2 + | u ( t ) | 2 )
x ( t + 1 ) = A x ( t ) + B u ( t ) + A 1 x ( t ) ω ( t )
where the coefficient matrices A, B, and A 1 are obtained as follows:
A = 1 Δ t 0 Δ t K 2 Δ t m 2 m 2 C 2 Δ t m 2 0 C 2 Δ t m 2 0 0 1 Δ t K 2 Δ t m 1 C 2 Δ t m 1 K 1 Δ t m 1 m 1 C 2 Δ t m 1 ,
B = 0 0 0 Δ t m 1 , A 1 = 0 0 0 0 0 0 0 0 0 0 Δ t 2 π G 0 U 0 0 0 0 0 0 ,
and Δ t is the time difference between this moment and the previous moment; u ( t ) = F t is the active control force as the control input; ω ( t ) represents the uncertainty input, generally white noise; ω ( t ) is a Markovian process; and its transition probability density function is p ( ξ , η ) = 1 2 π σ e ( η ξ ) 2 2 σ 2 .
J ( u ) represents the output of the LQ metric, and our goal is to find the optimal control u ( t ) that minimizes J ( u ) under the premise that x ( t ) is as small as possible.
According to Theorem 3, we can deduce that the optimal control of the system is
u ( t ) = ( I + B T R P t + 1 ( η ) p ( ξ , η ) d η B ) 1 × [ B T R P t + 1 ( η ) p ( ξ , η ) d η A + B T A 1 R P t + 1 ( η ) η p ( ξ , η ) d η ] x ( t )
where
P t ( ξ ) = I + A T R P t + 1 ( η ) p ( ξ , η ) d η A + A T R P t + 1 ( η ) η A 1 + A 1 T R P t + 1 ( η ) η p ( ξ , η ) d η A + A 1 T R P t + 1 ( η ) η 2 p ( ξ , η ) d η A 1 [ A T R P t + 1 ( η ) p ( ξ , η ) d η B + A 1 T R P t + 1 ( η ) η p ( ξ , η ) d η B ] × ( I + B T R P t + 1 ( η ) p ( ξ , η ) d η B ) 1 × [ B T R P t + 1 ( η ) p ( ξ , η ) d η A + B T A 1 R P t + 1 ( η ) η p ( ξ , η ) d η ] , P N = 0 , t = 0 , , N 1
So, we can obtain
inf u J ( u ) = E [ x T ( 0 ) P 0 x ( 0 ) ] .
According to [44,45], the coefficients of system (19) take the following values: Δ t = 0.1 , k 1 = 30 , k 2 = 159 , c 1 = 0.1 , c 2 = 1.1 , m 1 = 24 , m 2 = 400 , G 0 = 2.2 × 10 4 , and U 0 = 60 . By solving the Riccati Equation (21), the matrix P 0 is obtained as follows:
P 0 = 250.2033 44.4047 2.2024 45.9939 44.4047 93.3853 5.9328 40.0286 2.2024 5.9328 2.7048 6.5274 45.9939 40.0286 6.5274 45.3183 .
Figure 4 illustrates the profiles of the Markovian process with the transport probability density function p ( ξ , η ) and the trajectories of the optimal control u ( t ) in the optimal control problem in (18) and (19), with the optimal cost value J ( u ) = 3.079 .
Figure 5 illustrates the trajectories of the optimal control u ( t ) and the trajectories of x 1 ( t ) , x 2 ( t ) , x 3 ( t ) , and x 4 ( t ) , which are the components of the optimal states x ( t ) of the optimal control problem in (18) and (19), with the initial state x ( 0 ) = [ 0.1 , 0.1 , 0.1 , 0.1 ] T . This shows that under the effect of the optimal control, the fluctuations of the system’s states change in a small range, and less energy from the optimal controller is needed.
In the above two examples, Example 1 covers one-dimensional systems and Example 2 covers four-dimensional systems. These two examples show that the main results of this paper can be used in different dimensional systems. However, in Example 2, the computational simulations focus on a primitive model of a suspension, where the connections with Markov processes are completely artificial. The more demanding applications of these systems need further exploration and research.

5. Conclusions

A type of discrete-time stochastic system driven by general Markov processes, known as Markov-process-driven systems, is proposed to describe more complex noises. By using the properties of the probability distribution of Markov processes, the LQ problem of discrete-time stochastic systems driven by homogeneous Markovian processes is studied. The equivalent relationship between the well-posedness and attainability of the LQ problem, the solvability of the GDRE, and the feasibility of the LMI condition is established. By applying the completing square method to these linear systems, the relationship between the well-posedness of the LQ problem and the LMI condition is obtained: if there exists a series of positive definite matrices satisfying the LMI condition, the LQ problem is well-posed. These results extend the GDRE to the general forms in which the probability distributions of Markov processes are needed to describe the impacts of Markovian properties. In addition, the equivalent relationship between the well-posedness of the LQ problem and the solvability of the GDRE is also obtained, and the necessity of the LMI condition for the well-posedness is also proven. Moreover, by using the properties of Markov processes and the method of completing squares, we have proven that such LQ problems are attainable and optimal state-feedback control can be obtained. Finally, a numerical example and a practical example are used to illustrate the effectiveness and validity of the theory.

Author Contributions

Conceptualization, X.L. and R.Z.; methodology, X.L.; software, X.L. and D.R.; validation, L.S., X.L. and R.Z.; formal analysis, L.S. and X.L.; investigation, W.Z.; resources, X.L.; data curation, X.L.; writing—original draft preparation, X.L. and L.S.; writing—review and editing, X.L. and R.Z.; visualization, W.Z.; supervision, X.L. and W.Z.; project administration, X.L.; funding acquisition, W.Z. and R.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China, grant number 62273212; the Research Fund for the Taishan Scholar Project of Shandong Province of China; the Natural Science Foundation of Shandong Province of China, grant number ZR2020MF062.

Data Availability Statement

Not applicable.

Acknowledgments

We would like to thank the anonymous reviewers for their constructive suggestions to improve the quality of this paper.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kalman, R.E. Contributions to the theory of optimal control. Bol. Soc. Mex. 1960, 5, 102–119. [Google Scholar]
  2. Lewis, F.L. Optimal Control; John Wiley & Sons: New York, NY, USA, 1986. [Google Scholar]
  3. Wonham, W.M. On a matrix Riccati equation of stochastic control. SIAM J. Control. Optim. 1968, 6, 312–326. [Google Scholar] [CrossRef]
  4. Luenberger, D.G. Linear and Nonlinear Programming, 2nd ed.; Addision-Wesley: Reading, MA, USA, 1984. [Google Scholar]
  5. De Souza, C.E.; Fragoso, M.D. On the existence of maximal solution for generalized algebraic Riccati equations arising in stochastic control. Syst. Control. Lett. 1990, 14, 233–239. [Google Scholar] [CrossRef]
  6. Chen, S.P.; Zhou, X.Y. Stochastic linear quadratic regulators with indefinite control weight costs. SIAM J. Control Optim. 2000, 39, 1065–1081. [Google Scholar] [CrossRef]
  7. Rami, M.A.; Zhou, X.Y. Linear matrix inequalities, Riccati equations, and indefinite stochastic linear quadratic controls. IEEE Trans. Autom. Control 2000, 45, 1131–1143. [Google Scholar] [CrossRef]
  8. Yao, D.D.; Zhang, S.Z.; Zhou, X.Y. Stochastic linear quadratic control via semidefinite programming. SIAM J. Control Optim. 2001, 40, 801–823. [Google Scholar] [CrossRef]
  9. Rami, M.A.; Moore, J.B.; Zhou, X.Y. Indefinite stochastic linear quadratic control and generalized differential Riccati equation. SIAM J. Control Optim. 2001, 40, 1296–1311. [Google Scholar] [CrossRef]
  10. Rami, M.A.; Chen, X.; Zhou, X.Y. Discrete-time indefinite LQ control with state and control dependent noises. J. Glob. Optim. 2002, 23, 245–265. [Google Scholar] [CrossRef]
  11. Zhang, W. Study on generalized algebraic Riccati equation and optimal regulators. Control. Theory Appl. 2003, 20, 637–640. [Google Scholar]
  12. Zhang, W.; Chen, B.S. On stabilizability and exact observability of stochastic systems with their applications. Automatica 2004, 40, 87–94. [Google Scholar] [CrossRef]
  13. Huang, Y.; Zhang, W.; Zhang, H. Infinite horizon LQ optimal control for discrete-time stochastic systems. Asian J. Control 2008, 10, 608–615. [Google Scholar] [CrossRef]
  14. Li, G.; Zhang, W. Discrete-time indefinite stochastic linear quadratic optimal control: Inequality constraint case. In Proceedings of the 32nd Chinese Control Conference, Xi’an, China, 26 July 2013; pp. 2327–2332. [Google Scholar]
  15. Huang, H.; Wang, X. LQ stochastic optimal control of forward-backward stochastic control system driven by Lévy process. In Proceedings of the IEEE Advanced Information Management, Communicates, Electronic and Automation Control Conference, Xi’an, China, 3 October 2016; pp. 1939–1943. [Google Scholar]
  16. Tan, C.; Zhang, H.; Wong, W. Delay-dependent algebraic Riccati equation to stabilization of networked control systems: Continuous-time case. IEEE Trans. Cybern. 2018, 48, 2783–2794. [Google Scholar] [CrossRef]
  17. Tan, C.; Yang, L.; Zhang, F.; Zhang, Z.; Wong, W. Stabilization of discrete time stochastic system with input delay and control dependent noise. Syst. Control Lett. 2019, 123, 62–68. [Google Scholar] [CrossRef]
  18. Zhang, T.; Deng, F.; Sun, Y.; Shi, P. Fault estimation and fault-tolerant control for linear discrete time-varying stochastic systems. Sci. China Inf. Sci. 2021, 64, 200201. [Google Scholar] [CrossRef]
  19. Jiang, X.; Zhao, D. Event-triggered fault detection for nonlinear discrete-time switched stochastic systems: A convex function method. Sci. China Inf. Sci. 2021, 64, 200204. [Google Scholar] [CrossRef]
  20. Dashtdar, M.; Rubanenko, O.; Rubanenko, O.; Hosseinimoghadam, S.M.S.; Belkhier, Y.; Baiai, M. Improving the Differential Protection of Power Transformers Based on Fuzzy Systems. In Proceedings of the 2021 IEEE 2nd KhPI Week on Advanced Technology (KhPIWeek), Kharkiv, Ukraine, 13 September 2021; pp. 16–21. [Google Scholar]
  21. Belkhier, Y.; Nath Shaw, R.; Bures, M.; Islam, M.R.; Bajaj, M.; Albalawi, F.; Alqurashi, A.; Ghoneim, S.S.M. Robust interconnection and damping assignment energy-based control for a permanent magnet synchronous motor using high order sliding mode approach and nonlinear observer. Energy Rep. 2022, 8, 1731–1740. [Google Scholar] [CrossRef]
  22. Djouadi, H.; Ouari, K.; Belkhier, Y.; Lehouche, H.; Ibaouene, C.; Bajaj, M.; AboRas, K.M.; Khan, B.; Kamel, S. Non-linear multivariable permanent magnet synchronous machine control: A robust non-linear generalized predictive controller approach. IET Control Theory Appl. 2023, 2023, 1–15. [Google Scholar] [CrossRef]
  23. Lin, X.; Zhang, T.; Zhang, W.; Chen, B.S. New Approach to General Nonlinear Discrete-Time Stochastic H Control. IEEE Trans. Autom. Control 2019, 64, 1472–1486. [Google Scholar] [CrossRef]
  24. Lv, Q. Well-posedness of stochastic Riccati equations and closed-loop solvability for stochastic linear quadratic optimal control problems. J. Differ. Equ. 2019, 267, 180–227. [Google Scholar]
  25. Tang, C.; Li, X.Q.; Huang, T.M. Solvability for indefinite mean-field stochastic linear quadratic optimal control with random jumps and its applications. Optim. Control Appl. Methods 2020, 41, 2320–2348. [Google Scholar] [CrossRef]
  26. Chen, X.; Zhu, Y.G. Multistage uncertain random linear quadratic optimal control. J. Syst. Sci. Complex. 2020, 33, 1–26. [Google Scholar] [CrossRef]
  27. Zhang, W.; Zhang, L.Q. A BSDE approach to stochastic linear quadratic control problem. Optim. Control Appl. Methods 2021, 42, 1206–1224. [Google Scholar] [CrossRef]
  28. Meng, W.J.; Shi, J.T. Linear quadratic optimal control problems of delayed backward stochastic differential equations. Appl. Math. Optim. 2021, 84, 1–37. [Google Scholar] [CrossRef]
  29. Li, Y.B.; Wahlberg, B.; Hu, X.M. Identifiability and solvability in inverse linear quadratic optimal control problems. J. Syst. Sci. Complex. 2021, 34, 1840–1857. [Google Scholar] [CrossRef]
  30. Li, Y.C.; Ma, S.P. Finite and infinite horizon indefinite linear quadratic optimal control for discrete-time singular Markov jump systems. J. Frankl. Inst. 2021, 358, 8993–9022. [Google Scholar]
  31. Tan, C.; Zhang, S.; Wong, W.; Zhang, Z. Feedback stabilization of uncertain networked control systems over delayed and fading channels. IEEE Trans. Control Netw. Syst. 2021, 8, 260–268. [Google Scholar] [CrossRef]
  32. Tan, C.; Yang, L.; Wong, W. Learning based control policy and regret analysis for online quadratic optimization with asymmetric information structure. IEEE Trans. Cybern. 2022, 52, 4797–4810. [Google Scholar] [CrossRef]
  33. Bolzern, P.; Colaneri, P.; Nicolao, G.D. Almost sure stability of Markov jump linear systems with deterministic switching. IEEE Trans. Autom. Control 2013, 58, 209–214. [Google Scholar] [CrossRef]
  34. Dong, S.; Chen, G.; Liu, M.; Wu, Z.G. Cooperative adaptive H output regulation of continuous-time heterogeneous multi-agent Markov jump systems. IEEE Trans. Circuits Syst. II Express Briefs 2021, 68, 3261–3265. [Google Scholar]
  35. Wu, X.; Shi, P.; Tang, Y.; Mao, S.; Qian, F. Stability analysis of aemi-Markov jump atochastic nonlinear systems. IEEE Trans. Autom. Control 2022, 67, 2084–2091. [Google Scholar] [CrossRef]
  36. Øksendal, B. Stochastic Differential Equations: An Introduction with Applications; Springer: New York, NY, USA, 2005. [Google Scholar]
  37. Bertoin, J. Lévy Processes; Cambridge University Process: New York, NY, USA, 1996. [Google Scholar]
  38. Han, Y.; Li, Z. Maximum Principle of Discrete Stochastic Control System Driven by Both Fractional Noise and White Noise. Discret. Dyn. Nat. Soc. 2020, 2020, 1959050. [Google Scholar] [CrossRef]
  39. Ni, Y.H.; Li, X.; Zhang, J.F. Mean-field stochastic linear-quadratic optimal control with Markov jump parameters. Syst. Control Lett. 2016, 93, 69–76. [Google Scholar] [CrossRef]
  40. Rami, M.A.; Chen, X.; Moore, J.B.; Zhou, X.Y. Solvability and asymptotic behavior of generalized Riccati equations arising in indefinite stochastic LQ controls. IEEE Trans. Autom. Control 2001, 46, 428–440. [Google Scholar] [CrossRef]
  41. Albert, A. Conditions for positive and nonnegative definiteness in terms of pseudo-inverse. SIAM J. Appl. Math. 1969, 17, 434–440. [Google Scholar] [CrossRef]
  42. Yu, X.; Yin, J.; Khoo, S. Generalized Lyapunov criteria on finite-time stability of stochastic nonlinear systems. Automatica 2019, 107, 183–189. [Google Scholar] [CrossRef]
  43. Yin, J.; Khoo, S.; Man, Z.; Yu, X. Finite-time stability and instability of stochastic nonlinear systems. Automatica 2011, 47, 2671–2677. [Google Scholar] [CrossRef]
  44. Bu, X.F.; Xie, Y.H. Study on characteristics of electromagnetic hybrid active vehicle suspension based on mixed H2/H control. J. Manuf. Autom. 2018, 40, 129–133. [Google Scholar]
  45. Chen, M.; Long, H.Y.; Ju, L.Y.; Li, Y.G. Stochastic road roughness modeling and simulation in time domain. Mech. Eng. Autom. Chin. 2017, 201, 40–41. [Google Scholar]
Figure 1. Profiles of Markov process ω k and solutions of P k for the LQ problem in (13) and (14).
Figure 1. Profiles of Markov process ω k and solutions of P k for the LQ problem in (13) and (14).
Processes 11 02933 g001
Figure 2. Trajectories of u k and x k for the LQ problem in (13) and (14).
Figure 2. Trajectories of u k and x k for the LQ problem in (13) and (14).
Processes 11 02933 g002
Figure 3. Schematic diagram of hybrid active suspension mechanics.
Figure 3. Schematic diagram of hybrid active suspension mechanics.
Processes 11 02933 g003
Figure 4. Profiles of ω ( t ) and the trajectories of the optimal control u ( t ) of system (19).
Figure 4. Profiles of ω ( t ) and the trajectories of the optimal control u ( t ) of system (19).
Processes 11 02933 g004
Figure 5. Trajectories of the control u ( t ) of system (19) and the corresponding components of state x ( t ) = [ x 1 ( t ) , x 2 ( t ) , x 3 ( t ) , x 4 ( t ) ] .
Figure 5. Trajectories of the control u ( t ) of system (19) and the corresponding components of state x ( t ) = [ x 1 ( t ) , x 2 ( t ) , x 3 ( t ) , x 4 ( t ) ] .
Processes 11 02933 g005
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lin, X.; Song, L.; Rong, D.; Zhang, R.; Zhang, W. Linear Quadratic Optimal Control of Discrete-Time Stochastic Systems Driven by Homogeneous Markov Processes. Processes 2023, 11, 2933. https://doi.org/10.3390/pr11102933

AMA Style

Lin X, Song L, Rong D, Zhang R, Zhang W. Linear Quadratic Optimal Control of Discrete-Time Stochastic Systems Driven by Homogeneous Markov Processes. Processes. 2023; 11(10):2933. https://doi.org/10.3390/pr11102933

Chicago/Turabian Style

Lin, Xiangyun, Lifeng Song, Dehu Rong, Rui Zhang, and Weihai Zhang. 2023. "Linear Quadratic Optimal Control of Discrete-Time Stochastic Systems Driven by Homogeneous Markov Processes" Processes 11, no. 10: 2933. https://doi.org/10.3390/pr11102933

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop