Abstract
In this paper, we introduced controlled discrete-time semi-Markov random evolutions. These processes are random evolutions of discrete-time semi-Markov processes where we consider a control. applied to the values of random evolution. The main results concern time-rescaled weak convergence limit theorems in a Banach space of the above stochastic systems as averaging and diffusion approximation. The applications are given to the controlled additive functionals, controlled geometric Markov renewal processes, and controlled dynamical systems. We provide dynamical principles for discrete-time dynamical systems such as controlled additive functionals and controlled geometric Markov renewal processes. We also produce dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as controlled additive functionals, controlled geometric Markov renewal processes and controlled dynamical systems. As an example, we consider the solution of portfolio optimization problem by Merton for the limiting controlled geometric Markov renewal processes in diffusion approximation scheme. The rates of convergence in the limit theorems are also presented.
Keywords:
semi-Markov chain; controlled discrete-time semi-Markov random evolutions; averaging; diffusion approximation; diffusion approximation with equilibrium; rates of convergence; controlled additive functional; controlled dynamical systems; controlled geometric Markov renewal processes; HJB equation; Merton problem; Banach space 1. Introduction
Random evolutions were introduced over 40 years ago, see, e.g., in [1,2] and for its asymptotic theory in [3,4,5,6] and references therein. Discrete-time random evolutions, induced by discrete-time Markov chains, are introduced by Cohen [7] and Keepler [8], and discrete-time semi-Markov random evolutions (DTSMRE) by Limnios [9]. See also [10]. Koroliuk and Swishchuk [4], Swishchuk and Wu [5], Anisimov [11,12,13,14], and Koroliuk and Limnios [3], studied discrete-time random evolutions induced by the embedded Markov chains of continuous time semi-Markov processes. This is equivalent to discrete-time Markov random evolution stopped at random time (continuous). One of the examples of discrete-time random evolutions is the geometric Markov renewal process (GMRP). Applications of GMRP in finance have been considered in [15,16,17]. Optimal stopping of GMRP and pricing of European and American options for underlying assets modelled by GMRP have been studied in [18].
Discrete-time semi-Markov chains (SMC) have only recently been used in applications. Especially, in DNA analysis, image and speech processing, reliability theory, etc., see in [19] and references therein. These applications have stimulated a research effort in this area. While the literature in discrete-time Markov chains theory and applications is quite extensive, there is only a small amount of the literature on SMC and most of them are related to hidden semi-Markov models for estimation.
The present article is a continuation of our previous work [20]. Thus, we keep all our notation and definitions the same as in the latter paper. Compared with our previous work [20], where we studied random evolutions of semi-Markov chains, here we considered additionally a control on the random evolution, which we call controlled discrete-time semi-Markov random evolution (CDTSMRE) in a Banach space, and we presented time-rescaled convergence theorems. In particular, we get weak convergence theorems in Skorokhod space for càdlàg stochastic processes, see, e.g., in [21]. The limit theorems include averaging, diffusion approximation, and diffusion approximation with equilibrium. For the above limit theorems we also presented rates of convergence results. Finally, we give some applications regarding the above mentioned results, especially to controlled additive functionals (CAF), CGMRP, and controlled dynamical systems (CDS), and optimization problems.
Regarding the optimization problems, we provide dynamical principles for discrete-time dynamical systems such as CAF and CGMRPs (see Section 2.4), see, e.g., [22,23,24]. We also produce dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as CAF, CGMRP, and CDS. As an example, we consider the solution of portfolio optimization problem by Merton for the limiting CGMRP in DA (see Section 4.4). Merton problem, or Merton portfolio’s problem, is a problem in continuous-time finance associated with portfolio choice. In -security market, which consists of a stock and a risk-free asset, an investor must choose how much to consume, and must allocate his wealth between the stock and the risk-free asset in a such way that maximizes expected utility. The problem was formulated and first solved by Robert Merton in 1969, and published in 1971 [25].
Results presented here are new and deals with CDTSMRE on Banach spaces. This paper contains new and original results on dynamical principle for CDTSMRE and DPE (HJB equations) for the limiting processes in DA. One of the new remarkable results is the solution of Merton portfolio problem for the limiting CGMRP in DA. The method of proofs was based on the martingale approach together with convergence of transition operators of the extended semi-Markov chain via a solution of a singular perturbation problem [3,4,26]. As in our previous work [20], the tightness of these processes is proved via Sobolev’s embedding theorems [27,28,29]. It is worth mentioning that, as in the Markov case, the results presented here cannot be deduced directly from the continuous-time case. We should also note that that DTSMREs have been completely studied in [20]. For semi-Markov processes see, e.g., [30,31,32,33]. For Markov chains and additive functionals see, e.g., [34,35,36,37,38].
The paper is organized as follows. Definition and properties of discrete-time semi-Markov random evolutions and Controlled DTSMREs, as well as particular stochastic systems as applications, are introduced in Section 2. The main results of this paper, limit theorems of CDTSMRE, as averaging, diffusion approximation and diffusion approximation with equilibrium of controlled DTSMREs are considered in Section 3. In Section 4, we provide three applications of averaging, diffusion approximation, and diffusion approximation with equilibrium of controlled DTSMREs: controlled additive functionals, controlled GMRP, and controlled dynamical systems. Section 5 deals with the analysis of the rates of convergence in the limit theorems, presented in the previous sections, for controlled DTSMREs and for CAF and CGMRP. In Section 6, we give the proofs of theorems presented in the previous sections. The last section concludes the paper and indicates some future works.
2. Controlled Discrete-Time Semi-Markov Random Evolutions
2.1. Semi-Markov Chains
The aim of this section is to present some notation and to make this paper as autonomous as possible. The reader may refer to our article in [20] for more details.
Let be a measurable space with countably generated -algebra and be a stochastic basis on which we consider a Markov renewal process in discrete time , with state space . Notice that is the set of non-negative integer numbers. The semi-Markov kernel q is defined by (see, e.g., in [9,19]),
We will denote also , where . The process is the embedded Markov chain of the MRP with transition kernel . The semi-Markov kernel q is written as
where , the conditional distribution of the sojourn time in state x given that the next visited state is y.
Define also the counting process of jumps , and the discrete-time semi-Markov chain by , for . Define now the backward recurrence time process , , and the filtration , .
Let us consider a separable Banach space B of real-valued measurable functions defined on , endowed with the sup norm and denote by its Borel -algebra. The Markov chain , has the following transition probability operator on B
where , and its stationary distribution, if there exist, is given by
where
and is the stationary distribution of the EMC , , and . The probability measure defined by is the stationary probability of the SMC . Define also the r-th moment of holding time in state ,
Of course, , for any .
Define now the stationary projection operator on the null space of the (discrete) generating operator ,
where for any , and . This operator satisfies the equations
The potential operator of , denoted by , is defined by
2.2. General Definition and Properties of DTSMREs
We define here controlled discrete-time semi-Markov random evolutions. Let U denote a compact Polish space representing the control, and let be U-valued control process and we suppose that it is a Markov chain. We note that we could also define the process which is a semi-Markov control process, considered in many papers (see, e.g., in [39,40]). We suppose that homogeneous Markov chain is independent of , and transition probability kernel .
Let us consider a family of bounded contraction operators , defined on B, where the maps are -measurable, . Denote by I the identity operator on B. Let be the null space, and be the range values space of operator . We will suppose here that the Markov chain is uniformly ergodic, that is, , as , for any . In that case, the transition operator is reducible-invertible on B. Thus, we have , the direct sum of the two subspaces. The domain of an operator A on B is denoted by .
Definition 1.
A controlled discrete-time semi-Markov random evolution (CDTSMRE) , on the Banach space B, is defined by
for , and for any . Thus we have
The process is a Markov chain on , adapted to the filtration , . We also note that is a Markov chain on with discrete generator
where , and
The process defined by
on B, is an -martingale. The random evolution can be written as follows
and then, the martingale (5) can be written as follows,
or
Finally, as , one takes
2.3. Some Examples
Example 1.
Controlled Additive Functional or Markov Decision Process.
Let define the following controlled additive functional,
If we define the operator on in the following way,
then the controlled discrete-time semi-Markov random evolution has the following presentation,
Process is usually called in the literature the Markov decision process (see, e.g., in [41,42,43,44]).
Example 2.
Controlled geometric Markov renewal process.
The CGMRP is defined in the following way,
We suppose that
If we define the operator on in the following way,
then the controlled discrete-time semi-Markov random evolution can be given as follows,
To the authors opinion, this process is defined for the first time in the literature and the notion of controlled GMRP is a new one as well.
2.4. Dynamic Programming for Controlled Models
Here, we present dynamic programming for controlled models given in Examples in previous section. Let us consider a Markov control model (see in [45]) Here, E is the state space; A is the control or action set; Q is the transition kernel, i.e., a stochastic kernel on E given where ; and is a measurable function called the cost-per-stage function.
We are interested in is to minimize the finite-horizon performance criterion either (see Example 1)
or (see Example 2)
where is the terminal cost function, is the set of control policies.
In this way, denoting by the value function
the problem is to find a policy such that
Example 3.
Controlled Additive Functional.
Let us provide an algorithm for finding both the value function and an optimal policy for the example with function (see Example 1).
Let be the functions on E defined from to by (backwards)
and
Suppose that there is a selector such that attains the minimum in the above expression for for all meaning for any and
Then, the optimal policy is the deterministic Markov one , and the value function equals i.e.,
Example 4.
Controlled Geometric Markov Renewal Chain.
Let us provide an algorithm for finding both the value function and an optimal policy for the example with function (see Example 2). We will modify the expression for in Example 2. Let be a log-return, then
Thus, we are interested in minimizing the finite-horizon performance criterion for
Let be the functions on E defined from to by (backwards)
and
Suppose that there is a selector such that attains the minimum in the above expression for for all meaning for any and
Then, the deterministic Markov policy is optimal, and the value function equals i.e.,
3. Limit Theorems for Controlled Semi-Markov Random Evolutions
In this section, we present averaging, diffusion approximation, and diffusion approximation with equilibrium results for the controlled discrete-time semi-Markov random evolutions. It is worth noticing that the main scheme of results are almost the same as in our previous works in particular [20]. Nevertheless, the additional component of the control allows us to study more interesting problems.
3.1. Averaging of CDTSMREs
We consider here CDTSMREs defined in Section 2. Let us now set and consider the continuous time process
We will prove here asymptotic results for this process as .
The following assumptions are needed for averaging.
- A1:
- The MC is uniformly ergodic with ergodic distribution .
- A2:
- The moments , are uniformly integrable.
- A3:
- The perturbed operators have the following representation on Bwhere operators on B are closed and is dense in B, . Operators are negligible, i.e., for any .
- A4:
- We have (See A7.)
- A5:
- There exists Hilbert spaces H and such that compactly embedded in Banach spaces B and respectively, where is a dual space to
- A6:
- Operators and are contractive on Hilbert spaces H and respectively.
- A7:
- The MC , is independent of , and is uniformly ergodic with stationary distribution .
We note that if then is a Sobolev space, and and this embedding is compact (see [29]). For the spaces and the situation is the same.
We also note, that semi-Markov chain is uniformly ergodic on with stationary probabilities which follows from conditions A1 and A7.
Theorem 1.
Under Assumptions A1–A7, the following weak convergence takes place,
where the limit random evolution is determined by the following equation,
or, equivalently,
where the limit contracted operator is then given by
This result generalize the classical Krylov–Bogolyubov averaging principle [46] on a Banach and a controlled spaces.
3.2. Diffusion Approximation of DTSMREs
For the diffusion approximation of CDTSMREs, we will consider a different time-scaling and some additional assumptions.
- D1:
- Let us assume that the perturbed operators have the following representation in B,where operators on B are closed and is dense in B, ; operators are a negligible operator, i.e., .
- D2:
- The following balance condition holds,where
- D3:
- The moments , are uniformly integrable.
Theorem 2.
Under Assumptions A1, A5–A7 (see Section 3.1), and D1-D3, the following weak convergence takes place,
where the limit random evolution is a diffusion random evolution determined by the following generator
where
and
3.3. Diffusion Approximation with Equilibrium
The diffusion approximation with equilibrium or the normal deviation is obtained by considering the difference between the rescaled initial processes and the averaging limit process. This is of great interest when we have no balance condition as previously in the standard diffusion approximation scheme.
Consider now the controlled discrete-time semi-Markov random evolution averaged evolution (see Section 3.1) and the deviated evolution
Theorem 3.
Under Assumptions A1, A5–A6 (see Section 3.1), and D3, with operators in A3, instead of D1, the deviated controlled semi-Markov random evolution weakly convergence, when , to the diffusion random evolution defined by the following generator
where Π is defined in (9).
4. Applications to Stochastic Systems
In this section, we give two applications in connection with the above results: additive functionals that has many application, e.g., in storage, reliability, and risk theories (see, e.g., in [3,4,19,47]), and to geometric Markov renewal processes, that also have many application including finance (see [15,16,17,18]). Our main goal here is to get the limiting processes and apply optimal control methods to receive the solutions of optimization problems. The limiting results for MC such as LLN and CLT were considered in [11,12].
4.1. Controlled Additive Functionals
Let us consider here the CAF, , described previously in Example 1.
Averaging of CAF. Now, if we define the continuous time process
then from Theorem 1 it follows that this process has the following limit
where We suppose that
Diffusion Approximation of CAF. If we consider the continuous time process as follows
then under balance condition and we get that the limit process has the following form,
where , and
and is a standard Wiener process.
Diffusion Approximation with Equilibrium of CAF. Let us consider the following normalized additive functional,
Then, this process converges to the following process, where
and is a standard Wiener process.
In this way, the AF may be presented in the following approximated form,
4.2. Controlled Geometric Markov Renewal Processes
The CGMRP is defined in the following way (see in [15,16]),
We suppose that
If we define the operator on in the following way,
then the discrete-time semi-Markov random evolution has the following presentation,
Averaging of CGMRP. Now, define the following sequence of processes,
Then, under averaging conditions the limit process has the following form,
where
Diffusion Approximation of CGMRP. If we define the following sequence of processes,
then, in the diffusion approximation scheme, we have the following limit process,
where
It means that satisfies the following stochastic differential equation,
where is a standard Wiener process.
Diffusion Approximation with Equilibrium of CGMRP. Let us consider the following normalized GMRP:
It is worth noticing that in finance the expression represents the log-return of the underlying asset (e.g., stock)
Then, this process converges to the following process, where
and is a standard Wiener process.
In this way, the GMRP may be presented in the following approximated form,
4.3. Controlled Dynamical Systems
We consider here discrete-time CDS and their asymptotic behaviour in series scheme: average and diffusion approximation ([9]).
Define the measurable function C on . Let us consider the difference equation
switched by the SMC .
The perturbed operators , are defined now by
Averaging of CDS. Under averaging assumptions the following weak convergence takes place,
where is the solution of the following (deterministic) differential equation,
where .
Diffusion Approximation of CDS. Under diffusion approximation conditions the following weak convergence takes place
where , is a diffusion processes, with initial value , determined by the operator
provided that , and drift and diffusion coefficients are defined as follows,
with:
, ,
, where means transpose of the vector C,
, ,
, .
4.4. The Dynamic Programming Equations for Limiting Models in Diffusion Approximation
In this section, we consider the DPE, i.e., HJB Equations, for the limiting models in DA from Section 4.1, Section 4.2 andSection 4.3. As long as all limiting processes in DA in Section 4.1, Section 4.2 and Section 4.3 are diffusion processes, then we will set up a general approach to control for diffusion processes, see in [48].
Let be a diffusion process satisfying the following stochastic differential equation,
where is the control process, is a standard Wiener process. Let us also introduce the following performance criterion function,
where is a terminal reward function (uniformly bounded), is a running penalty/reward function (uniformly bounded), The problem is to maximize this performance criteria, i.e., to find the value function
where is the admissible set of strategies/controls which are -predictable, non-negative, and bounded.
The Dynamic Programming Principle (DPP) for diffusions states that the value function satisfies the DPP
for all
Moreover, the value function above satisfies the Dynamic Programming Equation (DPE) or Hamilton–Jacobi–Bellman (HJB) Equation:
where is an infinitesimal generator of the diffusion process above, i.e.,
•DPE/HJB Equation for the Limiting CAF in DA (see Section 4.1)
We remind that the limiting process in this case has the following form
where and
and is a standard Wiener process.
•DPE/HJB Equation for the Limiting CGMRP in DA (see Section 4.2)
We recall that we have the following limiting process in this case:
where
Furthermore, satisfies the following stochastic differential equation (SDE),
where is a standard Wiener process.
In this case, the DPE or HJB Equation (16) reads with the generator
and
•DPE/HJB Equation for the Limiting CDS in DA (see Section 4.3)
We remind that in the diffusion approximation the limiting process is a diffusion process with a generator
provided that , and drift and diffusion coefficients are defined as follows,
with
, ,
, where means transpose of the vector C,
, ,
, .
, ,
, where means transpose of the vector C,
, ,
, .
Remark 1.
Our construction here is equivalent to some extend to “Recurrent Processes of a semi-Markov type (RPSM)” studied first in [13,14] including limit theorems. Those results were described in more detail in [11,12]. In particular, “RPSM with Markov switching” reflects the case of independent Markov components and , and “General case of RPSM” reflects the case when is dependent on .
•The Merton Problem
This is an example of solution of DPE/HJB equation for the limiting CGMRP in DA. Let us consider the portfolio optimization problem proposed by Merton (1971), see in [25]. We will apply this approach to the limiting CGMRP in DA above. In this problem, the agent seeks to maximize expected wealth by trading in a risky asset and the risk-free bonds (or bank account). She/he places for a total wealth in the risky asset and looks to obtain the value function (performance criterion)
which depends on the current wealth x and asset price and the optimal trading strategy is the agent’s utility function (e.g., exponential or power ). We suppose that the asset price satisfies the following SDE
where
Here, represents the expected continuously compounded rate of growth of the traded asset, r is the continuously compounded rate of return of the risk-free asset (bond or bank account).
The wealth process follows the following SDE,
From the SDEs for and for above we conclude that the infinitesimal generator for the pair is
From HJB equation for the limiting CGRMP in DA it follows that the value function
should satisfy the equation
with terminal condition
The explicit solution of this PDE depends on the explicit form of the utility function Let us take the exponential utility function
In this case we can find that the optimal amount to invest in the risky asset is a deterministic function of time
5. Rates of Convergence in Averaging and Diffusion Approximations
The rate of convergence in a limit theorem is important in several ways, both theoretical and practical. We present here the rates of convergence of CDTSMRE in the averaging, diffusion approximation and diffusion approximation with equilibrium schemes and, as corollaries, we give the rates of convergence for CAF and CGMRP in the corresponding limits.
Proposition 1.
The Rate of Convergence of CDTSMRE in the Averaging has the following form,
where is a constant, and
The proof of this proposition is given in Section 6.4.
Proposition 2.
The Rate of Convergence of CDTSMRE in the Diffusion Approximation takes the following form,
where is a constant, and
Proposition 3.
The Rate of Convergence of CDTSMRE in Diffusion Approximation with Equilibrium has the following form,
where is a constant and
The proofs of the above Propositions 2 and 3 are similar as the proof of Proposition 1. We give in what follows some rate of convergence results (Corollaries 1 and 2) concerning applications.
Corollary 1.
The Rate of Convergence in the Limit Theorems for CAF:
- Rate of Convergence in Averaging:
where is a constant, and
- Rate of Convergence in Diffusion Approximation
where is a constant, and
- Rate of Convergence in diffusion approximation with equilibrium for CAF
where is a constant, and
Corollary 2.
The Rate of Convergence in the Limit Theorems for CGMRP:
- Rate of Convergence in Averaging
where is a constant, and
- Rate of Convergence in Diffusion Approximation
where is a constant, and
- Rate of Convergence in diffusion approximation with equilibrium
where is a constant, and
6. Proofs
The proofs here have almost the same general construction scheme as in our paper [20] except that we consider also the control process. Let be the space of B-valued continuous functions defined on .
6.1. Proof of Theorem 1
The proof of the relative compactness of CDTSMRE in the average approximation is based on the following four lemmas.
The CDTSMRE , see (3), is weakly compact in with limit points into .
Lemma 1.
Under Assumptions A1–A7, the limit points of , , as , belong to .
Proof.
Assumptions A5–A6 imply that the discrete-time semi-Markov random evolution is a contractive operator in H and, therefore, is a supermartingale for any where is a norm in Hilbert space H ([4,9]) Obviously, the same properties satisfy the following family Using Doob’s inequality for the supermartingale we obtain
where is a compact set in B and is any small number. It means that sequence is tight in Taking into account conditions A1–A6, we obtain that discrete-time semi-Markov random evolution is weakly compact in with limit points in
Let , and let be a compact set from compact containment condition . It is sufficient to show that weakly converges to zero. This is equivalent to the convergence of in probability as .
From the very definition of and A3, we obtain
where is the indicator of the set , and is the finite -set for . Then, for , we have
where , and
.
It is worth noticing that the operator is bounded when . So is the case for when .
Taking both and go to 0 we obtain the proof of the this lemma. □
Let us now consider the continuous time martingale
Lemma 2.
The process
is an -martingale.
Proof.
As long as , is a martingale is an -martingale. Here, we have which can be easily checked. □
Lemma 3.
The family is relatively compact for all , dual of the space .
Proof.
Let
Then,
As long as we obtain
Then,
as is bounded for any .
It means that the family , is relatively compact for any . □
Lemma 4.
The family is relatively compact for any , and any .
Proof.
It is worth noticing that the martingale can be represented in the form of the martingale differences
Then, using the equality
we get
for any . Now, from the above, we get
which proves the lemma. □
Now the proof of Theorem 1 is achieved as follows.
From Lemmas 2–4 and the representation (17) it follows that the family is relatively compact for any , and any .
Moreover, let , , be a family of perturbed operators defined on B as follows,
Then, the process
is an -martingale.
The following singular perturbation problem, for the non-negligible part of compensating operator, , denoted by ,
on the test functions , has the solution (see [3] Proposition 5.1): , , with , , and .
The limit operator is then given by
form which we get the contracted limit operator
We note that martingale has the following asymptotic representation,
where , as . The families and are weakly compact for all in a dense subset It means that family is also weakly compact. In this way, the sum converges, as , to the integral The quadratic variation of the martingale tends to zero when thus, when for any and for any Passing to the limit in (23), when , we get where is defined in (6).
The quadratic variation of the martingale , in the average approximation, is
where . Hence
and
Therefore,
Now, from (24) and (26) and from boundedness of all operators in (26) with respect to , it follows that goes to 0 when , and the quadratic variation of limit process , for the martingale , is equals to 0.
In this case, the limit martingale equals to 0. Therefore, the limit equation for has the form (6). As long as the solution of the martingale problem for operator is unique, then it follows that the solution of the Equation (6) is unique as well [49,50]. It is worth noticing that operator is a first order operator (, see (22)). Finally, the operator generates a semigroup, then and the latter representation is unique.
6.2. Proof of Theorem 2
We can prove the relative compactness of the family exactly on the same way, and following the same steps as above. However, in the case of diffusion approximation the limit continuous martingale for the martingale has quadratic variation that is not zero, that is,
and so , for .
Moreover, operator defined in Theorem 2 is a second-order kind operator as it contains operator and , compare with the first-order operator in (7).
Let , , be a family of perturbed operators defined on B as follows,
Then, the process
is an -martingale with mean value zero.
For the non-negligible part of compensating operator, , denoted by , consider the following singular perturbation problem,
where . The solution of this problem is realized by the vectors (see in [3], Proposition 5.2)
with . Finally, the negligible term is
Of course, .
Now the limit operator is given by
from which, the contracted operator on the null space is
Moreover, due to the balance condition (8) we get the limit operator.
We worth noticing that Assumptions A5–A7 and D1–D3 imply that discrete-time semi-Markov random evolution is a contractive operator in H and, therefore, is a supermartingale for any where is a norm in Hilbert space H ([4,9]). By Doob’s inequality for the supermartingale we obtain
where is a compact set in B and is any positive small real number.
We conclude that under Assumptions A5–A7 and D1–D3, the family is tight and is weakly compact in with limit points in
Moreover, under Assumptions A5–A6 and D1–D2, the martingale has the following asymptotic presentation:
where , as . The families and are weakly compact for all and It means that is also weakly compact and has a limit.
Let us denote the previous limit by , then the sum converges to the integral Let also be a limit martingale for when Then, from the previous steps and (32), we obtain
As long as martingale has mean value zero, the martingale has also mean value zero. If we take the mean value from both parts of (33) we get
or, solving it, we get
The last equality means that the operator generates a semigroup, namely, Now, the uniqueness of the limit evolution in diffusion approximation approximation follows from the uniqueness of solution of the martingale problem for (uniqueness of the limit process under weak compactness). As long as the solution of the martingale problem for operator is unique, then it follows that the solution of the Equation (34) is unique as well [49,50].
6.3. Proof of Theorem 3
We note that in (12) has the following presentation,
As the balance condition holds, then we apply the diffusion approximation algorithm (see Section 3.2), i.e., to the right-hand side of (36) with the following operators, and instead of It is worth mentioning that the family is weakly compact and the result is proved (see Section 6.1 and Section 6.2).
6.4. Proof of Proposition 1
The proof of this proposition is based on the estimation of
for any , where .
We note that
As long as , , Equation (37) has the solution in domain , .
In this way,
where is a potential operator of .
From here we obtain
as are contractive operators.
We note also that
where , . This follows from standard argument about the convergence of Riemann sums in Bochner integral (see Lemma 4.14, p. 161, [4]).
We note that
where we applied representation .
We also note that satisfies the equation
Let us introduce the following martingale,
This is of zero mean-value martingale
which comes directly from (43).
Again, from (43), we get the following asymptotic representation
where , as , for any .
Now, from Equation (6) and expressions (44) and (45), we obtain the following representation
where .
Let us estimate in (46).
where is a compact set, ,because satisfies compactness condition for any and any k.
In this way, we get from (46) that
Finally, from inequalities (38)–(41) and from (47)–(48), we obtain the desired rate of convergence of the CDTSMRE in averaging scheme
where the constant
and is defined in (48). Therefore, the proof of Proposition 1 is done.
Remark 2.
In a similar way, we can obtain the rate of convergence results in diffusion approximation (see Propositions 2–3).
7. Concluding Remarks and Future Work
In this paper, we introduced controlled semi-Markov random evolutions in discrete-time in Banach space. The main results concerned time-rescaled limit theorems, namely, averaging, diffusion approximation, and diffusion approximation with equilibrium by martingale weak convergence method. We applied these results to various important families of stochastic systems, i.e., the controlled additive functionals, controlled geometric Markov renewal processes, and controlled dynamical systems. We provided dynamical principles for discrete-time dynamical systems such as controlled additive functionals and controlled geometric Markov renewal processes. We also produced dynamic programming equations (Hamilton–Jacobi–Bellman equations) for the limiting processes in diffusion approximation such as CAF, CGMRP, and CDS. As an example, we considered the solution of portfolio optimization problem by Merton for the limiting CGMRP in DA. We also point out the importance of convergence rates and obtained them in the limit theorems for CDTSMRE and CAF, CGMRP, and CDS.
The future work will be associated with the study of optimal control for the initial, not limiting models, such as CAF in Section 4.1, CGMRP in Section 4.2, and CDS in Section 4.3. Other optimal control problems would be also interesting to consider for diffusion models with equilibrium, e.g., CAF in Section 4.1 and CGMRP in Section 4.2. In our future work, the latter models will be considered for solutions of Merton portfolio’s problems as well. We will also consider in our future research the case of dependent SMC and the MC .
Author Contributions
These authors contributed equally to this work. All authors have read and agreed to the published version of the manuscript.
Funding
The research of the first author is partially supported by NSERC.
Institutional Review Board Statement
Not applicable.
Informed Consent Statement
Informed consent was obtained from all subjects involved in the study.
Data Availability Statement
Not applicable.
Acknowledgments
We thank to four anonymous referees for valuable remarks and suggestions that improved the paper. The research of the first author is partially supported by NSERC. The first author also thanks to the Laboratory of Applied Mathematics of the Université de Technologie de Compiègne, Compiègne, France, very much for their hospitality during his visit.
Conflicts of Interest
The authors declare no conflict of interest.
Abbreviations
| SMC | Discrete-time semi-Markov chain; |
| DTSMRE | Discrete-time semi-Markov random evolution; |
| CDTSMRE | Controlled discrete-time semi-Markov random evolution; |
| CGMRP | Controlled geometric Markov renewal processes; |
| CAF | Controlled additive functionals; |
| CDS | Controlled dynamical systems; |
| HJB | Hamilton–Jacobi–Bellman (equation); |
| DPE | Dynamic programming equation; |
| DPP | Dynamic programming principle; |
| DA | Diffusion approximation; |
| SDE | Stochastic differential equation. |
References
- Griego, R.; Hersh, R. Random evolutions, Markov chains, and Systems of partial differential equations. Proc. Natl. Acad. Sci. USA 1969, 62, 305–308. [Google Scholar] [CrossRef] [PubMed]
- Hersh, R. Random evolutions: A Survey of results and problems. Rocky Mountain J. Math. 1974, 4, 443–477. [Google Scholar] [CrossRef]
- Koroliuk, V.S.; Limnios, N. Stochastic Systems in Merging Phase Space; World Scientific: Singapore, 2005. [Google Scholar]
- Korolyuk, V.S.; Swishchuk, A. Evolution of System in Random Media; CRC Press: Boca Raton, FL, USA, 1995. [Google Scholar]
- Swishchuk, A.; Wu, J. Evolution of Biological Systems in Random Media: Limit Theorems and Stability; Kluwer: Dordrecht, The Netherlands, 2003. [Google Scholar]
- Swishchuk, A.V. Random Evolutions and their Applications; Kluwer AP: Dordrecht, The Netherlands, 1995. [Google Scholar]
- Cohen, J.E. Random Evolutions in Discrete and Continuous Time. Stoch. Proc. Appl. 1979, 9, 245–251. [Google Scholar] [CrossRef]
- Keepler, M. Random evolutions processes induced by discrete time Markov chains. Port. Math. 1998, 55, 391–400. [Google Scholar]
- Limnios, N. Discrete-time random evolutions—Difference equations and additive functionals. Commun. Statist. Theor. Methods 2011, 40, 1–11. [Google Scholar] [CrossRef]
- Yin, G.G.; Zhang, Q. Discrete-Time Markov Chains. Two-Time-Scale Methods and Applications; Springer: New York, NY, USA, 2005. [Google Scholar]
- Anisimov, V.V. Switching processes: Averaging principle, diffusion approximation and applications. Acta Appl. Math. 1995, 40, 95–141. [Google Scholar] [CrossRef]
- Anisimov, V.V. Switching Processes in Queueing Models; J. Wiley & Sons: London, UK; ISTE: London, UK, 2008. [Google Scholar]
- Anisimov, V.V. Averaging principle for switching recurrent sequences. Theory Probab. Math. Stat. 1991, 45, 1–8. [Google Scholar]
- Anisimov, V.V. Averaging principle for switching processes. Theory Probab. Math. Stat. 1992, 46, 1–10. [Google Scholar]
- Swishchuk, A.V.; Islam, M.S. The Geometric Markov Renewal Processes with application to Finance. Stoch. Anal. Appl. 2010, 29, 4. [Google Scholar] [CrossRef]
- Swishchuk, A.V.; Islam, M.S. Diffusion Approximations of the Geometric Markov Renewal Processes and option price formulas. Int. J. Stoch. Anal. 2010, 2010, 347105. [Google Scholar] [CrossRef]
- Swishchuk, A.V.; Islam, M.S. Normal Deviation and Poisson Approximation of GMRP. Communic. Stat. Theory Methods 2013, 1488–1501. [Google Scholar] [CrossRef]
- Swishchuk, A.V.; Limnios, N. Optimal stopping of GMRP and pricing of European and American options. In Proceedings of the 15th Intern. Congress on Insurance: Mathematics and Economics (IME 2011), Trieste, Italy, 14–17 June 2011. [Google Scholar]
- Barbu, V.; Limnios, N. Semi-Markov Chains and Hidden Semi-Markov Models. Toward Applications. Their Use in Reliability and DNA Analysis; Lecture Notes in Statistics; Springer: New York, NY, USA, 2008; Volume 191. [Google Scholar]
- Limnios, N.; Swishchuk, A. Discrete-time Semi-Markov Random Evolutions and their Applications. Adv. Appl. Probab. 2013, 45, 214–240. [Google Scholar] [CrossRef]
- Jacod, J.; Shiryaev, A.N. Limit Theorems for Stochastic Processes; Springer: Berlin/Heidelberg, Germany, 1987. [Google Scholar]
- Bertsekas, D.; Shreve, S. Stochastic Optimal Control: The Discrete-Tiime Case; Athena Scientific Publisher: Belmont, MA, USA, 1996. [Google Scholar]
- Fleming, W.H.; Rishel, R.W. Deterministic and Stochastic Optimal Control; Springer: New York, NY, USA, 1975. [Google Scholar]
- Kushner, H.J. Introduction to Stochastic Control Theory; Holt, Rinehart, Winston: New York, NY, USA, 1971. [Google Scholar]
- Merton, R. Optimum consumption and portfolio rules in a continuous-time model. J. Econ. Theory 1971, 3, 373–413. [Google Scholar] [CrossRef]
- Sviridenko, M.N. Martingale approach to limit theorems for semi-Markov processes. Theor. Probab. Appl. 1986, 34, 540–545. [Google Scholar] [CrossRef]
- Adams, R. Sobolev Spaces; Academic Press: New York, NY, USA, 1979. [Google Scholar]
- Rudin, W. Functional Analysis; McGraw-Hill: New York, NY, USA, 1991. [Google Scholar]
- Sobolev, S. Some Applications of Functional Analysis in Mathematical Physics, 3rd ed.; American Mathematical Society: Providence, RI, USA, 1991; Volume 90. [Google Scholar]
- Pyke, R. Markov renewal processes: Definitions and preliminary properties. Ann. Math. Statist. 1961, 32, 1231–1242. [Google Scholar] [CrossRef]
- Pyke, R. Markov renewal processes with finitely many states. Ann. Math. Statist. 1961, 32, 1243–1259. [Google Scholar] [CrossRef]
- Pyke, R.; Schaufele, R. Limit theorems for Markov renewal processes. Ann. Math. Statist. 1964, 35, 1746–1764. [Google Scholar] [CrossRef]
- Shurenkov, V.M. On the theory of Markov renewal. Theor. Probab. Appl. 1984, 19, 247–265. [Google Scholar] [CrossRef]
- Maxwell, M.; Woodroofe, M. Central limit theorems for additive functionals of Markov chains. Ann. Probab. 2000, 28, 713–724. [Google Scholar]
- Nummelin, E. General Irreducible Markov Chains and Non-Negative Operators; Cambridge University Press: Cambridge, UK, 1984. [Google Scholar]
- Revuz, D. Markov Chains; North-Holland: Amsterdam, The Netherlands, 1975. [Google Scholar]
- Skorokhod, A.V. Asymptotic Methods in the Theory of Stochastic Differential Equations; AMS: Providence, RI, USA, 1989; Volume 78. [Google Scholar]
- Skorokhod, A.V.; Hoppensteadt, F.C.; Salehi, H. Random Perturbation Methods with Applications in Science and Engineering; Springer: New York, NY, USA, 2002. [Google Scholar]
- Jaśkiewicz, A.; Nowak, A. Average optimality for semi-Markov control processes. Morfismos 2007, 11, 15–36. [Google Scholar]
- Vega-Amaya, O.; Luque-Vásquez, F. Sample-path average cost optimality for semi-Markov control processes on Borel spaces: Unbounded cost and mean holding times. Appl. Math. 2000, 27, 343–367. [Google Scholar] [CrossRef]
- Altman, E.; Shwartz, A. Markov decision problems and state-action frequencies. SIAM J. Control. Optimikzation 1991, 29, 786–809. [Google Scholar] [CrossRef]
- Beutler, F.; Ross, K. Optimal policies for controlled Markov chains with a constraint. J. Mathem. Analysis Appl. 1985, 112, 236–252. [Google Scholar] [CrossRef]
- Borkar, V. Dynamic programming for ergodic control with partial observations. Stoch. Proc. Appl. 2003, 103, 293–310. [Google Scholar] [CrossRef]
- Boussement, M.; Limnios, N. Markov decision processes with asymptotic average failure rate constraint. Communic. Statis. Theory Methods 2004, 33, 1689–1714. [Google Scholar] [CrossRef]
- Hernández-Lerma, O.; Lasserre, J.B. Discrete-Time Markov Control Processes; Springer: New York, NY, USA, 1996. [Google Scholar]
- Krylov, N.; Bogolyubov, N. Introduction to Non-linear Mechanics; Princeton University Press: Princeton, NJ, USA, 1947. [Google Scholar]
- Limnios, N.; Oprişan, G. Semi-Markov Processes and Reliability; Birkhäuser: Boston, MA, USA, 2001. [Google Scholar]
- Cartea, A.; Jaimungal, S.; Penalva, J. It Algorithmic and High-Frequency Trading; Cambridge University Press: Cambridge, UK, 2015. [Google Scholar]
- Ethier, S.N.; Kurtz, T.G. Markov Processes: Characterization and Convergence; J. Wiley: New York, NY, USA, 1986. [Google Scholar]
- Stroock, D.W.; Varadhan, S.R.S. Multidimensional Diffusion Processes; Springer: Berlin/Heidelberg, Germany, 1979. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations. |
© 2021 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (http://creativecommons.org/licenses/by/4.0/).