1. Introduction
Differential games, which cannot be solved by application of the first-order solvability conditions, are called singular. For instance, a zero-sum differential game is called singular if it cannot be solved using the Isaacs MinMax principle [
1,
2] and the Bellman–Isaacs equation method [
1,
3]. Similarly, Nash equilibrium set of controls in a singular non-zero-sum differential game cannot be derived using the first-order variational method and the generalized Hamilton-Jacobi-Bellman equation method [
3,
4].
Singular differential games appear in various applications. For example, such games appear in pursuit-evasion problems (see, e.g., Ref. [
5]), in robust controllability problems (see, e.g., Ref. [
6]), in robust interception problems of maneuvering targets (see e.g., Ref. [
7]), in robust tracking problems (see, e.g., Ref. [
8]), in biology processes (see, e.g., Ref. [
9]), and in robust investment problems (see, e.g., Ref. [
10]).
Treating a singular differential game, one can try to use higher order solvability conditions. However, such conditions are useless for the game, which does not have an optimal control of at least one player in the class of regular (non-generalized) functions.
Singular zero-sum differential games were extensively analyzed in the literature by different methods (see, e.g., Refs. [
7,
11,
12,
13,
14,
15,
16,
17,
18] and references therein). Thus, in Refs [
7,
15,
16], various singular zero-sum differential games were solved by regularization method. In Reference [
11], a numerical method was proposed to solve one class of zero-sum differential games with singular control. In Reference [
12], a class of zero-sum differential games with singular arcs was considered. For this class of the games, sufficient conditions for the existence of a saddle-point solution were established. In Reference [
13], the Riccati matrix inequality was applied to establish the existence of an almost equilibria in a singular zero-sum differential game. In Reference [
14], a saddle-point solution of a singular zero-sum differential game was derived in the class of generalized functions. In Reference [
17], a class of zero-sum stochastic differential games was studied. Each player of this game has a control consisting of regular and singular parts. Necessary and sufficient saddle-point optimality conditions were derived for the considered game. In Reference [
18], a singular zero-sum linear-quadratic differential game was considered. This game was treated by its regularization and numerical solution of the regularized game.
Singular non-zero-sum Nash equilibrium differential games also were studied in the literature, but mostly in various stochastic settings (see, e.g., Refs. [
10,
19,
20,
21,
22] and references therein). Deterministic singular non-zero-sum Nash equilibrium differential games were studied only in few works. Thus, in Reference [
23], a two-person non-zero-sum differential game with a linear second order dynamics and scalar controls of both players was considered. Each player controls one equation of the dynamics. The infinite horizon quadratic functionals of the players do not contain control costs. The admissible class of controls for both players is the set of linear state-feedbacks. The notion of asymptotic (with respect to time)
-Nash equilibrium was introduced, and this equilibrium was designed subject to some condition. In Reference [
9], a finite-horizon two-person non-zero-sum differential game was studied. This game models a biological process. Its fourth-order dynamics is linear with respect to scalar controls of the players, and these controls are bounded. The players’ functionals depend only on the state variables, and this dependence is quadratic. For this singular game, a Nash equilibrium set of open-loop controls was derived in the class of regular functions. In Reference [
24], an infinite horizon two-person non-zero-sum differential game with
n-order linear dynamics and vector-valued unconstrained players’ controls was considered. Functionals of both players are quadratic, and these functionals do not contain control costs of one (the same) player. This singular game was solved by the regularization approach.
In the present paper, we consider a deterministic finite-horizon two-person non-zero-sum differential game. The dynamics of this game is linear and time-dependent. The controls of the players are unconstrained. Each player aims to minimize its own quadratic functional. We look for the Nash equilibrium in this game, and we treat the case where weight matrices in control costs of one player (the “singular” player) in both functionals are singular but non-zero. Such a feature means that the game under the consideration is singular. However, since the aforementioned singular weight matrices are non-zero, the control of the “singular” player contains both, singular and regular, coordinates. For this game, in general, the Nash equilibrium pair of controls, in which singular coordinates of the “singular” player’s control are regular (non-generalized) functions, does not exist. To the best of our knowledge, such a game has not yet been studied in the literature. The aims of the paper are the following: (A) to define the solution (the Nash equilibrium) of the considered game; (B) to derive this solution. Thus, we propose for the considered singular game a novel notion of the Nash equilibrium (a Nash equilibrium sequence). Based on this notion, we solve the game by application of the regularization method. Namely, we associate the original singular game with a new differential game. This new game has the same equation of the dynamics and a similar functional of the “singular” player augmented by a finite-horizon integral of the square of its singular control coordinates with a small positive weight (a small parameter). The functional of the other (“regular”) player remains unchanged. Thus, the new game is a finite-horizon regular linear-quadratic game.
The regularization method was applied for solution of singular optimal control problems in many works (see, e.g., Refs. [
25,
26,
27] and references therein). This method also was applied for solution of singular
control problems (see, e.g., Refs. [
28,
29] and references therein) and for solution of singular zero-sum differential games (see, e.g., Refs. [
7,
15,
16]). However, to the best of our knowledge, the application of the regularization method to analysis and solution of singular non-zero-sum differential games was considered only in two short conference papers [
24,
30]. In each of these papers, the study of the game was presented in a brief form and without detailed analysis and proofs of assertions.
The aforementioned new game, obtained by the regularization of the original singular game, is a partial cheap control game. Using the solvability conditions of a Nash equilibrium finite-horizon linear-quadratic regular game, the solution of this partial cheap control game is reduced to solution of a set of two matrix Riccati-type differential equations, singularly perturbed by the small parameter. Using an asymptotic solution of this set, a Nash equilibrium sequence of the pairs of the players’ state-feedback controls in the original singular game is constructed. The expressions for the optimal values of the players’ functionals in this game are obtained. Note that a particular case of the differential game, studied in the present paper, was considered briefly and without detailed proofs in the short conference paper [
30].
The paper is organized as follows. In the next section, the initial formulation of the singular differential game is presented. The main definitions also are formulated. The transformation of the initially formulated game is carried out in
Section 3. It is shown that the initially formulated game and the transformed game are equivalent to each other. Due to this equivalence, in the rest of the paper the transformed game is analyzed as an original singular differential game. The regularization of the original singular game, which is made in
Section 4, yields a partial cheap control regular game. Nash equilibrium solution of the latter is presented in
Section 5. Asymptotic analysis of the partial cheap control regular game is carried out in
Section 6. In
Section 7, the reduced differential game, associated with the original singular game, is presented along with its solvability conditions. The Nash equilibrium sequence for the original singular differential game and the expressions of the functionals’ optimal values of this game are derived in
Section 8. Two illustrative examples are considered in
Section 9.
Section 10 is devoted to concluding remarks. Some technically complicated proofs are placed in appendices.
The following main notations are used in the paper:
is the n-dimensional real Euclidean space.
The Euclidean norm of either a vector or a matrix is denoted by .
The upper index “T” denotes the transposition either of a vector x () or of a matrix A ().
denotes the identity matrix of dimension n.
denotes zero matrix of dimension ; however, if the dimension of zero matrix is clear, it is denoted as 0.
denotes the space of all functions square integrable in the interval .
, where , , denotes the column block-vector of the dimension with the upper block x and the lower block y, i.e., .
⊗ denotes the Kronecker product of matrices.
For a given -matrix A, means its vectorization, i.e., the -dimensional block vector in which the first (upper) block is the first (upper) row of A, the second block is the second row of A, and so on, the lower block of is the last (lower) row of A.
2. Initial Game Formulation
The game’s dynamics is described by the following system:
where
is a given final time instant;
is the state vector,
, (
),
are the players’ controls;
,
and
,
are given matrix-valued functions of corresponding dimensions;
is a given constant vector.
The functionals of the player “u” with the control
and the player “v” with the control
are, respectively,
where
and
are given symmetric positive semi-definite matrices of corresponding dimensions;
,
,
,
are given matrix-valued functions of corresponding dimensions; the matrix
is symmetric positive definite; the matrices
,
,
,
, and
are symmetric positive semi-definite.
In what follows, we assume that the weight matrices
and
of the costs of the control
in both functionals have the block form
where the matrices
and
are of the dimension
, (
); the matrix
is positive definite; the matrix
is positive semi-definite.
The player “u” aims to minimize the functional (
2) by a proper choice of the control
, while the player “v” aims to minimize the functional (
3) by a proper choice of the control
.
We study the game (
1)–(
3) with respect to its Nash equilibrium, and subject to the assumption that both players know perfectly the current game state.
Remark 1. Due to the assumption (4), the first-order Nash-equilibrium solvability conditions (see, e.g., Refs. [3,4]) cannot be applied to analysis and solution of the game (1)–(3), i.e., this game is singular. Moreover, this game does not have, in general, its solution (a Nash-equilibrium pair of controls) in the class of regular (non-generalized) functions. Consider the set of all functions , which are measurable w.r.t. for any fixed and satisfy the local Lipschitz condition w.r.t. uniformly in . In addition, consider the set of all functions with the same properties.
Definition 1. By , we denote the set of all pairs of functions satisfying the following conditions:
- (i)
, ;
- (ii)
the initial-value problem (1) for , and any has the unique absolutely continuous solution , ; - (iii)
;
- (iv)
.
In what follows, is called the set of all admissible pairs of players’ state-feedback controls (strategies) in the game (1)–(3). For any given functions
and
, we consider the sets
Consider the sequence of the pairs , .
Definition 2. The sequence is called a Nash equilibrium strategies’ sequence (or simply, a Nash equilibrium sequence) in the game (1)–(3) if: - (a)
for any , there exist finite limits and in the game (1)–(3); - (b)
for all ;
- (c)
for all .
The valuesandare called optimal values of the functionals (2) and (3), respectively, in the game (1)–(3). 3. Transformation of the Game (1)–(3)
Let us represent the matrix
in the block form
where the matrices
and
have the dimensions
and
, respectively.
In what follows, we assume:
AI. The matrix has full column rank r for all .
AII., .
AIII., .
AIV. The matrix-valued functions , , , , , , and are continuously differentiable in the interval .
AV. The matrix-valued functions and are twice continuously differentiable in the interval .
Let the -matrix be a complement matrix to in the interval , i.e., the block matrix is invertible for all . Therefore, the -matrix is a complement matrix to in the interval .
In what follows, we also assume:
AVI. The matrix-valued function is twice continuously differentiable in the interval .
Using the matrices
and
, we construct the following matrices:
Now, using the matrix
, we make the following transformation of the state variable
in the game (
1)–(
3):
where
is a new state variable.
Due to the results of Reference [
31], the transformation (
9) is invertible.
For the sake of the further analysis, we partition the matrix
into blocks as:
where the matrices
and
have the dimensions
and
, respectively.
Quite similarly to the results of References [
15,
29], we have the following assertion.
Proposition 1. Let the assumptions AI-AVI be valid. Then, the state transformation (9) converts the system (1) to the systemand the functionals (2), (3) to the functionalswhere The matrices and are symmetric positive semi-definite, while the matrix is symmetric positive definite for all . The matrices and are symmetric positive semi-definite. Moreover, the matrix-valued functions , , , , and are continuously differentiable in the interval .
Remark 2. In the new (transformed) game with the dynamics (11) and the functionals (12), (13), the player “u” aims to minimize the functional (12) by a proper choice of the control , while the player “v” aims to minimize the functional (13) by a proper choice of the control . Since in the game (1)–(3) both players know perfectly the current state , then due to the invertibility of the transformation (9), in the game (11)–(13) both players also know perfectly the current state . Like the game (1)–(3), the new game (11)–(13) also is singular. Consider the set of all functions , which are measurable w.r.t. for any fixed and satisfy the local Lipschitz condition w.r.t. uniformly in . In addition, consider the set of all functions with the same properties.
Definition 3. By , we denote the set of all pairs of functions satisfying the following conditions:
- (i)
, ;
- (ii)
the initial-value problem (11) for , and any has the unique absolutely continuous solution , ; - (iii)
;
- (iv)
.
In what follows, is called the set of all admissible pairs of players’ state-feedback controls (strategies) in the game (11)–(13). Corollary 1. Let the assumptions AI-AVI be valid. Let and , be the solution of the initial-value problem (1) generated by this pair of the players’ controls. Then, and , , where , is the unique solution of the initial-value problem (11) generated by the players’ controls , . Vice versa: let and , be the solution of the initial-value problem (11) generated by this pair of the players’ controls. Then, and , , where , is the unique solution of the initial-value problem (1) generated by the players’ controls , . Proof. The statements of the corollary directly follow from Definitions 1 and 3 and Proposition 1. □
For any given
and
, consider the sets
Consider the sequence of the pairs , .
Definition 4. The sequence is called a Nash equilibrium strategies’ sequence (or simply, a Nash equilibrium sequence) in the game (11)–(13) if: - (I)
for any , there exist finite limits and
in the game (11)–(13); - (II)
for all ;
- (III)
for all .
The valuesandare called optimal values of the functionals (12) and (13), respectively, in the game (11)–(13). Lemma 1. Let the assumptions AI-AVI be valid. Let be the Nash equilibrium sequence in the game (1)–(3). Then, is the Nash equilibrium sequence in the game (11)–(13). Vice versa: let be the Nash equilibrium sequence in the game (11)–(13). Then, is the Nash equilibrium sequence in the game (1)–(3). Corollary 2. Let the assumptions AI-AVI be valid. Then, the optimal values and of the functionals (2) and (3) in the game (1)–(3) coincide with the optimal values and of the corresponding functionals (12) and (13) in the game (11)–(13), i.e., and . Proof. The statement of the corollary is a direct consequence of the expressions for
,
,
and
(see Definitions 2 and 4), and the proof of Lemma 1 (see Equations (
A2) and (
A3)–(
A6) in
Appendix A). □
Remark 3. Due to Lemma 1 and Corollary 2, the initially formulated differential game (1)–(3) is equivalent to the new (transformed) differential game (11)–(13). Moreover, due to Proposition 1, the new game is simpler than the initial game. Due to this observation, in what follows of this paper, we consider the game (11)–(13) as an original game. We call this game the Singular Differential Game (SDG). 8. Nash Equilibrium Sequence of the SDG
For a given
, consider the following vector-valued function of
:
where
,
,
;
and
are given by (
56) and (
53), respectively.
Lemma 3. Let the assumptions AI-AVII be valid. Then, for any given , the pair is an admissible pair of the players’ state-feedback controls in the SDG (11)–(13), i.e., . Proof. The statement of the lemma directly follows from the linear dependence of on , on , and the continuity with respect to of the gain matrices in and . □
Lemma 4. Let the assumptions AI-AVII be valid. Then, in the SDG (11)–(13), the following limit equalities are satisfied:where and are given in (82). Let us substitute the control
instead of
into the system (
11) and the functional (
13). Due to this substitution, we obtain the following optimal control problem with the state variable
and the control
:
We seek the optimal control of the problem (
92) in the state-feedback form
among all such controls belonging to the set
, where
is given in (
20). Let
be the optimal value of the functional in the problem (
92).
Lemma 5. Let the assumptions AI-AVII be valid. Then, there exists a positive number such that, for all , the following inequality is satisfied: where is given in (82); is some constant independent of ε but depending on . Proof. The lemma is proven similarly to the results of Reference [
7] (see Lemmas 1, 4 and their proofs). □
Now, let us replace
with
in the system (
11) and the functional (
12). Due to such a replacement, we obtain the following optimal control problem with the state variable
and the control
:
We seek the infimum of the functional
in the optimal control problem (
93) for the state-feedback controls
belonging to the set
, where
is given in (
21). Let
be the infimum value of the functional in the problem (
93).
Lemma 6. Let the assumptions AI-AVII be valid. Then, the following equality is satisfied: where is given in (82). Proof. The lemma is proven similarly to the results of Reference [
7] (see Lemma 5 and its proof). □
Let , be a sequence of numbers such that: (i) , ...); (ii) .
Theorem 1. Let the assumptions AI-AVII be valid. Then, the sequence of the state-feedback controls , , where and are defined in (87) and (90), is the Nash equilibrium sequence in the SDG. Moreover, the optimal values and of the functionals in this game arewhere and are the optimal values of the functionals in the RDG (84)–(86) given by Equation (82). Proof. First of all let us note that, due to Lemma 3, the pair
is admissible in the SDG for any
. Therefore, to prove the first statement of the theorem, we should show the fulfillment of all the items of Definition 4 for the sequence
,
. Lemma 4 yields the fulfillment of the item (I) of this definition. The fulfillment of the item (II) directly follows from the first equality in (
91) and Lemma 6. The fulfillment of the item (III) follows immediately from the second equality in (
91) and Lemma 5. Namely, from this lemma, we have the inequality
, while, from the definition of the value
, we have the inequality
,
,
. The left-hand side of the first inequality, along with the second inequality and the second equality in (
91), yields
,
,
. Calculating
of both sides of the latter inequality, we obtain the fulfillment of the item (III) of Definition 4 for the sequence
,
. Thus, this sequence satisfies all the items of Definition 4.
The second statement of the theorem is a direct consequence of the expressions for and in Definition 4, as well as Lemma 4 and Proposition 3. □
Remark 7. Due to Theorem 1, to design the Nash equilibrium sequence in the SDG and to obtain the optimal values of its functionals, one has to solve the lower dimension regular RDG and to construct the gain matrices , , .
10. Concluding Remarks
CR1. In this paper, a finite-horizon two-person linear-quadratic Nash equilibrium differential game was studied. The game is singular because the weight matrices of the control costs of one player (the “singular” player) are singular in the functionals of both players. These singular weight matrices are positive semi-definite but non-zero. The weight matrix of the control cost of the other player (the ”regular” player) in its own functional is positive definite.
CR2. Subject to proper assumptions, the system of dynamics of this game was transformed to an equivalent system consisting of three modes. The first mode is controlled directly only by the ”regular” player. The second mode is controlled directly by the ”regular” player and the nonsingular control’s coordinates of the “singular” player. The third mode is controlled directly by the entire controls of both players. Due to this transformation, the initially formulated game was converted to an equivalent Nash equilibrium game. The new game, also being singular, is simpler than the initially formulated game. Therefore, the new game was considered as an original one.
CR3. For this game, a novel notion of the Nash equilibrium (the Nash equilibrium sequence) was proposed. To derive the Nash equilibrium sequence in the original singular game, the regularization method was applied. This method consists in the replacing the original singular game with a regular Nash equilibrium game depending on a small parameter . This regular game becomes the original singular game if we set formally . It should be noted that the regularization method was widely applied in the literature for analysis and solution of singular optimal control problems, singular control problems and zero-sum differential games. However, in the present paper, this method was applied for the first time in the literature to the rigorous and detailed analysis and solution of the general singular linear-quadratic Nash equilibrium differential game.
CR4. The regularized game is a partial cheap control game. Complete/partial cheap control problems were widely studied in the literature in the settings of an optimal control problem, an control problem and a zero-sum differential game. Non zero-sum differential games with a complete cheap control of one player also were considered in the literature, although in few works. However, in the present paper, for the first time in the literature, a non-zero-sum differential game with a partial cheap control of at least one player was analyzed.
CR5. Solvability conditions of the regularized (partial cheap control) game depend on the small parameter , which allowed us to analyze these conditions asymptotically with respect to . Using this analysis, the Nash equilibrium sequence in the original singular game was designed, and the expressions for the optimal values of the functionals were obtained.
CR6. It was established that the construction of the Nash equilibrium sequence in the original singular game and the obtaining the optimal values of its functionals are based on the solution of a lower dimension regular Nash equilibrium differential game (the reduced game). Namely, to solve the original singular game, one has to solve the lower dimension regular game and to calculate by explicit formulas two additional gain matrices.