1. Introduction
Since the first studies of Feldbaum [
1,
2], Pontryagin’s Principle of Maximum [
3], etc., the theory of linear time-optimal control problem has gained maturity—the main theoretical issues have been thoroughly studied and answered [
4,
5,
6,
7,
8,
9]. This historical evolution and facts provide a solid background of the progress in this field. The achieved state of knowledge in this field establishes the foundation for further exploration and advancement. Achieving a transition from one system state to another in a minimum time with maximum utilization of the available system resources—control within the constraints of both control inputs and state space variables—in a form of synthesis still presents an attractive topic for further research.
In synchrony with the above mentioned, the authors in the recently published book [
10] state, “there has been tremendous progress in numerical methods in optimal control over the past fifteen years that has led to the solutions of some specific and very difficult problems” and, in particular, the introduction of geometrical methods, more specifically—“a first illustration of the power of geometric methods that go well beyond the conditions of the maximum principle and lead to deep results about the structure of optimal solutions”. The geometric approach to the optimal control of a double integrator is also discussed in [
11,
12].
In a recent publication on the topic [
13], the authors say that, “this paper has proposed a global time optimal control law for triple integrator with input saturation and full state constraints”, and in terms of the results, “An analytical state feedback form control law has been synthesized based on the switching surfaces and curves”.
The authors also mention “there are plenty of researches trying to solve the problem analytically, while there is still no complete time optimal analytical solution for systems higher than second order”.
This is noteworthy considering Pontryagin’s original sources. In reference [
3] (Chapter 3, § 20, § 21, Example 3), the author and his colleagues describe the solution of the problem of a linear time-optimal control system fulfilling the condition of normality with real non-positive eigenvalues and one control input as follows.
The time-optimal control for such a type of linear system has maximum
(the order of the system) intervals of constancy, i.e., the number of switchings is maximum
; the state-space of the system is separated into manifold
…,
of dimensions, respectively, 1, 2, …,
n. The manifold
consists of all the points for which the time-optimal control has one interval of constancy. Supposing
, the trajectory of the system under the control +1 ending at the state-space origin is defined as
, while the trajectory of the system ending at the state-space origin but under the control -1 is defined as
. Together,
and
compose the switching curve
. The final stage of the time-optimal process represents a movement alongside
or
. All the trajectories of the system ending at a point of the curve
under the control +1 fill the surface
. Analogically, all the trajectories ending at a point of the curve
under the control −1 fill the surface
Combining
and
we obtain the switching surface
, so the last two stages of each time-optimal process are in
. In the same manner, the rest of the manifolds are constructed. The manifold
is of dimension
;
is entirely in
and divides it into two areas
and
;
consists of all the trajectories under the control +1 ending at a point of
, while
consists of all the trajectories under the control −1 ending at a point of
. The last manifold
M1 coincides with the whole state-space of the system. The synthesizing function is depicted as:
So, in order to synthesize the time-optimal control for a given system fulfilling the above conditions, one needs to describe properly the switching surfaces and .
Despite the progress in the field, finding a new solution for the problem discussed above by Pontryagin and others without the need of directly describing the respective manifolds and renders it more appealing by conducting a deeper investigation of the state-space geometric properties of this time-optimal control problem.
A novel method for synthesizing the time-optimal control for a class of controllable linear systems of any order with real non-positive simple eigenvalues and one input is developed and further explored in the dissertation [
14] and the following papers [
15,
16,
17]. It is founded on some new state-space properties of the considered linear time-optimal control problem and the exclusion of switching surfaces description serves as its main advantage. The study [
18] illustrates an example of a possible application of the method in practice.
Therefore, it is worthwhile trying to expand the thus developed solution of synthesizing the linear time-optimal control without the description of switching surfaces and curves to the more general case as the one described by Pontryagin and colleagues, in particular, a controllable linear system with one input and real non-positive eigenvalues, but not just non-positive simple eigenvalues.
The current paper is structured in the following way. In
Section 2, a new property of the linear time-optimal control problem is theoretically represented. In
Section 3, the author compares the classic solution of the time-optimal control problem of a double integrator to the alternatively suggested novel way by application of the new property.
Section 4 represents a detailed discussion of the obtained results.
2. Formulation of the Problem and Solution
Let us consider the following linear time-optimal control problem of order
,
. The system is described by the equations:
Let us suppose it is controllable as well as possessing real non-positive eigenvalues. It should be mentioned that every normal system with real eigenvalues could be transformed to such a type of presentation.
The initial state at the moment
of the system (1) is
and the target state at the moment
represents the origin of the system’s state-space where
is unspecified
The admissible control
is a piecewise continuous function that takes its values in the range of
which is continuous on the boundaries of the set of allowed values (4) and with regard to the points of discontinuity
we have
The problem is to find an admissible control
which transfers the system (1) from its initial state (2) to the final state (3) in minimum time, i.e., minimizing the performance index
Let us refer to this problem as “Problem P()”.
The form of the equations of the system (1) allows the introduction of the linear sub-system of order
with the state-space vector
and scalar output
where the matrices
are, respectively,
Thus, the system (1) could be represented by (7) in the following form which is also depicted in
Figure 1.
With regard to the sub-system (7), its initial state may be represented by
(11) and the relationship between the initial states of both the system and the sub-system may be described as (12).
Let us formulate the following linear time-optimal control problem of order
which we shall call “Problem
”. The system is defined by Equation (7). The initial state of the system (7) at the moment
is (11) and the target state at the moment
, which one should bear in mind is not initially specified, is the origin of the
-dimensional state-space of the system (7)
The admissible control
represents a piecewise continuous function that takes its values in the range of (4), which is continuous on the boundaries of the set of allowed values (4). Regarding the points of discontinuity
we have (5). The Problem
consists of synthesizing an admissible control
which on the one hand transfers the system (7) from its initial (11) to final state (13) and on the other hand, minimizes the performance index
Let us assume we have found the solution of Problem
and denote by
the optimal time defined as the minimum time of (14)
by
—the optimal control, and
—the optimal trajectory in the
-dimensional state-space of the system (7), which is described by
Let us denote the scalar output of the system (7) as a representation of the optimal vector-function
resulting as
In that case,
stands for
Let us define
(21) as an initial state of the
-th coordinate of the state-space vector
of the system (1) or (10) and consider the trajectory
in the
-dimensional state-space of Problem
with initial state in the point
and coordinates (20) and (21) under the optimal control
of Problem
.
Given the characteristics of the system as defined in (1), the vector-function
presented as (10) specifies (22). According to (16), the first
variables of the vector-function in (22) typify the optimal vector-function
of Problem
. Regarding the last
-th variable of
in (22), the function
depicts the scalar output of the system (7), which in this case is the result of the optimal vector-function
. Then, in terms of the above mentioned and in consonance with (18),
equals
. Thus, we obtain (23) for
(22).
With regard to
(23) at the moment
we obtain
Then, substituting
for (17) and
for (21), the following result is achieved.
Thus, we obtain that for Problem
the trajectory
in the
-dimensional state-space of the system (1) or (10) with initial point
(20) and (21) under the optimal control
of Problem
ends at the moment
at the origin of the
-dimensional state-space of Problem
. Taking into account that the function
represents the optimal control of Problem
and thereby it is a piecewise constant function with an amplitude
and a number of switchings maximum
, i.e., the number of intervals of constancy maximum is
[
7] (Chapter 2, §6, Theorem 2.11, p. 116), one comes to the conclusion that the trajectory
lies wholly on the switching hyper-surface of Problem
.
Let us now consider the trajectory
(28) in the
-dimensional state-space of Problem
with an initial point representing the initial state
(2) or (12) of Problem
under the optimal control
of Problem
. According to (16), the first
variables of the vector-function in (28) account for the optimal vector-function
of Problem
. With regard to the last variable of
in (28), the function
is the scalar output of the system (7), which is actually the result of the optimal vector-function
. In consonance with (18), the function
therefore denotes
in this case. Thus, we obtain (29) for
(28).
Let us consider now the difference between the two vector-functions
(29) and
(23). Thus, we obtain consecutively
As for the last -th coordinate of (31) for we could state that:
If , then the initial state (2) or (12) of Problem coincides with the point with coordinates (20)–(21). As already illustrated, represents a point of the switching hyper-surface of Problem and the trajectory with the initial point under the optimal control of Problem lies wholly on the switching hyper-surface of Problem and ends at the moment at the origin of the -dimensional state-space of the system (1) or (10) of Problem ;
If , then the initial state (2) or (12) of Problem does not coincide with the point with coordinates (20)–(21). The expression for does not change its sign and is not equal to zero because is a finite time. Thus, the trajectory with initial state (2) or (12) of Problem under the optimal control of Problem lies entirely above or below the switching hyper-surface of Problem nowhere intersecting it and ends at the moment at a point of the coordinate axis different from zero.
Thus, the following theorem has been proven.
Theorem 1. The trajectory of the system (1) or (10) with initial point in (2) under the optimal control of Problem lies wholly on the switching hyper-surface of Problem and ends at the moment at the origin of the n-dimensional state-space of the system (1) or (10) of Problem or lies entirely above or below the switching hyper-surface of Problem nowhere intersecting it and ends at the moment at a point of the coordinate axis different from zero.
3. Example
Let us consider the following example of synthesizing the time-optimal control of a double integrator (§ 3. Example. The problem of synthesis, p. 38) [
7]; (Chapter 7, Problem 7.1, p. 150) [
11,
12]. It is noteworthy to mention that the above problem of synthesis, as it is already an established example, has found a place in online optimal control courses on world platforms with video content [
19,
20,
21,
22]. It should be noted that these online resources are often volatile and unavailable after some time. In the first place, an illustration of this classical synthesis will be presented, and thereafter the synthesis as an expansion and update of the method [
14] by the new property.
The system is described by the variables
(position) and
(velocity) and represents
Let the constraints of the admissible control
(4), (5) be
.
3.1. Classical Synthesis
The switching curve
in the phase plane
is described by
The two pieces and of the switching curve are the parts of the parabolas representing the phase trajectories going through the origin of the phase plane in case of constant control or respectively.
The two areas
and
in the phase plane,
below and above the switching curve
(34), respectively, encompass the areas where the optimal control takes a value
with regard to the points of
and
with regard to the points of
. The areas
and
as well as the parts
and
of
are shown in the following
Figure 2.
The time-optimal control is synthesized in the form
After substitution of
and
for (35) as well as
and
for (34) in (36) the synthesized optimal control appears as
3.2. Synthesis Based on the New Property and the Method [14]
Let us now consider the synthesis in terms of the method developed in [
14] and expanded by the new property. One of the founding properties of the described method regards the trajectory in the state-space of a time-optimal control problem of higher order now being defined by the solution for the lower order, taking into consideration that all the time-optimal control problems of descending order are generated by the problem of the utmost order and form a class of problems. Thus, the method now allows a synthesis to be defined without the description of the switching hyper-surfaces. As we have shown here, the new property represents an expansion covering the general case of controllable linear systems with one input and real non-positive eigenvalues. Therefore, the simple non-positive system’s eigenvalues of the method demonstrated in [
14] is now omitted as an initial restriction. The example here considers a system of order two with double zero eigenvalue, so the synthesis is directly based on the solution of the problem of order one, which also allows the solution of the initial problem to be expressed analytically.
Step 1. First, we make a suitable change of variables through (38) and obtain a representation by
, which could also be performed by the matrix
(39) and (40) via (41).
Thus, we obtain (43) and (44) from the initial system (32) through its matrix representation (42).
The system (44) is now in the form (1). Then, (44) in form (10) is represented as (45) and (46) (as (1) in form (10)). The sub-system of (45) and (46) is (47) or (48).
Step 2. Solving Problem
. The eigenvalue of
is 0. The optimal control of Problem
,
for
is (49) and (50) [
14] (pp. 50–52).
Step 3. Calculating the value of the variable
. The variable
is defined in [
14] (pp. 39–40), [
15] (p. 320), [
16] (p. 41) and in the case of expanding the class of time-optimal control problems here, it represents at
the
-th coordinate of the vector
(29) at the moment
In case
the variable
represents (52). With regard to the system (47) or (48) of Problem
, the variable
(52) becomes (53) and after simplifying—(55).
Step 4. Applying the theorem for synthesizing the optimal function in the initial state [
14] (Theorem 3.2, pp. 40–43), [
15] (Theorem 3, p. 320), and [
16] (Theorem 3, p. 41). According to this theorem and its corollaries, the time-optimal control in the initial state of Problem
represents (56).
The variable
, respectively,
in (56), is a term introduced in [
14] (p. 38) and [
15] (pp. 319–320) and defines the relationship between the points on axis
of the state-space of the system of Problem
from the considered class of problems and the switching hyper-surface of the same Problem
. The value of the variable
is determined by a procedure called “axes initialization” (Chapter 3,
Section 3.3, pp. 60–88) [
14] and (pp. 41–45) [
16].
With regard to the example
Hence, this means that all the points of the negative semi-axis are above the switching curve of Problem P(2) and the optimal control value for them is while all the points of the positive semi-axis are below the switching curve of Problem P(2) and the optimal control value for them is .
Thus, after substitution
for (55) and
for (57) taking into consideration the initial state
based on (56), we obtain
So, the synthesized optimal function with regard to a state
is
Taking into account
according to (46), (59) becomes
Bearing in mind the relation (38) or (41) between and one can easily appreciate that the analytical expression of the synthesized here optimal control (60) is identical with the expression obtained by the classical synthesis (37).
3.3. Simulation Results
For instance, let us depict the following two initial states
The corresponding initial states in the state-space
of the system (44) are, respectively,
In Step 2, according to (49) and (50) and with regard to (63), we obtain
In Step 3, with regard to
according to (53), we obtain
Thus, in Step 4 in reference to (56) and (57), the result for the time-optimal control in the initial state (63) is
Analogically, in accord with the initial state (64) in Step 2, likewise the initial state (63) in Step 2, we again obtain (65), but in terms of
the result is
which leads to
Figure 3 shows the near time-optimal processes with an accuracy of
with regard to the considered initial states while the trajectories in the phase plane
of the system (32) are shown in
Figure 4a. The blue and red phase trajectories outline the initial states
and
, respectively. The near time-optimal trajectories relating to the corresponding initial states in the state-space
of (44)
and
are represented in the phase plane
of (44) in
Figure 4b. The blue trajectory concerns the state
, however, the red one—
The conversion of the trajectories shown in
Figure 4b by the relation (41) returns the identical result shown in
Figure 4a.
4. Discussion
If the assumption of real non-positive eigenvalues of the system, in particular the constraints on the eigenvalues of subsystem (7), is omitted as they lack any specific characteristics, then in accordance with the idea and derivation technique it will simply turn out that the trajectory of the system with the initial state (2), obtained under the action of the optimal control of Problem , will coincide with the trajectory with the initial point with coordinates (20) and (21) or that this trajectory will be completely below or above the last one in a vertical direction, determined by the axis and will end at a point on the axis different from the coordinate origin.
In order to serve the idea of synthesis this result is a matter of separate research. In case of not considering the spectral structure of the system matrix the number of switches or intervals of constancy, although finite, is not limited by the order of the system. At first glance, it might be appropriate to look at a certain area around the coordinate origin, if we do not deviate from the idea of the approach.
However, taking into account that the eigenvalues of the system are real non-positive, then on the basis of the theorem for the number of intervals of constancy [
7] (Chapter 2, §6, Theorem 2.11, p. 116), we obtain that the initial point
of Problem P(
) and the end point of the obtained trajectory located on the axis
have the same relationship to the switching surface.
In [
14,
15,
16], a novel property of the state space of the system has been defined, in particular that the positive and negative parts of the coordinate axes lie outside the switching hypersurface on opposite sides of the surface and the optimal control for the points of these axes is exactly with n number of intervals of constancy.
Then, finding the optimal control at the initial point
of Problem
becomes a significantly easier task because it only requires solving the easier Problem
, which significantly reduces the computational load and the knowledge of the relation of the positive or negative parts of the axis
to the switching hyper-surface. The latter can again be obtained by solving the not so difficult problem of the lower order, but under specific initial conditions [
14,
15,
16]. These data can be retrieved in advance and the process is called “axes initialization”. Relying on a straightforward geometrical concept, this advantage of the approach is of significant benefit when solving high-order problems by immersing the initial problem in a class of problems Problem
, Problem
, … Problem
and returning by reverse order to the initial Problem
.
Besides the property proved here, there is still a rigorous need to prove other properties of the problem in the case of its expansion. In [
18], the author presents several interesting results of numerical experiments for near time-optimal control of a scanning lidar system based on the described method. Furthermore, the numerical aspects of the currently developed technique imply a close connection with linear programming.