Next Article in Journal
Identification of the Four-Bar Linkage Size in a Beam Pumping Unit Based on Cubature Kalman Filter
Previous Article in Journal
Optimization of EDM Machinability of Hastelloy C22 Super Alloys
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Consistency of Approximation of Bernstein Polynomial-Based Direct Methods for Optimal Control

1
Department of Mechanical Engineering, University of Iowa, Iowa City, IA 52242, USA
2
Department of Mechanical and Aerospace Engineering, Naval Postgraduate School, Monterey, CA 93940, USA
3
Department of Electrical and Computer Engineering, University of Texas at San Antonio, San Antonio, TX 78249, USA
4
Department of Mechanical Science and Engineering, University of Illinois at Urbana-Champaign, Urbana, IL 61801, USA
5
Institute for Systems and Robotics (ISR), Instituto Superior Tecnico (IST), University of Lisbon, 1049-001 Lisbon, Portugal
*
Author to whom correspondence should be addressed.
Machines 2022, 10(12), 1132; https://doi.org/10.3390/machines10121132
Submission received: 28 September 2022 / Revised: 14 November 2022 / Accepted: 23 November 2022 / Published: 28 November 2022

Abstract

:
Bernstein polynomial approximation of continuous function has a slower rate of convergence compared to other approximation methods. “The fact seems to have precluded any numerical application of Bernstein polynomials from having been made. Perhaps they will find application when the properties of the approximant in the large are of more importance than the closeness of the approximation.”—remarked P.J. Davis in his 1963 book, Interpolation and Approximation. This paper presents a direct approximation method for nonlinear optimal control problems with mixed input and state constraints based on Bernstein polynomial approximation. We provide a rigorous analysis showing that the proposed method yields consistent approximations of time-continuous optimal control problems and can be used for costate estimation of the optimal control problems. This result leads to the formulation of the Covector Mapping Theorem for Bernstein polynomial approximation. Finally, we explore the numerical and geometric properties of Bernstein polynomials, and illustrate the advantages of the proposed approximation method through several numerical examples.

1. Introduction

Motion planning plays an important role in enabling robotic systems to accomplish tasks assigned to them autonomously, safely and reliably. Over the past decades, many approaches to generating trajectories have been proposed. Examples include bug algorithms, artificial potential functions, roadmap path planners, cell decomposition methods, and optimal control-based trajectory generation. The reader is referred to [1,2,3,4,5,6,7,8] and references therein for detailed discussions and comparisons of these methods. Each technique has different advantages and disadvantages, and is best suited to certain types of problems. Motion planning based on optimal control, i.e., optimal motion planning, is particularly suitable for applications that require the trajectory to optimize some costs while guaranteeing satisfaction of a complex set of vehicle and problem constraints. These applications include multi-robot road search [9], coordinated tracking [10], optimal and constrained formation control [11], and adversarial swarm defense [12].
Optimal control problems that arise from robotics and motion-planning applications are, in general, very complex. Finding a closed-form solution to these problems can be difficult or even impossible, and therefore they must be solved numerically. Numerical methods include indirect and direct methods [13]. Indirect methods solve the problems by converting them into boundary value problems. Then, the solutions are found by solving systems of differential equations. On the other hand, direct methods are based on transcribing optimal control problems into nonlinear programming problems (NLPs) using a discretization scheme [6,13,14,15]. These NLPs can be solved using ready-to-use NLP solvers (e.g., MATLAB, SNOPT, etc.) and do not require calculation of costate and adjoint variables, as indirect methods do.
A wide range of direct methods that use different discretization schemes have been developed, including direct single shootings, direct multiple shooting and direct collocation methods [6,13,14,15]. The software packages that implement some of these methods (e.g., PSOPT [16], NLOptControl [17], GPOPS II [18], PROPT [19], DIDO [20] and CasADi [21]) are particularly relevant; some of these have been applied successfully to solve a wide range of real-world problems [22,23,24,25,26,27,28]. Theoretical results in the literature on direct methods include those related to consistency of approximation theory; see [29], which provides a framework to assess the convergence properties of Euler and Range–Kutta discretization schemes. Motivated by the consistency of approximation theory, direct methods that use different discretization schemes have been developed, including Pseudospectral methods based on Legendre, Chebyshev and Lagrange polynomials [28]. One drawback of direct methods is that the costate of the original optimal control problem cannot be readily obtained from the approximated solution. Nevertheless, in several applications—such as motion planning and control for safety-critical robotic systems—the knowledge of the costate is important because it allows for the evaluation of the fulfillment of necessary conditions of optimality. This evaluation, in turn, provides important insights into the validity and optimality of the solution. Therefore, approaches for obtaining estimates of the costate from direct methods have been proposed in the literature on direct collocation [30,31,32] and direct shooting [33].
In [34] we presented a direct method based on Bernstein polynomials. We showed that the geometric properties of these polynomials allow for the implementation of efficient algorithms for the computation of state and input constraints, which are particularly useful for motion planning and trajectory generation applications [35,36]. Additional works that exploit the properties of Bernstein polynomials for nonlinear optimal control can be found in [37,38,39,40,41]. Furthermore, in [42] we used the approximation properties of Bernstein polynomials to derive consistency and convergence results for the proposed direct method. In the present paper, we propose an approximation scheme for primal and dual optimal control problems based on Bernstein polynomials. In particular, we propose an approach to approximate the costate of a general non-linear optimal control problem of Bolza type using the Lagrange multipliers of the Bernstein polynomial-based discrete approximation. We derive transformations that relate the Lagrange multipliers of the nonlinear programming problem to the costate of the original optimal control problem. These transformations are often referred to as covector mapping in the literature on direct methods for optimal control [28,29,43]. Finally, we demonstrate uniform convergence properties of the method.
The paper is structured as follows: in Section 2, we present the notation and the mathematical results, which will be used later in the paper. Section 3 introduces the optimal control problem of interest and some related assumptions, and presents the approximation method based on Bernstein approximation that approximates the optimal control problem into an NLP. In Section 4 we derive the Karush–Kuhn–Tucker (KKT) conditions associated with the NLP. Section 5 compares these conditions to the first-order optimality conditions for the original optimal control problem and states the Covector Mapping Theorem for Bernstein approximation. Numerical examples are discussed in Section 6, while Section 7 highlights the significance of the theoretical findings applied to a specific multi-robot simulation scenario, namely optimal defense against swarm attacks. The paper ends with conclusions in Section 8.

2. Notation and Mathematical Background

Vector-valued functions are denoted by bold letters, x ( t ) = [ x 1 ( t ) , , x n ( t ) ] , while vectors are denoted by bold letters with an upper bar, x ¯ = [ x 1 , , x n ] R n . The symbol C r denotes the space of functions with r continuous derivatives. C n r denotes the space of n-vector valued functions in C r . | | · | | denotes the Euclidean norm, | | x ¯ | | = x 1 2 + + x n 2 .
The Bernstein basis polynomials of degree N are defined as
b j , N ( t ) = N j t j ( 1 t ) N j , t [ 0 , 1 ] ,
for j = 0 , , N , with N j = N ! j ! ( N j ) ! . A Nth-order Bernstein polynomial x N : [ 0 , 1 ] R is a linear combination of N + 1 Bernstein basis polynomials of order N, i.e.,
x N ( t ) = j = 0 N x ¯ j b j , N ( t ) , t [ 0 , 1 ] ,
where x ¯ j R , j = 0 , , N , are referred to as Bernstein coefficients. For the sake of generality, and with a slight abuse of terminology, in this paper, we extend the definition of a Bernstein polynomial given above to a vector of Nth-order polynomials x N : [ 0 , 1 ] R n expressed in the following form
x N ( t ) = j = 0 N x ¯ j , N b j , N ( t ) , t [ 0 , 1 ] ,
where x ¯ 0 , N , , x ¯ N , N R n .
In what follows, we provide a review of numerical properties of Bernstein polynomials that are used throughout this paper. The derivative and integral of a Bernstein polynomial x N ( t ) can be easily computed as
x ˙ N ( t ) = N j = 0 N 1 ( x ¯ j + 1 , N x ¯ j , N ) b j , N 1 ( t )
and
0 1 x N ( t ) d t = w j = 0 N x ¯ j , N , w = 1 N + 1 ,
respectively.
Bernstein polynomials can be used to approximate smooth functions. Consider a n-vector valued function x : [ 0 , 1 ] R n . The Nth order Bernstein approximation of x ( t ) is a vector of Bernstein polynomials x N ( t ) computed as in (1) with x ¯ j , N = x ( t j ) and t j = j N for all j = 0 , , N . Namely,
x ( t ) x N ( t ) = j = 0 N x ( t j ) b j , N ( t ) , t j = j N .
The following results hold for Bernstein approximations.
Lemma 1
(Uniform convergence of Bernstein approximation). Let x ( t ) C n 0 on [ 0 , 1 ] , and let x N ( t ) be computed as in Equation (3). Then, for arbitrary order of approximation N Z + , the Bernstein approximation x N ( t ) satisfies
| | x N ( t ) x ( t ) | | C 0 W x ( N 1 2 ) ,
where C 0 is a positive constant satisfying C 0 < 5 n / 4 , and W x ( · ) is the modulus of continuity of x ( t ) in [ 0 , 1 ] [44,45,46]. Moreover, if x ( t ) C n 1 , then
x ˙ N ( t ) x ˙ ( t ) C 1 W x ( N 1 2 ) ,
where C 1 is a positive constant satisfying C 1 < 9 n / 4 and W x ( · ) is the modulus of continuity of x ˙ ( t ) in [ 0 , 1 ] [47].
Lemma 2
([48]). Assume x ( t ) C n r + 2 , r 0 , and let x N ( t ) be computed as in Equation (3). Let x ( r ) ( t ) denote the rth derivative of x ( t ) . Then, the following inequalities hold for all t [ 0 , 1 ] :
| | x N ( t ) x ( t ) | | C 0 N , | | x N ( r ) ( t ) x ( r ) ( t ) | | C r N ,
where C 0 , , C r are independent of N.
Lemma 3.
If x ( t ) C n 0 on [ 0 , 1 ] , then we have
0 1 x ( t ) d t w j = 0 N x j N C I W x ( N 1 2 ) ,
with w = 1 N + 1 , where C I > 0 is independent of N. Moreover, if x ( t ) C n 2 , then
0 1 x ( t ) d t w j = 0 N x j N C I N .
The Lemma above follows directly from Lemmas 1 and 2 and Equation (2).
The following property of Bernstein polynomials is relevant to this paper.
Property 1
(End point values). The Bernstein polynomial given by Equation (1) satisfies x N ( 0 ) = x ¯ 0 , N and x N ( 1 ) = x ¯ N , N .

3. Problem Formulation

This paper considers the following optimal control problem:
Problem 1
(Problem P). Determine x : [ 0 , 1 ] R n x and u : [ 0 , 1 ] R n u that minimize
I ( x ( t ) , u ( t ) ) = E ( x ( 0 ) , x ( 1 ) ) + 0 1 F ( x ( t ) , u ( t ) ) d t ,
subject to
x ˙ = f ( x ( t ) , u ( t ) ) , t [ 0 , 1 ] ,
e ( x ( 0 ) , x ( 1 ) ) = 0 ,
h ( x ( t ) , u ( t ) ) 0 , t [ 0 , 1 ] ,
where E : R n x × R n x R and F : R n x × R n u R are the terminal and running costs, respectively, f : R n x × R n u R n x describes the system dynamics, e : R n x × R n x R n e is the vector of boundary conditions, and h : R n x × R n u R n h is the vector of state and input constraints.
Next, we formulate a discretized version of Problem P, here referred to as Problem P N , where N denotes the order of approximation. This requires that we approximate the input and state functions, the cost function, the system dynamics and the equality and inequality constraints in Problem P. First, consider the following Nth-order vectors of Bernstein polynomials:
x N ( t ) = j = 0 N x ¯ j , N b j , N ( t ) , u N ( t ) = j = 0 N u ¯ j , N b j , N ( t ) ,
with x N : [ 0 , 1 ] R n x , u N : [ 0 , 1 ] R n u , x ¯ j , N R n x and u ¯ j , N R n u . Let x ¯ N R n x × ( N + 1 ) and u ¯ N R n u × ( N + 1 ) be defined as
x ¯ N = [ x ¯ 0 , N , , x ¯ N , N ] , u ¯ N = [ u ¯ 0 , N , , u ¯ N , N ] .
Let 0 = t 0 < t 1 < < t N = 1 be a set of equidistant time nodes, i.e., t j = j N . Then, Problem P N can be stated as follows:
Problem 2
(Problem P N ). Determine x ¯ N and u ¯ N that minimize
I N ( x ¯ N , u ¯ N ) = E ( x N ( 0 ) , x N ( t N ) ) + w j = 0 N F ( x N ( t j ) , u N ( t j ) ) ,
subject to
x ˙ N ( t j ) f ( x N ( t j ) , u N ( t j ) ) δ P N , j = 0 , , N ,
e ( x N ( 0 ) , x N ( t N ) ) = 0 ,
h ( x N ( t j ) , u N ( t j ) ) δ P N 1 , j = 0 , , N ,
where w = 1 N + 1 , and δ P N is a small positive number that depends on N and converges uniformly to 0, i.e., lim N δ P N = 0 .
Remark 1.
Compared to the constraints of Problem P, the dynamic and inequality constraints given by Equations (10) and (12) are relaxed. Motivated by previous work on consistency of approximation theory [29], the bound δ P N , referred to as relaxation bound, is introduced to guarantee that Problem P N has a feasible solution. As will become clear later, the relaxation bound can be made arbitrarily small by choosing a sufficiently large order of approximation N. Furthermore, note that when N , the right-hand sides of Equations (10) and (12) are equal to zero, i.e., the difference between the constraints imposed by Problems P and P N vanishes.
Remark 2.
The outcome of Problem P N is a set of optimal Bernstein coefficients x ¯ N * and u ¯ N * that determine the vectors of Bernstein polynomials x N * ( t ) and u N * ( t ) , i.e.,
x N * ( t ) = j = 0 N x ¯ j , N * b j , N ( t ) , u N * ( t ) = j = 0 N u ¯ j , N * b j , N ( t ) .
In our previous work, see [42], we provide theoretical results demonstrating: (i) the existence of a feasible solution to Problem P N , and (ii) the convergence of the pair ( x N * ( t ) , u N * ( t ) ) to the optimal solution of Problem P, given by ( x * ( t ) , u * ( t ) ) . Nevertheless, the present paper focuses on the existence and convergence of the estimates of the costates of Problem P, which are introduced next.

4. Costate Estimation for Problem P

4.1. First-Order Optimality Conditions of Problem P

We start by deriving the first-order necessary conditions for Problem P. Let λ ( t ) : [ 0 , 1 ] R n x be the costate trajectory, and let μ ( t ) : [ 0 , 1 ] R n h and ν R n e be the multipliers. By defining the Lagrangian of the Hamiltonian (also known as the D-form [49]) as
L ( x ( t ) , u ( t ) , λ ( t ) , μ ( t ) ) = H ( x ( t ) , u ( t ) , λ ( t ) ) + μ ( t ) h ( x ( t ) , u ( t ) ) ,
where the Hamiltonian H is given by
H ( x ( t ) , u ( t ) , λ ( t ) ) = F ( x ( t ) , u ( t ) ) + λ ( t ) f ( x ( t ) , u ( t ) ) ,
the dual of Problem P can be formulated as follows [49].
Problem 3
(Problem P λ ). Determine x ( t ) , u ( t ) , λ ( t ) , μ ( t ) and ν that for all t [ 0 , 1 ] satisfy Equations (5)–(7) and
μ ( t ) h ( x ( t ) , u ( t ) ) = 0 , μ ( t ) 0 ,
λ ˙ ( t ) + L x ( x ( t ) , u ( t ) , λ ( t ) , μ ( t ) ) = 0 ,
λ ( 0 ) = ν e x ( 0 ) ( x ( 0 ) , x ( 1 ) ) E x ( 0 ) ( x ( 0 ) , x ( 1 ) ) ,
λ ( 1 ) = ν e x ( 1 ) ( x ( 0 ) , x ( 1 ) ) + E x ( 1 ) ( x ( 0 ) , x ( 1 ) ) ,
L u ( x ( t ) , u ( t ) , λ ( t ) , μ ( t ) ) = 0 .
In the above problem, subscripts are used to denote partial derivatives, e.g., F x ( x , u ) = x F ( x , u ) .
The following assumptions are imposed onto Problem P λ .
Assumption 1.
E , F , f , e and h are continuously differentiable with respect to their arguments, and their gradients are Lipschitz continuous over the domain.
Assumption 2.
Solutions x * ( t ) , u * ( t ) , λ * ( t ) , μ * ( t ) and ν * of Problem P λ exist and satisfy x * ( t ) C n x 1 , u * ( t ) C n u 0 , λ * ( t ) C n x 1 and μ * ( t ) C n h 0 in [ 0 , 1 ] .
Remark 3.
Notice that Problem P λ implicitly assumes the absence of pure state constraints in Problem P. If the inequality constraint in Equation (7) is independent of u ( t ) , then the costate λ ( t ) must also satisfy the following jump condition [49]:
λ ( t e ) = λ ( t e + ) + h x ( t e ) η ,
where t e is the entry or exit time into a constrained arc in which the inequality constraint is active, t e and t e + denote the left-hand side and right-hand side limits of the trajectory, respectively, and η is a constant covector. For simplicity, the theoretical results that will be presented in Section 5 do not consider the jump conditions above, i.e., the inequality constraints are dependent on u ( t ) . Nevertheless, numerical examples will be presented in Section 6, showing the applicability of the discretization method to pure state-constrained problems.

4.2. KKT Conditions of Problem P N

Now, we derive the necessary conditions of Problem P N . Let us introduce the following Nth-order Bernstein polynomials:
λ N ( t ) = j = 0 N λ ¯ j , N b j , N ( t ) , μ N ( t ) = j = 0 N μ ¯ j , N b j , N ( t ) ,
with λ N : [ 0 , 1 ] R n x , μ N : [ 0 , 1 ] R n h , λ ¯ j , N R n x and μ ¯ j , N R n h , and the vector ν ¯ R n e . Finally, let λ ¯ N R n x × ( N + 1 ) and μ ¯ N R n u × ( N + 1 ) be defined as
λ ¯ N = [ λ ¯ 0 , N , , λ ¯ N , N ] , μ ¯ N = [ μ ¯ 0 , N , , μ ¯ N , N ] .
With the above notation, the Lagrangian for problem P N can be written as
L N = E ( x N ( 0 ) , x N ( t N ) ) + w j = 0 N F ( x N ( t j ) , u N ( t j ) ) + j = 0 N λ N ( t j ) ( x ˙ N ( t j ) + f ( x N ( t j ) , u N ( t j ) ) ) + j = 0 N μ N ( t j ) h ( x N ( t j ) , u N ( t j ) ) + ν ¯ e ( x N ( 0 ) , x N ( t N ) ) .
Then, the duality of Problem P N can be stated as follows:
Problem 4
(Problem P N λ ). Determine x ¯ N , u ¯ N , λ ¯ N , μ ¯ N and ν ¯ that satisfy the primal feasibility conditions, namely Equations (10)–(12), the complementary slackness and dual feasibility conditions
μ N ( t k ) h ( x N ( t k ) , u N ( t k ) ) N 1 δ D N , μ N ( t k ) N 1 δ D N 1 , k = 0 , , N ,
and the stationarity conditions
L N x ¯ k , N δ D N , L N u ¯ k , N δ D N , k = 0 , , N ,
where δ D N is a small positive number that depends on N and satisfies lim N δ D N = 0 .
At this point, similarly to most results on costate estimation [50,51,52], we introduce additional conditions that must be added to Equations (10)–(12), (20) and (21) in order to obtain consistent approximations of the solutions of Problem P λ . These conditions, often referred to as closure conditions in the literature, are given as follows:
λ N ( 0 ) w + ν ¯ e x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) + E x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) δ D N ,
λ N ( t N ) w ν ¯ e x ( 1 ) ( x N ( 0 ) , x N ( t N ) ) E x ( 1 ) ( x N ( 0 ) , x N ( t N ) ) δ D N .
In other words, the closure conditions are constraints that must be added to Problem P N λ so that the solution of this problem approximates the solution of Problem P λ . We notice that the conditions given above are discrete approximations of the conditions given by Equations (16) and (17). With this setup, we define the following problem:
Problem 5
(Problem P N λ c l o s ). Determine x ¯ N , u ¯ N , λ ¯ N , μ ¯ N and ν ¯ that satisfy the primal feasibility conditions, namely Equations (10)–(12), the complementary slackness and dual feasibility conditions (20), the stationarity conditions (21), and the closure conditions (22) and (23).
The solution of Problem P N λ c l o s presents a set of optimal Bernstein coefficients x ¯ N * , u ¯ N * , λ ¯ N * , μ ¯ N * (which determine the Bernstein polynomials x N * ( t ) , u N * ( t ) , λ N * ( t ) and μ N * ( t ) ) and a vector ν ¯ * .

5. Feasibility and Consistency of Problem P N λ c l o s

The objective of this section is to investigate the ability of the solutions of Problem P N λ c l o s to approximate the solutions of Problem P λ . In what follows, we first show the existence of a solution to Problem P N λ c l o s (feasibility). Second, we investigate the convergence properties of this solution as N (consistency). Third, by combining these two results, we finally formulate the covector mapping theorem for Bernstein approximations, which provides a map between the solution of Problem P N λ c l o s and the solution of Problem P λ . The main results of this section are reported in the three theorems below and summarized in Figure 1.
Theorem 1
(Feasibility). Let
δ D N = C D max { W x ( N 1 2 ) , W x ( N 1 2 ) , W u ( N 1 2 ) , W λ ( N 1 2 ) , W λ ( N 1 2 ) , W μ ( N 1 2 ) } ,
δ P N = C P max { W x ( N 1 2 ) , W x ( N 1 2 ) , W u ( N 1 2 ) } ,
where C D and C P are positive constants independent of N, and W x ( · ) , W x ( · ) , W u ( · ) , W λ ( · ) , W λ ( · ) and W μ ( · ) are the moduli of continuity of x ˙ ( t ) , x ( t ) , u ( t ) , λ ˙ ( t ) , λ ( t ) and μ ( t ) , respectively. Then Problem P N λ c l o s is feasible for arbitrary order of approximation N Z + .
Proof. 
This proof follows by constructing a solution for Problem P N λ c l o s , with δ D N given by Equation (24). To this end, let x ( t ) , u ( t ) , λ ( t ) , μ ( t ) and ν be a solution of Problem P λ , which exists by Assumption 2, and define
x ¯ j , N = x ( t j ) , u ¯ j , N = u ( t j ) ,
λ ¯ j , N = w λ ( t j ) , μ ¯ j , N = w μ ( t j ) , ν ¯ = ν ,
for all j = 0 , , N , t j = j N , w = 1 N + 1 , with corresponding Bernstein polynomials given by
x N ( t ) = j = 0 N x ¯ j , N b j , N ( t ) , u N ( t ) = j = 0 N u ¯ j , N b j , N ( t ) , λ N ( t ) = j = 0 N λ ¯ j , N b j , N ( t ) , μ N ( t ) = j = 0 N μ ¯ j , N b j , N ( t ) .
The remainder of this proof shows that x N ( t ) , u N ( t ) , λ N ( t ) , μ N ( t ) and ν ¯ given above satisfy Equations (20)–(23). The satisfaction of Equations (10)–(12) can be demonstrated using a proof similar to the one of [42], and is thus omitted. We start by defining the Bernstein coefficients λ ¯ ˜ j , N and μ ¯ ˜ j , N as follows
λ ¯ ˜ j , N = λ ¯ j , N w , μ ¯ ˜ j , N = μ ¯ j , N w ,
with corresponding Bernstein polynomials given by
λ ˜ N ( t ) = j = 0 N λ ¯ ˜ j , N b j , N ( t ) , μ ˜ N ( t ) = j = 0 N μ ¯ ˜ j , N b j , N ( t ) .
Notice that
λ ˜ N ( t ) = λ N ( t ) w , μ ˜ N ( t ) = μ N ( t ) w .
Combining Equations (26), (27) and (29) and using Assumption 2 and Lemma 1, we get
| | x N ( t ) x ( t ) | | C x W x ( N 1 2 ) , | | u N ( t ) u ( t ) | | C u W u ( N 1 2 ) , | | x ˙ N ( t ) x ˙ ( t ) | | C x W x ( N 1 2 ) , | | λ ˜ N ( t ) λ ( t ) | | C λ W λ ( N 1 2 ) , | | μ ˜ N ( t ) μ ( t ) | | C μ W μ ( N 1 2 ) , | | λ ˜ ˙ N ( t ) λ ˙ ( t ) | | C λ W λ ( N 1 2 ) ,
where C λ < 5 n x 4 , C μ < 5 n h 4 , C λ < 9 n x 4 and W λ ( · ) , W μ ( · ) and W μ ( · ) are the moduli of continuity of λ ( t ) , μ ( t ) and λ ˙ ( t ) , respectively.
Now, we show that the bound in Equation (20) is satisfied. Using Equation (30), and adding and subtracting w ( μ ( t k ) h ( x N ( t k ) , u N ( t k ) ) + μ ( t k ) h ( x ( t k ) , u ( t k ) ) ) , we get
μ N ( t k ) h ( x N ( t k ) , u N ( t k ) ) = w μ ˜ N ( t k ) h ( x N ( t k ) , u N ( t k ) ) | | w ( μ ˜ N ( t k ) μ ( t k ) ) h ( x N ( t k ) , u N ( t k ) ) + w μ ( t k ) h ( x ( t k ) , u ( t k ) ) + w μ ( t k ) ( h ( x N ( t k ) , u N ( t k ) ) h ( x ( t k ) , u ( t k ) )
Using Equation (14), the above inequality reduces to
μ N ( t k ) h ( x N ( t k ) , u N ( t k ) ) w ( μ ˜ N ( t k ) μ ( t k ) ) h ( x N ( t k ) , u N ( t k ) ) + w μ ( t k ) ( h ( x N ( t k ) , u N ( t k ) ) h ( x ( t k ) , u ( t k ) ) w | | h ( x N ( t k ) , u N ( t k ) ) | | C μ W μ ( N 1 2 ) + w μ ( t k ) L h ( C x W x ( N 1 2 ) + C u W u ( N 1 2 ) ) ,
where we used the bounds in Equation (31) together with the Lipschitz assumption on h (see Assumptions 1). Finally, from using Assumptions 1 and 2, it follows that h and μ are bounded on [ 0 , 1 ] with bounds h max and μ max , respectively. Therefore, we get
μ N ( t k ) h ( x N ( t k ) , u N ( t k ) ) w [ h max C μ W μ ( N 1 2 ) + μ max L h ( C x W x ( N 1 2 ) + C u W u ( N 1 2 ) ) ] ,
which implies that the bound in Equation (20) is satisfied with δ D N given by Equation (24) and C D > h max C μ + μ max L h ( C x + C u ) . Similarly,
μ N ( t k ) = w μ ˜ N ( t k ) w μ ( t k ) w μ ( t k ) μ ˜ N ( t k ) 1 N 1 C μ W μ ( N 1 2 ) 1 ,
which proves that Equation (20) holds.
Now, consider the left equation in (21). For k = 0 , we have
L N x ¯ 0 , N = E x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) + w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) + j = 0 N λ N ( t j ) f x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) b ˙ 0 , N ( t j ) + j = 0 N μ N ( t j ) h x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) + ν ¯ e x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) .
Substituting w λ ˜ N ( t j ) = λ N ( t j ) and w μ ˜ N ( t j ) = μ N ( t j ) , the equation above can be written as
L N x ¯ 0 , N = E x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) + w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) + w j = 0 N λ ˜ N ( t j ) f x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) b ˙ 0 , N ( t j ) + w j = 0 N μ ˜ N ( t j ) h x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) + ν ¯ e x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) .
Notice that the following inequalities are satisfied:
w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t C ¯ 1 ( N 1 2 + W x ( N 1 2 ) + W u ( N 1 2 ) ) ,
w j = 0 N λ ˜ N ( t j ) b ˙ 0 , N ( t j ) 0 1 λ ( t ) b ˙ 0 , N ( t ) d t C ¯ 2 ( N 1 2 + W λ ( N 1 2 ) ) ,
w j = 0 N λ ˜ N ( t j ) f x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 λ ( t ) f x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t C ¯ 3 ( N 1 2 + W λ ( N 1 2 ) + W x ( N 1 2 ) + W u ( N 1 2 ) ) ,
w j = 0 N μ ˜ N ( t j ) h x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 μ ( t ) h x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t C ¯ 4 ( N 1 2 + W μ ( N 1 2 ) + W x ( N 1 2 ) + W u ( N 1 2 ) ) ,
for some positive C ¯ 1 , C ¯ 2 , C ¯ 3 and C ¯ 4 independent of N. Proof of the above inequalities is given in Appendix A. Then, the combination of Equations (33) and (34) yields the following inequality
L N x ¯ 0 , N E x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t 0 1 λ ( t ) b ˙ 0 , N ( t ) d t + 0 1 λ ( t ) f x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t + 0 1 μ ( t ) h x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t + ν ¯ e x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + C ¯ max { N 1 2 , W x ( N 1 2 ) , W u ( N 1 2 ) , W λ ( N 1 2 ) , W μ ( N 1 2 ) } ,
with C ¯ 4 max { C ¯ 1 , C ¯ 2 , C ¯ 3 , C ¯ 4 } . Using integration by parts, we have 0 1 λ ( t ) b ˙ 0 , N ( t ) d t = 0 1 λ ˙ ( t ) b 0 , N ( t ) d t + [ λ ( t ) b 0 , N ( t ) ] 0 1 . Thus, since b 0 , N ( 0 ) = 1 , b N , N ( 0 ) = 0 , the above inequality becomes
L N x ¯ 0 , N E x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + λ ( 0 ) + ν e x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + 0 1 ( λ ˙ ( t ) + F x ( x ( t ) , u ( t ) ) + λ ( t ) f x ( x ( t ) , u ( t ) ) + μ ( t ) h x ( x ( t ) , u ( t ) ) ) b 0 , N ( t ) d t + C ¯ max { N 1 2 , W x ( N 1 2 ) , W u ( N 1 2 ) , W λ ( N 1 2 ) , W μ ( N 1 2 ) } .
Finally, using Equations (15) and (16), the above inequality reduces to the left condition in Equation (21) for k = 0 , with δ D N given by Equation (24) and C D C ¯ . The same condition for k = 1 , , N can be shown to be satisfied using an identical argument. The stationarity condition in the right of Equation (21) can also be verified similarly, and the computations are thus omitted. To show that the closure condition (22) is satisfied, we use the definitions in Equations (26) and (27) together with the end point values property of Bernstein polynomials, Property 1 in Section 2, which gives
λ N ( 0 ) w + ν ¯ e x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) + E x ( 0 ) ( x N ( 0 ) , x N ( t N ) ) λ ( 0 ) + ν e x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + E x ( 0 ) ( x ( 0 ) , x ( 1 ) ) = 0 ,
where the last equality follows from Equation (16). An identical argument can be used to show that the closure condition (23) holds, thus completing the proof of Theorem 1.
Corollary 1.
If solutions x * ( t ) , u * ( t ) , λ * ( t ) , μ * ( t ) and ν * of Problem P λ exist and satisfy x ˙ * ( t ) C n x 2 , u * ( t ) C n u 2 , λ ˙ * ( t ) C n x 2 , and μ * ( t ) C n h 2 in [ 0 , 1 ] , then Theorem 1 holds with δ P N = C P N 1 and δ D N = C D N 1 , where C P and C D are positive constants independent of the order of approximation, N.
Proof. 
The proof of Corollary 1 follows easily by applying Lemma 2 to the proof of Theorem 1.
Remark 4.
We notice that for arbitrarily small scalar ϵ D > 0 , there exists N 1 such that for all N N 1 , we have δ D N ϵ D ; i.e., the relaxation bound in Problem P N λ clos can be made arbitrarily small by choosing sufficiently large N.
Theorem 2 (Consistency)
Let { ( x ¯ N * , u ¯ u * , λ ¯ N * , μ ¯ N * , ν ¯ * ) } N = N 1 be a sequence of solutions of Problem P N λ clos . Consider the sequence of transformed solutions { ( x ¯ N * , u ¯ N * , λ ¯ ˜ N * , μ ¯ ˜ N * , ν ¯ * ) } N = N 1 , with
λ ¯ ˜ j , N * = λ ¯ j , N * w , μ ¯ ˜ j , N * = λ ¯ j , N * w ,
and the corresponding polynomial approximation { ( x N * ( t ) , u N * ( t ) , λ ˜ N * ( t ) , μ ˜ N * ( t ) , ν ¯ * ) } N = N 1 . Assume that the latter has a uniform accumulation point, i.e.,
lim N ( x N * ( t ) , u N * ( t ) , λ ˜ N * ( t ) , μ ˜ N * ( t ) , ν ¯ * ) = ( x ( t ) , u ( t ) , λ ˜ ( t ) , μ ˜ ( t ) , ν ¯ ) , t [ 0 , 1 ] ,
and assume x ˙ ( t ) , u ( t ) , λ ˜ ˙ ( t ) and μ ˜ ( t ) are continuous on [ 0 , 1 ] . Then,
( x ( t ) , u ( t ) , λ ˜ ( t ) , μ ˜ ( t ) , ν ¯ )
is a solution of Problem P λ .
Proof. 
The objective is to show that x ( t ) , u ( t ) , λ ˜ ( t ) , μ ˜ ( t ) and ν ¯ satisfy Equations (5)–(6) and (14)–(18). The satisfaction of Equations (5)–(7) has been demonstrated in ([42] [Proof of Theorem 2]). We start by showing Equation (14), and we do so using a proof by contradiction. Assume that x ( t ) , u ( t ) , μ ˜ ( t ) do not satisfy Equation (14). Then, there exists t [ 0 , 1 ] , such that
μ ˜ ( t ) h ( x ( t ) , u ( t ) ) > 0 .
Since the nodes { t k } k = 0 N are dense in [ 0 , 1 ] , there exists a sequence of indices { k N } N = 0 , such that
lim N t k N = t ,
which implies
lim N μ ˜ ( t ) μ ˜ ( t k N ) = 0 ,
lim N x ( t ) x ( t k N ) = 0 ,
lim N u ( t ) u ( t k N ) = 0 .
Then, we have
| | μ ˜ ( t ) h ( x ( t ) , u ( t ) ) | | lim N | | ( μ ˜ N * ( t ) μ ˜ N * ( t k N ) ) h ( x N * ( t ) , u N * ( t ) ) | | + lim N | | μ ˜ N * ( t k N ) ( h ( x N * ( t ) , u N * ( t ) ) h ( x N * ( t k N ) , u N * ( t k N ) ) ) | | + lim N | | μ ˜ N * ( t k N ) h ( x N * ( t k N ) , u N * ( t k N ) ) | | = lim N 1 w | | μ N * ( t k N ) h ( x N * ( t k N ) , u N * ( t k N ) ) | | = 0 ,
where we used Equation (20). This contradicts Equation (38). Similarly, we can show that μ ˜ ( t ) 0 , thus proving that x ( t ) , u ( t ) and μ ˜ ( t ) satisfy Equation (14).
Furthermore, we notice that if x ( t ) , u ( t ) , λ ˜ ( t ) , μ ˜ ( t ) and ν ¯ satisfy Equations (21)–(23), then the following holds for all k = 0 , , N :
λ ˜ ( 0 ) + ν ¯ e x ( 0 ) ( x ( 0 ) , x ( 1 ) ) + E x ( 0 ) ( x ( 0 ) , x ( 1 ) ) = 0 ,
λ ( 1 ) ν ¯ e x ( 1 ) ( x ( 0 ) , x ( 1 ) ) E x ( 1 ) ( x ( 0 ) , x ( 1 ) ) = 0 ,
0 1 [ λ ˜ ˙ ( t ) + F x ( x ( t ) , u ( t ) ) + λ ˜ ( t ) f x ( x ( t ) , u ( t ) ) + μ ˜ ( t ) h x ( x ( t ) , u ( t ) ) ] b k , N ( t ) d t = 0 ,
0 1 [ F u ( x ( t ) , u ( t ) ) + λ ˜ ( t ) f u ( x ( t ) , u ( t ) ) + μ ˜ ( t ) h u ( x ( t ) , u ( t ) ) ] b k , N ( t ) d t = 0 .
Since { b k , N ( t ) } k = 0 N is a linearly independent basis set, the last two equations above imply
λ ˜ ˙ ( t ) + F x ( x ( t ) , u ( t ) ) + λ ˜ ( t ) f x ( x ( t ) , u ( t ) ) + μ ˜ ( t ) h x ( x ( t ) , u ( t ) ) = 0 ,
F u ( x ( t ) , u ( t ) ) + λ ˜ ( t ) f u ( x ( t ) , u ( t ) ) + μ ˜ ( t ) h u ( x ( t ) , u ( t ) ) = 0 ,
for all t [ 0 , 1 ] . This proves that x ( t ) , u ( t ) , λ ˜ ( t ) , μ ˜ ( t ) and ν ¯ ( t ) satisfy Equations (15)–(18). □
Theorem 3
(Covector Mapping Theorem). Under the same assumptions of Theorems 1 and 2, when N , the covector mapping
x N * ( t ) x * ( t ) , u N * ( t ) u * ( t ) , λ N * ( t ) w λ * ( t ) , μ N * ( t ) w μ * ( t ) , ν ¯ * ν *
is a bijective mapping between the solution of Problem P N λ clos and the solution of Problem P λ .
Proof. 
The above result follows directly from Theorems 1 and 2. In fact, if
{ x * ( t ) , u * ( t ) , λ * ( t ) , μ * ( t ) , ν * }
is a solution to Problem P λ , which exists by Assumption 2, then from Theorem 1, it follows that { x * ( t ) , u * ( t ) , w λ * ( t ) , w μ * ( t ) , ν * } is a solution to Problem P N λ clos (see Equations (26)–(28)). Conversely, by using Equation (37), a solution
( x N * ( t ) , u N * ( t ) , λ N * ( t ) , μ N * ( t ) , ν ¯ * )
that solves Problem P N λ clos provides a solution
( x N * ( t ) , u N * ( t ) , λ N * ( t ) w , μ N * ( t ) w , ν ¯ * ) = ( x N * ( t ) , u N * ( t ) , λ ˜ N * ( t ) , μ ˜ N * ( t ) , ν ¯ * )
that converges to a solution to Problem P λ (see Theorem 2). □
Remark 5.
Define the Hamiltonian approximation
H N ( t ) = F ( x N * ( t ) , u N * ( t ) ) + λ N * ( t ) w f ( x N * ( t ) , u N * ( t ) ) ,
then, Theorem 3 implies lim N H N ( t ) = H ( t ) .

6. Numerical Examples

6.1. Example 1: 1D Minimum Time Problem

The first example we consider is the classical minimum time problem for a double integrator plant.
min u J = t f ,
subject to
x ˙ 1 = x 2 , x ˙ 2 = u , x 1 ( 0 ) = 1 , x 2 ( 0 ) = 1 x 1 ( t f ) = x 2 ( t f ) = 0 , | u ( t ) | 1 , t [ 0 , t f ] .
The analytical solution to this problem is well known: the optimal control input is bang-bang and the Hamiltonian along the optimal trajectories is equal to −1 [53]. Figure 2 includes plots of the state and control trajectories for this problem for N = 45 , as well as a graph of the analytical solution. It is clear that the Bernstein polynomial solution captures the switching time precisely and closely approximates the optimal control input solution. Figure 3 includes the plot of the Hamiltonian approximation H N computed using Covector Mapping Theorem. As expected, it is equal to −1 with the exception of a small bump around the switching time.

6.2. Example 2: 3D Minimum Time Problem

In this example, we consider a minimum-time problem for a simplified 3D model of a multi-rotor drone. The vehicle is asked to reach the origin in minimum time from a given initial condition with all the control inputs bounded by ± 1 . Unlike Example 1, there is no known analytical solution to this problem. However, we know that the Hamiltonian along the optimal trajectories is equal to −1 [53].
min u J = t f ,
subject to
x ˙ 1 = x 4 , x ˙ 2 = x 5 , x ˙ 3 = x 6 , x ˙ 4 = u 1 , x ˙ 5 = u 2 , x ˙ 6 = g + u 3 , x 10 = 1 x 20 = 2 , x 30 = 3 , x 40 = 1 , x 50 = 1 , x 60 = 1 , x 1 ( t f ) = x 2 ( t f ) = x 3 ( t f ) = 0 , x 4 ( t f ) = x 5 ( t f ) = x 6 ( t f ) = 0 , | u 1 ( t ) | 1 , | u 2 ( t ) | 1 , | g + u 3 ( t ) | 1 , t [ 0 , t f ] .
Figure 4 shows the 3D plot of the position of the vehicle. The vehicle clearly reaches the origin from the given initial condition. Figure 5 includes graphs of the control inputs. They satisfy the ± 10 bound imposed in the problem formulation. Finally, Figure 6 shows the plot of the Hamiltonian approximation H N , N = 45 . It is equal to −1, as predicted by theory, thus indicating that the solution obtained is indeed close to optimal.

7. Defense against a Swarm Attack

The numerical analysis presented here involves a scenario in which an enemy swarm is attempting to destroy an high-value unit (HVU). The HVU is defended by a number of defending agents whose trajectories are optimized to maximize the probability of the HVU survival. The attacking agents dynamics are defined using Leonard swarm dynamics model [54]. A virtual leader is guiding the attacking swarm towards the HVU’s position. The attacking and defending agents are equipped with weapons systems which allow them to inflict damage on each other. The attacking agents inflict damage on the defending agents and try to destroy the HVU. The defending agents inflict damage on the attacking agents and attempt to destroy them or herd them away.
Attacking agent   i { 1 , , N A } has position x i ( t ) R 3 and defending agent k { 1 , , N D } has position s k ( t ) R 3 . The equations of motion for attacker i is
x ¨ i = j i N f I ( x i j ) x i j x i j + k = 1 M f d ( s i k ) s i k s i k + K h i h i b x ˙ i ,
for i = 1 N . There are four terms in this equation, representing: (1) attractive and repulsive forces f I ( x i j ) from other attacking agents j, where x i j = x i x j is the distance between attackers i and j; (2) a constant “virtual leader” force with magnitude K pulling them toward the HVU’s position, where h i = h x i and h is the position of the HVU; (3) purely repulsive forces f d ( s i k ) due to defending agents, where s i k = x i s k is the distance between attacker i and defender k; and (4) a damping force proportional to the x ˙ i .
For the mathematical forms of f I and f d , we have chosen the Leonard model [54], i.e., f I and f d can be written as gradients of a scalar potential functions that depends only on x i j and s i k . The force f I is repulsive when x i j d 0 , attractive when d 0 < x i j d 1 and zero when x i j > d 1 . For f d , we only keep the repulsive term (since attackers should not be attracted to defenders), i.e., f d is repulsive when s i k s 0 and zero when s i k > s 0 .
Defending agent i’s dynamics are given by
s ¨ i = u i , i = 1 N D , s i ( t ) , u i ( t ) R 3 ,
where the absolute value of each element of u i , ( | u i j | , j = 1 , 2 , 3 ) is bounded by u max = 1 .
Mutual attrition model: for hostile swarm engagements, agents are equipped with some weapons systems. The likelihood of destruction of an agent depends on its position (how close it has come to enemies) as well as the positions of those enemy agents, since each agent’s ability to inflict damage is contingent on its own survival. To model this mutual attrition, we use a damage function to track the probability that defender k is destroyed by a shot from attacker i, and vice versa. We choose a cumulative normal distribution function, Φ , to model the damage function [55]. Next, we define (i) the attrition rate at which attacker i is destroyed due to defender k, d i k att , (ii) the attrition rate of defender k due to attacker i, d k i def , and (iii) the attrition rate of the HVU, d i hvu , as follows:
d i k att = λ d Φ s i k 2 σ d , d k i def = λ a Φ s i k 2 σ a , d i hvu = λ a Φ h i 2 σ a .
In the above equation σ d is a defender Poisson parameter that corresponds to the range of the defenders’ weapons, λ d is a defender Poisson parameter that corresponds to the defenders’ rate of fire, σ a is an attacker Poisson parameter that corresponds to the attackers’ range and λ a is an attacker Poisson parameter that corresponds to the attackers’ rate of fire. The probability of defender k destroying attacker i during a time interval of duration Δ t is weighted by the current survival probability of defender k, P k d ( t ) . Thus, the probability that defender k will destroy attacker i during a given time interval [ t , t + Δ t ] is P k d ( t ) d i k att Δ t . Assuming independence (i.e., defenders do not coordinate their fire), the expression k M ( 1 d i k att P k d ( t ) Δ t ) represents the probability that ith attacker would survive during a time interval [ t , t + Δ t ] due to attrition from all defenders. Therefore, the survival probability Q i ( t + Δ t ) of attacker i is governed by
Q i ( t + Δ t ) = Q i ( t ) k N D ( 1 d i k att P k d ( t ) Δ t ) ,
where we assumed that probabilities of attacker i survival Q i ( t 1 ) , Q i ( t 2 ) are independent for any t 1 , t 2 . Similarly, the survival probability P k d ( t ) of defender k and the survival probability P ( t ) of the HVU are governed by
P k d ( t + Δ t ) = P k d ( t ) i N A ( 1 d k i def Q i ( t ) Δ t ) , P ( t + Δ t ) = P ( t ) k N A ( 1 d k hvu Q k ( t ) Δ t ) .
Initial conditions are set to Q i ( 0 ) = P ( 0 ) = P k d ( 0 ) = 1 for all agents and the HVU.
Further rearranging Equations (43) and (44) and letting Δ t 0 , as derived in [56] we obtain:
Q ˙ i ( t ) = Q i ( t ) k N D ( 1 d i k att P k d ( t ) ) , P ˙ k d ( t ) = P k d ( t ) i N A ( 1 d k i def Q i ( t ) ) ,
P ˙ ( t ) = P ( t ) k N A ( 1 d k hvu Q k ( t ) ) .
The optimal control problem at hand can be expressed as Problem P by properly rescaling the time variable, i.e., τ = t / t f ; see Equations (4)–(7). In particular, we seek to maximize the probability of HVU survival at the terminal time t = t f ( τ = 1 ), i.e., minimize I = 1 P ( t f ) . The system’s state x ( t ) includes the attacker and the defender positions and velocities, as well as probabilities of the attacker and defender survivals and the probability of the HVU survival:
x = x 1 T , , x N A T , ( v 1 x ) T , , ( v N A x ) T , s 1 T , s N D T , ( v 1 s ) T , , ( v N D s ) T , Q 1 , , Q N A , P 1 d , , P N D d , P T ,
where v i x is the velocity of the i-th attacker and v k s is the velocity of the k-th defender. The control input vector is defined by stacking accelerations of each defender u = u 1 T , , u N D T T . Using definitions of the system’s state and control inputs the system dynamics function f ( . , . ) is given by concatenation of Equations (40), (41), (45), and (46). Finally, the function h ( . , . ) in our case becomes a function of defender control inputs only and includes ± u max .
Figure 7 shows results for an optimization with one defender protecting an HVU against a swarm of five attackers for N = 45 . The HVU is at the origin and the defender trajectory is color purple. The defender has a 50% larger weapons range ( σ d / σ a = 1.5 ), as well as double the fire rate with respect to the attackers ( λ d / λ a = 2 ). The defender initially herds the attackers on his right away from HVU than approaches the HVU and similarly herds the attackers to his left away from the HVU. Figure 8 illustrates the control inputs that drive the motion of the defender. Unlike minimum time problems, e.g., the previous example, in this case, the inequality constraints on the control input are never active. Figure 9 shows a sequence of the Hamiltonian approximations H N , N = 5 , , 45 . The sequence clearly converges to zero, indicating that the final numerical solution for N = 45 is indeed a close approximation of the true optimal solution.
The reader is referred to [57,58] for additional numerical examples.

8. Conclusions

This paper proposed a numerical method for costate estimation of nonlinear constrained optimal control problems using Bernstein polynomials. A rigorous analysis is provided that shows convergence of the costate estimates to the dual variables of the continuous-time problem. To this end, a set of conditions are derived under which the Karush–Kuhn–Tucker multipliers of the NLP converge to the costates of the optimal control problem. This led to the formulation of the Covector Mapping Theorem for Bernstein approximation. The theoretical findings are validated through several numerical examples.

Author Contributions

Conceptualization, V.C., I.K., A.P., C.W.; methodology, V.C., I.K., A.P., C.W.; software, V.C, I.K.; validation, V.C, I.K., C.W.; investigation, V.C., I.K., A.P., C.W.; resources, V.C., I.K., A.P., N.H.; writing—original draft preparation, V.C., I.K.; writing—review and editing, V.C, I.K.; supervision, V.C., I.K., A.P., C.W., N.H.; project administration, V.C.; funding acquisition, V.C., I.K., A.P., N.H. All authors have read and agreed to the published version of the manuscript.

Funding

Venanzio Cichella was supported by Amazon, by the Office of Naval Research (grants N000141912106 and N000142112091), and by the National Science Foundation (grant 2136298). Antonio Pascoal was supported by H2020-EU.1.2.2—FET Proactive RAMONES (grant GA 101017808) and LARSyS-FCT (grant UIDB/50009/2020). Isaac Kaminer was supported by the Office of Naval Research (grants N0001421WX01974, N0001419WX00155, N0001422WX01906) and NPS CRUSER. Naira Hovakimyan was supported by NASA ULI (grant #80NSSC22M0070) and NSF RI (grant 2133656).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A. Proof of Equation (34)

Let us focus on Equation (34a). Adding and subtracting 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t , we have
w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t + 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t + 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t .
Using Lemma 3 and continuity of F x ( x N ( t ) , u N ( t ) ) and b 0 , N ( t ) , the first term on the right-hand side of the inequality above satisfies
w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t C I W F x b 0 , N ( N 1 2 ) ,
where W F x b 0 , N ( · ) is used to denote the modulus of continuity of the product
F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) ) ,
with F x ( x N ( t ) , u N ( t ) ) being a bounded function due to its continuity over a bounded domain. Denote its bound as F x , max . Notice that b 0 , N ( t ) is bounded, as max t [ 0 , 1 ] b 0 , N ( t ) 1 . Then, using the properties of the modulus of continuity, we get
w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t C I F x , max W b 0 , N ( N 1 2 ) + C I W F x ( N 1 2 ) C I F x , max N 1 2 + C I W F x ( N 1 2 ) ,
where W F x ( · ) is the modulus of continuity of F x , and C I is a positive constant independent of N. Furthermore, we have
0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) d t 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t 0 1 F x ( x N ( t ) , u N ( t ) ) b 0 , N ( t ) F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t L F x ( C x W x ( N 1 2 ) + C u W u ( N 1 2 ) ) ,
where L F x is the Lipschitz constant of F x , C x < 5 n x / 4 , C u < 5 n u / 4 , and W x ( · ) and W u ( · ) are the moduli of continuity of x and u , respectively. Combining Equations (A2) and (A3) with Equation (A1) yields
w j = 0 N F x ( x N ( t j ) , u N ( t j ) ) b 0 , N ( t j ) 0 1 F x ( x ( t ) , u ( t ) ) b 0 , N ( t ) d t C I F x , max N 1 2 + C I W F x ( N 1 2 ) + L F x ( C x W x ( N 1 2 ) + C u W u ( N 1 2 ) ) ,
which proves the bound in Equation (34a). The bounds in Equation (34b–d) follow easily using an identical argument.

References

  1. Ng, J.; Bräunl, T. Performance comparison of bug navigation algorithms. J. Intell. Robot. Syst. 2007, 50, 73–84. [Google Scholar] [CrossRef]
  2. Khatib, O. Real-time obstacle avoidance for manipulators and mobile robots. In Autonomous Robot Vehicles; Springer: Berlin/ Heidelberg, Germany, 1986; pp. 396–404. [Google Scholar]
  3. Siegwart, R.; Nourbakhsh, I.R.; Scaramuzza, D. Introduction to Autonomous Mobile Robots; MIT Press: Cambridge, MA, USA, 2011. [Google Scholar]
  4. Choset, H.M. Principles of Robot Motion: Theory, Algorithms, and Implementation; MIT Press: Cambridge, MA, USA, 2005. [Google Scholar]
  5. Latombe, J.C. Robot Motion Planning; Springer Science & Business Media: Berlin/Heidelberg, Germany, 2012; Volume 124. [Google Scholar]
  6. Betts, J.T. Survey of numerical methods for trajectory optimization. J. Guid. Control Dyn. 1998, 21, 193–207. [Google Scholar] [CrossRef] [Green Version]
  7. Von Stryk, O.; Bulirsch, R. Direct and indirect methods for trajectory optimization. Ann. Oper. Res. 1992, 37, 357–373. [Google Scholar] [CrossRef]
  8. LaValle, S.M. Planning Algorithms; Cambridge University Press: Cambridge, UK, 2006. [Google Scholar]
  9. Cichella, V. Cooperative Autonomous Systems: Motion Planning and Coordinated Tracking Control for Multi-Vehicle Missions. Ph.D. Thesis, University of Illinois at Urbana-Champaign, Champaign, IL, USA, 2018. [Google Scholar]
  10. Cichella, V.; Kaminer, I.; Dobrokhodov, V.; Xargay, E.; Choe, R.; Hovakimyan, N.; Aguiar, A.P.; Pascoal, A.M. Cooperative Path-Following of Multiple Multirotors over Time-Varying Networks. IEEE Trans. Autom. Sci. Eng. 2015, 12, 945–957. [Google Scholar] [CrossRef]
  11. Sun, X.; Cassandras, C.G. Optimal dynamic formation control of multi-agent systems in constrained environments. Automatica 2016, 73, 169–179. [Google Scholar] [CrossRef] [Green Version]
  12. Walton, C.; Kaminer, I.; Gong, Q.; Clark, A.H.; Tsatsanifos, T. Defense against adversarial swarms with parameter uncertainty. Sensors 2022, 22, 4773. [Google Scholar] [CrossRef]
  13. Rao, A.V. A survey of numerical methods for optimal control. Adv. Astronaut. Sci. 2009, 135, 497–528. [Google Scholar]
  14. Betts, J.T. Practical Methods for Optimal Control and Estimation Using Nonlinear Programming; SIAM: Philadelphia, PA, USA, 2010. [Google Scholar]
  15. Conway, B.A. A survey of methods available for the numerical optimization of continuous dynamic systems. J. Optim. Theory Appl. 2012, 152, 271–306. [Google Scholar] [CrossRef]
  16. Becerra, V.M. Solving complex optimal control problems at no cost with PSOPT. In Proceedings of the 2010 IEEE International Symposium on Computer-Aided Control System Design, Yokohama, Japan, 8–10 September 2010; pp. 1391–1396. [Google Scholar]
  17. Febbo, H.; Jayakumar, P.; Stein, J.L.; Ersal, T. NLOptControl: A modeling language for solving optimal control problems. arXiv 2020, arXiv:2003.00142. [Google Scholar]
  18. Patterson, M.A.; Rao, A.V. GPOPS-II: A MATLAB software for solving multiple-phase optimal control problems using hp-adaptive Gaussian quadrature collocation methods and sparse nonlinear programming. ACM Trans. Math. Softw. (TOMS) 2014, 41, 1–37. [Google Scholar] [CrossRef] [Green Version]
  19. Rutquist, P.E.; Edvall, M.M. Propt—Matlab Optimal Control Software; Tomlab Optimization Inc.: Washington, DC, USA, 2010. [Google Scholar]
  20. Ross, I.M. Enhancements to the DIDO Optimal Control Toolbox. arXiv 2020, arXiv:2004.13112. [Google Scholar]
  21. Andersson, J.A.; Gillis, J.; Horn, G.; Rawlings, J.B.; Diehl, M. CasADi: A software framework for nonlinear optimization and optimal control. Math. Program. Comput. 2019, 11, 1–36. [Google Scholar] [CrossRef]
  22. Fahroo, F.; Ross, I.M. On discrete-time optimality conditions for pseudospectral methods. In Proceedings of the AIAA/AAS Astrodynamics Specialist Conference and Exhibit, Keystone, CO, USA, 21–24 August 2006; p. 6304. [Google Scholar]
  23. Bollino, K.; Lewis, L.R.; Sekhavat, P.; Ross, I.M. Pseudospectral optimal control: A clear road for autonomous intelligent path planning. In Proceedings of the AIAA Infotech@ Aerospace 2007 Conference and Exhibit, Rohnert Park, CA, USA, 7–10 May 2007; p. 2831. [Google Scholar]
  24. Gong, Q.; Lewis, R.; Ross, M. Pseudospectral motion planning for autonomous vehicles. J. Guid. Control. Dyn. 2009, 32, 1039–1045. [Google Scholar] [CrossRef]
  25. Bedrossian, N.S.; Bhatt, S.; Kang, W.; Ross, I.M. Zero-propellant maneuver guidance. IEEE Control. Syst. 2009, 29. [Google Scholar]
  26. Bollino, K.; Lewis, L.R. Collision-free multi-UAV optimal path planning and cooperative control for tactical applications. In Proceedings of the AIAA Guidance, Navigation and Control Conference and Exhibit, Honolulu, HI, USA, 18–21 August 2008; p. 7134. [Google Scholar]
  27. Bedrossian, N.; Bhatt, S.; Lammers, M.; Nguyen, L.; Zhang, Y. First Ever Flight Demonstration of Zero Propellant Maneuver (TM) Attitute Control Concept. In Proceedings of the AIAA Guidance, Navigation and Control Conference and Exhibit, Hilton Head, SC, USA, 20–23 August 2007; p. 76734. [Google Scholar]
  28. Ross, I.M.; Karpenko, M. A review of pseudospectral optimal control: From theory to flight. Annu. Rev. Control. 2012, 36, 182–197. [Google Scholar] [CrossRef]
  29. Polak, E. Optimization: Algorithms and Consistent Approximations; Springer: Berilin, Germany, 1997. [Google Scholar]
  30. Fahroo, F.; Ross, I.M. Costate estimation by a Legendre pseudospectral method. J. Guid. Control Dyn. 2001, 24, 270–277. [Google Scholar] [CrossRef]
  31. Darby, C.L.; Garg, D.; Rao, A.V. Costate estimation using multiple-interval pseudospectral methods. J. Spacecr. Rocket. 2011, 48, 856–866. [Google Scholar] [CrossRef]
  32. Hager, W.W. Runge-Kutta methods in optimal control and the transformed adjoint system. Numer. Math. 2000, 87, 247–282. [Google Scholar] [CrossRef]
  33. Grimm, W.; Markl, A. Adjoint estimation from a direct multiple shooting method. J. Optim. Theory Appl. 1997, 92, 263–283. [Google Scholar] [CrossRef]
  34. Cichella, V.; Kaminer, I.; Walton, C.; Hovakimyan, N. Optimal Motion Planning for Differentially Flat Systems Using Bernstein Approximation. IEEE Control Syst. Lett. 2018, 2, 181–186. [Google Scholar] [CrossRef]
  35. Kielas-Jensen, C.; Cichella, V. BeBOT: Bernstein polynomial toolkit for trajectory generation. In Proceedings of the 2019 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), Venetian Macao, Macau, 3–8 November 2019; pp. 3288–3293. [Google Scholar]
  36. Kielas-Jensen, C.; Cichella, V.; Berry, T.; Kaminer, I.; Walton, C.; Pascoal, A. Bernstein Polynomial-Based Method for Solving Optimal Trajectory Generation Problems. Sensors 2022, 22, 1869. [Google Scholar] [CrossRef] [PubMed]
  37. Ricciardi, L.A.; Vasile, M. Direct transcription of optimal control problems with finite elements on Bernstein basis. J. Guid. Control Dyn. 2018. [Google Scholar] [CrossRef]
  38. Choe, R.; Puig-Navarro, J.; Cichella, V.; Xargay, E.; Hovakimyan, N. Cooperative Trajectory Generation Using Pythagorean Hodograph Bézier Curves. J. Guid. Control Dyn. 2016, 39, 1744–1763. [Google Scholar] [CrossRef]
  39. Ghomanjani, F.; Farahi, M.H. Optimal control of switched systems based on Bezier control points. Int. J. Intell. Syst. Appl. 2012, 4, 16. [Google Scholar] [CrossRef]
  40. Huo, M.; Yang, L.; Peng, N.; Zhao, C.; Feng, W.; Yu, Z.; Qi, N. Fast costate estimation for indirect trajectory optimization using Bezier-curve-based shaping approach. Aerosp. Sci. Technol. 2022, 126, 107582. [Google Scholar] [CrossRef]
  41. Zhao, Z.; Kumar, M. Split-bernstein approach to chance-constrained optimal control. J. Guid. Control Dyn. 2017, 40, 2782–2795. [Google Scholar] [CrossRef]
  42. Cichella, V.; Kaminer, I.; Walton, C.; Hovakimyan, N.; Pascoal, A.M. Optimal Multi-Vehicle Motion Planning using Bernstein Approximants. IEEE Trans. Autom. Control 2020. [Google Scholar] [CrossRef]
  43. Schwartz, A.; Polak, E. Consistent approximations for optimal control problems based on Runge–Kutta integration. SIAM J. Control Optim. 1996, 34, 1235–1269. [Google Scholar] [CrossRef]
  44. Bojanic, R.; Cheng, F. Rate of convergence of Bernstein polynomials for functions with derivatives of bounded variation. J. Math. Anal. Appl. 1989, 141, 136–151. [Google Scholar] [CrossRef]
  45. Popoviciu, T. Sur l’approximation des fonctions convexes d’ordre supérieur. Mathematica 1935, 10, 49–54. [Google Scholar]
  46. Sikkema, P. Der wert einiger konstanten in der theorie der approximation mit Bernstein-Polynomen. Numer. Math. 1961, 3, 107–116. [Google Scholar] [CrossRef]
  47. Powell, M.J.D. Approximation Theory and Methods; Cambridge University Press: Cambridge, UK, 1981. [Google Scholar]
  48. Floater, M.S. On the convergence of derivatives of Bernstein approximation. J. Approx. Theory 2005, 134, 130–135. [Google Scholar] [CrossRef]
  49. Hartl, R.F.; Sethi, S.P.; Vickson, R.G. A survey of the maximum principles for optimal control problems with state constraints. SIAM Rev. 1995, 37, 181–218. [Google Scholar] [CrossRef]
  50. Garg, D.; Patterson, M.A.; Francolin, C.; Darby, C.L.; Huntington, G.T.; Hager, W.W.; Rao, A.V. Direct trajectory optimization and costate estimation of finite-horizon and infinite-horizon optimal control problems using a Radau pseudospectral method. Comput. Optim. Appl. 2011, 49, 335–358. [Google Scholar] [CrossRef] [Green Version]
  51. Gong, Q.; Ross, I.M.; Kang, W.; Fahroo, F. Connections between the covector mapping theorem and convergence of pseudospectral methods for optimal control. Comput. Optim. Appl. 2008, 41, 307–335. [Google Scholar] [CrossRef]
  52. Singh, B.; Bhattacharya, R.; Vadali, S.R. Verification of optimality and costate estimation using Hilbert space projection. J. Guid. Control Dyn. 2009, 32, 1345–1355. [Google Scholar] [CrossRef]
  53. Kirk. Optimal Control Theory: An Introduction; Prentice-Hall: Hoboken, NJ, USA, 1970. [Google Scholar]
  54. Ogren, P.; Fiorelli, E.; Leonard, N.E. Cooperative control of mobile sensor networks: Adaptive gradient climbing in a distributed environment. IEEE Trans. Autom. Control 2004, 49, 1292–1302. [Google Scholar] [CrossRef] [Green Version]
  55. Washburn, A.; Kress, M. Combat Modeling; Springer: Berlin/Heidelberg, Germany, 2009. [Google Scholar]
  56. Walton, C.; Lambrianides, P.; Kaminer, I.; Royset, J.; Gong, Q. Optimal motion planning in rapid-fire combat situations with attacker uncertainty. Naval Res. Logist. 2018, 65, 101–119. [Google Scholar] [CrossRef]
  57. Cichella, V.; Kaminer, I.; Walton, C.; Hovakimyan, N.; Pascoal, A. Bernstein approximation of optimal control problems. arXiv 2018, arXiv:1812.06132. [Google Scholar]
  58. Cichella, V.; Kaminer, I.; Walton, C.; Hovakimyan, N.; Pascoal, A.M. Consistent approximation of optimal control problems using Bernstein polynomials. In Proceedings of the 2019 IEEE 58th Conference on Decision and Control (CDC), Nice, France, 11–13 December 2019; pp. 4292–4297. [Google Scholar]
Figure 1. Diagram of the covector mapping principle for Bernstein approximation. The solution to Problem P N λ clos converges to that of Problem P λ as N .
Figure 1. Diagram of the covector mapping principle for Bernstein approximation. The solution to Problem P N λ clos converges to that of Problem P λ as N .
Machines 10 01132 g001
Figure 2. Example 1: state trajectories x 1 ( t ) and x 2 ( t ) , and control input u ( t ) . The control input approximates the optimal solution (bang-bang).
Figure 2. Example 1: state trajectories x 1 ( t ) and x 2 ( t ) , and control input u ( t ) . The control input approximates the optimal solution (bang-bang).
Machines 10 01132 g002
Figure 3. Example 1: the approximated Hamiltonian converges to the Hamiltonian of the problem, i.e., 1 . See Theorem 3 and Remark 5.
Figure 3. Example 1: the approximated Hamiltonian converges to the Hamiltonian of the problem, i.e., 1 . See Theorem 3 and Remark 5.
Machines 10 01132 g003
Figure 4. Example 2: 3D position plot, i.e., p = [ x 1 , x 2 , x 3 ] . The solid line represents the solution obtained with N = 45 , while the dashed line depicts the (near optimal) solution obtained with N = 250 .
Figure 4. Example 2: 3D position plot, i.e., p = [ x 1 , x 2 , x 3 ] . The solid line represents the solution obtained with N = 45 , while the dashed line depicts the (near optimal) solution obtained with N = 250 .
Machines 10 01132 g004
Figure 5. Example 2: Control inputs, i.e., vehicle’s acceleration along the three axis. The solid lines represent the solution obtained with N = 45 , while the dashed lines depict the (near optimal) solution obtained with N = 250 .
Figure 5. Example 2: Control inputs, i.e., vehicle’s acceleration along the three axis. The solid lines represent the solution obtained with N = 45 , while the dashed lines depict the (near optimal) solution obtained with N = 250 .
Machines 10 01132 g005
Figure 6. Example 2: the approximated Hamiltonian converges to the Hamiltonian of the problem, i.e., 1 . See Theorem 3 and Remark 5.
Figure 6. Example 2: the approximated Hamiltonian converges to the Hamiltonian of the problem, i.e., 1 . See Theorem 3 and Remark 5.
Machines 10 01132 g006
Figure 7. Defense against swarm attack. The plot shows optimal trajectory of one defender (purple) protecting a high-value unit (positioned at the origin) against five attackers.
Figure 7. Defense against swarm attack. The plot shows optimal trajectory of one defender (purple) protecting a high-value unit (positioned at the origin) against five attackers.
Machines 10 01132 g007
Figure 8. Defense against swarm attack. The plot shows the time history of the control input.
Figure 8. Defense against swarm attack. The plot shows the time history of the control input.
Machines 10 01132 g008
Figure 9. Hamiltonian Convergence. The plot shows a sequence of the Hamiltonian approximations H N , N = 5 , , 45 , indicating that the numerical solution converges to the true optimal solution.
Figure 9. Hamiltonian Convergence. The plot shows a sequence of the Hamiltonian approximations H N , N = 5 , , 45 , indicating that the numerical solution converges to the true optimal solution.
Machines 10 01132 g009
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Cichella, V.; Kaminer, I.; Walton, C.; Hovakimyan, N.; Pascoal, A. Consistency of Approximation of Bernstein Polynomial-Based Direct Methods for Optimal Control. Machines 2022, 10, 1132. https://doi.org/10.3390/machines10121132

AMA Style

Cichella V, Kaminer I, Walton C, Hovakimyan N, Pascoal A. Consistency of Approximation of Bernstein Polynomial-Based Direct Methods for Optimal Control. Machines. 2022; 10(12):1132. https://doi.org/10.3390/machines10121132

Chicago/Turabian Style

Cichella, Venanzio, Isaac Kaminer, Claire Walton, Naira Hovakimyan, and António Pascoal. 2022. "Consistency of Approximation of Bernstein Polynomial-Based Direct Methods for Optimal Control" Machines 10, no. 12: 1132. https://doi.org/10.3390/machines10121132

APA Style

Cichella, V., Kaminer, I., Walton, C., Hovakimyan, N., & Pascoal, A. (2022). Consistency of Approximation of Bernstein Polynomial-Based Direct Methods for Optimal Control. Machines, 10(12), 1132. https://doi.org/10.3390/machines10121132

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop