1. Introduction
In order to plan trajectories in aerospace missions, optimal control methods can deliver highly detailed trajectories that can serve to analyze system performance or as reference trajectories for tracking purposes. For increased safety, environmental or system-related uncertainties must be considered in trajectory planning. Robust optimal control methods take these uncertainties into account and provide robust trajectories that are less prone to disturbances, i.e., under disturbances in model parameters, the trajectory deviations are reduced. Two key techniques in application-oriented studies are the stochastic and deterministic approaches.
In statistical approaches, it is assumed that the uncertainty underlies a statistical distribution. Hence, probability or chance constraints can be imposed, requiring a particular event to happen with at least a given probability. Studies on this can be found in [
1], where robust transition trajectories of an electric Vertical Take-off and Landing (eVTOL) aircraft are calculated considering uncertainties in wind estimations. There, a stochastic collocation approach combined with general Polynomial Chaos (gPC) is used. In [
2], the authors propose to evaluate chance constraints via kernel density estimation, similarly as in [
3] coupled with Markov Chain Monte Carlo sampling. The main challenge in these methods is the high computational effort in calculating probabilities and their gradients.
In deterministic approaches, uncertainties are considered to appear deterministically in the dynamic system. One approach here is semi-infinite optimization [
4], wherein constraints are imposed on all realizations of an uncertain parameter, making the solution feasible for a set of disturbances and optimal for the ideal parameter value. However, these problems require high computational effort to be solved due to the complex constraints, including worst-case considerations.
Another approach is sensitivity minimization [
5,
6], wherein sensitivities are minimized within the cost function, reducing the dependency of a state of interest concerning an uncertain model parameter. Thereby, the system states’ or cost variations under parameter deviations are reduced, i.e., the robustness of the resulting trajectories is increased. In the mentioned studies, sensitivity minimization is applied to a Rayleigh system [
5] and to robustify reentry trajectories regarding the heat flux under uncertain air density [
6].
It is important to note that the mentioned studies are limited to open-loop optimal control. However, in many applications, it is essential to consider a closed-loop system in the optimization to describe a realistic behavior and exploit the full performance.
In gradient-based optimal control methods, incorporating a closed-loop system model is usually not suitable for the differentiability requirements of the optimization methods. Three main approaches to tackle this problem are discussed in [
7]. In summary, one approach may be to smooth a given controller for use in optimization. However, the controller behavior may not be perfectly reproduced after modification, as discussed in [
8], and the resulting trajectories are then limited to this one control strategy. Another approach is presented in [
9,
10], wherein an explicit controller is designed in the optimization process considering Lyapunov and eigenvalue stability constraints. The advantage here is that specific requirements for the controller can be set during optimization. Nevertheless, the procedure to obtain the optimal trajectory and control design highly depends on the considered system.
Another challenge is the consideration of stability of linear time-variant (LTV) systems. There are stability criteria, e.g., the fulfillment of the Riccati equation [
11,
12]. Nevertheless, the incorporation into an optimization problem is not straightforward since the Riccati equation needs to be solved.
An optimization-based state feedback approach is Model Predictive Control (MPC) [
13]. Here, cyclic re-optimization of the controls starting from the current state yields a feedback control law. By this, deviations from the optimal trajectory can be taken into account and can ideally be corrected. However, resolving the optimization problems can be computationally expensive. Instead of recalculating the optimal solution, post-optimal sensitivities can be used to linearly approximate the state and control history for solution updates as in [
14]. Here, the post-optimal sensitivities describe the change in the optimal solution regarding parameter value changes.
An approach combining sensitivity minimization with feedback is presented in [
15,
16]. Here, the open-loop sensitivity minimization approach is extended by control sensitivity terms regarding state feedback, such that the controls can be updated linearly according to state feedback. The control sensitivities are subject to optimization. The effectiveness of the method is presented in an academic example. The authors propose only the state feedback incorporation in case a feedforward cannot be determined, as parameter perturbations may not be measurable.
The approach of a linear control update with optimized gains is advantageous since, in the early development stages of new systems, performance assessment via trajectory optimization is possible without the actual usage of a fully developed controller. Furthermore, it provides a holistic optimization of trajectories and controller gains, which fully exploits system performance.
In the paper at hand, the sensitivity minimization approach is combined with state feedback and predicted feedback under the assumption that parameter values can be estimated during flight. In detail, generic control update terms considering state feedback and predicted feedback are introduced into the trajectory optimization problem appearing in the sensitivity differential equations. Thereby, the update terms improve sensitivity reduction and trajectory robustness compared to open-loop considerations. The proposed method is applied to a practically relevant UAV trajectory optimization problem of high dimension and is validated by closed-loop simulations.
This study is structured as follows. In
Section 2, the theoretical background of robust open-loop optimal control using sensitivity minimization is presented. Afterwards, the closed-loop sensitivity minimization approach incorporating control updates is introduced in
Section 3. An application example for UAV robust trajectory optimization is presented in
Section 4. This study is concluded in
Section 5.
2. Robust Open-Loop Optimal Control with Sensitivity Minimization
In this subsection, an open-loop optimal control problem formulation with sensitivity penalty is presented according to [
5,
6,
17]. The considered time horizon of the trajectories to be optimized is denoted by
. Let
be the state history and
the control history. The parameter vector is denoted by
. Then, the robust open-loop optimal control problem statement is as follows:
subject to the dynamic constraints
with
, the sensitivity differential equation for
with
according to [
17]
the initial and final boundary conditions
with
and inequality constraints
for all
with
Here,
J represents a Mayer cost function, which depends only on initial and final states and the free final time. The cost function here is formulated with a sensitivity penalty with weights
, which can be chosen to find a trade-off between optimality and robustness. The higher the sensitivity weights, the higher the robustness. Please note that increased robustness usually leads to increased nominal costs
J compared to the minimal cost value resulting from only minimizing
J. Solving the robust optimal control problem delivers an open-loop control history
, which can be utilized to obtain a state history
, which is less prone to deviations in the parameters
as studied in [
5,
6]. It is worth noting that the open-loop sensitivity does not include any terms describing the controls’ dependencies on parameters.
In the following section, the open-loop approach is extended to closed-loop modeling to increase robustness.
4. Application to UAV Trajectory Optimization
The proposed approach from
Section 2 is applied to a UAV climb trajectory optimization problem wherein flight time and sensitivities are minimized. The dynamic model of a fixed-wing UAV is based on [
18] and given by
where
describe the positional coordinates,
the course angle,
the flight path angle and
V the velocity. Furthermore,
m is the mass of the vehicle,
g the gravitational acceleration and
the bank angle. The thrust, lift and drag are given by
with
being the thrust lever position,
the maximum thrust of the vehicle,
the air density,
S the reference area,
the zero-lift drag coefficient,
k the induced drag factor and
the lift coefficient. The values of the model parameters are given in
Table 1.
The state and the controls are defined to be
4.1. Case 0: Nominal Problem Formulation
The maneuver is a climb in minimum time after transition to start a particular mission at a given altitude:
such that the initial and final boundaries in
Table 2 are fulfilled.
Furthermore, flight time is limited to 10 s to allocate enough flight time for the mission. This optimal control problem is solved with FALCON.m [
19], which is an optimal control framework in MATLAB. It is based on direct collocation methods [
17] using Trapezoidal integration.
Since aerodynamic parameters, in this case the drag coefficient , are usually estimated based on wind tunnel and CFD data, they may be subject to uncertainty. It is assumed that the drag coefficient can deviate by and that the most probable values of the realizations of are in the interval .
If the actual parameter values differ from the ideal values assumed in the model, the actual trajectory using the optimal controls may significantly change. For this reason, the robust optimal control methods presented in
Section 2 are applied. Robustness of along-track position may be essential for specific scenarios, e.g., to ensure separation between vehicles and reduce the collision risk. Therefore, the aim is to minimize the deviations in the positional state
x with respect to the parameter
by minimizing the sensitivity
For validation, simulations with worst-case deviations in are conducted. In detail, the following methods are applied and compared.
4.2. Case 1: Problem Formulation for Open-Loop Sensitivity Minimization
In order to robustify the trajectory in
Section 4.1, open-loop sensitivity minimization as presented in
Section 2 is conducted as a benchmark for the proposed methods. Therefore, the cost function is modeled as
The constraints regarding the states and controls are equal to those of case 0 in
Section 4.1.
4.3. Case 2: Problem Formulation for Closed-Loop Sensitivity Minimization with Feedback
To improve robustness, the method using state feedback as in
Section 3.1 is applied. The cost function is given by (
29) and the same as in the open-loop sensitivity minimization case 1. New controls are introduced, namely the time-dependent gains
. They are limited by the lower and upper bounds
in order to map the limitations of the controller. The post-optimal simulations are conducted with the updated control (
8). An estimation of feasible gain limits can be made by analyzing the magnitude of the estimated deviation
.
4.4. Case 3: Problem Formulation for Closed-Loop Sensitivity Minimization with Predicted Feedback
Under the assumption that the drag coefficient can be estimated during flight by energy estimations, the predicted feedback formulation is applied. Therefore, similarly to the feedback case in
Section 4.3, the gains
are introduced with the same bounds as in (
30). In this case, the control update in (
13) is utilized in the post-optimal simulations. Furthermore, control update constraints as in (
14) and (
15) are imposed with the same values as in
Table 2. Analogously, the same rate constraints as for the nominal controls are applied to the worst-case control updates.
An overview of the compared cases is given in
Table 3.
4.5. Numerical Results
The optimization using the open-loop and closed-loop sensitivity minimization leads to the optimal state and control trajectories depicted in
Figure 1 and
Figure 2, respectively. In the open-loop sensitivity minimization case 1, the optimization can find optimal controls that reduce the sensitivity of the
x-position and the other system states due to system interdependencies (see
Figure 3a). The sensitivity reduction can also be observed in the simulated trajectories under variations in
(see
Figure 1). In the nominal case 0, the variations in the
x-position are up to 3 m, whereas in the open-loop sensitivity minimal case 1, the values vary up to 1 m. The direct influence of
on
V, which directly influences the other system states, suggests that the optimization found a solution reducing the sensitivity of
V. Indeed, the solution with a sensitivity penalty on
shows the same maneuver structure. This suggests that a reduced velocity is more robust. To meet the final boundary conditions in
Table 2, the velocity increases by the end of the maneuver.
Compared to the nominal case 0, where the flight time is 9.05 s, flight time increases to the upper time limit of 10 s when minimizing sensitivities in the open-loop case 1 due to the trade-off between optimality and robustness. The significant increase in the final time can mathematically be explained by the sensitivity having a higher order of magnitude than the final time. For this reason, the sensitivity penalty considerably influences the overall cost (
29) and leads to an optimal solution, which may allow for an increased final time compared to the nominal case. Please note that, for increased time optimality, the weights of the sensitivities in (
29) could be decreased. For comparability to the subsequent closed-loop cases described in
Table 3, the weights are fixed for all cases.
In the closed-loop optimal solutions depicted in
Figure 2, the sensitivities can further be reduced by a magnitude of around 10 to 1000. The sensitivity reduction can also be observed in the simulations since the deviations between the simulated trajectories at perturbed values of
are decreased. In the feedback case 2 and the predicted feedback case 3, the deviations in the
x-position are reduced to be below 1 cm. Taking into account that a reasonable estimator can reduce the uncertainty in the drag estimation during flight to
, simulations employing disturbed parameter values for the control updates have shown to reduce the
x-position deviations up to 15 cm, hence still showing good performance in terms of robustness. Therefore, the simulations using an ideal parameter estimation are depicted for comparability with the benchmark method. Notably, the closed-loop robust optimal trajectories are similar to the nominal open-loop optimal trajectory without sensitivity minimization. Due to the nature of a closed-loop system, the overall magnitude of sensitivities is smaller, as depicted in
Figure 3b, having a negligible impact on the overall cost. Hence, a solution with a nearly minimum final time can be found. Specifically, the flight time for case 2 is 9.05 s and for case 3 is 9.07 s. This indicates that the
closed-loop solutions allow for increased optimality while being robust. Furthermore, both closed-loop solutions in
Figure 3b show high concordance. This indicates that the predicted feedback leads to the desired validity of the prediction and underscores the effectiveness of the proposed predicted feedback method. The effect of the control update constraints according to (
14) and (
15) and
Table 2 can be observed in the differences in the gains. Due to the fact that the updated controls in case 2 are already close to fulfilling these constraints, there is a slight difference to the updated controls in case 3. Although the gains vary rapidly over time, the constraints ensure that the updated controls are feasible.