1. Introduction
With the development of space science and technology, many different types of spacecraft have been launched into the sky. Spacecraft are widely used in communication, navigation, and positioning; resource exploration; and space science exploration, etc. [
1]. In recent years, a distributed array of small, low-cost, cooperative, and highly coordinated spacecraft has vigorously been pursued. Compared with large individual spacecraft, for these distributed spacecraft flying in formation, information collected by each spacecraft can be shared with the others in the formation via inter-spacecraft communication to allow for a more flexible onboard data-management scheme [
2]. With the ability to transfer information and measurements between member spacecraft, a more effective utilization of available computing resources can be achieved by techniques such as distributed computing and system load balancing.
There are various methods to realize spacecraft formation flying, such as the leader–follower strategy [
3], virtual structure approaches [
4], and behavior-based methods [
5]. Among different control techniques, the leader–follower strategy is the most commonly used since it has a hierarchical structure; i.e., in the leader–follower mode, each follower spacecraft is controlled in order to maintain its position and velocity in relation to a designated leader spacecraft, which flies along a predefined trajectory. Many studies related to spacecraft attitude control are based on this control strategy [
6]. For example, Wang and Hadagegh [
7] first applied this control strategy to control satellite attitude. Since then, many experts have studied satellite formations in the leader–follower mode. In [
8], the authors applied a feedback control algorithm to solve the control law for satellite attitude control. In [
9,
10], the application of robust control in satellite group attitude control was investigated. In [
11], a fast finite-time sliding mode control method was proposed for the attitude control problem of magnetically actuated satellite systems. A terminal sliding mode control approach as utilized in [
12] for the global stability and robust attitude tracking problem of a near-polar orbit satellite subject to unknown disturbances and uncertainties. In [
13], an adaptive, fuzzy, proportional–integral–derivative controller was designed for geostationary satellite attitude control. In [
14], a linear time-variant model predictive control was designed to control a satellite with two reaction wheels. In addition, there are some studies that use GPS-based state estimation methods to realize spacecraft formation flight [
15,
16].
The spacecraft rendezvous-control problem of a follower spacecraft in leader–follower control mode has attracted considerable attention. Generally, in the rendezvous phase, the interest is in the relative motion between two spacecraft, where relative motion control includes relative orbit control and relative attitude control. In [
17], angular velocity and attitude estimators were designed to deal with information delay. A relative attitude determination scheme and a relative attitude control system were developed based on line-of-sight observations between two spacecraft in [
18]. In [
19], the attitude-tracking problem was formulated as a nonlinear optimal control problem, after which the state-dependent Riccati equation controller and
-D controller were developed. For the presence of kinematic couplings and model uncertainties, a robust adaptive-relative position and relative attitude controllers were designed in [
20]. In [
21], for a non-cooperative target spacecraft fly-around mission, a modified nonsingular fast terminal sliding mode with bounded convergence time, regardless of the initial state, was designed. In [
22], a relative motion-control algorithm that achieved robustness against external disturbances with less measurement information was proposed. For the relative attitude tracking of rigid spacecraft with disturbance torques on a target and a follower, a computationally tractable method leveraging sum-of-squares optimization was proposed in [
23].
In the above methods, most of them are traditional control methods. They either only consider a single objective (i.e., orbital motion error) or consider both orbital motion error and energy consumption at the same time but often use a weighted sum strategy to transform the two-objective problem into a single-objective optimization problem. These methods can obtain only one control scheme. They did not consider the tradeoff between the orbital motion error and energy consumption and cannot provide decision makers with more choices in the case of different task requirements.
In this paper, we consider the problem of relative orbital motion-control in the leader–follower mode. In reality, in the leader–follower mode, tight control of the relative distances and phases between the follower spacecraft and the leader spacecraft is required. For the follower spacecraft rendezvous-control problem, not only the rendezvous-control time but also the energy consumption needs to be considered. Therefore, in this paper, we consider both the orbital motion error and the energy consumption simultaneously. We formulate the relative orbital motion-control problem of a follower spacecraft as a multi-objective optimization problem. Then, we propose a relative orbital motion-control algorithm based on ε-constraint to solve the problem. The proposed method can provide a set of optimal control schemes.
The rest of the paper is organized as follows: First, we introduce the relative position dynamics of spacecraft in
Section 2. Then, the relative orbital motion-control problem is described in
Section 3. After this, our proposed multi-objective methods based on elastic constraints are introduced in
Section 4. Next, in
Section 5, we show certain numerical examples. Finally, we give the conclusion in
Section 6.
2. Relative Position Dynamics
Consider the linear relative motion of a follower spacecraft orbiting the Earth in a near circular orbit with respect to a leader spacecraft, as shown in
Figure 1.
It is assumed that both the leader and the follower spacecraft can be seen as particles. In
Figure 1,
represents the center of the Earth, and
is the location of the leader spacecraft.
is the Earth-centered-inertial coordinate frame, whereas
is a moving coordinate system. The relative position of the follower spacecraft at the moment
t in the
coordinate system is
. The relative motion can be described by the Hill–Clohessy–Wiltshire (HCW) equation [
24]:
where
denotes the control input of the follower spacecraft, and
denotes the average orbital angular velocity of the leader spacecraft.
is calculated as
, where
represents the relative displacement of target from the center of the Earth, and
denotes the gravitational constant of the Earth.
We define the relative state between the follower spacecraft and the leader spacecraft by
, where
and
denote the position vector and velocity vector, respectively. Therefore, the state–space equation associated with (1) can be expressed as follows:
where
and
are constant coefficient matrices,
,
,
,
,
denotes the system state, and
denotes the control input. (
is a known tight convex set, which bounds the control input.)
3. Problem Description
In the rendezvous phase, the follower spacecraft is required to achieve a specified relative position with respect to the leader spacecraft as soon as possible within a terminal specified time, while at the same time, the energy consumption during the control process needs to be minimized.
Suppose that at the initial time
, the initial relative state between the follower spacecraft and the leader spacecraft is
Then, the state–space equation (Equation (2)) can be written as follows:
Furthermore, the rendezvous-control problem at terminal time
can be formulated as follows:
where
is the target position,
is the control force function,
denotes the distance between the state of the follower spacecraft and the target state, and
denotes the energy consumption. Obviously, the two objectives conflict with each other; thus, problem
is a bi-objective probem.
The goal of the problem is trying to control the follower spacecraft to reach a target state while at the same time attempting to keep the energy consumption as low as possible.
From problem
, we can see that the form of problem
is similar to the optimization problem of the LQR method [
25], which is as follows:
where
,
is target state, and
and
are two diagonal parameter matrices chosen by the decision maker.
However, general control methods (e.g., LQR, etc.) are mostly based on a single-objective optimization problem model and can only provide a single decision solution, while our method is designed for a multi-objective optimization problem, and we can provide multiple decision solutions for the user to choose from. In fact, the cost function of LQR can be considered as a weighted sum of , where and can be considered as the weights of two objectives; thus, the solution to LQR is encompassed by our solutions.
4. Methods
In order to numerically solve problem , we discretize the relative motion model system (4) into a discrete system and reformulate problem into a discretized multi–objective optimization model.
4.1. Discrete System
A continuous control system can be converted into a discrete one by numerical methods such as the zero-order hold method [
26], Euler methods, and Runge–Kutta methods. In this paper, the zero-order hold method of [
26] is used, where the signal values between two constitutive sampling are equal to the last sampling values. The zero-order hold method is widely used in autonomous vehicles [
27], the course-keeping control for ships [
28], and safety verification [
29], among other things.
With the zero-order hold method, System (3) can be converted into the following:
where
,
,
is the sampling period, and
is the total number of discretization steps. Recursively, by substituting
into system (6), we obtain the following:
Note that if ( is a constant), then the corresponding constraint can be written into the nonlinear form .
4.2. Multi-Objective Optimization Model of the Discretized Control Problem
The discretized version of problem
is as follows:
where
denotes the position of the relative motion between the follower spacecraft and the leader spacecraft at step
;
denotes the velocity of the relative motion between the follower spacecraft and the leader spacecraft at step
; and
is the total number of discretized steps. It is obvious that a set of control schemes
can be obtained by solving the multi-objective optimization problem (
). However, the following two difficulties often occur in solving a multi-objective optimization problem.
Firstly, solving a multi-objective optimization problem requires optimizing two or more objectives at the same time, and the objectives are often conflicting with each other. There is no single optimal solution that makes all objectives optimal at the same time. The solution of a multi-objective optimization problem is generally a set. This means that there may be an infinite number of solutions (control schemes), and only a finite number of solutions can be calculated. Therefore, a discrete representation of the whole solution set is often desirable.
Secondly, for a multi-objective optimization problem, each optimization objective has a different dimension; i.e., each objective has a different meaning in the actual problem, so it is meaningless to make a comparison without special treatment.
To solve the above problems, we introduce a multi-objective optimization method with a normalization strategy in the following subsections.
4.3. Multi-Objective Optimization Problem and ε-Constraint Method
Consider a multi-objective optimization problem (MOP) [
30]:
is the non-empty feasible set in decision (or variable) space . is a vector-valued objective function, is composed of real-valued objective functions. , where for . The image of the feasible set under the objective function mapping is denoted as .
Definition 1 (ref. [30]). For problem (MOP), a point is called a weakly (strictly) efficient solution in decision space if there is no such that for all . Correspondingly, point is called a weakly (strictly) nondominated point in objective space.
Definition 2 (ref. [30]). For problem (MOP), a point is called an efficient solution in decision space if there is no such that both for all and for some j. Correspondingly, point is called a nondominated point in objective space.
We denote the sets of weakly (strictly) efficient solutions by () and the sets of efficient solutions by . The images of () under are denoted by (), and the images of under are called the nondominated set (Pareto front), which is denoted by
Here, we introduce the ε-constraint method (
Figure 2). The ε-constraint method was proposed by Chankong and Haimes [
31]. It is based on a scalarization where one of the objective functions is minimized, while all the other objective functions are bounded from above by means of additional constraints.
where
and
.
We denote the feasible set of the ε-constraint problem
by the following:
It needs to be noted that
can be always chosen to make
feasible; i.e.,
. The following two main theorems [
31] concerning the ε-constraint method specify the relationship between problem
and (MOP):
Theorem 1. For any , the following statements hold:
If is an optimal solution of (), then ;
If is a unique optimal solution of (), then .
Theorem 2. A solution is efficient if and only if it is an optimal solution of for every , where for .
Therefore, for a multi-objective optimization problem, the Pareto front can be obtained by choosing a suitable set of and then solving a set of ε-constraint problems.
However, for a multi-objective orbital motion-control problem, constructing the complete Pareto front of its corresponding multi-objective optimization problem is difficult and less meaningful. Instead, a discrete representative set of the Pareto front is ideal to provide to the decision maker.
For the bi-objective optimization problem (), there are two dimensions in its objective space. For each dimension (objective), we have an individual minimal: they can be used to construct a set of equally spaced points. Furthermore, we can construct a set of ε-constraint problems accordingly. Then, solving this set of ε-constraint problems, a representative set of the Pareto front is obtained. The representative set has the property that the distance between two adjacent nondominated points in one dimension is equal.
4.4. Relative Orbital Motion Control Based on ε-Constraint
In this section, we propose a relative orbital motion-control algorithm based on the ε-constraint to solve problem . The ε-constraint method takes place in the normalized design metric space to avoid scaling deficiencies. Before introducing the algorithm, we first introduce the normalization method.
Let
and
, where
. Their corresponding objective space points are denoted as
and
, respectively. We define the Utopia point as
. Let
be the distances between
and the Utopia point
, and let
be the distances between
and
. That is,
With the above definitions, the normalized design metrics can be evaluated as follows:
By normalization, we eliminate the effect of different scales on different objectives. Based on this normalization method, when solving a multi-objective problem, we obtain a set of nondominated points such that each element value will fall into the interval [0,1]. The decision maker can compare different nondominated points more intuitively and thus choose a more appropriate control scheme according to different task situations.
Here, we show the relative orbital motion-control algorithm in Algorithm 1.
Algorithm 1. Relative Orbital Motion-Control Algorithm Based on ε-Constraint |
Input: , , , , , , , ; 1: Calculate , , and construct the discrete problem:
where , and (); 2: Set sets , and to be empty; 3: Calculate and , then calculate and ; 4: Normalize the objective functions by the normalized design metrics (13); 5: Divide the interval [0,1] into subintervals of equal length, and add the right values of the subintervals to set ; 6: For each , , find an optimal solution of ; i.e., , set ; 7: For each , , find an optimal solution of ; i.e., . Set ; Output: The points in sets and and their corresponding optimal control solutions . |
It should be noted that when Algorithm 1 terminates, it can obtain a set of discrete non-dominated points in . These nondominated points correspond to a set of efficient control schemes ; thus, a set of representative efficient control schemes can be obtained.
5. Numerical Examples
In this section, we use an example to demonstrate the effectiveness of our proposed algorithm. Consider the relative control problem of a follower spacecraft. Suppose that the initial and the terminal state of the follower spacecraft is known. Our simulation example is from [
32]. In the example, we assume that the initial moment
,
, where the initial position is
, and the initial velocity is
. Suppose that the target state of the spacecraft is
, the control variable feasible domain is
, the sampling period is
, and the termination moment is
.
We use the open-source package OPTI [
33] to solve the optimization problem in the algorithm. We use Algorithm 1 to solve the multi-objective problem (
). First, we solve
and
to obtain the payoff matrix for state control of the spacecraft in
Table 1.
Based on the above payoff matrix, the normalized design metrics can be evaluated as follows:
We divide
into 20 equal parts; thus, elasticity parameter variable
can be obtained. We run Algorithm 1 with the input of feasible set
and
to obtain a set of optimal solutions
as well as the corresponding sets
and
, which are shown in
Table 2.
It is known that the objective functions are convex and continuously differentiable and that the control variable feasible set is convex. Thus, the optimal solutions obtained by solving
are also efficient solutions of problem (MOP). Pareto front can be obtained by connecting the nondominated points in
Table 2, as shown in
Figure 3. In the figure, each nondominated point corresponds to a control scheme.
In
Figure 3, the two ending points are corresponding to two individual minima. The nondominated points shown by asterisks “
✳” are obtained by solving the ε-constraint problems with the orbital motion-control error as the objective function and the energy consumption bounded with an upper threshold
as a constraint. The nondominated points shown by asterisks “○” are obtained by solving the ε-constraint problems with the energy consumption as the objective function and the orbital motion-control error bounded with an upper threshold
as a constraint. Here, we select representative points
and
in
Figure 3 for further analysis.
For the nondominated point
in
Figure 3, we show the corresponding simulation results in
Figure 4.
Figure 4 shows the position change curve, velocity change curve, and control force change curve of the relative motion of the spacecraft. From
Figure 4, we can see that the spacecraft reaches termination state at about 7 s. At the same time, the control moment becomes zero, and the system reaches a steady state.
For the nondominated point
in
Figure 3, we show its corresponding simulation results in
Figure 5. We can see that it takes nearly 10 s for the spacecraft to reach the target state while maintaining the steady state and with the control moment being zero. Comparing
with
,
takes more time to achieve the target state than
.
In addition, it is worth noting that from
Figure 4 and
Figure 5, we can see that the state change trajectory of
and
are smooth and without sudden changes. This indicates that the control scheme we obtained is reasonable, and such a control scheme can also make the control system more stable.
We also show the energy consumptions (the second objective value of (
)) corresponding to
and
in
Figure 6.
It can be seen from
Figure 4,
Figure 5 and
Figure 6 that the control scheme corresponding to
reaches the target state faster than
but consumes more energy. Because
and
are two conflicting objectives, for a nondominated point, improving one objective makes another objective worse.
From the above, it can be seen that the control force sequence can be obtained by solving the multi-objective optimization problem. This method optimizes the orbital motion error and the energy consumption at the same time. The numerical results show the effectiveness of the method.
In summary, for the orbital motion-control problem () where the objective functions are convex continuously differentiable, and the feasible set of control variables is convex, the complete Pareto front can be obtained by using relative orbital motion-control algorithm based on ε-constraint. The decision maker can choose a more appropriate control scheme according to the specific situation.