1. Introduction
Dynamic positioning (DP) system is a critical piece of equipment in marine surface vessel systems to automatically maintain a desired horizontal position and heading or track the desired motion trajectory by using its own actuators [
1]. Examples of vessel types that employ dynamic positioning system include supply vessels, pipe-laying ships, oceanographic research vessels, and rescue ships [
2,
3,
4].
Robust control combined with a disturbance observer technique is a constructive method to suppress feedback gains [
5]. The core idea of the disturbance observer is to gather all unknown factors of the system into a single disturbance term and design an additional observer to estimate this unknown term. A finite-time sliding mode active disturbance rejection observer to estimate the unknown external disturbance of multiple ships was proposed in [
6]. In [
7], a fixed-time disturbance-observer-based controller was designed to handle actuator dead zones and disturbances for underactuated surface vehicles. An adaptive nonlinear disturbance observer was developed to estimate unknown time-varying uncertainties for dynamic positioning of ships in [
8]. In [
9], a fractional-order disturbance observer was designed by using fractional-order theory to estimate unknown disturbances in dynamic positioning systems. In [
10], an adaptive sliding mode control method was proposed. A continuous projection-based adaptive law was adopted to deal with uncertainty and external load thrust, but it was assumed that the upper bound of the uncertainty was known. Most of the methods described above require the use of upper-bound information of the derivative of the disturbance in the design of a disturbance observer.
Dynamic positioning concerns the control of marine surface vessels in the horizontal plane, i.e., the three axes: surge (longitudinal), sway (lateral), and yaw (rotation about the up/down axis). In order for the ship to more efficiently track the desired position, more actuators than necessary for ensuring controllability are generally equipped, in which case the system is referred as overactuated [
11]. Another reason for input redundancy is the need to guarantee fault tolerance in the control system [
12]. Redundant actuators might take over the tasks of defective ones if necessary. In fault-free cases, they can be used to support the overall actuation of the system. Although actuator redundancy can improve the fault tolerance of the system, it also increases the complexity of the closed-loop control system design. Due to input redundancy in general, infinite combinations of actuator actions result in the same effect on the system [
11]. The control-allocation-based method is one strategy to deal with this problem class. In this case, the first high-level control layer operates on the basis of a mathematical system model that neglects redundancy by replacing real actuators with a smaller number of artificial ones. Subsequently, the CA scheme is required to distribute the desired generalized force command from the high-level controller to the redundant actuators, so that the desired generalized force can be produced. The major difficulty of CA is that actuators always exhibit a limited operation range, leading to constrained optimization problems.
There are many methods to solve the control allocation problem (CAP), e.g., dynamic control allocation [
13], direct control allocation [
14], redistributed pseudoinverse [
15], and optimization-based allocation [
16]. A dynamic control allocation scheme was proposed in [
17], where the resulting control distribution depends both on the generalized force at the current moment and the control distribution in the previous sampling instant, so that the actuator rate constraints can be handled. Direct control allocation schemes can guarantee that the maximal attainable forces without losing the desired direction can be generated within actuator constraints. However, in most cases, direct and optimization-based control allocations often require the control system to have strong computational power; thus, it is usually impossible to be solved quickly online. Practically, the CA problem must be solved within a discrete-time environment with very limited computational resources. Consequently, computational complexity should be kept as small as possible, and efficient implementation is crucial for a successful application. The redistributed pseudoinverse (RPI) method has flexibility on the computational power requirement by limiting the number of iterations per step. It can improve performance even if some actuators are saturated, and it is relatively simple to implement. However, above a certain number of active actuator constraints, the RPI scheme often fails to achieve the desired generalized force signal. To address this problem, a novel enhanced redistributed pseudoinverse (ERPI) algorithm was proposed in [
18].
State constraints are ubiquitous in motion control systems. From the perspective of ship navigation safety, it is generally required that the DP system states operate under specific constraints. Ignoring these states constraints possibly degrades system control performance or even results in disasters [
19,
20]. Therefore, it is essential and desirable to take the system constraints into account in the motion control of DP ships. The barrier Lyapunov function (BLF) is a well-known approach for addressing state constraints [
21]. In the BLF-based method, Lyapunov function values grow to infinity if the states approach constraint boundaries. Therefore, it is possible to guarantee that the system state does not violate the constraints by making the BLF bounded. However, in order to prevent violation of state constraints, the virtual control law is always required to satisfy the so-called feasibility conditions. In general, ensuring feasibility conditions tends to be conservative, since the original state constraints are enforced indirectly by imposing transformed constraints on the tracking errors [
22]. Although the improved integral BLF-based method can avoid the conservativeness of feasibility conditions, the parameter design process of the integral BLF method is not only complex but also time-consuming [
23]. In this paper, instead of employing the commonly used BLF, states constraints are tackled gracefully by introducing unified barrier function (UBF) into the backstepping procedure [
24]. By introducing a new coordinate transformation technique, a novel backstepping control strategy which completely avoids the use of feasibility conditions is designed.
Motivated by the above observations, this paper presents a control-allocation-based robust tracking control method for DP ships subject to unknown disturbances, actuator constraints, and asymmetric state constraints. The specific contributions of this paper can be summarized as follows.
(1) A practical robust tracking control approach of the DP ship subject to environmental disturbances and asymmetric state constraints is proposed. The stability of the proposed method is proved, and simulation results are presented to verify its effectiveness.
(2) On the basis of sliding-mode techniques, a novel double-layer adaptive disturbance observer was designed to estimate the lumped uncertainty of the DP ship. Finite-time estimation of uncertainty can be achieved though upper-bound information of the uncertainty, and its derivative is unknown.
(3) By blending the unified barrier function technique into backstepping design, a control strategy completely obviating the feasibility condition was developed to address asymmetric state constraints.
(4) In practical application, actuators are subject to saturation constraints. In order to address this issue, an enhanced redistributed pseudoinverse algorithm was applied to control allocation for the dynamic positioning system.
The paper is organized as follows.
Section 2 presents the system description and problem statement. The double-layer adaptive sliding mode disturbance observer and DP controller are presented in
Section 3. Control allocation based on the enhanced redistributed pseudoinverse (ERPI) algorithm is proposed in
Section 4.
Section 5 highlights the controller performance through simulations. Lastly, concluding remarks are drawn in
Section 6.
For a given vector , all operations on ∨ are component-wise unless stated otherwise, and these operations operate on each component of vector ∨ and yield a new vector. Similarly, relational operators that are applied on vectors must be fulfilled element-wise. Further, given a vector ∨, we denote by ∨ diagonal matrix . The right null-space of matrix A is denoted by , and the left null-space is denoted by . and are the minimal and maximal eigenvalues of matrix A, respectively. Assignments in algorithms are indicated with the ← operator.
4. Formulation of Control Allocation
In this section, control allocation is solved by using the ERPI algorithm to determine the thrust and direction of each actuator. For DP ships, generalized force signal
is related to the control input through equation
with
where
denotes the angle between the force of the
i-th actuator and surge direction,
is the location of the
i-th actuator in the body-fixed frame, and
is the actuator commands.
For ships, additional control inputs
given by an azimuth thruster lead to a nonlinear optimization problem, which is nontrivial to solve. To this end, the extended thrust vector is constructed by decomposing the individual thrust vector in the horizontal plane according to
where subscript
denotes the
i-th azimuth thruster,
and
represent the equivalent control inputs for the
i-th azimuth thruster. Then, the generalized force vector is given by the linear equation
where
, with
stands for the thrust of the
i-th transverse tunnel thrusters.
and
denote the number of azimuth thrusters and transverse tunnel thrusters, respectively. Let
, then the matrices
and
.
Consider 3-dimensional (3-D) generalized force Equation (
54); then, the system is called input-redundant if
. Thus, for any
,
, we have
. In other words,
does not have full column rank, so perturbations of
u in null-space directions do not affect system dynamics. This enables an input matrix factorization of
into a virtual input matrix
and a control effectivity matrix
, i.e.,
with both of their ranks being
Then, we have
where
is called the virtual control vector, the control effectivity matrix
B characterizing the relationship between
v and
u. Typically, actuators are subject to constraints that define the feasible subset of
-D control space
Remark 5. Definition Ω nvagnitude and rate of actuator actions. All these can be treated as time-varying bounds in principle. Due to the fact that CA is carried out in each step, bounds are still considered to be constant during one execution cycle. The CA problem is to find a control command that fulfils (57) for given desired generalized force τ and virtual input matrix . However, this goal can be unachievable if the desired τ exceeds the capabilities of the actuators because of constraints. In such situations, the objective of the CA algorithm is to minimize the allocation error in some sense. 4.1. Weighted Pseudoinverse
As weighted pseudoinverse (WPI) algorithm is the basis of the ERPI control allocation method, the WPI is first introduced. Since the unconstrained CA problem theoretically has an infinite number of solutions, a reasonable choice is to choose the solution with the lowest energy consumption. Pseudoinverse algorithm is a common way to solve this CA problem and can be expressed as follows:
where
is a weighting matrix, and
is an offset vector. The purpose of
is to specify preferred control positions or to account for saturated actuators. To solve this constrained optimization problem, the Lagrangian function was designed as follows:
where
is the Lagrangian multiplier. Taking the partial derivatives of
L w.r.t.
u and
yields
Using (
61), virtual control
v can be obtained as follows:
From (
61) and (
63), the closed-form solution of CA problem is derived as
where
is weighted pseudoinverse; if
,
is the Moore–Penrose pseudoinverse.
4.2. Enhanced Redistributed Pseudo-Inverse
If all of the actuator constraints are inactive, the optimal solution is obtained by the weighted pseudoinverse (WPI) as in (
64). However, if inequality constraints are violated, that is, some elements of
u exceed the limits (
58), recalculation is necessary to obtain an optimal solution within limits.
For the convenience of describing the RPI algorithm, a new matrix variable
is defined that is modified at each step of the iteration. If an actuator triggers saturation, corresponding element
is set to its saturation value, the corresponding column of
is set to zero, and a reduced pseudoinverse
is computed. The offset vector
of the N-th iteration is calculated by
Then, the solution of RPI is obtained as
where
denotes the original control effectivity matrix.
From (
66), the resulting actual control after the application of RPI is obtained:
where
is the desired virtual control input vector. Obviously, sufficient conditions for the desired
can be achieved as
Let
j be the number of actuators that did not trigger saturation after executing the RPI algorithm, and
k be the generalized force space dimension (here,
). The authors in [
18] showed that, in the case of a lack of
and
, desired virtual control input
is not reached any more by the RPI algorithm because
. On this basis, a new enhanced redistributed pseudoinverse (ERPI) control allocation algorithm was proposed in [
18]. If an exact solution cannot be achieved due to actuator saturation, the errors of those components with high priority are forced to zero in preference. This goal is achieved by altering the factorization (
57) during execution. The following is a summary of the ERPI method:
Step (1) Find pseudoinverse resolution (
64) of the system. If no actuator triggers the saturation, the process stops, and the WPI solution (
64) is used.
Step (2) If any actuator triggers the saturation bound, recalculate modified matrix
and offset
according to (
65).
Step (3) If
, CA problem is solved again according to (
66). This process is repeated until no new saturation is triggered or
.
Step (4) If , let be the columns of corresponding to the actuators that do not trigger saturation. Select j rows of with indices equal to the first j entries of the priority list and refer to that matrix as . The remaining rows of are called .
Step (5) If
, the RPI solution (
66) is used.
Step (6) If , choose an arbitrary full column rank matrix and compute a basis of its left null-space . Evaluate to obtain the inverse transformation matrix with .
Step (7) Use
,
and
to transform
,
and
. Then,
is used to reformulate (
66) as:
Step (8) The algorithm is stopped and the ERPI solution (
69) is used if no new saturation is triggered or
. Otherwise, go to step 2 again.
Remark 6. Because modified matrix typically becomes singular at some point, it is necessary to replace it by a regularized matrix , where is a small regularization parameter.
5. Simulation
In this section, simulation results are presented to confirm the effectiveness of the proposed CA-based robust tracking control scheme. The numerical values of all hydrodynamic coefficients are given in
Table 1. The diagram of the thruster distribution is shown in
Figure 2. The ship has three azimuth thrusters
,
and
, and two transverse tunnel thruster
and
. It is easy to verify that this actuator configuration is designed with sufficient redundancy for the ship to have the capability to maintain position after any single fault in the thruster. It is assumed that each azimuth thruster can rotate 360 degrees. Moreover, taking
,
,
,
,
. For simplicity, the actual output limits of thrusters are given by
145,000,
. In addition, the initial values of ASMDO are selected as
,
and
. The initial position and velocity of the ship are given as
and
, respectively. The other control parameters are given in
Table 2.
In order to better show the superiority of the proposed ASMDO, the following adaptive nonlinear disturbance observer (ANDO) proposed in [
8] and sliding mode disturbance observer (SMDO) proposed in [
6] were constructed to compare performance.
where
is selected as
in the simulation.
where
, the design parameters of SMDO were selected to be
,
,
,
,
in the simulation. Moreover, to validate the performance of the ASMDO proposed in our paper, simulations are done in two different disturbance cases:
Continuous type disturbance. The disturbance vector is designed as follows
Step-type disturbance. Disturbance is represented by
, where
b is the output of the 1st order Markov process (
6). The band-limited white-noise block with sample time 2 [s] and noise power
was used in Simulink to generates the white noise. The initial value for Markov process was selected to be
. According to [
16], parameter matrices were chosen to be
and
.
Simulation results are presented in
Figure 3 and
Figure 4. First,
Figure 3 shows that, for continuous-type disturbance, although the adaptive nonlinear disturbance observer (ANDO) can achieve effective estimation, its estimation error is relatively large. This is because the ANDO can theoretically only guarantee that the estimation error is bounded [
8]. In addition, due to the use of a fixed switching gain
L in (
71), the chattering of the sliding mode disturbance observer (SMDO) is significant. This leads to a large chattering phenomenon in the sliding surface. The adaptive sliding mode disturbance observer (ASMDO) designed in this paper can both estimate unknown disturbance in finite time and, due to using the double-layer nested adaptive gain
rather than a fixed gain
L, ensure low chattering in the sliding mode, achieving an accurate estimation of disturbance, even if the bound of the derivative of the disturbance is unknown. Adaptive switching gain
is shown in
Figure 3, varying relative to
. In the simulation,
closely follows
, which is a sufficient condition to enforce a finite-time sliding motion in (
20). Further, from
Figure 4, we can learn that the proposed ASMDO can successfully achieve a better estimation than ANDO and SMDO, even for the discontinuous disturbance.
Command reference signal
is given by
To obtain the differentiable commands, desired reference command
is generated by the following filter
where
,
. Meanwhile, to satisfy trajectory tracking requirements, constraining functions
and
are selected by
where
Further, constraining functions
and
are selected by
To illustrate the performance of the proposed control method, simulation comparisons are presented by applying the backstepping controller (BS) and backstepping approach in conjunction with UBF (BSUBF). Simulation results are depicted in
Figure 5,
Figure 6,
Figure 7 and
Figure 8, where
Figure 5 and
Figure 6 show the trajectories of
and
with asymmetric full-state constraints.
Figure 5 shows that the introduction of UBF has a distinct effect on guaranteeing state constraints.
Figure 7 shows that both proposed controller and BSUBF controller could achieve relatively accurate tracking without violating the state constraints, but the tracking performance of the proposed controller may have been superior, as could be inferred from the quantitative comparison summarized in
Table 3. From the above analysis, it is clear that the control law proposed in this paper could maintain the desired performance in the presence of uncertain disturbances and asymmetric time-varying state constraints. In addition, due to the adoption of the ERPI control allocation scheme, the proposed control law can also address the saturation problem of a limited number of actuators to a certain extent. However, under the proposed CA-based control method, the ERPI control allocation scheme could not always perfectly solve the actuator saturation problem since the number of free actuators was less than 2 in the initial stage of the simulation (
Figure 8). How to better deal with the actuator saturation problem is also the focus of our future work.