Next Article in Journal
High Quality Factor Miniaturized CMOS Transmission Lines Using Cascaded T-Networks for 60 GHz Millimeter Wave Applications
Next Article in Special Issue
Optimal Control of a Semi-Active Suspension System Collaborated by an Active Aerodynamic Surface Based on a Quarter-Car Model
Previous Article in Journal
Path Planning in Complex Environments Using Attention-Based Deep Deterministic Policy Gradient
Previous Article in Special Issue
Multi-Objective Parameter Configuration Optimization of Hydrogen Fuel Cell Hybrid Power System for Locomotives
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MPC-TD3 Trajectory Tracking Control for Electrically Driven Unmanned Tracked Vehicles

China North Vehicle Research Institute, Beijing 100072, China
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(18), 3747; https://doi.org/10.3390/electronics13183747
Submission received: 25 August 2024 / Revised: 14 September 2024 / Accepted: 14 September 2024 / Published: 20 September 2024

Abstract

:
To address the trajectory tracking issue of unmanned tracked vehicles, the majority of studies employ the Model Predictive Control (MPC). The MPC imposes high demands on model accuracy. Due to factors such as environmental interference, actuator constraints, and the nonlinearity of vehicles under high-speed conditions, dynamic and kinematic models fail to accurately delineate the motion process of tracked vehicles. Aiming at the problem of insufficient trajectory tracking precision of unmanned tracked vehicles, a trajectory tracking controller jointly controlled by the Twin Delayed Deep Deterministic policy gradient (TD3) algorithm and the MPC algorithm is developed. During offline training, the agent acquires the discrepancies between the model and the environment under various working conditions and optimizes its own network; during online reasoning, the agent adaptively compensates the output of the MPC based on the vehicle state. The experimental results indicate that, compared with the pure MPC algorithm, the MPC algorithm compensated based on the TD3 algorithm reduces the lateral errors by 41.67% and 22.55%, respectively, in circular and double-lane-change trajectory conditions.

1. Introduction

In recent years, with the application of electric drive technology for tracked vehicles, high-speed tracked vehicles have been evolving toward intelligence and unmanned operation. Electric drive tracked vehicles exhibit numerous advantages in control, such as precise torque control, rapid response speed, braking energy recovery, etc. [1,2]. For the control system, the motor serves as a precise actuator. This establishes the foundation for the unmanned operation of tracked vehicles. Regarding manned tracked vehicles, the driver achieves the vehicle’s tracking of the target road or trajectory by manipulating the steering wheel angle. The objective of the controller is to guarantee safety while precisely implementing the control quantity. The majority of studies have concentrated on vehicle stability control [3,4,5,6]. For unmanned tracked vehicles, the key technologies can be generalized as environmental perception, positioning and navigation, and decision making and planning, as well as trajectory tracking [7]. Whereas trajectory tracking, as the ultimate safeguard of the maneuverability and driving safety of unmanned tracked vehicles, demands high tracking precision.
The traditional approaches to trajectory tracking can be classified into three categories: geometry-based methods, model-free methods, and model-based methods. Pure pursuit (PP) and Stanley methods are geometry-based approaches. These two methods determine the front wheel steering angle command based on the geometric relationship between the vehicle and the desired trajectory. Ahn J et al. [8] and AbdElmoniem A et al. [9] verified the effectiveness of these two methods through experiments. Nevertheless, the control accuracy of geometry-based methods is not satisfactory under high-speed and off-road conditions. Among the model-free control methods, the PID algorithm is widely employed due to its simplicity and extensive application scope [10,11]. However, since its parameter tuning methods largely rely on experience, in trajectory tracking control, inappropriate parameters can lead to vehicle instability, entailing high test costs and time consumption. LQR and MPC algorithms are typical representatives of model-based methods. The quality of their control effects hinges on the accuracy of the model. Zhao Z et al. [12] enhanced the trajectory tracking accuracy of the vehicle by employing the method of optimizing the model. Through observing the slip and skid amounts of the tracks on both sides of the tracked vehicle via the extended Kalman filter, they were utilized as parameters to compensate for the model of the MPC controller.
In recent years, reinforcement learning has gradually been applied in the trajectory tracking control of vehicles. Srikonda S et al. [13] utilized the Deep Deterministic Policy Gradient (DDPG) algorithm in place of the traditional controller to realize vehicle trajectory tracking. Liu M et al. [14], in an attempt to enhance the training speed, initially employed data for the agent to undertake imitation learning, and subsequently conducted reinforcement learning to increase exploration, achieving trajectory tracking on urban roads in wheeled vehicles. However, for scenarios not encountered during the reinforcement learning training, the output actions of the agent can readily cause the vehicle to lose control. Hence, some scholars endeavor to combine reinforcement learning with traditional control algorithms. Shan Y et al. [15] employed the Proximal Policy Optimization (PPO) algorithm to regulate the output proportion of the PID algorithm and the PP algorithm, dynamically integrating the advantages of the two algorithms. Nevertheless, this method remains confined within the framework of traditional control methods. Wang S et al. [16] utilized Q-learning to compensate for the output of the PID algorithm, but the Q-learning method is suitable for control in discrete state spaces and would give rise to the dimension curse for continuous state spaces. Chen I M et al. [17] employed the PPO algorithm to compensate for the output of the PP algorithm, demonstrating excellent real-time performance.
In the domain of trajectory tracking of unmanned tracked vehicles, the majority of studies still adopt traditional control methods [18,19,20]. Nevertheless, there exist the following challenges in the control of high-speed tracked vehicles: On the one hand, the slip and skid between the tracks of high-speed tracked vehicles and the ground significantly increase, resulting in intensified nonlinear characteristics of the tracked vehicles. On the other hand, compared to the steering of wheeled vehicles, the differential steering response speed of tracked vehicles is slower. In response to the aforementioned challenges, this paper proposes the MPC-TD3 control method. Herein, the MPC controller provides the control quantity based on the simplified kinematic model of the tracked vehicle. For the portion where the kinematic model inaccurately describes the tracked vehicle model, the TD3 algorithm is employed to compensate for the output of the MPC controller.

2. Tracked Vehicle Model Construction

2.1. Kinematic Model of Tracked Vehicle

The global coordinate system X O Y and the vehicle body coordinate system x o y are established as shown in Figure 1:
The position and attitude of the vehicle in the global coordinate system is ( X , Y , θ ) , where θ is the heading angle of the tracked vehicle. In the local vehicle body coordinate system, v x , v y , ω z are the longitudinal velocity, lateral velocity, and yaw rate of the vehicle, respectively; o c is the steering center of the tracked vehicle, o l is the instantaneous steering center of the left track of the tracked vehicle, and o r is the instantaneous steering center of the right track of the tracked vehicle. ( x c , y c ) , ( x l , y l ) , ( x r , y r ) are the coordinates of these three points, respectively. These three points are on the same straight line, and this straight line intersects the left track and the right track at points M and N , respectively. v s l , v s r are the winding speeds of the left and right tracks, respectively; and v M q , v N q are the entrainment velocities of the vehicle body at points M and N, respectively.
Then the kinematic model of the tracked vehicle is:
X ˙ Y ˙ θ ˙ = cos θ sin θ 0 sin θ cos θ 0 0 0 1 v x v y ω z = cos θ sin θ 0 sin θ cos θ 0 0 0 1 y l v s r y r v s l y l y r v s l v s r y l y r x c v s l v s r y l y r

2.2. Dynamic Model of the Tracked Vehicle

The force diagram of the tracked vehicle is shown in Figure 2 as follows:
The normal pressures imposed on the track plates on both sides of a tracked vehicle during steering can be represented as:
N l = 1 2 m g m B ( h ( v ˙ y + v x ω z ) c y g ) N r = 1 2 m g + m B ( h ( v ˙ y + v x ω z ) c y g ) P l , i = N 1 n 6 m L 2 n ( c x g + h ( v ˙ x v y ω z ) ) x l i P r , i = N 2 n 6 m L 2 n ( c x g + h ( v ˙ x v y ω z ) ) x r i
where N l , N r are the normal pressures exerted on the two sides of the tracks, respectively, P l i , P r i are the pressures exerted on each track plate on the two sides, respectively, and m is the mass of the tracked vehicle. L is the grounding length of the track, B is the center distance between the two sides of the tracks, n is the number of track plates on the single side of the track, c x , c y are the offsets of the center of mass of the tracked vehicle, respectively, and x l i , x r i is the abscissa of each track plate in the vehicle coordinate system, respectively.
In the vehicle coordinate system, the coordinates of each track plate are:
x k i = L 2 t 0 i t v s k r d t y k i = B 2
where v s k is the winding speed of the k th side track, and t 0 i is the moment when this track plate begins to contact the ground.
The absolute velocities of each track plate are:
v j x , k i = v s k ( v x ± w z ( B 2 c y ) ) v j y , k i = v y w z ( x k i c x )
Then, the shear displacement of each track plate is:
j x , k i = t 0 i t v j x , k i ( τ ) d t j y , k i = t 0 i t v j y , k i ( τ ) d t j k i = j x , k i 2 + j y , k i 2
Then, the driving forces received by the two sides of the tracks in the longitudinal and lateral directions are, respectively:
F x , k = i = 0 n μ P k i ( 1 exp ( j k i / K ) ) sin δ F y , k = i = 0 n μ P k i ( 1 exp ( j k i / K ) ) cos δ
where μ is the friction coefficient between the track and the ground, K is the ground shear modulus, and δ is the vehicle sideslip angle, expressed as:
δ = v x v x 2 + v y 2
The resistances received by the two sides of the tracks respectively are:
R f , k = f N k
where f is the rolling resistance.
The driving torque and resistant torque exerted on the tracked vehicle are:
T D = i = 0 n μ P l i ( B 2 c y ) ( 1 exp ( j l i / K ) ) sin δ + i = 0 n μ P r i ( B 2 c y ) ( 1 exp ( j r i / K ) ) sin δ T f = i = 0 n x l i μ P l i ( 1 exp ( j l i / K ) ) cos δ + i = 0 n x r i μ P r i ( 1 exp ( j r i / K ) ) cos δ
Based on the above formulas, the dynamic equilibrium equation of the tracked vehicle is obtained:
F x , l + F x , r ( R f , l + R f , r ) = m ( v ˙ x v y w z ) F y , l + F y , r = m ( v ˙ y v x w z ) T D T f ( R f , l + R f , r ) B 2 = I z ω ˙ z
The dynamics simulation model of the tracked vehicle is shown in Figure 3:

2.3. Verification of the Dynamics Model of Tracked Vehicles

Figure 4 and Figure 5 respectively present the comparison diagrams of the actual and simulation results of the driving trajectory and vehicle speed of a certain type of crawler vehicle on the dirt road. The vehicle weighs 24 tons, with L being 4.4 m and B being 2.71 m. In the simulation environment, μ is 0.8, K is 0.015, and f is 0.06. It is observable from the simulation results that this dynamic model can relatively accurately describe the state of the vehicle.

3. Construction of the MPC Controller

3.1. Construction of the Controller

The MPC controller utilizes the kinematic model to predict the vehicle’s position and posture within a future period, solves the optimal control sequence through quadratic programming, and outputs the first set of control quantities. During the process of quadratic programming solution, constraint conditions such as control increment limitations and control quantity magnitude limitations can be added, thereby considering the response capabilities of the actuators. The kinematic model of the tracked vehicle is a nonlinear equation and can be expressed as:
X ˙ = f ( X , u )
where the state quantities are X = [ X , Y , θ ] T , and the control quantities are u = [ v s l , v s r ] T .
Linearize and discretize Equation (4). Perform the Taylor expansion of the kinematic equation at the reference trajectory point and discretize it through the forward Euler method as follows:
X ˜ ( k + 1 ) = A X ˜ ( k ) + B u ˜ ( k )
where T s is the sampling time, v s d l , v s d r and are the reference values of the rotational speeds of the left and right driving wheels, and θ d is the reference value of the heading angle.
A = 1 0 T s ( y l v s d r y r v s d l y l y r sin θ d v s d l v s d r y l y r x c cos θ d ) 0 1 T s ( y l v s d r y r v s d l y l y r cos θ d v s d l v s d r y l y r x c sin θ d ) 0 0 1
B = T s ( y r y l y r cos θ d x c y l y r sin θ d ) T s ( y l y l y r cos θ d + x c y l y r sin θ d ) T s ( y r y l y r sin θ d + x c y l y r cos θ d ) T s ( y l y l y r sin θ d x c y l y r cos θ d ) T s y l y r T s y l y r
Considering the need to limit the acceleration of the winding speed, let:
ξ ( k ) = X ˜ ( k ) u ˜ ( k 1 )
The new state-space expression is:
ξ ( k + 1 ) = A ˜ ξ ( k ) + B ˜ Δ U ( k ) η ( k ) = C ˜ ξ ( k + 1 )
where A ˜ = A B 0 m × n I m ; B ˜ = B I m ;   n is the dimension of the state quantity; and m is the dimension of the control quantity.
Through iterative derivation, the predicted output expression of the system can be obtained:
Y ( k ) = ψ ξ ( k ) + Θ Δ U ( k )
The objective function of the model predictive controller is as follows:
J ( k ) = i = 1 N p η ( k + i t ) η r e f ( k + i t ) Q 2 + i = 1 N c 1 Δ U ( k + i t ) R 2 + ρ ε 2
where N p is the prediction horizon; N c is the control horizon; ρ is the weighting coefficient; and ε is the relaxation factor.
Transform it into the form of a quadratic form:
J = Δ U T ε Θ T Q Θ 0 0 ρ Δ U ε + 2 E T Q Θ 0 Δ U ε
The interior point method is adopted to solve it to obtain the control sequence, and the first group of control quantities is taken as the output of the controller.

3.2. Constraint Conditions

Considering the limited response capability of the motor and to prevent the vehicle from rollover, the increment of the rotational speed of the driving wheels is restricted, as well as the maximum value and the rotational speed difference between the left and right driving wheels.
The resistance received by the motor can be observed through the Luenberger observer. At the same time, according to the external characteristic curve of the motor, the maximum torque that the motor can output at the current rotational speed can be obtained, and the maximum rotational speed increment of the motors on both sides can be obtained:
Δ ω max l = T e max l T f l I c Δ ω max r = T e max r T f r I c
where I c is the moment of inertia of the entire vehicle mass equivalent to the rotation of the motor
The maximum rotational speed increment of the driving wheel is obtained through the coupling mechanism and the reducer:
Δ v s max l Δ v s max r = r i b 2 + k 2 ( 1 + k ) k 2 ( 1 + k ) k 2 ( 1 + k ) 2 + k 2 ( 1 + k ) Δ ω max l Δ ω max r
where k is the planetary row parameter of the coupling mechanism, i b is the total reduction ratio, and r is the radius of the driving wheel.
Bring it into the following formula as a constraint condition:
Δ u min ( t + k ) Δ u ( t + k ) Δ u max ( t + k ) , k = 0 , 1 , , N c 1
The maximum rotational speed and the rotational speed difference of the driving wheel are restricted; that is, it is ensured that in the model prediction process, the control quantity of each prediction solution is within the constraint conditions. The maximum rotational speed can be obtained through the following formula:
ω max = P max T e l + T e r
Similarly, through the coupling mechanism, bring it into the following constraint conditions:
A t 0 A t 0 A w t 0 A w t 0 Δ U 0 Δ U 0 U m a x U t U m a x + U t U w m a x U w t U w m a x + U w t
where:
A t = 1 0 0 1 1 0 0 1 1 1 1 N c × N c I m A w t = 1 1 1 1 I N c U t = 1 N c × 1 u ( k 1 ) U t = 1 N c × 1 u 1 ( k 1 ) u 2 ( k 1 ) u 1 ( k 1 ) u 2 ( k 1 )
This constraint condition can simultaneously constrain the rotational speed value and the rotational speed difference on the left and right sides. Among them, m is the dimension of the control quantity, and represents the Kronecker product. u ( k 1 ) is the control quantity at the previous moment, and u 1 ( k 1 ) u 2 ( k 1 ) represent the first and second elements in the control quantity, namely the rotational speeds of the left and right wheels. U m a x , U w m a x are the maximum rotational speed of the driving wheel and the maximum rotational speed difference between the left and right driving wheels, respectively.

4. Construction of the TD3 Agent Compensation Module

4.1. Design of State Space and Action Space

In this paper, the TD3 agent is adopted to compensate for the deviations in vehicle position and attitude caused by the inaccuracy of the MPC controller model. Firstly, the lateral deviation, longitudinal deviation, and heading angle deviation of the vehicle in the global coordinate system need to be observed, which are e X , e Y , e θ , respectively. At the same time, the situation where the input states are the same but the expected actions are different should be avoided as much as possible to avoid the non-convergence of the model. The change rates of yaw rate, lateral deviation, longitudinal deviation, and heading angle deviation of the vehicle need to be introduced, which are ω z , e ˙ X , e ˙ Y , e ˙ θ , respectively. Finally, considering the deviation between the trajectory predicted by the model and the actual trajectory, the deviations of the vehicle state at the current moment predicted by the vehicle state at the previous moment and the actual vehicle state at the current moment are introduced, which are e ˙ X p , e ˙ Y p , e ˙ θ p , respectively.
To sum up, the state space is designed as s = [ e X , e Y , e θ , ω z , e ˙ X , e ˙ Y , e ˙ θ , e ˙ X p , e ˙ Y p , e ˙ θ p ] , and the action space is designed as the compensation amount a = v ω for the speed difference of the left and right driving wheels.

4.2. Reward Function Design

The smaller the deviation of the vehicle’s position and attitude, the greater the reward value should be. In order to unify the magnitude of lateral deviation, longitudinal deviation, and yaw angle deviation, and to accelerate the convergence speed of the neural network, we set the deviation as an exponential function:
r X Y = 3 e 0.05 e X 2 0.05 e Y 2 r θ = 0.3 e 40 e θ 2
In order to keep the output compensation rotational speed increment of the agent within the constraint range and prevent the instability of the tracked vehicle caused by the sudden change in its output, when the variation in the neural network output is too large, a certain penalty should be given. The penalty is designed as a continuous value to avoid the problem of non-convergence of training caused by giving different rewards for the same state:
r u ˙ = 0.5 a ( k ) a ( k 1 )
The final reward function is designed as
r = r X Y + r θ + r u r X Y = 3 e 0.05 e X 2 0.05 e Y 2 r θ = 0.3 e 40 e θ 2 r u ˙ = 0.5 a ( k ) a ( k 1 )

4.3. Design of the TD3 Agent

The TD3 algorithm improves the DDPG algorithm in three aspects. Introducing dual Critic networks to alleviate the overestimation problem; adding random noise to the target action; delaying the update of the Actor network.
First of all, the TD3 algorithm introduces dual Critic networks to alleviate the overestimation problem. The DDPG algorithm adopts the Actor–Critic structure, in which the Critic network uses the Double Q-learning (DQN) algorithm. Therefore, the DDPG algorithm will have the same overestimation problem as the DQN algorithm. The neural network update process of DQN can be expressed as:
y = r + γ max a Q s , a ; w w w α y Q s , a ; w Q s , a ; w w
In the formula, y represents the target Q-value, r is the reward of the environment, γ is the discount factor that determines the priority of short-term rewards, α is the learning rate, Q s , a ; w represents the Q-value estimated by the Critic network at the current moment, s , a represent the state space and action space at the next moment, and Q s , a ; w represents the Q-value estimated by the Critic network at the next moment.
The maximization of Q-values and bootstrapping lead to the overestimation problem of Q-values. The training process is based on the method of experience replay, randomly extracting quadruples for update. In this process, the originally low Q-values may be overestimated, resulting in incorrect policies and reducing the training efficiency. The Target Network algorithm and the Double DQN algorithm alleviate the overestimation problem of Q-values to a certain extent, but the overestimation problem of Q-values in the DQN algorithm and DDPG still exists. The TD3 algorithm uses dual Critic networks. When calculating the Q-value estimated for the future at the next moment, it outputs the smaller output value among the two Target Critic networks:
y = r + γ min i = 1 , 2 Q i s , μ s ; ϕ ; w i Q i
In the formula, μ ; ϕ represents the Target Actor network, Q ; w Q represents the Target Critic network, μ s ; ϕ represents the output action at the next moment, and Q i s , a ; w i Q i represents the Q-value estimated at the next moment using the Target Critic network.
Secondly, the TD3 algorithm adds random noise to the target action. Using a deterministic policy network as the Actor network, when the Critic network is updated, the learning goal using the deterministic policy is easily affected by the error of function approximation, thereby increasing the variance of the target. This induced variance can be reduced through regularization. Therefore, the TD3 algorithm adds random noise to the target action:
y = r + γ Q s , μ s ; ϕ + ε ; w Q ε ~ c l i p N 0 , σ , c , c
Finally, TD3 adopts the method of delayed update of the Actor network, and updates the Actor network after the update of the Critic network is stable. This improves the stability of the Actor network update.
The complete control algorithm structure is shown in Figure 6. The parameter settings of the training process are shown in Table 1:
During the training process, the TD3 agent randomly samples a quadruple from the experience replay pool, which contains the current moment’s tracked vehicle’s position and other state information s , the action output by the Actor network a , the reward given by the environment r , and the next moment’s state information s . The agent inputs s , a into the Critic network to calculate the estimated Q-value at the current moment Q 1 , Q 2 . At the same time, it inputs the next moment’s state s and action value μ s ; ϕ + ε into the Target Critic network to calculate the estimated Q-value at the next moment Q 1 , Q 2 . Then, it takes the smaller of the two and adds it to the reward value to obtain the actual Q-value at the current moment. It separately calculates the TD error for each of the two Critic networks. Finally, it uses gradient descent to update the Critic network, and updates the Actor network using gradient ascent after the Critic network has been updated several times. The target network is updated using a soft update method.
During the online reasoning process, only the Actor network participates in the calculation. The agent calculates the compensation value in real time based on the current state information of the tracked vehicle.

5. Simulation and Analysis

5.1. Simulation Experiment

This article uses the dynamics model of a tracked vehicle built by MATLAB Simulink as the environment. The RL Agent module in the Simulink toolbox can be used for the training of agents, but the models trained thereby are difficult to deploy on the Huawei Atlas 200 DK controller. In this article, the Python language is used to write the MPC algorithm and train and infer the agent. Regarding the interaction problem between the agent and the environment, C code is generated by Simulink and then encapsulated as a DLL file to achieve the interaction between the agent and the controller in Python and the environment. As shown in Figure 7:
We validate the control effect using three trajectory conditions: straight-line, circular, and double-lane-change trajectory. The straight-line and circular trajectories can verify the vehicle’s track-following control effect under a constant desired turning radius. The double-lane-change trajectory is a variable-curvature trajectory, similar to actual driving scenarios such as overtaking and evading obstacles. It can verify the track-following control effect under this condition.

5.1.1. Straight-Line and Circular Trajectory Conditions in Simulation

Under this trajectory condition, the expected vehicle speed is set to 30 km/h. After driving in a straight line for 50 s, it tracks a circular path with a radius of 100 m. The simulation results are shown in Figure 8.
From the simulation results, it can be concluded that under the trajectory conditions of straight lines and circular trajectories, compared with the MPC controller, the mean values of the lateral deviation and the heading angle deviation of the MPC-TD3 controller proposed in this paper have decreased by 58.18% and 10.27% respectively. Moreover, it can be seen that the penalty for the output deviation value in the reward function has achieved a very good effect. The output value of the agent has no sudden change and has better stability.

5.1.2. Double-Lane-Change Trajectory Conditions

Under this trajectory condition, the expected vehicle speed is set at 30 km/h. After a period of linear acceleration, it enters the double-lane-change path. The simulation results are shown in Figure 9.
It can be obtained from the simulation results that under the double-lane-change trajectory condition, the maximum lateral deviation has decreased from 0.91 m to 0.29 m, reducing by 68.13%. The mean values of the lateral deviation and the heading angle deviation have decreased by 34.10% and 0.18%, respectively.

5.2. Hardware-in-the-Loop Experiment and Analysis

The hardware-in-the-loop simulation experiment is shown in Figure 10. The MPC-TD3 algorithm is deployed on the Atlas 200 DK development board produced by Huawei in Shenzhen, China, which adopts the CANN architecture. It is necessary to convert the agent saved by Pytorch into the OM model and deploy it on the development board, and conduct online reasoning using the AscendCL toolchain. The simulation environment runs in dSPACE, and the CAN bus is used for communication between the controller and dSPACE.

5.2.1. Straight-Line and Circular Trajectory Conditions in Hardware-in-the-Loop Experiment

It can be seen from Figure 11 that under this working condition with the vehicle speed of 30 km/h, when comparing the MPC-TD3 controller with the MPC controller, the mean values of the lateral deviation and the heading angle deviation have decreased by 41.67% and 37.51%, respectively. The output of the agent, due to communication interference and noise, fluctuates more than that in the simulation case, but it is still within the controllable range.

5.2.2. Double-Lane-Change Trajectory Conditions in Hardware-in-the-Loop Experiment

It can be seen from Figure 12 that under this trajectory condition with the vehicle speed of 30 km/h, when comparing the MPC-TD3 controller with the MPC controller, the maximum lateral deviation has decreased from 0.82 m to 0.36 m, a reduction of 56.10%. The mean value of the lateral deviation has decreased by 22.55%. It can be seen that the average calculation time of the MPC algorithm is 0.0448 s, and the average calculation time of the TD3 algorithm is 0.0008 s. The addition of the TD3 algorithm has almost no impact on the solution time.

6. Future Research Directions

The safety of tracked vehicle trajectory tracking control is of great importance. In this paper, the intelligent agent is constrained by using a penalty surrogate variable method, thus achieving vehicle stability. However, it is inevitable that the intelligent agent will output abrupt changes during the initial training. How to increase the exploratory nature of the intelligent agent’s training while ensuring safety is a very worthy research direction. In future research, we will try to combine human experience with the training of intelligent agents, and at the same time, impose safety constraints during training to enhance the safety of the training.

7. Results

In this paper, aiming at the problem of insufficient trajectory tracking accuracy of tracked vehicles under different trajectory conditions, an MPC-TD3 controller is proposed by combining the model-based and data-based methods, achieving an improvement in trajectory tracking accuracy. The following conclusions are obtained through simulation experiments and hardware-in-the-loop experiments: (1) The TD3 algorithm is used to adaptively compensate the output of the MPC controller, making up for the insufficient trajectory tracking accuracy caused by inaccurate vehicle models and environmental interferences, and achieving an improvement in trajectory tracking accuracy. (2) The designed reward function not only improves the control accuracy but also suppresses the problems of vehicle instability caused by the output mutation of the TD3 agent and the non-convergence of training. (3) The experimental verification under circular trajectory conditions and double-lane-change trajectory conditions is completed. Compared with the traditional MPC algorithm, the algorithm proposed in this paper reduces the lateral error by 41.67% and 22.55%, respectively, verifying the effectiveness of the algorithm.

Author Contributions

Conceptualization, Y.C.; Methodology, Y.C., J.G. and S.H.; Validation, J.G.; Investigation, S.H., H.L., C.C. and W.Z.; Resources, J.G.; Data curation, H.L., C.C. and W.Z.; Writing—original draft, Y.C.; Writing—review & editing, Y.C.; Supervision, J.G. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Wang, J.; Gao, S.; Wang, K.; Wang, Y.; Wang, Q. Wheel torque distribution optimization of four-wheel independent-drive electric vehicle for energy efficient driving. Control Eng. Pract. 2021, 110, 104779. [Google Scholar] [CrossRef]
  2. Husain, I.; Ozpineci, B.; Islam, M.S.; Gurpinar, E.; Su, G.J.; Yu, W.; Chowdhury, S.; Xue, L.; Rahman, D.; Sahu, R.; et al. Electric drive technology trends, challenges, and opportunities for future electric vehicles. Proc. IEEE 2021, 109, 1039–1059. [Google Scholar] [CrossRef]
  3. Yuan, Y.; Gai, J.; Zeng, G.; Zhou, G.; Li, X.; Ma, C. Analysis and Experimental Verification of Yaw Motion Response Characteristics of High-speed Tracked Vehicle. Acta Armamentarii 2024, 45, 1094–1107. [Google Scholar]
  4. Yuan, Y.; Gai, J.; Zhou, G.; Gao, X.; Li, X.; Ma, C. Analysis of High-Speed Electric Tracked Vehicle’s Handling Characteristics. Acta Armamentarii 2023, 44, 203–213. [Google Scholar]
  5. Hou, X.; Ma Yue Xiang, C. Research on Steering Stability Control of Electric Drive Tracked Vehicle. J. Mech. Eng. 2024, 60, 233–244. [Google Scholar]
  6. Zhang, J.; Wang, H.; Zheng, J.; Cao, Z.; Man, Z.; Yu, M.; Chen, L. Adaptive sliding mode-based lateral stability control of steer-by-wire vehicles with experimental validations. IEEE Trans. Veh. Technol. 2020, 69, 9589–9600. [Google Scholar] [CrossRef]
  7. Sun, W.; Wang, S. A Review of the Technical Content of Autonomous Vehicle. Int. J. Syst. Eng. 2018, 2, 42–46. [Google Scholar]
  8. AbdElmoniem, A.; Osama, A.; Abdelaziz, M.; Maged, S.A. Accurate path tracking by adjusting look-ahead point in pure pursuit method. Int. J. Automot. Technol. 2021, 22, 119–129. [Google Scholar]
  9. AbdElmoniem, A.; Osama, A.; Abdelaziz, M.; Maged, S.A. A path-tracking algorithm using predictive Stanley lateral controller. Int. J. Adv. Robot. Syst. 2020, 17, 1729881420974852. [Google Scholar] [CrossRef]
  10. Farag, W. Complex trajectory tracking using PID control for autonomous driving. Int. J. Intell. Transp. Syst. Res. 2020, 18, 356–366. [Google Scholar] [CrossRef]
  11. Xu, L.; Du, J.; Song, B.; Cao, M. A combined backstepping and fractional-order PID controller to trajectory tracking of mobile robots. Syst. Sci. Control Eng. 2022, 10, 134–141. [Google Scholar] [CrossRef]
  12. Zhao, Z.; Liu, H.; Chen, H.; Hu, J.; Guo, H. Kinematics-aware model predictive control for autonomous high-speed tracked vehicles under the off-road conditions. Mech. Syst. Signal Process. 2019, 123, 333–350. [Google Scholar] [CrossRef]
  13. Srikonda, S.; Norris, W.R.; Nottage, D.; Soylemezoglu, A. Deep Reinforcement Learning for Autonomous Dynamic Skid Steer Vehicle Trajectory Tracking. Robotics 2022, 11, 95. [Google Scholar] [CrossRef]
  14. Liu, M.; Zhao, F.; Yin, J.; Niu, J.; Liu, Y. Reinforcement-tracking: An effective trajectory tracking and navigation method for autonomous urban driving. IEEE Trans. Intell. Transp. Syst. 2021, 23, 6991–7007. [Google Scholar] [CrossRef]
  15. Shan, Y.; Zheng, B.; Chen, L.; Chen, L.; Chen, D. A reinforcement learning-based adaptive path tracking approach for autonomous driving. IEEE Trans. Veh. Technol. 2020, 69, 10581–10595. [Google Scholar] [CrossRef]
  16. Wang, S.; Yin, X.; Li, P.; Zhang, M.; Wang, X. Trajectory tracking control for mobile robots using reinforcement learning and PID. Iran. J. Sci. Technol. Trans. Electr. Eng. 2020, 44, 1059–1068. [Google Scholar] [CrossRef]
  17. Chen, I.M.; Chan, C.Y. Deep reinforcement learning based path tracking controller for autonomous vehicle. Proc. Inst. Mech. Eng. Part D J. Automob. Eng. 2021, 235, 541–551. [Google Scholar] [CrossRef]
  18. Sabiha, A.D.; Kamel, M.A.; Said, E.; Hussein, W.M. ROS-based trajectory tracking control for autonomous tracked vehicle using optimized backstepping and sliding mode control. Robot. Auton. Syst. 2022, 152, 104058. [Google Scholar] [CrossRef]
  19. Ruslan, N.A.I.; Amer, N.H.; Hudha, K.; Kadir, Z.A.; Ishak SA, F.M.; Dardin, S.M.F.S. Modelling and control strategies in path tracking control for autonomous tracked vehicles: A review of state of the art and challenges. J. Terramech. 2023, 105, 67–79. [Google Scholar] [CrossRef]
  20. Al-Jarrah, A.; Salah, M. Trajectory tracking control of tracked vehicles considering nonlinearities due to slipping while skid-steering. Syst. Sci. Control Eng. 2022, 10, 887–898. [Google Scholar] [CrossRef]
Figure 1. Kinematic model of tracked vehicle.
Figure 1. Kinematic model of tracked vehicle.
Electronics 13 03747 g001
Figure 2. Dynamic model of the tracked vehicle.
Figure 2. Dynamic model of the tracked vehicle.
Electronics 13 03747 g002
Figure 3. The dynamics simulation model of the tracked vehicle.
Figure 3. The dynamics simulation model of the tracked vehicle.
Electronics 13 03747 g003
Figure 4. Vehicle trajectory comparison diagram.
Figure 4. Vehicle trajectory comparison diagram.
Electronics 13 03747 g004
Figure 5. Vehicle speed comparison diagram.
Figure 5. Vehicle speed comparison diagram.
Electronics 13 03747 g005
Figure 6. Schematic diagram of the MPC-TD3 control algorithm.
Figure 6. Schematic diagram of the MPC-TD3 control algorithm.
Electronics 13 03747 g006
Figure 7. Simulation Experiment Data Interaction Diagram.
Figure 7. Simulation Experiment Data Interaction Diagram.
Electronics 13 03747 g007
Figure 8. Simulation diagrams of straight-line and circular trajectory conditions.
Figure 8. Simulation diagrams of straight-line and circular trajectory conditions.
Electronics 13 03747 g008
Figure 9. Simulation diagrams of double-lane-change trajectory conditions.
Figure 9. Simulation diagrams of double-lane-change trajectory conditions.
Electronics 13 03747 g009
Figure 10. Hardware-in-the-Loop Experimental System.
Figure 10. Hardware-in-the-Loop Experimental System.
Electronics 13 03747 g010
Figure 11. Hardware-in-the-Loop simulation diagrams of straight-line and circular trajectory conditions.
Figure 11. Hardware-in-the-Loop simulation diagrams of straight-line and circular trajectory conditions.
Electronics 13 03747 g011
Figure 12. Hardware-in-the-Loop simulation diagrams of double-lane-change trajectory conditions.
Figure 12. Hardware-in-the-Loop simulation diagrams of double-lane-change trajectory conditions.
Electronics 13 03747 g012
Table 1. Training parameters.
Table 1. Training parameters.
HyperparameterValue
The number of layers of the Actor network2
The number of neurons in each layer of the Actor 256
The learning rate of the Critic network0.001
The learning rate of the Actor network0.0001
discount factor0.99
The size of the experience replay buffer128
The update interval of the target network2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Gai, J.; He, S.; Li, H.; Cheng, C.; Zou, W. MPC-TD3 Trajectory Tracking Control for Electrically Driven Unmanned Tracked Vehicles. Electronics 2024, 13, 3747. https://doi.org/10.3390/electronics13183747

AMA Style

Chen Y, Gai J, He S, Li H, Cheng C, Zou W. MPC-TD3 Trajectory Tracking Control for Electrically Driven Unmanned Tracked Vehicles. Electronics. 2024; 13(18):3747. https://doi.org/10.3390/electronics13183747

Chicago/Turabian Style

Chen, Yuxuan, Jiangtao Gai, Shuai He, Huanhuan Li, Cheng Cheng, and Wujun Zou. 2024. "MPC-TD3 Trajectory Tracking Control for Electrically Driven Unmanned Tracked Vehicles" Electronics 13, no. 18: 3747. https://doi.org/10.3390/electronics13183747

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop