Next Article in Journal
A Hamilton–Jacobi Reachability-Based Minimum Separation Estimation of Integrated Manned and Unmanned Operation in Uncertain Environments
Previous Article in Journal
Lightweight and Efficient Tiny-Object Detection Based on Improved YOLOv8n for UAV Aerial Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

Penetration Strategy for High-Speed Unmanned Aerial Vehicles: A Memory-Based Deep Reinforcement Learning Approach

1
Unmanned System Research Institute, Northwestern Polytechnical University, Xi’an 710072, China
2
National Key Laboratory of Unmanned Aerial Vehicle Technology, Northwestern Polytechnical University, Xi’an 710072, China
3
Integrated Research and Development Platform of Unmanned Aerial Vehicle Technology, Xi’an 710072, China
4
Research Center for Unmanned System Strategy Development, Northwestern Polytechnical University, Xi’an 710072, China
5
Beijing Institute of Tracking and Telecommunication Technology, Beijing 100094, China
6
Science and Technology on Complex System Control and Intelligent Agent Cooperative Laboratory, Beijing 100074, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(7), 275; https://doi.org/10.3390/drones8070275
Submission received: 25 March 2024 / Revised: 6 June 2024 / Accepted: 8 June 2024 / Published: 21 June 2024

Abstract

With the development and strengthening of interception measures, the traditional penetration methods of high-speed unmanned aerial vehicles (UAVs) are no longer able to meet the penetration requirements in diversified and complex combat scenarios. Due to the advancement of Artificial Intelligence technology in recent years, intelligent penetration methods have gradually become promising solutions. In this paper, a penetration strategy for high-speed UAVs based on improved Deep Reinforcement Learning (DRL) is proposed, in which Long Short-Term Memory (LSTM) networks are incorporated into a classical Soft Actor–Critic (SAC) algorithm. A three-dimensional (3D) planar engagement scenario of a high-speed UAV facing two interceptors with strong maneuverability is constructed. According to the proposed LSTM-SAC approach, the reward function is designed based on the criteria for successful penetration, taking into account energy and flight range constraints. Then, an intelligent penetration strategy is obtained by extensive training, which utilizes the motion states of both sides to make decisions and generate the penetration overload commands for the high-speed UAV. The simulation results show that compared with the classical SAC algorithm, the proposed algorithm has a training efficiency improvement of 75.56% training episode reduction. Meanwhile, the LSTM-SAC approach achieves a successful penetration rate of more than 90% in hypothetical complex scenarios, with a 40% average increase compared with the conventional programmed penetration methods.
Keywords: memory-based DRL; LSTM-SAC; penetration method; high-speed UAV; intelligent decision making memory-based DRL; LSTM-SAC; penetration method; high-speed UAV; intelligent decision making

Share and Cite

MDPI and ACS Style

Zhang, X.; Guo, H.; Yan, T.; Wang, X.; Sun, W.; Fu, W.; Yan, J. Penetration Strategy for High-Speed Unmanned Aerial Vehicles: A Memory-Based Deep Reinforcement Learning Approach. Drones 2024, 8, 275. https://doi.org/10.3390/drones8070275

AMA Style

Zhang X, Guo H, Yan T, Wang X, Sun W, Fu W, Yan J. Penetration Strategy for High-Speed Unmanned Aerial Vehicles: A Memory-Based Deep Reinforcement Learning Approach. Drones. 2024; 8(7):275. https://doi.org/10.3390/drones8070275

Chicago/Turabian Style

Zhang, Xiaojie, Hang Guo, Tian Yan, Xiaoming Wang, Wendi Sun, Wenxing Fu, and Jie Yan. 2024. "Penetration Strategy for High-Speed Unmanned Aerial Vehicles: A Memory-Based Deep Reinforcement Learning Approach" Drones 8, no. 7: 275. https://doi.org/10.3390/drones8070275

Article Metrics

Back to TopTop