Next Article in Journal
Potential and Challenges in Airborne Automated External Defibrillator Delivery by Drones in a Mountainous Region
Previous Article in Journal
AMFEF-DETR: An End-to-End Adaptive Multi-Scale Feature Extraction and Fusion Object Detection Network Based on UAV Aerial Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Interception of a Single Intruding Unmanned Aerial Vehicle by Multiple Missiles Using the Novel EA-MADDPG Training Algorithm

1
School of Automation Science and Engineering, South China University of Technology, Guangzhou 510641, China
2
Key Laboratory of Autonomous Systems and Networked Control, Ministry of Education, Guangzhou 510641, China
3
Guangdong Engineering Technology Research Center of Unmanned Aerial Vehicle Systems, Guangzhou 510641, China
*
Author to whom correspondence should be addressed.
Drones 2024, 8(10), 524; https://doi.org/10.3390/drones8100524
Submission received: 7 August 2024 / Revised: 24 September 2024 / Accepted: 24 September 2024 / Published: 26 September 2024

Abstract

This paper proposes an improved multi-agent deep deterministic policy gradient algorithm called the equal-reward and action-enhanced multi-agent deep deterministic policy gradient (EA-MADDPG) algorithm to solve the guidance problem of multiple missiles cooperating to intercept a single intruding UAV in three-dimensional space. The key innovations of EA-MADDPG include the implementation of the action filter with additional reward functions, optimal replay buffer, and equal reward setting. The additional reward functions and the action filter are set to enhance the exploration performance of the missiles during training. The optimal replay buffer and the equal reward setting are implemented to improve the utilization efficiency of exploration experiences obtained through the action filter. In order to prevent over-learning from certain experiences, a special storage mechanism is established, where experiences obtained through the action filter are stored only in the optimal replay buffer, while normal experiences are stored in both the optimal replay buffer and normal replay buffer. Meanwhile, we gradually reduce the selection probability of the action filter and the sampling ratio of the optimal replay buffer. Finally, comparative experiments show that the algorithm enhances the agents’ exploration capabilities, allowing them to learn policies more quickly and stably, which enables multiple missiles to complete the interception task more rapidly and with a higher success rate.
Keywords: MADDPG; optimal replay buffer; equal reward setting; action filter MADDPG; optimal replay buffer; equal reward setting; action filter

Correction Statement

This article has been republished with a minor correction to the DURC Statement. This change does not affect the scientific content of the article.

Share and Cite

MDPI and ACS Style

Cai, H.; Li, X.; Zhang, Y.; Gao, H. Interception of a Single Intruding Unmanned Aerial Vehicle by Multiple Missiles Using the Novel EA-MADDPG Training Algorithm. Drones 2024, 8, 524. https://doi.org/10.3390/drones8100524

AMA Style

Cai H, Li X, Zhang Y, Gao H. Interception of a Single Intruding Unmanned Aerial Vehicle by Multiple Missiles Using the Novel EA-MADDPG Training Algorithm. Drones. 2024; 8(10):524. https://doi.org/10.3390/drones8100524

Chicago/Turabian Style

Cai, He, Xingsheng Li, Yibo Zhang, and Huanli Gao. 2024. "Interception of a Single Intruding Unmanned Aerial Vehicle by Multiple Missiles Using the Novel EA-MADDPG Training Algorithm" Drones 8, no. 10: 524. https://doi.org/10.3390/drones8100524

APA Style

Cai, H., Li, X., Zhang, Y., & Gao, H. (2024). Interception of a Single Intruding Unmanned Aerial Vehicle by Multiple Missiles Using the Novel EA-MADDPG Training Algorithm. Drones, 8(10), 524. https://doi.org/10.3390/drones8100524

Article Metrics

Back to TopTop