Next Article in Journal
CO2 Emissions Associated with Groundwater Storage Depletion in South Korea: Estimation and Vulnerability Assessment Using Satellite Data and Data-Driven Models
Previous Article in Journal
TCPSNet: Transformer and Cross-Pseudo-Siamese Learning Network for Classification of Multi-Source Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
This is an early access version, the complete PDF, HTML, and XML versions will be available soon.
Article

A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition

by
Hehao Liu
1,
Dong Li
1,*,
Ming Zhang
2,
Jun Wan
1,
Shuang Liu
1,
Hanying Zhu
1 and
Qinghua Liu
3
1
School of Microelectronics and Communication Engineering, Chongqing University, Chongqing 400044, China
2
South-West Institute of Electronics and Telecommunication Technology, Chengdu 610041, China
3
Guangxi Key Laboratory of Wireless Wideband Communication and Signal Processing, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(17), 3121; https://doi.org/10.3390/rs16173121 (registering DOI)
Submission received: 2 July 2024 / Revised: 18 August 2024 / Accepted: 21 August 2024 / Published: 23 August 2024

Abstract

With the continuous progress in drone and materials technology, numerous bionic drones have been developed and employed in various fields. These bionic drones are designed to mimic the shape of birds, seamlessly blending into the natural environment and reducing the likelihood of detection. However, such a high degree of similarity also poses significant challenges in accurately distinguishing between real birds and bionic drones. Existing methods attempt to recognize both using optical images, but the visual similarity often results in poor recognition accuracy. To alleviate this problem, in this paper, we propose a cross-modal semantic alignment and feature fusion (CSAFF) network to improve the recognition accuracy of bionic drones. CSAFF aims to introduce motion behavior information as an auxiliary cue to improve discriminability. Specifically, a semantic alignment module (SAM) was designed to explore the consistent semantic information between cross-modal data and provide more semantic cues for the recognition of bionic drones and birds. Then, a feature fusion module (FFM) was developed to fully integrate cross-modal information, which effectively enhances the representability of these features. Extensive experiments were performed on datasets containing bionic drones and birds, and the experimental results consistently show the effectiveness of the proposed CSAFF method in identifying bionic drones and bionic birds.
Keywords: cross-modal recognition; semantic alignment; feature fusion; bionic drones cross-modal recognition; semantic alignment; feature fusion; bionic drones

Share and Cite

MDPI and ACS Style

Liu, H.; Li, D.; Zhang, M.; Wan, J.; Liu, S.; Zhu, H.; Liu, Q. A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition. Remote Sens. 2024, 16, 3121. https://doi.org/10.3390/rs16173121

AMA Style

Liu H, Li D, Zhang M, Wan J, Liu S, Zhu H, Liu Q. A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition. Remote Sensing. 2024; 16(17):3121. https://doi.org/10.3390/rs16173121

Chicago/Turabian Style

Liu, Hehao, Dong Li, Ming Zhang, Jun Wan, Shuang Liu, Hanying Zhu, and Qinghua Liu. 2024. "A Cross-Modal Semantic Alignment and Feature Fusion Method for Bionic Drone and Bird Recognition" Remote Sensing 16, no. 17: 3121. https://doi.org/10.3390/rs16173121

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop