Next Article in Journal
An Improved Large Planar Point Cloud Registration Algorithm
Previous Article in Journal
Improved Command Filtered Backstepping Control for Uncertain Nonlinear Systems with Time-Delay
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Infrared Aircraft Detection Algorithm Based on Context Perception Feature Enhancement

1
College of Information Engineering, Henan University of Science and Technology, Luoyang 471023, China
2
School of Electrical, Computer and Telecommunications Engineering, University of Wollongong, Wollongong, NSW 2522, Australia
*
Author to whom correspondence should be addressed.
Electronics 2024, 13(14), 2695; https://doi.org/10.3390/electronics13142695
Submission received: 2 June 2024 / Revised: 4 July 2024 / Accepted: 6 July 2024 / Published: 10 July 2024
(This article belongs to the Section Artificial Intelligence)

Abstract

:
To address the issue of insufficient extraction of target features and the resulting impact on detection performance in long-range infrared aircraft target detection caused by small imaging area and weak radiation intensity starting from the idea of perceiving target context to enhance the features extracted by convolutional neural network, this paper proposes a detecting algorithm based on AWFGLC (adaptive weighted fusion of global–local context). Based on the mechanism of AWFGLC, the input feature map is randomly reorganized and partitioned along the channel dimension, resulting in two feature maps. One feature map is utilized by self-attention for global context modeling, establishing the correlation between target features and background features to highlight the salient features of the target, thereby enabling the detecting algorithm to better perceive the global features of the target. The other feature map is subjected to window partitioning, with max pooling and average pooling performed within each window to highlight the local features of the target. Subsequently, self-attention is applied to the pooled feature map for local context modeling, establishing the correlation between the target and its surrounding neighborhood, further enhancing the weaker parts of the target features, and enabling the detecting algorithm to better perceive the local features of the target. Based on the characteristics of the target, an adaptive weighted fusion strategy with learnable parameters is employed to aggregate the global context and local context feature maps. This results in a feature map containing more complete target information, enhancing the ability of the detection algorithm to distinguish between target and background. Finally, this paper integrates the mechanism of AWFGLC into YOLOv7 for the detection of infrared aircraft targets. The experiments indicate that the proposed algorithm achieves mAP50 scores of 97.8% and 88.7% on self-made and publicly available infrared aircraft datasets, respectively. Moreover, the mAP50:95 scores reach 65.7% and 61.2%, respectively. These results outperform those of classical target detection algorithms, indicating the effective realization of infrared aircraft target detection.

1. Introduction

ATR (automatic target recognition) is one of the key technologies in infrared imaging guidance systems. Generally, it is necessary to detect, identify, and track targets such as aircraft at long distances. In such situations, the target has a small area and weak radiation characteristics on the imaging plane, which makes the existing detection algorithms unable to fully extract the features of infrared target such as aircraft, resulting in poor detection performance. Currently, infrared aircraft target detection mainly relies on traditional methods that accurately model target features. However, precise modeling depends on predefined rules and prior knowledge. Once the operational environment exceeds the preset rules, traditional methods become ineffective. Deep learning network simulates the neural connectivity structure of the human brain, possessing powerful capabilities for autonomous learning and representation of target features. This approach is an effective means of addressing the problem of infrared aircraft target detection. Target detection methods based on deep learning are mainly divided into region-based two-stage detection algorithms, such as the R-CNN series and regression-based single-stage detection algorithms such as the YOLO series, SSD series, etc. Compared to two-stage algorithms, single-stage algorithms offer faster detection speeds, meeting the real-time requirements of detection [1].
The YOLO series algorithms directly regress the position coordinates and categories of targets, enabling high detection accuracy and fast detection speed with this end-to-end detection approach [2,3,4,5,6,7,8]. Compared to other algorithms in the YOLO series, YOLOv5 and YOLOv7 strike a good balance between detection efficiency and accuracy. Therefore, many researchers apply YOLOv5 and YOLOv7 to address practical problems. In terms of practical system applications, References [9] and [10] respectively design person detection algorithms for UAV (unmanned aerial vehicle) monitoring systems based on YOLOv5 and YOLOv7. Reference [11] combines active thermography with YOLOv5 to achieve surface defect detection. In terms of improving detection algorithms, scholars from various countries have conducted extensive research in various aspects such as expanding receptive fields [12], dense connection structure [13], attention mechanism [14,15,16], feature fusion [17,18,19,20,21], lightweight model [22,23], training strategy and loss [17,24,25], activation function [26], and so on. In reference [12], super-resolution reconstructed images are used as input for the detection algorithm. It integrates a multi-level receptive field structure into the feature fusion network of YOLOv5, enabling the detection algorithm to preserve the original information of small infrared targets. In Reference [13], dense connections are used to add the outputs of each layer of YOLOv5 together, better integrating the positional information of small infrared targets in both shallow and deep networks, thus improving the localization ability of YOLOv5. In reference [14], YOLOv5 serves as the baseline model, and standard rotation-equivariant convolution module and rotation residual module are designed. Through the SE (squeeze-and-excitation) attention mechanism, the importance of each channel is adaptively learned and applied to infrared target recognition for UAV. In Reference [15], SE module and inception structures are embedded in the feature extraction network of YOLOv5 to enhance the ability to extract target features in complex backgrounds. In Reference [16], a multi-scale feature extraction network based on ConvNext is constructed on top of YOLOv5. Then, a coordinate attention module is introduced in the ConvNeXt block to focus on targets and suppress the background. Additionally, a segmentation attention module is introduced in the neck to enhance feature fusion ability. Reference [17] constructs a feature pyramid fusion network that considers adaptive attention and feature enhancement and proposes a loss function based on angle regression to accelerate the convergence speed of model training. After fusion with YOLOv7, it achieves the detection of multi-scale targets in UAV images. In Reference [18], a weighted feature fusion pyramid network based on dilated convolution is proposed. Meanwhile, rotation detection is introduced to obtain more accurate detection boxes and ship navigation direction information. These are integrated into YOLOv7 to achieve the detection of infrared ship targets. In Reference [19], a dual-branch fusion detection network based on YOLOv5 is proposed, which can simultaneously input infrared and visible light images for object detection. In reference [20], a four-channel detection model is constructed using YOLOv5, integrating features from both visible light and infrared images to enhance the detection capabilities for person and vehicle by UAV in forest environments. Reference [21] uses YOLOv5 as baseline and utilizes image features from different modal such as infrared and visible light by constructing a layered residual connection. Based on this, a module of multi-level feature fusion is designed to achieve small object detection in multispectral images. In Reference [22], the MobileViT network is utilized as the feature extraction network for YOLOv7 to reduce the complexity of the detection algorithm and support the recognition of infrared vehicle targets by UAV. In Reference [23], starting from the lightweight YOLOv7-tiny network, improvements are made in terms of attention, receptive field and feature fusion, leading to the design of the YOLOv7-drone algorithm for recognizing small targets by drones. In Reference [24], based on the YOLOv5 framework, a novel training strategy using domain adaptation method is adopted to address the issue of imbalanced dataset distribution. Additionally, a new loss function based on the Wasserstein Distance is proposed to handle small objects by overcoming the sensitivity problem of intersection over union. In Reference [25], a hard sample mining loss function is utilized during the training of the YOLOv7 to guide the network in enhancing its learning of difficult samples, thus improving the classification capability of YOLOv7 for targets. In Reference [26], the activation function LeakyReLU in YOLOv7 is replaced with SiLU, which improves the convergence speed and generalization of the model during training. Additionally, a targeted detection layer for small objects is designed.
The above methods improve detection algorithms from the perspectives of enlarging receptive fields, dense connection structure, attention mechanism, feature fusion, lightweight design, training strategy, loss function, activation function, etc., but do not utilize the global context information of the target. However, incorporating contextual information into target detection tasks helps distinguish targets from the background, thereby enhancing the performance of detection algorithms. Reference [27] introduces a global context module to model the input feature map, enabling the detection algorithm to better distinguish between target and background. Reference [28] proposes a global context information aggregation module to enhance the global perception capability of the detection algorithm, improving its ability to perceive infrared target and background features under nighttime and occlusion conditions. Reference [29] utilizes multi-scale downsampling to obtain corresponding multi-scale features and selectively integrates spatial context information of multi-scale features, enhancing the perception capability of the model for target spatial position. Reference [30] aggregates contextual information using feature maps of different sizes, thereby enhancing the spatial correlation between the background and foreground. Reference [31] designs a position attention mechanism to capture global context information, thereby completing the task of semantic segmentation in scenes. Reference [32] proposes a lightweight global context modeling framework that is structurally similar to SENet. For the problem of infrared small target detection in complex backgrounds, Reference [33] proposes an attention-guided pyramid context network. This network captures and integrates global context information at different scales to achieve better feature representation. Reference [34] proposes a feature fusion module of a multi-scale receptive field to represent the global context information of deep feature maps, thereby expanding the receptive field of detection network for infrared small target.
References [27,28,29,30,31,32,33,34] utilize a global context mechanism to explore the relationship between target and other objects in the scene, highlighting the salient features of the target and enhancing the discriminative ability of the detection algorithm between target and background features. However, most global context mechanisms are based on modeling global information of the image without emphasizing the local information of the target, resulting in weak modeling capabilities for the target and its surrounding neighborhood.
Local context mechanisms establish the correlation between the target and its surrounding neighborhood, enhancing the weaker features of the target and allowing the detection algorithm to focus more on the local information surrounding the target, thereby enhancing the ability to extract local features of the target. Reference [35] introduces convolution-based local perception units into the Transformer structure, constructing local context to enhance the ability to represent local features of the target. On this basis, it proposes a Swin-transformer-based detector for arbitrarily oriented SAR (Synthetic Aperture Radar) ship detection. Reference [36] proposes a classification algorithm for polarization synthetic aperture radar data based on vision transformer (ViT). This algorithm uses a cellular neural network as the feature extractor and employs local window attention to model the local context of the target, thereby highlighting local features. Reference [37] proposes a multi-level vision transformer model that enhances target features by computing the context of non-overlapping local windows through a self-attention mechanism and applies it to vision detection task. Reference [38] proposes an attention-based context-aware network (ACANet), which enhances infrared small target features by clustering local features. Reference [39] proposes a deep network framework for infrared small target detection that encompasses feature enhancement, interaction, and comparison. This framework enhances local context information from different feature layers using a patch-based attention module. Inspired by the physical thermal diffusion model of infrared small target and human visual mechanism, Reference [40] designs a dense network of multi-level feature extraction and fusion. This network incorporates Gaussian saliency features and local context features into target feature representation to address the issue of sparse feature extraction.
It is essential to integrate both global and local context and prioritize them based on the characteristics of the target in practical applications. Based on the above analysis, this paper proposes a YOLOv7-based target detection algorithm for infrared aircraft using a novel adaptive weighted fusion mechanism that integrates both global and local context. The proposed mechanism can enhance the feature extraction capability of the detection algorithm for infrared aircraft target and improve the discriminative ability between infrared aircraft target and the background, thereby enhancing the performance of the detection algorithm. The main contributions of this paper are as follows:
(1)
The paper introduces a mechanism of adaptive weighted fusion of global–local context (AWFGLC) according to the characteristics of infrared aircraft target, providing more comprehensive target feature information for the detection algorithm.
(2)
The AWFGLC-YOLO (adaptive weighted fusion of global–local context—you only look once) algorithm is proposed by integrating the AWFGLC mechanism into different levels of the feature extraction network of YOLOv7. This approach effectively utilizes the physical information from shallow feature maps and the semantic information from deep feature maps. The algorithm is then applied to the detection of infrared aircraft target.
It should be further noted that the mechanism of AWFGLC proposed in this paper is universal and can be applied to other single-stage target detection algorithms. The remaining content of this paper is organized as follows: Section 2 elucidates the core idea and working principle of AWFGLC. Section 3 presents the AWFGLC-YOLO detection framework for infrared aircraft target, which integrates YOLOv7 with AWFGLC. Subsequently, Section 4 conducts experiments on two datasets and analyzes the results(Section 4 conducts experiments on two datasets and analyzes the results). Finally, Section 5 concludes the paper and provides prospects for future research.

2. Adaptive Weighted Fusion of Global-Local Context

The mechanism of AWFGLC can highlight target features, enhance the perception ability for target features, and thereby improve the discriminative ability of the detection algorithm between target and background. The proportion of global and local context in the fused feature map varies based on the characteristics of the target. When the target features are strong, the fused feature map should emphasize global context information, whereas it should focus more on local context when the target features are weaker. The proposed mechanism involves randomly reorganizing and partitioning the input feature map along the channel dimension to obtain two feature maps. One feature map is used for global context modeling. The other feature map is divided into windows and pooled. Then, local context modeling is performed within each pooled window, and the local context feature map is obtained through spatial rearrangement. The adaptive weighted strategy using learnable parameters is employed to aggregate global and local context information, enriching the feature representation and improving the feature extraction capability of the feature extraction network for infrared aircraft target. The overall process of the AWFGLC mechanism is illustrated in Figure 1, which includes five functional units: channel reorganization and division, window partitioning, global context modeling, local context modeling and weighted fusion, as depicted in Figure 2a–e respectively.
The detailed process of the AWFGLC mechanism is as follows:
(1) Assuming the input feature map is denoted as F R H × W × C . Firstly, F is randomly reorganized along the channel dimension to obtain an intermediate feature map F R H × W × C . Then, F is partitioned along the channel dimension to generate feature maps F 0 R H × W × C / 2 and F 1 R H × W × C / 2 , as illustrated in Figure 2a. Random reorganization takes the position index of each channel in feature map F, shuffles these indices, and concatenates the channels according to the shuffled indices. Different random numbers are set to randomly shuffle the input feature map along the channel indices in each epoch during iterative training. This diversifies the arrangement of input feature map. Compared to modeling the context on input feature maps based on specific permutation, modeling the global and local context on feature maps with different permutations helps the detection algorithm learn more diverse and comprehensive features.
To reduce the computational complexity of modeling the context for the input feature map, this paper divides the input feature map equally along the channel dimension and performs separately global and local context modeling. The calculation formulas for random shuffling and partitioning are described as:
F ( x 3 , x 1 , , x c , , x i ) = f r c ( F ( x 1 , x 2 , , x i , , x c ) )
F 0 , F 1 = f d 2 ( F ( x 3 , x 1 , , x c , , x i ) )
where c is the number of channels, f r c represents the channel reorganization of the input feature map, and F is obtained by changing the orders of x1, x2, x3, … xi, …xc. f d 2 is the partitioning for the intermediate feature map, and x i represents the ith channel of the input feature map. Compared to modeling global and local contexts based on raw feature maps, modeling global and local contexts based on channel partitioning reduces the computational complexity by half.
(2) Global context modeling utilizes self-attention to establish the dependency relationship between pixels in feature map F 0 in order to obtain the global context feature map F g , as illustrated in Figure 2c. MM denotes the dot product of vectors. More specifically, the dimension of F 0 is transformed into ( C / 2 ) × ( H × W ) and mapped to query vector F 0 q , key vector F 0 k , and value vector F 0 v through linear transformations. The similarity between the query vector F 0 q and the key vector F 0 k is computed using dot product and then divided by d , which is the square root of the input vector’s dimension. This ensures that the range of the dot product remains unchanged and reduces the impact of gradient changes, allowing for better backpropagation and optimization. Next, the obtained similarity scores are normalized using the Softmax function to obtain attention weights. These weights are then multiplied with the corresponding value vector F 0 v , and the resulting vector is reshaped to match the shape of vector F 0 , yielding the global context feature map F g . The entire calculation process is shown in Equations (3) and (4).
F 0 q = f L i n e a r q ( F 0 ) , F 0 k = f L i n e a r k ( F 0 ) , F 0 v = f L i n e a r v ( F 0 )
F g = S o f t max ( F 0 q F 0 k T d ) F 0 v
In the equations, f L i n e a r represents the linear transformation.
(3) To enhance the local features of the target, a window of h × w × ( C / 2 ) is slid over F 1 to divide it into i patches with dimension of h × w × ( C / 2 ) , as illustrated in Figure 2b. In each patch, max pooling and average pooling are performed to highlight the prominent and average features of the target region, respectively, resulting in the feature map p o o l max i and p o o l avg i . The pooled feature map is obtained by performing matrix addition between p o o l max i and p o o l avg i . The calculation process is illustrated by Equations (5)–(7).
p o o l max i = f max k × k × s ( p a t c h i )
p o o l a v g i = f avg k × k × s ( p a t c h i )
p o o l i = p o o l max i p o o l a v g i
In the equation, p a t c h i is the ith partitioned window, f a v g k × k × s represents average pooling performed using convolution with size k × k and stride s, f max k × k × s denotes max pooling performed using convolution with size k × k and stride s, is matrix addition, and p o o l i represents the ith pooled local feature map.
The ith pooled local feature map p o o l i is performed with self-attention modeling to adjust the correlation between pixels and obtain a t t i which is the pooled local context feature map. This can shift the focus of feature learning more towards the target and its vicinity. Reorganizing a t t i in spatial dimension yields the local context feature map F l R H × W × ( C / 2 ) , where the approach of self-attention modeling for each window is similar to F 0 , as shown in Equations (8)–(10).
p o o l i q = f L i n e a r q ( p o o l i ) , p o o l i k = f L i n e a r k ( p o o l i ) , p o o l i v = f L i n e a r v ( p o o l i )
a t t i = S o f t max ( p o o l i q p o o l i k T d ) p o o l i v
F l = f r s ( a t t 1 , a t t 2 , , a t t i )
In the equation, f r s represents the rearrangement of a t t i in the spatial dimension.
(4) With a learnable parameter-based adaptive weighting strategy, the global context feature map F g and the local context feature map F l are adaptively weighted and then channel-concatenated. Subsequently, convolution, batch normalization, and activation function are applied to obtain the feature map F g l containing both global and local contextual information. The process is illustrated in Figure 2e. Through the adaptive weighted fusion strategy, the detection algorithm can dynamically adjust the weights of global and local information in F g l based on the characteristics of the infrared aircraft target. When the target features are more prominent, greater emphasis should be placed on the global context, whereas when the target features are less distinct, more focus should be placed on the local context. This enables the detection algorithm to better perceive the surroundings in which the target is located and highlight the target features, thereby enhancing the ability to distinguish between the target and the background. Assuming that the learnable weighting parameters are denoted as α and β , the optimization process involves minimizing the loss function of the target detection algorithm, guiding the optimizer to update these parameters iteratively. The procedure of fusion is represented by Equation (11).
F g l = S I L U ( B N ( C o n v ( f c ( α F g , β F l ) ) ) )
In the equation, f c represents channel concatenation, Conv denotes a convolutional layer with kernel size and stride both equal to 1 × 1, BN stands for batch normalization, and SILU represents the activation function.
The calculation process of Formula (11) will be expanded below. Assuming the convolutional kernel in the convolutional layer is W = w 1 , w 2 , , w i , , w C , the aggregation feature map Fagg is obtained by integrating the weighted global and local context feature map, as shown in Equation (12). The aggregation method of contextual features for each channel feature map is shown in Equation (13). The BN layer normalizes the aggregation feature map to make its data distribution more stable, thereby helping the detection algorithm to converge faster, as shown in Equation (14). Meanwhile, the detection algorithm can capture and learn the nonlinear expression in the data by applying the activation function, thereby learning higher-level semantic information, as shown in Equation (15).
F a g g ( x g l 1 , x g l 2 , , x g l i , x g l C ) = = W [ α x g 1 , α x g C / 2 , β x l 1 , , β x l C / 2 ]
x g l i = w i α x g 1 + w i α x g 2 + + w i α x g C / 2 + w i β x l 1 + w i β x l 2 + + w i β x l C / 2
F a g g b n = γ F a g g μ σ 2 + ε + δ
F g l = F a g g b n 1 + e F a g g b n
In the equation, μ represents the mean of the aggregated feature map F a g g , σ is the variance of F a g g , γ and δ are the learnable parameters in the BN layer, and ε denotes the regularization parameter.

3. The Infrared Aircraft Detection Algorithm Based on Adaptive Weighted Fusion of Global–Local Context

The paper adopts YOLOv7 as the underlying framework and applies the proposed mechanism of adaptive weighted fusion of global–local context to design an infrared aircraft detection algorithm. This algorithm is named AWFGLC-YOLO (adaptive weighted fusion of global–local context—you only look once), and its overall structure is illustrated in Figure 3. The target detection algorithm of YOLOv7 mainly consists of three parts: feature extraction network, feature fusion network and detection head. The feature extraction network consists of basic convolutional module (CBS), extended efficient layer aggregation network (ELAN), and max pooling convolutional module (MP-C3). The feature fusion adopts a network structure of pathway aggregation, which integrates feature information of different scales through top-down and bottom-up approaches. The detection head utilizes the reparameterized convolution (RepConv) to predict results on feature maps, which are downsampled by 8 times, 16 times, and 32 times, respectively.
This paper integrates AWFGLC into the feature extraction network of YOLOv7. It models context information for feature maps which are downsampled by 4 times and 32 times, respectively, fully utilizing the physical information in shallow feature map and the semantic information in deep feature map. This enhances the ability of the detection algorithm to extract features of infrared aircraft target, thereby improving the discriminative ability between the target and the background.
To verify that the AWFGLC mechanism can highlight target features and improve the ability of the detection algorithm to distinguish between the target and the background, an example for the 4 times downsampled feature map of the feature extraction network is depicted in Figure 4. Figure 4a represents the original image, while Figure 4b represents the input feature map fed into the AWFGLC mechanism. Figure 4c illustrates the global context feature map, and Figure 4d,e respectively depict the window partitioning and pooled feature map as well as the local context feature map. Finally, Figure 4f illustrates the feature map after the AWFGLC mechanism. The content in the red box is an enlarged display of the target area in the image.
From Figure 4b, it can be observed that after the feature extraction and downsampling by the first four layers of the feature extraction network, the target features in the input feature map to the AWFGLC mechanism appear weak and relatively small in size. AWFGLC performs channel reorganization and division on the input feature map to obtain two separate feature maps, and then models these two feature maps with different contexts. From Figure 4c, it can be observed that the global context feature map which reflects the correlation between the target and other background pixels in the global image can highlight the brighter tail flame (yellow part) of the target, which helps to distinguish the target from the background. However, the global context feature map has a weaker ability to model the correlation between the target and the neighboring pixels, failing to highlight the darker fuselage of the aircraft. From Figure 4d,e, it can be observed that the feature map of local context based on pooling which reflects the correlation between the target features and the neighboring features can enhance the features of the aircraft fuselage, which helps the detection algorithm better perceive the local features of the target. From Figure 4f, it can be seen that the adaptive weighted fusion strategy effectively aggregates the global context feature map containing brighter tail flame information and the local context feature map with enhanced features of aircraft fuselage, thereby obtaining a feature map containing the overall contour information of the aircraft.

4. Experiment and Analysis

4.1. Experimental Environment

The hardware environment used in this paper includes a CPU with the Intel i7-8700K CPU @ 3.70 GHz, and a total of 32 GB of RAM. The GPU is NVIDIA 3090 Ti, with a memory size of 24 GB. The software environment includes Windows 10 operating system, PyTorch 1.13 deep learning framework, Python 3.7 programming language, and CUDA 11.1 for GPU acceleration.

4.2. Experiment on Self-Made Infrared Dataset

The self-made infrared aircraft dataset consists of 5831 frames, categorized into three poses: back, lateral, and backward. During the detection process, the fuselage and the tail-flame are distinguished simultaneously. Therefore, the manually annotated target categories consist of six classes: back fuselage (BAF), back tail flame (BAT), lateral fuselage (LAF), lateral tail flame (LAT), backward fuselage (BWF), and backward tail flame (BWT). The infrared aircraft dataset is randomly split, with 10% of the data reserved for the test set. The remaining data is subjected to cross-validation, with an 8:2 ratio used to partition the data into training and validation sets.

4.2.1. Experimental Parameters

During the process of model training, the Adam optimizer is employed as the stochastic gradient descent algorithm, with an initial learning rate set to 0.001. The momentum factor is set to 0.937, and the weight decay is set to 0.0005. The initial values for the learnable parameters α and β are both set to 0.5.
The curves showing the changes with the learnable α and β during the training process are depicted in Figure 5. It can be observed from the plot that the values of α and β gradually decrease with the increase of training iterations. According to the criterion of selecting the weights corresponding to the maximum value of the detection metric mAP50:95, the values of α and β are 0.168 and 0.151 respectively. Comparatively, the weight α , which acts on the global context feature map, tends to have slightly larger values than the weight β , which operates on the local context feature map. This indicates that the AWFGLC mechanism is more inclined to learn the global features of the target, especially on the self-made infrared aircraft dataset.

4.2.2. Comparison of Different Contextual Mechanisms

Using YOLOv7 as the benchmark detection algorithm, different contextual mechanisms are added to its feature extraction network at the positions of the 4-times-downsampled feature map and the 32-times-downsampled feature map, including global context (GC) based on simple NLNet [32], global context based on position attention (PA) [31], local context based on local transformer (LT) [37], and the proposed AWFGLC. The experimental results which compare the performance of these different contextual mechanisms in infrared aircraft object detection are shown in Table 1. From Table 1, it can be observed that compared to YOLOv7, the mAP50 of YOLOv7+GC and YOLOv7+LT decreases by 2.1% and 0.9%, while its mAP50:95 decreases by 1.2% and 0.6%, respectively. The mAP50 of YOLOv7+PA increases by 0.3%, with mAP50:95 increasing by 1.2%. The AWFGLC-YOLO detection algorithm, constructed by incorporating the AWFGLC mechanism into YOLOv7, achieves the highest mAP50 and mAP50:95, reaching 97.8% and 65.7%, respectively. This indicates that the AWFGLC mechanism effectively explores both global and local context information in the input feature map, highlighting target features and enhancing the ability of YOLOv7 to distinguish between the target and the background.
To visually observe how the AWFGLC-YOLO algorithm can better extract target features, this study utilizes Eigen-CAM [41] to visualize the features learned by the convolutional layers of different context mechanisms, as shown in Figure 6.
In Figure 6a, the radiance intensity of the fuselage is similar to that of the surrounding area, indicating limited usable information and partial obstruction of the fuselage by clouds. The content in the red box is the target area in the image. From Figure 6b–e—after incorporating GC, PA, and LT into the feature extraction network of YOLOv7—compared to the original YOLOv7 algorithm, some targets can be observed, but the focus is more on the background area. From Figure 6f, it can be observed that when incorporating the AWFGLC mechanism into the feature extraction network of YOLOv7, the detection algorithm pays more attention to the target and its surrounding neighborhood.

4.2.3. Comparison of Different Detection Algorithms

The experimental results comparing our algorithm with CenterNet [42], Sparse R-CNN [43], Efficientdet [44], Autoassign [45], YOLOF [46], Deformable DETR [47], YOLOv5 [3], YOLOv7 [6], YOLOv8 [8], and DDQ [48] are shown in Table 2. It can be seen from this that the mAP50 and mAP50:95 of Efficientdet and YOLOF are relatively low. Sparse R-CNN has good detection performance, with mAP50 and mAP50:95 reaching 90.8% and 59.7%, respectively. However, its model parameter count and floating-point computational complexity are relatively large, reaching 105.9M and 64.6G, respectively. The mAP50 and mAP50:95 of DDQ are 97.2% and 65.3%, respectively, but the computational complexity is relatively large. YOLOv5 demonstrates mAP50 score close to those of CenterNet, Autoassign, and Deformable DETR. However, YOLOv5 achieves mAP50:95 score 1.4% higher than the third-ranked YOLOv8, with smaller model parameters and floating-point computation, at 13.9 M and 16.4 G, respectively. Compared to YOLOv5 and YOLOv8, YOLOv7 achieves a higher mAP50 of 96.4%. However, the mAP50:95 of YOLOv7 is 2.3% lower than that of YOLOv5, and YOLOv7 has larger model parameters and computational requirements of 35.6 M and 105.4 G, respectively. Compared to other detection algorithms, AWFGLC-YOLO achieves the highest mAP50 and mAP50:95, at 97.8% and 65.7%, respectively. Due to the introduction of the AWFGLC mechanism, the parameter count and floating-point computational requirement of the proposed algorithm increase by 3.7% and 1.7%, respectively, compared to YOLOv7.
The partial visual detection results with different algorithms are shown in Figure 7. The content in the light green box at the bottom right of each image is an enlarged display of the detection result. From the first row of images in Figure 7, it can be seen that Deformable DETR, Sparse R-CNN, Autoassign, DDQ, and AWFGLC-YOLO are all capable of detecting targets. The proposed AWFGLC-YOLO algorithm detects back fuselage (BAF) and back tail flame (BAT) with the highest confidence, reaching values of 0.95 and 0.83, respectively. From the second row of images, it can be observed that Autoassign falsely detects two lateral fuselages (LAF), while other detection algorithms correctly identify the targets. However, AWFGLC-YOLO achieves the highest confidence in detection. As can be seen from the third row of images, Deformable DETR, Sparse R-CNN, Autoassign, and DDQ can correctly detect targets with an average confidence level of 0.710, 0.813, 0.848 and 0.823, respectively. The proposed algorithm in this paper also correctly detects the number, position, and category of the targets with an average confidence reaching the highest value of 0.85.
The target size is relatively large, and the contour information is obvious in this dataset. Integrating contextual information, especially global contextual information, into the object detection algorithm improves the detection accuracy of the algorithm. The experimental results confirm the effectiveness of the proposed AWFGLC mechanism.

4.3. Experiment on Infrared Small Target Dataset

The targets in this dataset are UAV on ground or sky backgrounds, with a total of 22 typical scenarios covering single targets on sky backgrounds, two targets in flight crossing on sky backgrounds, single targets on complex ground backgrounds, targets approaching from far to near, targets receding from near to far, and so on [49]. There are 16,177 frames of images and 16,944 targets in the dataset, with most of the target sizes being 3 × 3 and 5 × 5 pixels. Small targets typically carry limited feature information and lack distinctive appearance information to differentiate them from the background or similar objects. Incorporating contextual information around the target in natural scene-based target detection tasks can enrich the feature representation, aiding in distinguishing small targets and thereby enhancing the detection performance of the model. The training set, validation set, and test set of infrared small target data adopt the same partitioning strategy as the self-made infrared aircraft dataset. Meanwhile, the experimental parameters of this dataset are the same as those of the self-made infrared aircraft dataset.
The variation curves of weight α and β that can be learned during the training process are shown in Figure 8. From the graph, it can be seen that the values of learnable weights α and β gradually decrease with the increase of training iterations. The optimal weights learned are 0.065 for α and 0.102 for β . Compared to the weight α acting on the global context feature map, the weight β acting on the local context feature map has a larger value. This is because the targets are small in size and have blurred contours in the publicly available dataset, which makes the AWFGLC mechanism more inclined to learn the local features of the targets.
The detection results of different algorithms on the infrared small target dataset on ground or sky backgrounds are shown in Table 3. It can be observed that the mAP50 and mAP50:95 of the proposed algorithm reach 88.7% and 61.2%, respectively, which are higher than those of other detection algorithms.
The partial visual detection results with different algorithms are shown in Figure 9. The content within the blue box of each image represents the region where the target is located.From the first image in Figure 9a, it can be seen that the drone is located above the forest and that its radiation intensity is higher than that of the surrounding background. In the second image of Figure 9a, the drone is located above the forest, and its radiation intensity is similar to that of other background objects in the forest. In the third image of Figure 9a, the drone is located above complex ground backgrounds, such as roads and buildings, with weak target radiation intensity. From the first row of images in Figure 9b–f, it can be seen that except for the Autoassign algorithm, all other detection algorithms can detect the targets, and the confidence level of AWFGLC-YOLO is the highest, reaching 0.83. In the second row of images, Deformable DETR fails to detect the targets. DDQ detects two targets, with a confidence level of 0.79 for the true target and 0.68 for the false target. Other detection algorithms participating in the comparison are able to detect the targets, and the highest confidence level of AWFGLC-YOLO is 0.83. From the third row of images, it can be seen that all the detection algorithms participating in the comparison can detect the targets. The confidence level of AWFGLC-YOLO is 0.82, which is higher than the confidence level of other detection algorithms.
In this dataset, the target size is small and the contour information is blurry. Therefore, context information, especially local context information, is integrated into the detection algorithm to improve the ability to distinguish between the target and the background. The experimental results further confirm the role of the context fusion mechanism.

5. Conclusions

In response to the problem of insufficient extraction of target features due to small imaging area and weak radiation intensity in long-distance target detection for infrared aircraft, which affects detection performance, this paper proposes an infrared aircraft detection algorithm based on adaptive weighted fusion of global–local context (AWFGLC) mechanism. Starting from the idea of perceiving target context to enhance the features extracted by convolutional neural network, this mechanism applies self-attention to model both the global and local contexts of aircraft in infrared image separately. Then, it utilizes an adaptive weighting strategy to fuse the global and local context information based on the characteristics of the target, enhancing the feature representation of infrared aircraft target and improving the ability of the detection algorithm to distinguish between the target and the background. When the target features are more prominent, the AWFGLC mechanism tends to learn the global features of the target, and the fusion result focuses more on the global context. When the target features are weaker, the mechanism tends to learn the local features of the target, and the fusion result focuses more on the local context. Theoretical analysis and experimental results demonstrate that the proposed AWFGLC-YOLO algorithm is effective in detecting infrared aircraft targets. It should be noted that the AWFGLC mechanism proposed in this paper operates on feature maps extracted by convolutional neural networks, making it versatile for application in other single-stage target detection algorithms such as SSD (Single Shot MultiBox Detector).
The deep convolutional neural network has excellent performance, but its high complexity and non-linearity result in low transparency and poor interpretability. Using knowledge to guide the inference process of deep model to make the target detection process interpretable is a future research direction. Therefore, future research efforts could consider modeling knowledge about infrared aircraft radiation, motion, and other factors and integrate them into deep-learning-based target detection algorithms to further improve their performance and interpretability.

6. Patents

The research results of this article have been applied for a national invention patent in China (No. 202410588995.1).

Author Contributions

Conceptualization, G.L.; methodology, G.L., J.X., and J.T.; writing—original draft, G.L.; writing—review and editing, J.X. and J.T.; funding acquisition, J.X.; software, H.X.; validation, H.X. All authors have read and agreed to the published version of the manuscript.

Funding

This work is supported by the China Scholarship Council (No. [2022]20) and the Key Scientific Research Project of Higher Education Institutions in Henan Province (No. 21A520012).

Data Availability Statement

Data in this article can be downloaded at http://www.dx.doi.org/10.11922/sciencedb.902 (accessed on 17 August 2020).

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Cheng, X.; Song, C.; Shi, J.; Zhou, L.; Zhang, Y.; Zheng, Y. A review of general object detection based on deep learning. Acta Electron. Sin. 2021, 49, 1428–1438. [Google Scholar]
  2. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  3. Jocher, G.; Stoken, A.; Borovec, J.; Changyu, L.; Hogan, A.; Diaconu, L.; Ingham, F.; Poznanski, J.; Fang, J.; Yu, L.; et al. Ultralytics/yolov5: v3.1-Bug Fixes and Performance Improvements. 2020. Available online: https://github.com/ultralytics/yolov5 (accessed on 25 June 2020).
  4. Wang, C.Y.; Bochkovskiy, A.; Liao, H.Y.M. Scaled-YOLOv4: Scaling cross stage partial network. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 13029–13038. [Google Scholar]
  5. Ge, Z.; Liu, S.; Wang, F.; Li, Z.; Sun, J. YOLOx: Exceeding yolo series in 2021. arXiv 2021, arXiv:2107.08430. [Google Scholar]
  6. Wang, C.-Y.; Bochkovskiy, A.; Liao, H.-Y.M. YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors. In Proceedings of the 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 7464–7475. [Google Scholar]
  7. Wang, C.Y.; Yeh, I.H.; Liao, H.Y.M. You only learn one representation: Unified network for multiple tasks. J. Inf. Sci. Eng. 2023, 39, 691–709. [Google Scholar]
  8. Jocher, G.; Chaurasia, A.; Qiu, J. Ultralytics YOLO (Version 8.0.0). 2023. Available online: https://github.com/ultralytics/ultralytics (accessed on 10 January 2023).
  9. Mantau, A.J.; Widayat, I.W.; Leu, J.-S.; Köppen, M. A human-detection method based on YOLOv5 and transfer learning using thermal image data from UAV perspective for surveillance system. Drones 2022, 6, 290. [Google Scholar] [CrossRef]
  10. Guettala, W.; Sayah, A.; Kahloul, L.; Tibermacine, A. Real time human detection by unmanned aerial vehicles. In Proceedings of the 2022 International Symposium on iNnovative Informatics of Biskra (ISNIB), Biskra, Algeria, 7–8 December 2022; pp. 165–170. [Google Scholar]
  11. Lema, D.G.; Pedrayes, O.D.; Usamentiaga, R.; Venegas, P.; García, D.F. Automated detection of subsurface defects using active ther-mography and deep learning object detectors. IEEE Trans. Instrum. Meas. 2022, 71, 4503213. [Google Scholar] [CrossRef]
  12. Zhou, X.; Jiang, L.; Hu, C.; Lei, S.; Zhang, T.; Mou, X. YOLO-SASE: An improved YOLO algorithm for the small targets detection in complex backgrounds. Sensors 2022, 22, 4600. [Google Scholar] [CrossRef] [PubMed]
  13. Wang, K.; Li, S.; Niu, S.; Zhang, K. Detection of infrared small targets using feature fusion convolutional network. IEEE Access 2019, 7, 146081–146092. [Google Scholar] [CrossRef]
  14. Xiao, F.; Lu, H.; Zhang, W.; Huang, S.; Jiao, Y.; Lu, Z.; Li, Z. Aerial infrared image target recognition algorithm based on rotation equivariant convolution. Acta Armamentarii 2023, 1–9. Available online: https://link.cnki.net/urlid/11.2176.TJ.20231018.1031.004 (accessed on 18 October 2023).
  15. Zhou, W.; Wu, Z.; Zhang, Z.; Peng, L.; Xie, L. Lightweight small target detection method based on weak feature enhancement. Control. Decis. 2024, 39, 381–390. [Google Scholar]
  16. Zhou, J.; Zhang, B.; Yuan, X.; Lian, C.; Ji, L.; Zhang, Q.; Yue, J. YOLO-CIR:The network based on YOLO and ConvNeXt for infrared object detection. Infrared Phys. Technol. 2023, 131, 104703. [Google Scholar] [CrossRef]
  17. Zhang, H.; Shao, F.; He, X.; Chu, W.; Zhao, D.; Zhang, Z.; Bi, S. ATS-YOLOv7: A real -time multi-scale object detection method for UAV aerial images based on improved YOLOv7. Electronics 2023, 12, 4886. [Google Scholar] [CrossRef]
  18. Deng, H.; Zhang, Y. FMR-YOLO: Infrared ship rotating target detection based on synthetic fog and multiscale weighted feature fusion. IEEE Trans. Instrum. Meas. 2023, 73, 5001717. [Google Scholar] [CrossRef]
  19. Hou, Z.; Yang, C.; Sun, Y.; Ma, S.; Yang, X.; Fan, J. An object detection algorithm based on infrared-visible dual modal feature fusion. Infrared Phys. Technol. 2024, 137, 105107. [Google Scholar] [CrossRef]
  20. Marques, T.; Carreira, S.; Miragaia, R.; Ramos, J.; Pereira, A. Applying deep learning to real-time UAV-based forest monitoring: Leveraging multi-sensor imagery for improved results. Expert Syst. Appl. 2024, 245, 123107. [Google Scholar] [CrossRef]
  21. Sun, J.; Yin, M.; Wang, Z.; Xie, T.; Bei, S. Multispectral object detection based on multilevel feature fusion and dual feature modulation. Electronics 2024, 13, 443. [Google Scholar] [CrossRef]
  22. Zhao, X.; Xia, Y.; Zhang, W.; Zheng, C.; Zhang, Z. YOLO-ViT-based method for unmanned aerial vehicle infrared vehicle target detection. Remote Sens. 2023, 15, 3778. [Google Scholar] [CrossRef]
  23. Xue, S.; An, H.; Lv, Q.; Cao, G. Image object detection algorithm based on YOLOv7-tiny in complex background. Infrared Laser Eng. 2024, 53, 20230472-1–20230472-12. [Google Scholar]
  24. Kim, J.; Huh, J.; Park, I.; Bak, J.; Kim, D.; Lee, S. Small object detection in infrared images: Learning from imbalanced cross-domain data via domain adaptation. Appl. Sci. 2022, 12, 11201. [Google Scholar] [CrossRef]
  25. Hu, S.; Zhao, F.; Lu, H.; Deng, Y.; Du, J.; Shen, X. Improving YOLOv7-tiny for infrared and visible light image object detection on drones. Remote. Sens. 2023, 15, 3214. [Google Scholar] [CrossRef]
  26. Zhang, G.; Li, C.; Li, G.; Lu, W. Small target detection algorithm for UAV aerial images based on improved YOLOv7-tiny. Adv. Eng. Sci. 2023, 1–14. [Google Scholar] [CrossRef]
  27. Tan, J.; Yin, W.; Liu, L.; Wang, Y. DenseNet-siamese network with global context feature module for object tracking. J. Electron. Inf. Technol. 2021, 43, 179–186. [Google Scholar]
  28. Hou, Z.; Sun, Y.; Guo, H.; Li, J.; Ma, S.; Fan, J. M-YOLO: An object detector based on global context information for infrared images. J. Real-Time Image Process. 2022, 19, 1009–1022. [Google Scholar] [CrossRef]
  29. Wang, Z.; Xu, Z.; Xue, Y.; Lang, C.; Li, Z.; Wei, L. Global and spatial multi-scale contexts fusion for vehicle re-identification. J. Image Graph. 2023, 28, 471–482. [Google Scholar]
  30. Zhang, W.; Fu, C.; Xie, H.; Zhu, M.; Tie, M.; Chen, J. Global context aware RCNN for object detection. Neural Comput. Appl. 2021, 33, 11627–11639. [Google Scholar] [CrossRef]
  31. Fu, J.; Liu, J.; Tian, H.; Li, Y.; Bao, Y.; Fang, Z.; Lu, H. Dual attention network for scene segmentation. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 3141–3149. [Google Scholar]
  32. Cao, Y.; Xu, J.; Lin, S.; Wei, F.; Hu, H. GCNet: Non-local networks meet squeeze-excitation networks and beyond. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 1971–1980. [Google Scholar]
  33. Zhang, T.; Li, L.; Cao, S.; Pu, T.; Peng, Z. Attention-guided pyramid context networks for detecting infrared small target under complex background. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 4250–4261. [Google Scholar] [CrossRef]
  34. Zhang, Y.; Li, C.; Liu, Y.; Liu, Z.; Yang, R. Multi-scale feature fusion attention network for infrared small target detection. In Fourteenth International Conference on Graphics and Image Processing; SPIE: Bellingham, WA, USA, 2022; Volume 12705, pp. 34–43. [Google Scholar]
  35. Yang, Z.; Xia, X.; Liu, Y.; Wen, G.; Zhang, W.E.; Guo, L. LPST-Det: Local-perception-enhanced swin transformer for SAR ship detection. Remote Sens. 2024, 16, 483. [Google Scholar] [CrossRef]
  36. Jamali, A.; Roy, S.K.; Bhattacharya, A.; Ghamisi, P. Local window attention transformer for polarimetric SAR image classification. IEEE Geosci. Remote Sens. Lett. 2023, 20, 2735–2751. [Google Scholar] [CrossRef]
  37. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, QC, Canada, 10–17 October 2021; pp. 10012–10022. [Google Scholar]
  38. Ling, S.; Chen, L.; Wu, Y.; Zhang, Y.; Gao, Z. ACANet: Attention-based context-aware network for infrared small target detection. J. Supercomput. 2024, 4, 1–29. [Google Scholar] [CrossRef]
  39. Lv, G.; Dong, L.; Xu, W. Hierarchical interactive multi-granularity co-attention embedding to improve the small infrared target detection. Appl. Intell. 2023, 53, 27998–28020. [Google Scholar] [CrossRef]
  40. Ma, T.; Yang, Z.; Fan, Y.; Song, Y.-F.; Liang, J.; Wang, H. DMEF-Net: Lightweight infrared dim small target detection network for limited samples. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5626015. [Google Scholar] [CrossRef]
  41. Muhammad, M.B.; Yeasin, M. Eigen-cam: Class activation map using principal components. In Proceedings of the 2020 International Joint Conference on Neural Networks, Glasgow, UK, 19–24 July 2020; pp. 2131–2147. [Google Scholar]
  42. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  43. Sun, P.; Zhang, R.; Jiang, Y.; Kong, T.; Xu, C.; Zhan, W.; Tomizuka, M.; Li, L.; Yuan, Z.; Wang, C.; et al. Sparse R-CNN: End-to-end object detection with learnable proposals. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 14454–14463. [Google Scholar]
  44. Tan, M.; Pang, R.; Le, Q.V. Efficientdet: Scalable and efficient object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 10781–10790. [Google Scholar]
  45. Zhu, B.; Wang, J.; Jiang, Z.; Zong, F.; Liu, S.; Li, Z.; Sun, J. Autoassign: Differentiable label assignment for dense object detection. arXiv 2020, arXiv:2007.03496. [Google Scholar]
  46. Chen, Q.; Wang, Y.; Yang, T.; Zhang, X.; Cheng, J.; Sun, J. You only look one-level feature. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 13039–13048. [Google Scholar]
  47. Zhu, X.; Su, W.; Lu, L.; Li, B.; Wang, X.; Dai, J. Deformable DETR: Deformable transformers for end-to-end object detection. In Proceedings of the International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020; pp. 12254–12272. [Google Scholar]
  48. Zhang, S.; Wang, X.; Wang, J.; Pang, J.; Lyu, C.; Zhang, W.; Luo, P.; Chen, K. Dense distinct query for end-to-end object detection. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Vancouver, BC, Canada, 18–22 June 2023; pp. 7329–7338. [Google Scholar]
  49. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Ling, J.; Su, H.; Jin, W.; Zhang, Y.; et al. Dataset of weak small aircraft detection and tracking in ground/air background infrared images. China Sci. Data 2020, 5, 291–302. [Google Scholar]
Figure 1. Mechanism of AWFG.
Figure 1. Mechanism of AWFG.
Electronics 13 02695 g001
Figure 2. Function units of AWFGLC.
Figure 2. Function units of AWFGLC.
Electronics 13 02695 g002
Figure 3. Network architecture of AWFGLC-YOLO.
Figure 3. Network architecture of AWFGLC-YOLO.
Electronics 13 02695 g003
Figure 4. Effect of target feature enhancement.
Figure 4. Effect of target feature enhancement.
Electronics 13 02695 g004
Figure 5. Curves of learnable weight on self-made dataset.
Figure 5. Curves of learnable weight on self-made dataset.
Electronics 13 02695 g005
Figure 6. Visualization heatmaps of different context mechanisms.
Figure 6. Visualization heatmaps of different context mechanisms.
Electronics 13 02695 g006
Figure 7. Detection results of different algorithms on self-made infrared dataset.
Figure 7. Detection results of different algorithms on self-made infrared dataset.
Electronics 13 02695 g007
Figure 8. Curves of learnable weights on the public dataset.
Figure 8. Curves of learnable weights on the public dataset.
Electronics 13 02695 g008
Figure 9. Detection results of different algorithms on the public dataset.
Figure 9. Detection results of different algorithms on the public dataset.
Electronics 13 02695 g009
Table 1. Comparative experiment of different context mechanisms.
Table 1. Comparative experiment of different context mechanisms.
MethodmAP50/%mAP50:95/%Parameter/MFLOPs/G
YOLOv796.461.435.6105.4
YOLOv7+GC94.360.236.2105.9
YOLOv7+PA96.762.636.1106.1
YOLOv7+LT95.560.836.5106.7
AWFGLC-YOLO97.865.736.9107.2
Table 2. Comparative experiment of different detection algorithms on self-made dataset.
Table 2. Comparative experiment of different detection algorithms on self-made dataset.
MethodmAP50/%mAP50:95/%Params/MFLOPs/G
CenterNet92.459.120.414.2
Autoassign94.160.579.136.0
Efficientdet73.745.918.346.4
Sparse R-CNN90.859.7105.964.6
YOLOF80.649.342.341.2
Deformable DETR93.859.240.127.4
DDQ97.265.347.2203.9
YOLOv594.463.713.916.4
YOLOv796.461.435.6105.4
YOLOv896.062.325.079.1
AWFGLC-YOLO97.865.736.9107.2
Table 3. Comparative experiment of different detection algorithms on the public dataset.
Table 3. Comparative experiment of different detection algorithms on the public dataset.
MethodmAP50%mAP50:95%Params/MFLOPs/G
CenterNet79.650.820.414.2
Autoassign80.051.779.136.0
Efficientdet64.248.118.346.4
Sparse R-CNN76.450.2105.964.6
YOLOF54.220.442.341.2
Deformable DETR81.552.240.127.4
DDQ87.360.447.2203.9
YOLOv584.757.413.916.4
YOLOv785.356.735.6105.4
YOLOv885.757.225.079.1
AWFGLC-YOLO88.761.236.9107.2
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Liu, G.; Xi, J.; Tong, J.; Xu, H. An Infrared Aircraft Detection Algorithm Based on Context Perception Feature Enhancement. Electronics 2024, 13, 2695. https://doi.org/10.3390/electronics13142695

AMA Style

Liu G, Xi J, Tong J, Xu H. An Infrared Aircraft Detection Algorithm Based on Context Perception Feature Enhancement. Electronics. 2024; 13(14):2695. https://doi.org/10.3390/electronics13142695

Chicago/Turabian Style

Liu, Gang, Jiangtao Xi, Jun Tong, and Hongpeng Xu. 2024. "An Infrared Aircraft Detection Algorithm Based on Context Perception Feature Enhancement" Electronics 13, no. 14: 2695. https://doi.org/10.3390/electronics13142695

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop