Next Article in Journal
At Which Overpass Time Do ECOSTRESS Observations Best Align with Crop Health and Water Rights?
Previous Article in Journal
An Improved Average Acceleration Approach of Modelling Earth Gravity Field Based on K-Band Range-Rate Observations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Communication

An Efficient Knowledge Distillation-Based Detection Method for Infrared Small Targets

1
College of Computer Science and Technology, Nanjing University of Aeronautics and Astronautics, Nanjing 210016, China
2
HIWING Technology Academy, China Aerospace Science and Industry Corporation Limited (CASIC), Beijing 100074, China
3
School of Integrated Circuits, Beijing University of Posts and Telecommunications, Beijing 100876, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(17), 3173; https://doi.org/10.3390/rs16173173 (registering DOI)
Submission received: 13 June 2024 / Revised: 16 August 2024 / Accepted: 23 August 2024 / Published: 28 August 2024

Abstract

:
Infrared small-target detection is now commonly used in maritime surveillance, flight guidance, and other fields. However, extracting small targets from complex backgrounds remains a challenging task due to the small-target scale and complex imaging environment. Many studies are based on designing model structures to enhance the precision of target detection, and the number of Params and FLOPs has been significantly augmented. In this work, a knowledge distillation-based detection method (KDD) is proposed to overcome this challenge. KDD employs the small-target labeling information provided by a large-scale teacher model to refine the training process of students, thereby improving the performance and becoming lightweight. Specifically, we added efficient local attention (ELA), which can accurately identify areas of interest while avoiding dimensionality reduction. In addition, we also added the group aggregation bridge (GAB) module to connect low-level and high-level features for the fusion of different feature scales. Furthermore, a feature fusion loss was introduced to enhance the precision of target detection. Extensive evaluations have demonstrated that KDD performs better compared to several methods, achieving extremely low Params and FLOPs, as well as higher FPS.

1. Introduction

Infrared small-target detection (ISTD) identifies dim and small targets from infrared images full of clutter interference [1]. It is widely applied in traffic management, maritime surveillance, and flight guidance [2,3,4]. Compared with general target detection tasks, infrared small targets exhibit the following traits: (1) due to the characteristics of long-distance imaging, the target usually takes up a small number of pixels and the shape and texture characteristics are not obvious; and (2) infrared images are characterized by a low signal-to-clutter ratio (SCR) and a great deal of noise, so small targets are more likely to be submerged in the background and will be difficult to identify [5]. These characteristics result in the phenomenon of false alarms and miss detection in ISTD, which makes the task of detection challenging.
Researchers have continuously improved and optimized detection methods to solve the problems caused by the above characteristics. These methods can be categorized into traditional and deep learning-based. Among the traditional methods, filtering-based methods [6,7,8,9] are computationally small and can serve to suppress uniform backgrounds to some extent. Still, they cannot solve the problem of complex backgrounds, resulting in low detection rates. The human visual system (HVS) [10,11,12] mainly makes use of saliency maps, where the presence of small targets leads to a saliency change in the local texture of the image. HSV can be divided into two kinds of methods based on spectral residual and local contrast. However, spectral residuals cannot suppress clutter well. Local contrast-based methods have poor generalization ability and cannot adapt to complex background images. The low-rank sparse recovery [13,14,15,16,17] can be applied in scenes with high noise interference and low target contrast. It needs to consume large computational resources and time in practical applications.
Unlike traditional methods, deep learning-based methods acquire target features through data-driven learning processes. Liu et al. [18] constructed a five-layer detection network by utilizing multilayer perception. MDvsFA-cGAN [19] separates the ISTD task into subtasks for reducing false alarms and miss detections. Attentional local contrast network (ALCNet) [20] designs a cyclic shifting method for feature maps and utilizes bottom-up attention modulation. Dense nested attention network (DNANet) [21] achieves progressive interactions between the higher and lower layers that are prior through the dense nested network. Ma et al. [22] generated background components and reduced their interference via a generative adversarial network. Multiscale Feature Extraction U-Net [23] extracts rich multi-scale information to accurately detect targets in complex backgrounds. Li et al. [1] proposed a diffusion model framework, IRSTD-Diff. The above methods can also be considered as ones that convert ISTD to binary semantic segmentation, where the positive and negative samples exhibit extreme imbalance [24]. With the continuous development of YOLO, it is applied to various kinds of scenarios [25,26]. It has also been applied in infrared small-target detection. Liu et al. proposed a compact network based on YOLO to improve the accuracy and speed of detection [27]. Ciocarlan et al. introduced an inverse decision criterion in the training of a YOLO detector [28]. But YOLO still performs poorly in identifying tiny objects. Illuminatingly, some classic clustering methods for unbalanced data [29] may be used to detect infrared targets.
Influenced by the features of infrared small targets, deep learning-based methods tend to extract different levels of features by designing some specific modules or retaining the information of infrared small targets by some feature fusion strategies. These complex network models improve the detection accuracy but consume a great deal of storage space and computational resources. Therefore, we would like to construct a network to make use of the idea where the knowledge distillation in the model is a form of lightweight technology so that the network of high-precision complex network models assists the student model training. And we also designed relevant modules to make the positioning of small targets more accurate and would retain the different levels of characteristics of small targets.
Inspired by knowledge distillation [30], we proposed KDD, a knowledge distillation-based ISTD network. DNANet is utilized as the teacher model in KDD, providing soft labels to guide and assist the training process of the student model. U-Net [31] is used as the basic architecture of the student model. Infrared targets appear in relatively small areas of the image and have higher requirements for the localization ability of the network. Therefore, in the student model, we added the ELA [32] module, which can effectively utilize spatial information to accurately locate the position of interest without sacrificing the channel dimension. Infrared images result in small targets with different appearance characteristics due to the imaging environment. Acquiring multi-scale information can better understand the contextual information. We added the GAB [33] module to fuse the low-level and high-level features in a grouping for multi-scale information. KDD achieves the goal of being lightweight and it also improves accuracy. The main contributions of this article are summarized below:
  • We designed a novel ISTD method based on the idea of knowledge distillation for infrared target detection. KDD not only improves the detection accuracy, but also achieves a lightweight network.
  • KDD incorporates the ELA module and the GAB module. The small-target pixels account for a relatively small percentage of pixels. To enhance the localization ability, the ELA module was added to achieve precise localization of the region of interest without sacrificing the channel dimension. The GAB module was integrated to exploit grouping for multi-scale information capture and to make rational use of the contextual information.

2. Related Work

Deep learning has demonstrated significant benefits in the areas of computer vision, speech understanding, and various other domains. Deep learning AI relies on large pre-trained models to improve the performance of deep neural networks in different tasks. These models are complex and have numerous parameters, making them difficult to use on mobile hardware. Therefore, model compression and acceleration techniques have become one of the current research hotspots. Model quantization is one of the main effective methods. The popular methods are mainly divided into model pruning [34,35], knowledge distillation [30,36,37], neural network architecture search [38,39], and tensor decomposition [40,41]. Specifically, model pruning reduces model size by removing redundant parameters and connections while maintaining performance. Knowledge distillation uses trained teacher models to guide the training of student models by designing specific losses. Neural network architecture search automatically designs the neural network structure through algorithms. Tensor decomposition decomposes the tensor into small sizes to achieve the goal of reducing the storage space of the neural network model.
Large-scale language models (LLMs) like GPT-4 and Gemini have emerged as seminal techniques [42]. These models have become pioneering techniques for solving complex problems. However, their wide application has been hindered by the high cost of use and limited access. In contrast, some open-source models (e.g., LLaMa [43]) do not have cost and access restrictions, but their performance is lagging behind due to their small size. The emergence of knowledge distillation techniques has revolutionized this situation, allowing open-source models to learn from extensively trained and fine-tuned proprietary models, thereby improving performance. Knowledge distillation is widely used in the field of natural language processing [44]. The difference between proprietary and open-source models has gradually narrowed or even disappeared due to the use of knowledge distillation techniques, which has led to further improvements in both the performance and efficiency of open-source models.
Currently, knowledge distillation is categorized into logit-based distillation [45] and feature-based distillation [36]. To learn the decision process and probability distribution of the large model, the logit-based knowledge distillation takes the logit output of a teacher model as a soft target for training a student model. Neural networks commonly produce class probabilities by employing a softmax output layer, which transforms the logit value, z i , which is calculated for each class into a probability, q i , through a comparison with the logit values of other classes. It can be expressed as follows:
q i = exp ( z i / T ) j exp ( z j / T ) ,
where T denotes the temperature. In using the cross-entropy loss of soft targets, the performance of the student model can be further improved.
Feature-based knowledge distillation allows student models to improve the training process using the learned intermediate representation of the teacher as a cue. Since the intermediate hidden layer of the student is usually smaller than that of the teacher, additional parameters need to be introduced to ensure that the hidden layer maps of the student model with the hidden layer of the teacher model effectively produce accurate predictions. In general, the representation of hidden layer feature distillation can be described as follows:
L f e a = 1 N c = 1 C h = 1 H w = 1 W ( T c , h , w f ( S k , i , j ) ) 2 ,
where N denotes the sum of elements in the feature; T and S denote the teacher and student features, respectively; and f denotes the adaptation layer for the same dimensions of its teacher feature map. In summary, the idea of knowledge distillation is enlightening to explore for ISTD.

3. Methodology

3.1. Overall Architecture

As shown in Figure 1, the overall structure of KDD has two components: the teacher model and the student model. DNANet was selected as the teacher model. A single infrared image was used as the network input, and the process of detection was achieved through feature extraction, feature fusion, and clustering. The core of the DNANet network lies in its dense nested architecture, enabling interactions of different level features, coupled with the cascaded attention module that adaptively boosts multi-scale features. The student model uses the U-Net structure. Convolution, pooling, and interpolation operations are utilized in the shallow part for encoding and decoding operations, and the ELA module is used in the deep part for accurate localization of regions of interest. In the skip connection, we added the GAB module to extract information about the interaction features between different levels.

3.2. Network Details

(1) Logits-based distillation: The purpose of knowledge distillation is to allow students to obtain valuable potential knowledge from the teacher. The negative labeling information provided by the teacher model contains a certain amount of information that can assist the student model in rapidly acquiring an understanding of the reasoning process of the teacher model. The specific process of knowledge distillation using the logits method for infrared small targets is shown in Figure 2.
(2) Loss function: To better learn the information provided by the teacher model, we calculated the loss between the teacher and student model outputs. Soft Intersection over Union Loss (SoftIoULoss) was chosen as the loss function for when the smaller value indicates that the output detection result of the student model has a higher degree of overlap with the teacher model. It can be expressed as follows:
L S o f t I O U = 1 | O s O t | + θ s m o o t h | O s | + | O t | | O s O t | + θ s m o o t h ,
where O s and O t are the result of the student and teacher model, respectively.
We used BCEloss and Diceloss [46] to measure the difference between the results of the student model and the mask. BCEloss was used to judge the degree of the prediction results of a binary classification model, and Diceloss was used as a regionally correlated loss function. We also fused the feature maps of different stages by convolution and interpolation operations, and we generated the predicted mask for comparison with the original target. The loss representation is as follows:
L = L B C E ( O , M ) + L D i c e ( O , M ) ,
L h a r d = L ( O s , M ) + L ( O p g , M ) ,
where O p g denotes the result of feature fusion, and M denotes the ground truth.

3.3. Efficient Local Attention

U-Net realizes the encoding and decoding operations through convolutional operations, which may lose the feature information and have some impact on the subsequent segmentation or detection tasks. Although attention mechanisms can significantly improve the performance of models, existing implementations are often accompanied by a reduction in channel dimensions or an increase in network complexity. For example, the SE block attention [47] only considers the information between coding channels and ignores the potential benefits of spatial location information. The coordinate attention (CA) [48] method focuses on spatial information but has insufficient generalization ability.
In our approach, we use ordinary convolution to extract features in the shallow layer and replace the ordinary convolution operation with the ELA module in the deep layer. Infrared small targets are a relatively small proportion of the image and have higher requirements on the localization ability of the network. The ELA module combines the feature enhancement techniques of one-dimensional convolution and group normalization to help the network focus its attention on the region, and it may contain small targets without sacrificing the channel dimensions, thus improving the accuracy and precision of localization. Figure 3 shows the schematic diagrams of ELA.
The input feature map x was average pooled over two spatial scales, with the horizontal direction ( H , 1 ) and the vertical direction ( 1 , W ) , and the outputs of the c t h channel with height h and width w are denoted as follows:
z c h ( h ) = 1 W 0 i < W x c ( h , i ) ,
z c w ( w ) = 1 H 0 j < H x c ( j , w ) .
The obtained two types of feature vectors were augmented with information by one-dimensional convolution, and a group normalization operation was added to deal with the augmented positional information. Finally, the two spatial ranges of features were output through an activation function. The process is represented as follows:
f h = δ ( G N ( C o n v 1 d ( z h ) ) ) ,
f w = δ ( G N ( C o n v 1 d ( z w ) ) ) ,
where C o n v 1 d denotes 1D convolution operation, the kernel size is set to 7, and the padding is set to 3. G N denotes group normalization, the num group is set to 8, and δ denotes the Sigmoid function.
Finally, the horizontal and vertical outputs were multiplied with the inputs, and the outputs were fed into the 2D convolution to obtain the outputs.
y = C o n v 2 d ( f h f w x ) .

3.4. Group Aggregation Bridge

One of the advantages of U-Net is the incorporation of encoder information into the decoding process through hopping layer connections. In target detection tasks, it is necessary to obtain information about features at different scales. Due to the different imaging distances and angles, targets present different appearance characteristics and are easily disturbed by noise and other factors. The acquisition of multi-scale information can capture visual representations at different scales, as well as better understand contextual information to improve accuracy. Therefore, we introduced the GAB module in the skip connection. Figure 4 shows the schematic diagrams.
GAB received low-level features and high-level features. Firstly, the size of the high-level features was changed using 2D convolution and interpolation to match the bottom-level features.
y h = P ( C o n v ( x h ) ) ) ,
where P denotes the interpolation, and C o n v denotes the Conv2d.
Second, the different levels of features were divided into four groups along the channel direction, which makes the correlation between each group of features stronger. After that, we combined the high-level and low-level counterparts and performed layer normalization and 2D convolution for each combination to extract information at different scales.
z h = { y 1 h , y 2 h , y 3 h , y 4 h } , z l = { x 1 l , x 2 l , x 3 l , x 4 l } ,
α i = D C o n v ( L a y e r ( C ( y i h , x i l ) ) ) ,
where z h and z l denote the division results of high-level features and low-level features, respectively; C denotes the cat operation; L a y e r denotes the LayerNorm operation; and D C o n v denotes the Dilated Conv2d with different subgroups and dilation rates of 1, 2, 5, and 7.
Finally, the four combinations werre reconnected to realize the feature interaction by convolution.
β = C o n v ( L a y e r ( C ( α 1 , α 2 , α 3 , α 4 ) ) ) .

4. Experiments

4.1. Experimental Setup

(1) Datasets: We conducted experiments on three datasets, IRSTD-1k [4], SIRST [49], and MSISTD [50]. The IRSTD-1k dataset comprises one thousand infrared images covering various types of small targets including vehicles, aircraft, boats, etc. The background of the dataset covers a wide range, including cities, fields, mountains, oceans, rivers, and skies with heavy clutter and noise [4]. The SIRST dataset comprises 427 images. A significant proportion of these targets are low in brightness and are lost in cluttered and complex backgrounds. The MSISTD dataset consists of 1077 images with multi-scene, multi-scale, and lower SCR targets. It covers all types of scenes including sky, sea, land, forest, urban buildings, etc. The datasets are split into training and testing sets in a ratio of 8:2.
(2) Comparison methods: We compared KDD with some SOTA methods, including New TopHat [6], nonconvex rank approximation minimization (NRAM) [15], asymmetric contextual modulation (ACM) [49], DNANet [21], interior attention-aware network (IAANet) [51], attention-guided pyramid context networks (AGPCNet) [2], and lightweight IR small-target segmentation network (LW-IRSTNet) [52]. New TopHat improves on the classic top hat method by utilizing two distinct but related structural elements and by taking into account the differences between the target and its surroundings. NRAM employs non-convex, tighter-ranked surrogates, and a weighted l 1 criterion. ACM utilizes a bottom-up mechanism to preserve low-level features with fine details. DNANet is designed with densely nested interaction modules to maintain deep information about small targets, and it utilizes cascaded channels and spatial attention modules to enhance multilevel features. IAANet captures the coarse target area and first filters out the background, then performs fine detection. AGPCNet is designed with attention-guided context blocks to enhance the ability to sense small targets through using context pyramids and asymmetric fusion modules to retain small-target feature information. LW-IRSTNet integrates regular convolutions, depthwise separable convolutions, atrous convolutions, and asymmetric convolutions modules.
(3) Evaluation Metrics: Among the most commonly used metrics to assess the accuracy of detection are r e c a l l and p r e c i s i o n . The pursuit of high p r e c i s i o n or high r e c a l l alone is not a sufficiently comprehensive measure of detection method performance. For a more balanced assessment, we introduced the F 1 , I o U , and ROC. The calculation of F 1 is shown in Equation (15). I o U reflects the accuracy of localization through the intersection and concurrency ratio between prediction results and target, and the calculation of I o U is shown in Equation (16). A U C is the area under the ROC curve, and it is a quantitative assessment between the true target and misclassification.
F 1 = 2 × R e c a l l × P r e c i s i o n R e c a l l + P r e c i s i o n ,
I o U = A i / A u .
(4) Implementation Details: The process was implemented on PyTorch and the computer used an NVIDIA GeForce GTX 1650. The teacher model was trained according to the original parameters of DNANet, where the epoch was 1500 and the batch size was 4. In the student model, the AdamW optimizer was employed, and it was initialized with a learning rate of 0.001 and a weight decay rate of 0.01, which was adjusted using CosineAnnealingLR. In addition, the epoch was 150 and the batch size was 4.

4.2. Basic Experiments

(1) Visual Evaluation: In Figure 5, Figure 6 and Figure 7, some visualization results of the different methods on the SIRST and IRSTD-1k datasets are shown, where the red color means the positive detection results, green color indicates the miss detections, and the yellow color denotes the false alarms.
Figure 5 illustrates the detection results on the SIRST dataset. For Images (1)–(3), which have significant bright region effects, New Tophat, NRAM, and IAANet were prone to false alarms when detecting the target. AGPCNet was prone to miss detections of dark targets. However, KDD was able to avoid the interference of the highlighted regions and accurately detect targets that are very similar to the masks. For the targets in a simple background but very dark, e.g., Images (4)–(5), New Tophat and NRAM were able to detect the target location but the detection results were incomplete. IAANet and AGPCNet were prone to miss detections of dark targets. In the challenging building background (e.g., Image (6)), New Tophat, NRAM, ACM, IAANet, and AGPCNet all produced varying degrees of false alarms. In summary, KDD can perform relatively stable detection in different situations.
The results on the IRSTD-1k are shown in Figure 6. For images where small targets are more obvious and the background interference is small (e.g., Image (1)), all types of methods were able to accurately localize the target, but the New Tophat method still produced false detections. When the target was obvious but there was interference information in images such as highlighted areas in the background (e.g., Images (4), (5)), New Tophat, NRAM, ACM, IAANet, AGPCNet, and LW-IRSTNet all produced a certain degree of false detection for highlighted areas. In images where the target was not obvious and there was a great deal of distracting information in the background (e.g., Images (2), (3), and (6)), the detection was susceptible to background influence and was difficult to distinguish from the highlighted regions in the background. New tophat and NRAM produced a great deal of false alarms, and there was a leakage of the target detection by AGPCNet. From the detection results of the whole dataset, it can be concluded that the traditional methods have defects in the integrity of their detection of targets. The size of the detection result differed significantly from the mask when there was more interfering information in the background.
Figure 7 shows the visualized detection results for the MSISTD dataset. For small-target images with a simple sky background (e.g., Image (2)), all of the methods except NRAM were able to detect the small-target region but, due to the influence of clouds, New Tophat, IAANet, and LW-IRSTNet were prone to false detection. For infrared images with a relatively large proportion of small-target pixels (e.g., Image (4)), all of the methods except ACM were able to detect the targets, but they had some missing cases. In images with urban building backgrounds and more obvious targets (e.g., Images (1), (3), and (6)), the detection was affected by the highlighted areas of the buildings, and there were a large number of false detections in New Tophat and NRAM, and there were also a small number of false detections in IAANet and LW-IRSTNet. In images with complex backgrounds and unclear targets (e.g., Image (5)), NRAM and AGPCNet experienced missed detections, and New Tophat, NRAM, and IAANet produced false detections. Overall, KDD was able to obtain more satisfactory results in both simple and complex scenes.
(2) Numerical Evaluation: Table 1 lists the F 1 , I o U , and A U C for the different methods on three datasets, with the best results highlighted in red and the second-best results highlighted in green. Figure 8 shows the ROC curves.
(a)
SIRST: From Table 1, it is evident that deep learning methods are superior to traditional methods. KDD scored the highest in r e c a l l , F 1 , and I o U although DNANet had a high p r e c i s i o n . The r e c a l l performance was average so it led to the F 1 value of the combined metric of the two to not be good. The results of the traditional methods included false alarms, which increased in connected areas between the detection results and the area of concatenation between detection results and targets, resulting in smaller values for I o U . The A U C of KDD reached 0.9601 on the STRST dataset.
(b)
IRSTD-1k: In the quantitative comparison, KDD had the best I o U , which was 0.6359, and the ROC curve is shown in Figure 8b. Although DNANet performed the best on p r e c i s i o n and F 1 , the difference with the results of KDD was very small and was not the best on several other metrics. In addition, due to the large impact of background and clutter in the images, as well as the small-target size, this dataset produced a lower metric than the SIRST dataset. Overall, KDD still obtained the best detection performance.
(c)
MSISTD: In quantitative evaluation, KDD had the highest value of F 1 and I o U with 0.7571 and 0.6688, respectively. The ROC curve is shown in Figure 8c. DNANet also performed well on the MSISTD dataset and achieved the highest value on PRECISION. But the value of recall was lower than KDD, and the combined value of both F1 scores was not the best. IAANet achieved the highest value on AUC, and KDD was located at second.

4.3. Ablation Experiments

In KDD, the basic network structure of our model was composed of U-Net and was incorporated of three parts: the ELA module, the GAB module, and the feature fusion loss. To assess the effect of each part on the overall network performance, a set of ablation studies was conducted on two distinct datasets. Table 2 lists the test results of the ablation experiments. In the absence of the ELA module and lack of feature fusion loss, both datasets performed poorly in F 1 , I o U , and A U C . In the absence of the GAB module, although the results of r e c a l l were good and were able to reach 0.8264 and 0.8118 on the two datasets, the results of p r e c i s i o n were worse at 0.7957 and 0.6216, respectively, which led to a poorer equalization measure F 1 for both. When all three modules were present, the evaluation metrics of the experimental results had better performance. In conclusion, the additional modules incorporated into the fundamental structure proved to be beneficial in enhancing network performance.

4.4. Complexity Analysis

Currently, computational complexity is an important index. Table 3 provides a comparison of the Params, FLOPs, and FPS for various deep learning methods. Notably, KDD achieved optimal performance with the lowest values for both Params and FLOPs, followed by ACM in second place. The FPS value of New Tophat was higher among the traditional methods, but the method was simpler and there were more miss detections and false alarms in the results. Among the deep learning based methods, KDD had the highest FPS value because it is a lightweight network. The FPS values of DNANet and ACM were second and third, respectively.

5. Conclusions

This paper aims to propose a lightweight method based on knowledge distillation for ISTD. Utilizing the concept of knowledge distillation, KDD transfers labeling information from the teacher model to the student model, aiding the training process of the latter. Subsequently, the region of interest is identified by the ELA module, while the GAB module is used to obtain different scale features. Experimental results on three public datasets demonstrate that the KDD can accurately detect infrared small targets and has superior performance in several evaluation metrics. The ablation experiments demonstrate the efficacy of each module of KDD. Through the comparison of computational complexity, it can be seen that KDD achieved becoming a lightweight system. In missions such as search and rescue and surveillance, real-time monitoring capabilities and low-energy consumption characteristics are critical, and KDD fulfills these needs with a great deal of real-world applications. Although KDD shows superior performance, it still has certain limitations. In the second dataset, we can see that the performance of the detection results on the evaluation metrics was very close to the teacher model but not optimal. In future work, we intend to optimize the student model further by retaining more information about small-target features. To obtain better guidance from the teacher model, we can also consider introducing feature-based distillation methods, which could potentially lead to further improvements in performance.

Author Contributions

W.T.: conceptualization, methodology, formal analysis, and writing—review and editing; Q.D.: investigation, resources, writing—review and editing, and supervision; F.H.: software, validation, data curation, writing—original draft preparation, and visualization. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China under Grant no. 62476126.

Data Availability Statement

SIRST can be obtained from (https://github.com/YimianDai/sirst, accessed on 17 February 2024) and IRSTD-1k can be obtained from (https://github.com/RuiZhang97/ISNet, accessed on 17 February 2024). MSISTD can be obtained from (https://github.com/Crescent-Ao/MSISTD, accessed on 17 February 2024).

Conflicts of Interest

Author Wenjuan Tang was employed by the HIWING Technology Academy, China Aerospace Science and Industry Corporation Limited (CASIC). The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

References

  1. Li, H.; Yang, J.; Xu, Y.; Wang, R. Mitigate Target-level Insensitivity of Infrared Small Target Detection via Posterior Distribution Modeling. arXiv 2024, arXiv:2403.08380. [Google Scholar] [CrossRef]
  2. Zhang, T.; Li, L.; Cao, S.; Pu, T.; Peng, Z. Attention-Guided Pyramid Context Networks for Detecting Infrared Small Target Under Complex Background. IEEE Trans. Aerosp. Electron. Syst. 2023, 59, 4250–4261. [Google Scholar] [CrossRef]
  3. Teutsch, M.; Krüger, W. Classification of small boats in infrared images for maritime surveillance. In Proceedings of the 2010 International WaterSide Security Conference, Carrara, Italy, 3–5 November 2010; pp. 1–7. [Google Scholar]
  4. Zhang, M.; Zhang, R.; Yang, Y.; Bai, H.; Zhang, J.; Guo, J. ISNet: Shape Matters for Infrared Small Target Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 19–21 June 2022; pp. 867–876. [Google Scholar]
  5. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  6. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  7. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. In Proceedings of the Signal and Data Processing of Small Targets 1999, Denver, CO, USA, 20–22 July 1999; Volume 3809, pp. 74–83. [Google Scholar]
  8. Zeng, M.; Li, J.; Peng, Z. The design of top-hat morphological filter and application to infrared target detection. Infrared Phys. Technol. 2006, 48, 67–76. [Google Scholar] [CrossRef]
  9. Azimi-Sadjadi, M.; Pan, H. Two-dimensional block diagonal LMS adaptive filtering. IEEE Trans. Signal Process. 1994, 42, 2420–2429. [Google Scholar] [CrossRef]
  10. Chen, C.L.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A Local Contrast Method for Small Infrared Target Detection. IEEE Trans. Geosci. Remote Sens. 2014, 52, 574–581. [Google Scholar] [CrossRef]
  11. Han, J.; Ma, Y.; Zhou, B.; Fan, F.; Liang, K.; Fang, Y. A robust infrared small target detection algorithm based on human visual system. IEEE Geosci. Remote Sens. Lett. 2014, 11, 2168–2172. [Google Scholar]
  12. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  13. Zhu, H.; Liu, S.; Deng, L.; Li, Y.; Xiao, F. Infrared Small Target Detection via Low-Rank Tensor Completion With Top-Hat Regularization. IEEE Trans. Geosci. Remote Sens. 2020, 58, 1004–1016. [Google Scholar] [CrossRef]
  14. Dai, Y.; Wu, Y. Reweighted Infrared Patch-Tensor Model With Both Nonlocal and Local Priors for Single-Frame Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef]
  15. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared Small Target Detection via Non-Convex Rank Approximation Minimization Joint l2,1 Norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef]
  16. Lin, Z.; Ganesh, A.; Wright, J.; Wu, L.; Chen, M.; Ma, Y. Fast Convex Optimization Algorithms for Exact Recovery of a Corrupted Low-Rank Matrix; Coordinated Science Laboratory Report No. UILU-ENG-09-2214, DC-246; Coordinated Science Laboratory: Urbana, IL, USA, 2009. [Google Scholar]
  17. Zhu, H.; Ni, H.; Liu, S.; Xu, G.; Deng, L. TNLRS: Target-Aware Non-Local Low-Rank Modeling With Saliency Filtering Regularization for Infrared Small Target Detection. IEEE Trans. Image Process. 2020, 29, 9546–9558. [Google Scholar] [CrossRef]
  18. Liu, M.; Du, H.-y.; Zhao, Y.-j.; Dong, L.-q.; Hui, M. Image Small Target Detection based on Deep Learning with SNR Controlled Sample Generation. In Current Trends in Computer Science and Mechanical Automation Vol. 1; Wang, S.X., Ed.; De Gruyter Open Poland: Warsaw, Poland, 2022; pp. 211–220. [Google Scholar]
  19. Wang, H.; Zhou, L.; Wang, L. Miss Detection vs. False Alarm: Adversarial Learning for Small Object Segmentation in Infrared Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Repulic of Korea, 27 October–2 November 2019; pp. 8508–8517. [Google Scholar]
  20. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional Local Contrast Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  21. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense Nested Attention Network for Infrared Small Target Detection. IEEE Trans. Image Process. 2023, 32, 1745–1758. [Google Scholar] [CrossRef] [PubMed]
  22. Ma, Z.; Pang, S.; Hao, F. Generative Adversarial Differential Analysis for Infrared Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2024, 17, 6616–6626. [Google Scholar] [CrossRef]
  23. Wang, X.; Han, C.; Li, J.; Nie, T.; Li, M.; Wang, X.; Huang, L. Multiscale Feature Extraction U-Net for Infrared Dim- and Small-Target Detection. Remote Sens. 2024, 16, 643. [Google Scholar] [CrossRef]
  24. Kou, R.; Wang, C.; Peng, Z.; Zhao, Z.; Chen, Y.; Han, J.; Huang, F.; Yu, Y.; Fu, Q. Infrared small target segmentation networks: A survey. Pattern Recognit. 2023, 143, 109788. [Google Scholar] [CrossRef]
  25. Lan, W.; Dang, J.; Wang, Y.; Wang, S. Pedestrian Detection Based on YOLO Network Model. In Proceedings of the 2018 IEEE International Conference on Mechatronics and Automation (ICMA), Changchun, China, 5–8 August 2018; pp. 1547–1551. [Google Scholar]
  26. Xia, Z.; Ma, K.; Cheng, S.; Blackburn, T.; Peng, Z.; Zhu, K.; Zhang, W.; Xiao, D.; Knowles, A.J.; Arcucci, R. Accurate identification and measurement of the precipitate area by two-stage deep neural networks in novel chromium-based alloys. Phys. Chem. Chem. Phys. 2023, 25, 15970–15987. [Google Scholar]
  27. Liu, S.; Liu, Z.; Li, Y.; Liu, W.; Ge, C.; Liu, L. Design Compact YOLO based Network for Small Target Detection on Infrared Image. In Proceedings of the 2022 China Automation Congress (CAC), Xiamen, China, 25–27 November 2022; pp. 4991–4996. [Google Scholar]
  28. Ciocarlan, A.; Le Hegarat-Mascle, S.; Lefebvre, S.; Woiselle, A.; Barbanson, C. A Contrario Paradigm for Yolo-Based Infrared Small Target Detection. In Proceedings of the ICASSP 2024—2024 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Seoul, Republic of Korea, 14–19 April 2024; pp. 5630–5634. [Google Scholar]
  29. Zhang, Z.W.; Liu, Z.G.; Martin, A.; Zhou, K. BSC: Belief Shift Clustering. IEEE Trans. Syst. Man Cybern. Syst. 2023, 53, 1748–1760. [Google Scholar] [CrossRef]
  30. Hinton, G.; Vinyals, O.; Dean, J. Distilling the knowledge in a neural network. arXiv 2015, arXiv:1503.02531. [Google Scholar]
  31. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, 5–9 October 2015; Proceedings, Part III 18. Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  32. Xu, W.; Wan, Y. ELA: Efficient Local Attention for Deep Convolutional Neural Networks. arXiv 2024, arXiv:2403.01123. [Google Scholar]
  33. Ruan, J.; Xie, M.; Gao, J.; Liu, T.; Fu, Y. EGE-UNet: An Efficient Group Enhanced UNet for Skin Lesion Segmentation. In Proceedings of the Medical Image Computing and Computer Assisted Intervention—MICCAI 2023, Vancouver, BC, Canada, 8–12 October 2023; pp. 481–490. [Google Scholar]
  34. Dong, X.; Chen, S.; Pan, S.J. Learning to prune deep neural networks via layer-wise optimal brain surgeon. In Proceedings of the 31st Conference on Neural Information Processing Systems (NIPS 2017), Long Beach, CA, USA, 4–9 December 2017; pp. 4860–4874. [Google Scholar]
  35. Liu, Z.; Li, J.; Shen, Z.; Huang, G.; Yan, S.; Zhang, C. Learning Efficient Convolutional Networks through Network Slimming. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 2755–2763. [Google Scholar]
  36. Romero, A.; Ballas, N.; Kahou, S.E.; Chassang, A.; Gatta, C.; Bengio, Y. Fitnets: Hints for thin deep nets. arXiv 2014, arXiv:1412.6550. [Google Scholar]
  37. Ahn, S.; Hu, S.X.; Damianou, A.; Lawrence, N.D.; Dai, Z. Variational Information Distillation for Knowledge Transfer. In Proceedings of the 2019 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 9155–9163. [Google Scholar]
  38. Zoph, B.; Le, Q.V. Neural Architecture Search with Reinforcement Learning. arXiv 2017, arXiv:1611.01578. [Google Scholar]
  39. Alizadeh Vahid, K.; Prabhu, A.; Farhadi, A.; Rastegari, M. Butterfly Transform: An Efficient FFT Based Neural Architecture Design. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Seattle, WA, USA, 13–19 June 2020; pp. 12021–12030. [Google Scholar]
  40. Kim, Y.D.; Park, E.; Yoo, S.; Choi, T.; Yang, L.; Shin, D. Compression of Deep Convolutional Neural Networks for Fast and Low Power Mobile Applications. arXiv 2015, arXiv:1511.06530. [Google Scholar]
  41. Gusak, J.; Kholiavchenko, M.; Ponomarev, E.; Markeeva, L.; Blagoveschensky, P.; Cichocki, A.; Oseledets, I. Automated Multi-Stage Compression of Neural Networks. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 2501–2508. [Google Scholar]
  42. Xu, X.; Li, M.; Tao, C.; Shen, T.; Cheng, R.; Li, J.; Xu, C.; Tao, D.; Zhou, T. A Survey on Knowledge Distillation of Large Language Models. arXiv 2024, arXiv:2402.13116. [Google Scholar]
  43. Touvron, H.; Martin, L.; Stone, K.; Albert, P.; Almahairi, A.; Babaei, Y.; Bashlykov, N.; Batra, S.; Bhargava, P.; Bhosale, S.; et al. Llama 2: Open Foundation and Fine-Tuned Chat Models. arXiv 2023, arXiv:2307.09288. [Google Scholar]
  44. Zhao, W.X.; Zhou, K.; Li, J.; Tang, T.; Wang, X.; Hou, Y.; Min, Y.; Zhang, B.; Zhang, J.; Dong, Z.; et al. A Survey of Large Language Models. arXiv 2023, arXiv:2303.18223. [Google Scholar]
  45. Zhang, Y.; Xiang, T.; Hospedales, T.M.; Lu, H. Deep Mutual Learning. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 4320–4328. [Google Scholar]
  46. Milletari, F.; Navab, N.; Ahmadi, S.A. V-Net: Fully Convolutional Neural Networks for Volumetric Medical Image Segmentation. In Proceedings of the 2016 Fourth International Conference on 3D Vision (3DV), Stanford, CA, USA, 25–28 October 2016; pp. 565–571. [Google Scholar]
  47. Hu, J.; Shen, L.; Sun, G. Squeeze-and-Excitation Networks. In Proceedings of the 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–22 June 2018; pp. 7132–7141. [Google Scholar]
  48. Hou, Q.; Zhou, D.; Feng, J. Coordinate Attention for Efficient Mobile Network Design. In Proceedings of the 2021 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Nashville, TN, USA, 19–25 June 2021; pp. 13708–13717. [Google Scholar]
  49. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric Contextual Modulation for Infrared Small Target Detection. In Proceedings of the 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Waikoloa, HI, USA, 3–8 January 2021; pp. 949–958. [Google Scholar]
  50. Wang, A.; Li, W.; Huang, Z.; Wu, X.; Jie, F.; Tao, R. Prior-Guided Data Augmentation for Infrared Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 10027–10040. [Google Scholar] [CrossRef]
  51. Wang, K.; Du, S.; Liu, C.; Cao, Z. Interior Attention-Aware Network for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  52. Kou, R.; Wang, C.; Yu, Y.; Peng, Z.; Yang, M.; Huang, F.; Fu, Q. LW-IRSTNet: Lightweight Infrared Small Target Segmentation Network and Application Deployment. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
Figure 1. The overall structure of KDD, where Output-S is the result of the student model, Output-T is the result of the teacher, gt-pre is the result of feature fusion, and Mask is the ground truth.
Figure 1. The overall structure of KDD, where Output-S is the result of the student model, Output-T is the result of the teacher, gt-pre is the result of feature fusion, and Mask is the ground truth.
Remotesensing 16 03173 g001
Figure 2. The knowledge distillation process.
Figure 2. The knowledge distillation process.
Remotesensing 16 03173 g002
Figure 3. The schematic diagrams of ELA. The feature maps were subjected to one-dimensional convolution and normalization operations at different spatial scales, followed by combining the two using multiplication. Finally, the result was output using convolution.
Figure 3. The schematic diagrams of ELA. The feature maps were subjected to one-dimensional convolution and normalization operations at different spatial scales, followed by combining the two using multiplication. Finally, the result was output using convolution.
Remotesensing 16 03173 g003
Figure 4. The schematic diagrams of GAB, which was processed by grouping different level features.
Figure 4. The schematic diagrams of GAB, which was processed by grouping different level features.
Remotesensing 16 03173 g004
Figure 5. The target detection results obtained based on the SIRST dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Figure 5. The target detection results obtained based on the SIRST dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Remotesensing 16 03173 g005
Figure 6. The target detection results obtained based on the IRSTD-1k dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Figure 6. The target detection results obtained based on the IRSTD-1k dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Remotesensing 16 03173 g006
Figure 7. The target detection results obtained based on the MSISTD dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Figure 7. The target detection results obtained based on the MSISTD dataset. (a) Input image, (b) New Tophat, (c) NRAM, (d) ACM, (e) DNANet, (f) IAANet, (g) AGPCNet, (h) LW-IRSTNet, (i) KDD, (j) Ground truth.
Remotesensing 16 03173 g007
Figure 8. Performance of the ROC curves on different datasets.
Figure 8. Performance of the ROC curves on different datasets.
Remotesensing 16 03173 g008
Table 1. The p r e c i s i o n , r e c a l l , F 1 , I o U , and A U C for different methods on the three datasets.
Table 1. The p r e c i s i o n , r e c a l l , F 1 , I o U , and A U C for different methods on the three datasets.
SIRSTIRSTD-1kMSISTD
PrecisionRecallF1IoUAUCPrecisionRecallF1IoUAUCPrecisionRecallF1IoUAUC
New Tophat0.50980.69680.52180.22600.82840.22650.62630.27020.05350.66350.23410.52370.27020.04830.5861
NRAM0.85180.61420.68270.48310.79700.51990.35540.35870.15490.58670.55860.32700.36810.10400.5553
ACM0.90950.68950.76770.49210.93670.60750.75890.64900.58210.89080.78080.56430.62740.53120.8671
DNANet0.95060.64510.75490.59920.94240.88970.63700.71990.61720.90580.86300.70100.75220.65540.9159
IAANet0.85410.60830.67740.50970.94590.65670.78150.67630.50380.91160.81370.67210.69920.55520.9500
AGPCNet0.69430.49680.53960.44650.85310.48360.49830.46220.39100.81590.69610.43270.50420.39620.7836
LW-IRSTNet0.94060.62170.72960.61090.98090.66850.65750.62410.50390.93880.86200.50630.60890.49310.8660
KDD0.86410.76470.79490.66710.96010.72140.76910.71640.63590.93640.76160.79640.75710.66880.9261
Table 2. Ablation study of the ELA, GAB, and loss in different evaluation metrics. ✓and ✗ indicate that the module exists or does not exist respectively. The red number represents the highest value in each evaluation metric.
Table 2. Ablation study of the ELA, GAB, and loss in different evaluation metrics. ✓and ✗ indicate that the module exists or does not exist respectively. The red number represents the highest value in each evaluation metric.
DatasetELAGABLossPrecisionRecallF1IoUAUC
SIRST0.83910.77750.78490.64560.9501
0.79570.8264 0.78830.65670.9721
0.85540.73610.77650.66200.9659
0.8641 0.76470.7949 0.6671 0.9601
IRSTD-1k0.71350.78420.70010.57850.9042
0.62160.8118 0.66730.55400.9158
0.70480.76590.70570.59660.9007
0.7214 0.76910.7164 0.6359 0.9364
Table 3. The Params and FLOPs of the different methods. The red number represents the highest value in each evaluation metric.
Table 3. The Params and FLOPs of the different methods. The red number represents the highest value in each evaluation metric.
Params (M)FLOPs (G)FPS (Hz)
New Tophat28
NRAM0.6
ACM0.390.2222
DNANet2.618.5936
IAANet14.05408.9016
AGPCNet12.3633.068
LW-IRSTNet0.160.23218
KDD0.0410.056 41
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Tang, W.; Dai, Q.; Hao, F. An Efficient Knowledge Distillation-Based Detection Method for Infrared Small Targets. Remote Sens. 2024, 16, 3173. https://doi.org/10.3390/rs16173173

AMA Style

Tang W, Dai Q, Hao F. An Efficient Knowledge Distillation-Based Detection Method for Infrared Small Targets. Remote Sensing. 2024; 16(17):3173. https://doi.org/10.3390/rs16173173

Chicago/Turabian Style

Tang, Wenjuan, Qun Dai, and Fan Hao. 2024. "An Efficient Knowledge Distillation-Based Detection Method for Infrared Small Targets" Remote Sensing 16, no. 17: 3173. https://doi.org/10.3390/rs16173173

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop