Next Article in Journal
Deformation Behavior and Reactivation Mechanism of the Dandu Ancient Landslide Triggered by Seasonal Rainfall: A Case Study from the East Tibetan Plateau, China
Previous Article in Journal
A Modified Iteration-Free SPGA Based on Removing the Linear Phase
Previous Article in Special Issue
On-Board Geometric Rectification for Micro-Satellite Based on Lightweight Feature Database
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection

1
Science and Technology on Multi-Spectral Information Processing Laboratory, School of Electronic Information and Communications, Huazhong University of Science and Technology, Wuhan 430074, China
2
China Academy of Space Technology (Xi’an), Xi’an 710100, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(23), 5539; https://doi.org/10.3390/rs15235539
Submission received: 12 September 2023 / Revised: 10 November 2023 / Accepted: 26 November 2023 / Published: 28 November 2023

Abstract

:
Infrared small target detection technology is widely used in infrared search and tracking, infrared precision guidance, low and slow small aircraft detection, and other projects. Its detection ability is very important in terms of finding unknown targets as early as possible, warning in time, and allowing for enough response time for the security system. This paper combines the target characteristics of low-resolution infrared small target images and studies the infrared small target detection method under a complex background based on the attention mechanism. The main contents of this paper are as follows: (1) by sorting through and expanding the existing datasets, we construct a single-frame low-resolution infrared small target (SLR-IRST) dataset and evaluate the existing datasets on three aspects—target number, target category, and target size; (2) to improve the pixel-level metrics of low-resolution infrared small target detection, we propose a small target detection network with two stages and a corresponding method. Regarding the SLR-IRST dataset, the proposed method is superior to the existing methods in terms of pixel-level metrics and target-level metrics and has certain advantages in model processing speed.

1. Introduction

Compared with visible light imaging detection and active radar imaging detection, infrared imaging detection technology has the following characteristics [1]: unaffected by light conditions, works in all types of weather, works passively, has high imaging spatial resolution, adapts to various environments, has strong anti-electromagnetic interference ability, has a simple structure, is small in size, and easy to carry and hide. Benefiting from the above advantages, infrared detection and imaging technology has been widely used in infrared search and tracking, infrared precise guidance, low and slow small aircraft detection and identification, and other projects [2].
In some cases that need to be pre-judged, the target to be detected is far away from the infrared detection imaging system, and the image shows a dim and small target, often lacking texture information. The targets to be detected are usually aircrafts, drones, missiles, ships, vehicles, and other fast-moving objects [3,4], so the outlines of the imaging targets are fuzzy. In addition, as they are affected by the surrounding environment and detection equipment, small infrared targets are easily submerged in noise and complex backgrounds [5]. All these factors bring challenges to infrared small target detection.
At present, there are many infrared detection devices with low imaging resolution that are applied in various fields [6]. Therefore, it is of practical significance to design a method for small target detection in low-resolution infrared images to improve the small target detection performance. The number of pixels occupied by the target in the low-resolution infrared small target image is low [7], and a more accurate prediction of each pixel of the small target (that is, improving the pixel-level metrics of the low-resolution infrared image small target detection) can significantly improve the target detection performance.
Research on infrared small target detection is divided into single-frame image target detection and multi-frame image target detection [2]. This paper focuses on the former. Early researchers mainly proposed model-driven methods. Filter-based methods [8,9] require determining the filtering template in advance based on the structural characteristics of the image, so it has poor adaptability to complex background environments. Methods [10,11,12,13] based on local contrast are suitable for situations where there is a significant difference in grayscale between the target and surrounding background, but they are prone to missed detections and misjudgments. Low-rank-based [14,15] and tensor-based [16,17,18] methods can achieve good results, but the computational cost is high, and hyperparameters are more sensitive to image scenes.
With the development of deep learning, some data-driven methods and infrared small target datasets [7,19,20,21,22] have emerged in recent years. Considering the weak and small characteristics of infrared small targets, infrared small target detection is usually modeled as a semantic segmentation problem. In order to ensure that the features of small targets are not submerged, some methods [7,19,22,23] have been used to enhance the fusion of features at different layers of the network. Based on the small proportion of small targets in the overall image, some methods [24,25] solve the problem with infrared small target detection by suppressing the background area to make the network pay more attention to the target area. There are also some studies [26,27,28,29] that consider how to improve and innovate based on classic encoding and decoding structures.
The existing single-frame infrared small target detection methods [7,23,24] have problems in terms of poor adaptability and high false-alarm and missed detection rates when they are used to detect infrared small target images with low resolution. This is not only because the quality of the existing dataset is not high, which leads to unsatisfactory training of the network, but also related to the large number of parameters in the existing network structure or the insufficient local attention to small targets.
Therefore, in view of the problems in the existing datasets, we propose corresponding improvement strategies. Then, we construct a single-frame low-resolution infrared small target image dataset with high quality and a large amount of data called SLR-IRST. At the same time, we design a central point-guided circular region local self-attention network (CCRANet) for low-resolution infrared image small target detection. The CCRANet detects small targets in the image in two stages from coarse to fine. In the coarse stage, multiple circular regions of interest of fixed size are obtained by a center point-guided circular region of interest suggestion (CCRS) module. In the fine stage, the local feature information of small targets is further extracted in a share params local self-attention (SPSA) module, and each pixel is predicted. In the SLR-IRST dataset, the CCRANet can significantly improve the target detection effect on small targets. Compared with the present methods, the pixel-level metric, IoU, is improved by about 3%, the nIoU is improved by about 6%, and the target-level metric probability of detection (Pd) is improved by about 6%.

2. Related Works

2.1. Infrared Small Target Datasets

The Society of Photo-Optical Instrumentation Engineers (SPIE) defines infrared small targets as having a total spatial extent of less than 81 pixels (9 × 9) in a 256 × 256 image [30]—that is, the proportion of small targets in the entire image is less than 0.12%. In addition, the size of small infrared targets varies greatly, ranging from only one pixel (i.e., dot target) to dozens of pixels (i.e., expanded target) [29].
In recent years, some scholars have done a lot of work on the collection and production of infrared small target datasets and have publicly released these datasets, which include single-frame datasets [7,19,20,21,22] (see Table 1) and multi-frame datasets [31,32,33] (see Table 2).
In Table 1 and Table 2, it can be seen that the sample size of real single-frame infrared small target data is relatively small, but the sample size of multi-frame infrared small target data is rich, which can be used to expand single-frame data. Constructing a single-frame infrared small target dataset with a larger data volume and higher quality can promote the development of single-frame infrared small target detection.

2.2. Infrared Small Target Detection Methods

In recent years, deep learning has developed rapidly in terms of solving visual tasks such as image classification, object detection, and semantic segmentation. Some methods based on deep learning have also emerged for infrared small target detection.
Due to their “weak” and “small” characteristics, infrared small targets are easily overwhelmed by a network’s high-level features. However, if only low-level features are used, it is not possible to fully comprehend semantic information, making it easy to miss detection and raise false alarms. Therefore, some researchers have combined attention mechanisms to study methods for enhancing feature fusion at different layers. Dai et al. proposed a bottom-up channel attention modulation method [23] (ACM) to preserve and highlight infrared small target features in high-level layers. Thereafter, Dai et al. [19] modularized local contrast measurement methods [10] from traditional methods in the network to design a model-driven deep learning network (ALCNet). Li et al. [7] proposed the use of DNANet to achieve progressive information interaction between high-level and low-level features through densely nested modules (DNIM). Chen et al. [22] introduced the self-attention mechanism of a transformer into the designed IRSTFormer to extract multi-scale features from the input image through the overlapping block self-attention structure of the hierarchy.
Based on the small proportion of small targets in the overall image, some researchers solve the problem with infrared small target detection by suppressing the background area so that the network pays more attention to the target area. Wang et al. [24] proposed a coarse-fine two-stage network, IAANet. In the coarse stage, the candidate target regions are obtained by the region proposal network (RPN), and in the fine stage, the global features of all candidate target regions in the image are extracted by the attention encoder (AE). In IAANet, the hard decision method is used to suppress the background regions to the greatest extent. Cheng et al. [25] designed a supervised attention module trained by small target diffusion maps in the proposed LPNet to suppress most of the background pixels irrelevant to small target features in a soft decision manner.
It has been proved that the classical encoder-decoder structure can achieve better results in the semantic segmentation task [34], and some researchers have carried out work on the improvement and innovation of the classical codec structure. Tong et al. [26] proposed MSAFFNet, which introduced the EIFAM module containing edge information based on the codec structure and constructed multi-scale labels to focus on the details of target contour and internal features. Wu et al. [27] proposed UIU-Net (U-Net in U-Net). It imparts a tiny U-Net into a larger U-Net backbone network to realize multi-level and multi-scale representation learning of objects. Chen et al. [28] proposed a MultiTask-UNet (MTUNet) with both detection and segmentation heads. By sharing the backbone, the similar semantic features of the two tasks are fully utilized. Compared with the compound single-task model, MTUNet has fewer parameters and faster inference speed. Wu et al. [29] proposed a multi-level TransUNet (MTU-Net). In the encoder, the features extracted by convolution are passed through a multi-level ViT module (MVTM) to capture remote dependencies.
The networks that combine the attention mechanism and multi-scale feature fusion [7,19,22,23] enhance the network’s ability to extract image features, but the local attention to small targets in the image is not enough and the network’s ability to detect small targets is not enough to be improved. The networks that focus on the localized region of small targets [24,25] have a problem: the number of network parameters and computational amount are larger, and the network prediction speed is slower.

2.3. Evaluation Metrics

The output of the infrared small target detection network is pixel-level segmentation. Therefore, it is common to use semantic segmentation metrics to evaluate network performance, such as precision, recall, F1 score, ROC curve, and PR curve.
Precision and recall refer to the proportion of correctly predicted positive samples out of all predicted positive samples and all true positive samples, respectively. The F1-Score is the harmonic mean of precision and recall. The definitions of P (precision), R (recall), and F1 − P (F1 score) are as follows:
P = T P T P + F P R = T P T P + F N F 1 P = 2 P R P + R
where T, P, TP, FP, and FN denote the true, positive, true positive, false positive, and false negative, respectively.
The receiver operating characteristic (ROC) curve shows how the model performs across all classification thresholds. The horizontal coordinate of the ROC curve is the false positive rate (FPR), and the vertical coordinate is the true positive rate (TPR). It goes through the points (0, 0) and (1, 1). The horizontal coordinate of the precision-recall (PR) curve is the recall rate, which reflects the classifier’s ability to cover positive examples. The vertical coordinate is the precision, which reflects the accuracy of the classifier’s prediction of positive examples.
However, as a target detection task, some researchers have proposed pixel-level and target-level evaluation metrics based on existing metrics to better evaluate the detection performance of infrared small targets.
IoU and nIoU are pixel-level metrics. IoU represents the ratio of intersection and union between the predicted and true results:
IoU = T P T + P T P
nIoU [19] is the numerical result normalized by the IoU value of each target, as shown in (3), where N represents the total number of targets.
nIoU = 1 N i N T P [ i ] T [ i ] + P [ i ] T P [ i ]
Pd (probability of detection) and Fa (false-alarm rate) are target-level metrics [22]. Pd measures the ratio of the number of correctly predicted targets to the number of all targets. Fa measures the ratio of incorrectly predicted pixels to all pixels in the image.
P d = #   num   of   true   detections #   num   of   actual   targets
F a = #   num   of   false   predicted   pixels #   num   of   all   pixels

3. SLR-IRST Dataset

This section introduces the construction process and statistical characteristics of the single-frame low-resolution infrared small target image dataset (SLR-IRST).

3.1. Construction of the SLR-IRST Dataset

Unlike traditional model-driven methods, deep learning is a data-driven method. It requires a large amount of diverse data training to improve the generalization ability of the network [35]. At present, few real single-frame infrared small target images have been published publicly [19,20], and the resolution of the images is inconsistent. The published synthetic single-frame infrared small target images have high similarity and low synthesis quality [21,22]. All these problems affect the training coupling process and the final performance of the network [7]. Therefore, we constructed a single-frame low-resolution infrared small target image dataset (SLR-IRST) based on the existing infrared small target datasets and other infrared datasets through data collation and data expansion.
The generation process of the SLR-IRST dataset is as follows (as shown in Figure 1).
1.
Data Source: The data in the SLR-IRST dataset are from a single-frame infrared small target dataset (SIRST, IRSTD-1k, IRST640, SIRST-v2 [36]) and a multi-frame infrared small target dataset (IRDST [25]). There is a serious shortage of maritime target data in the existing dataset (see Table 3), so some waterborne target images were extended in the SLR-IRST dataset from this infrared dataset (http://iray.iraytek.com:7813/apply/Sea_shipping.html/ (accessed on 10 December 2021)).
2.
Data Cleaning: The collected images were first data cleaned. The images that did not meet the small target definition [30] or had obvious synthetic traces were removed. The frequency of image extraction in the multi-frame dataset, IRDST, is 1 for every 50 images. In order to ensure the diversity and balance of the data in the dataset, the structure similarity index measure (SSIM) value [37] was used to evaluate the similarity between the images. Highly similar images with SSIM values greater than 0.85 (for simple backgrounds) and 0.90 (for complex backgrounds) were discarded.
3.
Image Scaling: The resolution of the images in the existing infrared small target dataset is inconsistent. Under normal circumstances, the images in the dataset would be directly resized to the specified resolution before training and testing [19]. However, the direct resize operation causes the target to deform and the label to no longer be binarized, which introduces additional errors when network testing. Therefore, the resolution of images in SLR-IRST was unified by an undistorted method [38]. Zero was filled below or to the right of the original image to make it match the aspect ratio of the specified resolution, and then resized to the specified resolution. By referring to the common resolution of infrared detection equipment and parameters of other datasets [32], the unified resolution of images in SLR-IRST was 256 × 256.
4.
Data Label: The images, after the unified resolution, were re-labeled. The small target in the infrared image was fuzzy and the edge was difficult to define. The Canny function [39] helps to define the boundaries of small targets and reduce manual labeling errors. The SLR-IRST dataset has pixel-level labels, bounding box labels, and center point labels. The boundary box label and center point label are determined by calculating the boundary and center point of the pixel-level label.
5.
Data Check: Finally, all the marked data was checked again to ensure that the labels are correct and the data format is correct.
6.
Divide Dataset: When dividing the dataset, all the synthesized data was divided in the training set. All real data was evenly divided into training, validation, and test sets depending on the source of the data.

3.2. Construction and Characterization of the SLR-IRST Dataset

Table 4 presents some basic statistics of the SLR-IRST dataset and other mainstream infrared small target datasets. In addition, all single-frame infrared small target datasets are evaluated in terms of target number, target category, target size, etc.
The SLR-IRST dataset has the following characteristics.
  • More data: In Table 4, it can be seen that the SLR-IRST dataset has a much larger amount of data (2689 images, 3586 targets). As can be seen in Table 1, the scene types and target types in the SLR-IRST dataset are more abundant.
  • Smaller target: As can be seen in Table 4, targets in the SLR-IRST dataset are smaller, occupying only 15.48 pixels on average and sized 3.33 × 4.58 on average. As can be seen in Figure 2a, compared with other datasets, more targets in the SLR-IRST dataset accounted for less than 0.005% of the total image.
  • More balanced distribution of data categories: Figure 2b shows the statistical distribution of targets in different datasets against the real categories. The distribution of target categories in the SLR-IRST dataset is more balanced. The sample size of the waterborne targets was expanded from 87 (8.7%) in IRSTD-1k to 313 (11.3%) in SLR-IRST, which is a nearly three-fold increase. This greatly alleviates the problem of sample scarcity in the waterborne target category.
In addition, as can be seen in Figure 2c, images with no more than 3 small targets account for more than 98% of all datasets. It shows that the small target in the infrared image has sparsity. As can be seen in Figure 2a, the small target in the infrared image has more targets than 0.03% of the image. It indicates that the small target in the infrared image is very small.
The SCR value [14] is the normalized value of the difference between the gray value of the target and the surrounding background area. The SCR value can be used to describe the difficulty of infrared small target detection. The larger the SCR value, the easier the target is to detect, and the smaller the SCR value, the harder it is to detect. As seen in Table 4, the average SCR value across all datasets is less than 10. It indicates that there is a large number of weak targets that are difficult to detect in the infrared small target dataset—that is, the small targets in the infrared image are weak.
We selected state-of-the-art (SOTA) networks, DNANet and ALCNet, and compared their IoU metric results in the SLR-IRST dataset and the SIRST dataset. As shown in Table 5, both DNANet and ALCNet showed a large degree of performance degradation in the SLR-IRST dataset. This shows that the SLR-IRST dataset presents a new challenge. It is necessary to design a network according to the characteristics of low-resolution infrared small target images to improve the detection ability of the network.

4. Methodology

This section introduces specific information on the proposed CCRANet. First, Section 4.1 introduces the overall architecture of the CCRANet. Then, Section 4.2–4.4 introduce the specific structure of the network in detail.

4.1. Overall Architecture

Small targets in low-resolution infrared images are small in size, faint, and sparse. Combining these characteristics, we designed a central point-guided circular region local self-attention network (CCRANet) suitable for low-resolution infrared small target detection.
As shown in Figure 3, the CCRANet consists of a center point-guided circular region of interest suggestion (CCRS) module, a U-shaped feature extraction (U-FE) module, and a share params Local self-attention (SPSA) module. The entire working process is as follows: for the input image, X H × W , firstly, multiple central points of interest in the image are generated by the CCRS module, and the position information b of the circular regions of interest guided by the central points is generated. The non-interest region is the output as the background in the prediction result. At the same time, feature extraction is carried out on the input image through the U-FE module to obtain the feature map, X ^ . By combining b and X ^ , the local feature information, x, of multiple regions of interest is obtained. It will be fed into the SPSA module for self-attention calculations within each local region to obtain the feature map, x ^ . Finally, the feature map, x ^ , is used to predict the regions of interest using multi-layer perceptron (MLP). The three modules in the network are further described below.

4.2. Center Point-Guided Circular Region of Interest Suggestion Module

The small target in the infrared image only accounts for a very small part of the image, and most of the regions in the infrared image are redundant and interference information. At the same time, because of the long detection distance, the shapes of small targets in infrared images are almost round. Therefore, we designed a center point-guided circular region of interest suggestion (CCRS) module to obtain multiple centers of interest points in the image through a simple network. Under the guidance of these central points, multiple circular regions of interest are generated. Regions of non-interest are directly considered as the background.
In CCRS, the input image is progressively reduced by multiple convolutional layers (MCL) to obtain the feature map, x d h × w × c . xd has a step size of 16 relative to the input image. As shown in Figure 4, MCL has a similar structure to ResNet18 [40] and uses ResNet18’s pre-training weight information. Then, after a convolution with a kernel size of one, the number of channels of xd is reduced to three. Let (cx, cy) represent the spatial position of a point in the feature map, xd. s represents the subsampling multiple of the feature map, xd, corresponding to the original image. tx, ty, and to represent the three parameters corresponding to the three channels in the feature map, xd. The calculation method for the location information and confidence of the center point predicted by the network is as follows:
c ^ x = ( 2 σ ( t x ) 0.5 + c x ) × s c ^ y = ( 2 σ ( t y ) 0.5 + c y ) × s o ^ = σ ( t o )
where c ^ represents the predicted center point. ( c ^ x , c ^ y ) and o ^ represent the coordinates and confidence of the predicted central point, respectively. σ stands for the sigmoid function. For an input image with a resolution of 256 × 256, 256 predicted centers of interest will be generated.
In the CCRS module, there is a confidence threshold, conf. The center point with confidence above conf is considered as the interest center point that is needed. At the beginning of training, the number of retained centers of interest is variable. It may be the case that a large number of central points are retained or that no central points are retained. So, the CCRS module sets an additional hyperparameter, k, which represents the maximum number of retained center points of interest. When the number of center points higher than conf exceeds k or is 0, the first k center points with higher confidence are retained. In addition, there is another hyperparameter, l, which is the radius of the circular region of interest generated based on the central point guidance. The central coordinate information of the circular regions of interest generated by the CCRS is expressed as b b o x n × 2   ( 0 n k ) .

4.3. U-Shape Feature Extraction Module

The U-shape feature extraction (U-FE) module is parallel to the CCRS module, and its overall structure is U-shaped, which is similar to that of the U-Net network [34]. In the U-FE module, the downsampling times are reduced to two times to prevent the feature information of the small target being drowned in the deep feature. The downsampling operation is implemented using MaxPool, and the upsampling operation is implemented using ConvTranspose. As shown in Figure 5, for the input infrared image, X H × W , after passing through the U-shaped subnetwork, the feature map, X ^ H × W × D , is obtained. Among them, D = 64 .

4.4. Share Params Local Self-Attention Module

According to the output, b b o x n × 2 and X ^ H × W × D , of the first two stages, the characteristic information, x i π l 2 × D ( i = 1 , 2 , , n ) , of each circular region of interest is obtained first (see Figure 6). Then, all xi is spliced together along the new dimension to obtain x k × π l 2 × D , which is the feature information of all circular regions of interest in the image. (When n < k, the data size is kept consistent by filling 0 ( x 0 O π l 2 × D )).
x = concat ( x 1 + x 2 + + x n + x 0 + k )   ( 0 < n k )
The SPSA module uses the self-attention mechanism to further extract features from the regions of interest because the self-attention mechanism has a global receptive field and is better able to capture the internal correlation between features. The traditional self-attention computing process is proportional to the square of the length of the input feature sequence [41,42]. Considering that there are many feature sequences for self-attention calculation, the calculation amount is large. The SPSA adds scaling down operations to the traditional multi-head attention layer (MHA).
As shown in Figure 6, the feature sequence, x, is first restored to the form of a two-dimensional feature map along the spatial dimension. Since the region of interest is circular, when restoring along the spatial dimension, 0 is added around the circular region to make it square, and the feature map, x K k × 2 l × 2 l × D , is obtained. The scaling operation is carried out in the spatial dimension through a layer of convolution. Let the scaling reduction factor be s—that is, the kernel size and step size of the convolution are both s. The feature map, x s k × 2 l / s × 2 l / s × D , is obtained by scaling down the feature map, xK. The feature sequence, x K k × N × D   ( N = 4 l 2 ) , x s k × N × D   ( N = 4 l 2 / s 2 ) , is obtained by flattening xK, xs along the spatial dimension. The keyword (K) and the value (V) are generated from the scaled feature sequence, x s , through two sets of linear projections. The query (Q) is generated from the unscaled feature sequence, x , through a set of linear projections.
Q = Linear ( Flatten ( x K ) ) K = Linear ( Flatten ( Conv s   ( x K ) ) ) V = Linear ( Flatten ( Conv s   ( x K ) ) )
where, Flatten represents flattening feature maps along the spatial dimension, Convs represents a convolution of both kernel size and step size with s, and Linear represents a linear layer.
As shown in Figure 7, the calculation of multi-head self-attention is as follows:
MultiHead ( Q , K , V ) = concat ( head 1 , , head h ) W O where   head i = Attention ( Q i , K i , V i ) = softmax ( Q i K i T d k ) V i
and where W O D × D , Q i k × N × d , K i k × N × d , V i k × N × d , and d = D / h . h represents the number of heads. The scaling operation does not change the length and dimension of the output feature sequence but only reduces the computation amount in the process of self-attention calculation.
Other visual task algorithms [24,43,44] usually splice the features in the same image along the spatial dimension because it is believed that the features in the same image are related and influence each other. It is worth noting that in the SPSA, feature splicing of multiple regions of interest in the same image is performed in a new dimension. In this way, self-attention calculation is only confined to the interior of each region of interest, and no self-attention calculation is performed among features of different regions. We believe that the infrared small target task has different characteristics from other vision tasks. The infrared small targets in images are small and sparse and have obvious locality. The features of other regions in the image have little relevance to the prediction of small targets and are redundant information. Therefore, it is unnecessary to concatenate the features in the infrared small target image along the spatial dimension, as it will increase the amount of calculation. In addition, in the SPSA, the process of calculating self-attention for different regions of interest is parallel, and the parameters are shared. Since there is no need to calculate the self-attention between different areas of interest, the amount of computation and the number of parameters in the process of self-attention calculation are greatly reduced.
For k input regions of interest with radius l and number of channels D, the computational complexity of the calculation using the SPSA, or the standard multi-head self-attention mechanism (TE) are as follows:
Ω ( T E ) = 2 k 2 π 2 l 4 D + 4 k π l 2 D 2
Ω ( S P S A ) = 8 k l 4 D s 2 + 4 k l 2 D 2 s 2
Of the two items that make up computational complexity, the former contributes the vast majority of the computational effort. Using the SPSA for self-attention computation can greatly reduce the computation of the former and speed up the reasoning speed of the network.

5. Experiments

In this section, the experimental setup is introduced first, including implementation details and evaluation metrics. Then, the ablation experiments on the loss function, hyperparameter setting, and network structure design are shown to verify the rationality of the network design. Then, the proposed CCRANet is compared with other state-of-the-art (SOTA) methods to verify the effectiveness of the network. Finally, the CCRANet and other SOTA methods are compared in actually collected infrared small target images to show that the network has better generalization ability.

5.1. Experimental Setting

5.1.1. Implementation Details

In the network training stage, the optimizer is the adaptive gradient (AdaGrad) [45] method, and the method to initialize the weight and bias of the model is Xavier [46]. The CCRS module has a learning rate of 0.01, and the rest of the network has a learning rate of 0.001. The first epoch of network training trains only the CCRS module. The batch size and epoch are set to 16 and 500, respectively. All networks rely on PyTorch [47]. The computer used in the experiment has an Intel(R) Core (TM) i9-10920X @ 3.50 GHz CPU and a Nvidia RTX 3090 GPU. All experiments were trained and tested on the SLR-IRST dataset, and the number of images in the training set, verification set, and test set were 1616, 533, and 540, respectively, with a partition ratio of about 6:2:2. The SLR-IRST dataset and CCRANet will be made public on https://github.com/wangwjinggg/CCRANet.
Before the training began, data enhancement was performed on the training data. The input images were scaled at a random scale. The aspect ratio of the scaled image ranged from 0.58 to 1.85. The length of the longer side of the scaled image ranged from 0.5 to 2 times the original width and height. The scaled image was uniformly and arbitrarily cropped to a 256 × 256 size. In addition to this, data enhancement strategies also include random horizontal flipping and random superimposed Gaussian noise.

5.1.2. Evaluation Metrics

Infrared small target detection networks based on deep learning [7] usually use both pixel-level and target-level metrics to comprehensively evaluate the performance of the network.
Previously, the false-alarm rate (Fa) of the target-level metrics (see Formula (5)) represented the ratio of the wrongly predicted pixels to all pixels of the image and did not really measure the false-alarm situation of the network from the target-level. Therefore, we redefined the Fa as follows:
F a = #   num   of   false   detections #   num   of   predicted   targets
In addition, we also defined the Md (miss detection rate), the Pa (probability of accuracy), and the F1 − T (F1 score) of the target-level evaluation metrics:
M d = #   num   of   not   detections #   num   of   actual   targets
P a = #   num   of   true   detections #   num   of   predicted   targets
F 1 T = 2 P d P t P d + P t
where Pd represents the probability of detection (see Formula (4)). When the center of the predicted target is within d pixels of the center of the real target, the target is considered to be correctly detected. The threshold d is set to two in this paper. The F1 − T takes both recall and accuracy into account to measure target-level prediction. A method is considered good when high values are obtained on Pd (↑), Pa (↑), F1 − T (↑), and low values are obtained on Md (↓), Fa (↓).
The pixel-level metrics used for evaluation include IoU, nIoU, R (Recall), P (Precision), and F1 − P scores. In addition, the receiver operating characteristic (ROC) and precision-recall (PR) curves are used to further demonstrate the overall detection effect, target detection capability, and background suppression capability.

5.2. Loss Function Experiments

Loss in the CCRANet consists of two parts including the loss of regression learning from the predicted center point in the CCRS module and the loss of regression learning from the predicted pixel-level segmentation results. The regression learning of the predicted center point includes two parts: location information learning and confidence learning. The BCEWithLogits loss function is used to monitor the learning of confidence and the L1 loss function to monitor the learning of location information. The metrics results for monitoring learning of pixel-level segmentation results with four different loss functions are compared in this section, including BCEWithLogits, Soft-IoU [19], Focal [48], and Mixed. The Mixed loss function is defined as follows:
M i x e d = BCEWithLogits + Soft - IoU
In order to suppress the imbalance of positive and negative samples, the balance coefficient is set when using the BCEWithLogits loss function. Among them, the balance coefficient is set to a = 10 in central point regression learning and a = 3 in semantic segmentation learning.
It can be seen from Table 6 that BCE obtained better index results. It shows that the BCEWithLogits loss function, after the balanced adjustment of positive and negative sample proportions, has a better monitoring ability in terms of binary classification tasks. Based on the above experimental results, the loss function used in the CCRANet training is as follows:
= ( Mixed ) for   seg + 0.5 × ( L 1 + BCEWithLogits ) for   center

5.3. Hyperparameter Experiments

This section conducts comparative experiments on the values of the two hyperparameters k and l in the CCRS module, and the experimental results are compared in Table 7.
It can be seen that the metrics results are the best when k = 5 and l = 10. When k = 3, the metrics drop slightly, and when k = 10, they drop sharply. This is because when the number of circular regions of interest is too large, some background regions containing suspected target characteristics are also preserved. This introduces interference information, which causes the precision (P) and Pa of the network to decrease significantly. Metrics at l = 5 and l = 20 both decreased compared with those at l = 10. This indicates that when the circular region of interest is too small, some small targets are not fully wrapped, resulting in a significant decrease in the network pixel-level recall rate (R) metric. When the circular region of interest is too large, it contains more interference information, resulting in a significant decrease in the target-level probability of accuracy (Pa) metric of the network. To sum up, the values of the two hyperparameters in the CCRS module are set to k = 5 and l = 10.

5.4. Ablation Study

Some ablation studies were conducted to verify the rationality and validity of the CCRANet in terms of network structure. As shown in Table 8, the proposed CCRANet is used as the benchmark network, denoted as A. Based on A, the U-FE module is replaced with the FPN [49] structure, denoted as B1. The number of feature map channels output in B1 is set to 64, which is consistent with that in A. On the basis of A, the U-FE module is replaced with a 7-layer fully convolution FCN structure without downsampling, denoted as B2. The 7-layer convolution gradually increases the number of channels of the feature to 256, then maintains the 3-layer convolution, and then gradually reduces the number of channels to 64, which is consistent with the output in A.
Based on A, the SPSA module was replaced with traditional self-attention (TSA) [41,42], denoted as C1. In TSA, the feature sequences of multiple circular regions of interest are concatenated along spatial dimensions. Moreover, there is no scaling down operation through convolution in the multi-head self-attention of TSA. The self-attention calculation in TSA is performed among the features of all the circular regions of interest in the image. Based on A, the SPSA module was replaced with a 4-layer convolution (CONV) structure, denoted as C2. The CONV convolution has a kernel size of 3 × 3, step size of 2, and padding of 1 and does not change the number of channels of the original feature map. In keeping with the residual connection structure in the SPSA, a residual link is also added here for each layer of convolution. The metrics of the ablation experiment are shown in Table 9.
Compared with the pixel-level metrics results of A, B1, and B2 in Table 9, the U-FE module obtained the best results. The U-FE module performs more convolution operations on shallow features and the feature maps at shallow layers provide more detailed information, which makes U-FE perform better than FPN. The U-FE module performs two downsampling operations and the receptive field of the convolution kernel is larger and more global features are extracted, which makes U-FE perform better than FCN. The two sets of experiments show that both shallow and deep features are important for small target detection. More downsampling will drown the features of small targets in the deep network. Without downsampling, the convolution receptive field will be limited and semantic feature information will be missing. Therefore, in order to obtain a better detection effect, it is necessary to select a suitable number of downsampling operations.
Compared with the pixel-level metrics results of A, C1, and C2 in Table 9, the SPSA module obtained the best results. Compared to the SPSA, C1 using TSA decreased by about 2% and C2 using CONV decreased by about 7%. The superiority of the SPSA over TSA indicates the following. Firstly, the scaling down operation in multi-head self-attention does not result in performance loss. Secondly, the features of different regions of interest are mutually interfering information, and calculating self-attention locally in the region of interest can avoid interference from redundant features. The sharp decline of CONV indicates that compared with the convolution of the local receptive field, the global receptive field of self-attention can better excavate the internal correlation between features.
Comparing the results of the target-level metrics in Table 9, A achieves the optimum for Pd and Md and sub-optimum for Pa, Fa, and F1 − T. There is an interesting experimental result: the B1 network using FPN is poor in pixel-level metrics, but it is optimal in target-level metrics, Pa and Fa. We analyze that this may be related to the operation of fusing multi-layer features simultaneously in FPN.

5.5. Comparison with SOTA Methods

To demonstrate the superiority of the proposed CCRANet, we performed a comparison to four other methods: the low-rank-based method IPI [14] and three deep learning methods—ALCNet [19], DNANet [7], and IAANet [24]. The metrics results of the contrast experiments are shown in Table 10.
As can be seen from Table 10, compared with the model-based method IPI, deep learning methods have better ability to cope with complex and diverse infrared images. In deep learning methods, the CCRANet achieves optimality in almost all metrics, and the improvement is obviously relative to the suboptimal network. In terms of pixel-level metrics, the CCRANet outperforms the suboptimal DNANet by about 5% on both nIoU and IoU and by about 3% on F1 − P. In terms of target-level metrics, compared to the second-best DNANet, the CCRANet improves Pd by about 6%, Pa by about 1%, and F1 − T by about 3%.
In order to comprehensively measure the pixel-level prediction ability of different deep learning methods, here the PR curve and ROC curve of the CCRANet and the other three methods (see Figure 8) were plotted. As seen in Figure 8, the CCRANet PR curve can cover the PR curve of other methods. This shows that the CCRANet’s prediction accuracy is stronger for positive samples (i.e., the pixels corresponding to the target region in the image). After the ROC curve stabilized, the recall metric of CCRANet is better, and it is always better than the other networks under different false-alarm rates.
The model processing speeds of the CCRANet and other three deep learning methods are compared in Table 11. The average speed of different methods on the test set of the SLR-IRST was calculated using the same data computing platform. As seen in Table 11, the CCRANet is slower than ALCNet in model processing but faster than the other two methods. Considering the improvement of the CCRANet in terms of prediction indicators, it can be considered that the CCRANet also has advantages in model processing speed.
We undertook a comprehensive comparison of pixel-level metrics, target-level metrics, and the results of model processing speed. It showed that the CCRANet can stably obtain a better infrared small target detection performance in the SLR-IRST dataset compared with comparison methods.
In order to observe the prediction abilities of different networks more clearly and intuitively, the visual prediction results are shown in Figure 9. Three scenarios are selected randomly from the test dataset, as shown in Figrue 9a. The corresponding labels are shown in Figure 9b. The visual prediction results of the CCRANet and other four methods are presented in Figure 9c–g. In Figure 9, the green, red, and yellow circles represent correctly detected targets, missed targets, and false targets, respectively. The red box in the lower left corner is a partial enlargement of the small target.
In order to further evaluate the prediction performance of the CCRANet for low-resolution infrared small target images, we collected some real infrared small target images. The backgrounds of the collected infrared images include the sky, clouds, buildings, trees, tower cranes, rivers, etc. Acquisition angles include looking up and looking down. Small targets in the infrared images include small drones, birds, and other creatures. These images and those in the SLR-IRST were acquired using different devices in different scenarios with data from different sources, and the correlation is low.
The single-frame collected infrared small target images, and the dataset (SAC-IRST) with a resolution of 256 × 256 was obtained after the filling and scaling of the images. It can further evaluate the generalization ability of different network training models. There are 80 images in the SAC-IRST, including 40 infrared images with targets and 40 infrared images without targets. There is interference information such as small targets in images without targets. This can be used to evaluate the ability of different methods to suppress false alarms. The comparative experimental results are shown in Table 12 and Table 13.
As can be seen in Table 12 and Table 13, the CCRANet has a significant advantage over other methods in terms of images with targets. The CCRANet also shows better suppression of false alarms due to untargeted images. The above experiments show that the CCRANet has good generalization ability.
In order to observe the prediction performance of different methods on actual acquired images more clearly, two images with targets and two images without targets are randomly selected (see Figure 10).
As can be seen from the above results, infrared images from different sources present new challenges to all methods. Our proposed CCRANet still performs better in the SAC-IRST dataset due to its excellent network structure. However, false alarms still exist. As shown by the third line in Figure 10, false alarms are still not avoided for some target-like point clouds. Additionally, the first line in Figure 10 shows that the prediction of targets in uneven grayscale is incomplete. The improvement of the prediction integrity of gray uneven targets is a direction that needs to be considered later.

6. Conclusions

In this paper, we constructed a high-quality and large low-resolution infrared small target dataset by sorting, improving, and expanding existing datasets. At the same time, based on the data characteristics of low-resolution infrared small target images, a coarse-to-fine two-stage target detection method, CCRANet, is constructed. We fully considered the small, weak, and sparse characteristics of infrared small targets and designed a local self-attention network structure with shared parameters, which improves the prediction results of the method and reduces the number of parameters and calculations undertaken by the network. To better evaluate the target-level prediction results of the methods used for infrared small targets, we also proposed two sets of complete target-level evaluation metrics to measure the target-level prediction abilities of the methods in an all-around way.
For the CCRANet method, there are still some directions for further research. In the CCRS module, only the regions of interest were reserved for further feature extraction and prediction. This hard decision method increased the influence of the CCRS module prediction performance on network results. Therefore, in future research, a soft decision method should be tried, the weight of the interest regions should be increased, and the weight of the non-interest regions should be reduced rather than simply discarded. In addition, the time information in the multi-frame sequence should be introduced to further enhance the accuracy of network prediction.

Author Contributions

Conceptualization, W.W. and C.X.; methodology, simulation and validation, W.W., C.X. and H.D.; data curation, W.W., G.Z., Y.H. and Z.C.; writing—original draft preparation, W.W.; writing—review and editing, W.W., C.X., H.D., R.L. and H.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the Qian Xuesen Youth Innovation Fund 2020-QXSQNCXJJ-01, and in part by the Young Star of Science and Technology in Shaanxi Province 2023KJXX-104.

Data Availability Statement

The SIRST, IRSTD-1k, IRST640, SIRST-v2, IRDST, and sea ship infrared image data used to support the research is available from the websites https://github.com/YimianDai/sirst, accessed on 29 July 2020; https://github.com/RuiZhang97/ISNet, accessed on 20 March 2022; https://github.com/jzchenriver/IRST640, accessed on 17 August 2022; https://github.com/YimianDai/open-sirst-v2, accessed on 21 February 2023; http://xzbai.buaa.edu.cn/datasets.html, accessed on 2 April 2023; and http://iray.iraytek.com:7813/apply/Sea_shipping.html/, accessed on 10 December 2021.

Acknowledgments

The authors would like to thank D.Y., Z.M., C.G. and S.H. for providing the data. And the authors would like to thank InfiRay for providing the data, too.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Levenson, E.; Lerch, P.; Martin, M.C. Infrared imaging: Synchrotrons vs. arrays, resolution vs. speed. Infrared Phys. Technol. 2006, 49, 45–52. [Google Scholar] [CrossRef]
  2. Zhao, M.; Li, W.; Li, L.; Hu, J.; Ma, P.; Tao, R. Single-frame infrared small-target detection: A survey. IEEE Trans. Geosci. Remote Sens. 2022, 10, 87–119. [Google Scholar] [CrossRef]
  3. Sun, X.; Guo, L.; Zhang, W.; Wang, Z.; Yu, Q. Small aerial target detection for airborne infrared detection systems using LightGBM and trajectory constraints. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 9959–9973. [Google Scholar] [CrossRef]
  4. Yang, P.; Dong, L.; Xu, W. Infrared small maritime target detection based on integrated target saliency measure. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 2369–2386. [Google Scholar] [CrossRef]
  5. Qi, S.; Ma, J.; Tao, C.; Yang, C.; Tian, J. A robust directional saliency-based method for infrared small-target detection under various complex backgrounds. IEEE Geosci. Remote Sens. Lett. 2012, 10, 495–499. [Google Scholar]
  6. Rogalski, A.; Martyniuk, P.; Kopytko, M. Challenges of small-pixel infrared detectors: A review. Rep. Prog. Phys. 2016, 79, 046501. [Google Scholar] [CrossRef]
  7. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2023, 32, 1745–1758. [Google Scholar] [CrossRef]
  8. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  9. Comaniciu, D. An algorithm for data-driven bandwidth selection. IEEE Trans. Pattern Anal. Mach. Intell. 2003, 25, 281–288. [Google Scholar] [CrossRef]
  10. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  11. Qin, Y.; Li, B. Effective infrared small target detection utilizing a novel local contrast method. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1890–1894. [Google Scholar] [CrossRef]
  12. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  13. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Infrared small-target detection using multiscale gray difference weighted image entropy. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 60–72. [Google Scholar] [CrossRef]
  14. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  15. Zhu, H.; Ni, H.; Liu, S.; Xu, G.; Deng, L. Tnlrs: Target-aware non-local low-rank modeling with saliency filtering regularization for infrared small target detection. IEEE Trans. Image Process. 2020, 29, 9546–9558. [Google Scholar] [CrossRef]
  16. Pang, D.; Shan, T.; Li, W.; Ma, P.; Tao, R.; Ma, Y. Facet derivative-based multidirectional edge awareness and spatial–temporal tensor model for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1–15. [Google Scholar] [CrossRef]
  17. Guan, X.; Zhang, L.; Huang, S.; Peng, Z. Infrared small target detection via non-convex tensor rank surrogate joint local contrast energy. Remote Sens. 2020, 12, 1520. [Google Scholar] [CrossRef]
  18. Pang, D.; Ma, P.; Shan, T.; Li, W.; Tao, R.; Ma, Y.; Wang, T. STTM-SFR: Spatial–temporal tensor modeling with saliency filter regularization for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–18. [Google Scholar] [CrossRef]
  19. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional local contrast networks for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  20. Zhang, M.; Zhang, R.; Yang, Y.; Bai, H.; Zhang, J.; Guo, J. ISNet: Shape Matters for Infrared Small Target Detection. In Proceedings of the 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), New Orleans, LA, USA, 18–24 June 2022; pp. 867–876. [Google Scholar]
  21. Wang, H.; Zhou, L.; Wang, L. Miss Detection vs. False Alarm: Adversarial Learning for Small Object Segmentation in Infrared Images. In Proceedings of the 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Seoul, Republic of Korea, 27 October–2 November 2019; pp. 8508–8517. [Google Scholar]
  22. Chen, G.; Wang, W.; Tan, S. IRSTFormer: A Hierarchical Vision Transformer for Infrared Small Target Detection. Remote Sens. 2022, 14, 3258. [Google Scholar] [CrossRef]
  23. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Asymmetric contextual modulation for infrared small target detection. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, Waikoloa, HI, USA, 3–8 January 2021; pp. 950–959. [Google Scholar]
  24. Wang, K.; Du, S.; Liu, C.; Cao, Z. Interior Attention-Aware Network for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–13. [Google Scholar] [CrossRef]
  25. Chen, F.; Gao, C.; Liu, F.; Zhao, Y.; Zhou, Y.; Meng, D.; Zuo, W. Local patch network with global attention for infrared small target detection. IEEE Trans. Aerosp. Electron. Syst. 2022, 58, 3979–3991. [Google Scholar] [CrossRef]
  26. Tong, X.; Su, S.; Wu, P.; Guo, R.; Wei, J.; Zuo, Z.; Sun, B. MSAFFNet: A multi-scale label-supervised attention feature fusion network for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–16. [Google Scholar] [CrossRef]
  27. Wu, X.; Hong, D.; Chanussot, J. UIU-Net: U-Net in U-Net for infrared small object detection. IEEE Trans. Image Process. 2022, 32, 364–376. [Google Scholar] [CrossRef]
  28. Chen, Y.; Li, L.; Liu, X.; Su, X. A multi-task framework for infrared small target detection and segmentation. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–9. [Google Scholar] [CrossRef]
  29. Wu, T.; Li, B.; Luo, Y.; Wang, Y.; Xiao, C.; Liu, T.; Yang, J.; An, W.; Guo, Y. MTU-Net: Multilevel TransUNet for Space-Based Infrared Tiny Ship Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–15. [Google Scholar] [CrossRef]
  30. Zhang, W.; Cong, M.; Wang, L. Algorithms for optical weak small targets detection and tracking. In Proceedings of the International Conference on Neural Networks and Signal Processing (ICNNSP), Nanjing, China, 14–17 December 2003; pp. 643–647. [Google Scholar]
  31. Pang, D.; Ma, P.; Feng, Y.; Shan, T.; Tao, R.; Jin, Q. Tensor Spectral k-support Norm Minimization for Detecting Infrared Dim and Small Target against Urban Backgrounds. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  32. Hui, B.; Song, Z.; Fan, H.; Zhong, P.; Hu, W.; Zhang, X.; Ling, J.; Su, H.; Jin, W.; Zhang, Y. A dataset for infrared detection and tracking of dim-small aircraft targets under ground/air background. China Sci. Data 2020, 5, 291–302. [Google Scholar]
  33. Sun, H.; Bai, J.; Yang, F.; Bai, X. Receptive-field and Direction Induced Attention Network for Infrared Dim Small Target Detection with a Large-scale Dataset IRDST. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–13. [Google Scholar] [CrossRef]
  34. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional networks for biomedical image segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention (MICCAI), Munich, Germany, 5–9 October 2015; pp. 234–241. [Google Scholar]
  35. Sun, C.; Shrivastava, A.; Singh, S.; Gupta, A. Revisiting unreasonable effectiveness of data in deep learning era. In Proceedings of the 2017 IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 843–852. [Google Scholar]
  36. Dai, Y.; Li, X.; Zhou, F.; Qian, Y.; Chen, Y.; Yang, J. One-Stage Cascade Refinement Networks for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 1–17. [Google Scholar] [CrossRef]
  37. Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  38. Ultralytics. Yolov5. Available online: https://github.com/ultralytics/yolov5 (accessed on 18 May 2020).
  39. Canny, J. A computational approach to edge detection. IEEE Trans. Pattern Anal. Mach. Intell. 1986, 6, 679–698. [Google Scholar] [CrossRef]
  40. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the 2016 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  41. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 6000–6010. [Google Scholar]
  42. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  43. Wang, W.; Xie, E.; Li, X.; Fan, D.P.; Song, K.; Liang, D.; Lu, T.; Luo, P.; Shao, L. Pyramid vision transformer: A versatile backbone for dense prediction without convolutions. In Proceedings of the 2021 IEEE/CVF International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 10–17 October 2021; pp. 568–578. [Google Scholar]
  44. Zhang, Q.; Yang, Y.B. Rest: An efficient transformer for visual recognition. Adv. Neural Inf. Process. Syst. 2021, 34, 15475–15485. [Google Scholar]
  45. Duchi, J.C.; Hazan, E.; Singer, Y. Adaptive subgradient methods for online learning and stochastic optimization. J. Mach. Learn. Res. 2011, 12, 2121–2159. [Google Scholar]
  46. Glorot, X.; Bengio, Y. Understanding the difficulty of training deep feedforward neural networks. In Proceedings of the Thirteenth International Conference on Artificial Intelligence and Statistics (ICAIS), Sardinia, Italy, 13–15 May 2010; pp. 249–256. [Google Scholar]
  47. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Chintala, S.; et al. PyTorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  48. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the 2017 IEEE/CVF International Conference on Computer Vision (ICCV), Venice, Italy, 22–29 October 2017; pp. 318–327. [Google Scholar]
  49. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 2017 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
Figure 1. Construction process of the SLR-IRST dataset.
Figure 1. Construction process of the SLR-IRST dataset.
Remotesensing 15 05539 g001
Figure 2. Statistical properties of SLR-IRST. (a) Statistical properties of target quantity distribution. (b) Statistical properties of target class distribution. (c) Statistical properties of target size distribution.
Figure 2. Statistical properties of SLR-IRST. (a) Statistical properties of target quantity distribution. (b) Statistical properties of target class distribution. (c) Statistical properties of target size distribution.
Remotesensing 15 05539 g002
Figure 3. Overall architecture of the CCRANet.
Figure 3. Overall architecture of the CCRANet.
Remotesensing 15 05539 g003
Figure 4. Architecture of the CCRS module. The dashed part appears only in the first basic block in stage 2 and stage 3.
Figure 4. Architecture of the CCRS module. The dashed part appears only in the first basic block in stage 2 and stage 3.
Remotesensing 15 05539 g004
Figure 5. U-FE module architecture. DoubleConv consists of two 3 × 3 convolutions (including BN and ReLU).
Figure 5. U-FE module architecture. DoubleConv consists of two 3 × 3 convolutions (including BN and ReLU).
Remotesensing 15 05539 g005
Figure 6. Each step in the SPSA from the circular region of interest features to the matrices Q, K, V in self-attention. (nd) means that the feature matrix is n-dimensional (n = 2, 3, 4).
Figure 6. Each step in the SPSA from the circular region of interest features to the matrices Q, K, V in self-attention. (nd) means that the feature matrix is n-dimensional (n = 2, 3, 4).
Remotesensing 15 05539 g006
Figure 7. Architecture of the multi-head self-attention mechanism for scaling down in the SPSA.
Figure 7. Architecture of the multi-head self-attention mechanism for scaling down in the SPSA.
Remotesensing 15 05539 g007
Figure 8. The PR curve and ROC curve of the CCRANet and contrast deep learning methods. (a) PR curve. (b) ROC curve.
Figure 8. The PR curve and ROC curve of the CCRANet and contrast deep learning methods. (a) PR curve. (b) ROC curve.
Remotesensing 15 05539 g008
Figure 9. Visualization of test images for the SLR-IRST dataset. (a) image. (b) label. (c) IPI. (d) ALCNet. (e) DNANet. (f) IAANet. (g) CCRANet.
Figure 9. Visualization of test images for the SLR-IRST dataset. (a) image. (b) label. (c) IPI. (d) ALCNet. (e) DNANet. (f) IAANet. (g) CCRANet.
Remotesensing 15 05539 g009aRemotesensing 15 05539 g009b
Figure 10. Visualization of test images for the SAC-IRST dataset (the top two pairs are images with targets and the bottom two are images without targets). (a) image. (b) label. (c) IPI. (d) ALCNet. (e) DNANet. (f) IAANet. (g) CCRANet.
Figure 10. Visualization of test images for the SAC-IRST dataset (the top two pairs are images with targets and the bottom two are images without targets). (a) image. (b) label. (c) IPI. (d) ALCNet. (e) DNANet. (f) IAANet. (g) CCRANet.
Remotesensing 15 05539 g010
Table 1. Details on the present single-frame infrared small target datasets.
Table 1. Details on the present single-frame infrared small target datasets.
Data TypeDatasetImage NumImage SizeProvided LabelTarget True ClassBackground Type
RealSIRST [19]42796 × 135 to 388 × 418Pixel/BoxAircraft/Drone/Ship/VehicleCloud/Grass/River
IRSTD-1k [20]1001512 × 512PixelDrone/Bird/AnimalCloud/Building/Grass/River/Lake
SyntheticMFIRST [21]10,000128 × 128Pixel-Cloud/Road
NUDT-SIRST [7]1327256 × 256Pixel-Cloud/Building/Vegetation
IRST640 [22]1024640 × 512Pixel-Cloud/Building
real/syntheticSLR-IRST
(our)
2689256 × 256Pixel/Box/CenterAircraft/Drone/Ship/Vehicle/Bird/AnimalCloud/Building/Grass/River/Lake/Vegetation
Table 2. Details on the present multi-frame infrared small target datasets.
Table 2. Details on the present multi-frame infrared small target datasets.
Data TypeDatasetSequence NumImage NumImage SizeTarget True ClassBackground Type
realDataset in [31]6342318 × 256 to 540 × 398DroneCity/Building/Tower Hanger
ISATD [32]2216177256 × 256DroneSky/Field/Building
real/syntheticIRDST [33]401142727720 × 480/934 × 696Aircraft/DroneClouds/Trees/Lakes/Buildings
Table 3. Real target class statistics of the SLR-IRST dataset and other mainstream infrared small target datasets.
Table 3. Real target class statistics of the SLR-IRST dataset and other mainstream infrared small target datasets.
DatasetAirborne TargetWaterborne TargetGround Target
SIRST3941518
IRST640102400
IRSTD-1K51687398
SLR-IRST1960312417
Table 4. Basic statistics of the SLR-IRST dataset and other mainstream infrared small target datasets.
Table 4. Basic statistics of the SLR-IRST dataset and other mainstream infrared small target datasets.
DatasetSLR-IRSTSIRSTIRST640IRSTD-1K
image size256 × 25696 × 135 to 388 × 418640 × 512512 × 512
image num268942710241001
target num358653316621495
target pixel range
(average)
1~367
(15.48)
4~330
(32.86)
1~51
(27.73)
1~1065
(50.11)
target size range
(average)
1 × 1~14 × 34
(3.33 × 4.58)
2 × 2~14 × 34
(5.62 × 6.94)
1 × 1~9 × 8
(6.11 × 6.33)
1 × 1~56 × 53
(7.69 × 8.74)
target SCR range
(average)
0.004~70.35
(7.46)
0.17~42.81
(9.20)
0.04~70.34
(4.93)
0.004~68.76
(7.13)
Table 5. Comparison of IoU metrics between the SOTA networks in the SIRST and SLR-IRST datasets.
Table 5. Comparison of IoU metrics between the SOTA networks in the SIRST and SLR-IRST datasets.
MethodSIRSTSLR-IRST
ALCNet0.75700.7077
DNANet0.77570.7076
Table 6. Comparison of pixel-level metrics for networks trained with different loss functions. The best result values are marked in bold.
Table 6. Comparison of pixel-level metrics for networks trained with different loss functions. The best result values are marked in bold.
Seg LossCenter LossnIoU (↑)IoU (↑)R (↑)P (↑)F1 − P (↑)
BCEL1+BCE0.73400.72860.85620.82750.8416
FocalL1+BCE0.73980.72160.79880.87030.8330
Soft-IoUL1+BCE0.73590.71780.80690.83660.8215
MixedL1+BCE0.75790.73980.83240.86940.8505
Table 7. Comparison of pixel-level and target-level metrics of the CCRS module hyperparameters under different settings. The best result values are marked in bold.
Table 7. Comparison of pixel-level and target-level metrics of the CCRS module hyperparameters under different settings. The best result values are marked in bold.
klPixel-Level MetricTarget-Level Metric
nIoU (↑)IoU (↑)R (↑)P (↑)F1 − P (↑)Pd (↑)Md (↓)Pa (↑)Fa (↓)F1 − T (↑)
3100.73480.72710.86820.81190.83910.95870.04130.82560.17440.8872
10100.67640.69410.88530.74980.81190.95160.04840.72860.27140.8253
5100.75790.73980.83240.86940.85050.96590.03410.87730.12270.9194
550.72160.66580.76810.82870.79730.95160.04840.82530.17470.8840
5200.73950.72870.82340.82960.82650.95590.04410.79360.20640.8672
Table 8. Network settings for the ablation experiment.
Table 8. Network settings for the ablation experiment.
CCRSU-FEFPNFCNSPSATSACONVName
A (CCRANet)
B1
B2
C1
C2
Table 9. Comparison of pixel-level and target-level metrics for the ablation experiments. The best result values are marked in bold.
Table 9. Comparison of pixel-level and target-level metrics for the ablation experiments. The best result values are marked in bold.
MethodPixel-Level MetricTarget-Level Metric
nIoU (↑)IoU (↑)R (↑)P (↑)F1 − P (↑)Pd (↑)Md (↓)Pa (↑)Fa (↓)F1 − T (↑)
A0.75790.73980.83240.86940.85050.96590.03410.87730.12270.9194
B10.67950.69320.80150.81170.80660.93880.06120.93940.06160.9391
B20.72430.69020.83170.78890.80970.95160.04840.75970.24130.8449
C10.73480.72780.86840.81490.84080.94310.05690.83280.16720.8845
C20.66910.67250.80260.79350.79800.94590.05410.60810.39190.7403
Table 10. Comparison of pixel-level and target-level metrics for the contrast experiments. The best result values are marked in bold.
Table 10. Comparison of pixel-level and target-level metrics for the contrast experiments. The best result values are marked in bold.
MethodPixel-Level MetricTarget-Level Metric
nIoU (↑)IoU (↑)R (↑)P (↑)F1 − P (↑)Pd (↑)Md (↓)Pa (↑)Fa (↓)F1 − T (↑)
IPI0.19860.15360.16890.63010.26630.58460.41540.40140.59860.4760
ALCNet0.66210.70770.81500.84320.82880.89900.10100.83930.16070.8681
DNANet0.69160.70760.77510.89050.82880.90470.09530.85830.14170.8809
IAANet0.67240.64510.77630.79250.78430.76960.23040.60510.39490.6775
CCRANet0.75790.73980.83240.86940.85050.96590.03410.87730.12270.9194
Table 11. Comparison of model processing speeds between the CCRANet and contrast networks.
Table 11. Comparison of model processing speeds between the CCRANet and contrast networks.
MethodALCNetDNANetIAANetCCRANet
Fps (↑)111.0945.6824.4366.52
Table 12. Comparison of pixel-level and target-level metrics of different methods on the SAC-IRST dataset (with target images). The best result values are marked in bold.
Table 12. Comparison of pixel-level and target-level metrics of different methods on the SAC-IRST dataset (with target images). The best result values are marked in bold.
MethodPixel-Level MetricTarget-Level Metric
nIoU (↑)IoU (↑)R (↑)P (↑)F1 − P (↑)Pd (↑)Md (↓)Pa (↑)Fa (↓)F1 − T (↑)
IPI0.19070.14390.21650.30030.25160.80430.19570.26810.73190.4022
ALCNet0.51780.48860.55840.79630.65650.84780.15220.78000.22000.8125
DNANet0.53310.53980.58660.87140.70120.93480.06520.86000.14000.8958
IAANet0.53500.46730.51080.84590.63700.93480.06520.78180.21820.8515
CCRANet0.55550.58320.64500.85880.73670.95650.04350.88000.12000.9167
Table 13. Comparison of false-alarm number (Fa Num) of different methods on the SAC-IRST dataset (without target images). The best result values are marked in bold.
Table 13. Comparison of false-alarm number (Fa Num) of different methods on the SAC-IRST dataset (without target images). The best result values are marked in bold.
MethodIPIALCNetDNANetIAANetCCRANet
Fa Num(↓)25616146210
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, W.; Xiao, C.; Dou, H.; Liang, R.; Yuan, H.; Zhao, G.; Chen, Z.; Huang, Y. CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection. Remote Sens. 2023, 15, 5539. https://doi.org/10.3390/rs15235539

AMA Style

Wang W, Xiao C, Dou H, Liang R, Yuan H, Zhao G, Chen Z, Huang Y. CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection. Remote Sensing. 2023; 15(23):5539. https://doi.org/10.3390/rs15235539

Chicago/Turabian Style

Wang, Wenjing, Chengwang Xiao, Haofeng Dou, Ruixiang Liang, Huaibin Yuan, Guanghui Zhao, Zhiwei Chen, and Yuhang Huang. 2023. "CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection" Remote Sensing 15, no. 23: 5539. https://doi.org/10.3390/rs15235539

APA Style

Wang, W., Xiao, C., Dou, H., Liang, R., Yuan, H., Zhao, G., Chen, Z., & Huang, Y. (2023). CCRANet: A Two-Stage Local Attention Network for Single-Frame Low-Resolution Infrared Small Target Detection. Remote Sensing, 15(23), 5539. https://doi.org/10.3390/rs15235539

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop