Next Article in Journal
Tracking of Maneuvering Extended Target Using Modified Variable Structure Multiple-Model Based on Adaptive Grid Best Model Augmentation
Previous Article in Journal
Measuring Coastal Subsidence after Recent Earthquakes in Chile Central Using SAR Interferometry and GNSS Data
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Unsupervised Infrared Small-Object-Detection Approach of Spatial–Temporal Patch Tensor and Object Selection

Department of Research, Nanjing Research Institute of Electronic Technology, Nanjing 210039, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(7), 1612; https://doi.org/10.3390/rs14071612
Submission received: 8 March 2022 / Accepted: 23 March 2022 / Published: 28 March 2022
(This article belongs to the Section Remote Sensing Image Processing)

Abstract

:
In this study, an unsupervised infrared object-detection approach based on spatial–temporal patch tensor and object selection is proposed to fully use effective temporal information and maintain a balance between object-detection performance and computation time. Initially, a spatial–temporal patch tensor is proposed by performing median pooling function on patch tensors generated from consecutive frames to suppress sky or cloud clutter. Then, a contrast-boosted approach that incorporates morphological operations is proposed to improve the contrast between objects and background. Finally, an object-selection approach is proposed based on the cluster center derived from clustering locations and gray values, thereby decreasing the search scope of objects in the detection process. The experiments of five infrared sequence frames confirm that the proposed framework can obtain better results than most previous methods when handling heterogeneous scenes in terms of gray values. Experimental results of five real sequence frames also demonstrate that the spatial–temporal patch tensor, the contrast-boosted approach, and object-selection approach can increase the recall ratio by 6.7, 2.21, and 1.14 percentage units and the precision ratio by 1.61, 3.44, and 11.79 percentage units, respectively. Moreover, the proposed framework can achieve an average F1 score of 0.9804 with about 1.85 s of computation time, demonstrating that it can obtain satisfactory object-detection performance with relatively low computation time.

1. Introduction

Infrared search-and-track systems (ISTS) have been widely applied to military applications such as missile warning, precision guidance, and space surveillance [1,2,3]. Infrared small-object detection is one of the most important techniques that can influence the performance of ISTS [4]. Infrared small-object detection aims to detect objects including ships, airplanes, and vehicles on backgrounds of sea, sky, and land [5,6]. However, as shown in Figure 1, unmanned aerial vehicle (UAV) objects in infrared images may take up only several pixels, with little structural or texture information because they are far from the imaging sensor [7,8]; thus, they may reduce the detection robustness. Moreover, complex backgrounds, such as cloud clutter and sea clutter, may lead to low signal-to-clutter ratio (SCR) [9,10], which may increase the difficulty of detecting small objects in infrared images. For example, artificial heat sources and heavy clouds may increase the false-alarm rates [11]. Therefore, detecting objects with infrared image sequences is still a difficult problem.
Infrared object-detection methods usually consist of single-frame-based methods and sequence-based methods according to the number of frames used for detecting objects in infrared images [12]. Sequence-based methods including 3D matched filter [13], dynamic programming [14], pipeline filter [15], and Kalman filter [16], assume that the background is static between adjacent frames of the same image sequence. Nevertheless, the background usually changes rapidly in military applications because sensor platforms used for ISTS move fast even though the objects to be detected may remain static [17]. Therefore, sequence-based approaches are unsuitable for potential applications. Single-frame-based methods that detect objects in each frame by exploring the consistency between background pixels are investigated because they can represent the background information more accurately.
The prior information is one of the most important components in single-frame-based infrared object-detection methods [18]. The infrared object-detection approaches can be categorized into two types based on the type of prior information used, namely, local prior-based methods and nonlocal prior-based methods [19].
Local prior-based approaches exploit a local background consistent prior that assumes the background is slowly transitional and close background pixels are highly correlated [20]. However, real objects may break the local correlation hidden in the background pixels. Under this assumption, traditional local prior-based approaches include 2D least-mean-square filter [21], morphological filter [22], max-median filter [23], and their improved methods. Unfortunately, traditional filters improve the edges of objects as well as the sky–sea surface or heavy cloud clutter because these background structures also break the local correlation. Approaches based on the saliency map that computes the difference between objects and their neighborhood have been proposed to better distinguish objects and background structures [24]. Multiscale patch-based contrast measure [25], local contrast measure (LCM) [26], laplacian of Gaussian filter [27], multiscale gray difference weighted image entropy [28], nonnegativity constrained variational mode decomposition [29], and saliency in the Fourier domain [30] are used to obtain saliency maps. The local prior-based methods may suffer from a nonuniform or heterogeneous background that ruins the spatial consistency, thereby leading to a high false-alarm rate [31].
Different from local priors, nonlocal priors explore the nonlocal self-correlation property of background patches by taking advantage of target sparsity and low-rank of background pixels in infrared images [32]. Essentially, these types of methods model objects in infrared images as a sparse representation of the input data. A classical nonlocal prior-based approach called the infrared patch image (IPI) model [33] is proposed by taking advantage of the nonlocal relation between background images. However, it suffers from over-shrinking of targets and noise residual due to the nuclear norm of the low-rank regularization term. Therefore, some improved versions of IPI are proposed by better reconstructing a low-rank matrix. The patch tensor model [34] can explore the nonlocal information on the assumption of low-rank unfolding matrices. Dictionary learning and principal component pursuit approaches can separate the background and target matrix from the original image. However, they cannot effectively handle infrared images with complex background. Total variation should be combined with principal component pursuit to address this problem [35]. However, the result may be a local minimum because it approximates the l 0 n o r m [36] that represents the total number of nonzero elements as the l 1 n o r m [37], representing the sum of the magnitudes. Moreover, l p n o r m = ( i = 1 n | x i | p ) 1 p [38] is used to optimize the infrared small-object detection approaches to better reach the global minimum [39]. To accurately detect the small target located in a highly heterogeneous background, a low-rank and sparse representation model is proposed under the multi-subspace cluster assumption [40]. The nonlocal priors are more powerful and fit the real scenes effectively, but still suffer from sparse edges and noise [41]. Moreover, existing approaches are faced with two drawbacks, namely, target edges highlighted along with background edges, as shown in Figure 2a, and no clear edges, as shown in Figure 2b.
A novel infrared object-detection approach based on spatial–temporal patch tensor and object selection is proposed to fully use effective information in the temporal domain and to maintain a balance between the object-detection performance and computation time. Using the spatial–temporal patch tensor, color-boosted and object selection approach, UAV targets under sky or cloud background can be detected in target images reconstructed from original infrared images. The major contributions of this study are as follows:
  • The proposed framework is an unsupervised infrared object-detection method which can provide effective means for infrared small-object detection when no labeled information of true UAV targets is acquired.
  • The proposed spatial–temporal patch tensor can dig out the spatial and temporal evidence hidden in infrared image sequences by performing median pooling operations on three adjacent frames to further suppress the sky or cloud clutter and enable a better target-detection performance.
  • An object-selection approach is proposed to automatically extract objects from infrared images based on the cluster center derived from unsupervised clustering, which can decrease the search scope of objects and the false-alarm rates.
This paper is organized as follows. Section 2 introduces the proposed framework for infrared small-object detection consisting of constructing spatial–temporal patch tensor, contrast-boosted, and object-selection approach. Then, the dataset description and experimental setup are depicted in Section 3. Experimental results and their analysis are shown in Section 4. Section 5 concludes the study, and is then followed by future directions.

2. Materials and Methods

An unsupervised infrared UAV target-detection method based on spatial–temporal patch tensor and object selection is proposed in this study. Figure 3 shows that the proposed method can be divided into the following steps.
  • Constructing spatial–temporal patch tensor for each frame of infrared image sequences. For each frame of the infrared image sequence, its temporal window is constructed with three adjacent frames. Patch tensor is constructed for each frame of the temporal window with t sliding windows whose size is k × k . Median pooling is applied to the patch tensor of all frames in the temporal window to obtain the spatial–temporal patch tensor P R k × k × t .
  • Calculating prior weight map and its patch tensor. The prior weight maps that can reflect the background and object information in infrared images to some extent are calculated for each temporal frame by combining local and nonlocal priors. Patch tensor of prior weight map W P R k × k × t is constructed for each temporal frame with t sliding windows with a size of k × k .
  • Decomposing background and target patch tensor from the spatial–temporal patch tensor P and patch tensor of prior weight map W P . The patch tensor of temporal frame P R k × k × t is separated into a sparse patch tensor T R k × k × t and a low-rank patch tensor B R k × k × t with the constraint of W P , which can be considered a target and background patch tensor.
  • Reconstructing background and target images from background and target patch tensors. Background image I B and target image I T are reconstructed from the background patch tensor B and target patch tensor T .
  • Performing the proposed contrast-boosted method on reconstructed target images. To enhance the contrast between gray values of background pixels and target pixels, the contrast-boosted approach that combines Tophat operations with Bothat operations are performed on reconstructed target images to obtain enhanced target images.
  • Segmenting UAV targets from the enhanced target images by the object-selection approach based on the cluster center. Candidate pixels that satisfy the threshold of prior weight map are obtained for each frame. Then, the object-selection approach is proposed to determine the optimal object number by clustering the candidate object pixels. Each cluster center is considered a detected object.

2.1. Construction of Spatial–Temporal Patch Tensor and Prior Weight Map

Construction of spatial–temporal patch tensor: The existing infrared object-detection methods are usually based on the spatial patch tensor, which may ignore the temporal information hidden in infrared image sequences. Therefore, a spatial–temporal patch tensor is proposed to solve this problem.
The main idea of spatial–temporal patch tensor is shown in Figure 4. Given an infrared image sequence consisting of three frames I k 1 , I k , I k + 1 in a sliding window ω × h × 3 , a temporal window is initially constructed for each frame, namely, I k 1 , I k , I k + 1 . Then, a spatial patch tensor is constructed for each frame of the temporal window. Finally, median pooling is performed on all spatial patch tensors to obtain the spatial–temporal patch tensor.
The idea of constructing a spatial patch tensor should be explained prior to illustrating a spatial–temporal patch tensor. Overlapped local patches may contain spatial information hidden in pixels because different patches may include the same pixel in one image. Therefore, a spatial patch tensor from generated overlapped local patches in infrared images is constructed. The procedure can be illustrated as follows. Initially, overlapped local patches are generated from left to right and top to bottom in each frame of infrared image sequences. Then, gray values of each pixel in the overlapped local patches are vectorized as a column of a tensor. Finally, vectors of all local patches in the frame form the spatial patch tensor.
The details of constructing a spatial–temporal patch tensor can be illustrated as follows. A temporal window I k 1 , I k , I k + 1 is constructed for frame I k to fully use the temporal information. For each frame in the temporal windows, spatial patch tensors P k 1 , P k , P k + 1 are constructed. Three patch tensors in the temporal window that are generated by the local patch located in the same place of image are highly correlated because adjacent frames are usually highly correlated in image sequences. Therefore, the spatial patch tensors in the temporal may be redundant. To reduce the redundancy of the spatial patch tensor and fully use temporal information, median pooling is performed on three spatial patch tensors, as shown in Equation (1), as follows:
P m e d i a n = m e d i a n ( P k 1 , P k , P k + 1 ) ,
where the median value of corresponding elements of three spatial patch tensors is considered the value of the spatial–temporal patch tensor.
The size of the spatial–temporal patch tensor is the same as that of the spatial patch tensor for each frame but with richer temporal information. The spatial–temporal patch tensor can enrich the temporal information and better represent each frame of infrared image sequences.
Construction of prior weight map: Existing prior weight-based methods use structure tensors [42] to distinguish between image boundaries and real objects. Structure tensors are widely used in many partial differential equation (PDE)-based methods [43] to estimate the local structure information in the image, including edge orientation. To integrate the local information, the structure tensor is constructed based on a local regularization of a tensorial product, which is defined in Equation (2), as follows:
J α ( I σ ) = G α ( I σ I σ ) = ( J 11     J 12 J 21     J 22 )
where I σ is a Gaussian-smoothed version of a given image I. σ > 0 is the standard deviation of the Gaussian kernel; it denotes the noise scale, making the edge detector ignorant of small details. J α is a symmetric and positive semi-definite matrix.
Two highest eigenvalues λ 1 and λ 2 of the structure tensor can be used as two feature descriptors of the local geometry structure, which can be calculated as Equation (3).
λ 1 , λ 2 = ( J 11 + J 22 ) ± ( J 22 J 11 ) 2 + 4 J 12 2
A combination of eigenvalues can enhance image boundaries, which can distinguish image boundaries similar to those of targets. The existing local prior is designed in Equation (4).
W L P = exp ( h × ( L 1 L 2 ) d min d max d min )
where L 1 and L 2 can be calculated by applying Equations (2) and (3) to every pixel in the input image I, h is a weight-stretching parameter, d max and d min are the maximum and minimum of eigenvectors L 1 and L 2 , respectively.
However, λ 1 λ 2 operator cannot identify whether image boundaries are background or targets. As a result, objects located at corner regions disappear or over-shrink.
To address these problems, the prior weight map W P improves the corner strength function W c s [44], as shown in Equations (5) and (6).
W P ( x , y ) = max ( λ 1 , λ 2 ) W c s
W c s = λ 1 λ 2 λ 1 + λ 2
where ( x , y ) represents the location of pixel. The prior weight map W P consists of two parts. The first part max ( λ 1 , λ 2 ) replaces the λ 1 λ 2 operator with the maximum operator max ( λ 1 , λ 2 ) to suppress the problem in the λ 1 λ 2 operator to some extent. In the second part, λ 1 + λ 2 and λ 1 λ 2 denote the trace and determinant of the structure tensor, respectively. The second part, namely, W cs can not only highlight object information as expected, but also identify the objects located at corner regions.
Then, the prior weight map W P is normalized with Equation (7).
W P = W P W min W max W min ,
where W min and W max represent the minimum and the maximum of the prior weight map W P , respectively.

2.2. Reconstructing Background and Target Images from Spatial–Temporal Patch Tensor

Reconstructing background and target patch tensor: Existing methods impose tensor robust principal component analysis [45] to separate sparse and low-rank tensors, as shown in Equation (8).
min B , T   r a n k ( B ) + λ T 0   s . t . D = B + T ,
where D, B and T represent the input, background, and target patch tensor, respectively; λ represents the tradeoff parameter between the background patch tensor and target patch tensor.   0 represents the number of nonzero elements. However, the low-rank of tensors cannot be evaluated in a mathematical approach.
To approximate the rank of the patch tensor more accurately and to incorporate prior information into the model that separates low-rank and sparse matrices, the objective function is modified in Equation (9), as follows:
min B , T i = 1 n 3 i = N + 1 min ( n 1 , n 2 ) σ i ( B ) + λ T W r e c 1 , B R n 1 × n 2 × n 3
where represents the Hadamard product [46], and   1 represents the sum of absolute values of all elements in the tensor. Each element in W r e c represents the reciprocals of the corresponding element in W P . N represents the preserved target rank, and σ i ( B ) is the i-th largest singular value of B. The objective function Equation (9) can be solved with alternating direction method of multiplier (ADMM) solver [47] whose procedure is shown in Algorithm 1.
Algorithm 1.ADMM solver
Input: Patch tensor of original image D, prior weight map W p , tradeoff parameter λ , penalty factor μ 0 , preserved target rank N, stopping threshold ξ , and learning rate of penalty factor ρ .
Output: Patch tensor of target and background images T k and B k .
Initialize: B k = T k = 0, μ 0 = 3 × 10 3 , ρ = 1.1 , k = 0.
(1)    Calculate the Langrangian function of Equations (9) with Equation (10).
L ( B , T , W , M ) = i = 1 n 3 B p = N + λ T W r e c 1 + ( B + T D ) y + μ 2 B + T D F 2
    where y represents the langrangian multiplier. is the inner product of two tensors and F is the Frobenius norm.
(2)    While not end of convergence do
(3)      Calculate T k + 1 when fixing other parameters by solving Equation (11).
T k + 1 = arg min T λ T W r e c + μ k 2 B k + T D + y k μ k F 2
(4)      Calculate B k + 1 when fixing other parameters by solving Equation (12).
B k + 1 = arg min B i = 1 n 3 B p = N + μ k 2 B + T k + 1 D + y k μ k F 2
(5)      Calculate y k + 1 and μ k + 1 with Equations (13) and (14), respectively.
y k + 1 = y k + μ k ( D B k + 1 T k + 1 )
μ k + 1 = ρ μ k
(6)      Check the stopping criterion shown in Equation (15).
T k + 1 0 T k 0 T k 0 ξ
(7)      k = k + 1
(8)    End while
(9)  Return: background and target patch tensor B k + 1 and T k + 1
As shown in Figure 5, diverse types of patches may include the same pixel because local patches generated by the spatial patch tensor are usually overlapped. As a result, a pixel in infrared images may have several pixel values due to overlapped local patches. To determine pixel values from overlapped patches, a median function is introduced to reconstruct each pixel value as the median values of pixel values from overlapped patches, as shown in Equation (16).
v = m e d i a n ( x )
where v R and x R p are vectors including pixel gray values from p local patches.
Contrast-boosted approach for target images: When reconstructing target and background images from original images, gray values of objects are lost in the background image to some extent because parts of objects are similar to the backgrounds. Therefore, target images should be enhanced in terms of the contrast between object and background to better distinguish between target and background pixels. Therefore, a contrast-boosted approach that combines Tophat [48] with Bothat [49] operations is proposed.
The contrast-boosted approach is performed to achieve an enhanced target image T E , as shown in Equation (17), as follows:
T E = T + T t o p h a t T b o t h a t
where T t o p h a t and T b o t h a t can be computed as Equations (18) and (19).
T t o p h a t = T ( T b ) ÷ b ,
T b o t h a t = ( T ÷ b ) × b T ,
where b represents the structure elements, × represents the dilation operation, and ÷ represents the erosion operation.
As shown in Figure 6, the gray values of objects are improved after performing the contrast-boosted method shown in Equation (17) on target images. The enhanced target images can better distinguish between background pixels and object pixels, thereby decreasing the false-alarm rate.

2.3. Object-Selection Approach Based on the Cluster Center

After obtaining enhanced target images, objects should be detected. When detecting objects in target images, the number of objects should be determined. Therefore, an object-selection approach is proposed for improving target images based on the cluster center derived from clustering locations and gray values. The main idea of determining the optimal number of objects is shown in Figure 7. The object selection approach can be divided into the following steps:
A predefined threshold τ is used to select candidate object pixels because the prior weight maps calculated can reflect the possibility that one pixel belongs to the target.
Then, the optimal number of objects is determined by clustering candidate object pixels in terms of locations and gray values with k-means based on the assumption that close candidate object pixels are likely to belong to the same object. The procedure of determining the optimal number of objects is shown in Algorithm 2.
Algorithm 2.Determination of optimal object number
Input: Locations and gray values of candidate objects L c and G c , the stopping threshold τ 1 and τ 2 .
Output: The optimal object number O k .
Initialize: Cluster number k = 1.
(1)   Cluster locations and gray values of candidate object pixels with k-means and cluster number k, respectively.
(2)    While not end of convergence do
(3)       k = k + 1.
(4)        Calculate the distance between each cluster center derived from clustering locations and locations, respectively, with Equations (20) and (21).
D c l = c i l c j l 2
D c g = c i g c j g 2
             where c i l and c i g represent the i-th cluster center derived from locations and gray values of candidate object pixels, respectively.
(5)         Check the stopping criterion shown in Equation (22).
D c l τ 1   &   D c g τ 2
(6)     End while
(7)    Return: optimal object number k − 1
Finally, cluster locations of candidate object pixels with the determined optimal object number. Each cluster is considered a detected object.
The object-selection approach can decrease the false-alarm rate and the search scope of objects to obtain satisfactory object-detection performance with relatively low computation time.

3. Results

3.1. Description of the Datasets and Experimental Setup

3.1.1. Description of the Datasets

Five infrared approaches are employed to evaluate the performance of the proposed infrared UAV target-detection method. The first dataset [50] is aimed at detecting and tracking small UAV targets under the ground or sky background. This dataset is denoted as dataset 1. Four representative image sequences are selected from 22 image sequences to evaluate the proposed method under various scenes, namely, Sequence 3, Sequence 4, Sequence 16, and Sequence 18. Dataset 1 can be downloaded at the web address of http://www.dx.doi.org/10.11922/sciencedb.902 (accessed on 6 March 2022). Table 1 describes the detailed information about all five infrared image sequences, and Figure 8 shows the examples and ground truth of four infrared image sequences in dataset 1. All thermal images have the same size of 256 × 256. The spatial resolution of this dataset ranges from 10 m to 100 m. The distance to targets ranges from 50 m to 500 m. The object in this dataset ranges from 3 pixels to 10 pixels due to diverse distances to targets. The ground truth of the dataset provides each image frame with its number of objects and the center of object.
Another infrared image sequence that contains airplane targets under a complex sky background has been experimented on to evaluate the robustness of the proposed method. This sequence is denoted as dataset 2. All images in this dataset have a size of 256 × 200. All images have the same spatial resolution, and the object size is 6 pixels. Examples of the sequence along with their ground truth are shown in Figure 9.

3.1.2. Experimental Setup

To prove the superiority of the proposed unsupervised infrared small-object detection method, it is compared with seven publicly available infrared object-detection approaches that reconstruct the background and the target from original infrared images and two deep-learning-based approaches. The compared approaches include sequence-based approaches, mixture of Gaussians (MOG) [51], nonlocal prior weight-based approaches IPI model [33], non-convex rank approximation minimization (NRAM) [39], local prior weight-based approaches tri-layer local contrast measure (TLLCM) [52], weighted scale local contrast measure (WSLCM) [53], generalized structure tensor (GST) [54], and partial sum of the tensor nuclear norm (PSTNN) [55] that combines local and nonlocal prior weights and deep-learning-based approaches, you only look once (YOLO), and single-shot detector (SSD). The target images obtained from seven methods not based on deep-learning approaches and the proposed framework are used to detect objects by using Algorithm 2. A window with size of 5 × 5 is used after ADMM is solved.
Table 2 describes the computer settings of the proposed framework and compared methods. All experiments are implemented under MATLAB 2016a on a PC (Core i7-6700 CPU @ 3.40 GHz, 16.0 GB of memory) with Windows 7 x64 operating system.
The precision ratio, recall ratio, F1 score, and average computation time are used to evaluate the performance of infrared object-detection performance. The recall ratio (RR), precision ratio (PR), and F1 score can be calculated according to Equations (23)–(25).
R R = T P T P + F N ,
P R = T P T P + F P ,
F 1 = 2 T P 2 T P + F N + F P
where TP represents the number of objects that are correctly classified, FP represents the number of background objects that is classified as objects, and FN represents the number of undetected objects.

3.2. Comparison with State-of-the-Art Methods in Five Sequences

Table 3 and Table 4 show the recall ratio and precision ratio of the proposed framework compared with competitive infrared object detection approaches. As shown in Table 3 and Table 4, the proposed framework achieves the highest precision ratio and relatively high recall ratio, demonstrating that it is suitable for heterogeneous scenes, where the background can be distinguished from the targets. The proposed framework can improve the recall ratio by the proposed spatial–temporal patch tensor or the contrast-boosted approach and increase the precision ratio by the object-selection method. SSD and YOLO achieve higher recall ratio and lower precision ratio in most sequences because they do not have clutter samples. The MOG algorithm achieves the highest recall ratio in four sequences but the lowest precision ratio in sequences 3, 16, and 18, demonstrating that it is more suitable for scenes, where background is different from objects in gray values. The MOG approach delivers poor precision ratio because it considers the temporal information hidden in adjacent frames, which may enhance the background information and cause false alarms. PSTNN and GST approaches deliver poorer performance in terms of recall and precision ratio because they lose object information when suppressing background information. IPI and NRAM methods perform worse than the proposed framework in sequences 3, 4, and 16 but better than the proposed framework in sequence 18, demonstrating that they are more suitable for sequences, where each frame contains only one object, and unsuitable for scenes, where multiple objects or no object exists. TLLCM and WSLCM deliver acceptable object detection performance in sequences 3, 16, and 18 but unsatisfactory performance in sequence 4, demonstrating that they cannot handle scenes, where multiple objects exist.
Table 5 shows the comprehensive evaluation metric F1 score of the proposed framework compared with state-of-the-art infrared object-detection approaches. The proposed framework achieves an F1 score comparable or even higher than deep-learning-based approaches SSD and YOLO, demonstrating that it can adapt to sequences when faced with multiple objects and complex background. SSD and YOLO can achieve an F1-score comparable to the proposed framework because object samples are used to train the object-detection model. The MOG achieves the highest F1 score in sequence 4 and the lowest F1 score in other sequences, indicating that it is suitable for scenes, where objects are highly different from the background. PSTNN performs effectively in sequences 3, 4, and 16 but delivers poor performance in sequence 18, indicating that PSTNN is unsuitable for scenes, where objects are similar to backgrounds to some extent. The IPI approach delivers satisfactory performance in sequences 4, 16, and 18 but unsatisfactory performance in sequence 3. This condition is caused by the nonexistence of objects in some frames of sequence 3. NRAM achieves a relatively high precision ratio at the sacrifice of the recall ratio in five sequences. TLLCM and WSLCM are more appropriate for scenes, where only one object exists, and these methods may suffer from multiple objects and no objects in some frames of image sequences.

3.3. Ablation Experiments

Table 6 and Table 7 show the recall and precision ratio of ablation experiments. Table 6 shows that the spatial–temporal tensor, the contrast-boosted, and the object-selection approach can increase average recall ratio by 6.7, 2.21, and 1.14 percentage units, respectively, in five sequences. The spatial–temporal patch tensor increases the recall ratio the most because it considers the temporal information hidden in adjacent frames. The contrast-boosted approach increases the recall ratio the second most in 3, 4, and 16 and dataset 2 but reduces the recall ratio in sequence 18 because it may enhance the object and background information. Thus, the possibility of detecting objects may increase. The object-selection approach increases the recall ratio the least because it does not increase the possibility of detecting objects.
Table 7 indicates that the spatial–temporal tensor, the contrast-boosted approach, and the object-selection approach can improve an average precision ratio by 1.61, 3.44, and 11.79 percentage units, respectively. The object-selection approach can increase the precision ratio the most because it predicts the most suitable number of objects, thereby decreasing the possibility of detecting background pixels as objects. The contrast-boosted approach can increase the precision ratio because they can increase the possibility of detecting objects by enhancing the gray value of objects. The spatial–temporal tensor slightly increases precision ratio because more temporal information may enhance the object and background information, thereby increasing the possibility of detecting background as objects.

3.4. Computation Cost

Table 8 shows the computation cost of the proposed framework along with other state-of-the-art infrared object-detection methods. The computation time of deep-learning-based methods only consists of time for testing each frame in the image sequence without the time for training an object-detection model. Table 8 shows that the proposed framework achieves a relatively balanced computation time of 1.85 s compared with existing approaches. GST and PSTNN obtain lower computation time than the proposed framework but delivers unsatisfactory object-detection performance. The MOG, IPI, and WSLCM require higher computation time, which cannot satisfy the requirements of real-time object detection. NRAM and TLLCM have computation time comparable to the proposed framework, but their object-detection performance is worse. Although deep-learning-based approaches require less than 1 s to test, the training time is thousands of seconds.

4. Discussion

Figure 10 shows the performance comparison in the 27th frame of sequence 3 between the proposed framework and other existing methods. The proposed framework can detect the actual UAV objects because the contrast-boosted approach can enhance the object information when parts of object information are lacking. Figure 9 shows that NRAM, IPI, GST, and WSLCM approaches fail to detect UAV targets because they cannot effectively handle scenes, where part of UAV targets is missing. TLLCM MOG and PSTNN approaches can detect UAV targets but the MOG algorithm detects some background pixels as objects.
Figure 11 shows the performance of the proposed framework in the 165th frame of sequence 4 compared with other existing infrared object-detection methods. Figure 10 shows that the proposed framework can detect two true UAV targets because the object-selection approach can determine the number of objects accurately, even though two objects are close in images. NRAM, GST, TLLCM, WSLCM, and PSTNN approaches may miss one UAV target because two UAV targets are very close in the 165th frame. The proposed framework, MOG, and IPI approaches can accurately detect two UAV targets, whereas the IPI method detects one background object as object.
Figure 12 shows the performance of the proposed framework along with other existing infrared object detection approaches in the 28th frame of sequence 16. Figure 11 shows that all methods including the proposed framework can detect true UAV targets because the background can be suppressed effectively in this scene. However, IPI, MOG, and PSTNN approaches detect some background pixels as objects because they do not suppress background information effectively.
Figure 13 shows the error of the proposed framework in five sequences; it is caused by different reasons. For (a), part of UAV objects is missing, leading to incomplete object shape information. Objects of (b) are close in infrared images, which may lead to missing detection of UAV targets. For (c)–(f), objects and background are highly similar in gray values and may cause confusion between objects and background.

5. Conclusions

An unsupervised infrared object-detection framework based on a spatial–temporal patch tensor and object-selection approach is proposed to address the problems of ignoring the temporal information hidden in infrared image sequences, the low contrast between objects and background, and the imbalance between real-time processing and satisfactory detection results. The proposed framework mainly consists of three contributions, namely, the spatial–temporal patch tensor, the contrast-boosted, and the object-selection approach. The following conclusions can be drawn from experiments performed on five image sequences.
  • The proposed framework outperforms most previous infrared object-detection approaches that reconstruct the background and the target from original infrared images when handling scenes, where the background is heterogeneous compared with objects in terms of gray values.
  • Spatial–temporal patch tensor and the contrast-boosted approach can increase the possibility of detecting real objects by utilizing temporal information hidden in adjacent frames and enhancing the contrast between objects and background. The object-selection approach can decrease the possibility of detecting background as objects by determining the appropriate number of objects.
  • The proposed framework can achieve the average F1 score of 0.9804 with computation time of approximately 1.85 s, demonstrating that it can obtain satisfactory object-detection performance with relatively low computation time.
Section 4 shows that the proposed framework may cause confusion when part of UAV objects is missing, objects are too close in infrared images or objects are highly similar to background. More studies are required in the future to address the above-mentioned problems.

Author Contributions

Conceptualization, R.Z. and L.Z.; methodology, R.Z.; validation, R.Z. and L.Z.; writing—original draft preparation, R.Z.; writing—review and editing, R.Z. and L.Z.; supervision, L.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Dataset 1 can be downloaded at the web address of http://www.dx.doi.org/10.11922/sciencedb.902 (accessed on 6 March 2022). Dataset 2 can be downloaded at the web address of https://github.com/wxw211311/small-target (accessed on 6 March 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

AbbreviationsExplanation
ISTSinfrared search and track systems
UAVunmanned aerial vehicle
SCRsignal-to-clutter ratio
LCMlocal contrast measure
IPIinfrared patch image
ADMMalternating direction method of multipliers
MOGmixture of gaussians
NRAMnon-convex rank approximation minimization
TLLCMtri-layer local contrast measure
WSLCMweighted scale local contrast measure
GSTgeneralized structure tensor
PSTNNpartial sum of the tensor nuclear norm
t-SVDtensor-singular value decomposition
RRrecall ratio
PRprecision ratio
YOLOyou look only once
SSDsingle-shot detector

References

  1. Cao, X.; Rong, C.; Bai, X. Infrared small target detection based on derivative dissimilarity measure. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3101–3116. [Google Scholar] [CrossRef]
  2. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  3. Kristo, M.; Ivasic-Kos, M.; Pobar, M. Thermal Object Detection in Difficult Weather Conditions Using YOLO. IEEE Access 2020, 8, 125459–125476. [Google Scholar] [CrossRef]
  4. Gao, J.; Guo, Y.; Lin, Z.; An, W.; Li, J. Robust Infrared Small Target Detection Using Multiscale Gray and Variance Difference Measures. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 5039–5052. [Google Scholar] [CrossRef]
  5. Xue, W.; Qi, J.; Shao, G.; Xiao, Z.; Zhang, Y.; Zhong, P. Low-Rank Approximation and Multiple Sparse Constraint Modeling for Infrared Low-Flying Fixed-Wing UAV Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 4150–4166. [Google Scholar] [CrossRef]
  6. Rawat, S.; Verma, S.K.; Kumar, Y. Review on recent development in infrared small target detection algorithms. Procedia Comput. Sci. 2020, 167, 2496–2505. [Google Scholar] [CrossRef]
  7. Han, J.; Liu, C.; Liu, Y.; Luo, Z.; Zhang, X.; Niu, Q. Infrared Small Target Detection Utilizing the Enhanced Closest-Mean Background Estimation. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 14, 645–662. [Google Scholar] [CrossRef]
  8. Zhang, H.; Bai, J.; Li, Z.; Liu, Y.; Liu, K. Scale invariant SURF detector and automatic clustering segmentation for infrared small targets detection. Infrared Phys. Technol. 2017, 83, 7–16. [Google Scholar] [CrossRef]
  9. Dai, Y.; Wu, Y. Reweighted Infrared Patch-Tensor Model with Both Nonlocal and Local Priors for Single-Frame Small Target Detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef] [Green Version]
  10. Hu, Z.; Guan, Y.; Deng, L.; Li, Y. Infrared moving point target detection based on an anisotropic spatial-temporal fourth-order diffusion filter. Comput. Electr. Eng. 2018, 68, 550–556. [Google Scholar]
  11. Liu, D.; Cao, L.; Li, Z.; Liu, T.; Che, P. Infrared Small Target Detection Based on Flux Density and Direction Diversity in Gradient Vector Field. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2018, 11, 2528–2554. [Google Scholar] [CrossRef]
  12. Mansoori, A.A.S.A.; Swamidoss, I.N.; Sayadi, S.; Almarzooqi, A. Analysis of different tracking algorithms applied on thermal infrared imagery for maritime surveillance systems. In Proceedings of the Artificial Intelligence and Machine Learning in Defense Applications II, Online, 21–25 September 2020; Volume 11543, p. 1154308. [Google Scholar] [CrossRef]
  13. Reed, I.S.; Gagliardi, R.M.; Stotts, L.B. Optical moving target detection with 3-D matched filtering. IEEE Trans. Aerosp. Electron. Syst. 2002, 24, 327–336. [Google Scholar] [CrossRef]
  14. Huang, L.; Zhang, G.; Wang, X. Detecting of small infrared moving object based on dynamic programming algorithm. Infrared Laser Eng. 2004, 33, 303–306. [Google Scholar]
  15. Wang, B.; Xu, W.; Zhao, M.; Wu, H. Antivibration pipeline-filtering algorithm for maritime small target detection. Opt. Eng. 2014, 53, 113109. [Google Scholar] [CrossRef]
  16. Vaishnavi, R.; Unnikrishnan, G.; Raj, A.A.B. Implementation of algorithms for Point target detection and tracking in Infrared image sequences. In Proceedings of the 4th International Conference on Recent Trends on Electronics, Information, Communication & Technology (RTEICT), Bangalore, India, 17–18 May 2019; pp. 904–909. [Google Scholar] [CrossRef]
  17. Sun, Y.; Yang, J.; An, W. Infrared Dim and Small Target Detection via Multiple Subspace Learning and Spatial-Temporal Patch-Tensor Model. IEEE Trans. Geosci. Remote Sens. 2020, 59, 3737–3752. [Google Scholar] [CrossRef]
  18. Algarni, A.D. Efficient Object Detection and Classification of Heat Emitting Objects from Infrared Images Based on Deep Learning. Multimed. Tools Appl. 2020, 79, 13403–13426. [Google Scholar] [CrossRef]
  19. Li, W.; Zhao, M.; Deng, X.; Li, L.; Zhang, W. Infrared Small Target Detection Using Local and Nonlocal Spatial Information. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2019, 12, 3677–3689. [Google Scholar] [CrossRef]
  20. Bae, T.-W.; Zhang, F.; Kweon, I.-S. Edge directional 2D LMS filter for infrared small target detection. Infrared Phys. Technol. 2012, 55, 137–145. [Google Scholar] [CrossRef]
  21. Hadhoud, M.M.; Thomas, D.W. The two-dimensional adaptive LMS (TDLMS) algorithm. IEEE Trans. Circuits Syst. 1988, 35, 485–494. [Google Scholar] [CrossRef]
  22. Ming, Z.; Li, J.; Zhang, P. The design of Top-Hat morphological filter and application to infrared target detection. Infrared Phys. Technol. 2006, 48, 67–76. [Google Scholar]
  23. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. Signal and Data Processing of Small Targets. Int. Soc. Opt. Photonics 1999, 3809, 74–83. [Google Scholar]
  24. Nasiri, M.; Mosavi, M.R.; Mirzakuchaki, S. Infrared dim small target detection with high reliability using saliency map fusion. IET Image Process. 2016, 10, 524–533. [Google Scholar] [CrossRef]
  25. Chen, C.L.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A Local Contrast Method for Small Infrared Target Detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  26. Kim, S.; Yang, Y.; Lee, J.; Park, Y. Small Target Detection Utilizing Robust Methods of the Human Visual System for IRST. J. Infrared Millim. Terahertz Waves 2009, 30, 994–1011. [Google Scholar] [CrossRef]
  27. Deng, H.; Sun, X.; Liu, M.; Ye, C. Infrared small-target detection using multi-scale gray difference weighted image entropy. IEEE Trans. Aerosp. Electron. Syst. 2016, 52, 60–72. [Google Scholar] [CrossRef]
  28. Wang, X.; Peng, Z.; Zhang, P.; He, Y. Infrared Small Target Detection via Nonnegativity-Constrained Variational Mode De-composition. IEEE Geosci. Remote Sens. Lett. 2017, 10, 1700–1704. [Google Scholar] [CrossRef]
  29. Li, J.; Duan, L.; Chen, X.; Huang, T.; Tian, Y. Finding the Secret of Image Saliency in the Frequency Domain. IEEE Trans. Pattern Anal. Mach. Intell. 2015, 37, 2428–2440. [Google Scholar] [CrossRef]
  30. Chen, Y.; Xin, Y. An Efficient Infrared Small Target Detection Method Based on Visual Contrast Mechanism. IEEE Geosci. Remote Sens. Lett. 2016, 13, 962–966. [Google Scholar] [CrossRef]
  31. Rawat, S.S.; Verma, S.K.; Kumar, Y. Infrared small target detection based on non-convex triple tensor factorisation. IET Image Process. 2020, 15, 556–570. [Google Scholar] [CrossRef]
  32. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  33. Mancera, L.; Portilla, J. L0-Norm-Based Sparse Representation through Alternate Projections. In Proceedings of the International Conference on Image Processing, Atlanta, GA, USA, 8–11 October 2006; pp. 2089–2092. [Google Scholar] [CrossRef]
  34. Jajuga, K. L1-norm based fuzzy clustering. Fuzzy Sets Syst. 1991, 39, 43–50. [Google Scholar] [CrossRef]
  35. Debeye, H.W.J.; Van Riel, P. Lp-NORM DECONVOLUTION1. Geophys. Prospect. 1990, 38, 381–403. [Google Scholar] [CrossRef]
  36. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared Small Target Detection via Non-Convex Rank Approximation Min-imization Joint l2,1 Norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef] [Green Version]
  37. He, Y.; Li, M.; Zhang, J.; An, Q. Small infrared target detection based on low-rank and sparse representation. Infrared Phys. Technol. 2015, 68, 98–109. [Google Scholar] [CrossRef]
  38. Dai, Y.; Wu, Y.; Song, Y.; Guo, J. Non-negative infrared patch-image model: Robust target-background separation via partial sum minimization of singular values. Infrared Phys. Technol. 2017, 81, 182–194. [Google Scholar] [CrossRef]
  39. Bigun, J.; Granlund, G.; Wiklund, J. Multidimensional orientation estimation with applications to texture analysis and optical flow. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 775–790. [Google Scholar] [CrossRef]
  40. Weickert, J. Coherence-enhancing diffusion filtering. Int. J. Comput. Vis. 1999, 31, 111–127. [Google Scholar] [CrossRef]
  41. Brown, M.; Szeliski, R.; Winder, S. Multi-image matching using multi-scale oriented patches. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, San Diego, CA, USA, 20–25 June 2005; Volume 1, pp. 510–517. [Google Scholar]
  42. Lu, C.; Feng, J.; Chen, Y.; Liu, W.; Lin, Z.; Yan, S. Tensor Robust Principal Component Analysis with a New Tensor Nuclear Norm. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 42, 925–938. [Google Scholar] [CrossRef] [Green Version]
  43. Horn, R.A. The hadamard product. Proc. Symp. Appl. Math. 1990, 40, 87–169. [Google Scholar]
  44. Aslan, S.; Nikitin, V.; Ching, D.J.; Bicer, T.; Leyffer, S.; Gürsoy, D. Joint ptycho-tomography reconstruction through alternating direction method of multi-pliers. Opt. Express 2019, 27, 9128–9143. [Google Scholar] [CrossRef]
  45. Tom, V.T.; Peli, T.; Leung, M.; Bondaryk, J.E. Morphology-based algorithm for point target detection in infrared backgrounds. Signal and Data Processing of Small Targets. Int. Soc. Opt. Photonics 1993, 1954, 2–11. [Google Scholar]
  46. Zhao, T.; Gong, H. Realization of tophat transform and bothat transform of mathematical morphology. Inf. Technol. 2008, 5, 149–151. [Google Scholar]
  47. A Dataset for Infrared Image Dim-Small Aircraft Target Detection and Tracking Underground/air Background (V1). 28 October 2019. cstr:31253.11.sciencedb.902. Available online: https://datapid.cn/31253.11.sciencedb.902 (accessed on 6 March 2022).
  48. Gao, C.; Wang, L.; Xiao, Y.; Zhao, Q.; Meng, D. Infrared small-dim target detection based on Markov random field guided noise modeling. Pattern Recognit. 2018, 76, 463–475. [Google Scholar] [CrossRef]
  49. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A Local Contrast Method for Infrared Small-Target Detection Utilizing a Tri-Layer Window. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1822–1826. [Google Scholar] [CrossRef]
  50. Du, P.; Hamdulla, A. Infrared Small Target Detection Using Homogeneity-Weighted Local Contrast Measure. IEEE Geosci. Remote Sens. Lett. 2019, 17, 514–518. [Google Scholar] [CrossRef]
  51. Gao, C.-Q.; Tian, J.-W.; Wang, P. Generalised-structure-tensor-based infrared small target detection. Electron. Lett. 2008, 44, 1349–1351. [Google Scholar] [CrossRef]
  52. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef] [Green Version]
  53. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27 June 2016; pp. 779–788. [Google Scholar]
  54. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single Shot Multibox Detector. European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  55. Mo, N.; Yan, L.; Zhu, R.; Xie, H. Class-Specific Anchor Based and Context-Guided Multi-Class Object Detection in High Resolution Remote Sensing Imagery with a Convolutional Neural Network. Remote Sens. 2019, 11, 272. [Google Scholar] [CrossRef] [Green Version]
Figure 1. UAV objects in infrared images and their intensity distributions. (a) UAV object under the sky background. (b) UAV object under the ground background. (c) Intensity distribution of infrared image (a). (d) Intensity distribution of infrared image (b).
Figure 1. UAV objects in infrared images and their intensity distributions. (a) UAV object under the sky background. (b) UAV object under the ground background. (c) Intensity distribution of infrared image (a). (d) Intensity distribution of infrared image (b).
Remotesensing 14 01612 g001
Figure 2. Drawbacks of existing prior-based methods. (a) Target edges highlighted along with background edges. (b) No clear edges but the target is similar to the background.
Figure 2. Drawbacks of existing prior-based methods. (a) Target edges highlighted along with background edges. (b) No clear edges but the target is similar to the background.
Remotesensing 14 01612 g002
Figure 3. The overall architecture of the proposed framework.
Figure 3. The overall architecture of the proposed framework.
Remotesensing 14 01612 g003
Figure 4. Sketch map of constructing spatial–temporal patch tensor.
Figure 4. Sketch map of constructing spatial–temporal patch tensor.
Remotesensing 14 01612 g004
Figure 5. Reconstruction of pixel values from patch tensor.
Figure 5. Reconstruction of pixel values from patch tensor.
Remotesensing 14 01612 g005
Figure 6. Target image and enhanced target images.
Figure 6. Target image and enhanced target images.
Remotesensing 14 01612 g006
Figure 7. Determination of object number from candidate objects.
Figure 7. Determination of object number from candidate objects.
Remotesensing 14 01612 g007
Figure 8. Image samples of four infrared image sequences in dataset 1. Frames without objects contain no red rectangles.
Figure 8. Image samples of four infrared image sequences in dataset 1. Frames without objects contain no red rectangles.
Remotesensing 14 01612 g008
Figure 9. Image sample of one infrared image sequence in dataset 2.
Figure 9. Image sample of one infrared image sequence in dataset 2.
Remotesensing 14 01612 g009
Figure 10. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 27th frame of sequence 3. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, (i) and PSTNN.
Figure 10. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 27th frame of sequence 3. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, (i) and PSTNN.
Remotesensing 14 01612 g010
Figure 11. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 165th frame from sequence 4. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, and (i) PSTNN.
Figure 11. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 165th frame from sequence 4. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, and (i) PSTNN.
Remotesensing 14 01612 g011
Figure 12. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 28th frame from sequence 16. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, Additionally, (i) PSTNN.
Figure 12. Performance of the proposed method compared with state-of-the-art infrared object detection methods in the 28th frame from sequence 16. (a) Original image along with its ground truth, (b) proposed method, (c) IPI, (d) NRAM, (e) GST, (f) MOG, (g) TLLCM, (h) WSLCM, Additionally, (i) PSTNN.
Remotesensing 14 01612 g012
Figure 13. Major confusion of the proposed framework. (a) 28th frame in sequence 3, (b) 164th frame in sequence 4, (c) 207th frame in sequence 16, (d) 297th frame in sequence 16, (e) 25th frame in sequence 18, and (f) 321st frame in sequence 18.
Figure 13. Major confusion of the proposed framework. (a) 28th frame in sequence 3, (b) 164th frame in sequence 4, (c) 207th frame in sequence 16, (d) 297th frame in sequence 16, (e) 25th frame in sequence 18, and (f) 321st frame in sequence 18.
Remotesensing 14 01612 g013
Table 1. Details of five representative infrared image sequences.
Table 1. Details of five representative infrared image sequences.
SequencesNumber of FramesAverage SCR ValuesDescription of Scenes
31002.17Single object, close imaging distance, mixed sky and ground background, no objects in some scenes.
43993.75Two objects, close imaging distance, sky background, cross flight
164992.98Single extended object, fast moving speed of objects, from close to far imaging distance, ground background,
185003.32Single mobile object, from close to far imaging distance, ground background
1705.11Single mobile object, sky background, cloud clutter
Table 2. Parameter settings of state-of-the-art infrared small-object-detection approaches.
Table 2. Parameter settings of state-of-the-art infrared small-object-detection approaches.
TypeState-of-the-Art MethodsAbbreviationsParameter Settings
Nonlocal prior weightInfrared Patch-ImageIPIPatch size 50, slide step 10, λ = 1 min ( m , n ) ,   ε = 10 7
Non-Convex Rank Approximation MinimizationNRAMPatch size 50, slide step 10,   λ = 1 min ( m , n )
Sequence based approachesMixture of GaussiansMOGTemporal step 3, patch step 5, patch size 50, component number 3, maximum iterations 300
Local prior weightTri-Layer Local Contrast MeasureTLLCMGaussian kernel [ 1   2   1 2   4   2 1   2   1 ] 16
Weighted Scale Local Contrast MeasureWSLCMGaussian kernel [ 1   2   1 2   4   2 1   2   1 ] 16
Generalized Structure TensorGST σ 1 = 0.6 ,   σ 2 = 1.1 , boundary width 5, filter size 5
Deep learningYou Look Only OnceYOLOBatch size 32,initial learning rate 0.001,weight decreases 50% every 2000 iterations, maximum iterations = 12,000, momentum = 0.9, weight decay = 0.0005
Single-Shot DetectorSSDBatch size 32, initial learning rate 0.0005, weight decreased by 10% every 500 iterations, 2000 max iterations
Local and nonlocal prior weightsPartial Sum of the Tensor Nuclear NormPSTNNPatch size 40, slide step 40, λ = 0.7
The proposed framework-Patch size 40, slide step 40, temporal size 3, λ = 0.7 ,   τ 1 = 3.85 ,   τ 2 = 4.2 ,   τ = 0.5
Table 3. The recall ratio compared with state-of-the-art infrared object detection approaches.
Table 3. The recall ratio compared with state-of-the-art infrared object detection approaches.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
IPI0.82350.99250.98790.9420.986
NRAM0.840.90130.92120.8580.8861
GST0.920.68750.94540.9280.986
MOG110.99190.990.7778
TLLCM0.98670.77750.97980.9060.875
WSLCM0.82670.74880.96970.9260.9589
PSTNN0.97300.95250.96150.8680.986
SSD0.99410.97820.95800.95390.9857
YOLO0.99620.98870.96390.96330.9929
The proposed framework0.98650.99880.98380.9381
Table 4. The precision ratio compared with state-of-the-art infrared object detection approaches.
Table 4. The precision ratio compared with state-of-the-art infrared object detection approaches.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
IPI0.80460.99750.89890.90580.6863
NRAM0.98430.99860.86860.85970.7527
GST0.63300.96670.80140.76690.6863
MOG0.431010.46300.47780.7778
TLLCM0.831410.92910.85800.6667
WSLCM0.71260.99670.91430.88200.986
PSTNN10.99870.93270.90610.729
SSD0.99020.99130.91940.92660.9929
YOLO0.97730.99450.92660.91290.9857
The proposed framework110.98980.9380.986
Table 5. The F1 score compared with state-of-the-art infrared object detection approaches.
Table 5. The F1 score compared with state-of-the-art infrared object detection approaches.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
IPI0.81400.99500.94130.92350.8092
NRAM0.90650.94740.89410.90600.8139
GST0.750.80350.86750.83980.8092
MOG0.610.60330.64450.7778
TLLCM0.90240.87480.95380.88130.7568
WSLCM0.76540.85510.94120.92510.9722
PSTNN0.98630.97500.94690.88660.838
SSD0.99210.98540.93160.93740.9893
YOLO0.98870.99210.94710.94150.9893
The proposed framework0.99320.99940.97890.9380.9929
Table 6. The recall ratio of the ablation experiments.
Table 6. The recall ratio of the ablation experiments.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
Without spatial–temporal patch tensor0.90670.93620.93130.8880.9091
Without the contrast-boosted approach0.94670.970.97570.9580.9459
Without the object selection approach0.97330.98120.96760.9320.986
The proposed method0.98650.99880.98380.9381
Table 7. The precision ratio of the ablation experiments.
Table 7. The precision ratio of the ablation experiments.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
Without spatial–temporal patch tensor110.95270.92170.9589
Without the contrast-boosted approach10.99870.93060.87720.9333
Without object selection approach0.8690.90330.86150.82580.8642
The proposed method110.98980.9380.986
Table 8. The average computation time compared with state-of-the-art infrared object-detection approaches.
Table 8. The average computation time compared with state-of-the-art infrared object-detection approaches.
MethodDataset 1Dataset 2
Sequence 3Sequence 4Sequence 16Sequence 18
IPI28.10 s79.76 s84.68 s79.14 s55.99 s
NRAM1.32 s1.57 s1.10 s1.15 s1.21 s
GST0.05 s0.04 s0.02 s0.03 s0.02 s
MOG81.26 s33.21 s60.25 s111.81 s263.86 s
TLLCM1.49 s1.62 s1.54 s1.57 s2.22 s
WSLCM7.16 s6.79 s7.13 s7.25 s9.54 s
PSTNN0.37 s0.26 s0.39 s0.48 s0.31 s
YOLO0.921 s0.421 s0.755 s0.589 s0.673 s
SSD0.04 s0.04 s0.02 s0.03 s0.04 s
The proposed framework1.74 s1.46 s1.38 s1.41 s3.23 s
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Zhu, R.; Zhuang, L. Unsupervised Infrared Small-Object-Detection Approach of Spatial–Temporal Patch Tensor and Object Selection. Remote Sens. 2022, 14, 1612. https://doi.org/10.3390/rs14071612

AMA Style

Zhu R, Zhuang L. Unsupervised Infrared Small-Object-Detection Approach of Spatial–Temporal Patch Tensor and Object Selection. Remote Sensing. 2022; 14(7):1612. https://doi.org/10.3390/rs14071612

Chicago/Turabian Style

Zhu, Ruixi, and Long Zhuang. 2022. "Unsupervised Infrared Small-Object-Detection Approach of Spatial–Temporal Patch Tensor and Object Selection" Remote Sensing 14, no. 7: 1612. https://doi.org/10.3390/rs14071612

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop