Next Article in Journal
A Heterogeneity-Enhancement and Homogeneity-Restraint Network (HEHRNet) for Change Detection from Very High-Resolution Remote Sensing Imagery
Next Article in Special Issue
Infrared Small Dim Target Detection Using Group Regularized Principle Component Pursuit
Previous Article in Journal
Yield Estimation of Wheat Using Cropland Masks from European Common Agrarian Policy: Comparing the Performance of EVI2, NDVI, and MTCI in Spanish NUTS-2 Regions
Previous Article in Special Issue
LMSD-Net: A Lightweight and High-Performance Ship Detection Network for Optical Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Small-Target Detection Based on Background-Suppression Proximal Gradient and GPU Acceleration

1
National Key Laboratory of Optical Field Manipulation Science and Technology, Chinese Academy of Sciences, Chengdu 610209, China
2
Institute of Optics and Electronics, Chinese Academy of Sciences, Chengdu 610209, China
3
School of Electronic, Electrical and Communication Engineering, University of Chinese Academy of Sciences, Beijing 101408, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(22), 5424; https://doi.org/10.3390/rs15225424
Submission received: 1 October 2023 / Revised: 5 November 2023 / Accepted: 15 November 2023 / Published: 20 November 2023
(This article belongs to the Special Issue Remote Sensing of Target Object Detection and Identification II)

Abstract

:
Patch-based methods improve the performance of infrared small target detection, transforming the detection problem into a Low-Rank Sparse Decomposition (LRSD) problem. However, two challenges hinder the success of these methods: (1) The interference from strong edges of the background, and (2) the time-consuming nature of solving the model. To tackle these two challenges, we propose a novel infrared small-target detection method using a Background-Suppression Proximal Gradient (BSPG) and GPU parallelism. We first propose a new continuation strategy to suppress the strong edges. This strategy enables the model to simultaneously consider heterogeneous components while dealing with low-rank backgrounds. Then, the Approximate Partial Singular Value Decomposition (APSVD) is presented to accelerate solution of the LRSD problem and further improve the solution accuracy. Finally, we implement our method on GPU using multi-threaded parallelism, in order to further enhance the computational efficiency of the model. The experimental results demonstrate that our method out-performs existing advanced methods, in terms of detection accuracy and execution time.

1. Introduction

Infrared Small-Target Detection (ISTD) is an important component of infrared search and tracking, aiming to exploit the thermal radiation difference between a target and its background to achieve long-range target detection. According to the definition by the Society of Photo-Optical Instrumentation Engineers (SPIE), small targets typically refers to objects in a 256 × 256 image with an area of fewer than 80 pixels, accounting for approximately 0.12% of the total image area [1]. These small targets usually appear as faint, tiny points, characterized by their diminutive size and a lack of clear texture and shape features. Moreover, the background in infrared images is often affected by random noise, clutter, and environmental factors, making small targets vulnerable to interference. Furthermore, some practical applications have strict requirements on the real-time performance of detection algorithms. Therefore, the rapid and accurate detection of small targets in complex backgrounds poses a significant challenge.
Two primary methods are employed in ISTD for target detection: Tracking-Before-Detection (TBD) and Detection-Before-Tracking (DBT). TBD relies on the temporal information of consecutive frames to capture the movement and features of potential targets. It struggles with stationary or sporadically moving targets and is constrained by computational resources. On the other hand, DBT applies single-frame ISTD to infrared data, identifying potential targets based on features such as contrast and low-rank sparsity. Single-frame infrared small target detection has been widely concerned because of its simple data acquisition, low computational complexity, not affected by target motion, and wide applicability.
The categorization of single-frame ISTD can be determined by the structure of the image; that is, whether (1) the original image or (2) the patch image is used [2]. The first category detects the target directly from the original image; for example, by filtering or Human Vision System (HVS). Filter-based methods [3,4,5,6] have limited utility in ISTD, due to their strict requirements on the background variation and prior knowledge. Meanwhile, HVS-based methods [7,8,9,10,11] use the contrast mechanism to quantify the difference between the target and the background, thereby enhancing small targets. However, these methods are limited by the local saliency of the target, rendering them ineffective when detecting targets that are dark or similar to the background. Some deep learning technologies [12,13,14,15] have recently been applied to this category, but a lack of large data sets limits their performance.
The other category—namely, patch-based methods—transforms small target detection into a low-rank matrix recovery problem [16]. This transformation can circumvent the aforementioned limitations, such as the dependence on prior knowledge and target saliency, as well as the false detection of dark targets. The most popular method is Infrared Patch-Image (IPI) [17], which uses a sliding window technique to generate a corresponding patch image from the original image. Due to its outstanding performance, many studies [18,19,20,21,22,23,24,25,26] have been conducted on IPI, which typically yields superior results. However, patch-based methods still have two problems: (1) The misclassification of strong edges as sparse target components, and (2) the time-consuming nature of the method.
The above-mentioned misclassification arises from the limited ability of the model to distinguish strong edges from sparse components. To address this issue, we propose a Background Suppression Proximal Gradient (BSPG) method, incorporating a novel continuation strategy during the alternating updating of low-rank and sparse components. Our proposed continuation strategy can preserve more components while updating the low-rank matrix, while also reducing the update rate of sparse matrix. As strong edges frequently correspond to larger singular values than targets, the former facilitates the transition of strong edges from sparse components to low-rank components, thereby enabling the model to eliminate the affect of strong edges. Meanwhile, the latter ensures the convergence of the algorithm.
The time-consuming nature of patch-based methods is due to the complex nature of solving the method, mainly including solving the LRSD problem and constructing/reconstructing patch images. To address this issue, we utilize both algorithmic optimizations and hardware enhancements. At the algorithmic level, we propose an approximate partial SVD (APSVD) for efficiently solving the LRSD problem and use a rank estimation method to ensure the accuracy of the solution. At the hardware level, we propose the use of GPU multi-threaded parallelism strategies to expedite the construction and reconstruction modules, as these modules can be decomposed into repetitive and independent sub-tasks.
Our main contributions can be summarized as follows:
  • A novel continuation strategy based on the Proximal Gradient (PG) algorithm is introduced to suppress strong edges. This continuation strategy preserves heterogeneous backgrounds as low-rank components, hence reducing false alarms.
  • The APSVD is proposed for solving the LRSD problem, which is more efficient than the original SVD. Subsequently, parallel strategies are presented to accelerate the construction and reconstruction of patch images. These designs can reduce the computation time at the algorithmic and hardware levels, facilitating rapid and accurate solution.
  • Implementation of the proposed method on GPU is executed and experimentally validate its effectiveness with respect to the detection accuracy and computation time. The obtained results demonstrate that the proposed method out-performs nine state-of-the-art methods.
The remainder of this paper is organized as follows: Section 2 presents the related work. Section 3 details our proposed BSPG algorithm and GPU acceleration strategies. Section 4 introduces the data set and experimental settings, as well as providing the experimental results and analysis. Section 5 discusses the results of the experiments. Section 6 gives a summary of our method.

2. Related Work

2.1. HVS-Based Methods

HVS-based methods detect small targets by utilizing the contrast differences between the target region and its surrounding background. These methods can be categorized based on the type of information they use: grey scale information, gradient information, and a combination of both grey scale and gradient information. Local Contrast Measure (LCM) [1] proposes a novel method for detecting small targets by leveraging grey scale contrast. This method uses a contrast mechanism designed to enhance small targets while effectively suppressing the background noise. Based on the improvement of LCM algorithm, Relative Local Contrast Measure (RLCM) [8], Multiscale Patch-based Contrast Measure (MPCM) [9], Weighted Local Difference Measure (WLDM) [27] and other methods were proposed. Gradient-based contrast methods use first-order or second-order derivatives of the image to extract gradient information. They then utilize this information to design a gradient difference measure that effectively discriminates between small targets and the surrounding background. Building on this concept, Derivative Entropy-based Contrast Measure (DECM) [28] and Local Contrast-Weighted Multidirectional Derivative (LCWMD) [29] propose the use of multidirectional derivative to incorporate more gradient information. In addition, Local Intensity and Gradient (LIG) [30], Gradient-Intensity Joint Saliency Measure (GISM) [31] fuse gradient and intensity information to further highlight small targets. Although HVS-based methods can be effective in many scenarios, they are susceptible to missed detections and false positives in images characterized by low signal-to-clutter ratios and high-intensity backgrounds.

2.2. Deep Learning-Based Methods

In recent years, there has been a significant research focus on deep learning-based methods for infrared small target detection, which seek to achieve high-accuracy detection rates. These deep learning models are trained to discern features within infrared images using vast datasets, thereby enhancing their detection capabilities. To address the problem that infrared small target features are easily lost in deep neural networks, Attention Local Contrast Network (ALCNet) [32] proposes asymmetric contextual modulation to interact the feature information between the high and low levels. Dense Nested Attention Network (DNANet) [15] adequately fuses feature information through densely nested interaction modules to maintain small targets in deep layers. Miss Detection vs. False Alarm (MDvsFA) [33] proposes dual generative adversarial network models, trained inversely to decompose the detection challenge into sub-problems, aiming to strike a balance between miss detections and false alarms. While publicly available datasets have advanced deep learning for infrared small target detection, the scant features of small targets and the dependency on training samples limit the applicability of the model in varied real-world scenarios.

2.3. Patch-Based Methods

A significant amount of research has been conducted to improve the detection ability of IPI [17]. On one hand, some methods have used prior constraints, including Column-Weighted IPI (WIPI) [18], Non-negative IPI with Partial Sum (NIPPS) [20], and Re-Weighted IPI (ReWIPI) [21]. On the other hand, some studies have identified limitations in the nuclear norm and L1 norm and, so, alternative norms to achieve improved target representation and background suppression have been proposed; for example, Non-convex Rank Approximation Minimization (NRAM) [22] and Non-convex Optimization with Lp norm Constraint (NOLC) [23] introduce non-convex matrix rank approximation coupled with L2,1 norm and Lp norm regularization, while Total Variation Weighted Low-Rank (TVWLR) [24], Kernel Robust Principal Component Analysis (KRPCA) [25] introduce total variation regularization, High Local Variance (HLV) [26] method present LV* norm to constrain the background’s local variance. Patch-based methods mainly consider the low-rank nature of the background, affecting their performance in the presence of strong edges. However, our method pays additional attention to heterogeneous background suppression in low-rank constraints, in order to avoid this problem.

2.4. Acceleration Strategies for Patch-Based Methods

Acceleration strategies for patch-based methods can be categorized into algorithm-level and hardware-level acceleration. The first category mainly relies on the strategy of reducing the number of iterations. Self-Regularized Weighted Sparse (SRWS) [34] and NOLC [23] improve the iteration termination condition for acceleration, but still suffer from the time consumption associated to decomposing large matrices. The other category (i.e., hardware acceleration) relies on the use of computationally powerful hardware and efficient parallelization strategies. In [35], the researchers proposed Separable Convolutional Templates (SCT); however, this method has poor performance under complex backgrounds. In addition, extending the patch model to tensor space can also achieve acceleration [36,37,38,39,40,41]. Representative methods in this direction include Re-weighted Infrared Patch-Tensor (RIPT) [36], LogTFNN [39] and the Pareto Frontier Algorithm (PFA) [37]. However, unfolding the tensor into a two-dimensional matrix before decomposition increases the algorithm’s complexity. Partial Sum of the Tensor Nuclear Norm (PSTNN) [38] and Self-Adaptive and Non-Local Patch-Tensor Model (ANLPT) [42] utilize the t-SVD speed up tensor decomposition with t-SVD. However, these methods are limited by the complexity of finding the applicable constrained kernel norm. Our work investigates accelerated patch-based methods at both the algorithmic and hardware levels.

3. Method

In this section, we present the details and principles of the proposed method. First, a novel continuous strategy is proposed for the suppression of strong edges. Then, APSVD is used to accelerate solution of the LRSD problem. Finally, the integration of our proposed method on GPU is presented. The overall framework is shown in Figure 1.

3.1. BSPG Model

The infrared image is considered to be composed of the target image, background image, and noise image, formulated as
f D = f B + f T + f N ,
where f D , f B , f T , and f N represent the original infrared, background, target, and noise images, respectively. The IPI model uses a sliding window (from top-left to bottom-right) to convert the original image into a patch image. The IPI model can be formulated as
D = B + T + N ,
where D, B, T, and N represent the patch image, background patch image, target patch image, and noise patch image, respectively. Then, we can transform the small-target detection problem into the following convex optimization under broad conditions; that is,
m i n B , T B * + λ T 1 , s . t . D B T F δ ,
where . * represents the nuclear norm, . 1 represents the column sum norm, λ is a positive weighting parameter, and δ > 0 . This convex optimization problem is called robust principle component analysis (RPCA), which can recover low-rank and sparse parts of the data matrix even when a fraction of the entries are missing. Let f ( X ) = 1 2 D B T F 2 , P ( X ) = μ ( B * + λ T 1 ) , where μ is a relaxation parameter. Hence, we can express Equation (3) as
m i n F ( X ) = f ( X ) + P ( X ) .
The PG algorithm is an efficient method to solve the RPCA problem, which estimates the background image and the target image by minimizing the separable quadratic approximation sequence of Equation (4); that is,
Q ( X , Y ) f ( Y ) + f ( Y ) , X Y + τ 2 X Y F 2 + P ( X ) = τ 2 ( X G ) F 2 + P ( X ) + f ( Y ) 1 2 τ Y F 2 ,
where G = Y 1 L f f , L f is the Lipschitz constant (which is set to 2 in this problem), and τ > 0 is a given parameter. The following function has a unique optimal solution as Equation (5) is convex:
a r g m i n { Q τ ( X , Y | X d o m ( P ) ) } ,
where d o m ( P ) = X | P ( X ) < + . In our method, Equation (5) can be expressed as:
Q ( B , T , μ , Y B , Y T ) = τ 2 B Y B | F 2 + τ 2 T Y T | F 2 + f ( Y B , Y T ) + μ P ( B , T ) + 1 2 τ | f ( Y B , Y T ) | F 2 .
To solve Equation (6), the iterative process of the PG algorithm repeatedly sets X k + 1 = a r g m i n Q ( X , Y k ) , and Y k is obtained from X 0 , X 1 , X k . In our method, X K to be solved are ordered pairs ( B k , T k ) . Therefore, we set
Y k = X k + t k 1 1 t k ( X k X k 1 ) ,
where t k is the sequence satisfying t k + 1 2 t k + 1 t k 2 .
The closed-form expression of X k + 1 can be obtained by soft-thresholding the singular values. The soft-threshold operation is defined as
S ε [ x ] x ε , i f x > ε , x + ε , i f x < ε , 0 , o t h e r w i s e .
Then,
T k + 1 = S τ [ G k T ] = S λ μ L f [ G k T ] , B k + 1 = S τ [ G k B ] = U S μ L f V T ,
where U S V T is the SVD of G k B .
The continuation strategy in [43] can speed up the convergence of the PG. This technique employs a decreasing sequence to derive μ ¯ , where μ k is updated as follows:
μ k = m a x { η μ k 1 , μ ¯ } , 0 < η < 1 .
The value of η affects the convergence of the algorithm. A larger decrease results in more G k components to be retained, while fewer iterations to update μ ¯ results in an inability to separate the target. Conversely, a smaller η decrease has the opposite effect. Figure 2 shows examples of the target images at different iterations. It can be seen that, in the early iterations, the strong edges are separated first as the low component of the background is retained at the highlights. This leads to many false alarms at the strong edges when the target is separated. This phenomenon motivates us to use different decrement rates for G k B and G k T . We set a higher rate for μ B and set G k + 1 B to compute more singular values, in order to retain strong edges. We also set a lower rate for μ T , ensuring that the target is decomposed into sparse parts when the algorithm converges. Thus, μ k is updated as follows:
μ k T = m a x { α μ k 1 T , μ ¯ } , μ k B = m a x { β μ k T , μ ¯ } ,
where 0 < α , β < 1 . The solution of the BSPG algorithm is described in Algorithm 1.
The upper bound of the algorithm is discussed below. By denoting { X k , Y k , t k } as the sequence obtained by the algorithm with t k k + 2 2 , according to [44], for any k 1 , we have
F ( X k ) F ( X * ) 2 L f X * X 0 F 2 ( k + 1 ) 2 , X * 0 .
Then,
F ( X k ) F ( X * ) ϵ ,
when k > k 0 + L f ϵ X k 0 X * F , where the convergence accuracy is ϵ > 0 , yielding that the algorithm has O L f ϵ iteration complexity.
Algorithm 1: BSPG solution via APSVD
Remotesensing 15 05424 i001

3.2. APSVD

The most time-consuming step in each iteration of the PG algorithm is the execution of the full SVD. It is worth mentioning that the soft-threshold operation only leaves a portion of the singular values and vectors to participate in the subsequent calculations. In particular, few singular values are needed in early iterations. Therefore, it is feasible to replace full SVD with partial SVD. The crucial step of this strategy is rank estimation, which involves estimating the number of singular values and singular vectors participating in the computation after truncation. As the noise in infrared images is often not simply Gaussian distributed, estimation functions such as minimax estimator or the simple quantitative increase method proposed in [43] are not feasible. Due to the low-rank nature of the patch image, the singular value matrix has a clear trend of change, as shown in Figure 3. Thus, we estimate the rank by evaluating the degree of variation of the singular values. In the kth iteration, the pre-determined rank s v k 1 is initialized by the number of singular values in S greater than μ k / L f . We update s v k as follows:
s v k = s v k 1 + 5 , i f σ s v k σ s v k 1 < δ , s v k 1 + γ N , o t h e r w i s e ,
where δ is the threshold for measuring the degree of singular value variation, N is the width of the patch image, and δ and γ are empirically set to 0.95 and 0.1, respectively.
The truncated SVD of the patch image needs to satisfy two requirements: (1) Only a small number of large singular values are retained, and (2) the singular vectors (including the left and right) need to be computed. We approximate the SVD of the patch image A by solving the eigenvalue decomposition of its covariance matrix A T A . Given the slender nature of the patch matrix, the latter computation is considerably more straightforward compared to the former. Moreover, the symmetry of A T A guarantees its suitability for eigenvalue decomposition, leading to
A = U S V T , A T A = V S 2 V T .
It is apparent, from Equation (16), that the eigenvalues of A T A correspond to the squares of the singular values of A. Additionally, both the right singular vectors of A and the eigenvectors of A T A are unitary. This relationship further implies that the left singular vectors U follow the equation U = A V S 1 . The approximate SVD of matrix A satisfies
U T A V = S + ξ ,
where ξ depends on the rounding errors. The singular values and singular vectors of the eigendecomposition in CUDA can approximate the accuracy of SVD to machine zero [45]. Furthermore, the larger the singular value, the smaller the error of U.

3.3. GPU Parallel Implementation

The GPU implementation consists of three main parts: Constructing the patch image, solving the LRSD problem, and reconstructing the patch image. Our method uses GPU for implementation purposes, and CPU only for data transfer and GPU control.

3.3.1. Construction

In order to reduce the number of data transfers between the host memory and the device’s global memory, we first copy all the image data and hyperparameters read by the CPU to the GPU via PCI Express, and then execute the parallelization kernel functions. GPU parallelism mainly relies on data parallelism, i.e., performing the same operation on multiple data elements. Correspondingly, a patch image is constructed to change the storage location of each pixel in the original image. Therefore we build an index mapping between the original image and the patch image. This mapping allows a thread to manage the correspondence of a pixel position, facilitating the parallel processing of all pixels.
Let d w , d h denote the width and height of the sliding window, s x , s y denote the sliding step in the x, y directions, and p denotes the number of patches in the x direction. The mapping of the patch image pixel index I x p to the original image pixel index I x o is
I x o = I x p % p × s y + I y p d h , I y o = I x p p × s x + I y p % d h .
The execution of the kernel function requires the determination of the thread block and grid size, where the grid size is determined based on the number of processing subtasks and the thread block size. Suppose the output image size n of the kernel function is n x × n y and the thread block size k is set to k x × k y (where n x , n y denote the size of the image in the x and y direction, respectively, and k x , k y denote the number of threads per block in the x and y direction, respectively), then the grid size is determined as ( ( n x ) k x , ( n y ) k y ) , where operator ( ) k is defined as
( n ) k = k n k , k N , n R .
We set the thread block size for the construction kernel function based on the size of the patch image. Assuming that the patch image dimensions are p x × p y , we configure the thread block size as ( p x , 1024 / p x ) . For instance, given an image with a size of 200 × 150 , a sliding window with a size of 50 × 50 , and a sliding step of 10, the size of the resulting patch image would be 176 × 2500 . Accordingly, the thread block size is set to ( 176 , 5 ) and the grid dimensions are set to ( 1 , 500 ) . Consequently, a row of threads in the x direction corresponds to a row within the patch image. The construction process is carried out pixel-by-pixel. The processing time of each pixel is assumed to be t, for an M × N patch image, serial execution of the construction module takes M × N × t . In contrast, our method operates M × N threads in parallel, completing the process in time t. The theoretical speedup ratio is M × N .

3.3.2. Reconstruction

Figure 4 shows the steps for reconstructing the background and target patch images after LRSD. First, the target patch image is transformed into the pre-filtered image. We provide the pseudo-code for this transformation in Algorithm 2. Then, the indices of the first and last patches containing valid information are determined. Finally, filtering is performed on the valid portions of each row to obtain the target image. In summary, the reconstruction includes two parallel processes: One involving mapping from the patch image to the pre-filter image, and another entailing filtering.
Algorithm 2: The mapping of patch image and pre-filter image
Input: Patch image D, original image size w and h, patch size d w and d h , step s, patch number of per row p r
Output: pre-filter image F
 1 
Step I: Compute the index I D in the patch image by using the row and column numbers R D , C D .
 2 
R D = b l o c k I d x . y × b l o c k D i m . y + t h r e a d I d x . y ;
 3 
C D = b l o c k I d x . x × b l o c k D i m . x + t h r e a d I d x . x ;
 4 
I D = C D × d w × d h + R D ;
 5 
Step II: Compute the row index I p r and column index I p c of the patch by using patch index I p .
 6 
I p = I D / ( d w × d h ) ;
 7 
I p r = I p / p r ; I p c = I p % p r ;
 8 
Step III: Compute the index I F in F by using the row number R O and column number C O in the original image O.
 9 
R O = I p r × s + R D % d h ;
10 
C O = I p c × s + R D / d h ;
11 
I F = C D × w × h + C O × h + R O ;
12 
F [ I F ] = D [ I D ] .
We set the thread block and grid of the mapping kernel to the same size as the construction kernel. The filter kernel handles a much larger matrix, and in order to improve the resource usage, the thread block size is set to ( 32 , 32 ) , and the grid size is obtained according to Equation (19). Then, the indices of the first patch and the last patch I p f , I p l containing valid information can be expressed as:
I p f = I x o d w s x + 1 + ( I y o d h s y + 1 ) × p , I p l = I x o s x + 1 + ( I y o s y + 1 ) × p .
In NVIDIA’s GPU architecture, one warp typically consists of 32 threads, while our thread block contains 32 × 32 threads. This means that each warp can effectively execute an entire thread block. This high warp occupancy rate reaches 100%, efficiently harnessing the performance of the GPU.

3.3.3. APSVD Using CUDA

The key to implementing the APSVD is the exact eigendecomposition, which can be achieved using the Symmetric Eigenvalue Divide (SYEVD) function based on QR decomposition or the Symmetric Eigenvalue Jacobi (SYEVJ) function based on Jacobi decomposition [46]. SYEVD employs a divide-and-conquer method to decompose a symmetric matrix into smaller sub-problems and solves them recursively. Its runtime is primarily attributed to QR decomposition. QR decomposition can be expressed as
A m × n = Q m × n R n × n ,
where Q is an orthogonal matrix and R is an upper triangular matrix. QR decomposition usually requires Householder transformations for multiple iterations, and each transformation needs to manipulate all elements of the matrix, which becomes redundant for small matrices.
SYEVJ transforms the symmetric matrix A into a diagonal matrix D by performing a rotational transformation via the bilateral Jacobi method.
D = J 3 T ( J 2 T ( J 1 T A J 1 ) J 2 ) J 3 = ( J 3 T J 2 T J 1 T ) A ( J 1 J 2 J 3 ) ,
where J is denoted as J ( i , j , θ ) , contains the rotation angle θ and an index pair ( i , j ) , and satisfies J ( i , j , θ ) T J ( i , j θ ) = E . The Jacobi method, with its element-wise rotations, offers lower computational complexity, localized memory access, and parallelization potential, making it more efficient for small matrices.
To quantitatively analyze both methods, we introduce arithmetic intensity [46] which is a metric used to evaluate the performance of parallel computational tasks. Specifically, the arithmetic intensity I is defined as
I = FLOPs bytes loaded ,
where FLOPs represents the number of floating point operations and can measure the complexity of an algorithm, bytes loaded represents the number of bytes loaded from memory during kernel execution. Given an N × N matrix (single-precision), a Givens rotation typically requires 8 floating-point operations (2 trigonometric functions, 4 multiplications, 2 additions) and loads 2 N elements. This means that the number of FLOPS per iteration is 8, and the memory access requires loading 8 N bytes. In QR decomposition, the computational complexity of each iteration, which involves Householder transformations, is 2 N 3 , and it loads the entire matrix, including 4 N 2 bytes. The arithmetic intensity of the Jacobi kernel I J and the arithmetic intensity of the QR decomposition kernel I Q R can be expressed as
I J = 8 2 N , I Q R = 2 N 3 4 N 2 .
Therefore, from the perspective of arithmetic intensity, we choose the more efficient Jacobi method to perform eigenvalue decomposition. The Jacobi method typically has lower arithmetic intensity and is relatively memory-access efficient, whereas QR decomposition involves orthogonal transformations and matrix updates, often requiring more memory bandwidth and computation. QR decomposition has a higher arithmetic intensity, making it perform better on larger matrices where the high arithmetic intensity can be fully utilized, but not on small matrices.
Furthermore, the implementation of APSVD requires an efficient matrix multiplication function. The General Element-wise Matrix Multiply (GEMM) function in CUDA takes advantage of the GPU’s parallel computing capabilities and efficiently processes substantial amounts of data, thereby enhancing computational performance. To reduce routing errors, we use the double precision-controlled GEMM function, DGEMM, which has a time complexity of 2 M N 2 . Notably, other functions utilize single precision to strike a balance between instruction throughput and accuracy. Figure 5 illustrates the runtime ratios for each component of APSVD on matrices of varying size. Notably, the efficiency of SYEVJ is demonstrated, as it is unaffected by the matrix height and exhibits a decreasing ratio of time spent on eigendecomposition as the matrix size increases.

4. Experiments and Analysis

In this section, we provide an evaluation of our method in terms of detection accuracy and execution time.

4.1. Experimental Setup

Data. The experiments used the real single-frame infrared images provided in [32], selected from infrared sequences in a variety of scenes, including ocean, cloud, sky, and urban areas, as shown in Figure 6. The targets are marked with red boxes and magnified for convenient viewing in the bottom left corner of each image. It can be seen that the targets occupy very few pixels; most small targets lack shape and texture information and have low intensity. Images with poor imaging quality exhibit strong noise, such as SIR_2 and SIR_3. Some small targets are submerged in cloud or sea clutter as SIR_11 and SIR_13 and suffered from highlight backgrounds with strong edges as SIR_8 to SIR_14. Furthermore, we provide detailed information of the test images, including the background type, target type, Signal Clutter Ratio (SCR), target size, and detection challenges, in Table 1.
Hardware. We implemented our method on the embedded GPU Jetson AGX Xavier, which has 7764 MB of global memory and 48 KB of shared memory. The version of CUDA was 10.2. The experiments in MATLAB were based on an Intel(R) Core(TM) i7-8750H CPU with 8 GB RAM.
Baselines and parameter settings. We compared our proposed method to other state-of-the-art patch-based methods, including IPI [17], NIPPS [20], NRAM [22], NOLC [23], SRWS [34] and HLV [26]. As tensor-based methods have better performance in terms of computational efficiency, we also included three tensor-based methods for comparison, including RIPT [36], PSTNN [38], PFA [37], LogTFNN [39] and ANLPT [42]. The parameter settings are provided in Table 2. We employed a sliding window with a size of 100 × 100 and a step size of 30 on images with resolution equal to or exceeding 640 × 512 . This setting ensured that the execution time remained within the desired range.
Evaluation metrics. We used two quantitative analysis evaluation indicators commonly used for small-target detection to evaluate our method: Signal Clutter Ratio Gain (SCRG) and Background Suppress Factor (BSF). SCRG reflects the effect of increasing target saliency, and is defined as follows:
S C R G = S C R o u t S C R i n , S C R = | μ t μ b | σ b ,
where S C R i n and S C R o u t represent the signal-to-clutter ratio of the input and output images, respectively, μ t represents the average pixel gray value of the target, μ b represents the average pixel gray value of the local background around the target, and σ b represents the standard deviation of the gray pixel value of the local background around the target. BSF reflects the effect of suppressing background interference and is defined as follows:
B S F = σ i n σ o u t ,
where σ o u t and σ i n are the standard deviation values of the local background around the target in the output image and the original image, respectively.
We also analyzed the results using Receiver Operating Characteristic (ROC) curves. The ROC curve is plotted by assessing the True Positive Rate (TPR) and the False Positive Rate (FPR) at different classification thresholds, as defined below:
T P R = n u m b e r o f r e a l t a r g e t s d e t e c t e d n u m b e r o f r e a l t a r g e t s , F P R = n u m e r o f f a l s e t a r g e t s d e t e c t e d n u m b e r o f r e a l t a r g e t s .
To quantitatively compare the ROC curves, the area under the curve (AUC) can be used as an evaluation criterion; the larger the AUC, the more accurate the detection performance.

4.2. Visual Comparison with Baselines

The visualization results of twelve detection methods are shown in Figure 7 and Figure 8. IPI, RIPT, and NIPPS only successfully detect targets under simple backgrounds with high local patch similarity. In images with complex backgrounds, NIPPS present noticeable clutter, while IPI and RIPT exhibit noise in highlighted backgrounds. NRAM and PFA can detect the majority of targets, but they are prone to generating false alarms in regions with strong edges. NOLC and SRWS suffer from the sea surface background and can not detect weak dark targets. PSTNN has similar poor performance, with many false alarms under clutter. LogTFNN is poorly detected under high-intensity backgrounds, leaving a large amount of background residue. HLV and ANLPT perform poorly when detecting targets that are dark or have low contrast with the neighboring background. In the case of images SIR_7 to SIR_14, the highlighted backgrounds cause most methods to produce false alarms near strong edges, leading to inaccurate detection. However, our method excels in terms of effectively suppressing strong edges under such conditions. From the visualized detection results, our method exhibits robust detection performance.

4.3. Quantitative Evaluation and Analysis

The results of the quantitative comparison of the various methods on the test images are shown in Table 3. From the definitions of the two evaluation metrics, the larger the SCRG and BSF values, the better the detection results. In terms of BSF, our method performs better than the other methods on all images, indicating that our method excels in suppressing the background. In terms of SCRG, our method outperforms the other methods on most images. When the target is missed, the local background standard deviation is 0. Consequently, this leads to the SCRG appearing as a NaN result and BSF tending to infinity. In addition, we plotted the ROC curves corresponding to the experiments, in order to further validate the effectiveness of our method. The results in Figure 9 demonstrate that our method outperforms the other comparative methods on the test images. IPI and NIPPS are sensitive to clutter and high-intensity backgrounds due to the difficulty in distinguishing between targets and strong edges. NRAM and NOLC exhibit instability in detecting dim and weak targets. HLV experiences an increase in false alarms when dealing with strong clutter interference on the sea surface. Tensor-based detection methods show stable performance in simple backgrounds. However, RIPT and LogTFNN exhibit low detection accuracy in high-intensity backgrounds due to their high requirements for sparsity. PSTNN and PFA tend to erroneously reject targets in the presence of background clutter. SRWS demonstrates effective clutter suppression in high-intensity backgrounds but struggles to detect low-contrast weak targets. ANLPT exhibits weaker clutter suppression capabilities in high-intensity structured backgrounds. In contrast, our method achieves remarkable results in terms of detection accuracy and false alarm rate in a wide range of scenarios.
To evaluate the execution speed of our method, we compared our method with other patch-based methods. For fair comparison, we set these methods to share the same patch and step configurations. We implemented our method in two versions; that is, CPU and GPU versions. The GPU execution time includes the data transfer time between the host and the device. To ensure the reliability of the time statistics, we took the average time of 10 executions for each method. The comparison results are presented in Table 4. Most patch-based methods are time-consuming due to iteration and complex matrix decomposition, including IPI, NIPPS, and NRAM. RIPT and ANLPT relatively improve the detection efficiency, but the tensor decomposition is still complex. SRWS and HLV optimise the iterative termination conditions to achieve a faster detection speed. The results demonstrate that our method achieved impressive speed, particularly with significant acceleration when using the GPU. Combined with the previous detection accuracy evaluation, our method was found to achieve faster detection while maintaining higher accuracy.

4.4. Ablation Study

The effect of relaxation parameters. We explored the effects of the relaxation parameters α and β on the detection accuracy of our method using ROC curves. Figure 10 shows that excessively large or small values of α and β led to a decrease in the AUC value. Small α can retain more background components, but overly small values result in insufficient iteration, thereby failing to separate targets. Meanwhile, large α and β values can cause false alarms by decomposing strong edges into sparse target portions. Therefore, we set α and β to 0.4 and 0.7, respectively.
Comparison with other tensor-based methods. For a fair comparison, we studied the tensor-based methods under the same settings as used for ours. We evaluated their detection accuracy and execution time under various patch and step configurations. Table 5 shows the execution time results. When using the same patch and step settings, our method on GPU is faster than PFA, PSTNN, and LogTFNN. As illustrated in Figure 11, our method consistently achieves the best detection accuracy across most scenarios. However, the detection accuracy of PFA, PSTNN, and LogTFNN significantly diminishes between images with variations in the patch size and step values.
The acceleration effect of different strategies. To validate the speed enhancement due to our proposed APSVD, we compared its implementations on both MATLAB 2017b and CUDA 10.2 platforms with various SVD functions, as shown in Table 6. It can be observed that APSVD exhibits high efficiency on slender matrices. To explore the effectiveness of the proposed acceleration strategies, we conducted tests on their cumulative acceleration effects over the baseline method IPI. Table 7 demonstrates that our proposed acceleration strategies yield commendable speed increases. Notably, PASVD avoids the intricate decomposition of large matrices, thus significantly saving time; the new continuation strategy employs a greater decrement for the relaxation parameters, reducing the number of iterations and consequently accelerating the overall speed; and the GPU parallel strategies provide significant acceleration, especially for larger images.
The acceleration effects on images with different attributes. To validate the acceleration effect of the proposed method, we conducted experiments on images with varying attributes (i.e., resolution and background complexity). As shown in Figure 12, the acceleration effect of our method becomes more pronounced with increasing image resolution. On an image with a resolution of 1024 × 1020 , the execution speed of the proposed method is nearly 60 times faster than that of IPI. Due to the influence of image complexity on execution time, the acceleration effect varies slightly at the same resolution. Furthermore, we conducted a comparative analysis of the three stages of our method—namely, constructing a patch image, solving the LRSD problem, and reconstructing a patch image—as shown in Figure 13. It is evident that multi-threading parallelism and optimized memory access significantly reduce the time required for the construction and reconstruction modules. Additionally, the new continuation strategy and APSVD greatly contribute to reducing the time required to solve the LRSD problem.

5. Discussion

Patch-based methods are well-studied in single-frame infrared small target detection for their reliability. The classical IPI algorithm is a notable example. It transforms the original image into a patch image and leverages the non-local self-similarity of the background to enhance the low-rank property of the patch image. This method allows for effective infrared small target detection through low-rank and sparse decomposition.
In our comparison of small target detection methods, we observed that methods like NIPPS, NRAM, and NOLC aim to improve detection accuracy by enhancing the nuclear norm and l1 norm. However, these methods involve complex matrix decomposition and iterative processes, leading to time-consuming issues. These methods also struggle to differentiate between the edges and the actual targets due to the local sparsity of strong edges. Conversely, SRWS and HLV, with their proposed multi-subspace assumptions and high local variance constraints, generally perform well in most cases. They effectively suppress strong edges but may miss dark and weak targets. Additionally, these methods require complex matrix decomposition, making them time-consuming. RIPT expands the patch model into tensor space, adding to the computational burden as tensors are unfolded and decomposed. While tensor-based methods such as PSTNN, PFA, and logTFNN have accelerated detection somewhat, their effectiveness is limited by the challenges of accurately approximating nuclear norms within tensor models.
This paper aims to strike a balance between detection performance and time consumption. To address interference from strong edges, the BSPG method proposed in this paper introduces a novel continuous strategy in the alternating update process of low-rank and sparse components. This allows the model to mitigate the influence of strong edges by preserving more components while updating the low-rank matrix. For algorithm acceleration, a combined approach involving algorithm optimization and hardware enhancement is presented. On the algorithmic front, APSVD is introduced to expedite solving the LRSD problem. On the hardware front, we suggest utilizing GPU multi-thread parallel strategies to accelerate the construction and reconstruction of modules. This is possible as these modules can be decomposed into repetitive and independent subtasks. Visual and quantitative results from experiments demonstrate that our method outperforms other state-of-the-art methods. However, there is still room for improvement in terms of time performance, and in the future, we plan to explore even faster methods.

6. Conclusions

In this paper, we proposed a novel infrared small-target detection method using background-suppression proximal gradient and GPU parallelism. Considering that patch-based methods often result in false alarms at strong edges, we first proposed a novel continuation strategy to suppress such background interference. Then, we presented APSVD to accelerate the solution of the LRSD problem, which involves complex and time-consuming large matrix decomposition. Moreover, we employed GPU multi-threading parallelism to accelerate the construction and reconstruction of patch images. Finally, we optimized the proposed method on the GPU, ultimately achieving outstanding performance. The obtained experimental results demonstrated that our method out-performs nine state-of-the-art methods in terms of both detection accuracy and computational efficiency. The proposed GPU parallelism strategy can be applied to infrared motion sensors and other patch-based infrared small-target detection methods, thus facilitating their application in practical engineering.

Author Contributions

Conceptualization, X.H.; data curation, X.H.; investigation, T.L.; methodology, X.H.; software, X.H. and Y.L.; writing—original draft, X.H.; writing—review & editing, X.L. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the CAS “Light of West China” Program.

Data Availability Statement

The data presented in this study are cited within the article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chen, C.P.; Li, H.; Wei, Y.; Xia, T.; Tang, Y.Y. A local contrast method for small infrared target detection. IEEE Trans. Geosci. Remote Sens. 2013, 52, 574–581. [Google Scholar] [CrossRef]
  2. Zhang, C.; He, Y.; Tang, Q.; Chen, Z.; Mu, T. Infrared Small Target Detection via Interpatch Correlation Enhancement and Joint Local Visual Saliency Prior. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5001314. [Google Scholar] [CrossRef]
  3. Bai, X.; Zhou, F. Analysis of new top-hat transformation and the application for infrared dim small target detection. Pattern Recognit. 2010, 43, 2145–2156. [Google Scholar] [CrossRef]
  4. Zhao, Y.; Pan, H.; Du, C.; Peng, Y.; Zheng, Y. Bilateral two-dimensional least mean square filter for infrared small target detection. Infrared Phys. Technol. 2014, 65, 17–23. [Google Scholar] [CrossRef]
  5. Deshpande, S.D.; Er, M.H.; Venkateswarlu, R.; Chan, P. Max-mean and max-median filters for detection of small targets. In Proceedings of the Signal and Data Processing of Small Targets, Denver, CO, USA, 20–22 July 1999; Volume 3809, pp. 74–83. [Google Scholar]
  6. Liu, X.; Li, L.; Liu, L.; Su, X.; Chen, F. Moving dim and small target detection in multiframe infrared sequence with low SCR based on temporal profile similarity. IEEE Geosci. Remote Sens. Lett. 2022, 19, 7507005. [Google Scholar] [CrossRef]
  7. Qin, Y.; Li, B. Effective infrared small target detection utilizing a novel local contrast method. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1890–1894. [Google Scholar] [CrossRef]
  8. Han, J.; Liang, K.; Zhou, B.; Zhu, X.; Zhao, J.; Zhao, L. Infrared small target detection utilizing the multiscale relative local contrast measure. IEEE Geosci. Remote Sens. Lett. 2018, 15, 612–616. [Google Scholar] [CrossRef]
  9. Wei, Y.; You, X.; Li, H. Multiscale patch-based contrast measure for small infrared target detection. Pattern Recognit. 2016, 58, 216–226. [Google Scholar] [CrossRef]
  10. Chen, Y.; Zhang, G.; Ma, Y.; Kang, J.U.; Kwan, C. Small infrared target detection based on fast adaptive masking and scaling with iterative segmentation. IEEE Geosci. Remote Sens. Lett. 2021, 19, 7000605. [Google Scholar] [CrossRef]
  11. Cui, H.; Li, L.; Liu, X.; Su, X.; Chen, F. Infrared small target detection based on weighted three-layer window local contrast. IEEE Geosci. Remote Sens. Lett. 2021, 19, 7505705. [Google Scholar] [CrossRef]
  12. Du, J.; Lu, H.; Hu, M.; Zhang, L.; Shen, X. CNN-based infrared dim small target detection algorithm using target-oriented shallow-deep features and effective small anchor. IET Image Process. 2021, 15, 1–15. [Google Scholar] [CrossRef]
  13. Du, J.; Lu, H.; Zhang, L.; Hu, M.; Chen, S.; Deng, Y.; Shen, X.; Zhang, Y. A spatial-temporal feature-based detection framework for infrared dim small target. IEEE Trans. Geosci. Remote Sens. 2021, 60, 3000412. [Google Scholar] [CrossRef]
  14. Zhang, M.; Dong, L.; Ma, D.; Xu, W. Infrared target detection in marine images with heavy waves via local patch similarity. Infrared Phys. Technol. 2022, 125, 104283. [Google Scholar] [CrossRef]
  15. Li, B.; Xiao, C.; Wang, L.; Wang, Y.; Lin, Z.; Li, M.; An, W.; Guo, Y. Dense nested attention network for infrared small target detection. IEEE Trans. Image Process. 2022, 32, 1745–1758. [Google Scholar] [CrossRef]
  16. Zhong, S.; Zhou, H.; Cui, X.; Cao, X.; Zhang, F. Infrared small target detection based on local-image construction and maximum correntropy. Measurement 2023, 211, 112662. [Google Scholar] [CrossRef]
  17. Gao, C.; Meng, D.; Yang, Y.; Wang, Y.; Zhou, X.; Hauptmann, A.G. Infrared patch-image model for small target detection in a single image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef] [PubMed]
  18. Dai, Y.; Wu, Y.; Song, Y. Infrared small target and background separation via column-wise weighted robust principal component analysis. Infrared Phys. Technol. 2016, 77, 421–430. [Google Scholar] [CrossRef]
  19. Wang, X.; Peng, Z.; Kong, D.; Zhang, P.; He, Y. Infrared dim target detection based on total variation regularization and principal component pursuit. Image Vis. Comput. 2017, 63, 1–9. [Google Scholar] [CrossRef]
  20. Dai, Y.; Wu, Y.; Song, Y.; Guo, J. Non-negative infrared patch-image model: Robust target-background separation via partial sum minimization of singular values. Infrared Phys. Technol. 2017, 81, 182–194. [Google Scholar] [CrossRef]
  21. Guo, J.; Wu, Y.; Dai, Y. Small target detection based on reweighted infrared patch-image model. IET Image Process. 2018, 12, 70–79. [Google Scholar] [CrossRef]
  22. Zhang, L.; Peng, L.; Zhang, T.; Cao, S.; Peng, Z. Infrared small target detection via non-convex rank approximation minimization joint l 2, 1 norm. Remote Sens. 2018, 10, 1821. [Google Scholar] [CrossRef]
  23. Zhang, T.; Wu, H.; Liu, Y.; Peng, L.; Yang, C.; Peng, Z. Infrared small target detection based on non-convex optimization with Lp-norm constraint. Remote Sens. 2019, 11, 559. [Google Scholar] [CrossRef]
  24. Chen, X.; Xu, W.; Tao, S.; Gao, T.; Feng, Q.; Piao, Y. Total Variation Weighted Low-Rank Constraint for Infrared Dim Small Target Detection. Remote Sens. 2022, 14, 4615. [Google Scholar] [CrossRef]
  25. Yan, F.; Xu, G.; Wu, Q.; Wang, J.; Li, Z. Infrared small target detection using kernel low-rank approximation and regularization terms for constraints. Infrared Phys. Technol. 2022, 125, 104222. [Google Scholar] [CrossRef]
  26. Liu, Y.; Liu, X.; Hao, X.; Tang, W.; Zhang, S.; Lei, T. Single-Frame Infrared Small Target Detection by High Local Variance, Low-Rank and Sparse Decomposition. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5614317. [Google Scholar] [CrossRef]
  27. Deng, H.; Sun, X.; Liu, M.; Ye, C.; Zhou, X. Entropy-based window selection for detecting dim and small infrared targets. Pattern Recognit. 2017, 61, 66–77. [Google Scholar] [CrossRef]
  28. Bai, X.; Bi, Y. Derivative entropy-based contrast measure for infrared small-target detection. IEEE Trans. Geosci. Remote Sens. 2018, 56, 2452–2466. [Google Scholar] [CrossRef]
  29. Xu, Y.; Wan, M.; Zhang, X.; Wu, J.; Chen, Y.; Chen, Q.; Gu, G. Infrared Small Target Detection Based on Local Contrast-Weighted Multidirectional Derivative. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5000816. [Google Scholar] [CrossRef]
  30. Zhang, H.; Zhang, L.; Yuan, D.; Chen, H. Infrared small target detection based on local intensity and gradient properties. Infrared Phys. Technol. 2018, 89, 88–96. [Google Scholar] [CrossRef]
  31. Li, Y.; Li, Z.; Li, W.; Liu, Y. Infrared Small Target Detection Based on Gradient-Intensity Joint Saliency Measure. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2022, 15, 7687–7699. [Google Scholar] [CrossRef]
  32. Dai, Y.; Wu, Y.; Zhou, F.; Barnard, K. Attentional local contrast networks for infrared small target detection. IEEE Trans. Geosci. Remote Sens. 2021, 59, 9813–9824. [Google Scholar] [CrossRef]
  33. Wang, H.; Zhou, L.; Wang, L. Miss detection vs. false alarm: Adversarial learning for small object segmentation in infrared images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 8509–8518. [Google Scholar]
  34. Zhang, T.; Peng, Z.; Wu, H.; He, Y.; Li, C.; Yang, C. Infrared small target detection via self-regularized weighted sparse model. Neurocomputing 2021, 420, 124–148. [Google Scholar] [CrossRef]
  35. Wu, X.; Zhang, J.Q.; Huang, X.; Liu, D.L. Separable convolution template (SCT) background prediction accelerated by CUDA for infrared small target detection. Infrared Phys. Technol. 2013, 60, 300–305. [Google Scholar] [CrossRef]
  36. Dai, Y.; Wu, Y. Reweighted infrared patch-tensor model with both nonlocal and local priors for single-frame small target detection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 3752–3767. [Google Scholar] [CrossRef]
  37. Xu, L.; Wei, Y.; Zhang, H.; Shang, S. Robust and fast infrared small target detection based on pareto frontier optimization. Infrared Phys. Technol. 2022, 123, 104192. [Google Scholar] [CrossRef]
  38. Zhang, L.; Peng, Z. Infrared small target detection based on partial sum of the tensor nuclear norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  39. Kong, X.; Yang, C.; Cao, S.; Li, C.; Peng, Z. Infrared small target detection via nonconvex tensor fibered rank approximation. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5000321. [Google Scholar] [CrossRef]
  40. Wang, G.; Tao, B.; Kong, X.; Peng, Z. Infrared Small Target Detection Using Nonoverlapping Patch Spatial—Temporal Tensor Factorization With Capped Nuclear Norm Regularization. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5001417. [Google Scholar] [CrossRef]
  41. Li, J.; Zhang, P.; Zhang, L.; Zhang, Z. Sparse Regularization-Based Spatial-Temporal Twist Tensor Model for Infrared Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5000417. [Google Scholar] [CrossRef]
  42. Zhang, Z.; Ding, C.; Gao, Z.; Xie, C. ANLPT: Self-Adaptive and Non-Local Patch-Tensor Model for Infrared Small Target Detection. Remote Sens. 2023, 15, 1021. [Google Scholar] [CrossRef]
  43. Toh, K.C.; Yun, S. An accelerated proximal gradient algorithm for nuclear norm regularized linear least squares problems. Pac. J. Optim. 2010, 6, 15. [Google Scholar]
  44. Lin, Z.; Ganesh, A.; Wright, J.; Wu, L.; Chen, M.; Ma, Y. Fast Convex Optimization Algorithms for Exact Recovery of a Corrupted Low-Rank Matrix. Coordinated Science Laboratory Report no. UILU-ENG-09-2214, DC-246. 2009. Available online: https://people.eecs.berkeley.edu/~yima/matrix-rank/Files/rpca_algorithms.pdf (accessed on 30 September 2023).
  45. Ordóñez, Á.; Argüello, F.; Heras, D.B.; Demir, B. GPU-accelerated registration of hyperspectral images using KAZE features. J. Supercomput. 2020, 76, 9478–9492. [Google Scholar] [CrossRef]
  46. Seznec, M.; Gac, N.; Orieux, F.; Naik, A.S. Real-time optical flow processing on embedded GPU: An hardware-aware algorithm to implementation strategy. J. Real Time Image Process. 2022, 19, 317–329. [Google Scholar] [CrossRef]
Figure 1. Framework of the proposed infrared small-target detection method. Targets in the input and output images are highlighted with red boxes.
Figure 1. Framework of the proposed infrared small-target detection method. Targets in the input and output images are highlighted with red boxes.
Remotesensing 15 05424 g001
Figure 2. IPI target images at different iterations. Strong edges are preserved when the target is detected. Targets are shown in red boxes and strong edges are denoted by green circles.
Figure 2. IPI target images at different iterations. Strong edges are preserved when the target is detected. Targets are shown in red boxes and strong edges are denoted by green circles.
Remotesensing 15 05424 g002
Figure 3. Trend of rank with an increasing number of iterations. Image 1 to image 4 represent the patch images corresponding to four different infrared images from the used data set.
Figure 3. Trend of rank with an increasing number of iterations. Image 1 to image 4 represent the patch images corresponding to four different infrared images from the used data set.
Remotesensing 15 05424 g003
Figure 4. Three steps of the proposed reconstruction method.
Figure 4. Three steps of the proposed reconstruction method.
Remotesensing 15 05424 g004
Figure 5. The running time ratio of each component in APSVD. The matrix is taken from SIR_1 in the experiment with a fixed width of 32.
Figure 5. The running time ratio of each component in APSVD. The matrix is taken from SIR_1 in the experiment with a fixed width of 32.
Remotesensing 15 05424 g005
Figure 6. Test images SIR_1 to SIR_14. The targets are highlighted with red boxes and the binary mask of the target is given in the lower left corner of each image.
Figure 6. Test images SIR_1 to SIR_14. The targets are highlighted with red boxes and the binary mask of the target is given in the lower left corner of each image.
Remotesensing 15 05424 g006
Figure 7. Partial detection results for different methods. The correctly detected targets are highlighted with red boxes and enlarged in the top left corner of each target image. The incorrect targets are highlighted with green boxes and circles.
Figure 7. Partial detection results for different methods. The correctly detected targets are highlighted with red boxes and enlarged in the top left corner of each target image. The incorrect targets are highlighted with green boxes and circles.
Remotesensing 15 05424 g007
Figure 8. Partial detection results for different methods. The correctly detected targets are highlighted with red boxes and enlarged in the top left corner of each target image. The incorrect targets are highlighted with green boxes and circles.
Figure 8. Partial detection results for different methods. The correctly detected targets are highlighted with red boxes and enlarged in the top left corner of each target image. The incorrect targets are highlighted with green boxes and circles.
Remotesensing 15 05424 g008
Figure 9. ROC curves for the twelve methods on test images SIR_1 to SIR_14.
Figure 9. ROC curves for the twelve methods on test images SIR_1 to SIR_14.
Remotesensing 15 05424 g009
Figure 10. ROC curves of our method under different α and β values.
Figure 10. ROC curves of our method under different α and β values.
Remotesensing 15 05424 g010
Figure 11. ROC curves for PFA, PSTNN, LogTFNN and the proposed method (Ours). The patch size and step are labeled in the figure; for example, (25,25) means that the patch size was set to 25 × 25 and the step was set to 25.
Figure 11. ROC curves for PFA, PSTNN, LogTFNN and the proposed method (Ours). The patch size and step are labeled in the figure; for example, (25,25) means that the patch size was set to 25 × 25 and the step was set to 25.
Remotesensing 15 05424 g011
Figure 12. Comparison of execution time between IPI and the proposed method (Ours) for images of different resolution and complexity.
Figure 12. Comparison of execution time between IPI and the proposed method (Ours) for images of different resolution and complexity.
Remotesensing 15 05424 g012
Figure 13. Comparison of execution time between IPI and the proposed method (Ours) for the three parts.
Figure 13. Comparison of execution time between IPI and the proposed method (Ours) for the three parts.
Remotesensing 15 05424 g013
Table 1. Detailed information of the test images. The target size is expressed as the number of pixels.
Table 1. Detailed information of the test images. The target size is expressed as the number of pixels.
DataImage SizeTarget SizeSCRBackground TypeTarget TypeDetection Challenges
Strong Edge Low Contrast Heavy Noise Cloud Clutter
SIR_1256 × 172116.52cloud + skyIrregular shape
SIR_2256 × 23938.63building + skyWeak
SIR_3300 × 209121.04sea + skyLow intensity
SIR_4280 × 22823.09cloud + skyWeak, hidden
SIR_5320 × 240711.11cloud + skyHidden
SIR_6359 × 24966.14building + skyIrregular shape
SIR_7640 × 512410.52cloud + skyWeak, hidden
SIR_8320 × 25655.36sea + skyWeak
SIR_9283 × 18281.59cloud + seaHidden
SIR_10379 × 246310.57building + skyLow intensity
SIR_11315 × 20659.61cloud + skyLow intensity
SIR_12305 × 214178.43tree + skyIrregular shape
SIR_13320 × 25544.12cloud + skyLow intensity
SIR_14377 × 26162.38cloud + skyLow intensity
Table 2. Parameter settings. All methods used their original settings.
Table 2. Parameter settings. All methods used their original settings.
MethodPatch SizeStepParameter
IPI [17] 50 × 50 10 L = 1 , λ = L / min ( m , n ) , ε = 10 7
RIPT [36] 50 × 50 10 L = 1 , λ = L / min ( n 1 , n 2 , n 3 ) , ε = 10 7
NIPPS [20] 50 × 50 10 L = 1 , λ = L / min ( m , n ) , ε = 10 7
NRAM [22] 50 × 50 10 L = 1 , λ = L / min ( m , n ) , ε = 10 7
NOLC [23] 50 × 50 10 L = 1 , λ = L / min ( s i z e ( D ) ) , p = 0.5 , ε = 10 7
PSTNN [38] 40 × 40 40 L = 0.7 , λ = L / min ( n 1 , n 2 ) n 3 , ε = 10 7
SRWS [34] 50 × 50 10 L = 1 , λ = L / min ( m , n ) , γ = 0.09 / m i n ( m , n ) , ε = 10 7
PFA [37] 25 × 25 25 κ = 30 , τ 0 = 1 e + 5 , ε = 10 5
LogTFNN [39] 40 × 40 40 L = 1 , λ = L / min ( n 1 , n 2 ) × n 3 , β = 0.01 , μ = 200
HLV [26] 50 × 50 10 L = 1 , λ = L / max ( m , n ) , α = 1.3 , β = 2.5 , C = 8
ANLPT [42] 50 × 50 10 λ = s i g m o i d ( E / n 3 ) / min ( n 1 , n 2 ) × n 3 , E = e n t r o p y ( T )
Ours 50 × 50 10 L = 1 , λ = L / max ( m , n ) , ε = 10 7
Table 3. Comparison of SCRG and BSF under various methods. The best performance is indicated in bold.
Table 3. Comparison of SCRG and BSF under various methods. The best performance is indicated in bold.
   MethodsIPI [17]RIPT [17]NIPPS [20]NRAM [22]NOLC [23]PSTNN [38]SRWS [34]PFA [37]LogTFNN [39]HLV [26]ANLPT [42]Ours
SIR_1 SCRG2.082.550.052.762.581.812.780.031.562.85NaN20.67
 BSF1.512.263.452.821.981.315.544.501.142.14Inf32.40
SIR_2 SCRG3.292.381.172.89NaN3.135.200.911.824.243.4023.50
 BSF1.050.590.260.75Inf0.802.480.400.481.080.837.20
SIR_3 SCRG137.56NaN102.47235.38NaN90.23NaN32.4011.80NaNNaN151.21
 BSF11.39Inf5.9917.02Inf18.42Inf12.101.28InfInf19.48
SIR_4 SCRG16.3615.369.46Inf39.94Inf60.86NaNNaN16.74NaNInf
 BSF3.553.472.04Inf8.80Inf13.90InfInf3.61InfInf
SIR_5 SCRG2.185.600.684.794.961.536.572.391.411.820.017.81
 BSF0.772.070.161.631.720.492.490.800.710.610.683.59
SIR_6 SCRG28.9917.087.77Inf26.84NaNInfNaNNaN2.56NaNInf
 BSF32.216.181.96Inf8.08InfInfInfInf0.90InfInf
SIR_7 SCRG275.57InfInfNaNNaN5.36NaN2.133.42351.29NaNInf
 BSF169.41InfInfInfInf3.30Inf1.692.47215.97InfInf
SIR_8 SCRG7.5332.227.8917.0441.076.16NaN2.403.488.754.7690.97
 BSF3.9825.673.289.7743.574.50Inf192.281.824.882.3969.74
SIR_9 SCRG24.3425.5111.86InfNaN14.85Inf5.4118.0823.11NaNInf
 BSF12.9224.339.04InfInf7.95Inf10.009.4412.42InfInf
SIR_10 SCRG1.94Inf0.38Inf3.37Inf4.31NaNNaN2.392.02Inf
 BSF1.04Inf0.16Inf1.85Inf2.47InfInf1.301.36Inf
SIR_11 SCRG2.57NaN0.87NaNNaNNaN10.58NaN0.06Inf1.73Inf
 BSF0.28Inf0.07InfInfInf1.46Inf0.05Inf0.18Inf
SIR_12 SCRG1.47Inf1.42InfNaN1.911.11Inf0.521.751.14Inf
 BSF0.73Inf0.62InfInf1.021.75Inf0.250.910.55Inf
SIR_13 SCRG1.58Inf0.30InfInf5.67InfNaNNaN31.94InfInf
 BSF0.53Inf0.08InfInf3.40InfInfInf6.23InfInf
SIR_14 SCRG4.287.691.878.26Inf3.25Inf1.580.527.255.73Inf
 BSF1.552.790.483.09Inf1.14Inf0.630.192.882.06Inf
Table 4. Comparison of the execution time (s) across various patch-based methods. All baselines were tested on the CPU, while our methods were tested on both the CPU and GPU. The best time is denoted in bold, while the second-best is underlined.
Table 4. Comparison of the execution time (s) across various patch-based methods. All baselines were tested on the CPU, while our methods were tested on both the CPU and GPU. The best time is denoted in bold, while the second-best is underlined.
Image idSIR_1SIR_2SIR_3SIR_4SIR_5SIR_6SIR_7SIR_8SIR_9SIR_10SIR_11SIR_12SIR_13SIR_14
IPI [17]3.285.237.636.4512.5212.9312.6711.284.1215.327.887.2914.8718.72
RIPT [36]1.172.762.022.824.702.888.014.350.961.851.021.402.122.14
NIPPS [20]1.883.343.603.565.516.827.116.712.849.183.953.997.519.96
NRAM [22]2.172.141.552.612.993.882.384.791.444.202.092.273.944.20
NOLC [23]0.720.861.111.151.241.673.621.640.943.171.551.281.332.11
SRWS [34]2.012.011.103.122.122.603.651.630.781.571.011.291.461.77
HLV [26]1.131.762.321.552.864.514.263.541.444.472.302.274.016.09
ANLPT [42]1.531.791.911.732.052.188.072.571.532.291.992.152.522.80
Ours (CPU)0.490.760.940.931.551.942.101.290.531.770.860.871.641.89
Ours (GPU)0.340.420.540.520.870.980.540.900.360.840.470.420.820.85
Table 5. Execution time (s) of PFA, PSTNN, LogTFNN and the proposed method (Ours) under varying patch size and number of steps. The best is denoted in bold, while the second-best is underlined.
Table 5. Execution time (s) of PFA, PSTNN, LogTFNN and the proposed method (Ours) under varying patch size and number of steps. The best is denoted in bold, while the second-best is underlined.
MethodSIR_1SIR_2SIR_3
(Patch, Step) (25,25) (40,40) (50,10) (25,25) (40,40) (50,10) (25,25) (40,40) (50,10)
PFA [37]9.960.331.3912.680.261.690.330.262.19
PSTNN [38]0.040.051.150.060.073.900.160.061.44
LogTFNN [39]0.891.3315.061.221.8111.631.271.3826.92
Ours(CPU)0.120.130.490.190.170.760.160.140.94
Ours(GPU)0.020.020.340.040.020.420.020.010.54
Table 6. Execution times (ms) of SVD functions in MATLAB and CUDA on different matrices, with a fixed matrix width of 32 and a rank of 10 for partial SVD. The fastest time on MATLAB and CUDA is marked in bold.
Table 6. Execution times (ms) of SVD functions in MATLAB and CUDA on different matrices, with a fixed matrix width of 32 and a rank of 10 for partial SVD. The fastest time on MATLAB and CUDA is marked in bold.
Matrix HeightMATLABCUDA
SVD SVDS Lanczos RSVD APSVD SGESVD SGESVDJ APSVD
10001.036.684.597.670.539.075.751.06
10,0006.2719.9322.0810.051.8316.717.361.24
100,000280.12406.70298.7750.8211.82/24.069.58
Table 7. Cumulative acceleration effects at different sizes of SIR_1, obtained using the resize function.
Table 7. Cumulative acceleration effects at different sizes of SIR_1, obtained using the resize function.
Image SizeBase+PASVD+New Continuation+GPU Parallelism
200 × 150 1.420.860.290.09
280 × 228 6.234.911.120.41
320 × 256 12.779.241.990.74
640 × 512 13.607.332.310.59
1020 × 750 57.7934.67.322.38
1260 × 1024 207.13116.5822.203.65
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Hao, X.; Liu, X.; Liu, Y.; Cui, Y.; Lei, T. Infrared Small-Target Detection Based on Background-Suppression Proximal Gradient and GPU Acceleration. Remote Sens. 2023, 15, 5424. https://doi.org/10.3390/rs15225424

AMA Style

Hao X, Liu X, Liu Y, Cui Y, Lei T. Infrared Small-Target Detection Based on Background-Suppression Proximal Gradient and GPU Acceleration. Remote Sensing. 2023; 15(22):5424. https://doi.org/10.3390/rs15225424

Chicago/Turabian Style

Hao, Xuying, Xianyuan Liu, Yujia Liu, Yi Cui, and Tao Lei. 2023. "Infrared Small-Target Detection Based on Background-Suppression Proximal Gradient and GPU Acceleration" Remote Sensing 15, no. 22: 5424. https://doi.org/10.3390/rs15225424

APA Style

Hao, X., Liu, X., Liu, Y., Cui, Y., & Lei, T. (2023). Infrared Small-Target Detection Based on Background-Suppression Proximal Gradient and GPU Acceleration. Remote Sensing, 15(22), 5424. https://doi.org/10.3390/rs15225424

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop