Next Article in Journal
Influence of Potentiostat Hardware on Electrochemical Measurements
Previous Article in Journal
Flight Attitude Estimation with Radar for Remote Sensing Applications
Previous Article in Special Issue
Robust Learning with Noisy Ship Trajectories by Adaptive Noise Rate Estimation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Infrared Bilateral Polarity Ship Detection in Complex Maritime Scenarios

1
School of Electronic and Optical Engineering, Nanjing University of Science and Technology, Nanjing 210094, China
2
Jiangsu Key Laboratory of Spectral Imaging & Intelligent Sense, Nanjing University of Science and Technology, Nanjing 210094, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(15), 4906; https://doi.org/10.3390/s24154906 (registering DOI)
Submission received: 24 June 2024 / Revised: 23 July 2024 / Accepted: 26 July 2024 / Published: 29 July 2024
(This article belongs to the Special Issue Advanced Sensing Technologies for Marine Intelligent Systems)

Abstract

:
In complex maritime scenarios where the grayscale polarity of ships is unknown, existing infrared ship detection methods may struggle to accurately detect ships among significant interference. To address this issue, this paper first proposes an infrared image smoothing method composed of Grayscale Morphological Reconstruction (GMR) and a Relative Total Variation (RTV). Additionally, a detection method considering the grayscale uniformity of ships and integrating shape and spatiotemporal features is established for detecting bright and dark ships in complex maritime scenarios. Initially, the input infrared images undergo opening (closing)-based GMR to preserve dark (bright) blobs with the opposite suppressed, followed by smoothing the image with the relative total variation model to reduce clutter and enhance the contrast of the ship. Subsequently, Maximally Stable Extremal Regions (MSER) are extracted from the smoothed image as candidate targets, and the results from the bright and dark channels are merged. Shape features are then utilized to eliminate clutter interference, yielding single-frame detection results. Finally, leveraging the stability of ships and the fluctuation of clutter, true targets are preserved through a multi-frame matching strategy. Experimental results demonstrate that the proposed method outperforms ITDBE, MRMF, and TFMSER in seven image sequences, achieving accurate and effective detection of both bright and dark polarity ship targets.

1. Introduction

In navigation, searching, and tracking tasks under marine environments, infrared target detection technology plays a crucial role due to its unique advantages, such as long detection range, and high concealment [1,2,3]. Infrared imaging systems can obtain distance and shape information by receiving thermal radiation [4], and then produce infrared images for different tasks through subsequent processing. However, constrained by detector performance, complex weather conditions, and the inherent fluctuations of the marine, infrared images typically have only a narrow grayscale range, with a low contrast and restricted signal-to-noise ratio at long imaging distances, which significantly increases the difficulty of target detection.
In infrared ship detection tasks, targets can be roughly divided into point targets or small targets (having area less than 9 × 9 pixels, isotropic), area targets (typically having certain shape and contour information, lacking texture, but having a relatively uniform grayscale distribution), and larger targets (larger area, with rich texture and contour information). In point target detection tasks, researchers have delved into the isotropic shape characteristics and strong contrast of point targets, developing a series of efficient and reliable methods such as curvature filtering [5,6] and Local Contrast Measure (LCM) [7,8]. For larger targets, which usually occupy a significant area in the image and have rich contour and texture details, they are easier to detect compared to point and area targets, but face the challenge of how to completely extract the entire ship.
Area target ships always appear as patches with uniform grayscale distribution and regular shapes in infrared images. If the grayscale of the ship is greater (smaller) than its local area, it is called a bright (dark) polarity target. The existing detection methods of the area target ship can be roughly divided into histogram-based methods, background modeling methods, feature-based methods, and deep learning-based methods. Histogram-based detection methods rely on the grayscale distribution of pixels in the whole image, dividing the image into foreground and background categories under certain criteria. For different scenarios, many researchers have developed various histogram transformation methods to adjust the grayscale distribution of images [4,9,10], enhance the contrast of targets, and then use methods such as the Otsu [11,12], the maximum entropy [13,14], and various improved forms for segmentation. In addition, clustering methods [15,16] have been introduced into infrared ship detection tasks to achieve complete extraction of larger targets. Mean shift was utilized to smooth infrared images, enhancing the contrast of ship targets [17,18]. However, these methods often have strict limitations on the grayscale distribution of the image and the target, making it difficult to determine a reasonable segmentation threshold in complex scenarios with unknown target grayscale polarity, interference similar to real targets, or irregular histogram distribution. Background modeling methods estimate the background in various ways to separate the target from the background. The Infrared Patch Image (IPI) model [19] is based on the assumption that the target is sparse with the background low-rank; thus, by dividing, re-organizing, decomposing, and reconstructing the image, small-sized targets in a stable background can be detected. Subsequently, researchers have improved the reconstruction and decomposition methods of sub-images to enhance the detection capability of the IPI model in complex scenarios [20,21]. Apart from this, researchers use Gaussian mixture models to model the background [22,23,24], achieving prediction and reproduction of dynamic backgrounds. These methods have certain robustness in scenarios with severe fluctuations but struggle to distinguish irregular fish scale reflections, sun-glint, and other interferences in complex scenarios.
In infrared images, ships often have one or more features that make them distinguishable from the background. By quantitatively analyzing these features and segmenting images according to certain criteria, feature analysis-based detection algorithms can be formed. Top-Hat filtering was first introduced into infrared ship detection tasks [25], using morphological filters to suppress clutter in the image while preserving bright blobs, achieving the detection of bright ships. On this basis, many researchers have improved morphological operations, such as using annular structural elements to enhance the detection capability for small-sized targets [26,27] or introducing multi-scale algorithms to achieve adaptive detection of targets of different sizes [28]. In addition, features such as contours [29,30] and gradients [31,32,33] have also been widely used in ship target detection, and many researchers consider combining multiple features [15,34] to enhance the robustness of algorithms in different scenarios. For dark ships, many researchers have conducted in-depth studies. Dong et al. [35] calculated saliency maps through an inverse Gaussian difference filter, making dark blobs outstanding in the saliency map, and then extracted potential ships in the image by segmenting the saliency map with an adaptive threshold. However, this method struggles to distinguish narrow dark bands on the sea surface or small-sized fish scale patterns with clear edges.
By using Grayscale Morphological Reconstruction operations to preserve and suppress bright or dark blobs in infrared images, Li et al. [36] achieved parallel detection of bright and dark ships. However, this method also faces difficulties in determining segmentation thresholds in complex scenarios with “significant” interference. To address the interference of island and reef backgrounds, Chen et al. [37] calculated the improved structural tensor of the multi-scale grayscale morphological reconstruction results of the original infrared image as a guide, merging the prominent regions in the Gaussian-filtered image to detect bright polarity ships of different sizes. However, the improved structural tensor proposed by this method struggles to distinguish fish scale pattern interference similar in size to ships. Ding et al. [4] proposed an improved histogram equalization combined with gradient information (MHEEF) to preprocess infrared images with backlit scenes, enhancing the contrast of dark ships, and then a dual-scale, dual-mode Local Contrast Measure (LCMDSM) was utilized to extract targets. The above methods can all be summarized as using single-frame image information to detect ships in single-frame images. Considering that ships, as man-made objects, have temporal and spatial stability in continuously captured image sequences, based on this feature, Wang et al. [31] proposed an improved wavelet transform to suppress the time-varying background clutter and simultaneously track stable ships by using pipeline filtering. Similarly, Li et al. [38] first extracted the Maximally Stable Extremal Regions (MSER) in the infrared image, then suppressed clutter through region matching between adjacent frames, and finally stable bright and dark polarity ships were detected at the same time. However, when the sea surface fluctuates violently and the size of the ship is small, it will be challenging to achieve stable matching between the MSER regions containing ship targets directly extracted from different frames, which may lead to frequent missed detections. Zhang et al. [39] developed a “detection-tracking-detection” method for detecting small-sized bright ships in infrared images, first extracting regions in single-frame images where targets may exist by using difference of Gaussian filtering and adaptive threshold segmentation, then suppressing interference through continuous frame matching. Apart from this, a re-detection method for potential missed targets was designed to further improve the robustness of this method.
As a trend in many research fields, deep learning-based methods are data-driven, that is, through data annotation, reward and punishment mechanisms, as well as iterative training, researchers enable neural networks to mine and refine deeper, more abstract features from a large amount of data and finally achieve efficient and accurate detection. Early on, such methods were mainly applied to infrared image target detection tasks with a space-based observer [40,41]. In 2018, Zhou et al. [42] proposed a one-stage network to learn features from multi-resolution infrared images, achieving reliable detection of ships in large infrared images. In 2022, Long et al. [43] introduced a visual attention mechanism into the YOLOv5 network architecture and introduced dilated convolution to enhance the receptive field, achieving the recognition of infrared ships against the background of a gentle sea surface with island reefs. By combining a manually designed feature extractor and deep learning methods, Yao et al. [44] designed a multi-dimensional information fusion network to accurately identify small-sized bright ships in infrared images. In 2023, Deng et al. [45] published an infrared ship rotating target detection algorithm, FMR-YOLO, in which a Weighted Feature Pyramid Network Based on Extended Convolution (DWFPN) was proposed with rotation detection technology introduced and achieved an average accuracy of up to 92.7%. Considering the complexity of the deep learning method and the difficulty of deployment on small devices, Gao et al. [46] proposed a lightweight model for detecting infrared ships by replacing the backbone of YOLOv5 with the Mobilev3, which greatly improved the computational efficiency and achieved the same detection performance as the YOLOv5m model while reducing the parameter size by 83%. In 2024, an improved detection model based on YOLOv5s to detect infrared ship targets in coastal areas with high ship density and significant target scale differences was proposed by Wang et al. [47], in which a feature fusion module was designed to enhance the feature fusion of the network, with SPD-Conv and Soft-NMS adopted to improve the detection accuracy of small targets in low-resolution images and deal with the missed detection in the case of dense occlusion. In addition to improving the design of the model, Wang et al. [48] introduced infrared multi-band fusion technology to improve detection accuracy with fewer parameters, achieving inference speeds close to 60 frames per second on embedded devices. Apart from this, many researchers [49,50,51] have been applying deep learning-based methods into ship detection tasks in infrared remote sensing images, continuously improving the model performance and detection effect. However, for deep learning-based methods, a large amount of training data are needed to ensure the reliability of the neural network. For example, in [43], researchers mentioned using 4079 out of 4533 infrared images for training and then testing the remaining images. On the one hand, publicly available infrared sea surface image datasets vary greatly and are limited in number. At the same time, manually annotating a large number of infrared images with poor contrast, missing target details, and a low signal-to-noise ratio requires a lot of manpower and resources, still posing challenges for deep learning-based methods [37].
In summary, we may summarize the current challenges in detecting infrared ships. First, most existing methods are designed for relatively simple scenarios with smooth seas, and usually a single grayscale polarity of the target is assumed, while in actual sea surface scenarios, the polarity of ship targets in infrared images is often unknown due to the variation in sea conditions, illumination, and detector positions, thus may result in missing detections. Second, in complex scenarios, there may exist interferences of different sizes, such as islands, artificial structures, bright and dark bands, fish scale patterns, and even clouds, which may be more prominent than real targets in the saliency map of various features. As a result, it may struggle to determine the segmentation parameters for methods such as adaptive threshold segmentation or Otsu to achieve balance between accuracy and completeness. Finally, in some scenarios, the temporal and spatial stability of the ships may not be fully utilized, and these features may provide some assistance for infrared ship detection tasks.
In this paper, we make the following assumptions regarding area target ships in maritime scenarios:
(1)
Ship Polarity: Ships can exhibit either bright polarity or dark polarity. Specifically, their grayscale values are either relatively high (bright) or low (dark) than the local background.
(2)
Uniform Grayscale Distribution: The grayscale distribution of ships is uniform across the infrared image sequences.
(3)
Temporal and Spatial Stability: Ships demonstrate temporal and spatial stability in infrared image sequences. In other words, their grayscale distribution and shape remain nearly constant over time.
Addressing the issues above, this paper first proposes an infrared image smoothing method that combines GMR and RTV to suppress noises and enhance the contrast of ships. Subsequently, the Maximally Stable Extremal Regions in the image are extracted as candidate targets. Finally, shape features and spatiotemporal characteristics are integrated to discriminate between ships and interferences, achieving the detection of bright and dark ships in complex scenarios.

2. Materials and Methods

The framework of the proposed method is illustrated in Figure 1, primarily consisting of image smoothing based on the GMR and RTV, candidate region extraction based on MSER and shape features, and multi-frame matching based on spatiotemporal characteristics. In this section, the principles of the GMR and RTV algorithms used for image smoothing and sea clutter suppression are introduced first. Subsequently, we extract MSER as the candidate targets from the smoothed images. Finally, spatiotemporal features are introduced to detect ships and suppress interferences.

2.1. Grayscale Morphological Reconstruction

Grayscale Morphological Reconstruction (GMR) is widely used as a powerful tool in image preprocessing and segmentation due to its excellent performance in image feature extraction and image restoration [36]. Taking the results of the opening (closing) operation of the original image as the constraint of iterative geodesic dilation (erosion), the GMR can be divided into opening-based (OGMR) and closing-based (CGMR), which can be used to extract the connected domain in the image with a uniform gray distribution that is darker (brighter) than the surrounding pixels, respectively. The definition of OGMR is as follows:
O G M R I I o p e n = k 1 g d k ( I o p e n ) ,
g d 1 ( I o p e n ) = I o p e n b , I ,
g d k ( I o p e n ) = g d 1 g d k 1 ( I o p e n ) ,
where Iopen is the opening result of the original image I and k and b represent the iterations and the structure element of the geodesic dilation gd, respectively. and represent the pixel-wise maximum operation and minimum operation, respectively. With I the mask image, geodesic dilation is iteratively executive until stability is reached according to Equation (3), in which the bright blobs have been suppressed while the dark blobs persevered with their contour unchanged.
Replace the marker image with Iclose and perform iterative geodesic erosion gek; similarly, the definition of CGMR is as follows:
C G M R I I c l o s e = k 1 g e k ( I c l o s e ) ,
g e 1 ( I c l o s e ) = I c l o s e b , I ,
g e k ( I c l o s e ) = g e 1 g e k 1 ( I c l o s e ) ,
In contrast, the CGMR operation suppresses dark blobs while preserving bright blobs with almost unchanged contours. In this paper, a disc-shaped structural element with a radius of 15 pixels is employed in the opening and closing operation to obtain the marker image, followed by iterative geodesic dilation/erosion using a 3 × 3 pixels square structural element, b. The results, as shown in Figure 2, demonstrate that OGMR (CGMR) helps preserve dark (bright) polarity targets while suppressing the opposite. Given that ship targets in infrared images often manifest as bright or dark blobs, and with the intention of achieving both bright and dark ships in single-frame detection, this paper applies OGMR and CGMR separately to the input images to extract potential bright and dark polarity targets concurrently. However, in complex maritime scenarios, under the combined effects of sea wind, illumination, and other factors, there is a possibility of forming interference regions similar to real ships on one hand. And on the other hand, the uniformity of the ship’s grayscale distribution may be compromised, affecting the regularity of the contours. In such cases, methods like the adaptive threshold segmentation mentioned in [36] or the improved structural tensor used in [37] may struggle to distinguish between real ships and a large number of interferences. To address this issue, this paper introduces Relative Total Variation into the ship target detection process, aiming to effectively suppress sea clutter in infrared images.

2.2. Relative Total Variation

Relative Total Variation (RTV) is an algorithm proposed by Li et al. [52], which can achieve effectively smoothing textures within the input image while preserving and extracting the structure. The fundamental idea of RTV is to distinguish between larger-scale structures and smaller-scale textures by using the ratio of the sum of the weighted absolute gradient values (referred to as windowed total variations) to the absolute value of the sum of the weighted gradients (referred to as windowed inherent variations) within a sliding window. This ratio, that is, the relative total variation, is then used as a penalty term in the objective function. The definition of RTV is as follows:
arg min S p S p I p 2 + λ D x ( p ) L x ( p ) + ε + D y ( p ) L y ( p ) + ε ,
D x ( p ) = q N ( p ) g p , q x S q , D y ( p ) = q N ( p ) g p , q y S q ,
L x ( p ) = q N ( p ) g p , q x S q , L y ( p ) = q N ( p ) g p , q y S q ,
where S and I are the input and output-smoothed image, respectively. q is the index of the pixels in the sliding window with p as the center. λ is the parameter that controls the degree of smoothing, and ε a constant to prevent the denominator from being 0. Dx(p), Dy(p), and Lx(p), Ly(p) represent the windowed total variations and the windowed inherent variations of pixel p along x and y directions. The sliding window gp,q is in fact a Gaussian filter, with its standard deviation σ corresponding to the maximum scale of the unfiltered texture.
By transforming the original objective function into linear equations, the output image S can be obtained by iteratively solving those equations. In this paper, we set the parameters λ, σ, and the number of iterations t to 0.015, 3, and 3, respectively, in most cases. Figure 3 shows the RTV results obtained directly from the original image and the image performed by OGMR. It can be observed that in the results of windowed total variation, smaller-scale textures such as fish-scale patterns are quite prominent, whereas in the results of windowed inherent variation, larger-scale structures with consistent gradient orientations within the sliding window, such as ships, edges of reefs, and boundaries of bright and dark bands, are more notable. In the results of the reciprocal relative total variation, by taking the ratio of the former two, the differences between small-scale clutter and large-scale structures, that is, the differences between textures and structures, are further amplified. The more prominent parts reflect large-scale structures such as reefs, dark bands, and ships. Based on the aforementioned analysis, using the relative total variation model for iterative processing of infrared images can effectively suppress small-scale clutter in the background while enhancing the grayscale uniformity across different regions of the image. After performing OGMR (CGMR), connected domains of a single polarity are preserved, while those of the opposite polarity are suppressed, and the magnitude and directional characteristics of image gradient changes are restrained to some extent. Subsequent smoothing with the RTV model, on this basis, can achieve the desired smoothing effect with fewer iterations. At the same time, small-scale connected domains of the same polarity preserved in the reconstructed image are suppressed after smoothing, reducing the potential interference. The smoothed infrared image and its grayscale distribution are shown in Figure 4. This paper processes the input infrared images with two types of GMR as the input for detecting bright and dark ships, followed by RTV smoothing as the input for candidate extraction.
Apart from this, the smoothed results of infrared images in diverse scenarios are displayed in Figure 5, and the numerical indicators of the corresponding image sequences are also displayed in Table 1. Here, the input images were set as references to evaluate the effect of the smoothing method in this paper, and we adopted the peak signal-to-noise ratio (PSNR) and the mean Structural SIMilarity (SSIM) index [53] as numerical indicators. The definitions of the PSNR and mean SSIM are as follows:
P S N R = 10 log 10 255 2 M S E ,
m S S I M = 1 M × N i = 1 M j = 1 N S S I M ( x i , y j ) ,
where MSE is the mean-square error between input images and smoothed images, and SSIM is the Structural SIMilarity index of image patches calculated by using a Gaussian window with a size of 11 × 11 pixels and a standard deviation of 1.5, in which M × N is the number of image patches. As shown in Figure 5 and Table 1, the proposed smoothing method can effectively suppress interferences like the fish scale pattern and improve the signal-to-noise ratio, while inevitably resulting in the loss of image structure information.

2.3. Maximum Stable Extreme Region

In the results of the abovementioned smoothed images, the contrast between the ship and its local background is further enhanced, resulting in a more uniform gray distribution of whole images. Given the grayscale uniformity inherent to the ships, the Maximally Stable Extreme Region (MSER) algorithm can be adept at extracting the isolating connected domains ranging between [81, 1500] pixels as potential targets. The principle of extracting the MSER is to binarize the image by increasing (decreasing) the segmentation threshold step by step, and find out regions with minimal area variation, that is, the so-called maximum stable extreme region. Regions with more uniform grayscale distribution are more likely to maintain stability in area size under different thresholds, indicating a lower area change rate. The area change rate is defined as follows:
p ( i ) = Q j + Δ Q j Δ Q j ,
in which Q1, Q2, …, Qj, … represent a series of nested regions obtained as the threshold increases from 0 to 255 (or decreases from 255 to 0) in steps of Δ; |∙| is the area of a region, that is, the number of pixels. If one particular region Q(j) has an area change rate ρ(j) less than a predefined threshold , it is considered maximally stable. In this paper, the area change rate threshold is consistently maintained at 0.25. Research [38] suggests that typically, a smaller step size Δ results in a greater number of MSERs extracted per frame, while a larger step size will only extract those with a higher uniform grayscale distribution. On the one hand, based on the characteristic of ships having a more uniform grayscale distribution compared to their background, the range of step size Δ can be appropriately amplified to reduce the number of clutter regions mistakenly extracted. On the other hand, in complex maritime scenarios, there may exist situations where ships have rather low contrast against their background, in which smaller ships can easily blend into the background. If a smaller step size Δ is chosen in such scenarios, large areas containing ships are more likely to be extracted rather than the ships themselves, leading to potential misjudgments in subsequent analysis.
Figure 6 illustrates the pixel grayscale distribution within the local background of the ship before and post-smoothing. The results indicate that after smoothing, the grayscale within the actual ship target zone is markedly uniform, exhibiting a significant contrast with the background. This contrast facilitates the selection of a larger step size Δ, which not only accurately extracts the ship target but also minimizes the extraction of extraneous clutter. However, it has to be mentioned that while the smoothing process effectively suppresses smaller clutters, it may inadvertently homogenize certain regions of the sea surface background, previously contained clutters, before smoothing, rendering them as potential candidate targets. In this paper, the outputs from the two aforementioned channels are independently processed and subsequently combined to form the final candidate targets. Taking into account the relative spatiotemporal stability of the ship against the dynamic sea surface background and the inherent variability of the background clutter, we employ both shape features and a pipe filtering approach to further exploit the difference between the ship and the clutter. Interference is screened out during single-frame detection with a manually designed shape feature range, while concurrently eliminate false positives through the aggregation of single-frame detection results and inter-frame matching, thereby ensuring the retention of genuine ships.

2.4. Shape Feature-Based Target Extraction

Following the smoothing and extraction processes above, candidate targets are separated from the local background and rendered into binary form to establish distinct candidate target regions. Because their shape and contour features are largely preserved, they provide a reliable foundation for verifying real ships and interferences. Shape features play a pivotal role in infrared ship detection tasks, offering an effective means to identify ships amidst sea clutter, islands, reefs, clouds, and other forms of interference. Typically, ship targets exhibit a narrow form, with the upper portion being smaller than the lower, coupled with a relatively uniform contour. Leveraging these shape features, metrics such as aspect ratio, rectangularity, compactness, and the ratio of the upper to lower area are employed as shape features to authenticate targets. The definitions of these features are as follows:
R a t i o w i d t h & h e i g h t = w i d t h h e i g h t ,
C o m p a c t n e s s = p e r i m e t e r 2 a r e a ,
R e c t a n g u l a r i t y = a r e a a r e a R e c ,
R a t i o u p & d o w n = a r e a u p a r e a d o w n ,
where width and height are the size of the minimum external rectangular box of the candidate target. perimeter and area are the perimeter and area of the candidate target, respectively. areaRec is the area of the minimum external rectangular box, and areaup and areadown are the areas of the upper and lower parts of the vertically equally divided bounding box.
The aspect ratio serves as a criterion to filter out interferences that manifest as excessively narrow strips. Compactness and rectangularity are utilized to eliminate clutter with relatively irregular contours. Furthermore, this study focuses on ships with an area ranging from 81 to 1500 pixels. Consequently, an area threshold is established to eliminate any interference falling outside this range. We selected 264 different ships from the VAIS dataset [54], 200 ships from the IRay dataset, and 100 infrared ship images from the self-collection dataset for analysis and summarization and obtained the statistical parameters of shape features as presented in Table 2. The IRay dataset is an open-source infrared maritime ship dataset provided by Yantai Raytron Technology Co., Ltd. (Yantai, China), which consists of thousands of infrared images containing various types of ships and can be accessed at http://openai.iraytek.com/apply/Sea_shipping.html/ (accessed on 9 June 2021). Additionally, the self-collection dataset in this paper comprises image sequences of infrared ships with different polarities captured from multiple coastal viewpoints.
Figure 7 illustrates the results of further screening candidate targets by employing the previously defined shape feature parameters. These shape features can be used to eliminate a portion of the interference; however, in complex environments, certain clutter with intricate shapes may still conform to the specified criteria. Therefore, the differences in spatiotemporal stability between targets and interference are introduced to make further distinctions.

2.5. Spatiotemporal Stability-Based Multi-Frame Matching Strategy

Apart from the utilization of shape features, many researchers have developed multi-frame detection methods for small targets, capitalizing on target stability, as referenced in [6,55]. The constancy of area targets within image sequences, particularly for larger ships, has been thoroughly investigated as well [31,39], culminating in the development of various algorithms. Pipeline filtering, a “tracking before detection” (TBD) strategy, is widely employed in infrared image target detection tasks. This technique begins with each prospective target in the reference frame, determining an optimal pipeline radius—corresponding to the detection scope in adjacent frames—based on the target’s size. Concurrently, an appropriate pipeline length threshold is established, representing the requisite number of detections and matching occurrences for a potential target within the image sequence. A potential target is deemed to be a real target only if its detection frequency surpasses the threshold; otherwise, it is classified as a false target.
Inspired by the idea of pipeline filtering, this paper designs a simple multi-frame detection approach to further utilize the spatiotemporal stability of ship targets. This method significantly diminishes false alarm interference by accumulating the results of single-frame detections across image sequences. The conceptual framework of this method is depicted in the subsequent Figure 8, utilizing the outcomes of shape feature analysis as inputs and employing a sequence of 5 adjacent frames for multi-frame detection and matching. The process commences with the designation of the initial frame from the five-frame sequence as the reference frame, labeled as i = 1. The coordinates, width, and height of all potential targets Ci,1, …, Ci,t1, …, Ci,n1 are recorded, with each target’s detection number recorded as one. The subsequent frame is then utilized as the matching image, denoted as j = 2, where all potential targets Cj,1, …, Cj,t2, …, Cj,n2 are identified and their respective parameters recorded. For every potential target in the reference frame, the distance Dis(Ci,t1, Cj,t2), area change rate Ratioarea(Ci,t1, Cj,t2), and the variation of aspect ratio Ratiowidth&height(Ci,t1, Cj,t2) are calculated in relation to all potential targets in the matching frame. A successful match updates only if the distance, area change rate, and variation of aspect ratio meet the corresponding thresholds; as a result, the position and size information are updated, and detection numbers of Ci,t1 are increased by one. The information about these confirmed targets is then fed into the matching process for the subsequent five-frame unit. As the matching progresses, the threshold for match count is elevated, ensuring the continuous identification and output of authentic targets. Dis(,), Ratioarea(,), and Ratiowidth&height(,) are defined as follows:
D i s ( C i , t 1 , C j , t 2 ) = x i , t 1 x j , t 2 2 + y i , t 1 y j , t 2 2 ,
R a t i o a r e a ( C i , t 1 , C j , t 2 ) = a i , t 1 a j , t 2 2 / a i , t 1 × a j , t 2 ,
R a t i o w i d t h & h e i g h t ( C i , t 1 , C j , t 2 ) = R i , t 1 R j , t 2 2 / R i , t 1 × R j , t 2 ,
where (x, y) is the upper-left coordinate of the minimum external bounding box and a and R are the area and aspect ratio of the bounding box, respectively. The threshold of distance Tdis is set to 20 pixels, and the area change threshold TRarea and the threshold of variation of aspect ratio TRwh are set to 0.5.

3. Results

In this section, a series of comparative experiments were performed on infrared image sequences across seven diverse scenarios to test the effectiveness of the proposed method in identifying both bright and dark ships. As for comparison, we chose the several algorithms that are commonly utilized in infrared ship detection tasks, to be specific, a method proficient in detecting dark targets within intricate scenarios [35], as well as two methods capable of simultaneously detecting both bright and dark targets [36,38]. The comparative experiments were executed using MATLAB2018a.

3.1. Test Dataset

The test dataset consists of seven sequences of infrared images, labeled Seq1–7. Each sequence features a resolution of 640 × 512 pixels and contains 1 or 2 bright/dark ships, as depicted in Figure 9a. The sea surface background in sequences Seq2, Seq5, and Seq6 is rather gentle, whereas Seq3 and Seq4 are full of fish scale patterns of varying sizes. There are large-scale dark bands in Seq1–Seq3, and all the sequences contain islands and reefs of diverse shapes. The Table 3 lists the specific details of the seven sequence sets. Notably, a median filter with a 3 × 3 template size was employed in images from Seq3 and Seq4 to achieve preliminary noise reduction. Moreover, taking into account the diversity of ship target size, shape, and background complexity, the Intersection over Union (IOU) threshold has been consistently established at 0.4 to ensure a standardized assessment across all scenarios. In addition to Seq4, the smoothing parameters λ, σ, and the number of iterations were set to 0.015, 3, and 3, respectively. To avoid incorrectly eliminating the narrow ship in Seq4, the σ was set to 1. Apart from this, the step size Δ range of MSER extraction was experimentally set to [3.5, 8].

3.2. Qualitative Comparison

In this section, ITDBE [35], which is adept at detecting dark targets within multifaceted scenes, alongside MRMF [36] and TFMSER [38], both capable of discerning light and dark polar targets, was chosen to test the infrared sequences. The ITDBE approach commences with Gaussian differential processing to enhance target contrast, followed by the identification of candidate targets using a multi-scale central difference method inspired by the human visual attention model. The process culminates with adaptive threshold segmentation, with extra grayscale inversion applied to Seq5–7 to align with the algorithmic prerequisite. As shown in Figure 9b, ITDBE demonstrates proficiency in detecting ships when there is obvious contrast with the background and the target’s area is small. However, numerous small but prominent interferences were also inadvertently detected, and they performed worse when dealing with larger targets. Moreover, in more complex scenes, adaptive threshold segmentation struggles to achieve an optimal balance between completely extracting the target and minimizing interference. The MRMF innovatively incorporates grayscale morphological reconstruction into the task of infrared ship detection, which generates a saliency map for potential bright and dark targets by executing dual reconstruction operations on the input image, followed by repeated adaptive threshold segmentation. The results are further refined by considering the shape features and the maximal eigenvalue of the structural tensor for final discernment. As observed in Figure 9c, MRMF generally succeeds in detecting ship targets with pronounced contrast. Nevertheless, it falters when the numerical “salience” of the target is comparable to or less than the interference, leading to missing. The challenge is exacerbated in sequences Seq6 and Seq7, where targets with intermediate grayscale values may be overlooked. The TFMSER leverages the grayscale uniformity and spatiotemporal stability of ships, which begins by extracting MSERs in adjacent frames, then employs spatiotemporal features to extract stable candidate regions while suppressing clutter, and ultimately utilizes shape features to identify the real ships. The results shown in Figure 9d indicate that this method typically incurs the fewest false alarms and exhibits robust detection capabilities for stable targets in simple scenarios, which are determined by the strict spatiotemporal limits. However, detection may fail when the ship is small or set against a complex background; even a relatively stable ship target may experience significant alterations in the extracted MSERs from frame to frame, leading to mismatches and missing. Figure 9e show the single-frame detection results of the proposed method, illustrating its capacity to detect ships across all sequences with the utmost precision.
Moreover, it should be noted that, due to the influence of factors such as the performance of the infrared detectors and the complexity of the maritime environment, it is difficult for the existing methods to reasonably infer the potential bright and dark polarity and shape features of the target in the absence of sufficient prior knowledge. In order to ensure the detection rate, current methods (including those discussed in this paper and the proposed method) that simultaneously process bright and dark targets through dual channels often yield a substantial number of false alarms in the channel opposite to the actual target’s polarity. To address this, we augment the single-frame detection result with a multi-frame matching technique to more effectively extract true targets. As illustrated in Figure 10, the integration of the multi-frame matching strategy with the single-frame detection method notably eliminates fluctuation interferences. When evaluated across various sequences, the proposed algorithm demonstrates superior performance in comparison to existing methods.

3.3. Quantitative Comparison

To facilitate a more objective assessment of the methods, four key performance metrics for quantitative comparison were employed: the detection rate Dp, false alarm rate FAR, misclassification error ME, and relative foreground area error RAE. Specifically, Dp and FAR reflect precision in ship detection and resilience to interference, respectively. ME quantifies the proportion of background pixels misclassified as foreground, while RAE represents the area-based proximity of the detection results to the real ships. According to the above analysis, the higher the Dp, the stronger the detection ability of the algorithm. Conversely, a lower FAR indicates better resistance to interference. Additionally, lower ME and RAE values denote a more precise and comprehensive segmentation effect of the ship. The four metrics are defined as follows:
M E = 1 B O B T + F O F T B O + F O ,
R A E = A O A T A T , A O > A T A T A O A T , o t h e r s ,
D p = T P G T ,
F A R = F P F P + T P ,
where BO and FO denote the number of pixels of the background and the number of pixels of the ship target in the ground truth images, respectively. BT and FT represent the number of pixels of the background and the ship target in the detection results, respectively. AT and AO are the areas of the ship determined by the ground truths and the detection results, respectively. TP quantifies the numbers where a ship target is accurately detected, while FP includes the occurrences where non-target elements are identified as ships, and GT indicates the total number of real ships that have been manually annotated.
The results of the metrics of the selected methods are shown in the following Table 4, Table 5, Table 6 and Table 7. The results reveal that with the IOU set to 0.4, the proposed method outperforms others in Dp, and the ME and RAE indicate that the proposed method can extract the ships accurately and completely across most scenarios while maintaining a relatively reasonable FAR. Additionally, although the TFMSER maintains the lowest FAR in all scenarios, its strict inter-frame matching mechanism also seriously curtails its detection capability in complex scenarios. In contrast, the ITDBE exhibits relatively robust detection ability but is prone to generating an excessive number of false alarms that far exceed the number of targets detected. In complex scenarios, adaptive threshold segmentation may struggle to distinguish between “significant” interference and real targets, resulting in the limited detection and anti-interference capabilities of MRMF in complex scenes such as Seq3–4. Figure 11 exhibits the ROC curves of the selected methods in seven sequences. It can be seen that the ITDBE method shows excellent detection performance in scenarios with simple backgrounds and small ship sizes, while in certain situations, such as Seq1 and Seq4, the entire ship may be detected partially, resulting in a fairly high Dp at low IOU levels, which demonstrates that this method lacks the ability to suppress minor background clutter. The MRMF falls short in detecting small ships and those with intermediate grayscale values. Owing to the strict spatiotemporal feature constraints, the TFMSER exhibits the strongest interference suppression ability, but in turn, this also constrains its detection performances at the edge of the image (such as Seq6 and Seq7) and small ships in complex scenarios. The proposed single-frame detection method showcases the best detection performance in all scenarios. Despite this, the processing method based on dual types of grayscale reconstruction and RTV model smoothing also contributes to the increase in FAR. Nevertheless, as previously discussed, a spatiotemporal feature-based multi-frame matching strategy can effectively mitigate fluctuating false alarms.

4. Conclusions

In this paper, we introduce a novel infrared image smoothing technique composed of GMR and RTV. Additionally, a detection method considering the grayscale uniformity of ships and integrating shape and spatiotemporal features is established for detecting bright and dark ships in complex maritime scenarios. Initially, the input infrared images undergo OGMR(CGMR) to preserve dark (bright) blobs with the opposite suppressed, followed by smoothing the image with the RTV to reduce clutter and enhance the contrast of the ship. Subsequently, Maximally Stable Extremal Regions (MSERs) are extracted from the smoothed image as candidate targets, and the results from the bright and dark channels are merged. Shape features are then utilized to eliminate clutter interference, yielding single-frame detection results. Finally, utilizing the spatiotemporal stability of ships and the fluctuation of clutter, true targets are identified through a multi-frame matching strategy. Experimental results demonstrate that the proposed method outperforms ITDBE, MRMF, and TFMSER in seven image sequences, achieving accurate and effective detection of bright and dark polarity ship targets. Our method avoids the use of adaptive threshold segmentation, which may struggle in complex maritime scenes. Instead, the RTV method is introduced into the preprocessing process of infrared images to enhance the suppression effect of fish scale plates, improve the detection effect of ships, and combine underutilized features such as gray uniformity and spatiotemporal stability, hoping to provide new ideas for infrared ship detection tasks.
Despite the excellent detection results of the proposed method, there are still some shortcomings. Primarily, the proposed method is aimed at ships with distinct polarities—bright and dark—and may exhibit unsatisfactory preformation when encountering ships with an uneven grayscale distribution, potentially leading to incomplete or missed detections. Future amendments could incorporate methods like watershed segmentation [56] and region growth [57] to refine the detection of unevenly distributed ships. Additionally, the proposed method avoids the use of adaptive threshold segmentation due to its limited adaptability in complex scenarios. A simple and effective evaluation mechanism for describing the scenarios and adjusting the parameters of the methods still remains necessary. For example, an evaluation method based on the statistical results of the image blocks in [18] could adaptively guide mean drift filters. Therefore, we will also focus on the evaluation mechanism of the scenario as a key direction for future studies. Lastly, with the rapid advancements in deep learning, there is an aspiration to integrate deep learning-based methods into infrared ship target detection tasks. This may involve using deep learning to refine and summarize the shape features of targets, potentially replacing traditional manual designed features to achieve more effective and robust ship detection. Furthermore, leveraging deep learning methods to explore image features at a deeper and more abstract level enables the extraction of richer semantic information, which can facilitate distinguishing different components within complex maritime scenarios (such as separating sea surfaces, islands, and the sky) and constructing image models and field data that can accurately describe such scenarios. These insights may provide novel ideas for achieving more effective and robust detection.

Author Contributions

Conceptualization, D.L.; methodology, L.T. and J.T.; software, L.T. and J.T.; validation, M.W. and L.T.; formal analysis, M.W. and Z.T.; investigation, L.T. and J.T.; resources, D.L.; data curation, M.W.; writing—original draft preparation, L.T. and J.T.; writing—review and editing, D.L. and L.T.; visualization, Z.T.; supervision, D.L.; project administration, G.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

The data that support the findings of this study are available from the corresponding author upon reasonable request.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Zhang, M.; Dong, L.; Zheng, H.; Xu, W. Infrared Maritime Small Target Detection Based on Edge and Local Intensity Features. Infrared Phys. Technol. 2021, 119, 103940. [Google Scholar] [CrossRef]
  2. Wang, B.; Benli, E.; Motai, Y.; Dong, L.; Xu, W. Robust Detection of Infrared Maritime Targets for Autonomous Navigation. IEEE Trans. Intell. Veh. 2020, 5, 635–648. [Google Scholar] [CrossRef]
  3. Prasad, D.; Rajan, D.; Rachmawati, L.; Rajabaly, E.; Quek, C. Video Processing From Electro-Optical Sensors for Object Detection and Tracking in a Maritime Environment: A Survey. IEEE Trans. Intell. Transp. Syst. 2016, 18, 1993–2016. [Google Scholar] [CrossRef]
  4. Ding, C.; Luo, Z.; Hou, Y.; Chen, S.; Zhang, W. An Effective Method of Infrared Maritime Target Enhancement and Detection with Multiple Maritime Scene. Remote Sens. 2023, 15, 3623. [Google Scholar] [CrossRef]
  5. Zhao, Y.; Pan, H.; Du, C.; Zheng, Y. Principal Curvature for Infrared Small Target Detection. Infrared Phys. Technol. 2015, 69, 36–43. [Google Scholar] [CrossRef]
  6. Wang, F.; Qian, W.; Ren, K.; Wan, M.; Gu, G.; Chen, Q. Maritime Small Target Detection Based on Appearance Stability and Depth-Normalized Motion Saliency in Infrared Video with Dense Sunglints. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5605919. [Google Scholar] [CrossRef]
  7. Chen, C.; Li, H.; Wei, Y.; Xia, T.; Tang, Y. A Local Contrast Method for Small Infrared Target Detection. Geoscience and Remote Sensing. IEEE Trans. Geosci. Remote Sens. 2014, 52, 574–581. [Google Scholar] [CrossRef]
  8. Han, J.; Moradi, S.; Faramarzi, I.; Liu, C.; Zhang, H.; Zhao, Q. A Local Contrast Method for Infrared Small-Target Detection Utilizing a Tri-Layer Window. IEEE Geosci. Remote Sens. Lett. 2019, 17, 1822–1826. [Google Scholar] [CrossRef]
  9. Wang, B.; Dong, L.; Zhao, M.; Xu, W. Fast Infrared Maritime Target Detection: Binarization via Histogram Curve Transformation. Infrared Phys. Technol. 2017, 83, 32–44. [Google Scholar] [CrossRef]
  10. Ding, C.; Pan, X.; Gao, X.; Ning, L.; Wu, Z. Three Adaptive Sub-Histograms Equalization Algorithm for Maritime Image Enhancement. IEEE Access 2020, 8, 147983–147994. [Google Scholar] [CrossRef]
  11. Zhang, T.-X.; Zhao, G.-Z.; Wang, F.; Zhu, G.-X. Fast Recursive Algorithm for Infrared Ship Image Segmentation. J. Infrared Millim. Waves 2006, 25, 295–300. [Google Scholar]
  12. Xu, H.-X.; Cao, W.-H.; Chen, W.; Guo, L.-Y. Segmentation of Unmanned Aerial Vehicle infrared ship target. Comput. Eng. Appl. 2009, 45, 224–225. [Google Scholar] [CrossRef]
  13. Du, F.; Shi, W.; Chen, L.; Yong, D.; Zhu, Z. Infrared Image Segmentation with 2-D Maximum Entropy Method Based on Particle Swarm Optimization (PSO). Pattern Recognit. Lett. 2005, 26, 597–603. [Google Scholar] [CrossRef]
  14. Li, Y.; Li, Z.; Ding, Z.; Qin, T.; Xiong, W. Automatic Infrared Ship Target Segmentation Based on Structure Tensor and Maximum Histogram Entropy. IEEE Access 2020, 8, 44798–44820. [Google Scholar] [CrossRef]
  15. Wang, T.; Bai, X.; Zhang, Y. Multiple Features Based Low-Contrast Infrared Ship Image Segmentation Using Fuzzy Inference System. In Proceedings of the 2014 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Wollongong, NSW, Australia, 25–27 November 2014; IEEE: New York, NY, USA, 2015. [Google Scholar]
  16. Yang, F.; Liu, Z.; Bai, X.; Zhang, Y. An Improved Intuitionistic Fuzzy C-Means for Ship Segmentation in Infrared Images. IEEE Trans. Fuzzy Syst. 2020, 30, 332–344. [Google Scholar] [CrossRef]
  17. Liu, J.; Sun, C.; Bai, X.; Zhou, F. Infrared Ship Target Image Smoothing Based on Adaptive Mean Shift. In Proceedings of the 2014 International Conference on Digital Image Computing: Techniques and Applications (DICTA), Wollongong, NSW, Australia, 25–27 November 2014. [Google Scholar]
  18. Liu, J.; Bai, X.; Sun, C.; Zhou, F.; Li, Y. Multi-Modal Ship Target Image Smoothing Based on Adaptive Mean Shift. IEEE Access 2018, 6, 12573–12586. [Google Scholar] [CrossRef]
  19. Gao, C.; Deyu, M.; Yang, Y.; Wang, Y.; Hauptman, A. Infrared Patch-Image Model for Small Target Detection in a Single Image. IEEE Trans. Image Process. 2013, 22, 4996–5009. [Google Scholar] [CrossRef]
  20. Zhang, L.; Peng, Z. Infrared Small Target Detection Based on Partial Sum of the Tensor Nuclear Norm. Remote Sens. 2019, 11, 382. [Google Scholar] [CrossRef]
  21. Zhou, X.; Li, P.; Zhang, Y.; Lu, X.; Hu, Y. Deep Low-Rank and Sparse Patch-Image Network for Infrared Dim and Small Target Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5003014. [Google Scholar] [CrossRef]
  22. Rostianingsih, S.; Adipranata, R.; Wibisono, F. Adaptive Background Dengan Metode Gaussian Mixture Models Untuk Real-Time Tracking. J. Inform. 2009, 9, 68–77. [Google Scholar] [CrossRef]
  23. Wang, H.; Zou, Z.; Shi, Z.; Li, B. Detecting Ship Targets in Spaceborne Infrared Image Based on Modeling Radiation Anomalies. Infrared Phys. Technol. 2017, 85, 141–146. [Google Scholar] [CrossRef]
  24. Zhou, A.; Xie, W.; Pei, J. Background Modeling Combined With Multiple Features in the Fourier Domain for Maritime Infrared Target Detection. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4202615. [Google Scholar] [CrossRef]
  25. Zeng, M.; Li, J.; Peng, Z. The Design of Top-Hat Morphological Filter and Application to Infrared Target Detection. Infrared Phys. Technol. 2006, 48, 67–76. [Google Scholar] [CrossRef]
  26. Deng, L.Z.; Zhang, J.; Xu, G.; Zhu, H. Infrared Small Target Detection via Adaptive M-Estimator Ring Top-Hat Transformation. Pattern Recognit. 2020, 112, 107729. [Google Scholar] [CrossRef]
  27. Li, Y.; Li, Z.; Li, J.; Yang, J.; Siddique, A. Robust Small Infrared Target Detection Using Weighted Adaptive Ring Top-Hat Transformation. Signal Process. 2023, 217, 109339. [Google Scholar] [CrossRef]
  28. Tian, W.; Li, S. Infrared Dim Target Detection Algorithm Based on Frequency-Space Domain Mapping and Multi-Scale Top-Hat Transform. Guangxue Jishu/Opt. Tech. 2018, 44, 325–332. [Google Scholar]
  29. Fang, L.; Wang, X.; Wan, Y. Adaptable Active Contour Model with Applications to Infrared Ship Target Segmentation. J. Electron. Imaging 2016, 25, 041010. [Google Scholar] [CrossRef]
  30. Fang, L.; Zhao, W.; Li, X.; Wang, X. A Convex Active Contour Model Driven by Local Entropy Energy with Applications to Infrared Ship Target Segmentation. Opt. Laser Technol. 2017, 96, 166–175. [Google Scholar] [CrossRef]
  31. Wang, B.; Dong, L.; Zhao, M.; Wu, H.; Xu, W. Texture Orientation-Based Algorithm for Detecting Infrared Maritime Targets. Appl. Opt. 2015, 54, 4689. [Google Scholar] [CrossRef]
  32. Yang, P.; Dong, L.; Xu, W. Small Maritime Target Detection Using Gradient Vector Field Characterization of Infrared Image. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2023, 16, 1827–1841. [Google Scholar] [CrossRef]
  33. Yang, P.; Dong, L.; Xu, W. Detecting Small Infrared Maritime Targets Overwhelmed in Heavy Waves by Weighted Multidirectional Gradient Measure. IEEE Geosci. Remote Sens. Lett. 2021, 19, 7002005. [Google Scholar] [CrossRef]
  34. Liu, J.; Bai, X.; Sun, C.; Zhou, F.; Li, Y. Infrared Ship Target Segmentation through Integration of Multiple Feature Maps. Image Vis. Comput. 2016, 48–49, 14–25. [Google Scholar] [CrossRef]
  35. Dong, L.; Ma, D.; Qin, G.; Zhang, T.; Xu, W. Infrared Target Detection in Backlighting Maritime Environment Based on Visual Attention Model. Infrared Phys. Technol. 2019, 99, 93–200. [Google Scholar] [CrossRef]
  36. Li, Y.; Li, Z.; Zhu, Y.; Li, B.; Xiong, W.; Huang, Y. Thermal Infrared Small Ship Detection in Sea Clutter Based on Morphological Reconstruction and Multi-Feature Analysis. Appl. Sci. 2019, 9, 3786. [Google Scholar] [CrossRef]
  37. Chen, X.; Qiu, C.; Zhang, Z. A Multiscale Method for Infrared Ship Detection Based on Morphological Reconstruction and Two-Branch Compensation Strategy. Sensors 2023, 23, 7309. [Google Scholar] [CrossRef] [PubMed]
  38. Li, L.; Liu, G.; Li, Z.; Ding, Z.; Qin, T. Infrared Ship Detection Based on Time Fluctuation Feature and Space Structure Feature in Sun-Glint Scene. Infrared Phys. Technol. 2021, 115, 103693. [Google Scholar] [CrossRef]
  39. Zhang, M.; Dong, L.; Tian, C. DTD: “Detect–Track–Detect” Method for Infrared Marine Object Detection. IEEE Geosci. Remote Sens. Lett. 2023, 20, 7000905. [Google Scholar] [CrossRef]
  40. Jiang, B.; Ma, X.; Lu, Y.; Li, Y.; Feng, L.; Shi, Z. Ship Detection in Spaceborne Infrared Images Based on Convolutional Neural Networks and Synthetic Targets. Infrared Phys. Technol. 2018, 97, 229–234. [Google Scholar] [CrossRef]
  41. Wang, N.; Li, B.; Wei, X.; Wang, Y.; Yan, H. Ship Detection in Spaceborne Infrared Image Based on Lightweight CNN and Multisource Feature Cascade Decision. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4324–4339. [Google Scholar] [CrossRef]
  42. Zhou, M.; Jing, M.; Liu, D.; Xia, Z.; Zou, Z.; Shi, Z. Multi-Resolution Networks for Ship Detection in Infrared Remote Sensing Images. Infrared Phys. Technol. 2018, 92, 183–189. [Google Scholar] [CrossRef]
  43. Long, Y.; Jin, D.; Wu, Z.; Zuo, Z.; Wang, Y.; Kang, Z. Accurate Identification of Infrared Ship in Island-Shore Background Based on Visual Attention. In Proceedings of the 2022 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2022; p. 806. [Google Scholar]
  44. Yao, J.; Xiao, S.; Deng, Q.; Wen, G.; Tao, H.; Du, J. An Infrared Maritime Small Target Detection Algorithm Based on Semantic, Detail, and Edge Multidimensional Information Fusion. Remote Sens. 2023, 15, 4909. [Google Scholar] [CrossRef]
  45. Deng, H.; Zhang, Y. FMR-YOLO: Infrared Ship Rotating Target Detection Based on Synthetic Fog and Multi-Scale Weighted Feature Fusion. IEEE Trans. Instrum. Meas. 2023, 73, 5001717. [Google Scholar] [CrossRef]
  46. Gao, Z.; Zhang, Y.; Wang, S. Lightweight Small Ship Detection Algorithm Combined with Infrared Characteristic Analysis for Autonomous Navigation. J. Mar. Sci. Eng. 2023, 11, 1114. [Google Scholar] [CrossRef]
  47. Wang, Y.; Wang, B.; Huo, L.; Fan, Y. GT-YOLO: Nearshore Infrared Ship Detection Based on Infrared Images. J. Mar. Sci. Eng. 2024, 12, 213. [Google Scholar] [CrossRef]
  48. Wang, L.; Dong, Y.; Fei, C.; Liu, J.; Fan, S.; Liu, Y.; Yongfu, L.; Liu, Z.; Zhao, X. A Lightweight CNN for Multi-Source Infrared Ship Detection from Unmanned Marine Vehicles. Heliyon 2024, 10, e26229. [Google Scholar] [CrossRef] [PubMed]
  49. Wu, T.; Li, B.; Luo, Y.; Wang, Y.; Xiao, C.; Liu, T.; Yang, J.; An, W.; Guo, Y. MTU-Net: Multi-Level TransUNet for Space-Based Infrared Tiny Ship Detection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5601015. [Google Scholar]
  50. Guo, F.; Ma, H.; Li, L.; Lv, M.; Jia, Z. Multi-Attention Pyramid Context Network for Infrared Small Ship Detection. J. Mar. Sci. Eng. 2024, 12, 345. [Google Scholar] [CrossRef]
  51. Guo, F.; Ma, H.; Li, L.; Lv, M.; Jia, Z. FCNet: Flexible Convolution Network for Infrared Small Ship Detection. Remote Sens. 2024, 16, 2218. [Google Scholar] [CrossRef]
  52. Xu, L.; Yan, Q.; Xia, Y.; Jia, J. Structure Extraction from Texture via Relative Total Variation. ACM Trans. Graph. (TOG) 2012, 31, 139. [Google Scholar] [CrossRef]
  53. Wang, Z.; Simoncelli, E.P. Image quality assessment: From error visibility to structural similarity. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
  54. Zhang, M.; Choi, J.; Daniilidis, K.; Wolf, M.; Kanan, C. VAIS: A Dataset for Recognizing Maritime Imagery in the Visible and Infrared Spectrums. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, Boston, MA, USA, 7–12 June 2015; p. 16. [Google Scholar]
  55. Wang, B.; Dong, L.; Zhao, M.; Xu, W. A Small Dim Infrared Maritime Target Detection Algorithm Based on Local Peak Detection and Pipeline-Filtering. In Proceedings of the Seventh International Conference on Graphic and Image Processing (ICGIP 2015), Singapore, 23–25 October 2015. [Google Scholar]
  56. Vincent, L.; Soille, P. Watersheds in Digital Spaces: An Efficient Algorithm Based on Immersion Simulations. IEEE Trans. Pattern Anal. Mach. Intell. 1991, 13, 583–598. [Google Scholar] [CrossRef]
  57. Ning, J.; Zhang, L.; Zhang, D.; Wu, C. Interactive Image Segmentation by Maximal Similarity Based Region Merging. Pattern Recognit. 2010, 43, 445–456. [Google Scholar] [CrossRef]
Figure 1. The framework of the proposed method.
Figure 1. The framework of the proposed method.
Sensors 24 04906 g001
Figure 2. The input images and the results of OGMR(CGMR). (a1a3) and (b1b3) are the input dark target image, the result of OGMR processing, CGMR processing, and the corresponding three-dimensional view of gray distribution, respectively. (c1c3) and (d1d3) are the input bright target image, CGMR processed images, OGMR processed images, and the corresponding 3D views of gray distribution, respectively.
Figure 2. The input images and the results of OGMR(CGMR). (a1a3) and (b1b3) are the input dark target image, the result of OGMR processing, CGMR processing, and the corresponding three-dimensional view of gray distribution, respectively. (c1c3) and (d1d3) are the input bright target image, CGMR processed images, OGMR processed images, and the corresponding 3D views of gray distribution, respectively.
Sensors 24 04906 g002aSensors 24 04906 g002b
Figure 3. Taking the dark target image as an example, (a1a3) and (b1b3) are the results of windowed total variation, windowed inherent variation, and inverse relative total variation of the original infrared image (Figure 2(a1)) and the result of OGMR processing, respectively.
Figure 3. Taking the dark target image as an example, (a1a3) and (b1b3) are the results of windowed total variation, windowed inherent variation, and inverse relative total variation of the original infrared image (Figure 2(a1)) and the result of OGMR processing, respectively.
Sensors 24 04906 g003
Figure 4. The smoothed results of the image with a dark ship (Figure 2(a2)) and the image with a bright ship (Figure 2(c2)) with RTV. (a1,a2) are the grayscale distribution and the three-dimensional view of the dark ship image, respectively. (b1,b2) are the grayscale distribution and the three-dimensional view of the bright ship image, respectively.
Figure 4. The smoothed results of the image with a dark ship (Figure 2(a2)) and the image with a bright ship (Figure 2(c2)) with RTV. (a1,a2) are the grayscale distribution and the three-dimensional view of the dark ship image, respectively. (b1,b2) are the grayscale distribution and the three-dimensional view of the bright ship image, respectively.
Sensors 24 04906 g004
Figure 5. The smoothed results of images in different scenarios. (a) consists of the original input infrared images. (b,c) are the smoothed results of OGMR and CGMR, respectively.
Figure 5. The smoothed results of images in different scenarios. (a) consists of the original input infrared images. (b,c) are the smoothed results of OGMR and CGMR, respectively.
Sensors 24 04906 g005
Figure 6. Grayscale distribution of dark (light) ships and surroundings before and after smoothing. (a,c) are the grayscale distribution and corresponding 3D views of the dark ship and local background in the original infrared image and smoothed result by OGMR and RTV, respectively. (b,d) are the grayscale distribution and corresponding 3D views of the bright ship and local background in the original infrared image and smoothed result by CGMR and RTV, respectively.
Figure 6. Grayscale distribution of dark (light) ships and surroundings before and after smoothing. (a,c) are the grayscale distribution and corresponding 3D views of the dark ship and local background in the original infrared image and smoothed result by OGMR and RTV, respectively. (b,d) are the grayscale distribution and corresponding 3D views of the bright ship and local background in the original infrared image and smoothed result by CGMR and RTV, respectively.
Sensors 24 04906 g006
Figure 7. The results of the MSER extracted from the infrared image after smoothing and the results of the screening by shape features, with the step size Δ = 3.5 in the dark ship image and the step Δ = 8 in the bright ship image. (a1,a2) and (b1,b2) are the results of the dark ship image processed and smoothed by OGMR (CGMR) and the extracted candidate targets, respectively. (c1,c2) and (d1,d2) are the candidate targets extracted after CGMR (OGMR) processing and smoothing, respectively. (e,f) are the results of the merging of two channels of dark ship image and bright ship image, respectively.
Figure 7. The results of the MSER extracted from the infrared image after smoothing and the results of the screening by shape features, with the step size Δ = 3.5 in the dark ship image and the step Δ = 8 in the bright ship image. (a1,a2) and (b1,b2) are the results of the dark ship image processed and smoothed by OGMR (CGMR) and the extracted candidate targets, respectively. (c1,c2) and (d1,d2) are the candidate targets extracted after CGMR (OGMR) processing and smoothing, respectively. (e,f) are the results of the merging of two channels of dark ship image and bright ship image, respectively.
Sensors 24 04906 g007
Figure 8. The framework of multi-frame matching.
Figure 8. The framework of multi-frame matching.
Sensors 24 04906 g008
Figure 9. Detection results of different methods. (a) The original images of Seq1–Seq9. (b) The detection results of the ITDBE. (c) The detection results of the MRMF. (d) The detection results of the TFMSER. (e) The single-frame detection results of the proposed method. (f) The multi-frame matching results of the proposed method. The red rectangles represent the positions of targets, and the yellow rectangles represent the false alarm.
Figure 9. Detection results of different methods. (a) The original images of Seq1–Seq9. (b) The detection results of the ITDBE. (c) The detection results of the MRMF. (d) The detection results of the TFMSER. (e) The single-frame detection results of the proposed method. (f) The multi-frame matching results of the proposed method. The red rectangles represent the positions of targets, and the yellow rectangles represent the false alarm.
Sensors 24 04906 g009
Figure 10. Multi-frame matching result. (a) The first frame of each set of sequences. (a1a12) are the results of multi-frame matching of the 25th, 50th, …, 25 × ith, …, and 300th (i = 1, 2, ... , 12) frames of the seven sequences, respectively. (a13) The last frame of each set of sequences.
Figure 10. Multi-frame matching result. (a) The first frame of each set of sequences. (a1a12) are the results of multi-frame matching of the 25th, 50th, …, 25 × ith, …, and 300th (i = 1, 2, ... , 12) frames of the seven sequences, respectively. (a13) The last frame of each set of sequences.
Sensors 24 04906 g010aSensors 24 04906 g010b
Figure 11. The ROC curves of 7 sequences, with IOUs ranging from 0.1 to 1. (a1g1) and (a2g2) are the curves of Dp and FAR for selected methods in 7 sequences, respectively. The green triangle, blue triangle, yellow “x”, and red dots in the figures denote the results of ITDBE, MRMF, TFMSER, and the proposed single-frame detection method, respectively.
Figure 11. The ROC curves of 7 sequences, with IOUs ranging from 0.1 to 1. (a1g1) and (a2g2) are the curves of Dp and FAR for selected methods in 7 sequences, respectively. The green triangle, blue triangle, yellow “x”, and red dots in the figures denote the results of ITDBE, MRMF, TFMSER, and the proposed single-frame detection method, respectively.
Sensors 24 04906 g011aSensors 24 04906 g011bSensors 24 04906 g011c
Table 1. Quantitative evaluation of the proposed smoothing method.
Table 1. Quantitative evaluation of the proposed smoothing method.
SequencesPSNR (dB) mSSIM
Seq120.71340.7257
Seq222.29950.7258
Seq318.91570.3616
Seq420.02360.6681
Seq522.57590.6064
Seq620.41980.5993
Seq724.06670.4799
Table 2. Ranges of shape features.
Table 2. Ranges of shape features.
FeatureMinimumMaximum
Ratioheight&width1.611
Compactness12.9445317.1574
Rectangularity0.30080.9788
Ratioup&down<1
Area811500
Table 3. Details of the test sequences.
Table 3. Details of the test sequences.
SequencesFrames TargetBackground
Seq13001 dark ship moving to the leftComplex, with a large-scale dark band
Seq23001 dark ship moving to the leftRather gentle, with a large-scale dark band
Seq33001 stationary dark shipRather complex, with a dark band and islands
Seq43001 dark ship moving to the rightComplex, with a dark band and small islands
Seq53001 stationary bright shipGentle, with artificial buildings and a mountain
Seq63002 stationary bright shipsGentle, with artificial buildings and a mountain
Seq73002 bright targets: one stationary and one moving to the leftRather complex, with artificial buildings, an island, and reefs
Table 4. Average Dp of the selected methods on the test dataset. The bold emphasis represents the best results.
Table 4. Average Dp of the selected methods on the test dataset. The bold emphasis represents the best results.
SequenceITDBEMRMFTFMSER Proposed
Seq10.95000.98330.46430.9833
Seq21.00000.68331.00001.0000
Seq30.43330.81670.48211.0000
Seq40.91670.15000.16070.9500
Seq51.00001.00000.78571.0000
Seq60.96720.85240.65181.0000
Seq70.60000.49160.75001.0000
Table 5. Average FAR of the selected methods on the test dataset. The bold emphasis represents the best results.
Table 5. Average FAR of the selected methods on the test dataset. The bold emphasis represents the best results.
SequenceITDBEMRMFTFMSER Proposed
Seq10.89330.00000.00000.0000
Seq20.25000.22640.00000.0000
Seq30.99480.78320.03570.6296
Seq40.91710.55000.00000.7635
Seq50.95180.47410.27870.0000
Seq60.92100.79690.00000.5000
Seq70.82780.67040.20750.3064
Table 6. Average ME of the selected methods on the test dataset. The bold emphasis represents the best results.
Table 6. Average ME of the selected methods on the test dataset. The bold emphasis represents the best results.
SequenceITDBEMRMFTFMSER Proposed
Seq10.16870.12400.61140.1812
Seq20.39620.60870.08550.0910
Seq30.11990.34050.59440.1575
Seq40.12260.89940.84260.2176
Seq50.12770.05400.24460.0402
Seq60.48690.36840.49840.2466
Seq70.47300.59120.32520.1276
Table 7. Average RAE of the selected methods on the test dataset. The bold emphasis represents the best results.
Table 7. Average RAE of the selected methods on the test dataset. The bold emphasis represents the best results.
SequenceITDBEMRMFTFMSER Proposed
Seq10.93350.15750.61140.1487
Seq20.24360.66900.09780.0977
Seq30.99390.37390.55600.4281
Seq40.93420.78630.89400.6206
Seq50.94960.24540.37260.2011
Seq60.84020.49640.35020.3056
Seq70.93420.78630.92070.5939
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lu, D.; Teng, L.; Tan, J.; Wang, M.; Tian, Z.; Wang, G. Infrared Bilateral Polarity Ship Detection in Complex Maritime Scenarios. Sensors 2024, 24, 4906. https://doi.org/10.3390/s24154906

AMA Style

Lu D, Teng L, Tan J, Wang M, Tian Z, Wang G. Infrared Bilateral Polarity Ship Detection in Complex Maritime Scenarios. Sensors. 2024; 24(15):4906. https://doi.org/10.3390/s24154906

Chicago/Turabian Style

Lu, Dongming, Longyin Teng, Jiangyun Tan, Mengke Wang, Zechen Tian, and Guihua Wang. 2024. "Infrared Bilateral Polarity Ship Detection in Complex Maritime Scenarios" Sensors 24, no. 15: 4906. https://doi.org/10.3390/s24154906

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Article metric data becomes available approximately 24 hours after publication online.
Back to TopTop