You are currently viewing a new version of our website. To view the old version click .
Remote Sensing
  • Article
  • Open Access

29 September 2022

FLNet: A Near-shore Ship Detection Method Based on Image Enhancement Technology

,
,
and
1
Logistics Engineering College, Shanghai Maritime University, Shanghai 201306, China
2
Naval Academy, Brest Naval, Lanveoc-Poulmic, BP 600, 29240 Brest, France
3
School of Medical Information Engineering, Guangzhou University of Chinese Medicine, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.

Abstract

In the past few years, Synthetic Aperture Radar (SAR) has been widely used to detect marine ships due to its ability to work in various weather conditions. However, due to the imaging mechanism of SAR, there is a lot of background information and noise information similar to ships in the images, which seriously affects the performance of ship detection models. To solve the above problems, this paper proposes a new ship detection model called Feature enhancement and Land burial Net (FLNet), which blends traditional image processing methods with object detection approaches based on deep learning. We first design a SAR image threshold segmentation method, Salient Otsu (S-Otsu), according to the difference between the object and the noise background. To better eliminate noise in SAR images, we further combine image processing methods such as Lee filtering. These constitute a Feature Enhancement Module (FEM) that mitigates the impact of noise data on the overall performance of a ship detection model. To alleviate the influence of land information on ship detection, we design a Land Burial Module (LBM) according to the morphological differences between ships and land areas. Finally, these two modules are added to You Only Look Once V5 (YOLO V5) to form our FLNet. Experimental results on the SAR Ship Detection Dataset (SSDD) dataset show that FLNet comparison with YOLO V5 accuracy when performing object detection is improved by 7% and recall rate by 6.5%.

1. Introduction

As an active microwave imaging system for earth observation, Synthetic Aperture Radar (SAR) can work in any weather and at any time and is widely used in the field of ship detection [1,2,3]. In the early stages of ship detection, the primary goal is to better manually select some features to detect ships [4,5,6,7]. For instance, when detecting ships at the superpixel level, gray value differences between a ship’s area and neighboring regions is mainly used. Liu et al. used superpixels to segment object areas before applying a Constant False Positive Rate (CFAR) [8]. Wang et al. combined a superpixel segmentation with local contrast measurement [9]. Deng et al. first used the Simple Linear Iterative Clustering (SLIC) algorithm to generate superpixel regions and then used the corners for ship detection [10]. Wang et al. detect ship objects based on intensity differences between object and clutter pixels [11]. Liu et al. proposed the CFAR. The main principle of the CFAR is to apply a sliding window to the SAR image and then judge whether the window area is a ship according to the signal threshold [12]. A few extensions of the CFAR were progressively suggested including Smallest of CFAR, Order Statistic CFAR, Cell Averaging CFAR, and Greatest of CFAR [13,14,15,16]. Yang et al. developed ship detection based on the initial CFAR detection algorithm and improved it with a geometric optimization framework [17]. Wang et al. proposed a superpixel detector that improves the performance of ship target detection in SAR images via the local contrast of fisher vectors [18].
In the past few years, due to the rapid development of deep learning [19], a series of object detection methods based on deep learning have been applied to SAR images. These detection methods can be categorized into one-stage and two-stage [20]. Two-stage methods first extract a region of interest, then identify candidate objects, so the object detection accuracy is improved [21,22,23,24,25]. Unlike the two-stage algorithm, the one-stage algorithm performs in one step and is particularly appropriate for real-time detections. As one of the most advanced one-stage object detection algorithms, You Only Look Once V5 (YOLO V5) has the characteristics of fast detection speed and high accuracy, so it is selected as the baseline for our algorithm.
Some researchers have optimized the object detection algorithm based on deep learning according to the characteristics of SAR images. Li et al. proposed to use the K-means method to obtain the target scale distribution, optimize the selection of anchor boxes, and reduce the difficulty of network learning [26]. To reduce memory consumption and the computational complexity of convolutional neural networks, Xiong et al. proposed a lightweight detection model that utilizes channel shearing and knowledge distillation techniques, thereby reducing the YOLO V4 parameters [27]. Zhou et al. extended YOLO V5 by performing a K-means dimension clustering on the object frame and a mosaic enhancement image scale transformation and finally optimizing the loss function [28]. Lin et al. introduced a squeeze and excitation faster R-CNN that provides better detection results [29]. A Feature Pyramid Network (FPN) improved the performance of the detection network by utilizing the information between different feature layers [30]. Wei et al. proposed a high-resolution FPN for multiscale detection [31]. Cui et al. extended the convolutional network with an attention module that can learn weight coefficients between different channels [32]. Zhang et al. proposed using Neural Architecture Search (NAS) to automatically design a convolutional neural network structure on the dataset, which further improved the accuracy of ship detection [33]. Ke et al. fused the network feature layers to optimize the detection process [34].
Overall, these detection networks have achieved considerable performance when applied to SAR images, but so far, these SAR ship detection algorithms based on deep learning have not fully considered the differences between SAR ships and the surrounding environment. For instance, differences in the backscattering properties between ships and sea clutter result in the gray value of the ship is larger than that of the noise region. In addition, there are distinct geometrical features between the ship and the coast. Based on these differences, we introduced a combination of two modules: Feature Enhancement Module (FEM) and Land Burial Module (LBM). FEM is designed to enhance the characteristics of ships by eliminating noisy areas, while LBM eliminates coastal areas. The main contributions of this paper are as follows:
  • We introduce a Salient Otsu (S-Otsu) threshold segmentation method to deal with SAR images noise;
  • The FEM highlights ship features by suppressing background and noise information;
  • The LBM is added to YOLO V5, removing coastal areas, improving convolutional neural network training efficiency, and reducing the impact of coastal features on detection performance.

3. Materials and Methods

This section gives a detailed description of the two main modules and discusses their main principles. The overall framework of our proposed Feature enhancement and Land burial Net (FLNet) is shown in Figure 1. Firstly, the image noise is reduced by FEM and then the threshold is set according to the difference in the sum of gray values in the SAR images with and without land areas. The FEM processed image is delivered to the LBM if the total grayscale values of the image are greater than the threshold. LBM generates land masks through superpixel segmentation, which alleviates the impact of land areas on ship detection. Finally, YOLO V5 is applied to identify ship locations as well as their types.
Figure 1. Ship inspection network model.

3.1. Feature Enhancement Module (FEM)

3.1.1. Overall Framework of Feature Enhancement Modules

Common filtering methods include mean filtering, median filtering, local filtering, sigma filtering, Frost filtering, and gamma filtering [52]. Among these methods, mean and median filtering have poor edge retention capabilities. Although local filtering has better edge retention, the Signal-to-Noise Ratio and Peak Signal-to-Noise Ratio are very low. gamma, sigma and Frost filters have low effective coefficients and generate significant changes in the mean of the original image’s pixel values. Therefore, to reduce noise, we comprehensively choose the Lee filter. This module first performs a Lee filtering on Figure 2a to obtain Figure 2b. To further increase the contrast between the object and the image background, a non-linear transformation is performed on Figure 2b to obtain Figure 2d, the non-linear transformation process is given by
y = [ c ( x / 255 ) α ] × 255
where x is the pixel value of the original image; c is a constant valued by 1; α is the coefficient here valued by 1.5; and y is the pixel value at the corresponding position of x after non-linear processing.
Figure 2. Image processing engineering: (a) original image; (b) Lee filtered image; (c) threshold segmented image; (d) linear transformed image; (e) image after fusion of (d) and (c).
To weaken the loss of image information generated by the filtering and non-linear transformation, a threshold is applied to Figure 2a to obtain Figure 2c. It is then fused with Figure 2d.

3.1.2. Salient Otsu (S-Otsu) Segmentation

An improved S-Otsu is applied to automatically find the threshold because pixel differences are image dependent.
Otsu is an algorithm that automatically determines the image binarization threshold [53]. The algorithm sets a threshold, T, to make a difference between the image background and potential objects and searches for the T value that maximizes the variance between the two, and then T is the final threshold. The calculation of the between-class variance is given by
σ = w 0 × w 1 × ( u 0 u 1 )
where σ is the between-class variance between the background and the foreground; w 0 is the proportion of the background pixel in the entire image; w 1 is the proportion of the foreground pixel in the entire image; u 0 is the average gray value of the background; and u 1 is the average gray value of the foreground.
Although Otsu threshold segmentation can achieve good results for most images, the effect is not ideal for SAR images with a lot of noise, as shown in Figure 3. One can remark that ships with higher gray values in the picture are salient objects. To this end, we further design a new threshold segmentation method, which is mainly designed according to the significant difference between the noise area and the ship’s area. First, we define the difference of each pixel relative to other pixels, replace the original pixel value with this difference, and then calculate a new image segmentation threshold through Otsu.
Figure 3. SAR images under different noises: (a) SAR image with a lot of noise; (b) SAR image with less noise; (c) SAR image with a small amount of noise.
Since the ratio between the object and the background areas is uncertain, when the ratio of the object area is greater than the background area, the background area is mistaken as a significant area. To avoid this situation, the calculation of the gray value distance between two pixels is given by
D i = j = 1 , i j N P i P j , ( P i P j > 0 )
where P i denotes the gray value of the original pixel; D i represents the sum of the gray value differences between the original and other pixels; and ( P i P j ) considers positive values.
The new pixel value is given by
D i = D i D m i n D m a x D m i n × 255
where D i is the new gray value of the pixel; D m i n is the minimum value of the sum of the gray value differences between a pixel and the other pixels in the original image; and D m a x is the maximum value of the sum of the gray value differences between a pixel and the other pixels in the original image.
The pixel values of the image after threshold segmentation is given by
y ( i , j ) { x ( i , j ) , x ( i , j ) T 0   ,   x ( i , j ) < T
where y ( i , j ) represents the new gray value of the pixel (i, j);   x ( i , j ) represents the original gray value of the pixel (i, j); and T is the threshold obtained by S-Otsu.

3.1.3. Effect of S-Otsu

We selected three types of images (SAR images containing high noise regions, SAR images containing low-noise regions, SAR images containing land regions) from the SSDD dataset to evaluate the effect of our proposed S-Otsu. As shown in Figure 4, S-Otsu has no additional adverse effects compared to OTSU in SAR image maps that contain low-noise areas and SAR images that contain land areas. In images containing noise, S-Otsu can better segment objects from noisy regions compared to Otsu.
Figure 4. Segmentation effect of S-Otsu on SAR images in different scenarios: (a) Original SAR image; (b) Segmentation effect of Otsu on SAR images; (c) Segmentation effect of S-Otsu on SAR images.

3.1.4. Image Fusion

Finally, the gray values of similar location points in Figure 2c,d are scaled by a factor and added. The function of the coefficient is to prevent the disappearance of the gradient caused by excessively high pixel values in the image fusion process. The fusion of images can be defined as
E ij = α C ij + β D ji
where E ij represents the gray value of the pixel at the position (i, j) in the image e; C ij represents the gray value of the pixel at the position (i, j) in the image c; D ij represents the gray value of the pixel at the position (i, j) in the image d; and α and β are the hyperparameters. The entire image processing steps are illustrated in Figure 2.

3.2. Grayscale Histogram Selection

Because a SAR image does not necessarily contain a land area, we set a threshold by counting the gray value distribution of the two types of images to decide whether to send the FEM processed image to LBM. There are 1160 SAR images in the SSDD dataset, including 194 SAR images containing land and 966 SAR images excluding land areas. We first count the Figure 2c gray value after threshold segmentation in module one to determine whether the Figure 2e in module one enters module two. It appears that most of the gray values of the images containing land areas are concentrated above one million (Figure 5). For SAR images without land areas, most of the gray values are concentrated below one million, therefore the sum of the gray values of each image is valued by one million as a preliminary threshold. This speeds up the image processing process and reduces misjudgments of the subsequent contour screening.
Figure 5. Gray value statistics chart: (a) The gray value of the image that includes the land area; (b) The gray value of the image that does not include the land area.

3.3. Land Burial Module (LBM)

It appears that most SAR image segmentation algorithms often fail to obtain the desired mask when applied to sea and land segmentation contexts. Superpixel segmentation is widely applied in computer vision and includes different approaches such as the K-means algorithm, watershed algorithm, and SLIC generation algorithms [54,55]. The SLIC algorithm has a good edge fitting ability and efficient computation time. We designed this module based on the SLIC algorithm. The processing principle is shown in Figure 6.
Figure 6. Land burial module.

3.3.1. Superpixel Segmentation

The superpixel segmentation aggregates a given pixel-based image towards a regional-level image. After superpixel segmentation, regions of different sizes emerge. We can extract specific information from these regions, such as grayscale histograms and texture information. Superpixel segmentation divides a picture into many small parts along the object boundary to facilitate subsequent mask generations. A suitable k-value (i.e., number of superpixels) should be valued. An example of the k-value segmentation effect is shown in Figure 7. It appears that object edges of SAR images are often unclear, as ship objects are most often too small. We set the k-value to k = 1000 according to this specific case.
Figure 7. Segmentation effect under different k-values: (a) Splitting effect picture when K = 100; (b) Splitting effect picture when K = 1000.

3.3.2. Threshold Processing

Common object segmentation algorithms are based on individual pixels, but as most SAR images are prone to noise with gray values close to the object, straight pixel processing generally does not achieve satisfactory results. We consider superpixels as the basic unit of object binaryzation and take the average gray value of the superpixel area as the gray value of the pixel block. The calculation of the gray value of the pixel block is given by
P m e a n = i = 1 n P i N
where P m e a n denotes the pixel value of the superpixel;   P i is the pixel value of the ith pixel of the original image; and N is the number of pixels in the original image.
The gray value of the superpixel in the noise background area is averaged. If the superpixel is in the object area, the overall gray value of the object area does not change significantly, thus greatly increasing the object and noise background areas. The gray value contrast is more conducive to this threshold segmentation. T is calculated by S-Otsu proposed in the first module and then binarized the image is defined as
y ( i , j ) { 1   , x ( i , j ) T 0   ,   x ( i , j ) < T
where y ( i , j ) represents the new gray value of the pixel point (i, j) and x ( i , j ) represents the original gray value of the pixel point (i, j).
The final image after object segmentation is shown in Figure 8.
Figure 8. Object segmentation results: (a) Original image; (b) Binarized image; (c) After segmentation, it only contains images of coasts and ships.

3.3.3. Generate Mask

This section identifies geometric differences between ships and land areas and then eliminates these ships to generate image masks. Indeed, ships and land features have different aspect ratios and area. Accordingly, we first identify the smallest bounding rectangle of the land area (as shown in Figure 9a). Next, we use the aspect ratio and area of the bounding area rectangle to identify ship and land areas and the final result is shown in Figure 9b. Finally, we can use a mask to remove the land areas from the SAR image.
Figure 9. Geometric feature processing diagram: (a) schematic diagram of the smallest enclosing bounding area of the object; (b) image mask.

4. Results

This section first introduces the dataset and evaluation parameters used in our experiments. Then, we validate the performance of our proposed ship detection model on SSDD and HRISD. Finally, we demonstrate the ability of FLNet to detect ships in different scenarios.

4.1. Dataset Introduction

The SSDD dataset is a public dataset mainly used for SAR ship object detection [56]. The dataset is obtained by clipping the target area and manually annotating the target objects. There are 2578 ships in the SSDD dataset and the size distribution of these ships ranges from several pixels to several hundred pixels, therefore the robustness of the proposed ship detection model can be well verified. A detailed description of the SSDD dataset is briefly summarized in Table 1. Figure 10 shows ships in different environments in the SSDD dataset.
Table 1. Detail of SSDD and HRSID.
Figure 10. Ships in different situations in SSDD: (a) Small ships in SAR images; (b) Near-shore ships in SAR images; (c) Ships affected by high noise in SAR images.
In HRSID, high-resolution SAR images are applied to construct the dataset and they are cropped to 800 × 800 pixel SAR images for better implementation of the functions such as multi-scale training [57]. This dataset contains a large number of ships with complex backgrounds. It was widely used in instance segmentation and ship detection. The details of this dataset are shown in Figure 11 and Table 1.
Figure 11. Ships in different situations in HRSID: (a) Small ships in SAR images; (b) Near-shore ships in SAR images.

4.2. Metrics for Evaluating Model Performance

We use precision (P), average precision (AP), and recall rate (R) to evaluate the performance of our object detection model. The calculation methods of these three indicators are given by
                  P = TP TP + FP
AP = 0 1 P ( R ) dR
Recall = TP TP + FN
where TP represents the number of positive samples correctly identified as positive samples, FP represents the number of negative samples misidentified as positive samples, P(R) represents the P-R curve, FN represents false negative.

4.3. Experimental Results

We trained our proposed FLNet and the Yolov5 separately on the SSDD dataset. To ensure the fairness of the experiment, all experiments are processed on a computer equipped with an intel i5-10600 processor and a Tesla K80 graphics card with 12GB of video memory, CUDA Version: 11.2, the operating system is 64-bit Windows 10. We used the most popular deep learning framework Pytorch. The backbone networks of these two ship detection models use the same pre-trained backbone.
During the training process, we set the learning rate of all model training to 0.01 and the batch size to 64. After 530 rounds of iterative training, the comparison of the detection effects of the two algorithms is shown in Table 2, Figure 12 and Figure 13.
Table 2. Experimental results on SSDD dataset.
Figure 12. Comparison of the P-R curves of the two detection algorithms on SSDD: (a) P-R curve of FLNet on SSDD dataset; (b) P-R curve of YOLO V5 on SSDD dataset.
Figure 13. Performance of the two detection algorithms in the training process.
From Table 3 and Figure 12, it appears that our improved algorithm has a 7% increase in precision, a 6.5% increase in recall rate, and a 2.3% increase in AP value compared with the original algorithm. It can be seen from Figure 13 that the performance of the improved algorithm in the training process also far exceeds the one of YOLO V5.
Table 3. Experimental results on HRSID dataset.
To further verify the robustness of FLNet, we evaluated the detection capability of FLNet on ships using HRSID. The experimental results are shown in Figure 14 and Table 3. We can see from the experimental results that our proposed FLNet achieves SOTA on all metrics compared to YOLOV5.
Figure 14. Comparison of the P-R curves of the two detection algorithms on HRSID: (a) P-R curve of YOLO V5 on HRSID dataset; (b) P-R curve of FLNet on HRSID dataset.

4.4. Visualization of Ship Detection Effect

To more intuitively illustrate the effectiveness of our improvements, we compared FLNet and the baseline YOLO V5 in three scenarios. The three scenarios include detection of ships affected by high noise, detection of ships affected by coastal environments, and intensive detection of small ships.

4.4.1. Detection of Ships Affected by High Noise

As shown in Figure 15, due to the existence of a large amount of noise in the SAR image, the detection effect of the ship detection model on ships is seriously affected. In Figure 15b, YOLO V5 did not detect the presence of the ship, while our proposed FLNet correctly detected the position of the ship with a confidence of 0.94, as shown in Figure 15c.
Figure 15. Ship detection is affected by high noise in SAR images. (a) Correct position of the ship in the SAR image; (b) Ship detection effect of YOLO V5 in SAR images; (c) Ship detection effect of FLNet in SAR images.

4.4.2. Detection of Ships Affected by the Coastal Environment

As shown in Figure 16, since the gray value of the coast is close to the gray value of the ship and the ship is close to the coast, the detection effect of the ship detection model on the ship is seriously affected. In Figure 16b, YOLO V5 has seriously missed detections, while our proposed FLNet correctly detected three ships, as shown in Figure 16c.
Figure 16. Ship detection is affected by the coastal environment in SAR images. (a) Correct position of the ship in the SAR image; (b) Ship detection effect of YOLO V5 in SAR images; (c) Ship detection effect of FLNet in SAR images.

4.4.3. Detection of Dense Small Ships

The detection of small ships is a test of the performance of the ship detection model. As shown in Figure 17, there are a large number of small ships in the SAR image. While YOLO V5 missed one ship in the SAR image, (as shown in Figure 17b) our YOLOV5 correctly identified all ships (as shown in Figure 17c).
Figure 17. Detection of dense small ships in SAR images. (a) Correct position of the ship in the SAR image; (b) Ship detection effect of YOLO V5 in SAR images; (c) Ship detection effect of FLNet in SAR images.

4.5. Comparison to Related Works

FLNet was compared with other deep learning ship detection models (faster RCNN, SSD, YOLO V4, HR-SDNet) on the SSDD dataset. The comparison results are shown in Table 4. The experimental results show that, in comparison to other algorithms, FLNet achieved SOTA with all indicators.
Table 4. Comparison results with other algorithms.

5. Discussion

This section first discusses the impact of image fusion weights in FEM on detection results and then validates the effectiveness of our proposed module through ablation experiments. Finally, the failure cases of ship detection using FLNet are analyzed.

5.1. Influence of Image Fusion Weight in FEM on the Detection Result

This section discusses the influence of the image fusion weight in the FE module on the detection performance of FLNet. We set 0.25 + 0.25, 0.5 + 0.5, 0.75 + 0.75, 1 + 1 under the condition that other conditions are the same. The four sets of weights are tested. Experimental results are shown in the following Figure 18 and Table 5.
Figure 18. The influence of four different weights set during image fusion on the P-R curve. (a) P-R curve of FLNet on SSDD dataset when the image fusion weight is 0.25; (b) P-R curve of FLNet on SSDD dataset when the image fusion weight is 0.5; (c) P-R curve of FLNet on SSDD dataset when the image fusion weight is 0.75; (d) P-R curve of FLNet on SSDD dataset when the image fusion weight is 1.
Table 5. Experimental results of FLNet with different fusion weights on the SSDD dataset.
It can be seen from the experiments that the weight of image fusion impacts the detection capability of the network model. By appropriately lifting the weight of image fusion, the gray value difference between the ship and the background can be enhanced, which is more conducive to ship detection. However, a too large image fusion weight is likely to withdraw texture information of the ship itself, thus affecting the performance of the ship detection model. This indicates that not only the differences between the ship and the background but also the characteristics of the ship itself should be considered when performing ship detection.

5.2. Ablation Experiment

In order to verify the effectiveness of our proposed FEM and LBM, we used YOLO V5 as the baseline to verify the effects of different combinations of the two modules on the ship detection effect. The experimental results are shown in Table 6.
Table 6. Ablation experiment results on SSDD dataset.
When adding FEM to YOLO V5 alone, ship detection accuracy is improved by 4.4%, recall rate by 7%, and AP by 2.3%. As shown in Figure 15, and due to the significant presence of noise, applying YOLO V5 alone is not satisfactory. On the other hand, FLNet first enhances ship features in the SAR image by removing noise and then facilitates the identification of the ship type and location. When adding LBM to YOLO V5 alone, ship detection accuracy is improved by 3.9%, recall rate by 3%, and AP by 1.1%. The LBM objective is to mainly detect near-shore ships. As shown in Figure 16, because land and ship gray values are close to each other, this potentially leads to false detection when using YOLO V5 to detect them. By applying LBM, land area is removed according to the geometric feature difference between the ship and the coast. When both FEM and LBM are added to YOLOV5, all indicators of the improved ship detection model reach SOTA except the recall rate, which is slightly decreased compared to FEM alone. Overall, the experiment results show that the performance of the ship detection model can be effectively enhanced by removing the noise background area. In addition, the decline in recall rate may be caused by the fact that our LBM can not completely distinguish land and ship according to the geometric characteristics of land and coast.

5.3. Bad Case Analysis

Experimental results on the SSDD and HRSID datasets show our improvements’ effectiveness. However, FLNet may miss detection in some specific scenes (as shown in Figure 19). Because the ship in the image is too close to the coast or other ships, the LBM module may incorrectly identify the ship as the coast, thus increasing the missed detection rate of the ship. We will pay more attention to the identification of such ships in future research.
Figure 19. Case of ship detection failure. The green target box is the correct position of the ship and the red target box is the propagation position predicted by FLNet.

6. Conclusions

This paper introduces a novel SAR ship detection model named FLNet. FLNet combines common image processing algorithms with a deep learning-based object detection algorithm, YOLO V5, to effectively suppress the influence of background information and noise information. The feature enhancement module combines Lee filtering with our improved Otsu threshold method to effectively enhance the SAR image objects and reduce noise. The land burial module eliminates the influence of land areas on ship detection based on geometric and gray value differences. The experimental results show that the proposed FLNet significantly improves the accuracy and recall rate as compared with YOLO V5. FLNet still misidentified a few near-shore ships in the land burial module. These misidentified ships will be eliminated. Therefore, our next step will be to study how to use deep learning algorithms to remove coastal areas to increase the performance of the ship detection model.

Author Contributions

Conceptualization, H.Z.; software, H.Z.; validation, H.Z.; resources, H.Z.; writing, H.Z.; visualization, H.Z.; supervision, G.T. and C.C.; project administration, G.T.; funding, editing, S.M. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded in part by the National Natural Science Foundation of China under Grant 82004259, in part by the Guangdong Basic and Applied Basic Research Foundation under Grant 2020A1515110503, and in part by the Guangzhou Basic and Applied Basic Research Project under Grant 202102020674.

Institutional Review Board Statement

Not applicable for studies not involving humans or animals.

Data Availability Statement

The data and code that support the findings of this study are available on request from the second author.

Acknowledgments

This research was supported in part by the National Natural Science Foundation of China (NO. 82004259), in part by the Guangdong Basic and Applied Basic Research Foundation (NO. 2020A1515110503) and in part by the Guangzhou Basic and Applied Basic Research Project (NO. 202102020674).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Chang, Y.-L.; Anagaw, A.; Chang, L.; Wang, Y.C.; Hsiao, C.-Y.; Lee, W.-H. Ship Detection Based on YOLOv2 for SAR Imagery. Remote Sens. 2019, 11, 786. [Google Scholar] [CrossRef]
  2. Liu, T.; Zhang, J.; Gao, G.; Yang, J.; Marino, A. CFAR ship detection in polarimetric synthetic aperture radar images based on whitening filter. IEEE Trans. Geosci. Remote Sens. 2019, 58, 58–81. [Google Scholar] [CrossRef]
  3. Tang, G.; Liu, S.; Fujino, I.; Claramunt, C.; Wang, Y.; Men, S. H-YOLO: A Single-Shot Ship Detection Approach Based on Region of Interest Preselected Network. Remote Sens. 2020, 12, 4192. [Google Scholar] [CrossRef]
  4. Wang, X.; Chen, C.; Pan, Z.; Pan, Z. Fast and automatic ship detection for SAR imagery based on multiscale contrast measure. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1834–1838. [Google Scholar] [CrossRef]
  5. Li, H.; Cui, X.; Chen, S. PolSAR Ship Detection with Optimal Polarimetric Rotation Domain Features and SVM. Remote Sens. 2021, 13, 3932. [Google Scholar] [CrossRef]
  6. Ai, J.; Tian, R.; Luo, Q.; Jin, J.; Tang, B. Multi-scale rotation-invariant Haar-like feature integrated CNN-based ship detection algorithm of the multiple-target environment in SAR imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 10070–10087. [Google Scholar] [CrossRef]
  7. Zhao, D.; Zhu, C.; Qi, J.; Qi, X.; Su, Z.; Shi, Z. Synergistic Attention for Ship Instance Segmentation in SAR Images. Remote Sens. 2021, 13, 4384. [Google Scholar] [CrossRef]
  8. Liu, M.; Chen, S.; Lu, F.; Xing, M.; Wei, J. Realizing Target Detection in SAR Images Based on Multiscale Superpixel Fusion. Sensors 2021, 21, 1643. [Google Scholar] [CrossRef]
  9. Wang, X.; Chen, C.; Pan, Z.; Pan, Z. Superpixel-based LCM detector for faint ships hidden in strong noise background SAR imagery. IEEE Geosci. Remote Sens. Lett. 2018, 16, 417–421. [Google Scholar] [CrossRef]
  10. Deng, Y.; Wang, H.; Liu, S.; Sun, M.; Li, X. Analysis of the ship target detection in high-resolution SAR images based on information theory and Harris corner detection. J. Wirel. Com. Netw. 2018, 2018, 291. [Google Scholar] [CrossRef]
  11. Wang, Y.; Liu, H. PolSAR Ship Detection Based on Superpixel-Level Scattering Mechanism Distribution Features. IEEE Geosci. Remote Sens. Lett. 2015, 12, 1780–1784. [Google Scholar] [CrossRef]
  12. Liu, Y.; Zhang, S.; Suo, J.; Zhang, J.; Yao, T. Research on a new comprehensive CFAR (Comp-CFAR) processing method. IEEE Access 2019, 7, 19401–19413. [Google Scholar] [CrossRef]
  13. Kuttikkad, S.; Chellappa, R. Non-Gaussian CFAR techniques fortarget detection in high resolution SAR images. Proc. AICIP 1994, 1, 910–914. [Google Scholar]
  14. Hofele, F.X. An innovative CFAR algorithm. In 2001 CIE International Conference on Radar Proceedings (Cat No. 01TH8559); IEEE: Piscataway, NJ, USA, 2001; pp. 329–333. [Google Scholar]
  15. Novak, L.M.; Hesse, S.R. On the performance of order-statistics CFAR detectors. In Conference Record of the Twenty-Fifth Asilomar Conference on Signals, Systems & Computers; IEEE: Piscataway, NJ, USA, 1991; Volume 2, pp. 835–840. [Google Scholar]
  16. Di Bisceglie, M.; Galdi, C. CFAR detection of extended objects in high-resolution SAR images. IEEE Trans. Geosci. Remote Sens. 2005, 43, 833–843. [Google Scholar] [CrossRef]
  17. Yang, M.; Pei, D.; Ying, N.; Guo, C. An information-geometric optimization method for ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 4005305. [Google Scholar] [CrossRef]
  18. Wang, X.; Li, G.; Zhang, X.P.; He, Y. Ship detection in SAR images via local contrast of Fisher vectors. IEEE Trans. Geosci. Remote Sens. 2020, 58, 6467–6479. [Google Scholar] [CrossRef]
  19. Zhu, M.; Hu, G.; Zhou, H.; Wang, S.; Feng, Z.; Yue, S. A Ship Detection Method via Redesigned FCOS in Large-Scale SAR Images. Remote Sens. 2022, 14, 1153. [Google Scholar] [CrossRef]
  20. Xu, C.; Yin, C.; Wang, D.; Han, W. Fast ship detection combining visual saliency and a cascade CNN in SAR images. IET Radar Sonar Navig. 2020, 14, 1879–1887. [Google Scholar] [CrossRef]
  21. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. TPAMI 2015, 28, 1137–1149. [Google Scholar] [CrossRef]
  22. Tang, G.; Zhuge, Y.; Claramunt, C.; Men, S. N-YOLO: A SAR Ship Detection Using Noise-Classifying and Complete-Target Extraction. Remote Sens. 2021, 13, 871. [Google Scholar] [CrossRef]
  23. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  24. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.Y.; Berg, A.C. Ssd: Single shot multibox detector. In European Conference on Computer Vision; Springer: Cham, Switzerland, 2016; pp. 21–37. [Google Scholar]
  25. Jia, F.; Li, S.; Zuo, H.; Shen, J. Deep neural network ensemble for the intelligent fault diagnosis of machines under imbalanced data. IEEE Access 2020, 8, 120974–120982. [Google Scholar] [CrossRef]
  26. Li, Y.; Zhang, S.; Wang, W.Q. A lightweight faster R-CNN for ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2020, 19, 4006105. [Google Scholar] [CrossRef]
  27. Ma, X.; Ji, K.; Xiong, B.; Zhang, L.; Kuang, G. Light-YOLOv4: An Edge-Device Oriented Target Detection Method for Remote Sensing Images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 10808–10820. [Google Scholar] [CrossRef]
  28. Zhou, J.; Jiang, P.; Zou, A.; Chen, X.; Hu, W. Ship Target Detection Algorithm Based on Improved YOLOv5. J. Mar. Sci. Eng. 2021, 9, 908. [Google Scholar] [CrossRef]
  29. Lin, Z.; Ji, K.; Leng, X.; Kuang, G. Squeeze and excitation rank faster R-CNN for ship detection in SAR images. IEEE Geosci. Remote Sens. Lett. 2018, 16, 751–755. [Google Scholar] [CrossRef]
  30. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 2117–2125. [Google Scholar]
  31. Wei, S.; Su, H.; Ming, J.; Wang, C.; Yan, M.; Kumar, D.; Shi, J.; Zhang, X. Precise and Robust Ship Detection for High-Resolution SAR Imagery Based on HR-SDNet. Remote Sens. 2020, 12, 167. [Google Scholar] [CrossRef]
  32. Cui, Z.; Li, Q.; Cao, Z.; Liu, N. Dense attention pyramid networks for multi-scale ship detection in SAR images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8983–8997. [Google Scholar] [CrossRef]
  33. Zhang, P.; Tang, J.; Zhong, H.; Ning, M.; Liu, D.; Wu, K. Self-trained target detection of radar and sonar images using automatic deep learning. IEEE Trans. Geosci. Remote Sens. 2021, 60, 4701914. [Google Scholar] [CrossRef]
  34. Zhang, T.; Zhang, X.; Ke, X. Quad-FPN: A Novel Quad Feature Pyramid Network for SAR Ship Detection. Remote Sens. 2021, 13, 2771. [Google Scholar] [CrossRef]
  35. Gao, G.; Gao, S.; He, J.; Li, G. Ship detection using compact polarimetric SAR based on the notch filter. IEEE Trans. Geosci. Remote Sens. 2018, 56, 5380–5393. [Google Scholar] [CrossRef]
  36. Leng, X.; Ji, K.; Zhou, S.; Xing, X. Ship detection based on complex signal kurtosis in single-channel SAR imagery. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6447–6461. [Google Scholar] [CrossRef]
  37. Lang, H.; Xi, Y.; Zhang, X. Ship detection in high-resolution SAR images by clustering spatially enhanced pixel descriptor. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5407–5423. [Google Scholar] [CrossRef]
  38. Liu, G.; Zhang, X.; Meng, J. A Small Ship Target Detection Method Based on Polarimetric SAR. Remote Sens. 2019, 11, 2938. [Google Scholar] [CrossRef]
  39. Zhang, L.; Zhang, Z.; Lu, S.; Xiang, D.; Su, Y. Fast Superpixel-Based Non-Window CFAR Ship Detector for SAR Imagery. Remote Sens. 2022, 14, 2092. [Google Scholar] [CrossRef]
  40. Wang, X.; Li, G.; Plaza, A.; He, Y. Ship detection in SAR images by aggregating densities of Fisher vectors: Extension to a global perspective. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5206613. [Google Scholar] [CrossRef]
  41. Hu, J.; Shen, L.; Sun, G. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 7132–7141. [Google Scholar]
  42. Woo, S.; Park, J.; Lee, J.Y.; Kweon, I.S. Cbam: Convolutional block attention module. In Proceedings of the European Conference on Computer Vision (ECCV); Springer: Berlin/Heidelberg, Germany, 2018; pp. 3–19. [Google Scholar]
  43. Liu, T.; Luo, R.; Xu, L.; Feng, D.; Cao, L.; Liu, S.; Guo, J. ECA-Net: Efficient Channel Attention for Deep Convolutional Neural Networks. In 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR); IEEE: Piscataway, NJ, USA, 2020. [Google Scholar]
  44. Liu, S.; Qi, L.; Qin, H.; Shi, J.; Jia, J. Path aggregation network for instance segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 8759–8768. [Google Scholar]
  45. Chen, L.C.; Papandreou, G.; Kokkinos, I.; Murphy, K.; Yuille, A.L. Deeplab: Semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 40, 834–848. [Google Scholar] [CrossRef]
  46. Yang, X.; Zhang, X.; Wang, N.; Gao, X. A Robust One-Stage Detector for Multiscale Ship Detection with Complex Background in Massive SAR Images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5217712. [Google Scholar] [CrossRef]
  47. Guo, H.; Yang, X.; Wang, N.; Gao, X. A CenterNet plus plus model for ship detection in SAR images. Pattern Recognit. 2021, 112, 107787. [Google Scholar] [CrossRef]
  48. Zhang, T.; Zhang, X. ShipDeNet-20: An only 20 convolution layers and <1-MB lightweight SAR ship detector. IEEE Geosci. Remote Sens. Lett. 2020, 18, 1234–1238. [Google Scholar] [CrossRef]
  49. Sun, Y.; Sun, X.; Wang, Z.; Fu, K. Oriented ship detection based on strong scattering points network in large-scale SAR images. IEEE Trans. Geosci. Remote Sens. 2021, 60, 5218018. [Google Scholar] [CrossRef]
  50. Ge, J.; Wang, C.; Zhang, B.; Xu, C.; Wen, X. Azimuth-Sensitive Object Detection of High-Resolution SAR Images in Complex Scenes by Using a Spatial Orientation Attention Enhancement Network. Remote Sens. 2022, 14, 2198. [Google Scholar] [CrossRef]
  51. Yu, L.; Wu, H.; Zhong, Z.; Zheng, L.; Deng, Q.; Hu, H. TWC-Net: A SAR Ship Detection Using Two-Way Convolution and Multiscale Feature Mapping. Remote Sens. 2021, 13, 2558. [Google Scholar] [CrossRef]
  52. Sivaranjani, R.; Roomi, S.M.M.; Senthilarasi, M. Speckle noise removal in SAR images using Multi-Objective PSO (MOPSO) algorithm. Appl. Soft Comput. 2019, 76, 671–681. [Google Scholar] [CrossRef]
  53. Xu, S.; Liao, Y.; Yan, X.; Zhang, G. Change detection in SAR images based on iterative Otsu. Eur. J. Remote Sens. 2020, 53, 331–339. [Google Scholar] [CrossRef]
  54. Ren, X.; Malik, J. Learning a classification model for segmentation. In IEEE International Conference on Computer Vision; IEEE Computer Society: Piscataway, NJ, USA, 2003; Volume 2, p. 10. [Google Scholar]
  55. Achanta, R.; Shaji, A.; Smith, K.; Lucchi, A.; Fua, P.; Süsstrunk, S. SLIC superpixels compared to state-of-the-art superpixel methods. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 34, 2274–2282. [Google Scholar] [CrossRef] [PubMed]
  56. Zhang, T.; Zhang, X.; Li, J.; Xu, X.; Wang, B.; Zhan, X.; Xu, Y.; Ke, X.; Zeng, T.; Su, H.; et al. SAR Ship Detection Dataset (SSDD): Official Release and Comprehensive Data Analysis. Remote Sens. 2021, 13, 3690. [Google Scholar] [CrossRef]
  57. Wei, S.; Zeng, X.; Qu, Q.; Wang, M.; Su, H.; Shi, J. HRSID: A high-resolution SAR images dataset for ship detection and instance segmentation. IEEE Access 2020, 8, 120234–120254. [Google Scholar] [CrossRef]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.