Next Article in Journal
Simulating Two-Phase Seepage in Undisturbed Soil Based on Lattice Boltzmann Method and X-ray Computed Tomography Images
Previous Article in Journal
Research on the Influence of Core Sensing Components on the Performance of Galvanic Dissolved Oxygen Sensors
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A SAR Ship Detection Method Based on Adversarial Training

Naval Submarine Academy, Qingdao 264001, China
*
Author to whom correspondence should be addressed.
Sensors 2024, 24(13), 4154; https://doi.org/10.3390/s24134154
Submission received: 30 April 2024 / Revised: 12 June 2024 / Accepted: 24 June 2024 / Published: 26 June 2024
(This article belongs to the Section Remote Sensors)

Abstract

:
SAR (synthetic aperture radar) ship detection is a hot topic due to the breadth of its application. However, limited by the volume of the SAR image, the generalization ability of the detector is low, which makes it difficult to adapt to new scenes. Although many data augmentation methods—for example, clipping, pasting, and mixing—are used, the accuracy is improved little. In order to solve this problem, the adversarial training is used for data generation in this paper. Perturbation is added to the SAR image to generate new samples for training, and it can make the detector learn more abundant features and promote the robustness of the detector. By separating batch normalization between clean samples and disturbed images, the performance degradation on clean samples is avoided. By simultaneously perturbing and selecting large losses of classification and location, it can keep the detector adaptable to more confrontational samples. The optimization efficiency and results are improved through K-step average perturbation and one-step gradient descent. The experiments on different detectors show that the proposed method achieves 8%, 10%, and 17% AP (Average Precision) improvement on the SSDD, SAR-Ship-Dataset, and AIR-SARShip, compared to the traditional data augmentation methods.

1. Introduction

Synthetic aperture radar (SAR) is a kind of all-day and all-weather tool which can provide high-resolution images. It is widely used in civil and military areas. Among them, finding ships from SAR images is one of the most popular directions for this tool [1].
Traditionally, the Constant False Alarm Rate (CFAR) detector is usually adopted in this area [2]. It firstly models the pixel distribution of the SAR image, and then calculates the model parameters. The discrimination threshold is computed according to the false alarm rate and the target pixel value. Pixel values that are higher than the threshold will be regarded as ship pixels, and those lower than the threshold will be regarded as background. Finally, these high-amplitude pixels are extracted and merged into targets. In essence, CFAR is a kind of unsupervised method in machine learning view, and its performance must be worse than the supervised method.
In addition to CFAR, the machine-learning-based detectors are also used in SAR ship detection. Feature extracting and classifier designing are the two core steps, and they show advantages compared to CFAR, as they are strongly supervised. With the emergence of deep learning in 2012, and its excellent performance in various computer vision tasks, SAR ship detection also began to use deep learning. The beginning of this was the SSDD dataset and the great performance of deep learning shown in the following paper [3]. After that, a large number of researchers began to seek solutions from deep learning, and a lot of papers were published [4,5]. The deep learning models are data-hungry; thus, if the dataset is not big enough, the models may face the problem of over-fitting. This problem is serious, especially in the scenarios where it is difficult to collect data, for example, SAR ship detection.
Data augmentation is an important method to hold back over-fitting, which is usually used in general object detection. It can provide more diversified samples by changing the appearance of images, and can extract more information from them; thus, it can make the model more generalized. It can also be divided into basic and advanced data augmentations. The representative methods of the former are manipulation and erasing, while the representative methods of the latter are mixing and auto-augment. The above methods are based on experience, and the generalization and effectiveness of them are poor in general.
As we all know, deep learning models are highly susceptible to adversarial attack. Adding an imperceptible disturbance to the image can lead to a wrong prediction [6]. We think that if we use adversarial examples in the right way, they can improve the performance of the detector, as they can generate additional training images by adding noise, so that we can regard them as one way of data augmentation.
This can be realized by adversarial training. Adversarial training adds adversarial samples to the training set, making the model gradually adapt to adversarial attacks during training. In this process, the sample is mixed with some small perturbations, and then the deep learning models adapt to this change, thus becoming more robust [7].
Therefore, in order to improve the performance of the SAR ship detector, images generated by adversarial learning are used for data generation [8]. The clean and adversarial images are normalized separately to avoid their influence on the clean ones. Through the K-step gradient ascent and one-step descent method, the detector is trained finely. The high-quality adversarial images improve the classification and localization ability of the deep learning models. Experiments on several detectors and datasets are conducted, and they are shown to have great advantages over traditional data augmentation methods.
The achievements of the paper are as follows:
(1)
The adversarial learning is used in SAR ship detection, and the principles and formulas of adversarial learning are described in detail.
(2)
The optimization method based on separated batch normalization (BN), selecting largely from classification and localization losses, and the K-step gradient ascent with one-step gradient descent are used to solve the adversarial learning formulas.
(3)
The experiments on the classical detectors and public datasets are conducted, which show obvious superiority compared to the traditional data augmentation methods.

2. Methodology

An adversarial attack can make the deep learning models output a wrong prediction, which shows great risk to the application. However, if we add adversarial samples to the input and use them to update the parameters, the models can learn to resist them, and the robustness can be improved. This is the core idea of adversarial training, which can be used as a way of data generation. When conducting adversarial training, some small perturbations (random noise) will be mixed into the training data (changes are small, but may lead to the wrong prediction), and then the deep learning models will be adapted to such changes [9].
Suppose that the objective function of training is as follows:
arg min θ E x , y ~ D L θ , x , y
Here, D represents the distribution of the input, L represents the loss function, θ represents the parameters of the model, x is the training data, and y is its corresponding real label [10].
The adversarial training is designed to determine the best parameters to minimize the maximum risk, as follows:
arg min θ E x , y ~ D max ε S L θ , x + ε , y
Here, ε is the interference, and S is its interference range. The purpose of ε is to maximize the loss function L θ , x + ε , y , which means trying to make the models’ output a wrong prediction. ε should not be too big, and it is usually limited by norm, for example ε F c o n s t a n t . Every sample x can generate an adversarial sample x + ε . The training sample and label x + ε , y are used to update the parameters θ and minimize the loss function.
In the above equation, ‘max’ means that we need to find a group of adversarial samples with the maximum loss within the sample space. The outer ‘min’ function refers to the fact that, faced with the sample set of such data distribution, the expected loss of the model on the adversarial sample set should be minimized by updating the model parameters [11,12,13].
The min–max formula can improve the performance of adversarial examples. However, they cannot generalize well-to-clean images. This is because the adversarial examples and clean images have different underlying distributions, which prevents the network from accurately and efficiently extracting valuable features from these two domains. Therefore, we use their respective BNs during training. During the prediction, only the BN parameters of clean samples are used. The formula is as follows:
arg min θ E x , y ~ D L θ , x , y + max ε S   L θ , x + ε , y
This training method, handling the variation in clean and adversarial samples alone, results in a better learning performance.
The loss function of object detection can be represented as follows:
L θ , x , y = L c l s θ , x , y + w L l o c θ , x , y
Here, w is used for balancing the two losses. It includes classification subtask L c l s and localization subtask L l o c . Both of them can be attacked and result in wrong predictions. Therefore, the purpose can be accomplished by attacking one of the subtasks. In order to fully consider this point, the larger of the classification loss and localization loss is selected as the final loss. This method enables the model to adapt to more confrontational samples.
For deep learning models, the outer ‘min’ is non-convex, and the inner ‘max’ is non-concave. The SGD can be used for the external minimization, and the PGD (Projected Gradient descent) can be used for the internal maximization. The usually used K-step PGD requires K forward and backward passes across the network, whereas standard SGD updates only require one pass [14,15,16,17]. Thus, the generation procedure of adversarial training can increase the running time by an order of magnitude. For improving the training speed and increasing the accuracy, the following steps are proposed.
The gradient g of loss function is accumulated in K steps, as follows:
g t = g t 1 + 1 K E x , y ~ D [ θ L θ , x + ε , y ]
It uses the average gradient in K iterations. The disturbance is updated by gradient ascent, as follows:
ε t = | | ε | | F constant ε t 1 + α g a d v / | | g a d v | | F
In which α is ascent step size and g a d v = ε L θ θ , x + ε t 1 , y .
Then, the parameters of deep learning models are updated by gradient descent, as follows:
θ m = θ m 1 τ g K
In which τ is the learning rate.
The above method can be used to optimize Formula (1) quickly and accurately.
The overall architecture of the proposed method is summarized in Figure 1.

3. Experimental Results

3.1. Experiment Setup and Baseline

The experiments use the Ubuntu system, Intel (R) Core (TM) i9-10900K processor, GeForce RTX 3080 GPU (NVIDIA Corporate, Santa Clara, CA, USA). The Pytorch framework is used. The initial learning rate is 0.001, the minimum learning rate is 0.00001, and the stepwise learning rate is adopted. The four representative deep learning-based detectors are chosen to conduct the experiments, and they are as follows: two-stage detector Faster R-CNN [18], one-stage detector YOLOV3 [19], focal loss-based detector RetinaNet [20], and the anchor-free detector CenterNet [21]. The experiments are conducted on the three public datasets, SSDD, SAR-Ship-Dataset [22], and AIR-SARShip [23]. The data augmentation methods of cutout, random erasing, mixup, cutmix, augmix, gridmask, mosic, and copy/paste are selected as the traditional method [24,25,26,27,28,29,30,31,32]. The AP50 is used as the evaluation indicator.
The deep learning-based detector usually adopts some data augmentation methods. For example, Faster R-CNN uses resizing and random flipping; YOLOV3 uses resizing, random flipping, and photo metric distortion; CenterNet uses the resizing and random flipping; and RetinaNet uses the resizing and random flipping. We call them the build-in data augmentation methods. The performances of them are regarded as the baseline in this letter, as shown in Table 1.
In Table 1, FR, RN, Y3, CN, and AVE are the abbreviations of Faster R-CNN, RetinaNet, YOLOV3, CenterNet, and average, respectively. In Table 2, Table 3 and Table 4, DAM refers to the data augmentation method.

3.2. Performance on Different Datasets

3.2.1. Performance on SSDD

SSDD was first made public in 2017, and it introduced deep learning into the SAR ship detection in 2017. It has 1160 images and 2456 ships. It has many large sized ships, and the ships are densely arranged in some images. The samples of SSDD are shown in Figure 2.
The results of different data augmentation methods and detectors on SSDD are shown in Table 2.
For the traditional data augmentation methods, 32 experiments are conducted. Compared to the build-in methods, 11 of them were increased, with an average increase of 5%, and 21 of them were decreased, with an average decrease of 7.9%. The average AP50 of the 32 combination is 0.85. The average of the build-in data augmentation method in Table 1 is 0.89. This shows that the traditional data augmentation methods would decrease the performance of the detectors overall. Only the specific combination of the detector and data augmentation methods can boost the performance. For example, Faster R-CNN with augmix can boost the performance from 84% to 92% (with 8% improvement) on SSDD.
Compared to the mediocre ability of the traditional data augmentation method, the proposed adversarial training-based method can receive a huge improvement with all of the detectors. For example, it receives 0.94, 0.95, 0.89, and 0.95 AP50 on SSDD, respectively. The average value is 0.93. It receives an improvement of 0.04 and 0.08, compared to the build-in and traditional methods.
What is more, in SSDD, the number of images is small, and most of them are easy samples. The performance improvement of the proposed method is limited, compared to other datasets, as shown below.

3.2.2. Performance on SAR-Ship-Dataset

Compared to SSDD, the SAR-Ship-Dataset has more samples. It has 43,819 images and 59,535 ships in total. They are from Gaofen-3 and Sentinel-1. Samples in it have diversity scales and conditions. Figure 3 shows some examples of this.
The results of different data augmentation methods and detectors on the SAR-Ship-Dataset are shown in Table 3.
For the traditional data augmentation methods, 32 experiments are conducted. Compared to the build-in methods, 17 of them were increased, with an average AP increase of 2.9%, and 15 of them were decreased, with an average decrease of 7.2%. The average AP50 of the 32 combination is 0.83. The average of the build-in data augmentation method in Table 1 is 0.85. This shows that the traditional data augmentation methods would decrease the performance of the detectors overall. Only the specific combination of the detector and data augmentation methods can boost the performance. For example, YOLOV3 with mosic can boost the performance from 79% to 84% (with 5% improvement) on the SAR-Ship-Dataset.
Compared to the plain ability of the traditional data augmentation method, the proposed adversarial training-based method can receive a huge improvement with all of the detectors. For example, it receives 0.93, 0.94, 0.89, and 0.94 AP50 on the SAR-Ship-Dataset, respectively. The average value is 0.93. It receives an improvement of 0.08 and 0.10, compared to the build-in traditional methods.
From these experiments, we can find that the proposed adversarial training is effective on different detectors trained on the SAR-Ship-Dataset.
What is more, the SAR-Ship-Dataset has a big data volume, and there are a lot of hard samples. This is especially suitable for the proposed adversarial training-based data augmentation methods. This is because it can generate a lot of perturbing samples, and thus make the detectors robust to the hard samples in the SAR-Ship-Dataset.

3.2.3. Performance on AIR-SARShip

AIR-SARShip has 21 whole SAR images in total. Resolutions of them are 1 m and 3 m, and the image size of them is 3000 × 3000 pixels. The samples of AIR-SARShip are shown in Figure 4.
The results of different data augmentation methods and detectors on AIR-SARShip are shown in Table 4.
For the traditional data augmentation methods, 32 experiments are conducted. Compared to the build-in methods, 14 of them were increased, with an average increase of 5.7%, and 18 of them were decreased, with an average decrease of 8.4%. The average of them is 0.72. The average of the build-in data augmentation method in Table 1 is 0.74. This shows that the traditional data augmentation methods would decrease the performance of the detectors overall. Only the specific combination of the detector and data augmentation methods can boost the performance. For example, YOLOV3 with random erasing can boost the performance from 46% to 84% (with 38% improvement) on AIR-SARShip.
Compared to the plain ability of the traditional data augmentation method, the proposed adversarial training-based method can receive a huge improvement with all of the detectors. For example, it receives 0.95, 0.93, 0.73, and 0.93 AP50 on AIR-SARShip, respectively. The average value is 0.89. It receives an improvement of 0.15 and 0.17, compared to the build-in and traditional methods.
The AIR-SARShip dataset is designed to detect ships in large scene. The ships in large-width SAR images are sparse. There are a large number of samples in the dataset that are difficult to detect. The detector has a large performance improvement space compared to other datasets. The adversarial training is particularly suitable for this situation.

3.3. Ablation Experiments

From the above experiments, we can find that the traditional data augmentation methods cannot always promote the detection accuracy, and, in some cases, it would even decrease the performance. However, the proposed method in this paper can promote the accuracy with all of the detectors and datasets. In order to investigate the principle of the proposed method further, we conduct the following experiments. The results are shown in Table 5.
On SSDD, the plain adversarial training method can improve the AP50 from 0.85 to 0.88. The separate batch normalization, maxing, and optimization can also improve it to 0.89, 0.91, and 0.93. For the SAR-Ship-Dataset, the plain adversarial training method can improve the AP50 from 0.83 to 0.87. The separate BN, maxing, and optimization can also improve it to 0.89, 0.90, and 0.93. On AIR-SARShip, the plain adversarial training method can improve the AP50 from 0.72 to 0.81. The separate BN, maxing, and optimization can also improve it to 0.84, 0.86, and 0.89.
We can find that the plain adversarial training can receive a big improvement, especially on the AIR-SARShip-Dataset, as it includes a lot of hard samples. The separate BN can receive an improvement of 0.02 on average. This is due to the fact that the separate batch normalizations of clean samples and disturbing samples can avoid the interference of the clean samples. Thus, the overall performance is improved. The maxing can receive an improvement of 0.02 on average. This can make the model adapt to more confrontational samples. The proposed optimization method can make the model converge better and faster, which can also receive an improvement of 0.027 on average.

3.4. The Detection Results

The detected results of the proposed method on the SSDD, SAR-Ship-Dataset, and AIR-SARShip are displayed in Figure 5. We found that it can cope with both the easy and hard samples, which reveals that the adversarial training is efficient.

4. Conclusions

This paper proposes an adversarial learning-based data augmentation method to address the issues of insufficient data diversity and limited performance improvement of traditional methods. The proposed method generates adversarial images by appending small perturbations to the image, and achieves the generation of interference images and optimization of model parameters by modeling and solving the max–min problem. During the model solving process, clean samples and interfering images were normalized separately, and the larger interference of the localization and classification subtasks was selected. A multi-step average disturbance and single-step gradient update were used to increase the efficiency and accuracy of the adversarial learning. Experiments on three datasets and four detectors show that it has significant performance advantages compared to traditional methods.

Author Contributions

J.L. and Z.Y. conceived and designed the experiments; J.C. performed the experiments; H.J. analyzed the data; Z.Y. contributed reagents/materials/analysis tools; J.L. wrote the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Dataset available on request from the authors.

Acknowledgments

The authors would like to thank the anonymous reviewers for the constructive suggestions.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Li, H.C.; Hong, W.; Wu, Y.R.; Fan, P.Z. An efficient and flexible statistical model based on generalized Gamma distribution for amplitude SAR images. IEEE Trans. Geosci. Remote Sens. 2010, 48, 2711–2722. [Google Scholar]
  2. Achim, A.; Kuruoglu, E.E.; Zerubia, J. SAR image filtering based on the heavy-tailed Rayleigh model. IEEE Trans. Image Process. 2006, 15, 2686–2693. [Google Scholar] [CrossRef] [PubMed]
  3. Li, J.; Qu, C.; Shao, J. Ship detection in SAR images based on an improved faster R-CNN. In Proceedings of the SAR in Big Data Era: Models, Methods & Applications, Beijing, China, 13–14 November 2017. [Google Scholar]
  4. Li, J.; Xu, C.; Su, H.; Gao, L.; Wang, T. Deep Learning for SAR Ship Detection: Past, Present and Future. Remote Sens. 2022, 14, 2712. [Google Scholar] [CrossRef]
  5. Yang, S.; Xiao, W.; Zhang, M.; Guo, S.; Zhao, J.; Shen, F. Image Data Augmentation for Deep Learning: A Survey. arXiv 2022, arXiv:2204.08610. [Google Scholar] [CrossRef]
  6. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the IEEE Symposium on Security and Privacy, San Jose, CA, USA, 22–26 May 2017. [Google Scholar]
  7. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar] [CrossRef]
  8. Xie, Q.; Luong, M.-T.; Hovy, E.; Le, Q.V. Self-training with noisy student improves imagenet classification. In Proceedings of the Computer Vision and Pattern Recognition, Virtual, 14–19 June 2020. [Google Scholar]
  9. Goodfellow, I.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2015. [Google Scholar] [CrossRef]
  10. Yin, D.; Gontijo Lopes, R.; Shlens, J.; Cubuk, E.D.; Gilmer, J. A fourier perspective on model robustness in computer vision. arXiv 2019, arXiv:1906.08988. [Google Scholar]
  11. Zhang, T.; Zhu, Z. Interpreting adversarially trained convolutional neural networks. Int. Conf. Mach. Learn. PMLR 2019, 97, 7502–7511. [Google Scholar]
  12. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial machine learning at scale. arXiv 2016, arXiv:1611.01236. [Google Scholar]
  13. Zhang, H.; Wang, J. Towards adversarially robust object detection. In Proceedings of the International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019. [Google Scholar]
  14. Schwinn, L.; Raab, R.; Nguyen, A.; Zanca, D.; Eskofier, B. Exploring misclassifications of robust neural networks to enhance adversarial attacks. Appl. Intell. 2023, 53, 19843–19859. [Google Scholar] [CrossRef]
  15. Shafahi, A.; Najibi, M.; Ghiasi, M.A.; Xu, Z.; Dickerson, J.; Studer, C.; Davis, L.S.; Taylor, G.; Goldstein, T. Adversarial training for free! arXiv 2019, arXiv:1904.12843. [Google Scholar]
  16. Zhang, D.; Zhang, T.; Lu, Y.; Zhu, Z.; Dong, B. You only propagate once: Accelerating adversarial training via maximal principle. arXiv 2019, arXiv:1905.00877. [Google Scholar]
  17. Zhu, C.; Cheng, Y.; Gan, Z.; Sun, S.; Goldstein, T.; Liu, J. Freelb: Enhanced adversarial training for natural language understanding. arXiv 2019, arXiv:1909.11764. [Google Scholar]
  18. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster R-CNN: Towards real-time object detection with region proposal networks. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 1137–1149. [Google Scholar] [CrossRef] [PubMed]
  19. Redmon, J.; Farhadi, A. YOLOv3: An incremental improvement. arXiv 2018, arXiv:1804.02767. [Google Scholar]
  20. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal Loss for Dense Object Detection. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 99, 2999–3007. [Google Scholar]
  21. Duan, K.; Bai, S.; Xie, L.; Qi, H.; Huang, Q.; Tian, Q. Centernet: Keypoint triplets for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6569–6578. [Google Scholar]
  22. Wang, Y.; Wang, C.; Zhang, H.; Dong, Y.; Wei, S. A SAR Dataset of Ship Detection for Deep Learning under Complex Backgrounds. Remote Sens. 2019, 11, 765. [Google Scholar] [CrossRef]
  23. Sun, X.; Wang, Z.; Sun, Y.; Diao, W.; Zhang, Y.; Fu, K. AIR-SARShip-1.0: High-resolution SAR ship detection dataset. J. Radars 2019, 8, 852–862. [Google Scholar]
  24. Shorten, C.; Khoshgoftaar, T.M. A survey on image data augmentation for deep learning. J. Big Data 2019, 6, 1–48. [Google Scholar] [CrossRef]
  25. DeVries, T.; Taylor, G.W. Improved regularization of convolutional neural networks with cutout. arXiv 2017, arXiv:1708.04552. [Google Scholar]
  26. Zhong, Z.; Zheng, L.; Kang, G.; Li, S.; Yang, Y. Random erasing data augmentation. Proc. AAAI Conf. Artif. Intell. 2020, 34, 13001–13008. [Google Scholar] [CrossRef]
  27. Zhang, H.; Cisse, M.; Dauphin, Y.N.; Lopez-Paz, D. mixup: Beyond empirical risk minimization. arXiv 2017, arXiv:1710.09412. [Google Scholar]
  28. Yun, S.; Han, D.; Oh, S.J.; Chun, S.; Choe, J.; Yoo, Y. Cutmix: Regularization strategy to train strong classifiers with localizable features. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October–2 November 2019; pp. 6023–6032. [Google Scholar]
  29. Hendrycks, D.; Mu, N.; Cubuk, E.D.; Zoph, B.; Gilmer, J.; Lakshminarayanan, B. Augmix: A simple data processing method to improve robustness and uncertainty. arXiv 2019, arXiv:1912.02781. [Google Scholar]
  30. Chen, P.; Liu, S.; Zhao, H.; Jia, J. Gridmask data augmentation. arXiv 2020, arXiv:2001.04086. [Google Scholar]
  31. Cubuk, E.D.; Zoph, B.; Mane, D.; Vasudevan, V.; Le, Q.V. Autoaugment: Learning augmentation strategies from data. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Long Beach, CA, USA, 16–20 June 2019; pp. 113–123. [Google Scholar]
  32. Li, P.; Li, X.; Long, X. Fencemask: A data augmentation approach for pre-extracted image features. arXiv 2020, arXiv:2006.07877. [Google Scholar]
Figure 1. The overall architecture of the proposed method.
Figure 1. The overall architecture of the proposed method.
Sensors 24 04154 g001
Figure 2. Samples of SSDD.
Figure 2. Samples of SSDD.
Sensors 24 04154 g002
Figure 3. Samples of SAR-Ship-Dataset.
Figure 3. Samples of SAR-Ship-Dataset.
Sensors 24 04154 g003
Figure 4. Samples of AIR-SARShip.
Figure 4. Samples of AIR-SARShip.
Sensors 24 04154 g004aSensors 24 04154 g004b
Figure 5. The detection results on the datasets.
Figure 5. The detection results on the datasets.
Sensors 24 04154 g005aSensors 24 04154 g005b
Table 1. The baselines of the build-in data augmentation methods.
Table 1. The baselines of the build-in data augmentation methods.
DatasetsFRRNY3CNAVE
SSDD0.840.900.870.940.89
SAR-Ship-Dataset0.860.890.790.860.85
AIR-SARShip0.820.860.460.820.74
Table 2. The results of different data augmentation methods and detectors on SSDD.
Table 2. The results of different data augmentation methods and detectors on SSDD.
DAMFRRNY3CNAVE1AVE2
Cutout0.910.930.750.880.870.85
Random erasing0.910.910.780.860.87
Mixup0.910.970.820.920.91
Cutmix0.730.770.710.720.73
Augmix0.920.940.850.910.91
Gridmask0.900.900.840.850.87
Mosic0.880.890.800.870.86
Copy/paste0.820.910.650.880.82
Proposed 0.940.950.890.950.930.93
Table 3. The results of different data augmentation methods and detectors on SAR-Ship-Dataset.
Table 3. The results of different data augmentation methods and detectors on SAR-Ship-Dataset.
DAMFRRNY3CNAVE1AVE2
Cutout0.870.910.810.890.870.83
Random erasing0.860.880.840.830.85
Mixup0.890.930.800.900.88
Cutmix0.710.920.520.600.69
Augmix0.900.910.650.880.84
Gridmask0.850.890.780.840.84
Mosic0.880.860.840.860.86
Copy/paste0.860.900.640.890.82
Proposed0.930.940.890.940.930.93
Table 4. The results of different data augmentation methods and detectors on AIR-SARShip.
Table 4. The results of different data augmentation methods and detectors on AIR-SARShip.
DAMFRRNY3CNAVE1AVE2
Cutout0.910.880.400.810.750.72
Random erasing0.910.850.420.840.76
Mixup0.900.810.380.860.74
Cutmix0.620.510.300.830.57
Augmix0.900.860.410.890.77
Gridmask0.810.760.20.840.65
Mosic0.800.840.650.850.79
Copy/paste0.810.780.480.810.72
Proposed 0.950.930.730.930.890.89
Table 5. The ablation experiment results.
Table 5. The ablation experiment results.
MethodsSSDDSAR-Ship-DatasetAIR-SARShip
Traditional methods0.850.830.72
Plain adversarial training0.880.870.81
+Separate BN0.890.890.84
+Maxing0.910.900.86
+Optimization0.930.930.89
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Li, J.; Yu, Z.; Chen, J.; Jiang, H. A SAR Ship Detection Method Based on Adversarial Training. Sensors 2024, 24, 4154. https://doi.org/10.3390/s24134154

AMA Style

Li J, Yu Z, Chen J, Jiang H. A SAR Ship Detection Method Based on Adversarial Training. Sensors. 2024; 24(13):4154. https://doi.org/10.3390/s24134154

Chicago/Turabian Style

Li, Jianwei, Zhentao Yu, Jie Chen, and Hao Jiang. 2024. "A SAR Ship Detection Method Based on Adversarial Training" Sensors 24, no. 13: 4154. https://doi.org/10.3390/s24134154

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop