Next Article in Journal
The Efficiency of Foliar Kaolin Spray Assessed through UAV-Based Thermal Infrared Imagery
Next Article in Special Issue
Auto-Weighted Structured Graph-Based Regression Method for Heterogeneous Change Detection
Previous Article in Journal
Temperature Contributes More than Precipitation to Runoff in the High Mountains of Northwest China
Previous Article in Special Issue
Attention-Based Multi-Level Feature Fusion for Object Detection in Remote Sensing Images
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Empirical Study of Fully Black-Box and Universal Adversarial Attack for SAR Target Recognition

1
School of Electronic Science, National University of Defense Technology (NUDT), Changsha 410073, China
2
College of System Engineering, National University of Defense Technology (NUDT), Changsha 410073, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(16), 4017; https://doi.org/10.3390/rs14164017
Submission received: 15 July 2022 / Revised: 8 August 2022 / Accepted: 12 August 2022 / Published: 18 August 2022

Abstract

:
It has been demonstrated that deep neural network (DNN)-based synthetic aperture radar (SAR) automatic target recognition (ATR) techniques are extremely susceptible to adversarial intrusions, that is, malicious SAR images including deliberately generated perturbations that are imperceptible to the human eye but can deflect DNN inference. Attack algorithms in previous studies are based on direct access to a ATR model such as gradients or training data to generate adversarial examples for a target SAR image, which is against the non-cooperative nature of ATR applications. In this article, we establish a fully black-box universal attack (FBUA) framework to craft one single universal adversarial perturbation (UAP) against a wide range of DNN architectures as well as a large fraction of target images. It is of both high practical relevance for an attacker and a risk for ATR systems that the UAP can be designed by an FBUA in advance and without any access to the victim DNN. The proposed FBUA can be decomposed to three main phases: (1) SAR images simulation, (2) substitute model training, and (3) UAP generation. Comprehensive evaluations on the MSTAR and SARSIM datasets demonstrate the efficacy of the FBUA, i.e., can achieve an average fooling ratio of 64.6% on eight cutting-edge DNNs (when the magnitude of the UAP is set to 16/255). Furthermore, we empirically find that the black-box UAP mainly functions by activating spurious features which can effectively couple with clean features to force the ATR models to concentrate on several categories and exhibit a class-wise vulnerability. The proposed FBUA aligns with the non-cooperative nature and reveals the access-free adversarial vulnerability of DNN-based SAR ATR techniques, providing a foundation for future defense against black-box threats.

1. Introduction

Synthetic aperture radar (SAR) can actively emit microwaves and utilize the notion of synthetic aperture to improve azimuth resolution and eventually obtain high-resolution radar images of ground targets. Due to its all-weather as well as day-and-night working capabilities, SAR has various applications, including resource mapping, military surveillance, post-disaster assessment, and environment monitoring, etc. [1]. With the eye-catching advances of deep neural networks (DNNs) over the past decade [2,3], they have been widely implemented into remote sensing image processing tasks. Although the applications of DNNs are largely limited to optical imagery interpretation, they have also been introduced in SAR imagery interpretations such as automatic target recognition (ATR) [4,5,6] and terrain classification [7], attaining spectacular progress and gaining increasing interest [8].
SAR ATR is of great importance for homeland security and military applications. Therefore, SAR ATR techniques are required to be both highly accurate yet trustworthy. However, it has been demonstrated that DNN-based SAR ATR techniques are highly susceptible to adversarial intrusions in the form of clean inputs added with malicious perturbations [9,10]. These perturbations are intentionally designed to be invisible to human eyes but can dramatically manipulate DNNs’ prediction, which is what makes this adversarial vulnerability of DNNs so significant and striking. The lack of robustness of DNNs was initially reported in the field of optical image processing [11] back into 2014. Throughout the past decade, many attempts have been made to find mathematical, empirical, or geometrical explanations. Szegedy et al. noted that adversarial examples are located in low-probability pockets where regular data is unlikely to be sampled [11]. This intriguing vulnerability, according to Goodfellow et al., may be induced by the local linearity of DNNs [12]. Tanay et al. proposed that the non-robust behavior occurs when a tilting decision boundary in high-dimensional space permits a small vertical step to lead a data point past the boundary [13]. Fawzi et al. advocated utilizing adversarial perturbations to investigate the geometric properties of the DNNs’ high-dimensional feature space, having generated a large number of discoveries regarding high-dimensional classification [14]. Although the adversarial vulnerability has not been tackled yet, several studies focusing on finding, defending, and understanding adversarial examples have provided insights into robust and trustworthy DNNs [15].
It is important to find novel (or potential) attacks to further facilitate the defense and robustness in a corresponding threat scenario. In optical image processing, adversarial attacks can be classified as white-box or black-box, depending on whether the attacker has direct access to the victim model. White-box attacks allow attackers to gain full access to detailed information about the target model, such as gradient, training data, architecture, parameters, and so on [16]. Straightforward examples of the white-box attacks are the first-order methods such as the fast gradient sign method (FGSM) [12], which is basically a classification loss magnifier. In the more practical black-box scenario, attackers a have limited (query-based framework [17,18,19]) or no (transfer-based framework [20,21,22,23]) access to the victim model. Adversarial attack frameworks can also be decomposed to image-dependent and image-agnostic, depending on whether the generated perturbation is just sensitive to a specific target image or a set of images. The image-agnostic adversarial perturbation, i.e., universal adversarial perturbation (UAP) was first proposed by Moosavi–Dezfooli et al. [24], which iteratively updates the perturbation for fooling most of data points in the input space. Afterward, several attempts were made to utilize generative models for crafting UAPs [25,26]. The UAPs can also be crafted by manipulating the intermediate features [27,28,29,30]. Compared to image-dependent perturbation, the UAP is highly feasible because it can be crafted and deployed in advance. Both the black-box and universal adversarial attacks are arguably the most practical yet challenging attack scenarios and also the best surrogates to study the non-cooperative threats of DNN-based SAR ATR systems.
Recently, the adversarial vulnerability of DNNs has received increasing attention in the SAR ATR community. Huang et al. revealed that the DNN-based SAR ATR techniques are also vulnerable to adversarial examples [31]. Following studies made comprehensive examinations of various DNN structures and different databases, providing several conclusions. For instance, Chen et al. demonstrated that the data-dependent adversarial examples generalize well across different DNN models [9]. Li et al. concluded that network structure has a significant impact on robustness [10]. Specifically, simpler network structures result in more robustness against adversarial attacks. Further, Du et al. introduced generative models to accelerate the attack process and to refine the scattering features [32,33]. Most recently, Liu et al. illustrated the physical feasibility of the digital adversarial examples by exploiting the phase modulation jamming metasurface [34,35,36]. Meanwhile, Peng et al. proposed that regional restricted design also potential to be implemented in real-world [37]. It is worth noting that all the current studies comply with a default setting with which the attacker can access the same training data of the victim model. Such a setting is undoubtedly against the intrinsic nature of SAR ATR being extremely non-cooperative and leaves the question whether the SAR ATR needs to worry about adversarial examples.
By contrast, this article explores the adversarial vulnerability of DNN-based SAR ATR in the fully black-box and non-cooperative operational circumstance. With the specific intention of shedding light on the threat in the practical scenario, we take into consideration an appropriate yet also the most challenging setting in which the SAR ATR systems would deny any of the queries, requests or access to the model and training data. Thus, the attacker may just have very limited intelligence or public information about the SAR system such as the radar frequency, imaging modes or resolution, etc. Meanwhile, we purse a universal attack capability that a fixed and pre-calculated UAP can bias most of the target models and target images. Figure 1 depicts the proposed fully black-box universal attack (FBUA) framework, which can be decomposed into three stages. Firstly, we need to obtain a simulated SAR dataset that carries enough information and knowledge to train a substitute model. The generation process in Figure 1 refers to the generation process of [38] and simulates the real imaging results with CAD models and electromagnetic calculation. Secondly, we train the substitute model with the synthesized dataset. Finally, the existing universal attack algorithms are utilized to craft the UAP and attack the black-box ATR models.
Utilizing the proposed FBUA framework, we empirically investigated the non-cooperative adversarial threats of a DNN-based SAR ATR with eight cutting-edge DNN architectures and five state-of-the-art (SOTA) UAP generation algorithms based on the publically accessible MSTAR and SARSIM datasets. The main contributions and experimental results of this study are summarized as follows.
1.
We propose a novel FBUA framework that can effectively attack a DNN-based SAR ATR without any access to the model’s gradient, parameters, architecture, or even training data. In addition, the generated single adversarial perturbation is universal to attack a variety of models and a large fraction of target images. Comprehensive experimental evaluations on the MSTAR and SARSIM datasets prove the efficacy of our proposal.
2.
This article conducts a comprehensive evaluation to reveal the black-box adversarial threats of ATR systems using the proposed FBUA method. We find that popular DNN-based SAR ATR models are vulnerable to UAPs generated by an FBUA. Specifically, with a relatively small magnitude (16 out of 255), a single UAP is capable of achieving a fooling ratio of 64.6% averaged on eight DNN models.
3.
We find that the vulnerability of the target DNN model exhibits a high degree of class-wise variability; that is, data points within a class share the similar robustness to the UAP generated by an FBUA.
4.
We empirically demonstrate that the UAPs created by an FBUA primarily function by activating spurious features, which are then coupled with clean features to form robust features that support several dominant labels. Therefore, DNNs demonstrate class-wise vulnerability to UAPs; that is, classes that do not conform to dominant labels are easily induced, whilst other classes show robustness to UAPs.

2. Preliminaries and Experimental Settings

In this section, we provide details about the problem description, the studied UAP generation algorithms, datasets, DNN models, and implementation details.

2.1. Problem Description of Universal Adversarial Attack to SAR ATR

Let f θ : X Y represent a SAR ATR classifier with a set of trained parameters (weights) θ . General adversarial attacks aim at finding a perturbation δ for each SAR target image x X and its corresponding ground truth label y g t Y that bias the DNN inference: arg max f θ ( x ) = y gt arg max f θ ( x + δ ) y gt (untargeted attack) or arg max f θ ( x + δ ) = y target (targeted attack). At this time, we call the perturbed input x + δ as adversarial example, which is denoted by x adv . The goal of untargeted attacks is to interfere with the correct predictions, while the targeted attacks are calculated to force the predictions to concentrate on a specific target class y target . Unless otherwise specified, the UAP in this article is by default untargeted. An adversarial example is commonly restricted by a distance measurement function D ( · ) to satisfy that of being imperceptible to the human visual system. In this framework, different attack methods can be summarized as diverse solutions for the following optimization problem [24]:
arg max f θ ( x + δ ) y gt s . t . D ( x , x + δ ) ϵ ,
where ϵ cooperates with D ( · ) serve as a distance constraint. We follow most studied cases that adopt the l p -norm with radius ϵ as the constraint. Then, in our context, i.e., the universal attack scenario, the goal is to seek a single UAP such that
arg max f θ ( x + δ ) y gt for most x X s . t . δ p ϵ .

2.2. UAP Generation Algorithms

Various UAP generation algorithms were proposed to provide solutions to Equation (2) from different perspectives. In this section, five SOTA UAP generation algorithms will be studied in the experiments introduced.

2.2.1. DeepFool-UAP

DeepFool-UAP [24] runs over the available training images for several iterations until achieving a desired fooling ratio on the test images. Specifically, the minimum data-dependent perturbation for the i-th data point in the m-th iteration will be accumulated to the total perturbation
δ i , m + 1 = P p , ϵ ( δ i , m + DeepFool ( f θ , x i + δ i , m , y i gt ) ) ,
The whole process stops when the specific fooling ratio is achieved. The perturbation calculation operator DeepFool ( · ) [40] is an attack algorithm that generates the perturbation with minimum Euclidean distance away from the original data to fool the classifier. Across the DeepFool-UAP iterations, the accumulated UAP will be projected back to the l p -norm ball of radius ϵ when δ p > ϵ , where the projection function is defined as [24]
P p , ϵ ( δ ) = arg min δ δ δ 2 s . t . δ p ϵ .

2.2.2. Dominant Feature-UAP (DF-UAP)

DF-UAP [41] applies the general DNN training framework to optimize the UAP. Suppose a perturbation is zero-initialized: δ 0 , and then it will be progressively modified to maximize the classification loss L of each m mini-batch data points:
δ = arg min δ E x , y gt B L ( f θ , x + δ , y gt ) , where B X and B = m .
Optimization in batch training is essential to obtain the universal perturbation, and δ can be conveniently optimized by common out-of-the-box optimizers, such as Adam [42] and stochastic gradient descent (SGD) [43], when training DNNs. Consistent with the DeepFool-UAP, the perturbation also should be projected back into the given norm ball. Following the original paper, we utilize the cross-entropy loss function L C E as the classification loss item L in Equation (5).

2.2.3. CosSim-UAP

Zhang et al. propose to maximize the cosine similarity (CosSim) between the model output of clean input f θ ( x ) and adversarial examples f θ ( x + δ ) [44]. The objective is simply formulated as
L C S = C o s S i m f θ ( x ) , f θ ( x + δ ) .
The solution is similar to Equation (5). With the network parameters frozen, the perturbation is updated by batch gradient descent.

2.2.4. Fast Feature Fool (FFF) and Generalizable Data-Free UAP (GD-UAP)

Mopuri et al. propose to generate a UAP in data-free condition [27], i.e., the attacker has no access to training data to craft any data-dependent sub-perturbation. They propose an efficient objective to maximize spurious activations at each of the convolution layers. The objective is formulated as
L F F F = log i = 1 K l ¯ i ( δ ) ,
where l ¯ i ( δ ) is the mean value of the output at layer i after non-linear activation such as error. The L F F F aim at maximizing the total activation of f θ that ignited by δ . In their later published work GD-UAP [29], the loss is changed to calculate a Euclidean norm of each activation:
L G D = log i = 1 K l i ( δ 2 ) 2 .
FFF and GD-UAP intentionally cause over-fitting of the neurons to deteriorate the features that derived from normal inputs. The mathematical significance of L F F F and L G D is comparable. In practice, we find that the performance of the later objective is marginally superior to that of the former; hence the remainder of this work focuses primarily on the L G D .

2.3. Database

Recall the condition that the attacker does not have any data; thus, the substitute models are expected to be trained by the images that are similar enough to the target SAR system. In this article, we focus on the most studied MSTAR recognition task. Taking the public setting of the MSTAR measured data as the only information we could know, e.g., the frequency, resolution, spotlight imaging mode, and so on. We propose to generate spurious data to obtain the informative and cognitive substitute models. To encourage reproducible research, a publicly accessible simulated MSTAR-like SAR dataset, namely SARSIM, was selected to train the substitute models [38].

2.3.1. MSTAR Dataset

MSTAR dataset for the public research of SAR ATR was made available by the moving and stationary target acquisition and recognition (MSTAR) program, which was funded by the defense advanced research projects agency (DARPA) and the Air Force Research Laboratory (AFRL). The Sandia National Laboratory SAR sensor platform with X-Band imaging capability and 1-foot resolution was utilized to gather the SAR data. The resulting measured data consists of 128 × 128 pixel size images with 360° articulation, 1° spacing, and several depression angles. A total of 10 different types of ground vehicle targets are included in the MSTAR dataset. We follow the standard operation condition (SOC) that use the 17° data to train and 15° data to test the models. More details of the SOC subset are specified in Figure 2 and Table 1.

2.3.2. SARSIM Dataset

SARSIM data were calculated specifically with the public information about the MSTAR dataset [38,39]. Initially, the far-field radar cross section (RCS) in corresponding frequency, azimuth, and depression angle was calculated utilizing the CST Microwave Studio Asymptotic Solver. Then, three types of terrain clutter (grass, roads, and the mix of them) and thermal noise were statistically provided. Meanwhile, the electromagnetic shadowing was well-estimated by ray projection from the sensor position to each scatterer. The focusing of the simulated radar data was performed by a time-domain back-projection algorithm. The SARSIM dataset consists of seven types of ground targets, and two sub-objects are included in each of the seven types of targets. The whole dataset contains seven depression angles (15°, 17°, 25°, 30°, 40°, and 45°), and the granularity in azimuth angle is 5°. It should be noted that the SARSIM dataset has shown to be effective in providing valid knowledge about the MSTAR targets in recent works [5,39].
In the experiments, we selected 10 objects and followed the MSTAR SOC setting that 17° images are for training. All the three terrain clutters were included in our setting. Figure 3 and Table 2 display the specifics of the SARSIM setting in our experiment.

2.4. Implementation Details

2.4.1. Environment

The Python programming language (v3.6) and Pytorch deep learning framework (v1.10.1) were used to implement all the software parts of the evaluations, including the data processing, DNN, and attack implementations. All the evaluations were supported by a NVIDIA DGX-1 server which is equipped with eight Tesla-V100 GPUs and powered by a dual 20-core Intel Xeon E5-2698 v4 CPU.

2.4.2. DNN Models and Training Details

The evaluations included eight DNN models: AlexNet [2], VGG11 [45], ResNet50 [46], DenseNet121 [47], MobileNetV2 [48], AConvNet [4], ShuffleNetV2 [49], and SqueezeNet [50]. The first four are typical structures that serve as backbone in various deep learning applications, and the last four are lightweight designs. Note that the AConvNet was especially proposed to be applied for the SAR ATR. The well-known procedure in [4] was followed to train these models. Specifically, the images were randomly cropped to 88 × 88 for training and center-cropped for testing. The single-channel SAR images were treated as gray-scale images and normalized to [0, 1] for faster convergence. With an SGD optimizer [43] and multi-step learning rate adjustment strategy, the models were, respectively, trained on the two studied datasets. Details of the models can be found in Table 3. It is shown that with far fewer parameters and computational resource requirements, the lightweight models can achieve competitive performance on both the studied datasets, which are of value for the on-board or edge-device ATR scenarios such as the drone-borne SAR.

2.4.3. Implementation Details

All the aforementioned UAP generation algorithms were re-implemented using Pytorch according to the original papers. Each of the images in the SARSIM training set are available for DeepFool-UAP, DF-UAP, and CosSim-UAP, and the perturbation was generated for five epochs of the whole training set. For DF-UAP, CosSim-UAP, and GD-UAP, the Adam [42] algorithm was provided to optimize the perturbation with respect to their loss functions. The learning rate was set to 0.1, and maximum iterations were set to 300 (GD-UAP); other parameters were set by default. p = and ϵ = 16 / 255 were selected as constraint unless otherwise stated, e.g., the pixels’ absolute value of perturbation was not allowed to exceed 16 / 255 (for x [ 0 , 1 ] ). The l -norm is defined as
δ = max 0 i < h 0 j < w δ i j ,
where i , j are the pixel index of the perturbation image (matrix) with a pixel size of h × w . When the UAP transfers across models that have different input size requirements, it will be scaled by bilinear interpolation to fit the target size.

2.4.4. Metric

In light of the aforementioned definition of UAPs, the fooling ratio is the most frequently used metric for assessing the effectiveness of UAPs. Specifically, the fooling ratio is defined as the percentage of samples whose prediction changes after the UAP is applied:
Fooling ratio = i N test I arg max f θ ( x i + δ ) y i gt N test .
Herein, I ( · ) denotes the indicator function, and N test is the capacity of the test set to evaluate the attack performance. In the experiments, this test set is a uniformly extracted subset from the MSTAR test set and N test = 1000 . These 1000 images can be correctly classified by all the eight MSTAR models and contain 100 images for each of the classes.

3. Results

3.1. Quantitative Results

We first investigate the overall attack performance of the studied UAP generation algorithms. The fooling ratios under 256 attack scenarios, that is, diverse UAP generation algorithms and substitute model architectures versus different black-box target models, are available in Figure 4. Each element of these matrices indicates the fooling ratio against the target MSTAR model (column index) achieved by the UAP (caption) generated based on the substitute SARSIM model (row index). We observe that the DeepFool-UAP algorithm is unable to carry out a successful cross-dataset universal attack, while the other algorithms exhibit considerable attack capability. The DF-UAP, CosSim-UAP, and GD-UAP, respectively, achieve 58.4%, 62.4%, and 49.8% overall average fooling ratios. The MobileNetV2 is shown to be the most effective substitute model for DF-UAP and CosSim-UAP, and VGG11 is for GD-UAP, respectively, achieving 62.4%, 64.6%, and 61.3% fooling ratios. Therefore, these three attack settings are further investigated in the following experiments. To conclude, the CosSim-UAP algorithm exhibits the most robust and effective attack performance against a variety of target models with different substitute models. In the subsequent sections of this article, we will report the results of the CosSim-UAP algorithm. The results of DF-UAP and GD-UAP will be listed in the Appendix A.
At the same time, the vulnerability of different target models is less related to the choice of substitute model. Each column shows an approximate fooling ratio, indicating that the remained accuracy of the target model is mostly influenced by itself and the attack algorithm other than the substitute model. Moreover, several cases reveal that the UAP performs badly to specific target models, such as the DF-UAP and GD-UAP achieving worse fooling ratios against AConvNet and SqueezeNet. Among all the target models, AConvNet exhibits the best resistance to UAPs in the face of DF-UAP, CosSim-UAP, and GD-UAP attacks, which may be due to the rescaling of UAPs losing their aggressiveness.

3.2. Qualitative Results

The UAPs generated in all the aforementioned attack scenarios are visualized in Figure 5. Since the magnitude of the UAPs is relatively small, we amplify them for better observation. The generated perturbation patterns are observed to link to the generation algorithm and model architecture, which, however, is not interpretable by a human observer. Nonetheless, we report the perturbed images along with the UAP that constrained by various magnitude ϵ in Figure 6. It can be observed that the perturbed images are not easy to discriminate with naked eyes and without original images as reference.

3.3. Comparison to Random Noise

We further compare the UAP generated by an FBUA with random noise to study its efficacy. The UAPs, uniform noise, and Gaussian noise are generated with the same l -norm constraints. Please notice that we set the deviation of Gaussian noise as ϵ / 3 to make most of the pixel modifications lie in the range section of [ ϵ , ϵ ] . Table 4, Table A1, and Table A2 detail the attack performances of all the studied noise from which the following observations can be summarized. Firstly, with larger ϵ , the UAPs exhibit more powerful universal aggressivity. It achieves an average fooling ratio of 79.3 % with ϵ = 32 / 255 . Secondly, some models are sensitive to random noise, such as the DenseNet121, MobileNetV2, and ShuffleNetV2. However, compared to the universal nature of UAPs, the random noise exhibits severe particularity and failed to attack all the target models. The results of this comparison are aligned with the aforementioned efficacy of the UAPs.

3.4. Robustness of the UAP

Well-designed perturbations are challenged by many factors in the SAR imaging thread, such as the velocity shake of the airborne platform, atmospheric factors of the space-borne platform, as well as the speckle noise, among many others. Therefore, we conduct several perturbations to test the UAPs’ robustness, including additive white noise, multiplicative exponential noise, Gaussian filtering, and median filtering. Details and results are elaborated in Table 5, Table A3, and Table A4. The most destructive perturbation to UAPs is the multiplicative noise which generally models the speckle noise in SAR imagery. There are cases in which extra perturbation enhances the fooling ratio for particular models, such as median filtering for AConvNet and SqueezeNet. Overall, the UAPs exhibit impressive robustness against these perturbations when ϵ 8 / 255 , and when ϵ become larger, the perturbations would have less influence on the attack performance.

3.5. Analysis of the Class-Wise Vulnerability

In this section, we focus on investigating the vulnerability gap between different target models. Table 6, Table A5 and Table A6 elaborate the detailed fooling ratios, showing that the difference in fooling ratio between models is mainly determined by whether certain classes are effectively attacked. For example, the total difference between AlexNet and VGG 11 is 14.4%, mainly composed of the difference 8.8% for the ZSU234. Figure 7, Figure A1 and Figure A2 exhibit more specifics of the class transfer. In these figures, each element expresses the number of samples transferred to the adversarial category from the original category after being added with the UAP. Take the first matrix of Figure 7 as an example. The first row (index 9) indicates that there were 88 (out of 100) ZSU234 samples misclassified as 4 BMP2, 62 T72, and 22 ZIL131 after being attacked. It is shown that the UAP transfers most images to several dominant labels represented by bright cells in the vertical direction. There is shown a class-wise vulnerability across all the studied models that images of a single class are similarly vulnerable or robust to the UAP. The main properties of this class-wise vulnerability are summarized as follows.
  • The dominant labels where the misclassified data concentrate are meanwhile the robust classes where the clean data are not easy to be successfully attacked. Additionally, data within the robust classes also tend to transfer to one of the dominant labels, such as for DenseNet121, the BMP2 data is easy to be induced to T62.
  • The dominant labels of diverse models share a certain similarity, although they are caused by a single substitute model. For instance, BMP2 is shown as a dominant label for all eight models and T72 for seven models (except DenseNet121). It demonstrates a universal representation of all the target models
  • The class-wise vulnerability to UAPs is more universal than the attack selectivity reported in previous image-dependent attack studies [9,10]. When facing the data-dependent adversarial examples, dominant labels are highly class-dependent. For example, most of the adversarial examples of D7 would be recognized as ZIL131, ZSU234, and T62, while examples of other classes are misclassified to different labels. It also behaves differently from the universal attack for optical image classifiers which trained on 1000 categories of images that the adversarial images would classify as a single dominant class [24,44].
Based on the aforementioned facts, we hypothesize that the UAP gives rise to robust features when flowing through the DNNs. These spurious features tend to merge with the normal features to generate dominant evidence and further deflect the DNN reference. Composite features may not support a single normal label or the predicted label of the UAP itself but may support other labels which we call dominant labels. As a result, when the clean features do not comply with the UAP features, data of the entire class tend to be misclassified to the dominant labels, whereas the data are resilient when the class features conform to the UAP features.

3.6. Summary

In this section, the proposed FBUA framework was evaluated by comprehensive experiments involving eight cutting-edge DNNs, five SOTA UAP generation algorithms, and a pair of synthetic-real SAR datasets (SARSIM and MSTAR). The results show that the FBUA is of considerable efficacy in attacking DNN-based SAR ATR models. The UAPs generated by an FBUA also represent certain robustness against various interference, illustrating the potential threats from invisible adversaries. Furthermore, we demonstrated and analyzed the class-wise vulnerability of the ATR models when attacked by the black-box UAPs. The hypothesis of how the UAPs work was provided, that is, by igniting strong spurious features to destruct the original discriminative evidence.

4. Conclusions and Discussion

In this article, we have proposed the FBUA framework to fool various DNN-based SAR ATR models and SAR target images with a single UAP, in which scenario the attacker has no direct access to any of the target models’ information, including architecture, parameters, gradients, and training data. With extensive evaluations on the MSTAR and SARSIM datasets, we found that the substitute models trained with simulated SAR images could be utilized to craft UAPs that are very sensitive to the target models trained with real SAR data. This cross-dataset universal attack results in vigorous and spurious features for the target models which manifest as class-wise vulnerability; that is, there exist several dominant labels of which the attacked images are highly clustered and the original images are robust to the UAP. The class-wise vulnerability behaves differently from the hypothesis of dominant features reported in the intra-dataset universal attack for optical image classifiers. It may be caused by the different representations for SAR targets learned by models from the simulated data and real measured data.
This study demonstrates the existence of an access-free adversarial threat against DNN-based SAR ATR techniques. Due to the high stakes and the intrinsic non-cooperative nature of SAR ATR applications, the proposed FBUA is worth being further investigated in many aspects. For example, the key to a cross-dataset attack may be to model the knowledge of the structures and electromagnetic scattering behavior so perhaps this physical knowledge can be utilized directly to obtain efficient substitute models or directly to obtain aggressive adversarial examples. For another example, the UAP fools the target ATR system for SAR target images from all classes, which can lead to severe system misbehavior and raise suspicion for a human co-operator [51]. Therefore, a class-discriminative UAP may further facilitate the practical relevance that allows the attacker to have control over the classes to attack, such as a discriminative UAP may only deflect the tank targets to air defense units but has minimal influence on other types of targets [52]. In addition, the proposed framework can be directly exploited to improve the robustness of the ATR models against the black-box universal attacks. Specifically, current research has demonstrated that the UAP generation algorithm can be conventionally combined with the neural network training process to pursue the low-cost robustness against UAPs [53].
In future work, we plan to further facilitate the current attack with physical feasibility, such as by using electromagnetic metasurface [34]. We would also like to utilize the proposed method to construct robust SAR ATR models against black-box universal perturbations.

Author Contributions

Conceptualization, B.P. (Bowen Peng) and B.P. (Bo Peng); methodology, B.P. (Bowen Peng); software, B.P. (Bowen Peng); validation, B.P. (Bowen Peng), B.P. (Bo Peng) and L.L.; resources, B.P. (Bo Peng) and S.Y.; writing—original draft preparation, B.P. (Bowen Peng); writing—review and editing, B.P. (Bo Peng) and L.L.; supervision, S.Y.; funding acquisition, B.P. (Bo Peng) and L.L. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported partially by the National Key Research and Development Program of China under Grant 2021YFB3100800, National Natural Science Foundation of China under Grant 61921001, 62022091, and the Changsha Outstanding Innovative Youth Training Program under Grant kq2107002.

Data Availability Statement

The MSTAR dataset is available at https://www.sdms.afrl.af.mil/datasets/mstar/ (accessed on 29 October 2020). The SARSIM dataset is available at https://zenodo.org/record/573750 (accessed on 1 May 2022).

Conflicts of Interest

The authors declare no conflict of interest.

Abbreviations

The following abbreviations are used in this manuscript:
ATRAutomatic Target Recognition
BIMBasic Iterative Method
CosSimCosine Similarity
DFDominant Feature
DNNDeep Neural Network
FFFFast Feature Fool
FGSMFast Gradient Sign Method
GDGeneralizable Date-Free
MSTARMoving and Stationary Target Acquisition and Recognition
SARSynthetic Aperture Radar
UAPUniversal Adversarial Perturbation

Appendix A

Table A1. Fooling ratio (%) of DF-UAP with MobileNetV2 as substitute model and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 and the results of random noise are averaged on 10 runs.
Table A1. Fooling ratio (%) of DF-UAP with MobileNetV2 as substitute model and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 and the results of random noise are averaged on 10 runs.
ϵ NoiseAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Uniform3.60.20.427.57.4029.70.2
Gaussian0.60.10.31.01.304.00.1
UAP47.724.732.555.847.310.457.212.7
16 / 255 Uniform24.56.42.662.944.4072.90.2
Gaussian5.20.40.637.012.3038.60.2
UAP68.251.267.079.675.038.376.643.0
24 / 255 Uniform42.017.79.277.570.40.185.10.7
Gaussian16.728.31.459.036.2065.70.2
UAP74.558.777.989.980.757.382.260.6
32 / 255 Uniform52.827.218.885.984.22.189.36.3
Gaussian29.19.23.867.353.4078.10.2
UAP80.468.281.490.188.664.089.166.2
Table A2. Fooling ratio (%) of GD-UAP with VGG11 as substitute model and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 and the results of random noise are averaged on 10 runs.
Table A2. Fooling ratio (%) of GD-UAP with VGG11 as substitute model and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 and the results of random noise are averaged on 10 runs.
ϵ NoiseAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Uniform3.60.20.427.57.4029.70.2
Gaussian0.60.10.31.01.304.00.1
UAP44.924.934.835.448.616.240.417.9
16 / 255 Uniform24.56.42.662.944.4072.90.2
Gaussian5.20.40.637.012.3038.60.2
UAP65.052.465.973.182.731.877.042.4
24 / 255 Uniform42.017.79.277.570.40.185.10.7
Gaussian16.728.31.459.036.2065.70.2
UAP72.962.983.089.186.050.880.964.9
32 / 255 Uniform52.827.218.885.984.22.189.36.3
Gaussian29.19.23.867.353.4078.10.2
UAP80.970.587.490.087.461.182.576.1
Table A3. Robustness test of DF-UAP with MobileNetV2 as substitute model.
Table A3. Robustness test of DF-UAP with MobileNetV2 as substitute model.
ϵ InterferenceAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Clean47.724.732.555.847.310.457.212.7
Additive49.226.232.759.150.310.161.212.9
Multiplicative35.613.616.753.837.73.652.66.0
Gaussian40.518.228.117.726.511.122.611.3
Median48.433.943.136.943.620.838.421.5
16 / 255 Clean68.251.267.079.675.038.376.643.0
Additive67.150.265.677.173.635.776.140.5
Multiplicative56.838.742.670.163.115.273.219.7
Gaussian64.849.866.066.966.340.754.641.5
Median66.754.472.673.370.955.063.355.3
24 / 255 Clean74.558.777.989.980.757.382.260.6
Additive73.058.077.889.779.655.981.259.8
Multiplicative66.949.962.675.574.830.678.740.5
Gaussian68.558.475.478.975.359.570.760.8
Median74.162.877.877.779.366.673.566.2
32 / 255 Clean80.468.281.490.188.664.089.166.2
Additive80.167.081.390.187.763.588.966.3
Multiplicative69.555.068.987.379.441.680.150.8
Gaussian75.366.379.584.082.166.178.367.4
Median77.269.581.487.384.770.780.873.8
Table A4. Robustness test of of GD-UAP with VGG11 as substitute model.
Table A4. Robustness test of of GD-UAP with VGG11 as substitute model.
ϵ InterferenceAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Clean44.924.934.835.448.616.240.417.9
Additive40.020.028.031.642.411.535.313.3
Multiplicative26.05.112.522.121.34.126.46.1
Gaussian42.119.831.120.330.215.125.114.8
Median48.826.640.128.637.723.232.221.5
16 / 255 Clean65.052.465.973.182.731.877.042.4
Additive63.650.462.070.080.727.975.939.5
Multiplicative51.234.431.467.364.08.069.117.7
Gaussian62.448.758.663.365.931.452.937.3
Median67.754.668.176.071.142.756.548.1
24 / 255 Clean72.962.983.089.186.050.880.964.9
Additive72.261.882.388.885.749.381.163.2
Multiplicative64.247.354.271.682.417.880.434.5
Gaussian69.860.779.481.578.551.963.259.0
Median73.565.585.081.777.962.671.668.9
32 / 255 Clean80.970.587.490.087.461.182.576.1
Additive80.669.487.390.087.059.481.675.1
Multiplicative70.055.467.085.987.227.483.347.7
Gaussian76.969.586.682.083.064.975.372.9
Median79.976.888.389.085.475.382.879.4
Table A5. Robustness test of DF-UAP with MobileNetV2 as substitute model.
Table A5. Robustness test of DF-UAP with MobileNetV2 as substitute model.
ClassAlexNetVGG11ResNet50DenseNet121MobileNetV2AConvNetShuffleNetV2SqueezeNet
2S1100991001001007810085
BMP295484852627
BRDM28968778992439137
BTR7013731390772
BTR609988879898679882
D79887981001007610085
T72140110024020
T6210010091321005710064
ZIL1317862839892349658
ZSU23494078100100210010
Total682512670796750383766430
Table A6. Detailed fooling ratio (%) on each class. The attack is GD-UAP with VGG11 as substitute model.
Table A6. Detailed fooling ratio (%) on each class. The attack is GD-UAP with VGG11 as substitute model.
ClassAlexNetVGG11ResNet50DenseNet121MobileNetV2AConvNetShuffleNetV2SqueezeNet
2S1100991001001007110082
BMP2175602922609
BRDM28554678088198228
BTR70223411438089014
BTR609489799298599890
D7938595100997510083
T7290410076090
T62991008621004210051
ZIL1316958828785189149
ZSU234620759899010018
Total650524659731827318770424
Figure A1. Distribution of adversarial labels. The attack is DF-UAP with MobileNetV2 as substitute model. Ori. Category refers to the ground truth category of targets, Adv. Category denotes the misclassified category of the universal adversarial examples. The number from 0 to 9 respectively indicates the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, ZSU234.
Figure A1. Distribution of adversarial labels. The attack is DF-UAP with MobileNetV2 as substitute model. Ori. Category refers to the ground truth category of targets, Adv. Category denotes the misclassified category of the universal adversarial examples. The number from 0 to 9 respectively indicates the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, ZSU234.
Remotesensing 14 04017 g0a1
Figure A2. Distribution of adversarial labels. The attack is GD-UAP with VGG11 as substitute model. Ori. Category refers to the ground truth category of targets, Adv. Category denotes the misclassified category of the universal adversarial examples. The number from 0 to 9 respectively indicates the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, ZSU234.
Figure A2. Distribution of adversarial labels. The attack is GD-UAP with VGG11 as substitute model. Ori. Category refers to the ground truth category of targets, Adv. Category denotes the misclassified category of the universal adversarial examples. The number from 0 to 9 respectively indicates the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, ZSU234.
Remotesensing 14 04017 g0a2

References

  1. Yue, D.; Xu, F.; Frery, A.C.; Jin, Y. Synthetic Aperture Radar Image Statistical Modeling: Part One-Single-Pixel Statistical Models. IEEE Geosci. Remote Sens. Mag. 2021, 9, 82–114. [Google Scholar] [CrossRef]
  2. Krizhevsky, A.; Sutskever, I.; Hinton, G. ImageNet Classification with Deep Convolutional Neural Networks. In Proceedings of the Conference and Workshop on Neural Information Processing Systems (NeurIPS); NIPS: Lake Tahoe, NV, USA, 2012; pp. 1097–1105. [Google Scholar]
  3. Liu, L.; Chen, J.; Zhao, G.; Fieguth, P.; Chen, X.; Pietikäinen, M. Texture Classification in Extreme Scale Variations Using GANet. IEEE Trans. Image Proces. 2019, 18, 3910–3922. [Google Scholar] [CrossRef]
  4. Chen, S.; Wang, H.; Xu, F.; Jin, Y.Q. Target Classification Using the Deep Convolutional Networks for SAR Images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 4806–4817. [Google Scholar] [CrossRef]
  5. Zhang, L.; Leng, X.; Feng, S.; Ma, X.; Ji, K.; Kuang, G.; Liu, L. Domain Knowledge Powered Two-Stream Deep Network for Few-Shot SAR Vehicle Recognition. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–15. [Google Scholar] [CrossRef]
  6. Li, Y.; Du, L.; Wei, D. Multiscale CNN Based on Component Analysis for SAR ATR. IEEE Trans. Geosci. Remote Sens. 2022, 60, 1–12. [Google Scholar] [CrossRef]
  7. Ai, J.; Wang, F.; Mao, Y.; Luo, Q.; Yao, B.; Yan, H.; Xing, M.; Wu, Y. A Fine PolSAR Terrain Classification Algorithm Using the Texture Feature Fusion Based Improved Convolutional Autoencoder. IEEE Trans. Geosci. Remote Sens. 2021, 60, 1. [Google Scholar] [CrossRef]
  8. Zhu, X.; Montazeri, S.; Ali, M.; Hua, Y.; Wang, Y.; Mou, L.; Shi, Y.; Xu, F.; Bamler, R. Deep Learning Meets SAR: Concepts, Models, Pitfalls, and Perspectives. IEEE Geosci. Remote Sens. Mag. 2021, 9, 143–172. [Google Scholar] [CrossRef]
  9. Chen, L.; Xu, Z.; Li, Q.; Peng, J.; Wang, S.; Li, H. An Empirical Study of Adversarial Examples on Remote Sensing Image Scene Classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 7419–7433. [Google Scholar] [CrossRef]
  10. Li, H.; Huang, H.; Chen, L.; Peng, J.; Huang, H.; Cui, Z.; Mei, X.; Wu, G. Adversarial Examples for CNN-Based SAR Image Classification: An Experience Study. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2021, 14, 1333–1347. [Google Scholar] [CrossRef]
  11. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing Properties of Neural Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
  12. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and Harnessing Adversarial Examples. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  13. Tanay, T.; Griffin, L. A Boundary Tilting Persepective on the Phenomenon of Adversarial Examples; Cornell University: Ithaca, NY, USA, 2016. [Google Scholar] [CrossRef]
  14. Fawzi, A.; Moosavi-Dezfooli, S.M.; Frossard, P. The Robustness of Deep Networks: A Geometrical Perspective. IEEE Signal Process. Mag. 2017, 34, 50–62. [Google Scholar] [CrossRef]
  15. Ortiz-Jiménez, G.; Modas, A.; Moosavi-Dezfooli, S.M.; Frossard, P. Optimism in the Face of Adversity: Understanding and Improving Deep Learning through Adversarial Robustness. Proc. IEEE 2021, 109, 635–659. [Google Scholar] [CrossRef]
  16. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
  17. Ma, C.; Chen, L.; Yong, J.H. Simulating Unknown Target Models for Query-Efficient Black-Box Attacks. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Online, 21–25 June 2021; pp. 11835–11844. [Google Scholar]
  18. Chen, W.; Zhang, Z.; Hu, X.; Wu, B. Boosting Decision-Based Black-Box Adversarial Attacks with Random Sign Flip. In Proceedings of the European Conference on Computer Vision (ECCV), Glasgow, UK, 23–28 August 2020; pp. 276–293. [Google Scholar]
  19. Shi, Y.; Han, Y.; Hu, Q.; Yang, Y.; Tian, Q. Query-efficient Black-box Adversarial Attack with Customized Iteration and Sampling. IEEE Trans. Pattern Anal. Mach. Intell. 2022, 1. [Google Scholar] [CrossRef] [PubMed]
  20. Dong, Y.; Pang, T.; Su, H.; Zhu, J. Evading Defenses to Transferable Adversarial Examples by Translation-Invariant Attacks. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Long Beach, CA, USA, 15–20 June 2019; pp. 4312–4321. [Google Scholar]
  21. Lin, J.; Song, C.; He, K.; Wang, L.; Hopcroft, J. Nesterov Accelerated Gradient and Scale Invariance for Adversarial Attacks. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
  22. Xie, C.; Zhang, Z.; Zhou, Y.; Bai, S.; Wang, J.; Ren, Z.; Yuille, A.L. Improving Transferability of Adversarial Examples With Input Diversity. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Long Beach, CA, USA, 15–20 June 2019. [Google Scholar]
  23. Wang, X.; He, K. Enhancing the Transferability of Adversarial Attacks Through Variance Tuning. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Nashville, TN, USA, 20–25 June 2021; pp. 1924–1933. [Google Scholar]
  24. Moosavi-Dezfooli, S.M.; Fawzi, A.; Fawzi, O.; Frossard, P. Universal Adversarial Perturbations. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Honolulu, HI, USA, 21–26 July 2017. [Google Scholar]
  25. Poursaeed, O.; Katsman, I.; Gao, B.; Belongie, S. Generative Adversarial Perturbations. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  26. Mopuri, K.R.; Ojha, U.; Garg, U.; Babu, R.V. NAG: Network for Adversary Generation. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  27. Mopuri, K.R.; Garg, U.; Babu, R.V. Fast Feature Fool: A Data Independent Approach to Universal Adversarial Perturbations. arXiv 2017, arXiv:1707.05572. [Google Scholar]
  28. Khrulkov, V.; Oseledets, I. Art of Singular Vectors and Universal Adversarial Perturbations. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
  29. Mopuri, K.R.; Ganeshan, A.; Babu, R.V. Generalizable Data-Free Objective for Crafting Universal Adversarial Perturbations. IEEE Trans. Pattern Anal. Mach. Intell. 2019, 41, 2452–2465. [Google Scholar] [CrossRef] [PubMed]
  30. Zhang, C.; Benz, P.; Lin, C.; Karjauv, A.; Wu, J.; Kweon, I.S. A Survey on Universal Adversarial Attack. In Proceedings of the International Joint Conference on Artificial Intelligence (IJCAI), Online, 19–26 August 2021; pp. 4687–4694. [Google Scholar]
  31. Huang, T.; Zhang, Q.; Liu, J.; Hou, R.; Wang, X.; Li, Y. Adversarial Attacks on Deep-Learning-Based SAR Image Target Recognition. J. Netw. Comput. Appl. 2020, 162, 102632. [Google Scholar] [CrossRef]
  32. Du, C.; Zhang, L. Adversarial Attack for SAR Target Recognition Based on UNet-Generative Adversarial Network. Remote Sens. 2021, 13, 4358. [Google Scholar] [CrossRef]
  33. Du, C.; Huo, C.; Zhang, L.; Chen, B.; Yuan, Y. Fast C&W: A Fast Adversarial Attack Algorithm to Fool SAR Target Recognition With Deep Convolutional Neural Networks. IEEE Geosci. Remote Sens. Lett. 2022, 19, 1–5. [Google Scholar] [CrossRef]
  34. Liu, Z.; Xia, W.; Lei, Y. SAR-GPA: SAR Generation Perturbation Algorithm. In Proceedings of the 2021 3rd International Conference on Advanced Information Science and System (AISS 2021), Sanya, China, 26–28 November 2021; pp. 1–6. [Google Scholar]
  35. Xu, L.; Feng, D.; Zhang, R.; Wang, X. High-Resolution Range Profile Deception Method Based on Phase-Switched Screen. IEEE Antennas Wirel. Propag. Lett. 2016, 15, 1665–1668. [Google Scholar] [CrossRef]
  36. Xu, H.; Guan, D.F.; Peng, B.; Liu, Z.; Yong, S.; Liu, Y. Radar One-Dimensional Range Profile Dynamic Jamming Based on Programmable Metasurface. IEEE Antennas Wirel. Propag. Lett. 2021, 20, 1883. [Google Scholar] [CrossRef]
  37. Peng, B.; Peng, B.; Zhou, J.; Xia, J.; Liu, L. Speckle-Variant Attack: Towards Transferable Adversarial Attack to SAR Target Recognition. IEEE Geosci. Remote Sens. Lett. 2022, 19, 4509805. [Google Scholar] [CrossRef]
  38. Kusk, A.; Abulaitijiang, A.; Dall, J. Synthetic SAR Image Generation Using Sensor, Terrain and Target Models. In Proceedings of the 11th European Conference on Synthetic Aperture Radar (VDE), Hamburg, Germany, 6–9 June 2016; pp. 1–5. [Google Scholar]
  39. Malmgren-Hansen, D.; Kusk, A.; Dall, J.; Nielsen, A.A.; Engholm, R.; Skriver, H. Improving SAR Automatic Target Recognition Models with Transfer Learning from Simulated Data. IEEE Geosci. Remote Sens. Lett. 2017, 14, 1484–1488. [Google Scholar] [CrossRef]
  40. Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A Simple and Accurate Method to Fool Deep Neural Networks. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Las Vegas, NV, USA, 26–30 June 2016; pp. 2574–2582. [Google Scholar]
  41. Zhang, C.; Benz, P.; Imtiaz, T.; Kweon, I.S. Understanding Adversarial Examples From the Mutual Influence of Images and Perturbations. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Seoul, Korea, 14–19 June 2020. [Google Scholar]
  42. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
  43. Sutskever, I.; Martens, J.; Dahl, G.; Hinton, G. On the Importance of Initialization and Momentum in Deep Learning. In Proceedings of the International Conference on Machine Learning, Atlanta, GA, USA, 17–19 June 2013; pp. 1139–1147. [Google Scholar]
  44. Zhang, C.; Benz, P.; Karjauv, A.; Kweon, I.S. Data-Free Universal Adversarial Perturbation and Black-Box Attack. In Proceedings of the International Conference on Computer Vision (ICCV), Montreal, QC, Canada, 11–17 October 2021; pp. 7868–7877. [Google Scholar]
  45. Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
  46. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep Residual Learning for Image Recognition. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
  47. Huang, G.; Liu, Z.; Van Der Maaten, L.; Weinberger, K. Densely Connected Convolutional Networks. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 4700–4708. [Google Scholar]
  48. Sandler, M.; Howard, A.; Zhu, M.; Zhmoginov, A.; Chen, L.C. Mobilenetv2: Inverted Residuals and Linear Bottlenecks. In Proceedings of the Computer Vision and Pattern Recognition Conference (CVPR), Salt Lake City, UT, USA, 18–23 June 2018; pp. 4510–4520. [Google Scholar]
  49. Ma, N.; Zhang, X.; Zheng, H.T.; Sun, J. ShuffleNet V2: Practical Guidelines for Efficient CNN Architecture Design. In Proceedings of the European Conference on Computer Vision (ECCV), Munich, Germany, 8–14 September 2018. [Google Scholar]
  50. Iandola, F.N.; Han, S.; Moskewicz, M.W.; Ashraf, K.; Dally, W.J.; Keutzer, K. SqueezeNet: AlexNet-Level Accuracy with 50× Fewer Parameters and <0.5 MB Model Size. arXiv 2016, arXiv:1602.07360. [Google Scholar]
  51. Zhang, C.; Benz, P.; Imtiaz, T.; Kweon, I.S. CD-UAP: Class Discriminative Universal Adversarial Perturbation. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 6754–6761. [Google Scholar]
  52. Benz, P.; Zhang, C.; Imtiaz, T.; Kweon, I.S. Double Targeted Universal Adversarial Perturbations. In Proceedings of the Asian Conference on Computer Vision (ACCV), Kyoto, Japan, 30 November–4 December 2020. [Google Scholar]
  53. Shafahi, A.; Najibi, M.; Xu, Z.; Dickerson, J.; Davis, L.S.; Goldstein, T. Universal Adversarial Training. In Proceedings of the AAAI Conference on Artificial Intelligence, New York, NY, USA, 7–12 February 2020; Volume 34, pp. 5636–5643. [Google Scholar]
Figure 1. Workflow of the proposed FBUA for the SAR ATR. The simulated SAR images are from the publically accessible SARSIM dataset [38,39], and the real SAR images are from the MSTAR dataset. Please notice that the Parula color map is employed in this article to visualize the gray-scale SAR images.
Figure 1. Workflow of the proposed FBUA for the SAR ATR. The simulated SAR images are from the publically accessible SARSIM dataset [38,39], and the real SAR images are from the MSTAR dataset. Please notice that the Parula color map is employed in this article to visualize the gray-scale SAR images.
Remotesensing 14 04017 g001
Figure 2. Examples of the targets in the MSTAR dataset: (top) optical images; and (bottom) the corresponding SAR images.
Figure 2. Examples of the targets in the MSTAR dataset: (top) optical images; and (bottom) the corresponding SAR images.
Remotesensing 14 04017 g002
Figure 3. Examples of the targets in the SARSIM dataset.
Figure 3. Examples of the targets in the SARSIM dataset.
Remotesensing 14 04017 g003
Figure 4. Transferability of UAPs generated by the studied attack algorithms. The fooling ratio is evaluated by using the UAPs generated by substitute SARSIM models to attack the target MSTAR models.
Figure 4. Transferability of UAPs generated by the studied attack algorithms. The fooling ratio is evaluated by using the UAPs generated by substitute SARSIM models to attack the target MSTAR models.
Remotesensing 14 04017 g004
Figure 5. UAPs generated by different algorithms for the studied DNN models: (a) DeepFool-UAP; (b) DF-UAP; (c) CosSim-UAP; and (d) GD-UAP. All the UAPs are enlarged to [ 1 , 1 ] for observation.
Figure 5. UAPs generated by different algorithms for the studied DNN models: (a) DeepFool-UAP; (b) DF-UAP; (c) CosSim-UAP; and (d) GD-UAP. All the UAPs are enlarged to [ 1 , 1 ] for observation.
Remotesensing 14 04017 g005aRemotesensing 14 04017 g005b
Figure 6. Adversarial examples with various l constraint ϵ . (a) Ori. (b) ϵ = 8/255. (c) ϵ = 16/255. (d) ϵ = 24/255. (e) ϵ = 32/255.
Figure 6. Adversarial examples with various l constraint ϵ . (a) Ori. (b) ϵ = 8/255. (c) ϵ = 16/255. (d) ϵ = 24/255. (e) ϵ = 32/255.
Remotesensing 14 04017 g006
Figure 7. Distribution of adversarial label: Ori category refers to the ground truth category of targets; Adv. category denotes the misclassified category of the universal adversarial examples. The numbers from 0 to 9, respectively, indicate the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, and ZSU234. Results of DF-UAP and GD-UAP can be found in Figure A1 and Figure A2.
Figure 7. Distribution of adversarial label: Ori category refers to the ground truth category of targets; Adv. category denotes the misclassified category of the universal adversarial examples. The numbers from 0 to 9, respectively, indicate the following classes: 2S1, BMP2, BRDM2, BTR70, BTR60, D7, T72, T62, ZIL131, and ZSU234. Results of DF-UAP and GD-UAP can be found in Figure A1 and Figure A2.
Remotesensing 14 04017 g007
Table 1. Details of the SOC subset of MSTAR dataset.
Table 1. Details of the SOC subset of MSTAR dataset.
TypeClassSerial NumberTraining Set
(17°)
Test Set
(15°)
Rocket launcher2S1b01299274
Armored
personnel
carrier
BMP29563233195
BRDM2E71298274
BTR70c71233196
BTR60k10yt7532256195
BulldozerD792v13015299274
TankT72132232196
T62A51299273
TruckZIL131E12299274
Air defense unitZSU234d08299274
Total 27472425
Table 2. Details of the training set of the SARSIM dataset used in this work.
Table 2. Details of the training set of the SARSIM dataset used in this work.
TypeClassTraining Set (17°)Training Set (15°)
Bulldozer8020216216
13013216216
CarPeugeot607216216
Toyota216216
Humvee3663216216
9657216216
Tank65047216216
86347216216
Truck2096216216
2107216216
Total 21602160
Table 3. Details of the studied models.
Table 3. Details of the studied models.
ModelInput Size# Paras.FLOPs ( × 10 9 ) MSTAR Acc. (%)SARSIM Acc. (%)
AlexNet [2] 224 2 58,299,0821.0696.394.2
VGG11 [45] 224 2 128,814,1542.0298.196.5
ResNet50 [46] 224 2 23,522,2504.0496.492.8
DenseNet121 [47] 224 2 6,957,7062.8097.896.4
MobileNetV2 [48] 224 2 2,236,1060.3197.496.0
AConvNet [4] 88 2 303,4980.0498.194.5
ShuffleNetV2 [49] 224 2 1,263,4220.1496.996.6
SqueezeNet [50] 224 2 726,4740.2697.194.0
# Paras: number of trainable parameters; FLOPs: number of floating point operations when reasoning an input image.
Table 4. Fooling ratio (%) of CosSim-UAP with substitute model of MobileNetV2 and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 , and the results of random noise are averaged on 10 runs. Best results are highlighted in Bold. Results of DF-UAP and GD-UAP can be found in Table A1 and Table A2.
Table 4. Fooling ratio (%) of CosSim-UAP with substitute model of MobileNetV2 and various ϵ . The deviation of Gaussian noise is set to ϵ / 3 , and the results of random noise are averaged on 10 runs. Best results are highlighted in Bold. Results of DF-UAP and GD-UAP can be found in Table A1 and Table A2.
ϵ NoiseAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Uniform3.60.20.427.57.4029.70.2
Gaussian0.60.10.31.01.304.00.1
UAP47.327.129.857.049.910.659.113.7
16 / 255 Uniform24.56.42.662.944.4072.90.2
Gaussian5.20.40.637.012.3038.60.2
UAP67.453.069.281.274.445.377.448.7
24 / 255 Uniform42.017.79.277.570.40.185.10.7
Gaussian16.728.31.459.036.2065.70.2
UAP72.162.678.589.378.861.183.263.0
32 / 255 Uniform52.827.218.885.984.22.189.36.3
Gaussian29.19.23.867.353.4078.10.2
UAP77.873.882.390.083.665.788.872.6
Table 5. Fooling ratio (%) against various perturbations. The additive noise is white noise with a deviation of 10 2 , the multiplicative noise is truncated exponential noise with a tail of 2, the deviation of Gaussian filtering is 1. Kernel size of the filters is 7 × 7. The results of random noise are averaged on 10 runs. Results of DF-UAP and GD-UAP can be found in Table A3 and Table A4.
Table 5. Fooling ratio (%) against various perturbations. The additive noise is white noise with a deviation of 10 2 , the multiplicative noise is truncated exponential noise with a tail of 2, the deviation of Gaussian filtering is 1. Kernel size of the filters is 7 × 7. The results of random noise are averaged on 10 runs. Results of DF-UAP and GD-UAP can be found in Table A3 and Table A4.
ϵ InterferenceAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
8 / 255 Clean47.327.129.857.049.910.659.113.7
Additive41.021.425.154.243.88.254.210.4
Multiplicative27.47.012.141.625.82.637.65.1
Gaussian37.719.526.418.628.311.525.813.0
Median50.835.846.034.645.724.640.024.2
16 / 255 Clean67.453.069.281.274.445.377.448.7
Additive66.651.867.377.573.341.876.747.3
Multiplicative57.140.946.369.563.020.671.924.9
Gaussian64.752.167.567.467.746.258.948.6
Median66.954.472.673.071.355.764.155.5
24 / 255 Clean72.162.678.589.378.861.183.263.0
Additive70.761.777.588.878.060.082.562.0
Multiplicative66.850.664.175.974.236.477.844.4
Gaussian70.762.076.376.974.962.471.563.1
Median74.262.777.877.379.366.973.766.4
32 / 255 Clean77.873.882.390.083.665.788.872.6
Additive77.372.881.890.083.065.288.771.6
Multiplicative70.156.371.486.678.348.280.053.2
Gaussian76.271.279.383.980.166.580.273.0
Median77.370.281.287.684.371.080.575.5
Table 6. Detailed fooling ratio (%) on each class. The attack is CosSim-UAP with MobileNetV2 as the substitute model. Results of DF-UAP and GD-UAP can be found in Table A5 and Table A6.
Table 6. Detailed fooling ratio (%) on each class. The attack is CosSim-UAP with MobileNetV2 as the substitute model. Results of DF-UAP and GD-UAP can be found in Table A5 and Table A6.
ClassAlexNetVGG11ResNet50DenseMobileAConvNetShuffleSqueeze
2S1100991001001008610092
BMP210136663855427
BRDM28969809192469544
BTR7024837380802
BTR609989899898709886
D79889981001008110086
T72110010016000
T6210010089251005510072
ZIL1317767879892589768
ZSU23488075100100210010
Total674530692812744453774487
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Peng, B.; Peng, B.; Yong, S.; Liu, L. An Empirical Study of Fully Black-Box and Universal Adversarial Attack for SAR Target Recognition. Remote Sens. 2022, 14, 4017. https://doi.org/10.3390/rs14164017

AMA Style

Peng B, Peng B, Yong S, Liu L. An Empirical Study of Fully Black-Box and Universal Adversarial Attack for SAR Target Recognition. Remote Sensing. 2022; 14(16):4017. https://doi.org/10.3390/rs14164017

Chicago/Turabian Style

Peng, Bowen, Bo Peng, Shaowei Yong, and Li Liu. 2022. "An Empirical Study of Fully Black-Box and Universal Adversarial Attack for SAR Target Recognition" Remote Sensing 14, no. 16: 4017. https://doi.org/10.3390/rs14164017

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop