Next Article in Journal
An Efficient Lightweight SAR Ship Target Detection Network with Improved Regression Loss Function and Enhanced Feature Information Expression
Next Article in Special Issue
Effective Feature Selection Methods to Detect IoT DDoS Attack in 5G Core Network
Previous Article in Journal
Towards Personalization of Indoor Air Quality: Review of Sensing Requirements and Field Deployments
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Universal Detection Method for Adversarial Examples and Fake Images

1
School of Computer Science and Cyber Engineering, Guangzhou University, Guangzhou 510006, China
2
Pazhou Lab, Guangzhou 510330, China
3
Institute of Artificial Intelligence and Blockchain, Guangzhou University, Guangzhou 510006, China
*
Author to whom correspondence should be addressed.
Sensors 2022, 22(9), 3445; https://doi.org/10.3390/s22093445
Submission received: 17 March 2022 / Revised: 4 April 2022 / Accepted: 7 April 2022 / Published: 30 April 2022
(This article belongs to the Collection Cryptography and Security in IoT and Sensor Networks)

Abstract

:
Deep-learning technologies have shown impressive performance on many tasks in recent years. However, there are multiple serious security risks when using deep-learning technologies. For examples, state-of-the-art deep-learning technologies are vulnerable to adversarial examples that make the model’s predictions wrong due to some specific subtle perturbation, and these technologies can be abused for the tampering with and forgery of multimedia, i.e., deep forgery. In this paper, we propose a universal detection framework for adversarial examples and fake images. We observe some differences in the distribution of model outputs for normal and adversarial examples (fake images) and train the detector to learn the differences. We perform extensive experiments on the CIFAR10 and CIFAR100 datasets. Experimental results show that the proposed framework has good feasibility and effectiveness in detecting adversarial examples or fake images. Moreover, the proposed framework has good generalizability for the different datasets and model structures.

1. Introduction

In recent years, as one of the core technologies of artificial intelligence, deep learning has attracted unprecedented attention from academia and industry [1]. Compared with traditional machine learning methods, deep learning produces results with higher accuracy, does not require complex feature engineering, and has better adaptability. Hence, deep-learning technology has been gradually applied to various fields, such as computer vision, speech recognition, natural language processing, autonomous driving, etc., [2,3,4,5]. However, research shows that deep learning still has many problems in its security and privacy [6,7,8,9], such as adversarial examples and deep forgery [10,11].
Szegedy et al. first proposed the concept of adversarial examples [12]. Its basic principle is to add some specific subtle perturbations to the original data; the model would output error results with high confidence. The discovery of adversarial examples illustrates the fragility of deep-learning models. Since then, the researchers have researched adversarial examples and proposed many adversarial example generation methods, such as FGSM, C&W, DeepFool, etc., [13,14,15,16]. These methods can generate adversarial examples with extremely high success rates based on different attack scenarios and targets. Moreover, it was found that the adversarial examples are transferable, i.e., the adversarial examples generated for one model are effective for other similar models [17]. This aggravates the seriousness of deep learning security problems and greatly restricts the application of deep-learning technology in military, medical, financial, and other sophisticated fields [18,19,20].
Except for the security issues of the technology itself, deep learning has abuse problems, such as deep forgery. Deep forgery uses deep-learning algorithms [11], i.e., generative adversarial networks (GANs), to tamper with or forge original data so that observers mistakenly regard fake data as original data. These fake data are realistic, diversified, and challenging to identify. With the help of the ripple effect of online social media, fake data is likely to spread on a large scale, causing a severe social impact. More seriously, if criminals use deep forgery for political or malicious profit-making motives, it will result in many risks and severe threats to political, economic, social, and national security. Hence, the detection of adversarial examples and fake images is a hot issue in academia and industry.
To solve the above issue, we explore the detection methods for adversarial examples and fake images. Inspired by membership inference attacks that use the model output to determine whether a sample belongs to the training dataset [21], we observe that there are also differences between the model outputs of normal samples and adversarial examples (fake images). Specifically, the distribution of model outputs is fit using the kernel density estimation. We propose a universal detection method for adversarial examples and fake images based on this difference. This method includes the detector training algorithm and online detection algorithm. The detector training algorithm is used to construct the model output’s dataset and train the detector. This dataset consists of the model outputs of normal samples and adversarial examples (fake images). In addition, the detector is trained to learn the rule of the output distribution of normal samples and adversarial examples (fake images). The online detection algorithm is to obtain the model outputs of samples and calculate detection results. To the best of our knowledge, Li’s method is similar to ours [22]. However, he uses the middle-layer features of the Bayesian neural network to determine normal samples and adversarial examples, and we use the output of the deep neural network. Moreover, our method is not only suitable for the detection of adversarial examples, but also for the detection of fake images.
Our contributions consist of the following:
  • Based on the difference in the distribution of model outputs between normal samples and adversarial examples (fake images), we propose a universal detection method for adversarial examples and fake images.
  • We tested the detector’s performance using state-of-the-art generation algorithms of adversarial examples and fake images and proved the effectiveness of the detector.
  • We tested the proposed method on different datasets and neural network structures and proved the generalizability of the detector.
The rest of this paper is structured as follows. Section 2 introduces related research. Section 3 presents our method. Section 4 experimentally evaluates our method. Section 5 summarizes our work.

2. Related Work

In this section, we introduce the related work of adversarial examples and fake images from attack and defense.

2.1. Adversarial Examples

2.1.1. Attack

The generation methods of adversarial examples can be roughly divided into two types: gradient-based methods and optimization-based methods.
Gradient-based methods.The basic principle of this type of method is to add perturbations in the gradient direction, and these perturbations should not be easily noticed. FGSM is a one-step method that adds limited perturbations in the gradient direction to search for a similar image, which will cause the model to output wrong results [13]. BIM, also called I-FGSM, which is a multiple-iterations version of FGSM, extends FGSM by running a minor optimization in each iteration [23]. PGD is also an iteration extension of FGSM [24]. Different from BIM directly clipping pixel values, it first utilizes gradient projection to avoid excessive changes in the optimization process.
Optimization-based methods. Different from gradient-based methods, its basic principle is to take the generation process of adversarial examples as a constrained optimization problem, that is, to ensure that the model outputs wrong results without being easily noticed. DeepFool adds some minor perturbation to the normal image, causing the image to exceed the classification boundary through iterative calculation [15]. Similar to DeepFool, UAP also uses adversarial perturbations to make normal images exceed the classification boundary [25]. It can be added into most images to generate adversarial examples and have good generalization capabilities on other network architectures.

2.1.2. Defense

The defense methods against adversarial examples mainly adopt two strategies: robust classifier and detector-based methods.
Robust classifier methods. The robust classifier method improves the robustness of the model, such as defensive distillation [14] and adversarial training [13]. For defensive distillation, the model can be smoother than the original model by assigning a larger value of T during the student model training stage, which would reduce the sensitivity to perturbations and improve the robustness and generalizability of the model. Adversarial training is currently the most effective defense method. Its basic idea is to assign the correct labels to adversarial examples and then train the model with these samples to improve its robustness. It has also been proved that adversarial training can provide higher precision and regularization for models [13]. However, adversarial training also presents some challenges. Adversarial training needs to generate many adversarial examples, so the computational cost and time cost of the method are very high. The adversarial example generation methods are constantly updated, which leads to the continuous retraining of the model.
Detector-based methods. This strategy trains a detector to distinguish between adversarial examples and normal samples. Feinman et al. believe that the adversarial examples deviate from the manifold region of the real data, so they use the kernel density estimation function in the feature space of the middle layer to detect abnormal points that deviate from the data manifold [26]. Ma et al. observed that the Local Intrinsic Dimension of the hidden layer output differs between normal images and adversarial examples. They used this characteristic to detect adversarial examples [27]. Tian et al. found that image processing operations could invalidate the adversarial examples, which would not affect the classification of the normal images [28]. MagNet constructs multiple autoencoders and uses the reconstruction errors of autoencoders to detect adversarial examples based on cryptographic randomness [29]. SafetyNet detects adversarial examples that rely on neural activation patterns with SVM [30]. Li et al. found that the output distribution of the hidden layer of the adversarial examples was different from that of the normal images, so they used the Bayesian neural network to simulate the output distribution of the hidden layer and detect the adversarial examples using distribution dispersion [22].

2.2. Fake Images

2.2.1. Attack

Traditional image forgery methods are generated mainly by image-editing software. With the development of GANs, GANs-based generation methods for fake images have become the mainstream.
Software-based method. It is straightforward to use powerful image editing software to generate fake images without leaving perceptible artifacts. Those methods can be divided into the copy-move and splice methods, which create fake images without leaving traces by adding new content to the original image or perform image stitching, respectively. In addition, the copy-move method is the mainstream method that can change the entire meaning of the original image.
GANs-based method. GANs is a method of unsupervised learning that Goodfellow proposes. It consists of a generator and a discriminator. Taking the normal image or the generator’s output as input, the discriminator distinguishes the generator’s output from the real sample as much as possible. The generator randomly samples from the latent space as input and tries to imitate the real samples in the training set to fool the discriminator. Two neural networks learn by confronting each other and constantly adjusting the parameters. With the development of GANs technology, GANs are used for different scenarios, for example, generating high perceptual quality images, domain transfer, image to image translations, and so on [31].

2.2.2. Defense

Defense methods against fake images are roughly divided into defense against software-based fake images and defense against GANs-based fake images.
Defense against the software-based fake image. The methods of image-manipulation detection can be summarized into two types: (i) Active: people embed some additional information, i.e., digital watermarks or digital fingerprinting, into the image to determine the authenticity of the image [32]. However, active approaches have some shortcomings, such as difficulty in secondary propagation and single verification, which are confronted with overlooked challenges. (ii) Passive: passive approaches extract the features from the images and use these features for forgery detection [33]. For example, it can identify contrast enhancement, reveal image resampling, etc. To solve the problem that an image may use multiple tampering methods, the authors in [34] use various features in a steganalysis to detect the fake image and identify tampering types. However, these traditional approaches are mostly ineffective when identifying GANs-based fake images.
Defense against GANs-based fake image. Similar to adversarial-example detection, the most direct method of detecting fake images is to train the detector using real and fake data. Marra et al. show that a simple fake image detector could be constructed using an image translation network [35]. A three-channel co-occurrence matrix-based detection method was proposed in [36]. Dang et al. realized that the detection of fake face images and the location of the tampered region depend on an attention mechanism [37]. Some researchers believe that, due to the diversity of fake images and the continuous updating of generation methods, the detector could only distinguish the fake images in training. To enhance the generalizability of the detector, Zhang et al. proposed a fake image-generation method called AutoGAN [38]. This method uses a frequency-spectrum input instead of pixel-space input to train the detection model. On the contrary, Wang et al. found that only using a fake image and then performing data pre-processing or data enhancement to expand the training data set could improve the generalizability of the detector [39].

3. Method

In this paper we mainly focus on the detection of adversarial examples and GANs-based fake images, and we present our method in detail in this section.

3.1. Overview

3.1.1. Observation

We observed the difference in the distribution of model outputs between adversarial examples (fake images) and normal samples through experiments. Specifically, we first obtained the output of Googlenet on normal samples, adversarial examples, and fake images, then used kernel density estimation to fit the distribution of model outputs. The result is shown in the Figure 1.
In Figure 1, TOP1, TOP2, and TOP3 indicate the top 1 value, the top 2 values, and the top 3 values of the model outputs, respectively. NOR, ADV, and GBI represent normal data, adversarial examples, and fake images, respectively. ADV and GBI are generated by DeepFool [15] and WGAN [40], respectively. It can be seen from Figure 1 that there are significant differences in the distribution of model outputs between the normal images and the adversarial examples (fake images).

3.1.2. Framework

Based on the above observation, we propose a universal detection method for adversarial examples and fake images. The overall framework is shown in Figure 2.
Figure 2 shows that the method mainly includes two stages: detector training and online detection. The purpose of detector training is to learn the difference in the distribution of model outputs between the normal samples and adversarial examples (fake images) to distinguish the normal samples from adversarial examples or fake images. Therefore, we need to generate a certain number of adversarial examples or fake images in advance as training data. Moreover, to reduce the time-cost of training and improve the detection efficiency, we chose the first k values from the model output to represent the distribution of outputs. The purpose of online detection is to detect adversarial examples or fake images. The following is a detailed introduction to detector training and online detection.

3.2. Detector Training

The main work of detector training includes constructing the training dataset and training detector, and the basic process is shown in Algorithm 1.
Algorithm 1 Detector Training Algorithm.
Require: 
Normal Data D n , Generator G , Target Model F
  1:
D m G ( D n )
  2:
D Merge ( D m , D n )
  3:
D o u t F ( D )
  4:
D t r a i n Top ( k , D o u t )
  5:
D Train ( D , D t r a i n )
According to Algorithm 1, normal data, the malicious data generator, and the target model, F , should be obtained before training the detector. Specifically, we first use G to generate adversarial examples or fake images, as shown in Line 1. G represents some mainstream adversarial examples or fake-image-generation methods. Line 2 merges normal data and adversarial examples (fake images). Line 3 inputs the merged data D into F to obtain the output distribution dataset D o u t . In our experiments, F is the classifier trained on the CIFAR10 or CIFAR100 datasets. To reduce the complexity of the detector and improve the detection efficiency, we selected the top k values from the model output to represent the output distribution, as shown in Line 4. Line 5 uses the dataset D t r a i n to train the detector D . The detector can be trained offline and deployed online to reduce the time-cost.

3.3. Online Detection

The main work of online detection includes obtaining the model output of the untrusted data and computing the detection result. The basic process is shown in Algorithm 2.
Algorithm 2 Online Detection Algorithm.
Require: 
Untrusted Data D u , Target Model F , Detector D
  1:
D u o u t F ( D u )
  2:
D u k Top ( k , D u o u t )
  3:
R e s u l t D ( D u k )
According to Algorithm 2, Line 1 obtains the model output of untrusted data D u . Line 2 selects the top k values from D u o u t to obtain D u k . Finally, D u k is input into the detector D to obtain the detection result.

4. Experiments

In this section, we experimentally evaluate the proposed method using the CIFAR10 and CIFAR100 datasets. Both CIFAR10 and CIFAR100 are natural image datasets, and both include 50,000 training images and 10,000 test images. However, there are ten classes in CIFAR10 and 100 classes in CIFAR100. The experiments include performance experiments, generalizability experiments, and transferability experiments. Additionally, we use AUC as a measure of our detector’s performance in our experiments. Next, we will introduce the experiments in detail.

4.1. Performance Experiments

The performance experiments were used to test the ability of our method in adversarial example detection or fake image detection.

4.1.1. Performance in Adversarial Example Detection

We used nine state-of-the-art adversarial example-generation methods for detection, including FGSM, DeepFool, BIM, PGD, C&W, etc., [13,14,15,23,24]. The experimental results are shown in Table 1 and we mark out the best detection result using bold text for each type of adversarial example.
From Table 1, we know: (1) Given the generation methods of adversarial examples in training, if these methods are the same as the generation methods in the evaluation, the detector almost reached the highest level of detection accuracy; on the contrary, the detection accuracy decreased slightly. (2) On the whole, when the generation method in training is ZOO or NewtonFool, the detector reaches the best detection performance. Hence, in the real world, we could use ZOO or NewtonFool to generate adversarial examples in training.

4.1.2. Performance in Fake Image Detection

Considering that the mainstream generation methods of fake images are mainly based on GAN, we selected eight state-of-the-art GAN-based algorithms to test the detector’s performance. The experimental results are shown in Table 2 and we mark out the best detection result using bold text for each type of fake image.
It can be seen from Table 2 that the detection results of fake images are similar to the adversarial examples. Given the generation methods of fake images in training, if these methods are the same as the generation methods in the evaluation, the detector achieved the best detection performance; on the contrary, the detection accuracy decreased slightly. In addition, when the generation method in training was WGAN_DIV [40], the detector reached the best detection performance. Hence, WGAN_DIV is a candidate to generate fake images in training in the real world.

4.2. Generalizability Experiments

The generalizability of the detector mainly includes (1) the generalizability of datasets; (2) the generalizability of target model architecture. In addition, the generation methods of adversarial examples and fake images are DeepFool and WGAN [15,40], respectively. In addition, we have to describe the experiment case in the form of condition I–condition II. For example, CIFAR10–CIFAR100 represents that the detector is trained by CIFAR10 and evaluated by CIFAR100. Limited by the two time-scale update rule (TTUR), we only use AutoGAN [41] and TransGAN [42] generators on the CIFAR10 dataset. Below, we provide a detailed introduction to the experimental content.

4.2.1. Generalizability for Dataset

We alternately used CIFAR10 and CIFAR100 as the training dataset and used CIFAR100 and CIFAR10 as the evaluation dataset to verify the generalizability of the detector for the dataset. The experimental results are shown in Figure 3.
As shown in Figure 3, the AUC values of adversarial example detection are mainly distributed between 0.5 and 0.8; most of the AUC values of fake image detection are between 0.6 and 0.8. Hence, the generalizability of the fake image detector is slightly stronger than that of the adversarial example detector, but it still needs to be further improved.

4.2.2. Generalizability for Target-Model Architecture

We selected five neural network structures to verify the generalizability of the target model structure, including Resnet18, VGG11, Googlenet, Densenet121, and Inceptionv4. Specifically, we first selected one of the network structures as the target model, one after the other in turn, then used the output of the target model to train the detector, and finally used the remaining networks to evaluate the generalizability of the detector. The experimental results are shown in Figure 4.
From Figure 4, the adversarial example detector shows good generalizability for different model structures. However, there are a few cases below the average level. For example, the generalizability for Densenet121 on the CIFAR10 dataset, the generalizability for VGG11 and Googlenet on the CIFAR100 dataset, etc. Similar to the adversarial example detector, the generalizability of the fake image detector for the model structure is generally reasonable.

4.3. Transferability Experiments

The transferability experiments include (1) testing the ability of the adversarial example detector to detect fake images; (2) testing the ability of the fake image detector on adversarial examples. The experimental results are shown in Figure 5.
As shown in Figure 5, for the CIFAR10 dataset, the AUC values of the detectors are mainly distributed from 0.7 to 0.9, but the fake image detector is more stable. For the CIFAR100 dataset, the transferability of the fake image detectors is significantly better than that of adversarial example detectors, all distributed around 0.9. This shows that our method has good transferability, and a single detector can be used to simultaneously detect adversarial examples and fake images to a certain extent.

5. Conclusions

In this paper, we observe the difference in output distribution between normal samples and adversarial examples (fake images) and propose a universal detection method for adversarial examples and fake images. The method mainly includes two stages: detector training and online detection. In the detector-training stage, we used the output distribution of normal and adversarial samples (fake images) to train the adversarial example (fake image) detectors. After training the detector, we took the model output of untrusted data as the input of the detector to realize online detection of adversarial examples or fake images. We experimentally verified our method using CIFAR10 and CIFAR100 datasets, and the results show that: (1) the detector has a good detection ability for adversarial or fake images; (2) the detector has good generalizability for different model structures; and (3) the detector has good transferability, that is, the adversarial example detector and fake images detector can effectively detect fake images or adversarial examples. Hence, in the real world, our method is feasible and effective.

Author Contributions

Conceptualization, Methodology, Software, J.L.; Software, Y.H.; Writing—Reviewing and Editing, R.H.; Reviewing and Editing, X.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the National Natural Science Foundation of China (No. 62072127, No. 62002076), Project 6142111180404 and supported by the CNKLSTISS, Science and Technology Program of Guangzhou, China (No. 202002030131, No. 201904010493), Guangdong basic and applied basic research fund joint fund Youth Fund (No. 2019A1515110213), Natural Science Foundation of Guangdong Province (No. 2020A1515010423).

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Le Cun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  2. Wu, C.; Li, W. Enhancing intrusion detection with feature selection and neural network. Int. J. Intell. Syst. 2021, 36, 3087–3105. [Google Scholar] [CrossRef]
  3. Wang, X.; Liang, Z.; Koe, A.S.V.; Wu, Q.; Zhang, X.; Li, H.; Yang, Q. Secure and efficient parameters aggregation protocol for federated incremental learning and its applications. Int. J. Intell. Syst. 2021, 1–17. [Google Scholar] [CrossRef]
  4. Zhang, N.; Xue, J.; Ma, Y.; Zhang, R.; Liang, T.; Tan, Y.A. Hybrid sequence-based Android malware detection using natural language processing. Int. J. Intell. Syst. 2021, 36, 5770–5784. [Google Scholar] [CrossRef]
  5. Li, Y.; Yao, S.; Zhang, R.; Yang, C. Analyzing host security using D-S evidence theory and multisource information fusion. Int. J. Intell. Syst. 2021, 36, 1053–1068. [Google Scholar] [CrossRef]
  6. Wang, X.; Li, J.; Kuang, X.; Tan, Y.a.; Li, J. The security of machine learning in an adversarial setting: A survey. J. Parallel Distrib. Comput. 2019, 130, 12–23. [Google Scholar] [CrossRef]
  7. Mo, K.; Tang, W.; Li, J.; Yuan, X. Attacking Deep Reinforcement Learning with Decoupled Adversarial Policy. IEEE Trans. Dependable Secur. Comput. 2022. [Google Scholar] [CrossRef]
  8. Yan, H.; Hu, L.; Xiang, X.; Liu, Z.; Yuan, X. PPCL: Privacy-preserving collaborative learning for mitigating indirect information leakage. Inf. Sci. 2021, 548, 423–437. [Google Scholar] [CrossRef]
  9. Mo, K.; Liu, X.; Huang, T.; Yan, A. Querying little is enough: Model inversion attack via latent information. Int. J. Intell. Syst. 2021, 36, 681–690. [Google Scholar] [CrossRef]
  10. Ren, H.; Huang, T.; Yan, H. Adversarial examples: Attacks and defenses in the physical world. Int. J. Mach. Learn. Cybern. 2021, 12, 3325–3336. [Google Scholar] [CrossRef]
  11. Rao, Y.; Ni, J. Self-Supervised Domain Adaptation for Forgery Localization of JPEG Compressed Images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 15034–15043. [Google Scholar]
  12. Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. arXiv 2013, arXiv:1312.6199. [Google Scholar]
  13. Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. arXiv 2014, arXiv:1412.6572. [Google Scholar]
  14. Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 39–57. [Google Scholar]
  15. Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 26 June–1 July 2016; pp. 2574–2582. [Google Scholar]
  16. Guo, S.; Geng, S.; Xiang, T.; Liu, H.; Hou, R. ELAA: An efficient local adversarial attack using model interpreters. Int. J. Intell. Syst. 2021, 1–23. [Google Scholar] [CrossRef]
  17. Chen, H.; Lu, K.; Wang, X.; Li, J. Generating transferable adversarial examples based on perceptually-aligned perturbation. Int. J. Mach. Learn. Cybern. 2021, 12, 3295–3307. [Google Scholar] [CrossRef]
  18. Huang, T.; Chen, Y.; Yao, B.; Yang, B.; Wang, X.; Li, Y. Adversarial attacks on deep-learning-based radar range profile target recognition. Inf. Sci. 2020, 531, 159–176. [Google Scholar] [CrossRef]
  19. Huang, T.; Zhang, Q.; Liu, J.; Hou, R.; Wang, X.; Li, Y. Adversarial attacks on deep-learning-based SAR image target recognition. J. Netw. Comput. Appl. 2020, 162, 102632. [Google Scholar] [CrossRef]
  20. Chen, C.; Huang, T. Camdar-adv: Generating adversarial patches on 3D object. Int. J. Intell. Syst. 2021, 36, 1441–1453. [Google Scholar] [CrossRef]
  21. Shokri, R.; Stronati, M.; Song, C.; Shmatikov, V. Membership inference attacks against machine learning models. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017; pp. 3–18. [Google Scholar]
  22. Li, Y.; Tang, T.; Hsieh, C.J.; Lee, T. Detecting Adversarial Examples with Bayesian Neural Network. arXiv 2021, arXiv:2105.08620. [Google Scholar]
  23. Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. arXiv 2016, arXiv:1607.02533. [Google Scholar]
  24. Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. arXiv 2017, arXiv:1706.06083. [Google Scholar]
  25. Moosavi-Dezfooli, S.M.; Fawzi, A.; Fawzi, O.; Frossard, P. Universal adversarial perturbations. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1765–1773. [Google Scholar]
  26. Feinman, R.; Curtin, R.R.; Shintre, S.; Gardner, A.B. Detecting adversarial samples from artifacts. arXiv 2017, arXiv:1703.00410. [Google Scholar]
  27. Ma, X.; Li, B.; Wang, Y.; Erfani, S.M.; Wijewickrema, S.; Schoenebeck, G.; Song, D.; Houle, M.E.; Bailey, J. Characterizing adversarial subspaces using local intrinsic dimensionality. arXiv 2018, arXiv:1801.02613. [Google Scholar]
  28. Tian, S.; Yang, G.; Cai, Y. Detecting adversarial examples through image transformation. In Proceedings of the Thirty-Second AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018; pp. 4139–4146. [Google Scholar]
  29. Meng, D.; Chen, H. Magnet: A two-pronged defense against adversarial examples. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017; pp. 135–147. [Google Scholar]
  30. Lu, J.; Issaranon, T.; Forsyth, D. Safetynet: Detecting and rejecting adversarial examples robustly. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 446–454. [Google Scholar]
  31. Aggarwal, A.; Mittal, M.; Battineni, G. Generative adversarial network: An overview of theory and applications. Int. J. Inf. Manag. Data Insights 2021, 1, 100004. [Google Scholar] [CrossRef]
  32. Tang, W.; Li, B.; Barni, M.; Li, J.; Huang, J. An automatic cost learning framework for image steganography using deep reinforcement learning. IEEE Trans. Inf. Forensics Secur. 2020, 16, 952–967. [Google Scholar] [CrossRef]
  33. Li, H.; Luo, W.; Qiu, X.; Huang, J. Image forgery localization via integrating tampering possibility maps. IEEE Trans. Inf. Forensics Secur. 2017, 12, 1240–1252. [Google Scholar] [CrossRef]
  34. Qiu, X.; Li, H.; Luo, W.; Huang, J. A universal image forensic strategy based on steganalytic model. In Proceedings of the 2nd ACM Workshop on Information Hiding and Multimedia Security, New York, NY, USA, 14 July 2014; pp. 165–170. [Google Scholar]
  35. Marra, F.; Gragnaniello, D.; Cozzolino, D.; Verdoliva, L. Detection of gan-generated fake images over social networks. In Proceedings of the 2018 IEEE Conference on Multimedia Information Processing and Retrieval (MIPR), Miami, FL, USA, 10–12 April 2018; pp. 384–389. [Google Scholar]
  36. Nataraj, L.; Mohammed, T.M.; Manjunath, B.; Chandrasekaran, S.; Flenner, A.; Bappy, J.H.; Roy-Chowdhury, A.K. Detecting GAN generated fake images using co-occurrence matrices. Electron. Imaging 2019, 2019, 532-1–532-7. [Google Scholar] [CrossRef] [Green Version]
  37. Dang, H.; Liu, F.; Stehouwer, J.; Liu, X.; Jain, A.K. On the detection of digital face manipulation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 5781–5790. [Google Scholar]
  38. Zhang, X.; Karaman, S.; Chang, S.F. Detecting and simulating artifacts in gan fake images. In Proceedings of the 2019 IEEE International Workshop on Information Forensics and Security (WIFS), Delft, The Netherlands, 9–12 December 2019; pp. 1–6. [Google Scholar]
  39. Wang, S.Y.; Wang, O.; Zhang, R.; Owens, A.; Efros, A.A. CNN-generated images are surprisingly easy to spot… for now. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 14–19 June 2020; pp. 8695–8704. [Google Scholar]
  40. Arjovsky, M.; Chintala, S.; Bottou, L. Wasserstein generative adversarial networks. In Proceedings of the International Conference on Machine Learning, PMLR, Sydney, Australia, 6–11 August 2017; pp. 214–223. [Google Scholar]
  41. Gong, X.; Chang, S.; Jiang, Y.; Wang, Z. Autogan: Neural architecture search for generative adversarial networks. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Korea, 27 October–2 November 2019; pp. 3224–3234. [Google Scholar]
  42. Jiang, Y.; Chang, S.; Wang, Z. Transgan: Two pure transformers can make one strong gan, and that can scale up. In Proceedings of the Advances in Neural Information Processing Systems 34 (NeurIPS 2021), Virtual, 6–14 December 2021; Volume 34. [Google Scholar]
Figure 1. The difference of the distribution of model outputs between normal samples and adversarial examples (fake images).
Figure 1. The difference of the distribution of model outputs between normal samples and adversarial examples (fake images).
Sensors 22 03445 g001
Figure 2. The framework of our method.
Figure 2. The framework of our method.
Sensors 22 03445 g002
Figure 3. The generalizability for dataset.
Figure 3. The generalizability for dataset.
Sensors 22 03445 g003
Figure 4. The generalizability for target model architecture. (a,b) are the experimental results of ADV on CIFAR10 and CIFAR100. (c,d) are the experimental results of GBI on CIFAR10 and CIFAR100.
Figure 4. The generalizability for target model architecture. (a,b) are the experimental results of ADV on CIFAR10 and CIFAR100. (c,d) are the experimental results of GBI on CIFAR10 and CIFAR100.
Sensors 22 03445 g004
Figure 5. The result of transferability experiments. (a) is the experimental result on CIFAR10. (b) is the experimental result on CIFAR100.
Figure 5. The result of transferability experiments. (a) is the experimental result on CIFAR10. (b) is the experimental result on CIFAR100.
Sensors 22 03445 g005
Table 1. The results of adversarial example detection (The vertical axis and the horizontal axis represent training and evaluation, respectively).
Table 1. The results of adversarial example detection (The vertical axis and the horizontal axis represent training and evaluation, respectively).
CIFAR10
FGSMDeepFoolBIMPGDAutoPGDUPANewtonFoolZOOC&W
FGSM0.9050.9110.7960.7810.7570.8170.8730.8690.875
DeepFool0.8980.9160.7990.7840.7560.7960.8750.8660.871
BIM0.8840.8870.8330.8110.7760.8020.8710.8650.872
PGD0.8870.8900.8320.8130.8130.7800.8040.8760.865
AutoPGD0.8670.8770.7780.7690.7830.8710.8530.8460.862
UPA0.8600.8720.7440.7150.8800.8800.8350.8260.837
NewtonFool0.9030.9070.8170.8030.7740.8290.8890.8770.895
ZOO0.9020.9060.8200.8030.7740.8270.8880.8790.892
C&W0.9010.9050.8170.8000.7720.8410.8880.8760.896
CIFAR100
FGSMDeepFoolBIMPGDAutoPGDUPANewtonFoolZOOC&W
FGSM0.8820.9100.8690.8640.8750.9070.8800.8880.883
DeepFool0.8710.9220.8560.8550.8540.9100.8740.8700.871
BIM0.8790.9130.8740.8640.8770.9060.8890.8910.887
PGD0.8770.9030.8710.8730.8760.8940.8850.8900.890
AutoPGD0.8780.9080.8690.8640.8800.9020.8880.8880.887
UPA0.8650.8650.8430.8420.8480.9170.8550.8640.859
NewtonFool0.8730.9040.8690.8670.8670.8910.8890.8890.887
ZOO0.8800.9090.8700.8700.8770.9020.8850.8940.887
C&W0.8790.8790.8730.8710.8770.8950.8860.8910.892
Table 2. The results of fake image detection (The vertical axis and the horizontal axis represent training and evaluation, respectively).
Table 2. The results of fake image detection (The vertical axis and the horizontal axis represent training and evaluation, respectively).
CIFAR10
GANACGANWGANWGAN_GPWGAN_DIVDCGANAutoGANTransGAN
GAN0.7430.6800.7760.7890.7830.6800.5790.518
ACGAN0.4950.8550.7980.8920.8740.8270.6240.562
WGAN0.5480.7750.8300.8850.8750.7580.6020.555
WGAN_GP0.5050.8060.8080.9060.8860.7860.6040.545
WGAN_DIV0.5120.7870.8110.9000.8910.7800.5990.535
DCGAN0.5020.8440.7950.8780.8680.8390.6370.554
AutoGAN0.5510.7870.7780.8130.7960.7850.6470.585
TransGAN0.5790.7650.7610.7890.7640.7540.6380.586
CIFAR100
GANACGANWGANWGAN_GPWGAN_DIVDCGAN
GAN0.8220.8390.7400.8270.8420.825
ACGAN0.7160.8770.7110.8660.8770.809
WGAN0.7860.8370.7680.8450.8530.818
WGAN_GP0.7340.8690.7260.8710.8790.822
WGAN_DIV0.7390.8720.7260.8680.8850.824
DCGAN0.7620.8630.7470.8660.8770.847
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Lai, J.; Huo, Y.; Hou, R.; Wang, X. A Universal Detection Method for Adversarial Examples and Fake Images. Sensors 2022, 22, 3445. https://doi.org/10.3390/s22093445

AMA Style

Lai J, Huo Y, Hou R, Wang X. A Universal Detection Method for Adversarial Examples and Fake Images. Sensors. 2022; 22(9):3445. https://doi.org/10.3390/s22093445

Chicago/Turabian Style

Lai, Jiewei, Yantong Huo, Ruitao Hou, and Xianmin Wang. 2022. "A Universal Detection Method for Adversarial Examples and Fake Images" Sensors 22, no. 9: 3445. https://doi.org/10.3390/s22093445

APA Style

Lai, J., Huo, Y., Hou, R., & Wang, X. (2022). A Universal Detection Method for Adversarial Examples and Fake Images. Sensors, 22(9), 3445. https://doi.org/10.3390/s22093445

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop