DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention
Abstract
:1. Introduction
- (1)
- We propose an adversarial attack named DIPA, which is based on an attention mechanism and high-frequency information separation. The adversarial perturbation is generated by attacking the attention of neural network models combined with the separation of high-frequency information from image samples. Simultaneously, we employ the norm to limit adversarial perturbation, thereby reinforcing a high attack success rate and effectively concealing it.
- (2)
- We conduct extensive experiments on the ImageNet dataset, setting up two different scenarios. The experiment results show that, compared with the existing AoA method and one-pixel attack method, our method achieves better results on several evaluating metrics.
- (3)
- Finally, we use visualization and quantification to analyze the concealing effect of adversarial disturbance and compare it with many traditional adversarial attack methods, which verify that our method can generate imperceptible adversarial perturbation.
2. Related Work
3. Methodology
3.1. High-Frequency Information Separation Based on Discrete Wavelet Transform
3.2. Designing Attention Suppression Loss Function
3.3. Pixel-Level Attack Algorithm Based on Attention Mechanism
Algorithm 1 Adversarial Attack. |
|
4. Experiments
4.1. Experiment Setup
4.2. Evaluation Metrics
- Average root mean square error ():In the experiment, the deep neural network models generate multiple adversarial samples, which we evaluate using the average root mean square error () to determine the degree of change. Where represents the number of adversarial samples generated by our method, and represents the total number of pixels in the image sample. The metric is shown in Equation (13).
- Attack success rate ():Following the generation of adversarial samples, we input them into the model to gauge their capacity to deceive. Where represents the number of adversarial samples generated by our method that successfully attacks the target model, and variable N denotes the total number of image samples. The metric is shown in Equation (14).
- Average confidence ():
- Time complexity ():
- Number of disturbed pixels ():
- Learning-based perceptual similarity metrics (Lpips): Currently, the most common method to measure the similarity between two image samples is based on distance, such as SSIM [32] and FSIM [33] based on Euclidean distance, etc. These methods use a simple distance function to calculate the similarity directly. However, humans can easily and quickly assess the perceptual similarity between two images, but the process is highly complex. The method based on distance measurement does not consider the details of human perception and cannot fit well with the human perceptual similarity between two images. Therefore, we choose a learn-based perceptual similarity measurement (Lpips) [34] method that aligns with human perception judgment. The Lpips method visualizes the degree of change between the adversarial sample and the original sample by utilizing a perceptual distance space map. It also quantifies the concealing effect of adversarial perturbation using the Lpips perceptual loss metric. Thus, the advantages and characteristics of the DIPA method can be verified more reasonably.
4.3. Results Analysis in Single-Pixel Attack Scenario
4.4. Results Analysis in the White-Box Attack Scenario
4.5. Results Analysis in Perceptual Quality
4.5.1. Perceptual Quality Visualization Experiment
4.5.2. Comparison with Traditional Adversarial Attack Methods
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- Hinton, G.E.; Osindero, S.; The, Y. A fast learning algorithm for deep belief nets. Neural Comput. 2006, 18, 1527–1554. [Google Scholar] [CrossRef] [PubMed]
- Taigman, Y.; Yang, M.; Ranzato, M.; Wolf, L. Deepface: Closing the gap to human-level performance in face verification. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Columbus, OH, USA, 23–28 June 2014; pp. 1701–1708. [Google Scholar]
- Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. In Proceedings of the International Conference on Learning Representations (ICLR), Banff, AB, Canada, 14–16 April 2014; pp. 1–9. [Google Scholar]
- Moosavi-Dezfooli, S.M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2574–2582. [Google Scholar]
- Whitaker, T.A.; Simões-Franklin, C.; Newell, F.N. Vision and touch: Independent or integrated systems for the perception of texture? Brain Res. 2008, 1242, 59–72. [Google Scholar] [CrossRef] [PubMed]
- Barreno, M.; Nelson, B.; Joseph, A.D.; Tygar, J.D. The security of machine learning. Achine Learn. 2010, 81, 121–148. [Google Scholar] [CrossRef]
- Goodfellow, I.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 676–681. [Google Scholar]
- Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017; pp. 1–11. [Google Scholar]
- Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the IEEE Symposium on Security and Privacy (S&P), San Jose, CA, USA, 22–26 May 2017; pp. 39–57. [Google Scholar]
- Cheng, S.Y.; Miao, Y.B.; Dong, Y.P.; Yang, X.; Gao, X.S.; Zhu, J. Efficient Black-box Adversarial Attacks via Bayesian Optimization Guided by a Function Prior. In Proceedings of the International Conference on Machine Learning (ICML), Vienna, Austria, 21–27 July 2024; pp. 1–21. [Google Scholar]
- Duan, R.; Chen, Y.; Niu, D.; Yang, Y.; Qin, A.K.; He, Y. AdvDrop: Adversarial attack to DNNs by dropping information. In Proceedings of the IEEE/CVF Conference on International Conference on Computer Visual (ICCV), Montreal, BC, Canada, 11–17 October 2021; pp. 7486–7495. [Google Scholar]
- Liu, D.; Su, Q.; Peng, C.; Wang, N.; Gao, X. Imperceptible Face Forgery Attack via Adversarial Semantic Mask. arXiv 2024, arXiv:2406.10887. [Google Scholar]
- Ilyas, A.; Engstrom, L.; Madry, A. Prior convictions: Black-box adversarial attacks with bandits and priors. In Proceedings of the International Conference on Learning Representations (ICLR), New Orleans, LA, USA, 6–9 May 2019; pp. 1–13. [Google Scholar]
- Dong, Y.; Pang, T.; Su, H.; Zhu, J. Evading defenses to transferable adversarial examples by translation-invariant attacks. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Long Beach, CA, USA, 16–20 June 2019; pp. 4312–4321. [Google Scholar]
- Chen, S.; He, Z.; Sun, C.; Huang, X. Universal adversarial attack on attention and the resulting dataset damagenet. IEEE Trans. Pattern Anal. Mach. Intell. 2020, 44, 2188–2197. [Google Scholar] [CrossRef]
- Huang, L.F.; Zhuang, W.Z.; Liao, Y.X.; Liu, N. Black-box Adversarial Attack Method Based on Evolution Strategy and Attention Mechanism. J. Softw. 2021, 32, 3512–3529. [Google Scholar]
- Duan, M.; Qin, Y.; Deng, J.; Li, K.; Xiao, B. Dual Attention Adversarial Attacks with Limited Perturbations. IEEE Trans. Neural Netw. Learn. Syst. 2023; early access. [Google Scholar] [CrossRef] [PubMed]
- Lin, C.; Han, S.; Zhu, J.; Li, Q.; Shen, C.; Zhang, Y.; Guan, X. Sensitive region-aware black-box adversarial attacks. Inf. Sci. 2023, 637, 118929. [Google Scholar] [CrossRef]
- Liu, H.; Zhang, Z.H.; Xia, X.F.; Gao, T.G. A fast black box boundary attack algorithm based on geometric detection. J. Comput. Res. Dev. 2023, 60, 435–447. [Google Scholar]
- Su, J.; Vargas, D.V.; Kouichi, S. One pixel attack for fooling deep neural networks. IEEE Trans. Evol. Comput. 2019, 23, 828–841. [Google Scholar] [CrossRef]
- Papernot, N.; Mcdaniel, P.; Jha, S.; Fredrikson, M.; Swami, A.; Celik, Z.B. The limitations of deep learning in adversarial settings. In Proceedings of the IEEE European Symposium on Security and Privacy (EuroS&P), Saarbrucken, Germany, 21–24 March 2016; pp. 372–387. [Google Scholar]
- Combey, T.; Loison, A.; Faucher, M.; Hajri, H. Probabilistic jacobian-based saliency maps attacks. Mach. Learn. Knowl. Extr. 2020, 2, 558–578. [Google Scholar] [CrossRef]
- Liu, H.L.; Liu, J. PlAA: Pixel-level Adversarial Attack on Attention for Deep Neural Network. In Proceedings of the International Conference on Artificial Neural Networks (ICANN), Bristol, UK, 6–9 September 2022; pp. 611–623. [Google Scholar]
- Williams, P.N.; Li, K. Black-box sparse adversarial attack via multi-objective optimisation CVPR proceedings. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Vancouver, BC, Canada, 18–22 June 2023; pp. 12291–12301. [Google Scholar]
- Iwana, B.K.; Kuroki, R.; Uchida, S. Explaining convolutional neural networks using softmax gradient layer-wise relevance propagation. In Proceedings of the International Conference on Computer Vision Workshop (ICCVW), Seoul, Republic of Korea, 27–28 October 2019; pp. 4176–4185. [Google Scholar]
- Tay, Y.; Dehghani, M.; Gupta, J. Are Pretrained Convolutions Better than Pretrained Transformers? In Proceedings of the Annual Meeting of the Association for Computational Linguistics(ACL), Bangkok, Thailand, 1–6 August 2021; pp. 4349–4359. [Google Scholar]
- Jia, D.; Wei, D.; Socher, R.; Li, L.J.; Kai, L.; Li, F.F. Imagenet: A large-scale hierarchical image database. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Miami, FL, USA, 20–25 June 2009; pp. 248–255. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very deep convolutional networks for large-scale image recognition. In Proceedings of the International Conference on Learning Representations (ICLR), San Diego, CA, USA, 7–9 May 2015; pp. 398–406. [Google Scholar]
- Szegedy, C.; Vanhoucke, V.; Ioffe, S.; Shlens, J.; Wojna, Z. Rethinking the inception architecture for computer visual. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 2818–2826. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Las Vegas, NV, USA, 27–30 June 2016; pp. 770–778. [Google Scholar]
- Gao, H.; Zhuang, L.; Kilian, Q.W. Densely connected convolutional networks. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Honolulu, HI, USA, 21–26 July 2017; pp. 2261–2269. [Google Scholar]
- Wang, Z.; Bovik, A.C.; Sheikh, H.R.; Simoncelli, E.P. Image quality assessment. IEEE Trans. Image Process. 2004, 13, 600–612. [Google Scholar] [CrossRef]
- Zhang, L.; Zhang, L.; Mou, X.; Zhang, D. Fsim: A feature similarity index for image quality assessment. IEEE Trans. Image Process. 2011, 20, 2378–2386. [Google Scholar] [CrossRef] [PubMed]
- Zhang, R.; Isola, P.; Efros, A.A.; Shechtman, E.; Wang, O. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Visual and Pattern Recognition (CVPR), Salt Lake City, UT, USA, 18–22 June 2018; pp. 586–595. [Google Scholar]
- Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards deep learning models resistant to adversarial attacks. In Proceedings of the International Conference on Learning Representations (ICLR), Vancouver, BC, Canada, 30 April–3 May 2018; pp. 39–57. [Google Scholar]
- Croce, F.; Hein, M. Reliable evaluation of adversarial robustness with an ensemble of diverse parameter-free attacks. In Proceedings of the IEEE Conference on International Conference on Machine Learning (ICML), Vienna, Austria, 12–18 July 2020; pp. 2206–2216. [Google Scholar]
Metrics Victim | AlexNet_BVLC |
---|---|
ASR | 16.04% |
AvgConfidence | 22.91 |
AvgRMSE | 14.32 |
AvgTime (s) | - |
Metrics Victim | AlexNet_BVLC | VGG19 | VGG16 | IncV3 | RN50 | RN152 | DN121 | DN169 | DN201 |
---|---|---|---|---|---|---|---|---|---|
ASR | 43.2% | 50.2% | 43.5% | 56.8% | 16.7% | 23.3% | 13.1% | 12.4% | 15.2% |
AvgConfidence | 69.552 | 68.987 | 69.650 | 20.833 | 45.710 | 55.465 | 45.461 | 40.979 | 49.898 |
AvgRMSE | 16.437 | 16.534 | 16.310 | 12.433 | 18.434 | 18.839 | 18.116 | 18.629 | 18.977 |
AvgTime (s) | 1.4 | 2.4 | 1.5 | 6.0 | 6.4 | 14.8 | 35.7 | 44.6 | 48.1 |
Metrics Victim | VGG19 | VGG16 | IncV3 | RN50 | RN152 | DN121 | DN169 | DN201 |
---|---|---|---|---|---|---|---|---|
AoA | 99.99% | 99.85% | 89.84% | 93.94% | 86.78% | 96.14% | 94.09% | 93.44% |
DIPA | 98.8% | 98.5% | 96.4% | 97.6% | 95.6% | 84.7% | 88.4% | 93.2% |
Metrics Victim | VGG19 | VGG16 | IncV3 | RN50 | RN152 | DN121 | DN169 | DN201 |
---|---|---|---|---|---|---|---|---|
AvgPixels | 2.415 | 2.561 | 2.318 | 5.231 | 4.448 | 8.372 | 8.415 | 6.483 |
AvgConfidence | 69.940 | 70.260 | 21.054 | 43.977 | 58.185 | 43.497 | 47.696 | 43.910 |
AvgRMSE | 20.143 | 20.158 | 18.742 | 28.970 | 30.744 | 29.907 | 32.703 | 25.621 |
AvgTime (s) | 3.3 | 3.5 | 8.2 | 18.9 | 32.1 | 38.5 | 54.6 | 52.5 |
Model | Acc Top1 | Acc Top5 | Params |
---|---|---|---|
VGG16 | 73.360% | 91.516% | 138.4 M |
VGG19 | 74.218% | 91.842% | 143.7 M |
ResNet50 | 81.198% | 95.340% | 25.0 M |
ResNet152 | 82.284% | 96.002% | 60.2 M |
DenseNet121 | 74.434% | 91.972% | 8.0 M |
DenseNet169 | 75.600% | 92.806% | 14.1 M |
DenseNet201 | 76.896% | 93.37% | 20.0 M |
Model | Attack | Iteration | Run Time (s) | ASR | Lpips |
---|---|---|---|---|---|
VGG16 | FGSM | 1 | 22 | 96.2% | 0.306 |
BIM | 10 | 243 | 94.0% | 0.089 | |
PGD | 10 | 56 | 98.3% | 0.129 | |
C&W | 1000 | ≥10,000 | 99.2% | 0.394 | |
AA | 100 | 62 | 99.5% | 0.394 | |
Ours | 20 | 292 | 98.5% | 0.060 | |
VGG19 | FGSM | 1 | 24 | 94.1% | 0.301 |
BIM | 10 | 255 | 94.3% | 0.091 | |
PGD | 10 | 66 | 97.9% | 0.128 | |
C&W | 1000 | ≥10,000 | 99.3% | 0.395 | |
AA | 100 | 72 | 99.2% | 0.395 | |
Ours | 20 | 300 | 98.8% | 0.062 | |
ResNet50 | FGSM | 1 | 85 | 94.1% | 0.304 |
BIM | 10 | 248 | 92.6% | 0.091 | |
PGD | 10 | 187 | 99.9% | 0.107 | |
C&W | 1000 | ≥10,000 | 98.8% | 0.390 | |
AA | 100 | 74 | 96.8% | 0.396 | |
Ours | 20 | 1405 | 97.6% | 0.063 | |
ResNet152 | FGSM | 1 | 67 | 94.2% | 0.308 |
BIM | 10 | 320 | 91.7% | 0.096 | |
PGD | 10 | 257 | 99.7% | 0.126 | |
C&W | 1000 | ≥10,000 | 97.3% | 0.397 | |
AA | 100 | 192 | 97.1% | 0.396 | |
Ours | 20 | ≥1000 | 95.6% | 0.068 | |
DenseNet121 | FGSM | 1 | 63 | 98.5% | 0.307 |
BIM | 10 | 264 | 92.4% | 0.094 | |
PGD | 10 | 167 | 99.4% | 0.12 | |
C&W | 1000 | ≥10,000 | 98.3% | 0.402 | |
AA | 100 | 126 | 97.3% | 0.387 | |
Ours | 20 | ≥5000 | 90.7% | 0.070 | |
DenseNet169 | FGSM | 1 | 75 | 96.7% | 0.308 |
BIM | 10 | 229 | 90.8% | 0.092 | |
PGD | 10 | 231 | 99.7% | 0.126 | |
C&W | 1000 | ≥10,000 | 96.2% | 0.407 | |
AA | 100 | 182 | 97.1% | 0.384 | |
Ours | 20 | ≥5000 | 89.4% | 0.068 | |
DenseNet201 | FGSM | 1 | 131 | 94.0% | 0.304 |
BIM | 10 | 330 | 93.8% | 0.092 | |
PGD | 10 | 273 | 99.4 | 0.126 | |
C&W | 1000 | ≥10,000 | 95.4 | 0.408 | |
AA | 100 | 246 | 98.5 | 0.389 | |
Ours | 20 | ≥5000 | 93.2 | 0.069 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2024 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Liu, J.; Liu, H.; Wang, P.; Wu, Y.; Li, K. DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention. Information 2024, 15, 391. https://doi.org/10.3390/info15070391
Liu J, Liu H, Wang P, Wu Y, Li K. DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention. Information. 2024; 15(7):391. https://doi.org/10.3390/info15070391
Chicago/Turabian StyleLiu, Jing, Huailin Liu, Pengju Wang, Yang Wu, and Keqin Li. 2024. "DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention" Information 15, no. 7: 391. https://doi.org/10.3390/info15070391
APA StyleLiu, J., Liu, H., Wang, P., Wu, Y., & Li, K. (2024). DIPA: Adversarial Attack on DNNs by Dropping Information and Pixel-Level Attack on Attention. Information, 15(7), 391. https://doi.org/10.3390/info15070391