Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification
Abstract
:1. Introduction
- We introduce the Label-Feature Distribution Mismatch property of adversarial examples and point out that the Label-Feature Distribution Mismatch property is a reason that causes poor adversarial robustness generalization performance of DNNs against adversarial examples;
- We propose two novel adversarial defense algorithms named BCAT and BCAT+ that train DNNs to output the mixing ratio of two adversarial examples with different real labels, which impose effective regularization on the feature distribution of adversarial examples;
- We design extensive experiments to evaluate the proposed BCAT and BCAT+ algorithms under both white-box and black-box attacks on CIFAR-10, CIFAR-100, and SVHN datasets. The experimental results show that BCAT and BCAT+ achieve better global adversarial robustness generalization performance than the state-of-the-art adversarial defense methods.
2. Related Works
2.1. Adversarial Training
2.2. Regularization
3. Methods
3.1. Label-Feature Distribution Mismatch
3.2. Motivation
3.3. BCAT: Between-Class Adversarial Training
Algorithm 1 Pseudo code of BCAT | |
Input: | Dataset , initial weight parameters , training steps , batch size , PGD perturbation value , PGD step size , PGD number of steps |
Output: | |
1 | do |
2 | with corresponding examples having different class label |
3 | do |
4 | do |
5 | |
6 | |
7 | End for |
8 | End for |
9 | |
10 | do |
11 | |
12 | |
13 | End for |
14 | |
15 | End for |
3.4. BCAT+: A More Powerful Mixing Method
Algorithm 2 Pseudo code of BCAT+ | |
Input: | Dataset , initial weight parameters , training steps , batch size , PGD perturbation value , PGD step size , PGD number of steps |
Output: | weight parameters |
1 | do |
2 | with corresponding examples having different class labels |
3 | do |
4 | do |
5 | |
6 | |
7 | End for |
8 | End for |
9 | |
10 | |
11 | do |
12 | |
13 | |
14 | End for |
15 | |
16 | End for |
3.5. Real Label and Fake Label
4. Results and Discussion
4.1. Datasets
4.2. Threat Model
4.3. Training Parameters
4.4. Evaluation under White-Box Attacks
4.4.1. Feature Distribution
4.4.2. Robustness Generalization
4.4.3. Convergence Analysis
4.5. Evaluation under Black-Box Attacks
4.6. Ablation Study
4.6.1. BCAT (+) and BCATf (+)
4.6.2. Ablation on Data Augmentation
4.6.3. Ablation on Attack Steps
5. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef] [Green Version]
- Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Long, J.; Shelhamer, E.; Darrell, T. Fully convolutional networks for semantic segmentation. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–12 June 2015. [Google Scholar]
- Szegedy, C.; Zaremba, W.; Sutskever, I.; Bruna, J.; Erhan, D.; Goodfellow, I.; Fergus, R. Intriguing properties of neural networks. In Proceedings of the 2nd International Conference on Learning Representations, Banff, AB, Canada, 14–16 April 2014. [Google Scholar]
- Goodfellow, I.J.; Shlens, J.; Szegedy, C. Explaining and harnessing adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Eykholt, K.; Evtimov, I.; Fernandes, E.; Li, B.; Rahmati, A.; Xiao, C.; Prakash, A.; Kohno, T.; Song, D. Robust physical-world attacks on deep learning visual classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018. [Google Scholar]
- Xu, K.; Zhang, G.; Liu, S.; Fan, Q.; Sun, M.; Chen, H.; Chen, P.-Y.; Wang, Y.; Lin, X. Adversarial t-shirt! evading person detectors in a physical world. In Proceedings of the European Conference on Computer Vision, Glasgow, UK, 23–28 August 2020. [Google Scholar]
- Li, Z.; Dong, M.; Wen, S.; Hu, X.; Zhou, P.; Zeng, Z. CLU-CNNs: Object detection for medical images. Neurocomputing 2019, 350, 53–59. [Google Scholar] [CrossRef]
- Madry, A.; Makelov, A.; Schmidt, L.; Tsipras, D.; Vladu, A. Towards Deep Learning Models Resistant to Adversarial Attacks. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Kurakin, A.; Goodfellow, I.; Bengio, S. Adversarial examples in the physical world. In Proceedings of the 5th International Conference on Learning Representations, Toulon, France, 24–26 April 2017. [Google Scholar]
- Papernot, N.; McDaniel, P.; Jha, S.; Fredrikson, M.; Celik, Z.B.; Swami, A. The limitations of deep learning in adversarial settings. In Proceedings of the 2016 IEEE European Symposium on Security and Privacy (EuroS&P), Saarbruecken, Germany, 21–24 March 2016. [Google Scholar]
- Su, J.; Vargas, D.V.; Sakurai, K. One pixel attack for fooling deep neural networks. IEEE Trans. Evol. Comput. 2019, 23, 828–841. [Google Scholar] [CrossRef] [Green Version]
- Carlini, N.; Wagner, D. Towards evaluating the robustness of neural networks. In Proceedings of the 2017 IEEE Symposium on Security and Privacy (SP), San Jose, CA, USA, 22–26 May 2017. [Google Scholar]
- Moosavi-Dezfooli, S.-M.; Fawzi, A.; Frossard, P. Deepfool: A simple and accurate method to fool deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Chen, P.-Y.; Sharma, Y.; Zhang, H.; Yi, J.; Hsieh, C.-J. Ead: Elastic-net attacks to deep neural networks via adversarial examples. In Proceedings of the AAAI Conference on Artificial Intelligence, New Orleans, LA, USA, 2–7 February 2018. [Google Scholar]
- Papernot, N.; McDaniel, P.; Goodfellow, I.; Jha, S.; Celik, Z.B.; Swami, A. Practical black-box attacks against machine learning. In Proceedings of the 2017 ACM on Asia Conference on Computer and Communications Security, Abu Dhabi, United Arab Emirates, 2–6 April 2017. [Google Scholar]
- Xu, W.; Evans, D.; Qi, Y. Feature squeezing: Detecting adversarial examples in deep neural networks. In Proceedings of the 25th Annual Network and Distributed System Security Symposium, San Diego, CA, USA, 18–21 February 2018. [Google Scholar]
- Dhillon, G.S.; Azizzadenesheli, K.; Lipton, Z.C.; Bernstein, J.; Kossaifi, J.; Khanna, A.; Anandkumar, A. Stochastic activation pruning for robust adversarial defense. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Gu, S.; Rigazio, L. Towards deep neural network architectures robust to adversarial examples. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Samangouei, P.; Kabkab, M.; Chellappa, R. Defense-gan: Protecting classifiers against adversarial attacks using generative models. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Song, Y.; Kim, T.; Nowozin, S.; Ermon, S.; Kushman, N. Pixeldefend: Leveraging generative models to understand and defend against adversarial examples. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Meng, D.; Chen, H. Magnet: A two-pronged defense against adversarial examples. In Proceedings of the 2017 ACM SIGSAC Conference on Computer and Communications Security, Dallas, TX, USA, 30 October–3 November 2017. [Google Scholar]
- Athalye, A.; Carlini, N.; Wagner, D. Obfuscated gradients give a false sense of security: Circumventing defenses to adversarial examples. In Proceedings of the International Conference on Machine Learning, Stockholm, Sweden, 10–15 July 2018. [Google Scholar]
- Carlini, N.; Wagner, D. Magnet and “efficient defenses against adversarial attacks” are not robust to adversarial examples. arXiv 2017, arXiv:1711.08478. [Google Scholar]
- Schmidt, L.; Santurkar, S.; Tsipras, D.; Talwar, K.; Mądry, A. Adversarially robust generalization requires more data. In Proceedings of the Advances in Neural Information Processing Systems 31: Annual Conference on Neural Information Processing Systems, Montréal, QC, Canada, 3–8 December 2018. [Google Scholar]
- Krogh, A.; Hertz, J.A. A simple weight decay can improve generalization. In Proceedings of the 4th International Conference on Neural Information Processing Systems, Denver, CO, USA, 2–5 December 1991. [Google Scholar]
- Srivastava, N.; Hinton, G.; Krizhevsky, A.; Sutskever, I.; Salakhutdinov, R. Dropout: A simple way to prevent neural networks from overfitting. J. Mach. Learn. Res. 2014, 15, 1929–1958. [Google Scholar]
- Simonyan, K.; Zisserman, A. Very Deep Convolutional Networks for Large-Scale Image Recognition. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Sandfort, V.; Yan, K.; Pickhardt, P.J.; Summers, R.M. Data augmentation using generative adversarial networks (CycleGAN) to improve generalizability in CT segmentation tasks. Sci. Rep. 2019, 9, 1–9. [Google Scholar] [CrossRef] [PubMed] [Green Version]
- Tokozume, Y.; Ushiku, Y.; Harada, T. Learning from Between-class Examples for Deep Sound Recognition. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Tokozume, Y.; Ushiku, Y.; Harada, T. Between-class learning for image classification. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Computer Vision Foundation/IEEE Computer Society, Salt Lake City, UT, USA, 18–22 June 2018. [Google Scholar]
- Kannan, H.; Kurakin, A.; Goodfellow, I. Adversarial logit pairing. arXiv 2018, arXiv:1803.06373. [Google Scholar]
- Zhang, H.; Yu, Y.; Jiao, J.; Xing, E.P.; Ghaoui, L.E.; Jordan, M.I. Theoretically Principled Trade-off between Robustness and Accuracy. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
- Mao, C.; Zhong, Z.; Yang, J.; Vondrick, C.; Ray, B. Metric Learning for Adversarial Robustness. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
- Li, P.; Yi, J.; Zhou, B.; Zhang, L. Improving the Robustness of Deep Neural Networks via Adversarial Training with Triplet Loss. In Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, Macao, China, 10–16 August 2019. [Google Scholar]
- Zhang, H.; Wang, J. Defense against adversarial attacks using feature scattering-based adversarial training. Adv. Neural Inf. Process. Syst. 2019, 32, 1831–1841. [Google Scholar]
- Yu, Y.; Gao, X.; Xu, C.-Z. LAFEAT: Piercing Through Adversarial Defenses with Latent Features. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Computer Vision Foundation, Virtual, 19–25 June 2021. [Google Scholar]
- Chen, K.; Chen, Y.; Zhou, H.; Mao, X.; Li, Y.; He, Y.; Xue, H.; Zhang, W.; Yu, N. Self-supervised adversarial training. In Proceedings of the ICASSP 2020—2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), Barcelona, Spain, 4–8 May 2020. [Google Scholar]
- Liu, X.; Li, Y.; Wu, C.; Hsieh, C.-J. Adv-BNN: Improved Adversarial Defense through Robust Bayesian Neural Network. In Proceedings of the 7th International Conference on Learning Representations, New Orleans, LA, USA, 6–9 May 2019. [Google Scholar]
- Wang, Y.; Ma, X.; Bailey, J.; Yi, J.; Zhou, B.; Gu, Q. On the Convergence and Robustness of Adversarial Training. In Proceedings of the 36th International Conference on Machine Learning, Long Beach, CA, USA, 9–15 June 2019. [Google Scholar]
- Rice, L.; Wong, E.; Kolter, Z. Overfitting in adversarially robust deep learning. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020. [Google Scholar]
- Yu, X.; Smedemark-Margulies, N.; Aeron, S.; Koike-Akino, T.; Moulin, P.; Brand, M.; Parsons, K.; Wang, Y. Improving adversarial robustness by learning shared information. Pattern Recognit. 2023, 134, 109054. [Google Scholar] [CrossRef]
- Zhang, J.; Xu, X.; Han, B.; Niu, G.; Cui, L.; Sugiyama, M.; Kankanhalli, M. Attacks which do not kill training make adversarial learning stronger. In Proceedings of the 37th International Conference on Machine Learning, Virtual Event, 13–18 July 2020. [Google Scholar]
- Ye, S.; Xu, K.; Liu, S.; Cheng, H.; Lambrechts, J.-H.; Zhang, H.; Zhou, A.; Ma, K.; Wang, Y.; Lin, X. Adversarial Robustness vs. Model Compression, or Both? In Proceedings of the IEEE/CVF International Conference on Computer Vision, Seoul, Republic of Korea, 27 October—2 November 2019. [Google Scholar]
- Sehwag, V.; Wang, S.; Mittal, P.; Jana, S. HYDRA: Pruning Adversarially Robust Neural Networks. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
- Shafahi, A.; Najibi, M.; Ghiasi, A.; Xu, Z.; Dickerson, J.P.; Studer, C.; Davis, L.S.; Taylor, G.; Goldstein, T. Adversarial training for free! In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems 2019, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
- Zhang, D.; Zhang, T.; Lu, Y.; Zhu, Z.; Dong, B. You Only Propagate Once: Accelerating Adversarial Training via Maximal Principle. In Proceedings of the Advances in Neural Information Processing Systems 32: Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019. [Google Scholar]
- Wong, E.; Rice, L.; Kolter, J.Z. Fast is better than free: Revisiting adversarial training. In Proceedings of the 8th International Conference on Learning Representations, Addis Ababa, Ethiopia, 26–30 April 2020. [Google Scholar]
- Vivek, B.; Babu, R.V. Single-step adversarial training with dropout scheduling. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), Computer Vision Foundation, Seattle, WA, USA, 13–19 June 2020. [Google Scholar]
- Wu, T.; Liu, Z.; Huang, Q.; Wang, Y.; Lin, D. Adversarial Robustness under Long-Tailed Distribution. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Computer Vision Foundation, Virtual, 19–25 June 2021. [Google Scholar]
- Ng, A.Y. Feature selection, L 1 vs. L 2 regularization, and rotational invariance. In Proceedings of the Twenty-First International Conference on Machine Learning, Association for Computing Machinery, Banff, AB, Canada, 4–8 July 2004. [Google Scholar]
- Xu, Y.; Zhong, Z.; Yang, J.; You, J.; Zhang, D. A new discriminative sparse representation method for robust face recognition via L2 regularization. IEEE Trans. Neural Netw. Learn. Syst. 2016, 28, 2233–2242. [Google Scholar] [CrossRef] [PubMed]
- Caruana, R.; Lawrence, S.; Giles, L. Overfitting in neural nets: Backpropagation, conjugate gradient, and early stopping. Adv. Neural Inf. Process. Syst. 2000, 13, 402–408. [Google Scholar]
- Chen, T.; Kornblith, S.; Swersky, K.; Norouzi, M.; Hinton, G.E. Big Self-Supervised Models are Strong Semi-Supervised Learners. In Proceedings of the Advances in Neural Information Processing Systems 33: Annual Conference on Neural Information Processing Systems, Virtual, 6–12 December 2020. [Google Scholar]
- Ndirango, A.; Lee, T. Generalization in multitask deep neural classifiers: A statistical physics approach. In Advances in Neural Information Processing Systems 32, Proceedings of the Annual Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 8–14 December 2019; Curran Associates Inc.: Red Hook, NY, USA, 2019. [Google Scholar]
- Noh, H.; You, T.; Mun, J.; Han, B. Regularizing Deep Neural Networks by Noise: Its Interpretation and Optimization. In Proceedings of the Advances in Neural Information Processing Systems 30: Annual Conference on Neural Information Processing Systems, Long Beach, CA, USA, 4–9 December 2017. [Google Scholar]
- He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016. [Google Scholar]
- Pang, T.; Yang, X.; Dong, Y.; Su, H.; Zhu, J. Bag of Tricks for Adversarial Training. In Proceedings of the 9th International Conference on Learning Representations, Virtual, 3–7 May 2021. [Google Scholar]
- Göpfert, C.; Göpfert, J.P.; Hammer, B. Adversarial Robustness Curves. In Proceedings of the Joint European Conference on Machine Learning and Knowledge Discovery in Databases, Würzburg, Germany, 16–20 September 2019. [Google Scholar]
Network | Defense | 0 | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|---|
ResNet34 | AT | 8/255 | 0.863 | 0.809 | 0.739 | 0.666 | 0.602 |
freeAT | 8/255 | 0.838 | 0.786 | 0.730 | 0.658 | 0.597 | |
ALP | 8/255 | 0.872 | 0.815 | 0.748 | 0.673 | 0.600 | |
TRADES | 8/255 | 0.813 | 0.774 | 0.720 | 0.671 | 0.623 | |
ATLSI | 8/255 | 0.879 | 0.818 | 0.739 | 0.654 | 0.586 | |
BCAT | 8/255 | 0.863 | 0.814 | 0.752 | 0.682 | 0.617 | |
BCAT+ | 8/255 | 0.863 | 0.817 | 0.761 | 0.691 | 0.628 | |
ResNet18 | AT | 8/255 | 0.850 | 0.799 | 0.739 | 0.673 | 0.604 |
freeAT | 8/255 | 0.823 | 0.770 | 0.712 | 0.644 | 0.582 | |
ALP | 8/255 | 0.862 | 0.805 | 0.738 | 0.663 | 0.592 | |
TRADES | 8/255 | 0.789 | 0.744 | 0.701 | 0.649 | 0.607 | |
ATLSI | 8/255 | 0.872 | 0.815 | 0.739 | 0.650 | 0.570 | |
BCAT | 8/255 | 0.841 | 0.794 | 0.735 | 0.671 | 0.610 | |
BCAT+ | 8/255 | 0.841 | 0.795 | 0.741 | 0.678 | 0.611 |
Network | Defense | 0 | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|---|
ResNet34-5 | AT | 8/255 | 0.644 | 0.565 | 0.488 | 0.427 | 0.395 |
freeAT | 8/255 | 0.628 | 0.559 | 0.482 | 0.418 | 0.360 | |
ALP | 8/255 | 0.642 | 0.658 | 0.499 | 0.436 | 0.388 | |
TRADES | 8/255 | 0.583 | 0.513 | 0.447 | 0.388 | 0.349 | |
ATLSI | 8/255 | 0.659 | 0.574 | 0.496 | 0.430 | 0.400 | |
BCAT | 8/255 | 0.675 | 0.612 | 0.548 | 0.474 | 0.411 | |
BCAT+ | 8/255 | 0.678 | 0.614 | 0.544 | 0.480 | 0.418 |
Network | Defense | 0 | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|---|
ResNet34 | AT | 12/255 | 0.925 | 0.881 | 0.827 | 0.766 | 0.718 |
freeAT | 12/255 | 0.927 | 0.470 | 0.339 | 0.282 | 0.256 | |
ALP | 12/255 | 0.929 | 0.891 | 0.838 | 0.775 | 0.720 | |
TRADES | 12/255 | 0.938 | 0.907 | 0.864 | 0.813 | 0.758 | |
ATLSI | 12/255 | 0.942 | 0.897 | 0.845 | 0.807 | 0.786 | |
BCAT | 12/255 | 0.946 | 0.915 | 0.873 | 0.815 | 0.755 | |
BCAT+ | 12/255 | 0.948 | 0.917 | 0.876 | 0.819 | 0.759 | |
ResNet18 | AT | 12/255 | 0.929 | 0.891 | 0.842 | 0.782 | 0.727 |
freeAT | 12/255 | 0.924 | 0.864 | 0.782 | 0.672 | 0.600 | |
ALP | 12/255 | 0.934 | 0.900 | 0.853 | 0.795 | 0.737 | |
TRADES | 12/255 | 0.933 | 0.904 | 0.863 | 0.812 | 0.760 | |
ATLSI | 12/255 | 0.943 | 0.902 | 0.842 | 0.771 | 0.709 | |
BCAT | 12/255 | 0.943 | 0.913 | 0.872 | 0.816 | 0.763 | |
BCAT+ | 12/255 | 0.943 | 0.914 | 0.873 | 0.822 | 0.766 |
Source/Target Model | Defense | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|
ResNet34/ResNet34 | AT | 8/255 | 0.862 | 0.863 | 0.856 | 0.855 |
freeAT | 8/255 | 0.836 | 0.835 | 0.832 | 0.828 | |
ALP | 8/255 | 0.858 | 0.856 | 0.854 | 0.854 | |
TRADES | 8/255 | 0.811 | 0.811 | 0.807 | 0.809 | |
BCAT | 8/255 | 0.861 | 0.860 | 0.855 | 0.855 | |
BCAT+ | 8/255 | 0.864 | 0.861 | 0.858 | 0.858 |
Source/Target Model | Defense | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|
ResNet34-5/ResNet34-5 | AT | 8/255 | 0.640 | 0.636 | 0.636 | 0.634 |
freeAT | 8/255 | 0.622 | 0.619 | 0.618 | 0.617 | |
ALP | 8/255 | 0.639 | 0.634 | 0.635 | 0.634 | |
TRADES | 8/255 | 0.583 | 0.579 | 0.575 | 0.578 | |
BCAT | 8/255 | 0.673 | 0.668 | 0.667 | 0.664 | |
BCAT+ | 8/255 | 0.677 | 0.673 | 0.668 | 0.669 |
Source/Target Model | Defense | 2/255 | 4/255 | 6/255 | 8/255 | |
---|---|---|---|---|---|---|
ResNet18/ResNet18 | AT | 12/255 | 0.923 | 0.918 | 0.914 | 0.910 |
freeAT | 12/255 | 0.917 | 0.912 | 0.908 | 0.904 | |
ALP | 12/255 | 0.929 | 0.923 | 0.918 | 0.914 | |
TRADES | 12/255 | 0.928 | 0.923 | 0.918 | 0.913 | |
BCAT | 12/255 | 0.938 | 0.932 | 0.927 | 0.922 | |
BCAT+ | 12/255 | 0.938 | 0.932 | 0.927 | 0.923 |
Dataset | Defense | w/o | 0 | 2/255 | 4/255 | 6/255 | 8/255 |
---|---|---|---|---|---|---|---|
CIFAR-10 | BCAT | with | 0.863 | 0.814 | 0.752 | 0.682 | 0.617 |
without | 0.839 | 0.763 | 0.678 | 0.582 | 0.500 | ||
BCAT+ | with | 0.863 | 0.817 | 0.761 | 0.691 | 0.628 | |
without | 0.841 | 0.764 | 0.680 | 0.583 | 0.506 | ||
standard | with | 0.863 | 0.809 | 0.739 | 0.666 | 0.602 | |
without | 0.802 | 0.727 | 0.653 | 0.595 | 0.557 | ||
CIFAR-100 | BCAT | with | 0.675 | 0.612 | 0.548 | 0.474 | 0.411 |
without | 0.613 | 0.520 | 0.435 | 0.352 | 0.292 | ||
BCAT+ | with | 0.678 | 0.614 | 0.544 | 0.480 | 0.418 | |
without | 0.620 | 0.531 | 0.437 | 0.355 | 0.298 | ||
standard | with | 0.644 | 0.565 | 0.488 | 0.427 | 0.395 | |
without | 0.570 | 0.491 | 0.419 | 0.363 | 0.325 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Wang, D.; Jin, W.; Wu, Y. Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification. Sensors 2023, 23, 3252. https://doi.org/10.3390/s23063252
Wang D, Jin W, Wu Y. Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification. Sensors. 2023; 23(6):3252. https://doi.org/10.3390/s23063252
Chicago/Turabian StyleWang, Desheng, Weidong Jin, and Yunpu Wu. 2023. "Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification" Sensors 23, no. 6: 3252. https://doi.org/10.3390/s23063252
APA StyleWang, D., Jin, W., & Wu, Y. (2023). Between-Class Adversarial Training for Improving Adversarial Robustness of Image Classification. Sensors, 23(6), 3252. https://doi.org/10.3390/s23063252