Abstract
In recent researches on adversarial robustness boosting, the trade-off between standard and robust generalization has been widely concerned, in which margin, the average distance from samples to the decision boundary, has become the bridge between the two ends. In this paper, the problems of the existing methods to improve the adversarial robustness by maximizing the margin are discussed and analyzed. On this basis, a new method to approximate the margin from a global point of view through the Wasserstein Distance of distribution of representation is proposed, which is called Global Wasserstein Margin. By maximizing the Global Wasserstein Margin in the process of adversarial training, the generalization capability of the model can be improved, reflected as the standard and robust accuracy advantages on the latest baseline of adversarial training.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Szegedy C, Zaremba W, Sutskever I (2014) Intriguing properties of neural networks. In: International conference on learning representations (ICLR)
Goodfellow I, Shlens J, Szegedy C (2015) Explaining and harnessing adversarial examples. In: International conference on learning representations (ICLR)
Zhang H, Yu Y, Jiao J, Xing E, El Ghaoui L, Jordan M (2019) Theoretically principled trade-off between robustness and accuracy. In: International conference on machine learning. PMLR, pp 7472–7482
Cisse M, Bojanowski P, Grave E, Dauphin Y, Usunier N (2017) Parseval networks: Improving robustness to adversarial examples. In: International conference on machine learning. PMLR, pp 854–863
Wen Y, Li S, Jia K (2020) Towards understanding the regularization of adversarial robustness on neural networks. In: International conference on machine learning. PMLR, pp 10225–10235
Arjovsky M, Chintala S, Bottou L (2017) Wasserstein generative adversarial networks. In: International conference on machine learning. PMLR, pp 214–223
Madry A, Makelov A, Schmidt L (2018) Towards deep learning models resistant to adversarial attacks. In: International conference on learning representations (ICLR)
Kannan H, Kurakin A, Goodfellow I (2018) Adversarial logit pairing. arXiv:1803.06373
Miyato T, Maeda S. -i., Koyama M, Ishii S (2018) Virtual adversarial training: a regularization method for supervised and semi-supervised learning. IEEE Trans Pattern Anal Mach Intell 41(8):1979–1993
Tramèr F, Boneh D, Kurakin A, Goodfellow I, Papernot N, McDaniel P (2018) Ensemble adversarial training: Attacks and defenses. In: 6th International Conference on Learning Representations
Sokolic J, Giryes R, Sapiro G, Rodrigues M (2017) Generalization error of invariant classifiers. In: Artificial intelligence and statistics. PMLR, pp 1094–1103
Cao K, Wei C, Gaidon A, Arechiga N, Ma T (2019) Learning imbalanced datasets with label-distribution-aware margin loss. In: 33Rd conference on neural information processing systems, (neurIPS
Jetley S, Lord NA, Torr PH (2018) With friends like these, who needs adversaries?. In: Proceedings of the 32nd International Conference on Neural Information Processing Systems, pp 10772–10782
Fawzi A, Fawzi O, Frossard P (2018) Analysis of classifiers robustness to adversarial perturbations. Mach Learn 107(3):481–508
Ding GW, Sharma Y, Lui KYC, Huang R (2020) Mma training: Direct input space margin maximization through adversarial training. In: International conference on learning representations
Chengzhi Mao ZZ (2019) Metric learning for adversarial robustness Advances in neural information processing systems
Hoffer E, Ailon N (2015) Deep metric learning using triplet network. In: International workshop on similarity-based pattern recognition. Springer, pp 84–92
Mustafa A, Khan SH, Hayat M, Goecke R, Shen J, Shao L (2020) Deeply supervised discriminative learning for adversarial defense. IEEE Trans Pattern Anal Mach Intell:1–1
Wen Y, Zhang K, Li Z, Qiao Y (2016) A discriminative feature learning approach for deep face recognition. In: European conference on computer vision. Springer, pp 499–515
Wong E, Kolter Z (2018) Provable defenses against adversarial examples via the convex outer adversarial polytope. In: International conference on machine learning. PMLR, pp 5286–5295
Stanforth R, Gowal S, Mann T, Kohli P, et al. (2018) A dual approach to scalable verification of deep networks. arXiv:1803.06567
Boopathy A, Weng T-W, Chen P-Y, Liu S, Daniel L (2019) Cnn-cert: an efficient framework for certifying robustness of convolutional neural networks. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 33, pp 3240- -3247
Wang Y, Zou D, Yi J, Bailey J, Ma X, Gu Q (2020) Improving adversarial robustness requires revisiting misclassified examples. In: International conference on learning representations
Xu H, Liu X, Li Y, Jain A, Tang J (2021) To be robust or to be fair: Towards fairness in adversarial training. In: International conference on machine learning. PMLR, pp 11492–11501
Zhang S, Huang K, Zhu J, Liu Y (2021) Manifold adversarial training for supervised and semi-supervised learning. Neural Netw 140:282–293
Rade R, Moosavi-Dezfooli S-M (2021) Helper-based adversarial training: Reducing excessive margin to achieve a better accuracy vs. robustness trade-off. In: ICML 2021 Workshop on Adversarial Machine Learning
Wong E, Rice L, Kolter JZ (2019) Fast is better than free: Revisiting adversarial training. In: International Conference on Learning Representations
Kim H, Lee W, Lee J (2021) Understanding catastrophic overfitting in single-step adversarial training. In: Proceedings of the AAAI Conference on Artificial Intelligence, vol 35, pp 8119–8127
Andriushchenko M, Flammarion N (2020) Understanding and improving fast adversarial training. Adv Neural Inf Process Syst 33
Vivek B, Babu RV (2020) Single-step adversarial training with dropout scheduling. In: 2020 IEEE/CVF Conference on computer vision and pattern recognition (CVPR). IEEE, pp 947–956
Roth K, Kilcher Y, Hofmann T (2019) Adversarial training generalizes data-dependent spectral norm regularization
Soudry D, Hoffer E, Nacson MS, Gunasekar S, Srebro N (2018) The implicit bias of gradient descent on separable data. J Mach Learn Res 19(1):2822–2878
Nacson MS, Lee J, Gunasekar S, Savarese PHP, Srebro N, Soudry D (2019) Convergence of gradient descent on separable data. In: The 22nd international conference on artificial intelligence and statistics. PMLR, pp 3420–3428
LeCun Y, Bottou L, Bengio Y, Haffner P (1998) Gradient-based learning applied to document recognition. Proc IEEE 86(11):2278–2324
Krizhevsky A, Hinton G (2009) Learning multiple layers of features from tiny images. Handb Syst Autoimmune Diseas 1:4
Reading digits in natural images with unsupervised feature learning. nips workshop on deep learning & unsupervised feature learning (2011)
Dong Y, Liao F, Pang T, Su H, Zhu J, Hu X, Li J (2018) Boosting adversarial attacks with momentum. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 9185–9193
He K, Zhang X, Ren S, Sun J (2016) Deep residual learning for image recognition. In: Proceedings of the IEEE conference on computer vision and pattern recognition, pp 770–778
Acknowledgements
This work is supported by the National Defense Basic Scientific Research Program of China (Grant Number: JCKY2018603B006)
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Yu, T., Wang, S. & Yu, X. Global Wasserstein Margin maximization for boosting generalization in adversarial training. Appl Intell 53, 11490–11504 (2023). https://doi.org/10.1007/s10489-022-03480-w
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-022-03480-w