Sliced Wasserstein adversarial training for improving adversarial robustness
- Authors
- Lee, W.; Lee, S.; Kim, H.; Lee, J.
- Issue Date
- May-2024
- Publisher
- Springer-Verlag GmbH Germany
- Keywords
- Adversarial attack; Adversarial defense; Adversarial training; Sliced Wasserstein Distance
- Citation
- Journal of Ambient Intelligence and Humanized Computing, v.15, no.8, pp 1 - 14
- Pages
- 14
- Indexed
- SCOPUS
- Journal Title
- Journal of Ambient Intelligence and Humanized Computing
- Volume
- 15
- Number
- 8
- Start Page
- 1
- End Page
- 14
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/22165
- DOI
- 10.1007/s12652-024-04791-1
- ISSN
- 1868-5137
1868-5145
- Abstract
- Recently, deep-learning-based models have achieved impressive performance on tasks that were previously considered to be extremely challenging. However, recent works have shown that various deep learning models are susceptible to adversarial data samples. In this paper, we propose the sliced Wasserstein adversarial training method to encourage the logit distributions of clean and adversarial data to be similar to each other. We capture the dissimilarity between two distributions using the Wasserstein metric and then align distributions using an end-to-end training process. We present the theoretical background of the motivation for our study by providing generalization error bounds for adversarial data samples. We performed experiments on three standard datasets and the results demonstrate that our method is more robust against white box attacks compared to previous methods. © The Author(s) 2024.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.