Towards Better Understanding of Training Certifiably Robust Models against Adversarial Examples
- Authors
- Lee, S.; Lee, W.; Park, J.; Lee, J.
- Issue Date
- 2021
- Publisher
- Neural information processing systems foundation
- Citation
- Advances in Neural Information Processing Systems, v.2, pp 953 - 964
- Pages
- 12
- Indexed
- SCOPUS
- Journal Title
- Advances in Neural Information Processing Systems
- Volume
- 2
- Start Page
- 953
- End Page
- 964
- URI
- https://scholarworks.dongguk.edu/handle/sw.dongguk/5590
- ISSN
- 1049-5258
- Abstract
- We study the problem of training certifiably robust models against adversarial examples. Certifiable training minimizes an upper bound on the worst-case loss over the allowed perturbation, and thus the tightness of the upper bound is an important factor in building certifiably robust models. However, many studies have shown that Interval Bound Propagation (IBP) training uses much looser bounds but outperforms other models that use tighter bounds. We identify another key factor that influences the performance of certifiable training: smoothness of the loss landscape. We find significant differences in the loss landscapes across many linear relaxation-based methods, and that the current state-of-the-arts method often has a landscape with favorable optimization properties. Moreover, to test the claim, we design a new certifiable training method with the desired properties. With the tightness and the smoothness, the proposed method achieves a decent performance under a wide range of perturbations, while others with only one of the two factors can perform well only for a specific range of perturbations. Our code is available at https://github.com/sungyoon-lee/LossLandscapeMatters. © 2021 Neural information processing systems foundation. All rights reserved.
- Files in This Item
- There are no files associated with this item.
- Appears in
Collections - College of Advanced Convergence Engineering > Department of Computer Science and Artificial Intelligence > 1. Journal Articles

Items in ScholarWorks are protected by copyright, with all rights reserved, unless otherwise indicated.