作者
Ali Shafahi, W Ronny Huang, Christoph Studer, Soheil Feizi, Tom Goldstein
发表日期
2018/9/6
期刊
arXiv preprint arXiv:1809.02104
简介
A wide range of defenses have been proposed to harden neural networks against adversarial attacks. However, a pattern has emerged in which the majority of adversarial defenses are quickly broken by new attacks. Given the lack of success at generating robust defenses, we are led to ask a fundamental question: Are adversarial attacks inevitable? This paper analyzes adversarial examples from a theoretical perspective, and identifies fundamental bounds on the susceptibility of a classifier to adversarial attacks. We show that, for certain classes of problems, adversarial examples are inescapable. Using experiments, we explore the implications of theoretical guarantees for real-world problems and discuss how factors such as dimensionality and image complexity limit a classifier's robustness against adversarial examples.
引用总数
20182019202020212022202320244477372535727
学术搜索中的文章
A Shafahi, WR Huang, C Studer, S Feizi, T Goldstein - arXiv preprint arXiv:1809.02104, 2018