On the robustness of the cvpr 2018 white-box adversarial example defenses

A Athalye, N Carlini - arXiv preprint arXiv:1804.03286, 2018 - arxiv.org
Neural networks are known to be vulnerable to adversarial examples. In this note, we
evaluate the two white-box defenses that appeared at CVPR 2018 and find they are …

Towards deep neural network architectures robust to adversarial examples

S Gu, L Rigazio - arXiv preprint arXiv:1412.5068, 2014 - arxiv.org
Recent work has shown deep neural networks (DNNs) to be highly susceptible to well-
designed, small perturbations at the input layer, or so-called adversarial examples. Taking …

Deep defense: Training dnns with improved adversarial robustness

Z Yan, Y Guo, C Zhang - Advances in Neural Information …, 2018 - proceedings.neurips.cc
Despite the efficacy on a variety of computer vision tasks, deep neural networks (DNNs) are
vulnerable to adversarial attacks, limiting their applications in security-critical systems …

Universal adversarial attack on attention and the resulting dataset damagenet

S Chen, Z He, C Sun, J Yang… - IEEE Transactions on …, 2020 - ieeexplore.ieee.org
Adversarial attacks on deep neural networks (DNNs) have been found for several years.
However, the existing adversarial attacks have high success rates only when the information …

Flooding-X: Improving BERT's resistance to adversarial attacks via loss-restricted fine-tuning

Q Liu, R Zheng, B Rong, J Liu, Z Liu… - Proceedings of the …, 2022 - aclanthology.org
Adversarial robustness has attracted much attention recently, and the mainstream solution is
adversarial training. However, the tradition of generating adversarial perturbations for each …

Limitations of the lipschitz constant as a defense against adversarial examples

T Huster, CYJ Chiang, R Chadha - … 2018, SoGood 2018, IWAISe 2018, and …, 2019 - Springer
Several recent papers have discussed utilizing Lipschitz constants to limit the susceptibility
of neural networks to adversarial examples. We analyze recently proposed methods for …

Why should adversarial perturbations be imperceptible? rethink the research paradigm in adversarial nlp

Y Chen, H Gao, G Cui, F Qi, L Huang, Z Liu… - arXiv preprint arXiv …, 2022 - arxiv.org
Textual adversarial samples play important roles in multiple subfields of NLP research,
including security, evaluation, explainability, and data augmentation. However, most work …

A survey on transferability of adversarial examples across deep neural networks

J Gu, X Jia, P de Jorge, W Yu, X Liu, A Ma… - arXiv preprint arXiv …, 2023 - arxiv.org
The emergence of Deep Neural Networks (DNNs) has revolutionized various domains,
enabling the resolution of complex tasks spanning image recognition, natural language …

Searching for a search method: Benchmarking search algorithms for generating NLP adversarial examples

JY Yoo, JX Morris, E Lifland, Y Qi - arXiv preprint arXiv:2009.06368, 2020 - arxiv.org
We study the behavior of several black-box search algorithms used for generating
adversarial examples for natural language processing (NLP) tasks. We perform a fine …

Interpreting adversarially trained convolutional neural networks

T Zhang, Z Zhu - International conference on machine …, 2019 - proceedings.mlr.press
We attempt to interpret how adversarially trained convolutional neural networks (AT-CNNs)
recognize objects. We design systematic approaches to interpret AT-CNNs in both …