Contextualized perturbation for textual adversarial attack

D Li, Y Zhang, H Peng, L Chen, C Brockett… - arXiv preprint arXiv …, 2020 - arxiv.org
Adversarial examples expose the vulnerabilities of natural language processing (NLP)
models, and can be used to evaluate and improve their robustness. Existing techniques of …

Bridge the gap between cv and nlp! a gradient-based textual adversarial attack framework

L Yuan, Y Zhang, Y Chen, W Wei - arXiv preprint arXiv:2110.15317, 2021 - arxiv.org
Despite recent success on various tasks, deep learning techniques still perform poorly on
adversarial examples with small perturbations. While optimization-based methods for …

Generating fluent adversarial examples for natural languages

H Zhang, H Zhou, N Miao, L Li - arXiv preprint arXiv:2007.06174, 2020 - arxiv.org
Efficiently building an adversarial attacker for natural language processing (NLP) tasks is a
real challenge. Firstly, as the sentence space is discrete, it is difficult to make small …

Openattack: An open-source textual adversarial attack toolkit

G Zeng, F Qi, Q Zhou, T Zhang, Z Ma, B Hou… - arXiv preprint arXiv …, 2020 - arxiv.org
Textual adversarial attacking has received wide and increasing attention in recent years.
Various attack models have been proposed, which are enormously distinct and …

Multi-granularity textual adversarial attack with behavior cloning

Y Chen, J Su, W Wei - arXiv preprint arXiv:2109.04367, 2021 - arxiv.org
Recently, the textual adversarial attack models become increasingly popular due to their
successful in estimating the robustness of NLP models. However, existing works have …

Is bert really robust? a strong baseline for natural language attack on text classification and entailment

D Jin, Z Jin, JT Zhou, P Szolovits - Proceedings of the AAAI conference on …, 2020 - aaai.org
Abstract Machine learning algorithms are often vulnerable to adversarial examples that have
imperceptible alterations from the original counterparts but can fool the state-of-the-art …

Generating natural language attacks in a hard label black box setting

R Maheshwary, S Maheshwary, V Pudi - Proceedings of the AAAI …, 2021 - ojs.aaai.org
We study an important and challenging task of attacking natural language processing
models in a hard label black box setting. We propose a decision-based attack strategy that …

Adversarial glue: A multi-task benchmark for robustness evaluation of language models

B Wang, C Xu, S Wang, Z Gan, Y Cheng, J Gao… - arXiv preprint arXiv …, 2021 - arxiv.org
Large-scale pre-trained language models have achieved tremendous success across a
wide range of natural language understanding (NLU) tasks, even surpassing human …

Adversarial attack and defense technologies in natural language processing: A survey

S Qiu, Q Liu, S Zhou, W Huang - Neurocomputing, 2022 - Elsevier
Recently, the adversarial attack and defense technology has made remarkable
achievements and has been widely applied in the computer vision field, promoting its rapid …

T3: Tree-autoencoder constrained adversarial text generation for targeted attack

B Wang, H Pei, B Pan, Q Chen, S Wang, B Li - arXiv preprint arXiv …, 2019 - arxiv.org
Adversarial attacks against natural language processing systems, which perform seemingly
innocuous modifications to inputs, can induce arbitrary mistakes to the target models …