Natural language adversarial defense through synonym encoding

X Wang, J Hao, Y Yang, K He - Uncertainty in Artificial …, 2021 - proceedings.mlr.press
In the area of natural language processing, deep learning models are recently known to be
vulnerable to various types of adversarial perturbations, but relatively few works are done on …

[PDF][PDF] Defense against synonym substitution-based adversarial attacks via Dirichlet neighborhood ensemble

Y Zhou, X Zheng, CJ Hsieh, KW Chang… - Association for …, 2021 - par.nsf.gov
Although deep neural networks have achieved prominent performance on many NLP tasks,
they are vulnerable to adversarial examples. We propose Dirichlet Neighborhood Ensemble …

Natural language adversarial attack and defense in word level

X Wang, H Jin, K He - 2019 - openreview.net
Up until very recently, inspired by a mass of researches on adversarial examples for
computer vision, there has been a growing interest in designing adversarial attacks for …

Defense of word-level adversarial attacks via random substitution encoding

Z Wang, H Wang - … 13th International Conference, KSEM 2020, Hangzhou …, 2020 - Springer
The adversarial attacks against deep neural networks on computer vision tasks have
spawned many new technologies that help protect models from avoiding false predictions …

Rmlm: A flexible defense framework for proactively mitigating word-level adversarial attacks

Z Wang, Z Liu, X Zheng, Q Su… - Proceedings of the 61st …, 2023 - aclanthology.org
Adversarial attacks on deep neural networks keep raising security concerns in natural
language processing research. Existing defenses focus on improving the robustness of the …

[PDF][PDF] Towards Semantics-and Domain-Aware Adversarial Attacks.

J Zhang, YC Huang, W Wu, MR Lyu - IJCAI, 2023 - ijcai.org
Abstract Language models are known to be vulnerable to textual adversarial attacks, which
add humanimperceptible perturbations to the input to mislead DNNs. It is thus imperative to …

Searching for an effective defender: Benchmarking defense against adversarial word substitution

Z Li, J Xu, J Zeng, L Li, X Zheng, Q Zhang… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent studies have shown that deep neural networks are vulnerable to intentionally crafted
adversarial examples, and various methods have been proposed to defend against …

Defense against adversarial attacks in nlp via dirichlet neighborhood ensemble

Y Zhou, X Zheng, CJ Hsieh, K Chang… - arXiv preprint arXiv …, 2020 - arxiv.org
Despite neural networks have achieved prominent performance on many natural language
processing (NLP) tasks, they are vulnerable to adversarial examples. In this paper, we …

Reevaluating adversarial examples in natural language

JX Morris, E Lifland, J Lanchantin, Y Ji, Y Qi - arXiv preprint arXiv …, 2020 - arxiv.org
State-of-the-art attacks on NLP models lack a shared definition of a what constitutes a
successful attack. We distill ideas from past work into a unified framework: a successful …

TextGuise: Adaptive adversarial example attacks on text classification model

G Chang, H Gao, Z Yao, H Xiong - Neurocomputing, 2023 - Elsevier
Adversarial examples greatly compromise the security of deep learning models. The key to
improving the robustness of a natural language processing (NLP) model is to study attacks …