A survey of adversarial defenses and robustness in nlp

S Goyal, S Doddapaneni, MM Khapra… - ACM Computing …, 2023 - dl.acm.org
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …

Better diffusion models further improve adversarial training

Z Wang, T Pang, C Du, M Lin… - … on Machine Learning, 2023 - proceedings.mlr.press
It has been recognized that the data generated by the denoising diffusion probabilistic
model (DDPM) improves adversarial training. After two years of rapid development in …

Backdoor learning for nlp: Recent advances, challenges, and future research directions

M Omar - arXiv preprint arXiv:2302.06801, 2023 - arxiv.org
Although backdoor learning is an active research topic in the NLP domain, the literature
lacks studies that systematically categorize and summarize backdoor attacks and defenses …

Prada: Practical black-box adversarial attacks against neural ranking models

C Wu, R Zhang, J Guo, M De Rijke, Y Fan… - ACM Transactions on …, 2023 - dl.acm.org
Neural ranking models (NRMs) have shown remarkable success in recent years, especially
with pre-trained language models. However, deep neural models are notorious for their …

Bridging the gap between indexing and retrieval for differentiable search index with query generation

S Zhuang, H Ren, L Shou, J Pei, M Gong… - arXiv preprint arXiv …, 2022 - arxiv.org
The Differentiable Search Index (DSI) is an emerging paradigm for information retrieval.
Unlike traditional retrieval architectures where index and retrieval are two different and …

Certified robustness to text adversarial attacks by randomized [mask]

J Zeng, J Xu, X Zheng, X Huang - Computational Linguistics, 2023 - direct.mit.edu
Very recently, few certified defense methods have been developed to provably guarantee
the robustness of a text classifier to adversarial synonym substitutions. However, all the …

Model-tuning Via Prompts Makes NLP Models Adversarially Robust

M Raman, P Maini, J Kolter, ZC Lipton… - Proceedings of the …, 2023 - aclanthology.org
In recent years, NLP practitioners have converged on the following practice:(i) import an off-
the-shelf pretrained (masked) language model;(ii) append a multilayer perceptron atop the …

Textual manifold-based defense against natural language adversarial examples

DN Minh, AT Luu - Proceedings of the 2022 Conference on …, 2022 - aclanthology.org
Despite the recent success of large pretrained language models in NLP, they are
susceptible to adversarial examples. Concurrently, several studies on adversarial images …

Dabert: Dual attention enhanced bert for semantic matching

S Wang, D Liang, J Song, Y Li, W Wu - arXiv preprint arXiv:2210.03454, 2022 - arxiv.org
Transformer-based pre-trained language models such as BERT have achieved remarkable
results in Semantic Sentence Matching. However, existing models still suffer from insufficient …

Rmlm: A flexible defense framework for proactively mitigating word-level adversarial attacks

Z Wang, Z Liu, X Zheng, Q Su… - Proceedings of the 61st …, 2023 - aclanthology.org
Adversarial attacks on deep neural networks keep raising security concerns in natural
language processing research. Existing defenses focus on improving the robustness of the …