Using Random Perturbations to Mitigate Adversarial Attacks on NLP Models

A Swenor - Proceedings of the AAAI Conference on Artificial …, 2022 - ojs.aaai.org
Deep learning models have excelled in solving many problems in Natural Language
Processing, but are susceptible to extensive vulnerabilities. We offer a solution to this …

Defense against adversarial attacks in nlp via dirichlet neighborhood ensemble

Y Zhou, X Zheng, CJ Hsieh, K Chang… - arXiv preprint arXiv …, 2020 - arxiv.org
Despite neural networks have achieved prominent performance on many natural language
processing (NLP) tasks, they are vulnerable to adversarial examples. In this paper, we …

Towards improving adversarial training of NLP models

JY Yoo, Y Qi - arXiv preprint arXiv:2109.00544, 2021 - arxiv.org
Adversarial training, a method for learning robust deep neural networks, constructs
adversarial examples during training. However, recent methods for generating NLP …

[PDF][PDF] Defense against synonym substitution-based adversarial attacks via Dirichlet neighborhood ensemble

Y Zhou, X Zheng, CJ Hsieh, KW Chang… - Association for …, 2021 - par.nsf.gov
Although deep neural networks have achieved prominent performance on many NLP tasks,
they are vulnerable to adversarial examples. We propose Dirichlet Neighborhood Ensemble …

Word level robustness enhancement: Fight perturbation with perturbation

P Huang, Y Yang, F Jia, M Liu, F Ma… - Proceedings of the AAAI …, 2022 - ojs.aaai.org
State-of-the-art deep NLP models have achieved impressive improvements on many tasks.
However, they are found to be vulnerable to some perturbations. Before they are widely …

Adversarial attack and defense of structured prediction models

W Han, L Zhang, Y Jiang, K Tu - arXiv preprint arXiv:2010.01610, 2020 - arxiv.org
Building an effective adversarial attacker and elaborating on countermeasures for
adversarial attacks for natural language processing (NLP) have attracted a lot of research in …

Adversarial attacks on deep-learning models in natural language processing: A survey

WE Zhang, QZ Sheng, A Alhazmi, C Li - ACM Transactions on Intelligent …, 2020 - dl.acm.org
With the development of high computational devices, deep neural networks (DNNs), in
recent years, have gained significant popularity in many Artificial Intelligence (AI) …

Concealed data poisoning attacks on NLP models

E Wallace, TZ Zhao, S Feng, S Singh - arXiv preprint arXiv:2010.12563, 2020 - arxiv.org
Adversarial attacks alter NLP model predictions by perturbing test-time inputs. However, it is
much less understood whether, and how, predictions can be manipulated with small …

Improving the reliability of deep neural networks in NLP: A review

B Alshemali, J Kalita - Knowledge-Based Systems, 2020 - Elsevier
Deep learning models have achieved great success in solving a variety of natural language
processing (NLP) problems. An ever-growing body of research, however, illustrates the …

Certified robustness against natural language attacks by causal intervention

H Zhao, C Ma, X Dong, AT Luu… - International …, 2022 - proceedings.mlr.press
Deep learning models have achieved great success in many fields, yet they are vulnerable
to adversarial examples. This paper follows a causal perspective to look into the adversarial …