Threats to pre-trained language models: Survey and taxonomy

S Guo, C Xie, J Li, L Lyu, T Zhang - arXiv preprint arXiv:2202.06862, 2022 - arxiv.org
Pre-trained language models (PTLMs) have achieved great success and remarkable
performance over a wide range of natural language processing (NLP) tasks. However, there …

A unified evaluation of textual backdoor learning: Frameworks and benchmarks

G Cui, L Yuan, B He, Y Chen… - Advances in Neural …, 2022 - proceedings.neurips.cc
Textual backdoor attacks are a kind of practical threat to NLP systems. By injecting a
backdoor in the training phase, the adversary could control model predictions via predefined …

Backdoors against natural language processing: A review

S Li, T Dong, BZH Zhao, M Xue, S Du… - IEEE Security & …, 2022 - ieeexplore.ieee.org
Backdoors Against Natural Language Processing: A Review Page 1 50 September/October
2022 Copublished by the IEEE Computer and Reliability Societies 1540-7993/22©2022IEEE …

Piccolo: Exposing complex backdoors in nlp transformer models

Y Liu, G Shen, G Tao, S An, S Ma… - 2022 IEEE Symposium …, 2022 - ieeexplore.ieee.org
Backdoors can be injected to NLP models such that they misbehave when the trigger words
or sentences appear in an input sample. Detecting such backdoors given only a subject …

Towards practical deployment-stage backdoor attack on deep neural networks

X Qi, T Xie, R Pan, J Zhu, Y Yang… - Proceedings of the …, 2022 - openaccess.thecvf.com
One major goal of the AI security community is to securely and reliably produce and deploy
deep learning models for real-world applications. To this end, data poisoning based …

Constrained optimization with dynamic bound-scaling for effective nlp backdoor defense

G Shen, Y Liu, G Tao, Q Xu, Z Zhang… - International …, 2022 - proceedings.mlr.press
Modern language models are vulnerable to backdoor attacks. An injected malicious token
sequence (ie, a trigger) can cause the compromised model to misbehave, raising security …

Physical backdoor attacks to lane detection systems in autonomous driving

X Han, G Xu, Y Zhou, X Yang, J Li… - Proceedings of the 30th …, 2022 - dl.acm.org
Modern autonomous vehicles adopt state-of-the-art DNN models to interpret the sensor data
and perceive the environment. However, DNN models are vulnerable to different types of …

A survey on backdoor attack and defense in natural language processing

X Sheng, Z Han, P Li, X Chang - 2022 IEEE 22nd International …, 2022 - ieeexplore.ieee.org
Deep learning is becoming increasingly popular in real-life applications, especially in
natural language processing (NLP). Users often choose training outsourcing or adopt third …

A roadmap for big model

S Yuan, H Zhao, S Zhao, J Leng, Y Liang… - arXiv preprint arXiv …, 2022 - arxiv.org
With the rapid development of deep learning, training Big Models (BMs) for multiple
downstream tasks becomes a popular paradigm. Researchers have achieved various …

Incremental learning, incremental backdoor threats

W Jiang, T Zhang, H Qiu, H Li… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Class incremental learning from a pre-trained DNN model is gaining lots of popularity.
Unfortunately, the pre-trained model also introduces a new attack vector, which enables an …