Revisiting the assumption of latent separability for backdoor defenses

X Qi, T Xie, Y Li, S Mahloujifar… - The eleventh international …, 2023 - openreview.net
Recent studies revealed that deep learning is susceptible to backdoor poisoning attacks. An
adversary can embed a hidden backdoor into a model to manipulate its predictions by only …

Detecting backdoors in pre-trained encoders

S Feng, G Tao, S Cheng, G Shen… - Proceedings of the …, 2023 - openaccess.thecvf.com
Self-supervised learning in computer vision trains on unlabeled data, such as images or
(image, text) pairs, to obtain an image encoder that learns high-quality embeddings for input …

Notable: Transferable backdoor attacks against prompt-based nlp models

K Mei, Z Li, Z Wang, Y Zhang, S Ma - arXiv preprint arXiv:2305.17826, 2023 - arxiv.org
Prompt-based learning is vulnerable to backdoor attacks. Existing backdoor attacks against
prompt-based models consider injecting backdoors into the entire embedding layers or word …

Plmmark: a secure and robust black-box watermarking framework for pre-trained language models

P Li, P Cheng, F Li, W Du, H Zhao, G Liu - Proceedings of the AAAI …, 2023 - ojs.aaai.org
The huge training overhead, considerable commercial value, and various potential security
risks make it urgent to protect the intellectual property (IP) of Deep Neural Networks (DNNs) …

Prompt injection attacks and defenses in llm-integrated applications

Y Liu, Y Jia, R Geng, J Jia, NZ Gong - arXiv preprint arXiv:2310.12815, 2023 - arxiv.org
Large Language Models (LLMs) are increasingly deployed as the backend for a variety of
real-world applications called LLM-Integrated Applications. Multiple recent works showed …

Attention-enhancing backdoor attacks against bert-based models

W Lyu, S Zheng, L Pang, H Ling, C Chen - arXiv preprint arXiv:2310.14480, 2023 - arxiv.org
Recent studies have revealed that\textit {Backdoor Attacks} can threaten the safety of natural
language processing (NLP) models. Investigating the strategies of backdoor attacks will help …

Red alarm for pre-trained models: Universal vulnerability to neuron-level backdoor attacks

Z Zhang, G Xiao, Y Li, T Lv, F Qi, Z Liu, Y Wang… - Machine Intelligence …, 2023 - Springer
The pre-training-then-fine-tuning paradigm has been widely used in deep learning. Due to
the huge computation cost for pre-training, practitioners usually download pre-trained …

Privacy in large language models: Attacks, defenses and future directions

H Li, Y Chen, J Luo, Y Kang, X Zhang, Q Hu… - arXiv preprint arXiv …, 2023 - arxiv.org
The advancement of large language models (LLMs) has significantly enhanced the ability to
effectively tackle various downstream NLP tasks and unify these tasks into generative …

Multi-target backdoor attacks for code pre-trained models

Y Li, S Liu, K Chen, X Xie, T Zhang, Y Liu - arXiv preprint arXiv …, 2023 - arxiv.org
Backdoor attacks for neural code models have gained considerable attention due to the
advancement of code intelligence. However, most existing works insert triggers into task …

" Get in Researchers; We're Measuring Reproducibility": A Reproducibility Study of Machine Learning Papers in Tier 1 Security Conferences

D Olszewski, A Lu, C Stillman, K Warren… - Proceedings of the …, 2023 - dl.acm.org
Reproducibility is crucial to the advancement of science; it strengthens confidence in
seemingly contradictory results and expands the boundaries of known discoveries …