Badclip: Dual-embedding guided backdoor attack on multimodal contrastive learning

S Liang, M Zhu, A Liu, B Wu, X Cao… - Proceedings of the …, 2024 - openaccess.thecvf.com
While existing backdoor attacks have successfully infected multimodal contrastive learning
models such as CLIP they can be easily countered by specialized backdoor defenses for …

Nearest is not dearest: Towards practical defense against quantization-conditioned backdoor attacks

B Li, Y Cai, H Li, F Xue, Z Li… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Model quantization is widely used to compress and accelerate deep neural
networks. However recent studies have revealed the feasibility of weaponizing model …

Where did i come from? origin attribution of ai-generated images

Z Wang, C Chen, Y Zeng, L Lyu… - Advances in neural …, 2024 - proceedings.neurips.cc
Image generation techniques have been gaining increasing attention recently, but concerns
have been raised about the potential misuse and intellectual property (IP) infringement …

Distributed backdoor attacks on federated graph learning and certified defenses

Y Yang, Q Li, J Jia, Y Hong, B Wang - Proceedings of the 2024 on ACM …, 2024 - dl.acm.org
Federated graph learning (FedGL) is an emerging federated learning (FL) framework that
extends FL to learn graph data from diverse sources without accessing the data. FL for non …

Django: Detecting trojans in object detection models via gaussian focus calibration

G Shen, S Cheng, G Tao, K Zhang… - Advances in …, 2023 - proceedings.neurips.cc
Object detection models are vulnerable to backdoor or trojan attacks, where an attacker can
inject malicious triggers into the model, leading to altered behavior during inference. As a …

Towards secure tuning: Mitigating security risks arising from benign instruction fine-tuning

Y Du, S Zhao, J Cao, M Ma, D Zhao, F Fan… - arXiv preprint arXiv …, 2024 - arxiv.org
Instruction Fine-Tuning (IFT) has become an essential method for adapting base Large
Language Models (LLMs) into variants for professional and private use. However …

Ta-cleaner: A fine-grained text alignment backdoor defense strategy for multimodal contrastive learning

Y Xun, S Liang, X Jia, X Liu, X Cao - arXiv preprint arXiv:2409.17601, 2024 - arxiv.org
Pre-trained large models for multimodal contrastive learning, such as CLIP, have been
widely recognized in the industry as highly susceptible to data-poisoned backdoor attacks …

Data Poisoning based Backdoor Attacks to Contrastive Learning

J Zhang, H Liu, J Jia, NZ Gong - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Contrastive learning (CL) pre-trains general-purpose encoders using an unlabeled pre-
training dataset which consists of images or image-text pairs. CL is vulnerable to data …

Transtroj: Transferable backdoor attacks to pre-trained models via embedding indistinguishability

H Wang, T Xiang, S Guo, J He, H Liu… - arXiv preprint arXiv …, 2024 - arxiv.org
Pre-trained models (PTMs) are extensively utilized in various downstream tasks. Adopting
untrusted PTMs may suffer from backdoor attacks, where the adversary can compromise the …

DMGNN: Detecting and Mitigating Backdoor Attacks in Graph Neural Networks

H Sui, B Chen, J Zhang, C Zhu, D Wu, Q Lu… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent studies have revealed that GNNs are highly susceptible to multiple adversarial
attacks. Among these, graph backdoor attacks pose one of the most prominent threats …