Federated learning for generalization, robustness, fairness: A survey and benchmark

W Huang, M Ye, Z Shi, G Wan, H Li… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Federated learning has emerged as a promising paradigm for privacy-preserving
collaboration among different parties. Recently, with the popularity of federated learning, an …

Defending against weight-poisoning backdoor attacks for parameter-efficient fine-tuning

S Zhao, L Gan, LA Tuan, J Fu, L Lyu, M Jia… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently, various parameter-efficient fine-tuning (PEFT) strategies for application to
language models have been proposed and successfully implemented. However, this raises …

Defenses in adversarial machine learning: A survey

B Wu, S Wei, M Zhu, M Zheng, Z Zhu, M Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Adversarial phenomenon has been widely observed in machine learning (ML) systems,
especially in those using deep neural networks, describing that ML systems may produce …

Ufid: A unified framework for input-level backdoor detection on diffusion models

Z Guan, M Hu, S Li, A Vullikanti - arXiv preprint arXiv:2404.01101, 2024 - arxiv.org
Diffusion Models are vulnerable to backdoor attacks, where malicious attackers inject
backdoors by poisoning some parts of the training samples during the training stage. This …

Universal detection of backdoor attacks via density-based clustering and centroids analysis

W Guo, B Tondi, M Barni - IEEE Transactions on Information …, 2023 - ieeexplore.ieee.org
We propose a Universal Defence against backdoor attacks based on Clustering and
Centroids Analysis (CCA-UD). The goal of the defence is to reveal whether a Deep Neural …

Backdoor attack on hash-based image retrieval via clean-label data poisoning

K Gao, J Bai, B Chen, D Wu, ST Xia - arXiv preprint arXiv:2109.08868, 2021 - arxiv.org
A backdoored deep hashing model is expected to behave normally on original query
images and return the images with the target label when a specific trigger pattern presents …

Mendata: A Framework to Purify Manipulated Training Data

Z Huang, N Gong, MK Reiter - arXiv preprint arXiv:2312.01281, 2023 - arxiv.org
Untrusted data used to train a model might have been manipulated to endow the learned
model with hidden properties that the data contributor might later exploit. Data purification …

Flatness-aware Sequential Learning Generates Resilient Backdoors

H Pham, TA Ta, A Tran, KD Doan - arXiv preprint arXiv:2407.14738, 2024 - arxiv.org
Recently, backdoor attacks have become an emerging threat to the security of machine
learning models. From the adversary's perspective, the implanted backdoors should be …

Mitigating Backdoor Attack by Injecting Proactive Defensive Backdoor

S Wei, H Zha, B Wu - arXiv preprint arXiv:2405.16112, 2024 - arxiv.org
Data-poisoning backdoor attacks are serious security threats to machine learning models,
where an adversary can manipulate the training dataset to inject backdoors into models. In …

Augmented Neural Fine-Tuning for Efficient Backdoor Purification

N Karim, AA Arafat, U Khalid, Z Guo… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent studies have revealed the vulnerability of deep neural networks (DNNs) to various
backdoor attacks, where the behavior of DNNs can be compromised by utilizing certain …