Shared adversarial unlearning: Backdoor mitigation by unlearning shared adversarial examples

S Wei, M Zhang, H Zha, B Wu - Advances in Neural …, 2023 - proceedings.neurips.cc
Backdoor attacks are serious security threats to machine learning models where an
adversary can inject poisoned samples into the training set, causing a backdoored model …

Sample efficient detection and classification of adversarial attacks via self-supervised embeddings

M Moayeri, S Feizi - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Adversarial robustness of deep models is pivotal in ensuring safe deployment in real world
settings, but most modern defenses have narrow scope and expensive costs. In this paper …

Reverse Engineering of Deceptions on Machine-and Human-Centric Attacks

Y Yao, X Guo, V Asnani, Y Gong, J Liu… - … and Trends® in …, 2024 - nowpublishers.com
This work presents a comprehensive exploration of Reverse Engineering of Deceptions
(RED) in the field of adversarial machine learning. It delves into the intricacies of machine …

Adaptive smoothness-weighted adversarial training for multiple perturbations with its stability analysis

J Xiao, Z Qin, Y Fan, B Wu, J Wang, ZQ Luo - arXiv preprint arXiv …, 2022 - arxiv.org
Adversarial Training (AT) has been demonstrated as one of the most effective methods
against adversarial examples. While most existing works focus on AT with a single type of …

Knowing is Half the Battle: Enhancing Clean Data Accuracy of Adversarial Robust Deep Neural Networks via Dual-model Bounded Divergence Gating

H Aboutalebi, MJ Shafiee, CEA Tai, A Wong - IEEE Access, 2023 - ieeexplore.ieee.org
Significant advances have been made in recent years in improving the robustness of deep
neural networks, particularly under adversarial machine learning scenarios where the data …

Towards Universal Certified Robustness with Multi-Norm Training

E Jiang, G Singh - arXiv preprint arXiv:2410.03000, 2024 - arxiv.org
Existing certified training methods can only train models to be robust against a certain
perturbation type (eg $ l_\infty $ or $ l_2 $). However, an $ l_\infty $ certifiably robust model …

Activation Control of Vision Models for Sustainable AI Systems

J Burton-Barr, B Fernando… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
As AI systems become more complex and widespread, they require significant
computational power, increasing energy consumption. Addressing this challenge is …

Can Adversarial Examples Be Parsed to Reveal Victim Model Information?

Y Yao, J Liu, Y Gong, X Liu, Y Wang, X Lin… - arXiv preprint arXiv …, 2023 - arxiv.org
Numerous adversarial attack methods have been developed to generate imperceptible
image perturbations that can cause erroneous predictions of state-of-the-art machine …

More or less (mol): Defending against multiple perturbation attacks on deep neural networks through model ensemble and compression

H Cheng, K Xu, Z Li, P Zhao, C Wang… - 2022 IEEE/CVF …, 2022 - ieeexplore.ieee.org
Deep neural networks (DNNs) have been adopted in many application domains due to their
superior performance. However, their susceptibility under test-time adversarial perturbations …

Poseidon: A NAS-Based Ensemble Defense Method Against Multiple Perturbations

Y Su, S Zhang, Z Lin, X Wang, L Zhao, D Meng… - … on Multimedia Modeling, 2024 - Springer
Deep learning models have been proven to be severely affected by adversarial examples,
which limit the widespread deployment of deep learning models. Prior research largely …