Recent advances in adversarial training for adversarial robustness

T Bai, J Luo, J Zhao, B Wen, Q Wang - arXiv preprint arXiv:2102.01356, 2021 - arxiv.org
Adversarial training is one of the most effective approaches defending against adversarial
examples for deep learning models. Unlike other defense strategies, adversarial training …

Adversarial training methods for deep learning: A systematic review

W Zhao, S Alwidian, QH Mahmoud - Algorithms, 2022 - mdpi.com
Deep neural networks are exposed to the risk of adversarial attacks via the fast gradient sign
method (FGSM), projected gradient descent (PGD) attacks, and other attack algorithms …

Towards efficient adversarial training on vision transformers

B Wu, J Gu, Z Li, D Cai, X He, W Liu - European Conference on Computer …, 2022 - Springer
Abstract Vision Transformer (ViT), as a powerful alternative to Convolutional Neural Network
(CNN), has received much attention. Recent work showed that ViTs are also vulnerable to …

Subspace adversarial training

T Li, Y Wu, S Chen, K Fang… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Single-step adversarial training (AT) has received wide attention as it proved to be both
efficient and robust. However, a serious problem of catastrophic overfitting exists, ie, the …

Prior-guided adversarial initialization for fast adversarial training

X Jia, Y Zhang, X Wei, B Wu, K Ma, J Wang… - European Conference on …, 2022 - Springer
Fast adversarial training (FAT) effectively improves the efficiency of standard adversarial
training (SAT). However, initial FAT encounters catastrophic overfitting, ie, the robust …

Make some noise: Reliable and efficient single-step adversarial training

P de Jorge Aranda, A Bibi, R Volpi… - Advances in …, 2022 - proceedings.neurips.cc
Abstract Recently, Wong et al.(2020) showed that adversarial training with single-step FGSM
leads to a characteristic failure mode named catastrophic overfitting (CO), in which a model …

Boosting fast adversarial training with learnable adversarial initialization

X Jia, Y Zhang, B Wu, J Wang… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Adversarial training (AT) has been demonstrated to be effective in improving model
robustness by leveraging adversarial examples for training. However, most AT methods are …

Adversarial attacks and defenses in deep learning for image recognition: A survey

J Wang, C Wang, Q Lin, C Luo, C Wu, J Li - Neurocomputing, 2022 - Elsevier
In recent years, researches on adversarial attacks and defense mechanisms have obtained
much attention. It's observed that adversarial examples crafted with small malicious …

[HTML][HTML] Understanding and combating robust overfitting via input loss landscape analysis and regularization

L Li, M Spratling - Pattern Recognition, 2023 - Elsevier
Adversarial training is widely used to improve the robustness of deep neural networks to
adversarial attack. However, adversarial training is prone to overfitting, and the cause is far …

The enemy of my enemy is my friend: Exploring inverse adversaries for improving adversarial training

J Dong, SM Moosavi-Dezfooli… - Proceedings of the …, 2023 - openaccess.thecvf.com
Although current deep learning techniques have yielded superior performance on various
computer vision tasks, yet they are still vulnerable to adversarial examples. Adversarial …