Recent advances in adversarial training for adversarial robustness

T Bai, J Luo, J Zhao, B Wen, Q Wang - arXiv preprint arXiv:2102.01356, 2021 - arxiv.org
Adversarial training is one of the most effective approaches defending against adversarial
examples for deep learning models. Unlike other defense strategies, adversarial training …

A survey of adversarial defenses and robustness in nlp

S Goyal, S Doddapaneni, MM Khapra… - ACM Computing …, 2023 - dl.acm.org
In the past few years, it has become increasingly evident that deep neural networks are not
resilient enough to withstand adversarial perturbations in input data, leaving them …

3d common corruptions and data augmentation

OF Kar, T Yeo, A Atanov… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
We introduce a set of image transformations that can be used as corruptions to evaluate the
robustness of models as well as data augmentation mechanisms for training neural …

The pitfalls of simplicity bias in neural networks

H Shah, K Tamuly, A Raghunathan… - Advances in …, 2020 - proceedings.neurips.cc
Several works have proposed Simplicity Bias (SB)---the tendency of standard training
procedures such as Stochastic Gradient Descent (SGD) to find simple models---to justify why …

Dverge: diversifying vulnerabilities for enhanced robust generation of ensembles

H Yang, J Zhang, H Dong… - Advances in …, 2020 - proceedings.neurips.cc
Recent research finds CNN models for image classification demonstrate overlapped
adversarial vulnerabilities: adversarial attacks can mislead CNN models with small …

[HTML][HTML] Adversarial training methods for deep learning: A systematic review

W Zhao, S Alwidian, QH Mahmoud - Algorithms, 2022 - mdpi.com
Deep neural networks are exposed to the risk of adversarial attacks via the fast gradient sign
method (FGSM), projected gradient descent (PGD) attacks, and other attack algorithms …

Evading the simplicity bias: Training a diverse set of models discovers solutions with superior ood generalization

D Teney, E Abbasnejad, S Lucey… - Proceedings of the …, 2022 - openaccess.thecvf.com
Neural networks trained with SGD were recently shown to rely preferentially on linearly-
predictive features and can ignore complex, equally-predictive ones. This simplicity bias can …

An adaptive model ensemble adversarial attack for boosting adversarial transferability

B Chen, J Yin, S Chen, B Chen… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
While the transferability property of adversarial examples allows the adversary to perform
black-box attacks ie, the attacker has no knowledge about the target model), the transfer …

Adversarial attacks and defenses in deep learning for image recognition: A survey

J Wang, C Wang, Q Lin, C Luo, C Wu, J Li - Neurocomputing, 2022 - Elsevier
In recent years, researches on adversarial attacks and defense mechanisms have obtained
much attention. It's observed that adversarial examples crafted with small malicious …

Towards a robust deep neural network against adversarial texts: A survey

W Wang, R Wang, L Wang, Z Wang… - ieee transactions on …, 2021 - ieeexplore.ieee.org
Deep neural networks (DNNs) have achieved remarkable success in various tasks (eg,
image classification, speech recognition, and natural language processing (NLP)). However …