Attacks which do not kill training make adversarial learning stronger

J Zhang, X Xu, B Han, G Niu, L Cui… - International …, 2020 - proceedings.mlr.press
Adversarial training based on the minimax formulation is necessary for obtaining adversarial
robustness of trained models. However, it is conservative or even pessimistic so that it …

Second-order adversarial attack and certifiable robustness

B Li, C Chen, W Wang, L Carin - 2018 - openreview.net
Adversarial training has been recognized as a strong defense against adversarial attacks. In
this paper, we propose a powerful second-order attack method that reduces the accuracy of …

Analysis and applications of class-wise robustness in adversarial training

Q Tian, K Kuang, K Jiang, F Wu, Y Wang - Proceedings of the 27th ACM …, 2021 - dl.acm.org
Adversarial training is one of the most effective approaches to improve model robustness
against adversarial examples. However, previous works mainly focus on the overall …

Improving adversarial robustness requires revisiting misclassified examples

Y Wang, D Zou, J Yi, J Bailey, X Ma… - … conference on learning …, 2019 - openreview.net
Deep neural networks (DNNs) are vulnerable to adversarial examples crafted by
imperceptible perturbations. A range of defense techniques have been proposed to improve …

WAT: improve the worst-class robustness in adversarial training

B Li, W Liu - Proceedings of the AAAI conference on artificial …, 2023 - ojs.aaai.org
Abstract Deep Neural Networks (DNN) have been shown to be vulnerable to adversarial
examples. Adversarial training (AT) is a popular and effective strategy to defend against …

Are labels required for improving adversarial robustness?

JB Alayrac, J Uesato, PS Huang… - Advances in …, 2019 - proceedings.neurips.cc
Recent work has uncovered the interesting (and somewhat surprising) finding that training
models to be invariant to adversarial perturbations requires substantially larger datasets …

Understanding catastrophic overfitting in single-step adversarial training

H Kim, W Lee, J Lee - Proceedings of the AAAI Conference on Artificial …, 2021 - ojs.aaai.org
Although fast adversarial training has demonstrated both robustness and efficiency, the
problem of" catastrophic overfitting" has been observed. This is a phenomenon in which …

To be robust or to be fair: Towards fairness in adversarial training

H Xu, X Liu, Y Li, A Jain, J Tang - … conference on machine …, 2021 - proceedings.mlr.press
Adversarial training algorithms have been proved to be reliable to improve machine learning
models' robustness against adversarial examples. However, we find that adversarial training …

Rademacher complexity for adversarially robust generalization

D Yin, R Kannan, P Bartlett - International conference on …, 2019 - proceedings.mlr.press
Many machine learning models are vulnerable to adversarial attacks; for example, adding
adversarial perturbations that are imperceptible to humans can often make machine …

Towards understanding fast adversarial training

B Li, S Wang, S Jana, L Carin - arXiv preprint arXiv:2006.03089, 2020 - arxiv.org
Current neural-network-based classifiers are susceptible to adversarial examples. The most
empirically successful approach to defending against such adversarial examples is …