Improved generalization bounds for robust learning

I Attias, A Kontorovich… - Algorithmic Learning …, 2019 - proceedings.mlr.press
We consider a model of robust learning in an adversarial environment. The learner gets
uncorrupted training data with access to possible corruptions that may be effected by the …

Sparsity winning twice: Better robust generalization from more efficient training

T Chen, Z Zhang, P Wang, S Balachandra… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent studies demonstrate that deep networks, even robustified by the state-of-the-art
adversarial training (AT), still suffer from large robust generalization gaps, in addition to the …

Adversarial robustness of supervised sparse coding

J Sulam, R Muthukumar… - Advances in neural …, 2020 - proceedings.neurips.cc
Several recent results provide theoretical insights into the phenomena of adversarial
examples. Existing results, however, are often limited due to a gap between the simplicity of …

Achieving adversarial robustness via sparsity

N Liao, S Wang, L Xiang, N Ye, S Shao, P Chu - Machine Learning, 2022 - Springer
Network pruning has been known to produce compact models without much accuracy
degradation. However, how the pruning process affects a network's robustness and the …

On achieving optimal adversarial test error

JD Li, M Telgarsky - arXiv preprint arXiv:2306.07544, 2023 - arxiv.org
We first elucidate various fundamental properties of optimal adversarial predictors: the
structure of optimal adversarial convex predictors in terms of optimal adversarial zero-one …

Improved generalization bounds for adversarially robust learning

I Attias, A Kontorovich, Y Mansour - Journal of Machine Learning Research, 2022 - jmlr.org
We consider a model of robust learning in an adversarial environment. The learner gets
uncorrupted training data with access to possible corruptions that may be affected by the …

Benign Overfitting in Adversarial Training of Neural Networks

Y Wang, K Zhang, R Arora - Forty-first International Conference on Machine … - openreview.net
Benign overfitting is the phenomenon wherein none of the predictors in the hypothesis class
can achieve perfect accuracy (ie, non-realizable or noisy setting), but a model that …

[PDF][PDF] Advances in Robust Statistical Learning Theory

I Attias - 2024 - tau.ac.il
Abstract Machine learning techniques were initially designed for stationary and benign
environments, where the training and test data are assumed to be generated from the same …

Stability and Generalization of Adversarial Training for Shallow Neural Networks with Smooth Activation

K Zhang, Y Wang, R Arora - The Thirty-eighth Annual Conference on … - openreview.net
Adversarial training has emerged as a popular approach for training models that are robust
to inference-time adversarial attacks. However, our theoretical understanding of why and …

Robustness analysis of deep neural networks in the presence of adversarial perturbations and noisy labels

ER Balda Canizares, R Mathar, B Leibe - 2020 - publications.rwth-aachen.de
In dieser Arbeit untersuchen wir die Robustheit und Verallgemeinerungseigenschaften von
Deep Neural Networks (DNNs) unter verschiedenen rauschbehafteten Bedingungen, die …