Cross-entropy loss functions: Theoretical analysis and applications

A Mao, M Mohri, Y Zhong - International conference on …, 2023 - proceedings.mlr.press
Cross-entropy is a widely used loss function in applications. It coincides with the logistic loss
applied to the outputs of a neural network, when the softmax is used. But, what guarantees …

Adversarial weight perturbation helps robust generalization

D Wu, ST Xia, Y Wang - Advances in neural information …, 2020 - proceedings.neurips.cc
The study on improving the robustness of deep neural networks against adversarial
examples grows rapidly in recent years. Among them, adversarial training is the most …

Adversarial training for free!

A Shafahi, M Najibi, MA Ghiasi, Z Xu… - Advances in neural …, 2019 - proceedings.neurips.cc
Adversarial training, in which a network is trained on adversarial examples, is one of the few
defenses against adversarial attacks that withstands strong attacks. Unfortunately, the high …

Bridging mode connectivity in loss landscapes and adversarial robustness

P Zhao, PY Chen, P Das, KN Ramamurthy… - arXiv preprint arXiv …, 2020 - arxiv.org
Mode connectivity provides novel geometric insights on analyzing loss landscapes and
enables building high-accuracy pathways between well-trained neural networks. In this …

Adversarial defense by restricting the hidden space of deep neural networks

A Mustafa, S Khan, M Hayat… - Proceedings of the …, 2019 - openaccess.thecvf.com
Deep neural networks are vulnerable to adversarial attacks which can fool them by adding
minuscule perturbations to the input images. The robustness of existing defenses suffers …

Relating adversarially robust generalization to flat minima

D Stutz, M Hein, B Schiele - Proceedings of the IEEE/CVF …, 2021 - openaccess.thecvf.com
Adversarial training (AT) has become the de-facto standard to obtain models robust against
adversarial examples. However, AT exhibits severe robust overfitting: cross-entropy loss on …

Exploring the differences in adversarial robustness between ViT-and CNN-based models using novel metrics

J Heo, S Seo, P Kang - Computer Vision and Image Understanding, 2023 - Elsevier
Deep-learning models have demonstrated remarkable performance in a variety of fields,
owing to advancements in computational power and the availability of extensive datasets for …

Deeply supervised discriminative learning for adversarial defense

A Mustafa, SH Khan, M Hayat, R Goecke… - IEEE transactions on …, 2020 - ieeexplore.ieee.org
Deep neural networks can easily be fooled by an adversary with minuscule perturbations
added to an input image. The existing defense techniques suffer greatly under white-box …

Rethinking the effect of data augmentation in adversarial contrastive learning

R Luo, Y Wang, Y Wang - arXiv preprint arXiv:2303.01289, 2023 - arxiv.org
Recent works have shown that self-supervised learning can achieve remarkable robustness
when integrated with adversarial training (AT). However, the robustness gap between …

Inverse-reference priors for fisher regularization of Bayesian neural networks

K Kim, EY Ma, J Choi, H Kim - Proceedings of the AAAI Conference on …, 2023 - ojs.aaai.org
Recent studies have shown that the generalization ability of deep neural networks (DNNs) is
closely related to the Fisher information matrix (FIM) calculated during the early training …