Robust overfitting may be mitigated by properly learned smoothening

T Chen, Z Zhang, S Liu, S Chang… - … Conference on Learning …, 2020 - openreview.net
A recent study (Rice et al., 2020) revealed overfitting to be a dominant phenomenon in
adversarially robust training of deep networks, and that appropriate early-stopping of …

Training your sparse neural network better with any mask

AK Jaiswal, H Ma, T Chen, Y Ding… - … on Machine Learning, 2022 - proceedings.mlr.press
Pruning large neural networks to create high-quality, independently trainable sparse masks,
which can maintain similar performance to their dense counterparts, is very desirable due to …

Relative flatness and generalization

H Petzka, M Kamp, L Adilova… - Advances in neural …, 2021 - proceedings.neurips.cc
Flatness of the loss curve is conjectured to be connected to the generalization ability of
machine learning models, in particular neural networks. While it has been empirically …

Why flatness does and does not correlate with generalization for deep neural networks

S Zhang, I Reid, GV Pérez, A Louis - arXiv preprint arXiv:2103.06219, 2021 - arxiv.org
The intuition that local flatness of the loss landscape is correlated with better generalization
for deep neural networks (DNNs) has been explored for decades, spawning many different …

Reparameterization invariance in approximate Bayesian inference

H Roy, M Miani, CH Ek, P Hennig, M Pförtner… - arXiv preprint arXiv …, 2024 - arxiv.org
Current approximate posteriors in Bayesian neural networks (BNNs) exhibit a crucial
limitation: they fail to maintain invariance under reparameterization, ie BNNs assign different …

Understanding Adversarially Robust Generalization via Weight-Curvature Index

Y Xu, X Zhang - arXiv preprint arXiv:2410.07719, 2024 - arxiv.org
Despite extensive research on adversarial examples, the underlying mechanisms of
adversarially robust generalization, a critical yet challenging task for deep learning, remain …

[PDF][PDF] Studying Solution Quality for Ill-Posed Optimization Problems

C Horváth - 2024 - diva-portal.org
This thesis investigates state-of-the-art methods and underlying theories in supervised deep
learning, aiming to apply insights from these techniques to broader inverse problem settings …

Network Parameterisation and Activation Functions in Deep Learning

M Trimmel - 2023 - portal.research.lu.se
Deep learning, the study of multi-layered artificial neural networks, has received tremendous
attention over the course of the last few years. Neural networks are now able to outperform …

[PDF][PDF] How does training deep neural networks on a biased dataset affect the loss landscape of the network?

A Mohammadi, AK Rana - 2021 - amitrana001.github.io
Many studies show a positive correlation between the generalization ability of a deep neural
network and the flatness of the minima in its loss landscape. Inspired by this statement …