Sam as an optimal relaxation of bayes

T Möllenhoff, ME Khan - arXiv preprint arXiv:2210.01620, 2022 - arxiv.org
arXiv preprint arXiv:2210.01620, 2022arxiv.org
Sharpness-aware minimization (SAM) and related adversarial deep-learning methods can
drastically improve generalization, but their underlying mechanisms are not yet fully
understood. Here, we establish SAM as a relaxation of the Bayes objective where the
expected negative-loss is replaced by the optimal convex lower bound, obtained by using
the so-called Fenchel biconjugate. The connection enables a new Adam-like extension of
SAM to automatically obtain reasonable uncertainty estimates, while sometimes also …
Sharpness-aware minimization (SAM) and related adversarial deep-learning methods can drastically improve generalization, but their underlying mechanisms are not yet fully understood. Here, we establish SAM as a relaxation of the Bayes objective where the expected negative-loss is replaced by the optimal convex lower bound, obtained by using the so-called Fenchel biconjugate. The connection enables a new Adam-like extension of SAM to automatically obtain reasonable uncertainty estimates, while sometimes also improving its accuracy. By connecting adversarial and Bayesian methods, our work opens a new path to robustness.
arxiv.org
以上显示的是最相近的搜索结果。 查看全部搜索结果