High-dimensional asymptotics of feature learning: How one gradient step improves the representation

J Ba, MA Erdogdu, T Suzuki, Z Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the first gradient descent step on the first-layer parameters $\boldsymbol {W} $ in a
two-layer neural network: $ f (\boldsymbol {x})=\frac {1}{\sqrt {N}}\boldsymbol {a}^\top\sigma …

Task arithmetic in the tangent space: Improved editing of pre-trained models

G Ortiz-Jimenez, A Favero… - Advances in Neural …, 2024 - proceedings.neurips.cc
Task arithmetic has recently emerged as a cost-effective and scalable approach to edit pre-
trained models directly in weight space: By adding the fine-tuned weights of different tasks …

A structured dictionary perspective on implicit neural representations

G Yüce, G Ortiz-Jiménez… - Proceedings of the …, 2022 - openaccess.thecvf.com
Implicit neural representations (INRs) have recently emerged as a promising alternative to
classical discretized representations of signals. Nevertheless, despite their practical …

Initialization matters: Privacy-utility analysis of overparameterized neural networks

J Ye, Z Zhu, F Liu, R Shokri… - Advances in Neural …, 2024 - proceedings.neurips.cc
We analytically investigate how over-parameterization of models in randomized machine
learning algorithms impacts the information leakage about their training data. Specifically …

Feature-learning networks are consistent across widths at realistic scales

N Vyas, A Atanasov, B Bordelon… - Advances in …, 2024 - proceedings.neurips.cc
We study the effect of width on the dynamics of feature-learning neural networks across a
variety of architectures and datasets. Early in training, wide neural networks trained on …

What can the neural tangent kernel tell us about adversarial robustness?

N Tsilivis, J Kempe - Advances in Neural Information …, 2022 - proceedings.neurips.cc
The adversarial vulnerability of neural nets, and subsequent techniques to create robust
models have attracted significant attention; yet we still lack a full understanding of this …

Limitations of the ntk for understanding generalization in deep learning

N Vyas, Y Bansal, P Nakkiran - arXiv preprint arXiv:2206.10012, 2022 - arxiv.org
The``Neural Tangent Kernel''(NTK)(Jacot et al 2018), and its empirical variants have been
proposed as a proxy to capture certain behaviors of real neural networks. In this work, we …

Spectral evolution and invariance in linear-width neural networks

Z Wang, A Engel, AD Sarwate… - Advances in Neural …, 2024 - proceedings.neurips.cc
We investigate the spectral properties of linear-width feed-forward neural networks, where
the sample size is asymptotically proportional to network width. Empirically, we show that the …

Learning sparse features can lead to overfitting in neural networks

L Petrini, F Cagnetta… - Advances in Neural …, 2022 - proceedings.neurips.cc
It is widely believed that the success of deep networks lies in their ability to learn a
meaningful representation of the features of the data. Yet, understanding when and how this …

Evolution of neural tangent kernels under benign and adversarial training

N Loo, R Hasani, A Amini… - Advances in Neural …, 2022 - proceedings.neurips.cc
Two key challenges facing modern deep learning is mitigating deep networks vulnerability
to adversarial attacks, and understanding deep learning's generalization capabilities …