Overview frequency principle/spectral bias in deep learning

ZQJ Xu, Y Zhang, T Luo - Communications on Applied Mathematics and …, 2024 - Springer
Understanding deep learning is increasingly emergent as it penetrates more and more into
industry and science. In recent years, a research line from Fourier analysis sheds light on …

A survey on scheduling techniques in computing and network convergence

S Tang, Y Yu, H Wang, G Wang, W Chen… - … Surveys & Tutorials, 2023 - ieeexplore.ieee.org
The computing demand for massive applications has led to the ubiquitous deployment of
computing power. This trend results in the urgent need for higher-level computing resource …

Gradient descent on two-layer nets: Margin maximization and simplicity bias

K Lyu, Z Li, R Wang, S Arora - Advances in Neural …, 2021 - proceedings.neurips.cc
The generalization mystery of overparametrized deep nets has motivated efforts to
understand how gradient descent (GD) converges to low-loss solutions that generalize well …

Provable guarantees for neural networks via gradient feature learning

Z Shi, J Wei, Y Liang - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Neural networks have achieved remarkable empirical performance, while the current
theoretical analysis is not adequate for understanding their success, eg, the Neural Tangent …

Understanding multi-phase optimization dynamics and rich nonlinear behaviors of relu networks

M Wang, C Ma - Advances in Neural Information Processing …, 2024 - proceedings.neurips.cc
The training process of ReLU neural networks often exhibits complicated nonlinear
phenomena. The nonlinearity of models and non-convexity of loss pose significant …

Initialization matters: Privacy-utility analysis of overparameterized neural networks

J Ye, Z Zhu, F Liu, R Shokri… - Advances in Neural …, 2024 - proceedings.neurips.cc
We analytically investigate how over-parameterization of models in randomized machine
learning algorithms impacts the information leakage about their training data. Specifically …

Robustness in deep learning: The good (width), the bad (depth), and the ugly (initialization)

Z Zhu, F Liu, G Chrysos… - Advances in neural …, 2022 - proceedings.neurips.cc
We study the average robustness notion in deep neural networks in (selected) wide and
narrow, deep and shallow, as well as lazy and non-lazy training settings. We prove that in …

Empirical phase diagram for three-layer neural networks with infinite width

H Zhou, Z Qixuan, Z Jin, T Luo… - Advances in Neural …, 2022 - proceedings.neurips.cc
Substantial work indicates that the dynamics of neural networks (NNs) is closely related to
their initialization of parameters. Inspired by the phase diagram for two-layer ReLU NNs with …

Embedding principle: a hierarchical structure of loss landscape of deep neural networks

Y Zhang, Y Li, Z Zhang, T Luo, ZQJ Xu - arXiv preprint arXiv:2111.15527, 2021 - arxiv.org
We prove a general Embedding Principle of loss landscape of deep neural networks (NNs)
that unravels a hierarchical structure of the loss landscape of NNs, ie, loss landscape of an …

Embedding principle of loss landscape of deep neural networks

Y Zhang, Z Zhang, T Luo, ZJ Xu - Advances in Neural …, 2021 - proceedings.neurips.cc
Understanding the structure of loss landscape of deep neural networks (DNNs) is obviously
important. In this work, we prove an embedding principle that the loss landscape of a DNN" …