Provable guarantees for neural networks via gradient feature learning

Z Shi, J Wei, Y Liang - Advances in Neural Information …, 2023 - proceedings.neurips.cc
Neural networks have achieved remarkable empirical performance, while the current
theoretical analysis is not adequate for understanding their success, eg, the Neural Tangent …

Provable guarantees for nonlinear feature learning in three-layer neural networks

E Nichani, A Damian, JD Lee - Advances in Neural …, 2024 - proceedings.neurips.cc
One of the central questions in the theory of deep learning is to understand how neural
networks learn hierarchical features. The ability of deep networks to extract salient features …

Learning hierarchical polynomials with three-layer neural networks

Z Wang, E Nichani, JD Lee - arXiv preprint arXiv:2311.13774, 2023 - arxiv.org
We study the problem of learning hierarchical polynomials over the standard Gaussian
distribution with three-layer neural networks. We specifically consider target functions of the …

Rethink depth separation with intra-layer links

FL Fan, ZY Li, H Xiong, T Zeng - arXiv preprint arXiv:2305.07037, 2023 - arxiv.org
The depth separation theory is nowadays widely accepted as an effective explanation for the
power of depth, which consists of two parts: i) there exists a function representable by a …

Depth Separations in Neural Networks: Separating the Dimension from the Accuracy

I Safran, D Reichman, P Valiant - arXiv preprint arXiv:2402.07248, 2024 - arxiv.org
We prove an exponential separation between depth 2 and depth 3 neural networks, when
approximating an $\mathcal {O}(1) $-Lipschitz target function to constant accuracy, with …