XW Chen, X Lin - IEEE access, 2014 - ieeexplore.ieee.org
Deep learning is currently an extremely active research area in machine learning and pattern recognition society. It has gained huge successes in a broad area of applications …
We perform an analysis of the average generalization dynamics of large neural networks trained using gradient descent. We study the practically-relevant “high-dimensional” regime …
We study the connection between the highly non-convex loss function of a simple model of the fully-connected feed-forward neural network and the Hamiltonian of the spherical spin …
Understanding the reasons for the success of deep neural networks trained using stochastic gradient-based methods is a key open problem for the nascent theory of deep learning. The …
The convergence of back-propagation learning is analyzed so as to explain common phenomenon observedb y practitioners. Many undesirable behaviors of backprop can be …
J Pennington, P Worah - Advances in neural information …, 2017 - proceedings.neurips.cc
Neural network configurations with random weights play an important role in the analysis of deep learning. They define the initial loss landscape and are closely related to kernel and …
Understanding the impact of data structure on the computational tractability of learning is a key challenge for the theory of neural networks. Many theoretical works do not explicitly …
S Lee, S Goldt, A Saxe - International Conference on …, 2021 - proceedings.mlr.press
Continual learning {—} the ability to learn many tasks in sequence {—} is critical for artificial learning systems. Yet standard training methods for deep networks often suffer from …
A recent series of theoretical works showed that the dynamics of neural networks with a certain initialisation are well-captured by kernel methods. Concurrent empirical work …