The modern mathematics of deep learning

J Berner, P Grohs, G Kutyniok… - arXiv preprint arXiv …, 2021 - cambridge.org
We describe the new field of the mathematical analysis of deep learning. This field emerged
around a list of research questions that were not answered within the classical framework of …

The unbearable shallow understanding of deep learning

A Plebe, G Grasso - Minds and Machines, 2019 - Springer
This paper analyzes the rapid and unexpected rise of deep learning within Artificial
Intelligence and its applications. It tackles the possible reasons for this remarkable success …

Deep learning without poor local minima

K Kawaguchi - Advances in neural information processing …, 2016 - proceedings.neurips.cc
In this paper, we prove a conjecture published in 1989 and also partially address an open
problem announced at the Conference on Learning Theory (COLT) 2015. For an expected …

Where is the information in a deep neural network?

A Achille, G Paolini, S Soatto - arXiv preprint arXiv:1905.12213, 2019 - arxiv.org
Whatever information a deep neural network has gleaned from training data is encoded in
its weights. How this information affects the response of the network to future data remains …

Sharp minima can generalize for deep nets

L Dinh, R Pascanu, S Bengio… - … Conference on Machine …, 2017 - proceedings.mlr.press
Despite their overwhelming capacity to overfit, deep learning architectures tend to
generalize relatively well to unseen data, allowing them to be deployed in practice …

Theoretical issues in deep networks

T Poggio, A Banburski, Q Liao - Proceedings of the …, 2020 - National Acad Sciences
While deep learning is successful in a number of applications, it is not yet well understood
theoretically. A theoretical characterization of deep learning should answer questions about …

Geometry of optimization and implicit regularization in deep learning

B Neyshabur, R Tomioka, R Salakhutdinov… - arXiv preprint arXiv …, 2017 - arxiv.org
We argue that the optimization plays a crucial role in generalization of deep learning models
through implicit regularization. We do this by demonstrating that generalization ability is not …

[HTML][HTML] Depth with nonlinearity creates no bad local minima in ResNets

K Kawaguchi, Y Bengio - Neural Networks, 2019 - Elsevier
In this paper, we prove that depth with nonlinearity creates no bad local minima in a type of
arbitrarily deep ResNets with arbitrary nonlinear activation functions, in the sense that the …

Full error analysis for the training of deep neural networks

C Beck, A Jentzen, B Kuckuck - Infinite Dimensional Analysis …, 2022 - World Scientific
Deep learning algorithms have been applied very successfully in recent years to a range of
problems out of reach for classical solution paradigms. Nevertheless, there is no completely …

[图书][B] The principles of deep learning theory

DA Roberts, S Yaida, B Hanin - 2022 - cambridge.org
This textbook establishes a theoretical framework for understanding deep learning models
of practical relevance. With an approach that borrows from theoretical physics, Roberts and …