A high-bias, low-variance introduction to machine learning for physicists

P Mehta, M Bukov, CH Wang, AGR Day, C Richardson… - Physics reports, 2019 - Elsevier
Abstract Machine Learning (ML) is one of the most exciting and dynamic areas of modern
research and application. The purpose of this review is to provide an introduction to the core …

Statistical mechanics of deep learning

Y Bahri, J Kadmon, J Pennington… - Annual Review of …, 2020 - annualreviews.org
The recent striking success of deep neural networks in machine learning raises profound
questions about the theoretical principles underlying their success. For example, what can …

[HTML][HTML] High-dimensional dynamics of generalization error in neural networks

MS Advani, AM Saxe, H Sompolinsky - Neural Networks, 2020 - Elsevier
We perform an analysis of the average generalization dynamics of large neural networks
trained using gradient descent. We study the practically-relevant “high-dimensional” regime …

Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks

A Canatar, B Bordelon, C Pehlevan - Nature communications, 2021 - nature.com
A theoretical understanding of generalization remains an open problem for many machine
learning models, including deep networks where overparameterization leads to better …

Bayes-optimal learning of deep random networks of extensive-width

H Cui, F Krzakala, L Zdeborová - … Conference on Machine …, 2023 - proceedings.mlr.press
We consider the problem of learning a target function corresponding to a deep, extensive-
width, non-linear neural network with random Gaussian weights. We consider the asymptotic …

Dynamics of stochastic gradient descent for two-layer neural networks in the teacher-student setup

S Goldt, M Advani, AM Saxe… - Advances in neural …, 2019 - proceedings.neurips.cc
Deep neural networks achieve stellar generalisation even when they have enough
parameters to easily fit all their training data. We study this phenomenon by analysing the …

[图书][B] Random fields for spatial data modeling

DT Hristopulos - 2020 - Springer
The series aims to: present current and emerging innovations in GIScience; describe new
and robust GIScience methods for use in transdisciplinary problem solving and decision …

Organizing memories for generalization in complementary learning systems

W Sun, M Advani, N Spruston, A Saxe… - Nature …, 2023 - nature.com
Memorization and generalization are complementary cognitive processes that jointly
promote adaptive behavior. For example, animals should memorize safe routes to specific …

Geometry of neural network loss surfaces via random matrix theory

J Pennington, Y Bahri - International conference on machine …, 2017 - proceedings.mlr.press
Understanding the geometry of neural network loss surfaces is important for the
development of improved optimization algorithms and for building a theoretical …

A precise high-dimensional asymptotic theory for boosting and minimum--norm interpolated classifiers

T Liang, P Sur - The Annals of Statistics, 2022 - projecteuclid.org
A precise high-dimensional asymptotic theory for boosting and minimum-l1-norm
interpolated classifiers Page 1 The Annals of Statistics 2022, Vol. 50, No. 3, 1669–1695 …