[HTML][HTML] Cleaning large correlation matrices: tools from random matrix theory

J Bun, JP Bouchaud, M Potters - Physics Reports, 2017 - Elsevier
This review covers recent results concerning the estimation of large covariance matrices
using tools from Random Matrix Theory (RMT). We introduce several RMT methods and …

The stability–complexity relationship at age 40: a random matrix perspective

S Allesina, S Tang - Population Ecology, 2015 - Wiley Online Library
Since the work of Robert May in 1972, the local asymptotic stability of large ecological
systems has been a focus of theoretical ecology. Here we review May's work in the light of …

High-dimensional asymptotics of feature learning: How one gradient step improves the representation

J Ba, MA Erdogdu, T Suzuki, Z Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the first gradient descent step on the first-layer parameters $\boldsymbol {W} $ in a
two-layer neural network: $ f (\boldsymbol {x})=\frac {1}{\sqrt {N}}\boldsymbol {a}^\top\sigma …

The generalization error of random features regression: Precise asymptotics and the double descent curve

S Mei, A Montanari - Communications on Pure and Applied …, 2022 - Wiley Online Library
Deep learning methods operate in regimes that defy the traditional statistical mindset.
Neural network architectures often contain more parameters than training samples, and are …

[HTML][HTML] Surprises in high-dimensional ridgeless least squares interpolation

T Hastie, A Montanari, S Rosset, RJ Tibshirani - Annals of statistics, 2022 - ncbi.nlm.nih.gov
Interpolators—estimators that achieve zero training error—have attracted growing attention
in machine learning, mainly because state-of-the art neural networks appear to be models of …

Randomized numerical linear algebra: Foundations and algorithms

PG Martinsson, JA Tropp - Acta Numerica, 2020 - cambridge.org
This survey describes probabilistic algorithms for linear algebraic computations, such as
factorizing matrices and solving linear systems. It focuses on techniques that have a proven …

[HTML][HTML] High-dimensional dynamics of generalization error in neural networks

MS Advani, AM Saxe, H Sompolinsky - Neural Networks, 2020 - Elsevier
We perform an analysis of the average generalization dynamics of large neural networks
trained using gradient descent. We study the practically-relevant “high-dimensional” regime …

[图书][B] Free probability and random matrices

JA Mingo, R Speicher - 2017 - Springer
This book is an invitation to the world of free probability theory. Free probability is a quite
young mathematical theory with many avatars. It owes its existence to the visions of one …

An introduction to matrix concentration inequalities

JA Tropp - Foundations and Trends® in Machine Learning, 2015 - nowpublishers.com
Random matrices now play a role in many areas of theoretical, applied, and computational
mathematics. Therefore, it is desirable to have tools for studying random matrices that are …

Spectral bias and task-model alignment explain generalization in kernel regression and infinitely wide neural networks

A Canatar, B Bordelon, C Pehlevan - Nature communications, 2021 - nature.com
A theoretical understanding of generalization remains an open problem for many machine
learning models, including deep networks where overparameterization leads to better …