[HTML][HTML] Cleaning large correlation matrices: tools from random matrix theory

J Bun, JP Bouchaud, M Potters - Physics Reports, 2017 - Elsevier
This review covers recent results concerning the estimation of large covariance matrices
using tools from Random Matrix Theory (RMT). We introduce several RMT methods and …

On simplicity and complexity in the brave new world of large-scale neuroscience

P Gao, S Ganguli - Current opinion in neurobiology, 2015 - Elsevier
Technological advances have dramatically expanded our ability to probe multi-neuronal
dynamics and connectivity in the brain. However, our ability to extract a simple conceptual …

High-dimensional asymptotics of feature learning: How one gradient step improves the representation

J Ba, MA Erdogdu, T Suzuki, Z Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the first gradient descent step on the first-layer parameters $\boldsymbol {W} $ in a
two-layer neural network: $ f (\boldsymbol {x})=\frac {1}{\sqrt {N}}\boldsymbol {a}^\top\sigma …

[HTML][HTML] High-dimensional dynamics of generalization error in neural networks

MS Advani, AM Saxe, H Sompolinsky - Neural Networks, 2020 - Elsevier
We perform an analysis of the average generalization dynamics of large neural networks
trained using gradient descent. We study the practically-relevant “high-dimensional” regime …

[HTML][HTML] Complex diffusion-weighted image estimation via matrix recovery under general noise models

L Cordero-Grande, D Christiaens, J Hutter, AN Price… - Neuroimage, 2019 - Elsevier
We propose a patch-based singular value shrinkage method for diffusion magnetic
resonance image estimation targeted at low signal to noise ratio and accelerated …

The Optimal Hard Threshold for Singular Values is

M Gavish, DL Donoho - IEEE Transactions on Information …, 2014 - ieeexplore.ieee.org
We consider recovery of low-rank matrices from noisy data by hard thresholding of singular
values, in which empirical singular values below a threshold λ are set to 0. We study the …

The gaussian equivalence of generative models for learning with shallow neural networks

S Goldt, B Loureiro, G Reeves… - Mathematical and …, 2022 - proceedings.mlr.press
Understanding the impact of data structure on the computational tractability of learning is a
key challenge for the theory of neural networks. Many theoretical works do not explicitly …

A statistical model for tensor PCA

E Richard, A Montanari - Advances in neural information …, 2014 - proceedings.neurips.cc
Abstract We consider the Principal Component Analysis problem for large tensors of
arbitrary order k under a single-spike (or rank-one plus noise) model. On the one hand, we …

Theoretical foundations of t-sne for visualizing high-dimensional clustered data

TT Cai, R Ma - Journal of Machine Learning Research, 2022 - jmlr.org
This paper investigates the theoretical foundations of the t-distributed stochastic neighbor
embedding (t-SNE) algorithm, a popular nonlinear dimension reduction and data …

Stochastic collapse: How gradient noise attracts sgd dynamics towards simpler subnetworks

F Chen, D Kunin, A Yamamura… - Advances in Neural …, 2024 - proceedings.neurips.cc
In this work, we reveal a strong implicit bias of stochastic gradient descent (SGD) that drives
overly expressive networks to much simpler subnetworks, thereby dramatically reducing the …