[HTML][HTML] Cleaning large correlation matrices: tools from random matrix theory

J Bun, JP Bouchaud, M Potters - Physics Reports, 2017 - Elsevier
This review covers recent results concerning the estimation of large covariance matrices
using tools from Random Matrix Theory (RMT). We introduce several RMT methods and …

Random matrix theory in statistics: A review

D Paul, A Aue - Journal of Statistical Planning and Inference, 2014 - Elsevier
We give an overview of random matrix theory (RMT) with the objective of highlighting the
results and concepts that have a growing impact in the formulation and inference of …

High-dimensional asymptotics of feature learning: How one gradient step improves the representation

J Ba, MA Erdogdu, T Suzuki, Z Wang… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the first gradient descent step on the first-layer parameters $\boldsymbol {W} $ in a
two-layer neural network: $ f (\boldsymbol {x})=\frac {1}{\sqrt {N}}\boldsymbol {a}^\top\sigma …

Learning in the presence of low-dimensional structure: a spiked random matrix perspective

J Ba, MA Erdogdu, T Suzuki… - Advances in Neural …, 2024 - proceedings.neurips.cc
We consider the learning of a single-index target function $ f_*:\mathbb {R}^ d\to\mathbb {R}
$ under spiked covariance data: $$ f_*(\boldsymbol {x})=\textstyle\sigma_*(\frac {1}{\sqrt …

Fundamental limits in structured principal component analysis and how to reach them

J Barbier, F Camilli, M Mondelli… - Proceedings of the …, 2023 - National Acad Sciences
How do statistical dependencies in measurement noise influence high-dimensional
inference? To answer this, we study the paradigmatic spiked matrix model of principal …

[HTML][HTML] High-dimensional dynamics of generalization error in neural networks

MS Advani, AM Saxe, H Sompolinsky - Neural Networks, 2020 - Elsevier
We perform an analysis of the average generalization dynamics of large neural networks
trained using gradient descent. We study the practically-relevant “high-dimensional” regime …

SC3: consensus clustering of single-cell RNA-seq data

VY Kiselev, K Kirschner, MT Schaub, T Andrews… - Nature …, 2017 - nature.com
Single-cell RNA-seq enables the quantitative characterization of cell types based on global
transcriptome profiles. We present single-cell consensus clustering (SC3), a user-friendly …

High-dimensional limit theorems for sgd: Effective dynamics and critical scaling

G Ben Arous, R Gheissari… - Advances in Neural …, 2022 - proceedings.neurips.cc
We study the scaling limits of stochastic gradient descent (SGD) with constant step-size in
the high-dimensional regime. We prove limit theorems for the trajectories of summary …

[HTML][HTML] Entrywise eigenvector analysis of random matrices with low expected rank

E Abbe, J Fan, K Wang, Y Zhong - Annals of statistics, 2020 - ncbi.nlm.nih.gov
Recovering low-rank structures via eigenvector perturbation analysis is a common problem
in statistical machine learning, such as in factor analysis, community detection, ranking …

Factors that fit the time series and cross-section of stock returns

M Lettau, M Pelger - The Review of Financial Studies, 2020 - academic.oup.com
We propose a new method for estimating latent asset pricing factors that fit the time series
and cross-section of expected returns. Our estimator generalizes principal component …