High-dimensional analysis of double descent for linear regression with random projections

F Bach - SIAM Journal on Mathematics of Data Science, 2024 - SIAM
We consider linear regression problems with a varying number of random projections,
where we provably exhibit a double descent curve for a fixed prediction problem, with a high …

Generalized equivalences between subsampling and ridge regularization

P Patil, JH Du - Advances in Neural Information Processing …, 2024 - proceedings.neurips.cc
We establish precise structural and risk equivalences between subsampling and ridge
regularization for ensemble ridge estimators. Specifically, we prove that linear and quadratic …

Bagging provides assumption-free stability

JA Soloff, RF Barber, R Willett - Journal of Machine Learning Research, 2024 - jmlr.org
Bagging is an important technique for stabilizing machine learning models. In this paper, we
derive a finite-sample guarantee on the stability of bagging for any model. Our result places …

Fine-grained Analysis and Faster Algorithms for Iteratively Solving Linear Systems

M Dereziński, D LeJeune, D Needell… - arXiv preprint arXiv …, 2024 - arxiv.org
While effective in practice, iterative methods for solving large systems of linear equations
can be significantly affected by problem-dependent condition number quantities. This makes …

Asymptotically free sketched ridge ensembles: Risks, cross-validation, and tuning

P Patil, D LeJeune - arXiv preprint arXiv:2310.04357, 2023 - arxiv.org
We employ random matrix theory to establish consistency of generalized cross validation
(GCV) for estimating prediction risks of sketched ridge regression ensembles, enabling …

Optimal Ridge Regularization for Out-of-Distribution Prediction

P Patil, JH Du, RJ Tibshirani - arXiv preprint arXiv:2404.01233, 2024 - arxiv.org
We study the behavior of optimal ridge regularization and optimal ridge risk for out-of-
distribution prediction, where the test distribution deviates arbitrarily from the train …

Bagging provides assumption-free stability

JA Soloff, RF Barber, R Willett - arXiv preprint arXiv:2301.12600, 2023 - arxiv.org
Bagging is an important technique for stabilizing machine learning models. In this paper, we
derive a finite-sample guarantee on the stability of bagging for any model. Our result places …

Distributed Least Squares in Small Space via Sketching and Bias Reduction

S Garg, K Tan, M Dereziński - arXiv preprint arXiv:2405.05343, 2024 - arxiv.org
Matrix sketching is a powerful tool for reducing the size of large data matrices. Yet there are
fundamental limitations to this size reduction when we want to recover an accurate estimator …

[PDF][PDF] Overparametrized Regression: Ridgeless Interpolation

R Tibshirani - stat.berkeley.edu
Overparametrized Regression: Ridgeless Interpolation Page 1 Overparametrized
Regression: Ridgeless Interpolation Advanced Topics in Statistical Learning, Spring 2024 …

[PDF][PDF] A Complete Bibliography of Publications in the SIAM Journal on Mathematics of Data Science

NHF Beebe - 2024 - netlib.org
A Complete Bibliography of Publications in the SIAM Journal on Mathematics of Data Science
Page 1 A Complete Bibliography of Publications in the SIAM Journal on Mathematics of Data …