The difficulty of computing stable and accurate neural networks: On the barriers of deep learning and Smale's 18th problem

MJ Colbrook, V Antun… - Proceedings of the …, 2022 - National Acad Sciences
Deep learning (DL) has had unprecedented success and is now entering scientific
computing with full force. However, current DL methods typically suffer from instability, even …

Matrix factorization techniques in machine learning, signal processing, and statistics

KL Du, MNS Swamy, ZQ Wang, WH Mow - Mathematics, 2023 - mdpi.com
Compressed sensing is an alternative to Shannon/Nyquist sampling for acquiring sparse or
compressible signals. Sparse coding represents a signal as a sparse linear combination of …

Regularized M-estimators with nonconvexity: Statistical and algorithmic theory for local optima

PL Loh, MJ Wainwright - Advances in Neural Information …, 2013 - proceedings.neurips.cc
We establish theoretical results concerning all local optima of various regularized M-
estimators, where both loss and penalty functions are allowed to be nonconvex. Our results …

[图书][B] Introduction to high-dimensional statistics

C Giraud - 2021 - taylorfrancis.com
Praise for the first edition:"[This book] succeeds singularly at providing a structured
introduction to this active field of research.… it is arguably the most accessible overview yet …

[PDF][PDF] partykit: A modular toolkit for recursive partytioning in R

T Hothorn, A Zeileis - The Journal of Machine Learning Research, 2015 - jmlr.org
The R package partykit provides a flexible toolkit for learning, representing, summarizing,
and visualizing a wide range of tree-structured regression and classification models. The …

A survey on sparse learning models for feature selection

X Li, Y Wang, R Ruiz - IEEE transactions on cybernetics, 2020 - ieeexplore.ieee.org
Feature selection is important in both machine learning and pattern recognition.
Successfully selecting informative features can significantly increase learning accuracy and …

Distributionally robust learning

R Chen, IC Paschalidis - Foundations and Trends® in …, 2020 - nowpublishers.com
This monograph develops a comprehensive statistical learning framework that is robust to
(distributional) perturbations in the data using Distributionally Robust Optimization (DRO) …

Sparse CCA: Adaptive estimation and computational barriers

C Gao, Z Ma, HH Zhou - The Annals of Statistics, 2017 - JSTOR
Canonical correlation analysis is a classical technique for exploring the relationship
between two sets of variables. It has important applications in analyzing high dimensional …

Simultaneous feature selection and clustering based on square root optimization

H Jiang, S Luo, Y Dong - European Journal of Operational Research, 2021 - Elsevier
The fused least absolute shrinkage and selection operator (LASSO) simultaneously
pursuing the joint sparsity of coefficients and their successive differences has attracted …

Sign-constrained least squares estimation for high-dimensional regression

N Meinshausen - 2013 - projecteuclid.org
Many regularization schemes for high-dimensional regression have been put forward. Most
require the choice of a tuning parameter, using model selection criteria or cross-validation …