Optimal stochastic non-smooth non-convex optimization through online-to-non-convex conversion

A Cutkosky, H Mehta… - … Conference on Machine …, 2023 - proceedings.mlr.press
We present new algorithms for optimizing non-smooth, non-convex stochastic objectives
based on a novel analysis technique. This improves the current best-known complexity for …

Faster gradient-free algorithms for nonsmooth nonconvex stochastic optimization

L Chen, J Xu, L Luo - International Conference on Machine …, 2023 - proceedings.mlr.press
We consider the optimization problem of the form $\min_ {x\in\mathbb {R}^ d} f
(x)\triangleq\mathbb {E}[F (x;\xi)] $, where the component $ F (x;\xi) $ is $ L $-mean-squared …

On the complexity of deterministic nonsmooth and nonconvex optimization

MI Jordan, T Lin, M Zampetakis - arXiv preprint arXiv:2209.12463, 2022 - arxiv.org
In this paper, we present several new results on minimizing a nonsmooth and nonconvex
function under a Lipschitz condition. Recent work shows that while the classical notion of …

Random scaling and momentum for non-smooth non-convex optimization

Q Zhang, A Cutkosky - arXiv preprint arXiv:2405.09742, 2024 - arxiv.org
Training neural networks requires optimizing a loss function that may be highly irregular,
and in particular neither convex nor smooth. Popular training algorithms are based on …

The cost of nonconvexity in deterministic nonsmooth optimization

S Kong, AS Lewis - Mathematics of Operations Research, 2023 - pubsonline.informs.org
We study the impact of nonconvexity on the complexity of nonsmooth optimization,
emphasizing objectives such as piecewise linear functions, which may not be weakly …

No dimension-free deterministic algorithm computes approximate stationarities of lipschitzians

L Tian, AMC So - Mathematical Programming, 2024 - Springer
We consider the oracle complexity of computing an approximate stationary point of a
Lipschitz function. When the function is smooth, it is well known that the simple deterministic …

Oracle complexity of single-loop switching subgradient methods for non-smooth weakly convex functional constrained optimization

Y Huang, Q Lin - Advances in Neural Information …, 2023 - proceedings.neurips.cc
We consider a non-convex constrained optimization problem, where the objective function is
weakly convex and the constraint function is either convex or weakly convex. To solve this …

Testing Stationarity Concepts for ReLU Networks: Hardness, Regularity, and Robust Algorithms

L Tian, AMC So - arXiv preprint arXiv:2302.12261, 2023 - arxiv.org
We study the computational problem of the stationarity test for the empirical loss of neural
networks with ReLU activation functions. Our contributions are: Hardness: We show that …

Lipschitz minimization and the Goldstein modulus

S Kong, AS Lewis - arXiv preprint arXiv:2405.12655, 2024 - arxiv.org
Goldstein's 1977 idealized iteration for minimizing a Lipschitz objective fixes a distance-the
step size-and relies on a certain approximate subgradient. That" Goldstein subgradient" is …

Some Problems in Conic, Nonsmooth, and Online Optimization

S Padmanabhan - 2023 - search.proquest.com
Some Problems in Conic, Nonsmooth, and Online Optimization Page 1 Some Problems in
Conic, Nonsmooth, and Online Optimization Swati Padmanabhan A dissertation submitted in …