Stability and convergence of stochastic gradient clipping: Beyond lipschitz continuity and smoothness

VV Mai, M Johansson - International Conference on …, 2021 - proceedings.mlr.press
Stochastic gradient algorithms are often unstable when applied to functions that do not have
Lipschitz-continuous and/or bounded gradients. Gradient clipping is a simple and effective …

An adaptive stochastic sequential quadratic programming with differentiable exact augmented lagrangians

S Na, M Anitescu, M Kolar - Mathematical Programming, 2023 - Springer
We consider solving nonlinear optimization problems with a stochastic objective and
deterministic equality constraints. We assume for the objective that its evaluation, gradient …

A trust region method for noisy unconstrained optimization

S Sun, J Nocedal - Mathematical Programming, 2023 - Springer
Classical trust region methods were designed to solve problems in which function and
gradient information are exact. This paper considers the case when there are errors (or …

Gradient descent in the absence of global lipschitz continuity of the gradients

V Patel, AS Berahas - SIAM Journal on Mathematics of Data Science, 2024 - SIAM
Gradient descent (GD) is a collection of continuous optimization methods that have achieved
immeasurable success in practice. Owing to data science applications, GD with diminishing …

Adaptive stochastic optimization: A framework for analyzing stochastic optimization algorithms

FE Curtis, K Scheinberg - IEEE Signal Processing Magazine, 2020 - ieeexplore.ieee.org
Optimization lies at the heart of machine learning (ML) and signal processing (SP).
Contemporary approaches based on the stochastic gradient (SG) method are nonadaptive …

Trust Region Methods for Nonconvex Stochastic Optimization beyond Lipschitz Smoothness

C Xie, C Li, C Zhang, Q Deng, D Ge, Y Ye - Proceedings of the AAAI …, 2024 - ojs.aaai.org
In many important machine learning applications, the standard assumption of having a
globally Lipschitz continuous gradient may fail to hold. This paper delves into a more …

Stochastic mesh adaptive direct search for blackbox optimization using probabilistic estimates

C Audet, KJ Dzahini, M Kokkolaras… - Computational …, 2021 - Springer
We present a stochastic extension of the mesh adaptive direct search (MADS) algorithm
originally developed for deterministic blackbox optimization. The algorithm, called StoMADS …

Fully stochastic trust-region sequential quadratic programming for equality-constrained optimization problems

Y Fang, S Na, MW Mahoney, M Kolar - SIAM Journal on Optimization, 2024 - SIAM
We propose a trust-region stochastic sequential quadratic programming algorithm (TR-
StoSQP) to solve nonlinear optimization problems with stochastic objectives and …

A trust region method for the optimization of noisy functions

S Sun, J Nocedal - arXiv preprint arXiv:2201.00973, 2022 - arxiv.org
Classical trust region methods were designed to solve problems in which function and
gradient information are exact. This paper considers the case when there are bounded …

Stochastic Optimization for Nonconvex Problem With Inexact Hessian Matrix, Gradient, and Function

L Liu, X Liu, CJ Hsieh, D Tao - IEEE Transactions on Neural …, 2023 - ieeexplore.ieee.org
Trust region (TR) and adaptive regularization using cubics (ARC) have proven to have some
very appealing theoretical properties for nonconvex optimization by concurrently computing …