A Review of multilayer extreme learning machine neural networks

JA Vásquez-Coronel, M Mora, K Vilches - Artificial Intelligence Review, 2023 - Springer
Abstract The Extreme Learning Machine is a single-hidden-layer feedforward learning
algorithm, which has been successfully applied in regression and classification problems in …

Acceleration methods

A d'Aspremont, D Scieur, A Taylor - Foundations and Trends® …, 2021 - nowpublishers.com
This monograph covers some recent advances in a range of acceleration techniques
frequently used in convex optimization. We first use quadratic optimization problems to …

From Nesterov's estimate sequence to Riemannian acceleration

K Ahn, S Sra - Conference on Learning Theory, 2020 - proceedings.mlr.press
We propose the first global accelerated gradient method for Riemannian manifolds. Toward
establishing our results, we revisit Nesterov's estimate sequence technique and develop a …

Characterizing the exact behaviors of temporal difference learning algorithms using Markov jump linear system theory

B Hu, U Syed - Advances in neural information processing …, 2019 - proceedings.neurips.cc
In this paper, we provide a unified analysis of temporal difference learning algorithms with
linear function approximators by exploiting their connections to Markov jump linear systems …

Robust and structure exploiting optimisation algorithms: an integral quadratic constraint approach

S Michalowsky, C Scherer… - International Journal of …, 2021 - Taylor & Francis
We consider the problem of analysing and designing gradient-based discrete-time
optimisation algorithms for a class of unconstrained optimisation problems having strongly …

Stochastic approximation beyond gradient for signal processing and machine learning

A Dieuleveut, G Fort, E Moulines… - IEEE Transactions on …, 2023 - ieeexplore.ieee.org
Stochastic Approximation (SA) is a classical algorithm that has had since the early days a
huge impact on signal processing, and nowadays on machine learning, due to the necessity …

Scheduled restart momentum for accelerated stochastic gradient descent

B Wang, T Nguyen, T Sun, AL Bertozzi… - SIAM Journal on Imaging …, 2022 - SIAM
Stochastic gradient descent (SGD) algorithms, with constant momentum and its variants
such as Adam, are the optimization methods of choice for training deep neural networks …

A universally optimal multistage accelerated stochastic gradient method

NS Aybat, A Fallah… - Advances in neural …, 2019 - proceedings.neurips.cc
We study the problem of minimizing a strongly convex, smooth function when we have noisy
estimates of its gradient. We propose a novel multistage accelerated algorithm that is …

Reproducibility in optimization: Theoretical framework and limits

K Ahn, P Jain, Z Ji, S Kale… - Advances in Neural …, 2022 - proceedings.neurips.cc
We initiate a formal study of reproducibility in optimization. We define a quantitative measure
of reproducibility of optimization procedures in the face of noisy or error-prone operations …

Robust accelerated primal-dual methods for computing saddle points

X Zhang, NS Aybat, M Gürbüzbalaban - SIAM Journal on Optimization, 2024 - SIAM
We consider strongly-convex-strongly-concave saddle point problems assuming we have
access to unbiased stochastic estimates of the gradients. We propose a stochastic …