Proxskip: Yes! local gradient steps provably lead to communication acceleration! finally!

K Mishchenko, G Malinovsky, S Stich… - International …, 2022 - proceedings.mlr.press
We introduce ProxSkip—a surprisingly simple and provably efficient method for minimizing
the sum of a smooth ($ f $) and an expensive nonsmooth proximable ($\psi $) function. The …

[图书][B] Large-scale convex optimization: algorithms & analyses via monotone operators

EK Ryu, W Yin - 2022 - books.google.com
Starting from where a first course in convex optimization leaves off, this text presents a
unified analysis of first-order optimization methods–including parallel-distributed algorithms …

Proximal splitting algorithms for convex optimization: A tour of recent advances, with new twists

L Condat, D Kitahara, A Contreras, A Hirabayashi - SIAM Review, 2023 - SIAM
Convex nonsmooth optimization problems, whose solutions live in very high dimensional
spaces, have become ubiquitous. To solve them, the class of first-order algorithms known as …

Optimal and practical algorithms for smooth and strongly convex decentralized optimization

D Kovalev, A Salim, P Richtárik - Advances in Neural …, 2020 - proceedings.neurips.cc
We consider the task of decentralized minimization of the sum of smooth strongly convex
functions stored across the nodes of a network. For this problem, lower bounds on the …

Can 5th generation local training methods support client sampling? yes!

M Grudzień, G Malinovsky… - … Conference on Artificial …, 2023 - proceedings.mlr.press
The celebrated FedAvg algorithm of McMahan et al.(2017) is based on three components:
client sampling (CS), data sampling (DS) and local training (LT). While the first two are …

Lower bounds and optimal algorithms for smooth and strongly convex decentralized optimization over time-varying networks

D Kovalev, E Gasanov, A Gasnikov… - Advances in Neural …, 2021 - proceedings.neurips.cc
We consider the task of minimizing the sum of smooth and strongly convex functions stored
in a decentralized manner across the nodes of a communication network whose links are …

RandProx: Primal-dual optimization algorithms with randomized proximal updates

L Condat, P Richtárik - arXiv preprint arXiv:2207.12891, 2022 - arxiv.org
Proximal splitting algorithms are well suited to solving large-scale nonsmooth optimization
problems, in particular those arising in machine learning. We propose a new primal-dual …

DADAO: Decoupled accelerated decentralized asynchronous optimization

A Nabli, E Oyallon - International Conference on Machine …, 2023 - proceedings.mlr.press
This work introduces DADAO: the first decentralized, accelerated, asynchronous, primal, first-
order algorithm to minimize a sum of $ L $-smooth and $\mu $-strongly convex functions …

An optimal algorithm for strongly convex minimization under affine constraints

A Salim, L Condat, D Kovalev… - … conference on artificial …, 2022 - proceedings.mlr.press
Optimization problems under affine constraints appear in various areas of machine learning.
We consider the task of minimizing a smooth strongly convex function F (x) under the affine …

DISA: A dual inexact splitting algorithm for distributed convex composite optimization

L Guo, X Shi, S Yang, J Cao - IEEE Transactions on Automatic …, 2023 - ieeexplore.ieee.org
In this article, we propose a novel dual inexact splitting algorithm (DISA) for distributed
convex composite optimization problems, where the local loss function consists of a smooth …