Communication-efficient distributed learning: An overview

X Cao, T Başar, S Diggavi, YC Eldar… - IEEE journal on …, 2023 - ieeexplore.ieee.org
Distributed learning is envisioned as the bedrock of next-generation intelligent networks,
where intelligent agents, such as mobile devices, robots, and sensors, exchange information …

Activated gradients for deep neural networks

M Liu, L Chen, X Du, L Jin… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Deep neural networks often suffer from poor performance or even training failure due to the
ill-conditioned problem, the vanishing/exploding gradient problem, and the saddle point …

Network topology and communication-computation tradeoffs in decentralized optimization

A Nedić, A Olshevsky, MG Rabbat - Proceedings of the IEEE, 2018 - ieeexplore.ieee.org
In decentralized optimization, nodes cooperate to minimize an overall objective function that
is the sum (or average) of per-node private objective functions. Algorithms interleave local …

Tutorial on dynamic average consensus: The problem, its applications, and the algorithms

SS Kia, B Van Scoy, J Cortes… - IEEE Control …, 2019 - ieeexplore.ieee.org
Technological advances in ad hoc networking and the availability of low-cost reliable
computing, data storage, and sensing devices have made scenarios possible where the …

A linear algorithm for optimization over directed graphs with geometric convergence

R Xin, UA Khan - IEEE Control Systems Letters, 2018 - ieeexplore.ieee.org
In this letter, we study distributed optimization, where a network of agents, abstracted as a
directed graph, collaborates to minimize the average of locally known convex functions …

A decentralized proximal-gradient method with network independent step-sizes and separated convergence rates

Z Li, W Shi, M Yan - IEEE Transactions on Signal Processing, 2019 - ieeexplore.ieee.org
This paper proposes a novel proximal-gradient algorithm for a decentralized optimization
problem with a composite objective containing smooth and nonsmooth terms. Specifically …

A general framework for decentralized optimization with first-order methods

R Xin, S Pu, A Nedić, UA Khan - Proceedings of the IEEE, 2020 - ieeexplore.ieee.org
Decentralized optimization to minimize a finite sum of functions, distributed over a network of
nodes, has been a significant area within control and signal-processing research due to its …

Conditional gradient methods

G Braun, A Carderera, CW Combettes… - arXiv preprint arXiv …, 2022 - arxiv.org
The purpose of this survey is to serve both as a gentle introduction and a coherent overview
of state-of-the-art Frank--Wolfe algorithms, also called conditional gradient algorithms, for …

A dual approach for optimal algorithms in distributed optimization over networks

CA Uribe, S Lee, A Gasnikov… - 2020 Information theory …, 2020 - ieeexplore.ieee.org
We study dual-based algorithms for distributed convex optimization problems over networks,
where the objective is to minimize a sum Σ i= 1 mfi (z) of functions over in a network. We …

Distributed heavy-ball: A generalization and acceleration of first-order methods with gradient tracking

R Xin, UA Khan - IEEE Transactions on Automatic Control, 2019 - ieeexplore.ieee.org
We study distributed optimization to minimize a sum of smooth and strongly-convex
functions. Recent work on this problem uses gradient tracking to achieve linear convergence …