LENA: Communication-efficient distributed learning with self-triggered gradient uploads

HS Ghadikolaei, S Stich… - … Conference on Artificial …, 2021 - proceedings.mlr.press
In distributed optimization, parameter updates from the gradient computing node devices
have to be aggregated in every iteration on the orchestrating server. When these updates …

LAG: Lazily aggregated gradient for communication-efficient distributed learning

T Chen, G Giannakis, T Sun… - Advances in neural …, 2018 - proceedings.neurips.cc
This paper presents a new class of gradient methods for distributed machine learning that
adaptively skip the gradient calculations to learn with reduced communication and …

Communication-efficient distributed learning via lazily aggregated quantized gradients

J Sun, T Chen, G Giannakis… - Advances in Neural …, 2019 - proceedings.neurips.cc
The present paper develops a novel aggregated gradient approach for distributed machine
learning that adaptively compresses the gradient communication. The key idea is to first …

Local exact-diffusion for decentralized optimization and learning

SA Alghunaim - IEEE Transactions on Automatic Control, 2024 - ieeexplore.ieee.org
Distributed optimization methods with local updates have recently attracted a lot of attention
due to their potential to reduce the communication cost of distributed methods. In these …

Distributed learning with compressed gradients

S Khirirat, HR Feyzmahdavian… - arXiv preprint arXiv …, 2018 - arxiv.org
Asynchronous computation and gradient compression have emerged as two key techniques
for achieving scalability in distributed optimization for large-scale machine learning. This …

Distributed training with heterogeneous data: Bridging median-and mean-based algorithms

X Chen, T Chen, H Sun, SZ Wu… - Advances in Neural …, 2020 - proceedings.neurips.cc
Recently, there is a growing interest in the study of median-based algorithms for distributed
non-convex optimization. Two prominent examples include signSGD with majority vote, an …

Anytime minibatch: Exploiting stragglers in online distributed optimization

N Ferdinand, H Al-Lawati, SC Draper… - arXiv preprint arXiv …, 2020 - arxiv.org
Distributed optimization is vital in solving large-scale machine learning problems. A widely-
shared feature of distributed optimization techniques is the requirement that all nodes …

Error compensated quantized SGD and its applications to large-scale distributed optimization

J Wu, W Huang, J Huang… - … Conference on Machine …, 2018 - proceedings.mlr.press
Large-scale distributed optimization is of great importance in various applications. For data-
parallel based distributed learning, the inter-node gradient communication often becomes …

Stochastic distributed learning with gradient quantization and double-variance reduction

S Horváth, D Kovalev, K Mishchenko… - Optimization Methods …, 2023 - Taylor & Francis
We consider distributed optimization over several devices, each sending incremental model
updates to a central server. This setting is considered, for instance, in federated learning …

Communication-efficient distributed optimization in networks with gradient tracking and variance reduction

B Li, S Cen, Y Chen, Y Chi - Journal of Machine Learning Research, 2020 - jmlr.org
There is growing interest in large-scale machine learning and optimization over
decentralized networks, eg in the context of multi-agent learning and federated learning …