Optimal client sampling for federated learning

W Chen, S Horvath, P Richtarik - arXiv preprint arXiv:2010.13723, 2020 - arxiv.org
It is well understood that client-master communication can be a primary bottleneck in
Federated Learning. In this work, we address this issue with a novel client subsampling …

Communication compression techniques in distributed deep learning: A survey

Z Wang, M Wen, Y Xu, Y Zhou, JH Wang… - Journal of Systems …, 2023 - Elsevier
Nowadays, the training data and neural network models are getting increasingly large. The
training time of deep learning will become unbearably long on a single machine. To reduce …

Local sgd: Unified theory and new efficient methods

E Gorbunov, F Hanzely… - … Conference on Artificial …, 2021 - proceedings.mlr.press
We present a unified framework for analyzing local SGD methods in the convex and strongly
convex regimes for distributed/federated training of supervised machine learning models …

Large-scale deep learning optimizations: A comprehensive survey

X He, F Xue, X Ren, Y You - arXiv preprint arXiv:2111.00856, 2021 - arxiv.org
Deep learning have achieved promising results on a wide spectrum of AI applications.
Larger datasets and models consistently yield better performance. However, we generally …

A better alternative to error feedback for communication-efficient distributed learning

S Horváth, P Richtárik - arXiv preprint arXiv:2006.11077, 2020 - arxiv.org
Modern large-scale machine learning applications require stochastic optimization
algorithms to be implemented on distributed compute systems. A key bottleneck of such …

Adaptive gradient quantization for data-parallel sgd

F Faghri, I Tabrizian, I Markov… - Advances in neural …, 2020 - proceedings.neurips.cc
Many communication-efficient variants of SGD use gradient quantization schemes. These
schemes are often heuristic and fixed over the course of training. We empirically observe …

DoCoFL: Downlink compression for cross-device federated learning

R Dorfman, S Vargaftik… - … on Machine Learning, 2023 - proceedings.mlr.press
Many compression techniques have been proposed to reduce the communication overhead
of Federated Learning training procedures. However, these are typically designed for …

Federated learning with regularized client participation

G Malinovsky, S Horváth, K Burlachenko… - arXiv preprint arXiv …, 2023 - arxiv.org
Federated Learning (FL) is a distributed machine learning approach where multiple clients
work together to solve a machine learning task. One of the key challenges in FL is the issue …

RATQ: A universal fixed-length quantizer for stochastic optimization

P Mayekar, H Tyagi - International Conference on Artificial …, 2020 - proceedings.mlr.press
Abstract We present Rotated Adaptive Tetra-iterated Quantizer (RATQ), afixed-length
quantizer for gradients in first order stochasticoptimization. RATQ is easy to implement and …

Moshpit sgd: Communication-efficient decentralized training on heterogeneous unreliable devices

M Ryabinin, E Gorbunov… - Advances in …, 2021 - proceedings.neurips.cc
Training deep neural networks on large datasets can often be accelerated by using multiple
compute nodes. This approach, known as distributed training, can utilize hundreds of …