Communication-efficient distributed learning: An overview

X Cao, T Başar, S Diggavi, YC Eldar… - IEEE journal on …, 2023 - ieeexplore.ieee.org
Distributed learning is envisioned as the bedrock of next-generation intelligent networks,
where intelligent agents, such as mobile devices, robots, and sensors, exchange information …

Database meets deep learning: Challenges and opportunities

W Wang, M Zhang, G Chen, HV Jagadish, BC Ooi… - ACM Sigmod …, 2016 - dl.acm.org
Deep learning has recently become very popular on account of its incredible success in
many complex datadriven applications, including image classification and speech …

Sparsity in deep learning: Pruning and growth for efficient inference and training in neural networks

T Hoefler, D Alistarh, T Ben-Nun, N Dryden… - Journal of Machine …, 2021 - jmlr.org
The growing energy and performance costs of deep learning have driven the community to
reduce the size of neural networks by selectively pruning components. Similarly to their …

Cocktailsgd: Fine-tuning foundation models over 500mbps networks

J Wang, Y Lu, B Yuan, B Chen… - International …, 2023 - proceedings.mlr.press
Distributed training of foundation models, especially large language models (LLMs), is
communication-intensive and so has heavily relied on centralized data centers with fast …

Advances and open problems in federated learning

P Kairouz, HB McMahan, B Avent… - … and trends® in …, 2021 - nowpublishers.com
Federated learning (FL) is a machine learning setting where many clients (eg, mobile
devices or whole organizations) collaboratively train a model under the orchestration of a …

A survey of federated learning for edge computing: Research problems and solutions

Q Xia, W Ye, Z Tao, J Wu, Q Li - High-Confidence Computing, 2021 - Elsevier
Federated Learning is a machine learning scheme in which a shared prediction model can
be collaboratively learned by a number of distributed nodes using their locally stored data. It …

EF21: A new, simpler, theoretically better, and practically faster error feedback

P Richtárik, I Sokolov… - Advances in Neural …, 2021 - proceedings.neurips.cc
Error feedback (EF), also known as error compensation, is an immensely popular
convergence stabilization mechanism in the context of distributed training of supervised …

Decentralized federated learning with unreliable communications

H Ye, L Liang, GY Li - IEEE journal of selected topics in signal …, 2022 - ieeexplore.ieee.org
Decentralized federated learning, inherited from decentralized learning, enables the edge
devices to collaborate on model training in a peer-to-peer manner without the assistance of …

Exponential graph is provably efficient for decentralized deep training

B Ying, K Yuan, Y Chen, H Hu… - Advances in Neural …, 2021 - proceedings.neurips.cc
Decentralized SGD is an emerging training method for deep learning known for its much
less (thus faster) communication per iteration, which relaxes the averaging step in parallel …

[PDF][PDF] Communication-Efficient Stochastic Gradient Descent Ascent with Momentum Algorithms.

Y Zhang, M Qiu, H Gao - IJCAI, 2023 - ijcai.org
Numerous machine learning models can be formulated as a stochastic minimax optimization
problem, such as imbalanced data classification with AUC maximization. Developing …