AI-based fog and edge computing: A systematic review, taxonomy and future directions

S Iftikhar, SS Gill, C Song, M Xu, MS Aslanpour… - Internet of Things, 2023 - Elsevier
Resource management in computing is a very challenging problem that involves making
sequential decisions. Resource limitations, resource heterogeneity, dynamic and diverse …

Communication compression techniques in distributed deep learning: A survey

Z Wang, M Wen, Y Xu, Y Zhou, JH Wang… - Journal of Systems …, 2023 - Elsevier
Nowadays, the training data and neural network models are getting increasingly large. The
training time of deep learning will become unbearably long on a single machine. To reduce …

Efficient sparse collective communication and its application to accelerate distributed deep learning

J Fei, CY Ho, AN Sahu, M Canini, A Sapio - Proceedings of the 2021 …, 2021 - dl.acm.org
Efficient collective communication is crucial to parallel-computing applications such as
distributed training of large-scale recommendation systems and natural language …

Rethinking gradient sparsification as total error minimization

A Sahu, A Dutta, AM Abdelmoniem… - Advances in …, 2021 - proceedings.neurips.cc
Gradient compression is a widely-established remedy to tackle the communication
bottleneck in distributed training of large deep neural networks (DNNs). Under the error …

Empirical analysis of federated learning in heterogeneous environments

AM Abdelmoniem, CY Ho, P Papageorgiou… - Proceedings of the 2nd …, 2022 - dl.acm.org
Federated learning (FL) is becoming a popular paradigm for collaborative learning over
distributed, private datasets owned by non-trusting entities. FL has seen successful …

RandProx: Primal-dual optimization algorithms with randomized proximal updates

L Condat, P Richtárik - arXiv preprint arXiv:2207.12891, 2022 - arxiv.org
Proximal splitting algorithms are well suited to solving large-scale nonsmooth optimization
problems, in particular those arising in machine learning. We propose a new primal-dual …

Gradient compression supercharged high-performance data parallel dnn training

Y Bai, C Li, Q Zhou, J Yi, P Gong, F Yan… - Proceedings of the …, 2021 - dl.acm.org
Gradient compression is a promising approach to alleviating the communication bottleneck
in data parallel deep neural network (DNN) training by significantly reducing the data …

EF-BV: A unified theory of error feedback and variance reduction mechanisms for biased and unbiased compression in distributed optimization

L Condat, K Yi, P Richtárik - Advances in Neural …, 2022 - proceedings.neurips.cc
In distributed or federated optimization and learning, communication between the different
computing units is often the bottleneck and gradient compression is widely used to reduce …

Deepreduce: A sparse-tensor communication framework for federated deep learning

H Xu, K Kostopoulou, A Dutta, X Li… - Advances in …, 2021 - proceedings.neurips.cc
Sparse tensors appear frequently in federated deep learning, either as a direct artifact of the
deep neural network's gradients, or as a result of an explicit sparsification process. Existing …

Federated learning with flexible control

S Wang, J Perazzone, M Ji… - IEEE INFOCOM 2023 …, 2023 - ieeexplore.ieee.org
Federated learning (FL) enables distributed model training from local data collected by
users. In distributed systems with constrained resources and potentially high dynamics, eg …