Icebreaker: Warming serverless functions better with heterogeneity

RB Roy, T Patel, D Tiwari - Proceedings of the 27th ACM International …, 2022 - dl.acm.org
Serverless computing, an emerging computing model, relies on" warming up" functions prior
to its anticipated execution for faster and cost-effective service to users. Unfortunately …

Distributed training of large language models

F Zeng, W Gan, Y Wang… - 2023 IEEE 29th …, 2023 - ieeexplore.ieee.org
The advent of large language models (LLMs), like ChatGPT ushers in revolutionary
opportunities that bring a vast variety of applications (such as healthcare, law, and …

Sancus: staleness-aware communication-avoiding full-graph decentralized training in large-scale graph neural networks

J Peng, Z Chen, Y Shao, Y Shen, L Chen… - Proceedings of the VLDB …, 2022 - dl.acm.org
Graph neural networks (GNNs) have emerged due to their success at modeling graph data.
Yet, it is challenging for GNNs to efficiently scale to large graphs. Thus, distributed GNNs …

Defending batch-level label inference and replacement attacks in vertical federated learning

T Zou, Y Liu, Y Kang, W Liu, Y He, Z Yi… - … Transactions on Big …, 2022 - ieeexplore.ieee.org
In a vertical federated learning (VFL) scenario where features and models are split into
different parties, it has been shown that sample-level gradient information can be exploited …

Fastermoe: modeling and optimizing training of large-scale dynamic pre-trained models

J He, J Zhai, T Antunes, H Wang, F Luo, S Shi… - Proceedings of the 27th …, 2022 - dl.acm.org
The current trend in deep learning is to scale models to extremely large sizes with the
objective of increasing their accuracy. Mixture-of-Expert (MoE) is the most popular pre …

Adaptive configuration for heterogeneous participants in decentralized federated learning

Y Liao, Y Xu, H Xu, L Wang… - IEEE INFOCOM 2023-IEEE …, 2023 - ieeexplore.ieee.org
Data generated at the network edge can be processed locally by leveraging the paradigm of
edge computing (EC). Aided by EC, decentralized federated learning (DFL), which …

DecentLaM: Decentralized momentum SGD for large-batch deep training

K Yuan, Y Chen, X Huang, Y Zhang… - Proceedings of the …, 2021 - openaccess.thecvf.com
The scale of deep learning nowadays calls for efficient distributed training algorithms.
Decentralized momentum SGD (DmSGD), in which each node averages only with its …

Stability-based generalization analysis of the asynchronous decentralized SGD

X Deng, T Sun, S Li, D Li - Proceedings of the AAAI Conference on …, 2023 - ojs.aaai.org
The generalization ability often determines the success of machine learning algorithms in
practice. Therefore, it is of great theoretical and practical importance to understand and …

Asynchronous sgd on graphs: a unified framework for asynchronous decentralized and federated optimization

M Even, A Koloskova… - … Conference on Artificial …, 2024 - proceedings.mlr.press
Decentralized and asynchronous communications are two popular techniques to speedup
communication complexity of distributed machine learning, by respectively removing the …

Heterogeneity-aware distributed machine learning training via partial reduce

X Miao, X Nie, Y Shao, Z Yang, J Jiang, L Ma… - Proceedings of the 2021 …, 2021 - dl.acm.org
All-reduce is the key communication primitive used in distributed data-parallel training due
to the high performance in the homogeneous environment. However, All-reduce is sensitive …