16 federated knowledge distillation

H Seo, J Park, S Oh, M Bennis, SL Kim - Machine Learning and …, 2022 - cambridge.org
… The aforementioned limitation of FL has motivated to the development of federated distillation
(FD)[… Finally, to transfer the downloaded global knowledge into local models, each worker …

Data-free knowledge distillation for heterogeneous federated learning

Z Zhu, J Hong, J Zhou - International conference on machine …, 2021 - proceedings.mlr.press
… model using aggregated knowledge from heterogeneous users, … Moreover, the ensemble
knowledge is not fully utilized to … In this work, we propose a data-free knowledge distillation

Preservation of the global knowledge by not-true distillation in federated learning

G Lee, M Jeong, Y Shin, S Bae… - Advances in Neural …, 2022 - proceedings.neurips.cc
… the knowledge from previous rounds, and the local training induces forgetting the knowledge
… To this end, we propose a novel and effective algorithm, Federated Not-True Distillation (…

Local-global knowledge distillation in heterogeneous federated learning with non-iid data

D Yao, W Pan, Y Dai, Y Wan, X Ding, H Jin… - arXiv preprint arXiv …, 2021 - arxiv.org
… For all the CV tasks, we set the channels of each group as 16. The optimizer of client training
used is SGD, we tune the learning rate in {0.1, 0.05, 0.01} for FedAvg and set the learning …

Dafkd: Domain-aware federated knowledge distillation

H Wang, Y Li, W Xu, R Li, Y Zhan… - Proceedings of the …, 2023 - openaccess.thecvf.com
… of existing federated distillation methods, we in this paper propose a novel federated distillation …
training with a proximal term in the model objective [16]. FEDDFUSION is a data-based …

Fine-tuning global model via data-free knowledge distillation for non-iid federated learning

L Zhang, L Shen, L Ding, D Tao… - Proceedings of the …, 2022 - openaccess.thecvf.com
… (2) for one class, the importance of knowledge are different among local models of clients.
To facilitate more effective knowledge distillation, we propose customized label sampling and …

FedDKD: Federated learning with decentralized knowledge distillation

X Li, B Chen, W Lu - Applied Intelligence, 2023 - Springer
knowledge distillation process (FedDKD). In FedDKD, we introduce a decentralized knowledge
distillation (DKD) module to distill the knowledge of local … speed of the algorithms [16]. …

Communication-efficient federated learning via knowledge distillation

C Wu, F Wu, L Lyu, Y Huang, X Xie - Nature communications, 2022 - nature.com
… We compare FedPAQ with 16-bit or 8-bit precision levels. … for personalized learning
and knowledge distillation. In addition, … -efficient than other compared federated learning-based …

Parameterized knowledge transfer for personalized federated learning

J Zhang, S Guo, X Ma, H Wang… - Advances in Neural …, 2021 - proceedings.neurips.cc
… Motivated by the paradigm of Knowledge Distillation (KD) [12–16] that knowledge can be
transferred from a neural network to another via exchanging soft predictions instead of using …

Ensemble attention distillation for privacy-preserving federated learning

X Gong, A Sharma, S Karanam, Z Wu… - Proceedings of the …, 2021 - openaccess.thecvf.com
distillation algorithm that aggregates structural knowledge with … inherent heterogeneity of
decentralized federated learning. … address this issue by distilling on public data [16, 22, 4], they …