A comprehensive survey of forgetting in deep learning beyond continual learning

Z Wang, E Yang, L Shen… - IEEE Transactions on …, 2024 - ieeexplore.ieee.org
Forgetting refers to the loss or deterioration of previously acquired knowledge. While
existing surveys on forgetting have primarily focused on continual learning, forgetting is a …

Learning to retain while acquiring: Combating distribution-shift in adversarial data-free knowledge distillation

G Patel, KR Mopuri, Q Qiu - … of the IEEE/CVF Conference on …, 2023 - openaccess.thecvf.com
Abstract Data-free Knowledge Distillation (DFKD) has gained popularity recently, with the
fundamental idea of carrying out knowledge transfer from a Teacher neural network to a …

Distribution shift matters for knowledge distillation with webly collected images

J Tang, S Chen, G Niu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Knowledge distillation aims to learn a lightweight student network from a pre-
trained teacher network. In practice, existing knowledge distillation methods are usually …

DFRD: Data-Free Robustness Distillation for Heterogeneous Federated Learning

S Wang, Y Fu, X Li, Y Lan… - Advances in Neural …, 2024 - proceedings.neurips.cc
Federated Learning (FL) is a privacy-constrained decentralized machine learning paradigm
in which clients enable collaborative training without compromising private data. However …

Data-free hard-label robustness stealing attack

X Yuan, K Chen, W Huang, J Zhang… - Proceedings of the AAAI …, 2024 - ojs.aaai.org
The popularity of Machine Learning as a Service (MLaaS) has led to increased concerns
about Model Stealing Attacks (MSA), which aim to craft a clone model by querying MLaaS …

Sampling to distill: Knowledge transfer from open-world data

Y Wang, Z Chen, J Zhang, D Yang, Z Ge, Y Liu… - Proceedings of the …, 2024 - dl.acm.org
Data-Free Knowledge Distillation (DFKD) is a novel task that aims to train high-performance
student models using only the pre-trained teacher network without original training data …

Data-free Knowledge Distillation for Fine-grained Visual Categorization

R Shao, W Zhang, J Yin… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Data-free knowledge distillation (DFKD) is a promising approach for addressing issues
related to model compression, security privacy, and transmission restrictions. Although the …

DFRD: data-free robustness distillation for heterogeneous federated learning

K Luo, S Wang, Y Fu, X Li, Y Lan, M Gao - arXiv preprint arXiv:2309.13546, 2023 - arxiv.org
Federated Learning (FL) is a privacy-constrained decentralized machine learning paradigm
in which clients enable collaborative training without compromising private data. However …

Direct distillation between different domains

J Tang, S Chen, G Niu, H Zhu, JT Zhou, C Gong… - … on Computer Vision, 2025 - Springer
Abstract Knowledge Distillation (KD) aims to learn a compact student network using
knowledge from a large pre-trained teacher network, where both networks are trained on …

Mind the Gap Between Synthetic and Real: Utilizing Transfer Learning to Probe the Boundaries of Stable Diffusion Generated Data

L Hennicke, CM Adriano, H Giese, JM Koehler… - arXiv preprint arXiv …, 2024 - arxiv.org
Generative foundation models like Stable Diffusion comprise a diverse spectrum of
knowledge in computer vision with the potential for transfer learning, eg, via generating data …