C Wang, Y Lu, Y Mu, Y Hu, T Xiao, J Zhu - arXiv e-prints, 2023 - ui.adsabs.harvard.edu
Abstract Knowledge distillation addresses the problem of transferring knowledge from a
teacher model to a student model. In this process, we typically have multiple types of …