T Kim, J Oh, NY Kim, S Cho, SY Yun - arXiv e-prints, 2021 - ui.adsabs.harvard.edu
Abstract Knowledge distillation (KD), transferring knowledge from a cumbersome teacher
model to a lightweight student model, has been investigated to design efficient neural …