J Lv, H Yang, P Li - The Thirty-eighth Annual Conference on Neural … - openreview.net
Since pioneering work of Hinton et al., knowledge distillation based on Kullback-Leibler
Divergence (KL-Div) has been predominant, and recently its variants have achieved …