Knowledge distillation from a stronger teacher

T Huang, S You, F Wang, C Qian… - Advances in Neural …, 2022 - proceedings.neurips.cc
Unlike existing knowledge distillation methods focus on the baseline settings, where the
teacher models and training strategies are not that strong and competing as state-of-the-art …

Knowledge diffusion for distillation

T Huang, Y Zhang, M Zheng, S You… - Advances in …, 2023 - proceedings.neurips.cc
The representation gap between teacher and student is an emerging topic in knowledge
distillation (KD). To reduce the gap and improve the performance, current methods often …

Logit standardization in knowledge distillation

S Sun, W Ren, J Li, R Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Knowledge distillation involves transferring soft labels from a teacher to a student
using a shared temperature-based softmax function. However the assumption of a shared …

Localization distillation for dense object detection

Z Zheng, R Ye, P Wang, D Ren… - Proceedings of the …, 2022 - openaccess.thecvf.com
Abstract Knowledge distillation (KD) has witnessed its powerful capability in learning
compact models in object detection. Previous KD methods for object detection mostly focus …

A good student is cooperative and reliable: CNN-transformer collaborative learning for semantic segmentation

J Zhu, Y Luo, X Zheng, H Wang… - Proceedings of the …, 2023 - openaccess.thecvf.com
In this paper, we strive to answer the question'how to collaboratively learn convolutional
neural network (CNN)-based and vision transformer (ViT)-based models by selecting and …

Improved feature distillation via projector ensemble

Y Chen, S Wang, J Liu, X Xu… - Advances in Neural …, 2022 - proceedings.neurips.cc
In knowledge distillation, previous feature distillation methods mainly focus on the design of
loss functions and the selection of the distilled layers, while the effect of the feature projector …

A review of AI edge devices and lightweight CNN deployment

K Sun, X Wang, X Miao, Q Zhao - Neurocomputing, 2024 - Elsevier
Abstract Artificial Intelligence of Things (AIoT) which integrates artificial intelligence (AI) and
the Internet of Things (IoT), has attracted increasing attention recently. With the remarkable …

Feature map distillation of thin nets for low-resolution object recognition

Z Huang, S Yang, MC Zhou, Z Li… - IEEE Transactions on …, 2022 - ieeexplore.ieee.org
Intelligent video surveillance is an important computer vision application in natural
environments. Since detected objects under surveillance are usually low-resolution and …

Rethinking feature-based knowledge distillation for face recognition

J Li, Z Guo, H Li, S Han, JW Baek… - Proceedings of the …, 2023 - openaccess.thecvf.com
With the continual expansion of face datasets, feature-based distillation prevails for large-
scale face recognition. In this work, we attempt to remove identity supervision in student …

Amd: Automatic multi-step distillation of large-scale vision models

C Han, Q Wang, SA Dianat, M Rabbani… - … on Computer Vision, 2025 - Springer
Transformer-based architectures have become the de-facto standard models for diverse
vision tasks owing to their superior performance. As the size of these transformer-based …