Curriculum temperature for knowledge distillation

Z Li, X Li, L Yang, B Zhao, R Song, L Luo, J Li… - Proceedings of the …, 2023 - ojs.aaai.org
Most existing distillation methods ignore the flexible role of the temperature in the loss
function and fix it as a hyper-parameter that can be decided by an inefficient grid search. In …

Disentangling voice and content with self-supervision for speaker recognition

T Liu, KA Lee, Q Wang, H Li - Advances in Neural …, 2023 - proceedings.neurips.cc
For speaker recognition, it is difficult to extract an accurate speaker representation from
speech because of its mixture of speaker traits and content. This paper proposes a …

Shadow knowledge distillation: Bridging offline and online knowledge transfer

L Li, Z Jin - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
Abstract Knowledge distillation can be generally divided into offline and online categories
according to whether teacher model is pre-trained and persistent during the distillation …

Logit standardization in knowledge distillation

S Sun, W Ren, J Li, R Wang… - Proceedings of the IEEE …, 2024 - openaccess.thecvf.com
Abstract Knowledge distillation involves transferring soft labels from a teacher to a student
using a shared temperature-based softmax function. However the assumption of a shared …

Few-shot image generation via adaptation-aware kernel modulation

Y Zhao, K Chandrasegaran… - Advances in …, 2022 - proceedings.neurips.cc
Few-shot image generation (FSIG) aims to learn to generate new and diverse samples given
an extremely limited number of samples from a domain, eg, 10 training samples. Recent …

Re-thinking model inversion attacks against deep neural networks

NB Nguyen, K Chandrasegaran… - Proceedings of the …, 2023 - openaccess.thecvf.com
Abstract Model inversion (MI) attacks aim to infer and reconstruct private training data by
abusing access to a model. MI attacks have raised concerns about the leaking of sensitive …

Discovering transferable forensic features for cnn-generated images detection

K Chandrasegaran, NT Tran, A Binder… - European Conference on …, 2022 - Springer
Visual counterfeits (We refer to CNN-generated images as counterfeits throughout this
paper.) are increasingly causing an existential conundrum in mainstream media with rapid …

Exploring incompatible knowledge transfer in few-shot image generation

Y Zhao, C Du, M Abdollahzadeh… - Proceedings of the …, 2023 - openaccess.thecvf.com
Few-shot image generation (FSIG) learns to generate diverse and high-fidelity images from
a target domain using a few (eg, 10) reference samples. Existing FSIG methods select …

Boosting knowledge distillation via intra-class logit distribution smoothing

C Li, G Cheng, J Han - … on Circuits and Systems for Video …, 2023 - ieeexplore.ieee.org
Previous arts built an intimate link between knowledge distillation (KD) and label smoothing
(LS) that they both impose regularization on the model training. In this paper, we delve …

CReg-KD: Model refinement via confidence regularized knowledge distillation for brain imaging

Y Yang, X Guo, C Ye, Y Xiang, T Ma - Medical Image Analysis, 2023 - Elsevier
One of the core challenges of deep learning in medical image analysis is data insufficiency,
especially for 3D brain imaging, which may lead to model over-fitting and poor …