Visual tuning

BXB Yu, J Chang, H Wang, L Liu, S Wang… - ACM Computing …, 2024 - dl.acm.org
Fine-tuning visual models has been widely shown promising performance on many
downstream visual tasks. With the surprising development of pre-trained visual foundation …

Edge-cloud polarization and collaboration: A comprehensive survey for ai

J Yao, S Zhang, Y Yao, F Wang, J Ma… - … on Knowledge and …, 2022 - ieeexplore.ieee.org
Influenced by the great success of deep learning via cloud computing and the rapid
development of edge chips, research in artificial intelligence (AI) has shifted to both of the …

Camel: Communicative agents for" mind" exploration of large language model society

G Li, H Hammoud, H Itani… - Advances in Neural …, 2023 - proceedings.neurips.cc
The rapid advancement of chat-based language models has led to remarkable progress in
complex task-solving. However, their success heavily relies on human input to guide the …

Knowledge distillation with the reused teacher classifier

D Chen, JP Mei, H Zhang, C Wang… - Proceedings of the …, 2022 - openaccess.thecvf.com
Abstract Knowledge distillation aims to compress a powerful yet cumbersome teacher model
into a lightweight student model without much sacrifice of performance. For this purpose …

Knowledge distillation: A survey

J Gou, B Yu, SJ Maybank, D Tao - International Journal of Computer Vision, 2021 - Springer
In recent years, deep neural networks have been successful in both industry and academia,
especially for computer vision tasks. The great success of deep learning is mainly due to its …

Knowledge distillation via the target-aware transformer

S Lin, H Xie, B Wang, K Yu, X Chang… - Proceedings of the …, 2022 - openaccess.thecvf.com
Abstract Knowledge distillation becomes a de facto standard to improve the performance of
small neural networks. Most of the previous works propose to regress the representational …

Exploring inter-channel correlation for diversity-preserved knowledge distillation

L Liu, Q Huang, S Lin, H Xie, B Wang… - Proceedings of the …, 2021 - openaccess.thecvf.com
Abstract Knowledge Distillation has shown very promising ability in transferring learned
representation from the larger model (teacher) to the smaller one (student). Despite many …

Generalizable heterogeneous federated cross-correlation and instance similarity learning

W Huang, M Ye, Z Shi, B Du - IEEE Transactions on Pattern …, 2023 - ieeexplore.ieee.org
Federated learning is an important privacy-preserving multi-party learning paradigm,
involving collaborative learning with others and local updating on private data. Model …

Clipping: Distilling clip-based models with a student base for video-language retrieval

R Pei, J Liu, W Li, B Shao, S Xu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Pre-training a vison-language model and then fine-tuning it on downstream tasks have
become a popular paradigm. However, pre-trained vison-language models with the …

Self-regulated feature learning via teacher-free feature distillation

L Li - European Conference on Computer Vision, 2022 - Springer
Abstract Knowledge distillation conditioned on intermediate feature representations always
leads to significant performance improvements. Conventional feature distillation framework …