Awt: Transferring vision-language models via augmentation, weighting, and transportation

Y Zhu, Y Ji, Z Zhao, G Wu, L Wang - arXiv preprint arXiv:2407.04603, 2024 - arxiv.org
Pre-trained vision-language models (VLMs) have shown impressive results in various visual
classification tasks. However, we often fail to fully unleash their potential when adapting …

Caps-adapter: Caption-based multimodal adapter in zero-shot classification

Q Wang, G Liu, B Wang - Proceedings of the 32nd ACM International …, 2024 - dl.acm.org
Recent advances in vision-language foundational models, such as CLIP, have
demonstrated significant strides in zero-shot classification. However, the extensive …

Prompt-induced prototype alignment for few-shot unsupervised domain adaptation

Y Li, S Long, S Wang, X Zhao, Y Li - Expert Systems with Applications, 2025 - Elsevier
Abstract Unsupervised Domain Adaptation excels in transferring predictive models from a
labeled source domain to an unlabeled target domain. However, acquiring sufficient source …

Mutual Prompt Leaning for Vision Language Models

S Long, Z Zhao, J Yuan, Z Tan, J Liu, J Feng… - International Journal of …, 2024 - Springer
Large pre-trained vision language models (VLMs) have demonstrated impressive
representation learning capabilities, but their transferability across various downstream …

A Slim Prompt-Averaged Consistency prompt learning for vision-language model

S He, S Wang, S Long - Knowledge-Based Systems, 2025 - Elsevier
Recent advancements in prompt tuning have enhanced the adaptation of large pre-trained
models to target tasks. However, existing methods struggle to establish an effective balance …

M-Tuning: Prompt Tuning with Mitigated Label Bias in Open-Set Scenarios

N Liao, X Zhang, M Cao, J Yan - IEEE Transactions on Circuits …, 2024 - ieeexplore.ieee.org
In realistic open-set scenarios where labels of a part of testing data are totally unknown,
when vision-language (VL) prompt learning methods encounter inputs related to unknown …