Open-vocabulary multi-label classification via multi-modal knowledge transfer

S He, T Guo, T Dai, R Qiao, X Shu, B Ren… - Proceedings of the AAAI …, 2023 - ojs.aaai.org
Real-world recognition system often encounters the challenge of unseen labels. To identify
such unseen labels, multi-label zero-shot learning (ML-ZSL) focuses on transferring …

Multi-Label Knowledge Distillation

P Yang, MK Xie, CC Zong, L Feng… - Proceedings of the …, 2023 - openaccess.thecvf.com
Existing knowledge distillation methods typically work by imparting the knowledge of output
logits or intermediate feature maps from the teacher network to the student network, which is …

Diverse and tailored image generation for zero-shot multi-label classification

K Zhang, Z Yuan, T Huang - Knowledge-Based Systems, 2024 - Elsevier
Recently, zero-shot multi-label classification has garnered considerable attention owing to
its capacity to predict unseen labels without human annotations. Nevertheless, prevailing …

Hierarchical Prompt Learning Using CLIP for Multi-label Classification with Single Positive Labels

A Wang, H Chen, Z Lin, Z Ding, P Liu, Y Bao… - Proceedings of the 31st …, 2023 - dl.acm.org
Collecting full annotations to construct multi-label datasets is difficult and labor-consuming.
As an effective solution to relieve the annotation burden, single positive multi-label learning …

A Versatile Multimodal Learning Framework For Zero-shot Emotion Recognition

F Qi, H Zhang, X Yang, C Xu - IEEE Transactions on Circuits …, 2024 - ieeexplore.ieee.org
Multi-modal Emotion Recognition (MER) aims to identify various human emotions from
heterogeneous modalities. With the development of emotional theories, there are more and …

Hierarchical multi-task learning via task affinity groupings

S Srivastava, S Bhugra, V Kaushik… - 2023 IEEE International …, 2023 - ieeexplore.ieee.org
Multi-task learning (MTL) permits joint task learning based on a shared deep learning
architecture and multiple loss functions. Despite the recent advances in MTL, one loss often …

TTD: Text-Tag Self-Distillation Enhancing Image-Text Alignment in CLIP to Alleviate Single Tag Bias

S Jo, S Ryu, S Kim, E Yang, K Kim - arXiv preprint arXiv:2404.00384, 2024 - arxiv.org
We identify a critical bias in contemporary CLIP-based models, which we denote as\textit
{single tag bias}. This bias manifests as a disproportionate focus on a singular tag (word) …

Functionally Similar Multi-Label Knowledge Distillation

B Chen, J Hu, X Zheng, W Lin… - ICASSP 2024-2024 …, 2024 - ieeexplore.ieee.org
Existing multi-label knowledge distillation methods simply use regression or single-label
classification methods without fully exploiting the essence of multi-label classification …

[HTML][HTML] Associating multiple vision transformer layers for fine-grained image representation

F Sun, HC Ngo, YW Sek, Z Meng - AI Open, 2023 - Elsevier
Accurate discriminative region proposal has an important effect for fine-grained image
recognition. The vision transformer (ViT) brings about a striking effect in computer vision due …

Determined Multi-Label Learning via Similarity-Based Prompt

M Wei, Z Li, P Ying, Y Zhou, X Xu - arXiv preprint arXiv:2403.16482, 2024 - arxiv.org
In multi-label classification, each training instance is associated with multiple class labels
simultaneously. Unfortunately, collecting the fully precise class labels for each training …