J Liang, R He, T Tan - International Journal of Computer Vision, 2024 - Springer
Abstract Machine learning methods strive to acquire a robust model during the training process that can effectively generalize to test samples, even in the presence of distribution …
Incrementally learning new information from a non-stationary stream of data, referred to as 'continual learning', is a key feature of natural intelligence, but a challenging problem for …
Test-time domain adaptation aims to adapt a source pre-trained model to a target domain without using any source data. Existing works mainly consider the case where the target …
Continual learning aims to enable a single model to learn a sequence of tasks without catastrophic forgetting. Top-performing methods usually require a rehearsal buffer to store …
In this paper, we study dataset distillation (DD), from a novel perspective and introduce a\emph {dataset factorization} approach, termed\emph {HaBa}, which is a plug-and-play …
The landscape of publicly available vision foundation models (VFMs) such as CLIP and SAM is expanding rapidly. VFMs are endowed with distinct capabilities stemming from their …
Deep models, eg, CNNs and Vision Transformers, have achieved impressive achievements in many vision tasks in the closed world. However, novel classes emerge from time to time in …
W Huang, M Ye, B Du - … of the IEEE/CVF Conference on …, 2022 - openaccess.thecvf.com
Federated learning has emerged as an important distributed learning paradigm, which normally involves collaborative updating with others and local updating on private data …
B Zhu, Y Niu, Y Han, Y Wu… - Proceedings of the IEEE …, 2023 - openaccess.thecvf.com
Thanks to the large pre-trained vision-language models (VLMs) like CLIP, we can craft a zero-shot classifier by discrete prompt design, eg, the confidence score of an image …