Learn from model beyond fine-tuning: A survey

H Zheng, L Shen, A Tang, Y Luo, H Hu, B Du… - arXiv preprint arXiv …, 2023 - arxiv.org
Foundation models (FM) have demonstrated remarkable performance across a wide range
of tasks (especially in the fields of natural language processing and computer vision) …

Multiway-adapater: Adapting large-scale multi-modal models for scalable image-text retrieval

Z Long, G Killick, R McCreadie… - arXiv preprint arXiv …, 2023 - arxiv.org
As the size of Large Multi-Modal Models (LMMs) increases consistently, the adaptation of
these pre-trained models to specialized tasks has become a computationally and memory …

Evaluating parameter-efficient transfer learning approaches on sure benchmark for speech understanding

Y Li, A Mehrish, R Bhardwaj… - ICASSP 2023-2023 …, 2023 - ieeexplore.ieee.org
Fine-tuning is widely used as the default algorithm for transfer learning from pre-trained
models. Parameter inefficiency can however arise when, during transfer learning, all the …

Audio-Adapterfusion: A Task-Id-Free Approach for Efficient and Non-Destructive Multi-Task Speech Recognition

H Ngai, R Agrawal, N Gaur, R Huang… - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
Adapters are an efficient, composable alternative to full fine-tuning of pre-trained models
and help scale the deployment of large ASR models to many tasks. In practice, a task ID is …

Low-rank adaptation method for wav2vec2-based fake audio detection

C Wang, J Yi, X Zhang, J Tao, L Xu, R Fu - arXiv preprint arXiv:2306.05617, 2023 - arxiv.org
Self-supervised speech models are a rapidly developing research topic in fake audio
detection. Many pre-trained models can serve as feature extractors, learning richer and …

AWE: Adaptive weight-space ensembling for few-shot fine-tuning

JC Gagnon-Audet, RP Monti… - ICLR 2023 Workshop on …, 2023 - openreview.net
In this paper, we introduce a new transfer learning approach called Adaptive Weight-space
Ensembling (AWE) that effectively adapts large pre-trained models for downstream tasks …

Ymir: A Scheduler for Foundation Model Fine-tuning Workloads in Datacenters

W Gao, W Zhuang, M Li, P Sun, Y Wen… - Proceedings of the 38th …, 2024 - dl.acm.org
The breakthrough of foundation models makes foundation model fine-tuning (FMF)
workloads prevalent in modern GPU datacenters. However, existing schedulers tailored for …

SRIL: Selective Regularization for Class-Incremental Learning

J Han, J Na, W Hwang - arXiv preprint arXiv:2305.05175, 2023 - arxiv.org
Human intelligence gradually accepts new information and accumulates knowledge
throughout the lifespan. However, deep learning models suffer from a catastrophic forgetting …

Parameter-Efficient Tuning with Adaptive Bottlenecks for Automatic Speech Recognition

G Vanderreydt, A Prasad, D Khalil… - 2023 IEEE Automatic …, 2023 - ieeexplore.ieee.org
Transfer learning from large multilingual pretrained models, like XLSR, has become the new
paradigm for Automatic Speech Recognition (ASR). Considering their ever-increasing size …

Unsupervised Online Continual Learning for Automatic Speech Recognition

SV Eeckt - arXiv preprint arXiv:2406.12503, 2024 - arxiv.org
Adapting Automatic Speech Recognition (ASR) models to new domains leads to
Catastrophic Forgetting (CF) of previously learned information. This paper addresses CF in …