A comprehensive survey on test-time adaptation under distribution shifts

J Liang, R He, T Tan - International Journal of Computer Vision, 2024 - Springer
Abstract Machine learning methods strive to acquire a robust model during the training
process that can effectively generalize to test samples, even in the presence of distribution …

Adaptersoup: Weight averaging to improve generalization of pretrained language models

A Chronopoulou, ME Peters, A Fraser… - arXiv preprint arXiv …, 2023 - arxiv.org
Pretrained language models (PLMs) are trained on massive corpora, but often need to
specialize to specific domains. A parameter-efficient adaptation method suggests training an …

Adapters: A unified library for parameter-efficient and modular transfer learning

C Poth, H Sterz, I Paul, S Purkayastha… - arXiv preprint arXiv …, 2023 - arxiv.org
We introduce Adapters, an open-source library that unifies parameter-efficient and modular
transfer learning in large language models. By integrating 10 diverse adapter methods into a …

mmt5: Modular multilingual pre-training solves source language hallucinations

J Pfeiffer, F Piccinno, M Nicosia, X Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Multilingual sequence-to-sequence models perform poorly with increased language
coverage and fail to consistently generate text in the correct target language in few-shot …

Adapter-based fine-tuning of pre-trained multilingual language models for code-mixed and code-switched text classification

H Rathnayake, J Sumanapala, R Rukshani… - … and Information Systems, 2022 - Springer
Code-mixing and code-switching are frequent features in online conversations.
Classification of such text is challenging if one of the languages is low-resourced. Fine …

Modular and parameter-efficient fine-tuning for nlp models

S Ruder, J Pfeiffer, I Vulić - … of the 2022 Conference on Empirical …, 2022 - aclanthology.org
State-of-the-art language models in NLP perform best when fine-tuned even on small
datasets, but due to their increasing size, fine-tuning and downstream usage have become …

Survey on publicly available sinhala natural language processing tools and research

N De Silva - arXiv preprint arXiv:1906.02358, 2019 - arxiv.org
Sinhala is the native language of the Sinhalese people who make up the largest ethnic
group of Sri Lanka. The language belongs to the globe-spanning language tree, Indo …

Modularized transfer learning with multiple knowledge graphs for zero-shot commonsense reasoning

YJ Kim, B Kwak, Y Kim, RK Amplayo, S Hwang… - arXiv preprint arXiv …, 2022 - arxiv.org
Commonsense reasoning systems should be able to generalize to diverse reasoning cases.
However, most state-of-the-art approaches depend on expensive data annotations and …

Domain generalisation of NMT: Fusing adapters with leave-one-domain-out training

TT Vu, S Khadivi, D Phung… - Annual Meeting of the …, 2022 - research.monash.edu
Generalising to unseen domains is under-explored and remains a challenge in neural
machine translation. Inspired by recent research in parameter-efficient transfer learning from …

Model ensemble instead of prompt fusion: a sample-specific knowledge transfer method for few-shot prompt tuning

X Peng, C Xing, PK Choubey, CS Wu… - arXiv preprint arXiv …, 2022 - arxiv.org
Prompt tuning approaches, which learn task-specific soft prompts for a downstream task
conditioning on frozen pre-trained models, have attracted growing interest due to its …