Domain specialization as the key to make large language models disruptive: A comprehensive survey

C Ling, X Zhao, J Lu, C Deng, C Zheng, J Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have significantly advanced the field of natural language
processing (NLP), providing a highly useful, task-agnostic foundation for a wide range of …

ScaLearn: Simple and Highly Parameter-Efficient Task Transfer by Learning to Scale

M Frohmann, C Holtermann, S Masoudian… - arXiv preprint arXiv …, 2023 - arxiv.org
Multi-task learning (MTL) has shown considerable practical benefits, particularly when using
pre-trained language models (PLMs). While this is commonly achieved by simultaneously …

[PDF][PDF] Beyond One-Model-Fits-All: A Survey of Domain Specialization for Large Language Models

X ZHAO, J LU, C DENG, CAN ZHENG… - arXiv preprint arXiv …, 2023 - academia.edu
2 Ling, et al. in very recent years on the domain specialization of LLMs, which, however,
calls for a comprehensive and systematic review to better summarizes and guide this …

Modular and Parameter-efficient Fine-tuning of Language Models

J Pfeiffer - 2023 - tuprints.ulb.tu-darmstadt.de
Transfer learning has recently become the dominant paradigm of natural language
processing. Models pre-trained on unlabeled data can be fine-tuned for downstream tasks …

[PDF][PDF] Information Propagation in Modular Language Modeling and Web Tracking

Z Su - 2024 - di.ku.dk
Abstract Information propagation is the process through which data are transmitted within a
system. The growth of large-scale web datasets has led to explosive growth in information …

Preventing Forgetting and Promoting Transfer in Continual Learning

L Pagé-Caccia - 2023 - search.proquest.com
In this thesis, we explore the development of methods for efficientcontinual learning, to
enable the sequential acquisition of knowledge by artificial intelligence (AI) systems as they …