Parameter-efficient fine-tuning for large models: A comprehensive survey

Z Han, C Gao, J Liu, SQ Zhang - arXiv preprint arXiv:2403.14608, 2024 - arxiv.org
Large models represent a groundbreaking advancement in multiple application fields,
enabling remarkable achievements across various tasks. However, their unprecedented …

Scaling down to scale up: A guide to parameter-efficient fine-tuning

V Lialin, V Deshpande, A Rumshisky - arXiv preprint arXiv:2303.15647, 2023 - arxiv.org
This paper presents a systematic overview and comparison of parameter-efficient fine-tuning
methods covering over 40 papers published between February 2019 and February 2023 …

Parameter-efficient fine-tuning methods for pretrained language models: A critical review and assessment

L Xu, H Xie, SZJ Qin, X Tao, FL Wang - arXiv preprint arXiv:2312.12148, 2023 - arxiv.org
With the continuous growth in the number of parameters of transformer-based pretrained
language models (PLMs), particularly the emergence of large language models (LLMs) with …

Exploring parameter-efficient fine-tuning techniques for code generation with large language models

M Weyssow, X Zhou, K Kim, D Lo… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) possess impressive capabilities to generate meaningful
code snippets given natural language intents in zero-shot, ie, without the need for specific …

Llm-adapters: An adapter family for parameter-efficient fine-tuning of large language models

Z Hu, L Wang, Y Lan, W Xu, EP Lim, L Bing… - arXiv preprint arXiv …, 2023 - arxiv.org
The success of large language models (LLMs), like GPT-4 and ChatGPT, has led to the
development of numerous cost-effective and accessible alternatives that are created by …

Increlora: Incremental parameter allocation method for parameter-efficient fine-tuning

F Zhang, L Li, J Chen, Z Jiang, B Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
With the increasing size of pre-trained language models (PLMs), fine-tuning all the
parameters in the model is not efficient, especially when there are a large number of …

Memory-efficient fine-tuning of compressed large language models via sub-4-bit integer quantization

J Kim, JH Lee, S Kim, J Park, KM Yoo… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large language models (LLMs) face the challenges in fine-tuning and deployment due to
their high memory demands and computational costs. While parameter-efficient fine-tuning …

Parameter-efficient fine-tuning without introducing new latency

B Liao, Y Meng, C Monz - arXiv preprint arXiv:2305.16742, 2023 - arxiv.org
Parameter-efficient fine-tuning (PEFT) of pre-trained language models has recently
demonstrated remarkable achievements, effectively matching the performance of full fine …

Compacter: Efficient low-rank hypercomplex adapter layers

R Karimi Mahabadi, J Henderson… - Advances in Neural …, 2021 - proceedings.neurips.cc
Adapting large-scale pretrained language models to downstream tasks via fine-tuning is the
standard method for achieving state-of-the-art performance on NLP benchmarks. However …

Delta tuning: A comprehensive study of parameter efficient methods for pre-trained language models

N Ding, Y Qin, G Yang, F Wei, Z Yang, Y Su… - arXiv preprint arXiv …, 2022 - arxiv.org
Despite the success, the process of fine-tuning large-scale PLMs brings prohibitive
adaptation costs. In fact, fine-tuning all the parameters of a colossal model and retaining …