Fine-tuning can distort pretrained features and underperform out-of-distribution

A Kumar, A Raghunathan, R Jones, T Ma… - arXiv preprint arXiv …, 2022 - arxiv.org
When transferring a pretrained model to a downstream task, two popular methods are full
fine-tuning (updating all the model parameters) and linear probing (updating only the last …

Do prompt-based models really understand the meaning of their prompts?

A Webson, E Pavlick - arXiv preprint arXiv:2109.01247, 2021 - arxiv.org
Recently, a boom of papers has shown extraordinary progress in zero-shot and few-shot
learning with various prompt-based models. It is commonly argued that prompts help models …

Shortcut learning of large language models in natural language understanding

M Du, F He, N Zou, D Tao, X Hu - Communications of the ACM, 2023 - dl.acm.org
Shortcut Learning of Large Language Models in Natural Language Understanding Page 1 110
COMMUNICATIONS OF THE ACM | JANUARY 2024 | VOL. 67 | NO. 1 research IMA GE B Y …

Stretching sentence-pair NLI models to reason over long documents and clusters

T Schuster, S Chen, S Buthpitiya, A Fabrikant… - arXiv preprint arXiv …, 2022 - arxiv.org
Natural Language Inference (NLI) has been extensively studied by the NLP community as a
framework for estimating the semantic relation between sentence pairs. While early work …

Fast trainable projection for robust fine-tuning

J Tian, YC Liu, JS Smith, Z Kira - Advances in Neural …, 2024 - proceedings.neurips.cc
Robust fine-tuning aims to achieve competitive in-distribution (ID) performance while
maintaining the out-of-distribution (OOD) robustness of a pre-trained model when …

Out-of-distribution generalization in natural language processing: Past, present, and future

L Yang, Y Song, X Ren, C Lyu, Y Wang… - Proceedings of the …, 2023 - aclanthology.org
Abstract Machine learning (ML) systems in natural language processing (NLP) face
significant challenges in generalizing to out-of-distribution (OOD) data, where the test …

Mind the biases: Quantifying cognitive biases in language model prompting

R Lin, HT Ng - Findings of the Association for Computational …, 2023 - aclanthology.org
We advocate the importance of exposing uncertainty on results of language model
prompting which display bias modes resembling cognitive biases, and propose to help …

Towards unified prompt tuning for few-shot text classification

J Wang, C Wang, F Luo, C Tan, M Qiu, F Yang… - arXiv preprint arXiv …, 2022 - arxiv.org
Prompt-based fine-tuning has boosted the performance of Pre-trained Language Models
(PLMs) on few-shot text classification by employing task-specific prompts. Yet, PLMs are …

Feature-level debiased natural language understanding

Y Lyu, P Li, Y Yang, M de Rijke, P Ren… - Proceedings of the …, 2023 - ojs.aaai.org
Natural language understanding (NLU) models often rely on dataset biases rather than
intended task-relevant features to achieve high performance on specific datasets. As a …

Evaluating the robustness of discrete prompts

Y Ishibashi, D Bollegala, K Sudoh… - arXiv preprint arXiv …, 2023 - arxiv.org
Discrete prompts have been used for fine-tuning Pre-trained Language Models for diverse
NLP tasks. In particular, automatic methods that generate discrete prompts from a small set …