Few-shot parameter-efficient fine-tuning is better and cheaper than in-context learning

H Liu, D Tam, M Muqeeth, J Mohta… - Advances in …, 2022 - proceedings.neurips.cc
Few-shot in-context learning (ICL) enables pre-trained language models to perform a
previously-unseen task without any gradient-based training by feeding a small number of …

Efficient few-shot learning without prompts

L Tunstall, N Reimers, UES Jo, L Bates, D Korat… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent few-shot methods, such as parameter-efficient fine-tuning (PEFT) and pattern
exploiting training (PET), have achieved impressive results in label-scarce settings …

Opportunities and challenges in neural dialog tutoring

J Macina, N Daheim, L Wang, T Sinha, M Kapur… - arXiv preprint arXiv …, 2023 - arxiv.org
Designing dialog tutors has been challenging as it involves modeling the diverse and
complex pedagogical strategies employed by human tutors. Although there have been …

Thinking about gpt-3 in-context learning for biomedical ie? think again

BJ Gutierrez, N McNeal, C Washington, Y Chen… - arXiv preprint arXiv …, 2022 - arxiv.org
The strong few-shot in-context learning capability of large pre-trained language models
(PLMs) such as GPT-3 is highly appealing for application domains such as biomedicine …

Natural language based context modeling and reasoning with llms: A tutorial

H Xiong, J Bian, S Yang, X Zhang, L Kong… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have become phenomenally surging, since 2018--two
decades after introducing context-awareness into computing systems. Through taking into …

Breaking physical and linguistic borders: Multilingual federated prompt tuning for low-resource languages

W Zhao, Y Chen, R Lee, X Qiu, Y Gao… - The Twelfth …, 2024 - openreview.net
Pretrained large language models (LLMs) have emerged as a cornerstone in modern
natural language processing, with their utility expanding to various applications and …

Instruction tuned models are quick learners

H Gupta, SA Sawant, S Mishra, M Nakamura… - arXiv preprint arXiv …, 2023 - arxiv.org
Instruction tuning of language models has demonstrated the ability to enhance model
generalization to unseen tasks via in-context learning using a few examples. However …

Automatically Inspecting Thousands of Static Bug Warnings with Large Language Model: How Far Are We?

C Wen, Y Cai, B Zhang, J Su, Z Xu, D Liu… - ACM Transactions on …, 2024 - dl.acm.org
Static analysis tools for capturing bugs and vulnerabilities in software programs are widely
employed in practice, as they have the unique advantages of high coverage and …

Federated few-shot learning for mobile nlp

D Cai, S Wang, Y Wu, FX Lin, M Xu - Proceedings of the 29th Annual …, 2023 - dl.acm.org
Natural language processing (NLP) sees rich mobile applications. To support various
language understanding tasks, a foundation NLP model is often fine-tuned in a federated …

A survey of methods for addressing class imbalance in deep-learning based natural language processing

S Henning, W Beluch, A Fraser, A Friedrich - arXiv preprint arXiv …, 2022 - arxiv.org
Many natural language processing (NLP) tasks are naturally imbalanced, as some target
categories occur much more frequently than others in the real world. In such scenarios …