Palm: Scaling language modeling with pathways

A Chowdhery, S Narang, J Devlin, M Bosma… - Journal of Machine …, 2023 - jmlr.org
Large language models have been shown to achieve remarkable performance across a
variety of natural language tasks using few-shot learning, which drastically reduces the …

Making pre-trained language models better few-shot learners

T Gao, A Fisch, D Chen - arXiv preprint arXiv:2012.15723, 2020 - arxiv.org
The recent GPT-3 model (Brown et al., 2020) achieves remarkable few-shot performance
solely by leveraging a natural-language prompt and a few task demonstrations as input …

Tuning language models as training data generators for augmentation-enhanced few-shot learning

Y Meng, M Michalski, J Huang… - International …, 2023 - proceedings.mlr.press
Recent studies have revealed the intriguing few-shot learning ability of pretrained language
models (PLMs): They can quickly adapt to a new task when fine-tuned on a small amount of …

Cutting down on prompts and parameters: Simple few-shot learning with language models

RL Logan IV, I Balažević, E Wallace, F Petroni… - arXiv preprint arXiv …, 2021 - arxiv.org
Prompting language models (LMs) with training examples and task descriptions has been
seen as critical to recent successes in few-shot learning. In this work, we show that …

Entailment as few-shot learner

S Wang, H Fang, M Khabsa, H Mao, H Ma - arXiv preprint arXiv …, 2021 - arxiv.org
Large pre-trained language models (LMs) have demonstrated remarkable ability as few-shot
learners. However, their success hinges largely on scaling model parameters to a degree …

Revisiting self-training for few-shot learning of language model

Y Chen, Y Zhang, C Zhang, G Lee, R Cheng… - arXiv preprint arXiv …, 2021 - arxiv.org
As unlabeled data carry rich task-relevant information, they are proven useful for few-shot
learning of language model. The question is how to effectively make use of such data. In this …

Language models are few-shot learners

T Brown, B Mann, N Ryder… - Advances in neural …, 2020 - proceedings.neurips.cc
We demonstrate that scaling up language models greatly improves task-agnostic, few-shot
performance, sometimes even becoming competitive with prior state-of-the-art fine-tuning …

Few-shot learning with multilingual language models

XV Lin, T Mihaylov, M Artetxe, T Wang, S Chen… - arXiv preprint arXiv …, 2021 - arxiv.org
Large-scale generative language models such as GPT-3 are competitive few-shot learners.
While these models are known to be able to jointly represent many different languages, their …

Few-shot learning with multilingual generative language models

XV Lin, T Mihaylov, M Artetxe, T Wang… - Proceedings of the …, 2022 - aclanthology.org
Large-scale generative language models such as GPT-3 are competitive few-shot learners.
While these models are known to be able to jointly represent many different languages, their …

RAFT: A real-world few-shot text classification benchmark

N Alex, E Lifland, L Tunstall, A Thakur… - arXiv preprint arXiv …, 2021 - arxiv.org
Large pre-trained language models have shown promise for few-shot learning, completing
text-based tasks given only a few task-specific examples. Will models soon solve …