Compositional exemplars for in-context learning

J Ye, Z Wu, J Feng, T Yu… - … Conference on Machine …, 2023 - proceedings.mlr.press
Large pretrained language models (LMs) have shown impressive In-Context Learning (ICL)
ability, where the model learns to do an unseen task simply by conditioning on a prompt …

Learning visually grounded words and syntax for a scene description task

DK Roy - Computer speech & language, 2002 - Elsevier
A spoken language generation system has been developed that learns to describe objects
in computer-generated visual scenes. The system is trained by a 'show-and-tell'procedure in …

[PDF][PDF] Taskprompter: Spatial-channel multi-task prompting for dense scene understanding

H Ye, D Xu - The Eleventh International Conference on Learning …, 2022 - drive.google.com
Learning effective representations simultaneously from multiple tasks in a unified network
framework is a fundamental paradigm for multi-task dense visual scene understanding. This …

The inductive bias of in-context learning: Rethinking pretraining example design

Y Levine, N Wies, D Jannai, D Navon… - arXiv preprint arXiv …, 2021 - arxiv.org
Pretraining Neural Language Models (NLMs) over a large corpus involves chunking the text
into training examples, which are contiguous text segments of sizes processable by the …

Mammut: A simple architecture for joint learning for multimodal tasks

W Kuo, AJ Piergiovanni, D Kim, X Luo, B Caine… - arXiv preprint arXiv …, 2023 - arxiv.org
The development of language models have moved from encoder-decoder to decoder-only
designs. In addition, we observe that the two most popular multimodal tasks, the generative …

Understanding in-context learning via supportive pretraining data

X Han, D Simig, T Mihaylov, Y Tsvetkov… - arXiv preprint arXiv …, 2023 - arxiv.org
In-context learning (ICL) improves language models' performance on a variety of NLP tasks
by simply demonstrating a handful of examples at inference time. It is not well understood …

Focused transformer: Contrastive training for context scaling

S Tworkowski, K Staniszewski… - Advances in …, 2024 - proceedings.neurips.cc
Large language models have an exceptional capability to incorporate new information in a
contextual manner. However, the full potential of such an approach is often restrained due to …

Learning to prompt for vision-language models

K Zhou, J Yang, CC Loy, Z Liu - International Journal of Computer Vision, 2022 - Springer
Large pre-trained vision-language models like CLIP have shown great potential in learning
representations that are transferable across a wide range of downstream tasks. Different …

Counterfactual vision-and-language navigation: Unravelling the unseen

A Parvaneh, E Abbasnejad, D Teney… - Advances in neural …, 2020 - proceedings.neurips.cc
The task of vision-and-language navigation (VLN) requires an agent to follow text
instructions to find its way through simulated household environments. A prominent …

Icl-d3ie: In-context learning with diverse demonstrations updating for document information extraction

J He, L Wang, Y Hu, N Liu, H Liu… - Proceedings of the …, 2023 - openaccess.thecvf.com
Large language models (LLMs), such as GPT-3 and ChatGPT, have demonstrated
remarkable results in various natural language processing (NLP) tasks with in-context …