The cot collection: Improving zero-shot and few-shot learning of language models via chain-of-thought fine-tuning

S Kim, SJ Joo, D Kim, J Jang, S Ye, J Shin… - arXiv preprint arXiv …, 2023 - arxiv.org
Language models (LMs) with less than 100B parameters are known to perform poorly on
chain-of-thought (CoT) reasoning in contrast to large LMs when solving unseen tasks. In this …

Complex QA and language models hybrid architectures, Survey

X Daull, P Bellot, E Bruno, V Martin… - arXiv preprint arXiv …, 2023 - arxiv.org
This paper reviews the state-of-the-art of language models architectures and strategies for"
complex" question-answering (QA, CQA, CPS) with a focus on hybridization. Large …

A comprehensive survey on instruction following

R Lou, K Zhang, W Yin - arXiv preprint arXiv:2303.10475, 2023 - arxiv.org
Task semantics can be expressed by a set of input-output examples or a piece of textual
instruction. Conventional machine learning approaches for natural language processing …

Muffin: Curating multi-faceted instructions for improving instruction following

R Lou, K Zhang, J Xie, Y Sun, J Ahn, H Xu… - The Twelfth …, 2023 - openreview.net
In the realm of large language models (LLMs), enhancing instruction-following capability
often involves curating expansive training data. This is achieved through two primary …

On measurement validity and language models: Increasing validity and decreasing bias with instructions

M Laurer, W van Atteveldt, A Casas… - … Methods and Measures, 2024 - Taylor & Francis
Language models like BERT or GPT are becoming increasingly popular measurement tools,
but are the measurements they produce valid? Literature suggests that there is still a …

Building Efficient Universal Classifiers with Natural Language Inference

M Laurer, W van Atteveldt, A Casas… - arXiv preprint arXiv …, 2023 - arxiv.org
Generative Large Language Models (LLMs) have become the mainstream choice for
fewshot and zeroshot learning thanks to the universality of text generation. Many users …

X-Shot: A Unified System to Handle Frequent, Few-shot and Zero-shot Learning Simultaneously in Classification

H Xu, M Chen, L Huang, S Vucetic, W Yin - arXiv preprint arXiv …, 2024 - arxiv.org
In recent years, few-shot and zero-shot learning, which learn to predict labels with limited
annotated instances, have garnered significant attention. Traditional approaches often treat …

Enabling Natural Zero-Shot Prompting on Encoder Models via Statement-Tuning

A Elshabrawy, Y Huang, I Gurevych, AF Aji - arXiv preprint arXiv …, 2024 - arxiv.org
While Large Language Models (LLMs) exhibit remarkable capabilities in zero-shot and few-
shot scenarios, they often require computationally prohibitive sizes. Conversely, smaller …

Large Language Model Instruction Following: A Survey of Progresses and Challenges

R Lou, K Zhang, W Yin - Computational Linguistics, 2024 - direct.mit.edu
Task semantics can be expressed by a set of input-output examples or a piece of textual
instruction. Conventional machine learning approaches for natural language processing …

SparseCL: Sparse Contrastive Learning for Contradiction Retrieval

H Xu, Z Lin, Y Sun, KW Chang, P Indyk - arXiv preprint arXiv:2406.10746, 2024 - arxiv.org
Contradiction retrieval refers to identifying and extracting documents that explicitly disagree
with or refute the content of a query, which is important to many downstream applications …