Trusting your evidence: Hallucinate less with context-aware decoding

W Shi, X Han, M Lewis, Y Tsvetkov… - arXiv preprint arXiv …, 2023 - arxiv.org
Language models (LMs) often struggle to pay enough attention to the input context, and
generate texts that are unfaithful or contain hallucinations. To mitigate this issue, we present …

Gpt-re: In-context learning for relation extraction using large language models

Z Wan, F Cheng, Z Mao, Q Liu, H Song, J Li… - arXiv preprint arXiv …, 2023 - arxiv.org
In spite of the potential for ground-breaking achievements offered by large language models
(LLMs)(eg, GPT-3), they still lag significantly behind fully-supervised baselines (eg, fine …

Thinking about gpt-3 in-context learning for biomedical ie? think again

BJ Gutierrez, N McNeal, C Washington, Y Chen… - arXiv preprint arXiv …, 2022 - arxiv.org
The strong few-shot in-context learning capability of large pre-trained language models
(PLMs) such as GPT-3 is highly appealing for application domains such as biomedicine …

Multitask prompt tuning enables parameter-efficient transfer learning

Z Wang, R Panda, L Karlinsky, R Feris, H Sun… - arXiv preprint arXiv …, 2023 - arxiv.org
Prompt tuning, in which a base pretrained model is adapted to each task via conditioning on
learned prompt vectors, has emerged as a promising approach for efficiently adapting large …

Exposing attention glitches with flip-flop language modeling

B Liu, J Ash, S Goel… - Advances in Neural …, 2024 - proceedings.neurips.cc
Why do large language models sometimes output factual inaccuracies and exhibit
erroneous reasoning? The brittleness of these models, particularly when executing long …

Promptagent: Strategic planning with language models enables expert-level prompt optimization

X Wang, C Li, Z Wang, F Bai, H Luo, J Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
Highly effective, task-specific prompts are often heavily engineered by experts to integrate
detailed instructions and domain insights based on a deep understanding of both instincts of …

Amortizing intractable inference in large language models

EJ Hu, M Jain, E Elmoznino, Y Kaddar, G Lajoie… - arXiv preprint arXiv …, 2023 - arxiv.org
Autoregressive large language models (LLMs) compress knowledge from their training data
through next-token conditional distributions. This limits tractable querying of this knowledge …

Coco: Coherence-enhanced machine-generated text detection under data limitation with contrastive learning

X Liu, Z Zhang, Y Wang, H Pu, Y Lan… - arXiv preprint arXiv …, 2022 - arxiv.org
Machine-Generated Text (MGT) detection, a task that discriminates MGT from Human-
Written Text (HWT), plays a crucial role in preventing misuse of text generative models …

Stay on topic with classifier-free guidance

G Sanchez, H Fan, A Spangher, E Levi… - arXiv preprint arXiv …, 2023 - arxiv.org
Classifier-Free Guidance (CFG) has recently emerged in text-to-image generation as a
lightweight technique to encourage prompt-adherence in generations. In this work, we …

Coco: Coherence-enhanced machine-generated text detection under low resource with contrastive learning

X Liu, Z Zhang, Y Wang, H Pu, Y Lan… - Proceedings of the 2023 …, 2023 - aclanthology.org
Abstract Machine-Generated Text (MGT) detection, a task that discriminates MGT from
Human-Written Text (HWT), plays a crucial role in preventing misuse of text generative …