From task structures to world models: what do LLMs know?

I Yildirim, LA Paul - Trends in Cognitive Sciences, 2024 - cell.com
In what sense does a large language model (LLM) have knowledge? We answer by
granting LLMs 'instrumental knowledge': knowledge gained by using next-word generation …

Tree of thoughts: Deliberate problem solving with large language models

S Yao, D Yu, J Zhao, I Shafran… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Language models are increasingly being deployed for general problem solving
across a wide range of tasks, but are still confined to token-level, left-to-right decision …

A contrastive framework for neural text generation

Y Su, T Lan, Y Wang, D Yogatama… - Advances in Neural …, 2022 - proceedings.neurips.cc
Text generation is of great importance to many natural language processing applications.
However, maximization-based decoding methods (eg, beam search) of neural language …

Internet-augmented language models through few-shot prompting for open-domain question answering

A Lazaridou, E Gribovskaya, W Stokowiec… - arXiv preprint arXiv …, 2022 - arxiv.org
In this work, we aim to capitalize on the unique few-shot capabilities of large-scale language
models (LSLMs) to overcome some of their challenges with respect to grounding to factual …

Controlled text generation with natural language instructions

W Zhou, YE Jiang, E Wilcox… - International …, 2023 - proceedings.mlr.press
Large language models can be prompted to pro-duce fluent output for a wide range of tasks
without being specifically trained to do so. Nevertheless, it is notoriously difficult to control …

Planning with large language models for code generation

S Zhang, Z Chen, Y Shen, M Ding… - arXiv preprint arXiv …, 2023 - arxiv.org
Existing large language model-based code generation pipelines typically use beam search
or sampling algorithms during the decoding process. Although the programs they generate …

Branch-solve-merge improves large language model evaluation and generation

S Saha, O Levy, A Celikyilmaz, M Bansal… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) are frequently used for multi-faceted language generation
and evaluation tasks that involve satisfying intricate user constraints or taking into account …

Interactive natural language processing

Z Wang, G Zhang, K Yang, N Shi, W Zhou… - arXiv preprint arXiv …, 2023 - arxiv.org
Interactive Natural Language Processing (iNLP) has emerged as a novel paradigm within
the field of NLP, aimed at addressing limitations in existing frameworks while aligning with …

Naturalprover: Grounded mathematical proof generation with language models

S Welleck, J Liu, X Lu, H Hajishirzi… - Advances in Neural …, 2022 - proceedings.neurips.cc
Theorem proving in natural mathematical language–the mixture of symbolic and natural
language used by humans–plays a central role in mathematical advances and education …

I Can't Believe There's No Images! Learning Visual Tasks Using only Language Supervision

S Gu, C Clark, A Kembhavi - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Many high-level skills that are required for computer vision tasks, such as parsing questions,
comparing and contrasting semantics, and writing descriptions, are also required in other …