Retrieval-augmented generation for large language models: A survey

Y Gao, Y Xiong, X Gao, K Jia, J Pan, Y Bi, Y Dai… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) demonstrate powerful capabilities, but they still face
challenges in practical applications, such as hallucinations, slow knowledge updates, and …

Evaluating large language models: A comprehensive survey

Z Guo, R Jin, C Liu, Y Huang, D Shi, L Yu, Y Liu… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have demonstrated remarkable capabilities across a broad
spectrum of tasks. They have attracted significant attention and been deployed in numerous …

Principle-driven self-alignment of language models from scratch with minimal human supervision

Z Sun, Y Shen, Q Zhou, H Zhang… - Advances in …, 2024 - proceedings.neurips.cc
Recent AI-assistant agents, such as ChatGPT, predominantly rely on supervised fine-tuning
(SFT) with human annotations and reinforcement learning from human feedback (RLHF) to …

Language models can solve computer tasks

G Kim, P Baldi, S McAleer - Advances in Neural Information …, 2024 - proceedings.neurips.cc
Agents capable of carrying out general tasks on a computer can improve efficiency and
productivity by automating repetitive tasks and assisting in complex problem-solving. Ideally …

Can language models solve graph problems in natural language?

H Wang, S Feng, T He, Z Tan, X Han… - Advances in Neural …, 2024 - proceedings.neurips.cc
Large language models (LLMs) are increasingly adopted for a variety of tasks with implicit
graphical structures, such as planning in robotics, multi-hop question answering or …

Interleaving retrieval with chain-of-thought reasoning for knowledge-intensive multi-step questions

H Trivedi, N Balasubramanian, T Khot… - arXiv preprint arXiv …, 2022 - arxiv.org
Prompting-based large language models (LLMs) are surprisingly powerful at generating
natural language reasoning steps or Chains-of-Thoughts (CoT) for multi-step question …

Demonstrate-search-predict: Composing retrieval and language models for knowledge-intensive nlp

O Khattab, K Santhanam, XL Li, D Hall, P Liang… - arXiv preprint arXiv …, 2022 - arxiv.org
Retrieval-augmented in-context learning has emerged as a powerful approach for
addressing knowledge-intensive tasks using frozen language models (LM) and retrieval …

Freshllms: Refreshing large language models with search engine augmentation

T Vu, M Iyyer, X Wang, N Constant, J Wei, J Wei… - arXiv preprint arXiv …, 2023 - arxiv.org
Most large language models (LLMs) are trained once and never updated; thus, they lack the
ability to dynamically adapt to our ever-changing world. In this work, we perform a detailed …

Large language model is not a good few-shot information extractor, but a good reranker for hard samples!

Y Ma, Y Cao, YC Hong, A Sun - arXiv preprint arXiv:2303.08559, 2023 - arxiv.org
Large Language Models (LLMs) have made remarkable strides in various tasks. Whether
LLMs are competitive few-shot solvers for information extraction (IE) tasks, however, remains …

Query2doc: Query expansion with large language models

L Wang, N Yang, F Wei - arXiv preprint arXiv:2303.07678, 2023 - arxiv.org
This paper introduces a simple yet effective query expansion approach, denoted as
query2doc, to improve both sparse and dense retrieval systems. The proposed method first …