Openassistant conversations-democratizing large language model alignment

A Köpf, Y Kilcher, D von Rütte… - Advances in …, 2024 - proceedings.neurips.cc
Aligning large language models (LLMs) with human preferences has proven to drastically
improve usability and has driven rapid adoption as demonstrated by ChatGPT. Alignment …

Aligning large language models with human: A survey

Y Wang, W Zhong, L Li, F Mi, X Zeng, W Huang… - arXiv preprint arXiv …, 2023 - arxiv.org
Large Language Models (LLMs) trained on extensive textual corpora have emerged as
leading solutions for a broad array of Natural Language Processing (NLP) tasks. Despite …

Aligning large language models through synthetic feedback

S Kim, S Bae, J Shin, S Kang, D Kwak, KM Yoo… - arXiv preprint arXiv …, 2023 - arxiv.org
Aligning large language models (LLMs) to human values has become increasingly
important as it enables sophisticated steering of LLMs. However, it requires significant …

Beyond imitation: Leveraging fine-grained quality signals for alignment

G Guo, R Zhao, T Tang, WX Zhao, JR Wen - arXiv preprint arXiv …, 2023 - arxiv.org
Alignment with human preference is a desired property of large language models (LLMs).
Currently, the main alignment approach is based on reinforcement learning from human …

Alignbench: Benchmarking chinese alignment of large language models

X Liu, X Lei, S Wang, Y Huang, Z Feng, B Wen… - arXiv preprint arXiv …, 2023 - arxiv.org
Alignment has become a critical step for instruction-tuned Large Language Models (LLMs)
to become helpful assistants. However, effective evaluation of alignment for emerging …

Black-box prompt optimization: Aligning large language models without model training

J Cheng, X Liu, K Zheng, P Ke, H Wang, Y Dong… - arXiv preprint arXiv …, 2023 - arxiv.org
Large language models (LLMs) have shown impressive success in various applications.
However, these models are often not well aligned with human intents, which calls for …

Peering through preferences: Unraveling feedback acquisition for aligning large language models

H Bansal, J Dang, A Grover - arXiv preprint arXiv:2308.15812, 2023 - arxiv.org
Aligning large language models (LLMs) with human values and intents critically involves the
use of human or AI feedback. While dense feedback annotations are expensive to acquire …

The unlocking spell on base llms: Rethinking alignment via in-context learning

BY Lin, A Ravichander, X Lu, N Dziri… - The Twelfth …, 2023 - openreview.net
Alignment tuning has become the de facto standard practice for enabling base large
language models (LLMs) to serve as open-domain AI assistants. The alignment tuning …

Language model alignment with elastic reset

M Noukhovitch, S Lavoie, F Strub… - Advances in Neural …, 2024 - proceedings.neurips.cc
Finetuning language models with reinforcement learning (RL), eg from human feedback
(HF), is a prominent method for alignment. But optimizing against a reward model can …

From Instructions to Intrinsic Human Values--A Survey of Alignment Goals for Big Models

J Yao, X Yi, X Wang, J Wang, X Xie - arXiv preprint arXiv:2308.12014, 2023 - arxiv.org
Big models, exemplified by Large Language Models (LLMs), are models typically pre-
trained on massive data and comprised of enormous parameters, which not only obtain …