Reinforcement learning algorithms: A brief survey

AK Shakya, G Pillai, S Chakrabarty - Expert Systems with Applications, 2023 - Elsevier
Reinforcement Learning (RL) is a machine learning (ML) technique to learn sequential
decision-making in complex problems. RL is inspired by trial-and-error based human/animal …

Concepts of artificial intelligence for computer-assisted drug discovery

X Yang, Y Wang, R Byrne, G Schneider… - Chemical …, 2019 - ACS Publications
Artificial intelligence (AI), and, in particular, deep learning as a subcategory of AI, provides
opportunities for the discovery and development of innovative drugs. Various machine …

Rlaif: Scaling reinforcement learning from human feedback with ai feedback

H Lee, S Phatale, H Mansoor, KR Lu, T Mesnard… - 2023 - openreview.net
Reinforcement learning from human feedback (RLHF) is an effective technique for aligning
large language models (LLMs) to human preferences, but gathering high-quality human …

Red teaming language models with language models

E Perez, S Huang, F Song, T Cai, R Ring… - arXiv preprint arXiv …, 2022 - arxiv.org
Language Models (LMs) often cannot be deployed because of their potential to harm users
in hard-to-predict ways. Prior work identifies harmful behaviors before deployment by using …

Slic-hf: Sequence likelihood calibration with human feedback

Y Zhao, R Joshi, T Liu, M Khalman, M Saleh… - arXiv preprint arXiv …, 2023 - arxiv.org
Learning from human feedback has been shown to be effective at aligning language models
with human preferences. Past work has often relied on Reinforcement Learning from Human …

Foundation models for decision making: Problems, methods, and opportunities

S Yang, O Nachum, Y Du, J Wei, P Abbeel… - arXiv preprint arXiv …, 2023 - arxiv.org
Foundation models pretrained on diverse data at scale have demonstrated extraordinary
capabilities in a wide range of vision and language tasks. When such models are deployed …

Teaching language models to support answers with verified quotes

J Menick, M Trebacz, V Mikulik, J Aslanides… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent large language models often answer factual questions correctly. But users can't trust
any given claim a model makes without fact-checking, because language models can …

Learning to summarize with human feedback

N Stiennon, L Ouyang, J Wu… - Advances in …, 2020 - proceedings.neurips.cc
As language models become more powerful, training and evaluation are increasingly
bottlenecked by the data and metrics used for a particular task. For example, summarization …

Offline reinforcement learning with fisher divergence critic regularization

I Kostrikov, R Fergus, J Tompson… - … on Machine Learning, 2021 - proceedings.mlr.press
Many modern approaches to offline Reinforcement Learning (RL) utilize behavior
regularization, typically augmenting a model-free actor critic algorithm with a penalty …

Fine-tuning language models from human preferences

DM Ziegler, N Stiennon, J Wu, TB Brown… - arXiv preprint arXiv …, 2019 - arxiv.org
Reward learning enables the application of reinforcement learning (RL) to tasks where
reward is defined by human judgment, building a model of reward by asking humans …