Z Yuan, H Yuan, C Tan, W Wang, S Huang… - arXiv e …, 2023 - ui.adsabs.harvard.edu
Abstract Reinforcement Learning from Human Feedback (RLHF) facilitates the alignment of
large language models with human preferences, significantly enhancing the quality of …