J Ji,
D Hong, B Zhang, B Chen,
J Dai, B Zheng… - arXiv preprint arXiv …, 2024 - arxiv.org
In this work, we introduce the PKU-SafeRLHF dataset, designed to promote research on
safety alignment in large language models (LLMs). As a sibling project to SafeRLHF and …