X Yi, Y Li, L Wang, X Wang,
L He - arXiv preprint arXiv:2501.10639, 2025 - arxiv.org
Ensuring safety alignment has become a critical requirement for large language models
(LLMs), particularly given their widespread deployment in real-world applications. However …