C-rag: Certified generation risks for retrieval-augmented language models

M Kang, NM Gürel, N Yu, D Song, B Li - arXiv preprint arXiv:2402.03181, 2024 - arxiv.org
Despite the impressive capabilities of large language models (LLMs) across diverse
applications, they still suffer from trustworthiness issues, such as hallucinations and …

-Guard: Robust Reasoning Enabled LLM Guardrail via Knowledge-Enhanced Logical Reasoning

M Kang, B Li - arXiv preprint arXiv:2407.05557, 2024 - arxiv.org
As LLMs become increasingly prevalent across various applications, it is critical to establish
safety guardrails to moderate input/output content of LLMs. Existing guardrail models treat …

Legend: Leveraging Representation Engineering to Annotate Safety Margin for Preference Datasets

D Feng, B Qin, C Huang, Y Huang, Z Zhang… - arXiv preprint arXiv …, 2024 - arxiv.org
The success of the reward model in distinguishing between responses with subtle safety
differences depends critically on the high-quality preference dataset, which should capture …

A Survey of Generative Techniques for Spatial-Temporal Data Mining

Q Zhang, H Wang, C Long, L Su, X He, J Chang… - arXiv preprint arXiv …, 2024 - arxiv.org
This paper focuses on the integration of generative techniques into spatial-temporal data
mining, considering the significant growth and diverse nature of spatial-temporal data. With …

Adaptive Token Biaser: Knowledge Editing via Biasing Key Entities

B Bi, S Liu, Y Wang, L Mei, H Gao, Y Xu… - arXiv preprint arXiv …, 2024 - arxiv.org
The parametric knowledge memorized by large language models (LLMs) becomes outdated
quickly. In-context editing (ICE) is currently the most effective method for updating the …