Towards Safety and Helpfulness Balanced Responses via Controllable Large Language Models

YL Tuan, X Chen, EM Smith, L Martin, S Batra… - arXiv preprint arXiv …, 2024 - arxiv.org
As large language models (LLMs) become easily accessible nowadays, the trade-off
between safety and helpfulness can significantly impact user experience. A model that …

A gradient analysis framework for rewarding good and penalizing bad examples in language models

YL Tuan, WY Wang - arXiv preprint arXiv:2408.16751, 2024 - arxiv.org
Beyond maximum likelihood estimation (MLE), the standard objective of a language model
(LM) that optimizes good examples probabilities, many studies have explored ways that also …

Achieving Human-like Chatbots from Reasoning and Optimization Perspectives

YL Tuan - 2024 - search.proquest.com
Human-like chatbots–machines that can act as humans to chat about any topic–need to
listen, understand, reason, respond, and interactively learn to optimize the whole process …

Understanding and Improving Language Models Through a Data-Centric Lens

A Albalak - 2024 - search.proquest.com
Training data has played a major role in the rise of large deep learning models. In particular,
the scale and diversity of training data has led to incredible new capabilities in large …