Robust conversational agents against imperceptible toxicity triggers

N Mehrabi, A Beirami, F Morstatter… - arXiv preprint arXiv …, 2022 - arxiv.org
Warning: this paper contains content that maybe offensive or upsetting. Recent research in
Natural Language Processing (NLP) has advanced the development of various toxicity …

Response generation in multi-modal dialogues with split pre-generation and cross-modal contrasting

L Li, D Zhang, S Zhu, S Li, G Zhou - Information Processing & Management, 2024 - Elsevier
Due to the natural multi-modal occurrence format (text, audio, vision) of the dialogues,
textual response generation in dialogues should rely on the multi-modal contexts beyond …

Explaining large language model-based neural semantic parsers (student abstract)

D Rai, Y Zhou, B Wang, Z Yao - … of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
While large language models (LLMs) have demonstrated strong capability in structured
prediction tasks such as semantic parsing, few amounts of research have explored the …

A systematic review of toxicity in large language models: Definitions, datasets, detectors, detoxification methods and challenges

G Villate-Castillo, JDS Lorente, BS Urquijo - 2024 - researchsquare.com
The emergence of the transformer architecture has ushered in a new era of possibilities,
showcasing remarkable capabilities in generative tasks exemplified by models like GPT4o …

Towards Safety and Helpfulness Balanced Responses via Controllable Large Language Models

YL Tuan, X Chen, EM Smith, L Martin, S Batra… - arXiv preprint arXiv …, 2024 - arxiv.org
As large language models (LLMs) become easily accessible nowadays, the trade-off
between safety and helpfulness can significantly impact user experience. A model that …

Learning towards selective data augmentation for dialogue generation

X Chen, M Li, J Zhang, X Xia, C Wei, J Cui… - Proceedings of the …, 2023 - ojs.aaai.org
As it is cumbersome and expensive to acquire a huge amount of data for training neural
dialog models, data augmentation is proposed to effectively utilize existing training samples …

Attribution and Alignment: Effects of Local Context Repetition on Utterance Production and Comprehension in Dialogue

A Molnar, J Jumelet, M Giulianelli, A Sinclair - arXiv preprint arXiv …, 2023 - arxiv.org
Language models are often used as the backbone of modern dialogue systems. These
models are pre-trained on large amounts of written fluent language. Repetition is typically …

DDImage: an image reduction based approach for automatically explaining black-box classifiers

M Jiang, C Tang, XY Zhang, Y Zhao, Z Ding - Empirical Software …, 2024 - Springer
Due to the prevalent application of machine learning (ML) techniques and the intrinsic black-
box nature of ML models, the need for good explanations that are sufficient and necessary …

Understanding the Effect of Algorithm Transparency of Model Explanations in Text-to-SQL Semantic Parsing

D Rai, RR Weiland, KMG Herrera, TH Shaw… - arXiv preprint arXiv …, 2024 - arxiv.org
Explaining the decisions of AI has become vital for fostering appropriate user trust in these
systems. This paper investigates explanations for a structured prediction task called``text-to …

An Investigation of Neuron Activation as a Unified Lens to Explain Chain-of-Thought Eliciting Arithmetic Reasoning of LLMs

D Rai, Z Yao - arXiv preprint arXiv:2406.12288, 2024 - arxiv.org
Large language models (LLMs) have shown strong arithmetic reasoning capabilities when
prompted with Chain-of-Thought (CoT) prompts. However, we have only a limited …