A review of the explainability and safety of conversational agents for mental health to identify avenues for improvement

S Sarkar, M Gaur, LK Chen, M Garg… - Frontiers in Artificial …, 2023 - frontiersin.org
Virtual Mental Health Assistants (VMHAs) continuously evolve to support the overloaded
global healthcare system, which receives approximately 60 million primary care visits and 6 …

Reframing human-AI collaboration for generating free-text explanations

S Wiegreffe, J Hessel, S Swayamdipta, M Riedl… - arXiv preprint arXiv …, 2021 - arxiv.org
Large language models are increasingly capable of generating fluent-appearing text with
relatively little task-specific supervision. But can these models accurately explain …

Explanations from large language models make small reasoners better

S Li, J Chen, Y Shen, Z Chen, X Zhang, Z Li… - arXiv preprint arXiv …, 2022 - arxiv.org
Integrating free-text explanations to in-context learning of large language models (LLM) is
shown to elicit strong reasoning capabilities along with reasonable explanations. In this …

Measuring association between labels and free-text rationales

S Wiegreffe, A Marasović, NA Smith - arXiv preprint arXiv:2010.12762, 2020 - arxiv.org
In interpretable NLP, we require faithful rationales that reflect the model's decision-making
process for an explained instance. While prior work focuses on extractive rationales (a …

Local interpretations for explainable natural language processing: A survey

S Luo, H Ivison, SC Han, J Poon - ACM Computing Surveys, 2024 - dl.acm.org
As the use of deep learning techniques has grown across various fields over the past
decade, complaints about the opaqueness of the black-box models have increased …

Assessing the quality of student-generated short answer questions using GPT-3

S Moore, HA Nguyen, N Bier, T Domadia… - European conference on …, 2022 - Springer
Generating short answer questions is a popular form of learnersourcing with benefits for
both the students' higher-order thinking and the instructors' collection of assessment items …

Towards faithful model explanation in nlp: A survey

Q Lyu, M Apidianaki, C Callison-Burch - Computational Linguistics, 2024 - direct.mit.edu
End-to-end neural Natural Language Processing (NLP) models are notoriously difficult to
understand. This has given rise to numerous efforts towards model explainability in recent …

Receval: Evaluating reasoning chains via correctness and informativeness

A Prasad, S Saha, X Zhou, M Bansal - arXiv preprint arXiv:2304.10703, 2023 - arxiv.org
Multi-step reasoning ability is fundamental to many natural language tasks, yet it is unclear
what constitutes a good reasoning chain and how to evaluate them. Most existing methods …

Chain of explanation: New prompting method to generate quality natural language explanation for implicit hate speech

F Huang, H Kwak, J An - Companion Proceedings of the ACM Web …, 2023 - dl.acm.org
Recent studies have exploited advanced generative language models to generate Natural
Language Explanations (NLE) for why a certain text could be hateful. We propose the Chain …

How to do human evaluation: A brief introduction to user studies in NLP

H Schuff, L Vanderlyn, H Adel, NT Vu - Natural Language …, 2023 - cambridge.org
Many research topics in natural language processing (NLP), such as explanation
generation, dialog modeling, or machine translation, require evaluation that goes beyond …