Waitgpt: Monitoring and steering conversational llm agent in data analysis with on-the-fly code visualization

L Xie, C Zheng, H Xia, H Qu, C Zhu-Tian - Proceedings of the 37th …, 2024 - dl.acm.org
Large language models (LLMs) support data analysis through conversational user
interfaces, as exemplified in OpenAI's ChatGPT (formally known as Advanced Data Analysis …

Relational composition in neural networks: a survey and call to action

M Wattenberg, FB Viégas - arXiv preprint arXiv:2407.14662, 2024 - arxiv.org
Many neural nets appear to represent data as linear combinations of" feature vectors."
Algorithms for discovering these vectors have seen impressive recent success. However, we …

Stereotype or personalization? user identity biases chatbot recommendations

A Kantharuban, J Milbauer, E Strubell… - arXiv preprint arXiv …, 2024 - arxiv.org
We demonstrate that when people use large language models (LLMs) to generate
recommendations, the LLMs produce responses that reflect both what the user wants and …

Comparing Criteria Development Across Domain Experts, Lay Users, and Models in Large Language Model Evaluation

A Szymanski, SA Gebreegziabher, O Anuyah… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) are increasingly utilized for domain-specific tasks, yet
integrating domain expertise into evaluating their outputs remains challenging. A common …

LatentQA: Teaching LLMs to Decode Activations Into Natural Language

A Pan, L Chen, J Steinhardt - arXiv preprint arXiv:2412.08686, 2024 - arxiv.org
Interpretability methods seek to understand language model representations, yet the outputs
of most such methods--circuits, vectors, scalars--are not immediately human-interpretable. In …

ChatGPT Doesn't Trust Chargers Fans: Guardrail Sensitivity in Context

VR Li, Y Chen, N Saphra - arXiv preprint arXiv:2407.06866, 2024 - arxiv.org
While the biases of language models in production are extensively documented, the biases
of their guardrails have been neglected. This paper studies how contextual information …

Proactive Agents for Multi-Turn Text-to-Image Generation Under Uncertainty

M Hahn, W Zeng, N Kannen, R Galt, K Badola… - arXiv preprint arXiv …, 2024 - arxiv.org
User prompts for generative AI models are often underspecified, leading to sub-optimal
responses. This problem is particularly evident in text-to-image (T2I) generation, where …

GenUP: Generative User Profilers as In-Context Learners for Next POI Recommender Systems

W Wongso, H Xue, FD Salim - arXiv preprint arXiv:2410.20643, 2024 - arxiv.org
Traditional POI recommendation systems often lack transparency, interpretability, and
scrutability due to their reliance on dense vector-based user embeddings. Furthermore, the …

Enhancing Customer Support in the Telecommunications Industry through AI-Driven Chatbots: A Telecom-Specific Approach

DT Thoutam, TS Jalasri - 2024 - diva-portal.org
The telecommunications industry is increasingly reliant on human agents to handle complex
customer support inquiries. With the rapid evolution of technologies like 5G and IoT …

Risk Sources and Risk Management Measures in Support of Standards for General-Purpose AI Systems

R Gipiškis, AS Joaquin, ZS Chin, A Regenfuß… - arXiv preprint arXiv …, 2024 - arxiv.org
There is an urgent need to identify both short and long-term risks from newly emerging types
of Artificial Intelligence (AI), as well as available risk management measures. In response …