Getting aligned on representational alignment

I Sucholutsky, L Muttenthaler, A Weller, A Peng… - arXiv preprint arXiv …, 2023 - arxiv.org
Biological and artificial information processing systems form representations that they can
use to categorize, reason, plan, navigate, and make decisions. How can we measure the …

Human uncertainty in concept-based ai systems

KM Collins, M Barker, M Espinosa Zarlenga… - Proceedings of the …, 2023 - dl.acm.org
Placing a human in the loop may help abate the risks of deploying AI systems in safety-
critical settings (eg, a clinician working with a medical AI system). However, mitigating risks …

Binary classification with confidence difference

W Wang, L Feng, Y Jiang, G Niu… - Advances in …, 2024 - proceedings.neurips.cc
Recently, learning with soft labels has been shown to achieve better performance than
learning with hard labels in terms of model generalization, calibration, and robustness …

Judgment sieve: Reducing uncertainty in group judgments through interventions targeting ambiguity versus disagreement

QZ Chen, AX Zhang - Proceedings of the ACM on Human-Computer …, 2023 - dl.acm.org
When groups of people are tasked with making a judgment, the issue of uncertainty often
arises. Existing methods to reduce uncertainty typically focus on iteratively improving …

On the informativeness of supervision signals

I Sucholutsky, RM Battleday… - Uncertainty in …, 2023 - proceedings.mlr.press
Supervised learning typically focuses on learning transferable representations from training
examples annotated by humans. While rich annotations (like soft labels) carry more …

Ambiguous images with human judgments for robust visual event classification

K Sanders, R Kriz, A Liu… - Advances in Neural …, 2022 - proceedings.neurips.cc
Contemporary vision benchmarks predominantly consider tasks on which humans can
achieve near-perfect performance. However, humans are frequently presented with visual …

Interpreting Predictive Probabilities: Model Confidence or Human Label Variation?

J Baan, R Fernández, B Plank, W Aziz - arXiv preprint arXiv:2402.16102, 2024 - arxiv.org
With the rise of increasingly powerful and user-facing NLP systems, there is growing interest
in assessing whether they have a good representation of uncertainty by evaluating the …

The empty signifier problem: Towards clearer paradigms for operationalising" alignment" in large language models

HR Kirk, B Vidgen, P Röttger, SA Hale - arXiv preprint arXiv:2310.02457, 2023 - arxiv.org
In this paper, we address the concept of" alignment" in large language models (LLMs)
through the lens of post-structuralist socio-political theory, specifically examining its parallels …

Learning personalized decision support policies

U Bhatt, V Chen, KM Collins, P Kamalaruban… - arXiv preprint arXiv …, 2023 - arxiv.org
Individual human decision-makers may benefit from different forms of support to improve
decision outcomes. However, a key question is which form of support will lead to accurate …

Subjective crowd disagreements for subjective data: Uncovering meaningful CrowdOpinion with population-level learning

TC Weerasooriya, S Luger, S Poddar… - arXiv preprint arXiv …, 2023 - arxiv.org
Human-annotated data plays a critical role in the fairness of AI systems, including those that
deal with life-altering decisions or moderating human-created web/social media content …