Fairness testing: A comprehensive survey and analysis of trends

Z Chen, JM Zhang, M Hort, M Harman… - ACM Transactions on …, 2024 - dl.acm.org
Unfair behaviors of Machine Learning (ML) software have garnered increasing attention and
concern among software engineers. To tackle this issue, extensive research has been …

BBQ: A hand-built bias benchmark for question answering

A Parrish, A Chen, N Nangia, V Padmakumar… - arXiv preprint arXiv …, 2021 - arxiv.org
It is well documented that NLP models learn social biases, but little work has been done on
how these biases manifest in model outputs for applied tasks like question answering (QA) …

Memory-assisted prompt editing to improve GPT-3 after deployment

A Madaan, N Tandon, P Clark, Y Yang - arXiv preprint arXiv:2201.06009, 2022 - arxiv.org
Large LMs such as GPT-3 are powerful, but can commit mistakes that are obvious to
humans. For example, GPT-3 would mistakenly interpret" What word is similar to good?" to …

Theories of “gender” in nlp bias research

H Devinney, J Björklund, H Björklund - … of the 2022 ACM conference on …, 2022 - dl.acm.org
The rise of concern around Natural Language Processing (NLP) technologies containing
and perpetuating social biases has led to a rich and rapidly growing area of research …

Fewer errors, but more stereotypes? the effect of model size on gender bias

Y Tal, I Magar, R Schwartz - arXiv preprint arXiv:2206.09860, 2022 - arxiv.org
The size of pretrained models is increasing, and so is their performance on a variety of NLP
tasks. However, as their memorization capacity grows, they might pick up more social …

Evaluating gender bias of pre-trained language models in natural language inference by considering all labels

P Anantaprayoon, M Kaneko, N Okazaki - arXiv preprint arXiv:2309.09697, 2023 - arxiv.org
Discriminatory social biases, including gender biases, have been found in Pre-trained
Language Models (PLMs). In Natural Language Inference (NLI), recent bias evaluation …

How gender debiasing affects internal model representations, and why it matters

H Orgad, S Goldfarb-Tarrant, Y Belinkov - arXiv preprint arXiv:2204.06827, 2022 - arxiv.org
Common studies of gender bias in NLP focus either on extrinsic bias measured by model
performance on a downstream task or on intrinsic bias found in models' internal …

On measuring social biases in prompt-based multi-task learning

AF Akyürek, S Paik, MY Kocyigit, S Akbiyik… - arXiv preprint arXiv …, 2022 - arxiv.org
Large language models trained on a mixture of NLP tasks that are converted into a text-to-
text format using prompts, can generalize into novel forms of language and handle novel …

What social attitudes about gender does BERT encode? Leveraging insights from psycholinguistics

J Watson, B Beekhuizen… - Proceedings of the 61st …, 2023 - aclanthology.org
Much research has sought to evaluate the degree to which large language models reflect
social biases. We complement such work with an approach to elucidating the connections …

Evaluating the robustness of discrete prompts

Y Ishibashi, D Bollegala, K Sudoh… - arXiv preprint arXiv …, 2023 - arxiv.org
Discrete prompts have been used for fine-tuning Pre-trained Language Models for diverse
NLP tasks. In particular, automatic methods that generate discrete prompts from a small set …