[PDF][PDF] Survey on sociodemographic bias in natural language processing

V Gupta, PN Venkit, S Wilson… - arXiv preprint arXiv …, 2023 - researchgate.net
Deep neural networks often learn unintended bias during training, which might have harmful
effects when deployed in realworld settings. This work surveys 214 papers related to …

The bias amplification paradox in text-to-image generation

P Seshadri, S Singh, Y Elazar - arXiv preprint arXiv:2308.00755, 2023 - arxiv.org
Bias amplification is a phenomenon in which models increase imbalances present in the
training data. In this paper, we study bias amplification in the text-to-image domain using …

Robust Pronoun Fidelity with English LLMs: Are they Reasoning, Repeating, or Just Biased?

V Gautam, E Bingert, D Zhu, A Lauscher… - Transactions of the …, 2024 - direct.mit.edu
Robust, faithful, and harm-free pronoun use for individuals is an important goal for language
model development as their use increases, but prior work tends to study only one or two of …

Detectors for safe and reliable llms: Implementations, uses, and limitations

S Achintalwar, AA Garcia, A Anaby-Tavor… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) are susceptible to a variety of risks, from non-faithful output
to biased and toxic generations. Due to several limiting factors surrounding LLMs (training …

Bias and fairness in large language models: A survey

IO Gallegos, RA Rossi, J Barrow, MM Tanjim… - Computational …, 2024 - direct.mit.edu
Rapid advancements of large language models (LLMs) have enabled the processing,
understanding, and generation of human-like text, with increasing integration into systems …

ViSAGe: A global-scale analysis of visual stereotypes in text-to-image generation

A Jha, V Prabhakaran, R Denton, S Laszlo… - Proceedings of the …, 2024 - aclanthology.org
Recent studies have shown that Text-to-Image (T2I) model generations can reflect social
stereotypes present in the real world. However, existing approaches for evaluating …

Socialstigmaqa: A benchmark to uncover stigma amplification in generative language models

M Nagireddy, L Chiazor, M Singh… - Proceedings of the AAAI …, 2024 - ojs.aaai.org
Current datasets for unwanted social bias auditing are limited to studying protected
demographic features such as race and gender. In this work, we introduce a comprehensive …

[PDF][PDF] Metrics for what, metrics for whom: assessing actionability of bias evaluation metrics in NLP

P Delobelle, G Attanasio, D Nozza… - Proceedings of the …, 2024 - iris.unibocconi.it
This paper introduces the concept of actionability in the context of bias measures in natural
language processing (NLP). We define actionability as the degree to which a …

Are Models Biased on Text without Gender-related Language?

CG Belém, P Seshadri, Y Razeghi, S Singh - arXiv preprint arXiv …, 2024 - arxiv.org
Gender bias research has been pivotal in revealing undesirable behaviors in large
language models, exposing serious gender stereotypes associated with occupations, and …

Sociodemographic bias in language models: A survey and forward path

V Gupta, PN Venkit, S Wilson… - Proceedings of the 5th …, 2024 - aclanthology.org
Sociodemographic bias in language models (LMs) has the potential for harm when
deployed in real-world settings. This paper presents a comprehensive survey of the past …