Fairness in deep learning: A survey on vision and language research

O Parraga, MD More, CM Oliveira, NS Gavenski… - ACM Computing …, 2023 - dl.acm.org
Despite being responsible for state-of-the-art results in several computer vision and natural
language processing tasks, neural networks have faced harsh criticism due to some of their …

Debiasing methods for fairer neural models in vision and language research: A survey

O Parraga, MD More, CM Oliveira, NS Gavenski… - arXiv preprint arXiv …, 2022 - arxiv.org
Despite being responsible for state-of-the-art results in several computer vision and natural
language processing tasks, neural networks have faced harsh criticism due to some of their …

Rethinking data augmentation for robust visual question answering

L Chen, Y Zheng, J Xiao - European conference on computer vision, 2022 - Springer
Data Augmentation (DA)—generating extra training samples beyond the original training set—
has been widely-used in today's unbiased VQA models to mitigate language biases. Current …

Be flexible! learn to debias by sampling and prompting for robust visual question answering

J Liu, CF Fan, F Zhou, H Xu - Information Processing & Management, 2023 - Elsevier
Recent studies point out that VQA models tend to rely on the language prior in the training
data to answer the questions, which prevents the VQA model from generalization on the out …

Question-conditioned debiasing with focal visual context fusion for visual question answering

J Liu, GX Wang, CF Fan, F Zhou, HJ Xu - Knowledge-Based Systems, 2023 - Elsevier
Abstract Existing Visual Question Answering models suffer from the language prior, where
the answers provided by the models overly rely on the correlations between questions and …

Robust visual question answering via polarity enhancement and contrast

D Peng, Z Li - Neural Networks, 2024 - Elsevier
Abstract The Visual Question Answering (VQA) task is an important research direction in the
field of artificial intelligence, which requires a model that can simultaneously understand …

Digging out discrimination information from generated samples for robust visual question answering

Z Wen, Y Wang, M Tan, Q Wu, Q Wu - Findings of the Association …, 2023 - aclanthology.org
Abstract Visual Question Answering (VQA) aims to answer a textual question based on a
given image. Nevertheless, recent studies have shown that VQA models tend to capture the …

Improving visual question answering models through robustness analysis and in-context learning with a chain of basic questions

JH Huang, M Alfadly, B Ghanem, M Worring - arXiv preprint arXiv …, 2023 - arxiv.org
Deep neural networks have been critical in the task of Visual Question Answering (VQA),
with research traditionally focused on improving model accuracy. Recently, however, there …

Towards Deconfounded Visual Question Answering via Dual-causal Intervention

D Peng, W Wei - Proceedings of the 33rd ACM International Conference …, 2024 - dl.acm.org
The Visual Question Answering (VQA) task has recently become notorious because models
are prone to predicting well-educated" guesses" as answers rather than deriving them …

Overcoming language priors via shuffling language bias for robust visual question answering

J Zhao, Z Yu, X Zhang, Y Yang - IEEE Access, 2023 - ieeexplore.ieee.org
Recent research has revealed the notorious language prior problem in visual question
answering (VQA) tasks based on visual-textual interaction, which indicates that well …