Lpf: A language-prior feedback objective function for de-biased visual question answering

Z Liang, H Hu, J Zhu - Proceedings of the 44th international ACM SIGIR …, 2021 - dl.acm.org
Most existing Visual Question Answering (VQA) systems tend to overly rely on the language
bias and hence fail to reason from the visual clue. To address this issue, we propose a novel …

Quantifying and alleviating the language prior problem in visual question answering

Y Guo, Z Cheng, L Nie, Y Liu, Y Wang… - Proceedings of the 42nd …, 2019 - dl.acm.org
Benefiting from the advancement of computer vision, natural language processing and
information retrieval techniques, visual question answering (VQA), which aims to answer …

Debiased Visual Question Answering via the perspective of question types

T Huai, S Yang, J Zhang, J Zhao, L He - Pattern Recognition Letters, 2024 - Elsevier
Abstract Visual Question Answering (VQA) aims to answer questions according to the given
image. However, current VQA models tend to rely solely on textual information from the …

Overcoming language priors for visual question answering via loss rebalancing label and global context

R Cao, Z Li - Uncertainty in Artificial Intelligence, 2023 - proceedings.mlr.press
Despite the advances in Visual Question Answering (VQA), many VQA models currently
suffer from language priors (ie generating answers directly from questions without using …

Show, ask, attend, and answer: A strong baseline for visual question answering

V Kazemi, A Elqursh - arXiv preprint arXiv:1704.03162, 2017 - arxiv.org
This paper presents a new baseline for visual question answering task. Given an image and
a question in natural language, our model produces accurate answers according to the …

Debiased visual question answering from feature and sample perspectives

Z Wen, G Xu, M Tan, Q Wu… - Advances in Neural …, 2021 - proceedings.neurips.cc
Visual question answering (VQA) is designed to examine the visual-textual reasoning ability
of an intelligent agent. However, recent observations show that many VQA models may only …

Overcoming language priors with self-supervised learning for visual question answering

X Zhu, Z Mao, C Liu, P Zhang, B Wang… - arXiv preprint arXiv …, 2020 - arxiv.org
Most Visual Question Answering (VQA) models suffer from the language prior problem,
which is caused by inherent data biases. Specifically, VQA models tend to answer questions …

Rubi: Reducing unimodal biases for visual question answering

R Cadene, C Dancette, M Cord… - Advances in neural …, 2019 - proceedings.neurips.cc
Abstract Visual Question Answering (VQA) is the task of answering questions about an
image. Some VQA models often exploit unimodal biases to provide the correct answer …

Overcoming language priors via shuffling language bias for robust visual question answering

J Zhao, Z Yu, X Zhang, Y Yang - IEEE Access, 2023 - ieeexplore.ieee.org
Recent research has revealed the notorious language prior problem in visual question
answering (VQA) tasks based on visual-textual interaction, which indicates that well …

From superficial to deep: Language bias driven curriculum learning for visual question answering

M Lao, Y Guo, Y Liu, W Chen, N Pu… - Proceedings of the 29th …, 2021 - dl.acm.org
Most Visual Question Answering (VQA) models are faced with language bias when learning
to answer a given question, thereby failing to understand multimodal knowledge …