Improving Data Augmentation for Robust Visual Question Answering with Effective Curriculum Learning

Y Zheng, Z Wang, L Chen - … of the 2024 International Conference on …, 2024 - dl.acm.org
Being widely used in learning unbiased visual question answering (VQA) models, Data
Augmentation (DA) helps mitigate language biases by generating extra training samples …

Suppressing biased samples for robust VQA

N Ouyang, Q Huang, P Li, Y Cai, B Liu… - IEEE Transactions …, 2021 - ieeexplore.ieee.org
Most existing visual question answering (VQA) models strongly rely on language bias to
answer questions, ie, they always tend to fit question-answer pairs on the train split and …

Overcoming Language Priors in Visual Question Answering with Cumulative Learning Strategy

A Mao, F Chen, Z Ma, K Lin - Available at SSRN 4740502 - papers.ssrn.com
The performance of visual question answering (VQA) has witnessed great progress over the
last few years. However, many current VQA models tend to rely on superficial linguistic …

Lpf: A language-prior feedback objective function for de-biased visual question answering

Z Liang, H Hu, J Zhu - Proceedings of the 44th international ACM SIGIR …, 2021 - dl.acm.org
Most existing Visual Question Answering (VQA) systems tend to overly rely on the language
bias and hence fail to reason from the visual clue. To address this issue, we propose a novel …

Overcoming language priors for visual question answering via loss rebalancing label and global context

R Cao, Z Li - Uncertainty in Artificial Intelligence, 2023 - proceedings.mlr.press
Despite the advances in Visual Question Answering (VQA), many VQA models currently
suffer from language priors (ie generating answers directly from questions without using …

Overcoming language priors via shuffling language bias for robust visual question answering

J Zhao, Z Yu, X Zhang, Y Yang - IEEE Access, 2023 - ieeexplore.ieee.org
Recent research has revealed the notorious language prior problem in visual question
answering (VQA) tasks based on visual-textual interaction, which indicates that well …

Towards robust visual question answering: Making the most of biased samples via contrastive learning

Q Si, Y Liu, F Meng, Z Lin, P Fu, Y Cao, W Wang… - arXiv preprint arXiv …, 2022 - arxiv.org
Models for Visual Question Answering (VQA) often rely on the spurious correlations, ie, the
language priors, that appear in the biased samples of training set, which make them brittle …

From superficial to deep: Language bias driven curriculum learning for visual question answering

M Lao, Y Guo, Y Liu, W Chen, N Pu… - Proceedings of the 29th …, 2021 - dl.acm.org
Most Visual Question Answering (VQA) models are faced with language bias when learning
to answer a given question, thereby failing to understand multimodal knowledge …

Overcoming language priors with self-supervised learning for visual question answering

X Zhu, Z Mao, C Liu, P Zhang, B Wang… - arXiv preprint arXiv …, 2020 - arxiv.org
Most Visual Question Answering (VQA) models suffer from the language prior problem,
which is caused by inherent data biases. Specifically, VQA models tend to answer questions …

HCCL: H ierarchical C ounterfactual C ontrastive L earning for Robust Visual Question Answering

D Hao, Q Wang, X Zhu, J Liu - ACM Transactions on Multimedia Computing … - dl.acm.org
Despite most state-of-the-art models having achieved amazing performance in visual
question answering (VQA), they usually utilize biases to answer the question. Recently …