Fairness in deep learning: A survey on vision and language research

O Parraga, MD More, CM Oliveira, NS Gavenski… - ACM Computing …, 2023 - dl.acm.org
Despite being responsible for state-of-the-art results in several computer vision and natural
language processing tasks, neural networks have faced harsh criticism due to some of their …

Harnessing the power of llms in practice: A survey on chatgpt and beyond

J Yang, H Jin, R Tang, X Han, Q Feng, H Jiang… - ACM Transactions on …, 2024 - dl.acm.org
This article presents a comprehensive and practical guide for practitioners and end-users
working with Large Language Models (LLMs) in their downstream Natural Language …

Deep image captioning: A review of methods, trends and future challenges

L Xu, Q Tang, J Lv, B Zheng, X Zeng, W Li - Neurocomputing, 2023 - Elsevier
Image captioning, also called report generation in medical field, aims to describe visual
content of images in human language, which requires to model semantic relationship …

Dall-eval: Probing the reasoning skills and social biases of text-to-image generation models

J Cho, A Zala, M Bansal - Proceedings of the IEEE/CVF …, 2023 - openaccess.thecvf.com
Recently, DALL-E, a multimodal transformer language model, and its variants including
diffusion models have shown high-quality text-to-image generation capabilities. However …

Understanding and evaluating racial biases in image captioning

D Zhao, A Wang… - Proceedings of the IEEE …, 2021 - openaccess.thecvf.com
Image captioning is an important task for benchmarking visual reasoning and for enabling
accessibility for people with vision impairments. However, as in many machine learning …

Uncurated image-text datasets: Shedding light on demographic bias

N Garcia, Y Hirota, Y Wu… - Proceedings of the …, 2023 - openaccess.thecvf.com
The increasing tendency to collect large and uncurated datasets to train vision-and-
language models has raised concerns about fair representations. It is known that even small …

A case-based interpretable deep learning model for classification of mass lesions in digital mammography

AJ Barnett, FR Schwartz, C Tao, C Chen… - Nature Machine …, 2021 - nature.com
Interpretability in machine learning models is important in high-stakes decisions such as
whether to order a biopsy based on a mammographic exam. Mammography poses …

Quantifying societal bias amplification in image captioning

Y Hirota, Y Nakashima… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
We study societal bias amplification in image captioning. Image captioning models have
been shown to perpetuate gender and racial biases, however, metrics to measure, quantify …

Visual abductive reasoning

C Liang, W Wang, T Zhou… - Proceedings of the IEEE …, 2022 - openaccess.thecvf.com
Abductive reasoning seeks the likeliest possible explanation for partial observations.
Although abduction is frequently employed in human daily reasoning, it is rarely explored in …

Large language models can be lazy learners: Analyze shortcuts in in-context learning

R Tang, D Kong, L Huang, H Xue - arXiv preprint arXiv:2305.17256, 2023 - arxiv.org
Large language models (LLMs) have recently shown great potential for in-context learning,
where LLMs learn a new task simply by conditioning on a few input-label pairs (prompts) …