Auto-debias: Debiasing masked language models with automated biased prompts

Y Guo, Y Yang, A Abbasi - … of the 60th Annual Meeting of the …, 2022 - aclanthology.org
Human-like biases and undesired social stereotypes exist in large pretrained language
models. Given the wide adoption of these models in real-world applications, mitigating such …

Nbias: A natural language processing framework for BIAS identification in text

S Raza, M Garg, DJ Reji, SR Bashir, C Ding - Expert Systems with …, 2024 - Elsevier
Bias in textual data can lead to skewed interpretations and outcomes when the data is used.
These biases could perpetuate stereotypes, discrimination, or other forms of unfair …

On measuring social biases in prompt-based multi-task learning

AF Akyürek, S Paik, MY Kocyigit, S Akbiyik… - arXiv preprint arXiv …, 2022 - arxiv.org
Large language models trained on a mixture of NLP tasks that are converted into a text-to-
text format using prompts, can generalize into novel forms of language and handle novel …

Fairframe: a fairness framework for bias detection and mitigation in news

D Sallami, E Aïmeur - AI and Ethics, 2024 - Springer
In the realm of digital information, ensuring the fairness and neutrality of textual content,
especially news, is paramount. This paper introduces FairFrame, a novel framework …

Labels Need Prompts Too: Mask Matching for Natural Language Understanding Tasks

B Li, W Ye, Q Wang, W Zhao, S Zhang - Proceedings of the AAAI …, 2024 - ojs.aaai.org
Textual label names (descriptions) are typically semantically rich in many natural language
understanding (NLU) tasks. In this paper, we incorporate the prompting methodology, which …

Less learn shortcut: Analyzing and mitigating learning of spurious feature-label correlation

Y Du, J Yan, Y Chen, J Liu, S Zhao, Q She… - arXiv preprint arXiv …, 2022 - arxiv.org
Recent research has revealed that deep neural networks often take dataset biases as a
shortcut to make decisions rather than understand tasks, leading to failures in real-world …

Projective Methods for Mitigating Gender Bias in Pre-trained Language Models

H Dawkins, I Nejadgholi, D Gillis, J McCuaig - arXiv preprint arXiv …, 2024 - arxiv.org
Mitigation of gender bias in NLP has a long history tied to debiasing static word
embeddings. More recently, attention has shifted to debiasing pre-trained language models …