Glam: Efficient scaling of language models with mixture-of-experts

N Du, Y Huang, AM Dai, S Tong… - International …, 2022 - proceedings.mlr.press
Scaling language models with more data, compute and parameters has driven significant
progress in natural language processing. For example, thanks to scaling, GPT-3 was able to …

" I'm sorry to hear that": Finding New Biases in Language Models with a Holistic Descriptor Dataset

EM Smith, M Hall, M Kambadur, E Presani… - arXiv preprint arXiv …, 2022 - arxiv.org
As language models grow in popularity, it becomes increasingly important to clearly
measure all possible markers of demographic identity in order to avoid perpetuating existing …

ROBBIE: Robust bias evaluation of large generative language models

D Esiobu, X Tan, S Hosseini, M Ung, Y Zhang… - arXiv preprint arXiv …, 2023 - arxiv.org
As generative large language models (LLMs) grow more performant and prevalent, we must
develop comprehensive enough tools to measure and improve their fairness. Different …

Unmasking nationality bias: A study of human perception of nationalities in ai-generated articles

P Narayanan Venkit, S Gautam… - Proceedings of the …, 2023 - dl.acm.org
We investigate the potential for nationality biases in natural language processing (NLP)
models using human evaluation methods. Biased NLP models can perpetuate stereotypes …

Theory-grounded measurement of US social stereotypes in English language models

YT Cao, A Sotnikova, H Daumé III, R Rudinger… - arXiv preprint arXiv …, 2022 - arxiv.org
NLP models trained on text have been shown to reproduce human stereotypes, which can
magnify harms to marginalized groups when systems are deployed at scale. We adapt the …

Exploring social biases of large language models in a college artificial intelligence course

S Kolisko, CJ Anderson - Proceedings of the AAAI Conference on …, 2023 - ojs.aaai.org
Large neural network-based language models play an increasingly important role in
contemporary AI. Although these models demonstrate sophisticated text generation …

Which examples should be multiply annotated? active learning when annotators may disagree

C Baumler, A Sotnikova… - Findings of the Association …, 2023 - aclanthology.org
Linguistic annotations, especially for controversial topics like hate speech detection, are
frequently contested due to annotator backgrounds and positionalities. In such situations …

SODAPOP: Open-ended discovery of social biases in social commonsense reasoning models

H An, Z Li, J Zhao, R Rudinger - arXiv preprint arXiv:2210.07269, 2022 - arxiv.org
A common limitation of diagnostic tests for detecting social biases in NLP models is that they
may only detect stereotypic associations that are pre-specified by the designer of the test …

Measuring machine learning harms from stereotypes: requires understanding who is being harmed by which errors in what ways

A Wang, X Bai, S Barocas, SL Blodgett - arXiv preprint arXiv:2402.04420, 2024 - arxiv.org
As machine learning applications proliferate, we need an understanding of their potential for
harm. However, current fairness metrics are rarely grounded in human psychological …

[PDF][PDF] Measuring stereotype harm from machine learning errors requires understanding who is being harmed by which errors in what ways

A Wang, X Bai, S Barocas… - ACM Conference on …, 2023 - conference2023.eaamo.org
Fig. 1. We distinguish between types of machine learning errors which are likely to have
different harms. The first split is between errors made on labels which are associated with …