Handling bias in toxic speech detection: A survey

T Garg, S Masud, T Suresh, T Chakraborty - ACM Computing Surveys, 2023 - dl.acm.org
Detecting online toxicity has always been a challenge due to its inherent subjectivity. Factors
such as the context, geography, socio-political climate, and background of the producers …

[HTML][HTML] Leaving traces behind: Using social media digital trace data to study adolescent wellbeing

M Sultan, C Scholz, W van den Bos - Computers in Human Behavior …, 2023 - Elsevier
Adolescents spend a significant amount of time on social media and there is a great public
worry, from parents to policy makers, about the effect of social media on healthy …

The risk of racial bias in hate speech detection

M Sap, D Card, S Gabriel, Y Choi… - Proceedings of the 57th …, 2019 - aclanthology.org
We investigate how annotators' insensitivity to differences in dialect can lead to racial bias in
automatic hate speech detection models, potentially amplifying harm against minority …

A survey of race, racism, and anti-racism in NLP

A Field, SL Blodgett, Z Waseem, Y Tsvetkov - arXiv preprint arXiv …, 2021 - arxiv.org
Despite inextricable ties between race and language, little work has considered race in NLP
research and development. In this work, we survey 79 papers from the ACL anthology that …

[图书][B] Challenges in automated debiasing for toxic language detection

X Zhou - 2020 - search.proquest.com
Biased associations have been a challenge in the development of classifiers for detecting
toxic language, hindering both fairness and accuracy. As potential solutions, we investigate …

On measures of biases and harms in NLP

S Dev, E Sheng, J Zhao, A Amstutz, J Sun… - arXiv preprint arXiv …, 2021 - arxiv.org
Recent studies show that Natural Language Processing (NLP) technologies propagate
societal biases about demographic groups associated with attributes such as gender, race …

Fairness in language models beyond English: Gaps and challenges

K Ramesh, S Sitaram, M Choudhury - arXiv preprint arXiv:2302.12578, 2023 - arxiv.org
With language models becoming increasingly ubiquitous, it has become essential to
address their inequitable treatment of diverse demographic groups and factors. Most …

Multilingual twitter corpus and baselines for evaluating demographic bias in hate speech recognition

X Huang, L Xing, F Dernoncourt, MJ Paul - arXiv preprint arXiv …, 2020 - arxiv.org
Existing research on fairness evaluation of document classification models mainly uses
synthetic monolingual data without ground truth for author demographic attributes. In this …

Algorithmic fairness datasets: the story so far

A Fabris, S Messina, G Silvello, GA Susto - Data Mining and Knowledge …, 2022 - Springer
Data-driven algorithms are studied and deployed in diverse domains to support critical
decisions, directly impacting people's well-being. As a result, a growing community of …

Exploring the role of grammar and word choice in bias toward african american english (aae) in hate speech classification

C Harris, M Halevy, A Howard, A Bruckman… - Proceedings of the 2022 …, 2022 - dl.acm.org
Language usage on social media varies widely even within the context of American English.
Despite this, the majority of natural language processing systems are trained only on …