Towards generalisable hate speech detection: a review on obstacles and solutions

W Yin, A Zubiaga - PeerJ Computer Science, 2021 - peerj.com
Hate speech is one type of harmful online content which directly attacks or promotes hate
towards a group or an individual member based on their actual or perceived aspects of …

Handling bias in toxic speech detection: A survey

T Garg, S Masud, T Suresh, T Chakraborty - ACM Computing Surveys, 2023 - dl.acm.org
Detecting online toxicity has always been a challenge due to its inherent subjectivity. Factors
such as the context, geography, socio-political climate, and background of the producers …

Language (technology) is power: A critical survey of" bias" in nlp

SL Blodgett, S Barocas, H Daumé III… - arXiv preprint arXiv …, 2020 - arxiv.org
We survey 146 papers analyzing" bias" in NLP systems, finding that their motivations are
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …

Trustworthy ai: A computational perspective

H Liu, Y Wang, W Fan, X Liu, Y Li, S Jain, Y Liu… - ACM Transactions on …, 2022 - dl.acm.org
In the past few decades, artificial intelligence (AI) technology has experienced swift
developments, changing everyone's daily life and profoundly altering the course of human …

Is your toxicity my toxicity? exploring the impact of rater identity on toxicity annotation

N Goyal, ID Kivlichan, R Rosen… - Proceedings of the ACM …, 2022 - dl.acm.org
Machine learning models are commonly used to detect toxicity in online conversations.
These models are trained on datasets annotated by human raters. We explore how raters' …

A survey of race, racism, and anti-racism in NLP

A Field, SL Blodgett, Z Waseem, Y Tsvetkov - arXiv preprint arXiv …, 2021 - arxiv.org
Despite inextricable ties between race and language, little work has considered race in NLP
research and development. In this work, we survey 79 papers from the ACL anthology that …

Quantifying social biases in NLP: A generalization and empirical comparison of extrinsic fairness metrics

P Czarnowska, Y Vyas, K Shah - Transactions of the Association for …, 2021 - direct.mit.edu
Measuring bias is key for better understanding and addressing unfairness in NLP/ML
models. This is often done via fairness metrics, which quantify the differences in a model's …

Benchmarking intersectional biases in NLP

JP Lalor, Y Yang, K Smith, N Forsgren… - Proceedings of the …, 2022 - aclanthology.org
There has been a recent wave of work assessing the fairness of machine learning models in
general, and more specifically, on natural language processing (NLP) models built using …

[HTML][HTML] Offensive, aggressive, and hate speech analysis: From data-centric to human-centered approach

J Kocoń, A Figas, M Gruza, D Puchalska… - Information Processing …, 2021 - Elsevier
Abstract Analysis of subjective texts like offensive content or hate speech is a great
challenge, especially regarding annotation process. Most of current annotation procedures …

Deep learning models for multilingual hate speech detection

SS Aluru, B Mathew, P Saha, A Mukherjee - arXiv preprint arXiv …, 2020 - arxiv.org
Hate speech detection is a challenging problem with most of the datasets available in only
one language: English. In this paper, we conduct a large scale analysis of multilingual hate …