Towards generalisable hate speech detection: a review on obstacles and solutions
Hate speech is one type of harmful online content which directly attacks or promotes hate
towards a group or an individual member based on their actual or perceived aspects of …
towards a group or an individual member based on their actual or perceived aspects of …
Handling bias in toxic speech detection: A survey
Detecting online toxicity has always been a challenge due to its inherent subjectivity. Factors
such as the context, geography, socio-political climate, and background of the producers …
such as the context, geography, socio-political climate, and background of the producers …
Language (technology) is power: A critical survey of" bias" in nlp
We survey 146 papers analyzing" bias" in NLP systems, finding that their motivations are
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …
often vague, inconsistent, and lacking in normative reasoning, despite the fact that …
Trustworthy ai: A computational perspective
In the past few decades, artificial intelligence (AI) technology has experienced swift
developments, changing everyone's daily life and profoundly altering the course of human …
developments, changing everyone's daily life and profoundly altering the course of human …
Is your toxicity my toxicity? exploring the impact of rater identity on toxicity annotation
N Goyal, ID Kivlichan, R Rosen… - Proceedings of the ACM …, 2022 - dl.acm.org
Machine learning models are commonly used to detect toxicity in online conversations.
These models are trained on datasets annotated by human raters. We explore how raters' …
These models are trained on datasets annotated by human raters. We explore how raters' …
A survey of race, racism, and anti-racism in NLP
Despite inextricable ties between race and language, little work has considered race in NLP
research and development. In this work, we survey 79 papers from the ACL anthology that …
research and development. In this work, we survey 79 papers from the ACL anthology that …
Quantifying social biases in NLP: A generalization and empirical comparison of extrinsic fairness metrics
Measuring bias is key for better understanding and addressing unfairness in NLP/ML
models. This is often done via fairness metrics, which quantify the differences in a model's …
models. This is often done via fairness metrics, which quantify the differences in a model's …
Benchmarking intersectional biases in NLP
There has been a recent wave of work assessing the fairness of machine learning models in
general, and more specifically, on natural language processing (NLP) models built using …
general, and more specifically, on natural language processing (NLP) models built using …
[HTML][HTML] Offensive, aggressive, and hate speech analysis: From data-centric to human-centered approach
Abstract Analysis of subjective texts like offensive content or hate speech is a great
challenge, especially regarding annotation process. Most of current annotation procedures …
challenge, especially regarding annotation process. Most of current annotation procedures …
Deep learning models for multilingual hate speech detection
Hate speech detection is a challenging problem with most of the datasets available in only
one language: English. In this paper, we conduct a large scale analysis of multilingual hate …
one language: English. In this paper, we conduct a large scale analysis of multilingual hate …