Toxicity Detection is NOT all you Need: Measuring the Gaps to Supporting Volunteer Content Moderators

YT Cao, LF Domingo, SA Gilbert, M Mazurek… - arXiv preprint arXiv …, 2023 - arxiv.org
Extensive efforts in automated approaches for content moderation have been focused on
developing models to identify toxic, offensive, and hateful content--with the aim of lightening …

A Critical Reflection on the Use of Toxicity Detection Algorithms in Proactive Content Moderation Systems

M Warner, A Strohmayer, M Higgs… - arXiv preprint arXiv …, 2024 - arxiv.org
Toxicity detection algorithms, originally designed with reactive content moderation in mind,
are increasingly being deployed into proactive end-user interventions to moderate content …

Multilingual content moderation: A case study on Reddit

M Ye, K Sikka, K Atwell, S Hassan, A Divakaran… - arXiv preprint arXiv …, 2023 - arxiv.org
Content moderation is the process of flagging content based on pre-defined platform rules.
There has been a growing need for AI moderators to safeguard users as well as protect the …

How We Define Harm Impacts Data Annotations: Explaining How Annotators Distinguish Hateful, Offensive, and Toxic Comments

A Schöpke-Gonzalez, S Wu, S Kumar… - arXiv preprint arXiv …, 2023 - arxiv.org
Computational social science research has made advances in machine learning and natural
language processing that support content moderators in detecting harmful content. These …

Algorithmic Arbitrariness in Content Moderation

JF Gomez, C Machado, LM Paes… - The 2024 ACM Conference …, 2024 - dl.acm.org
Machine learning (ML) is widely used to moderate online content. Despite its scalability
relative to human moderation, the use of ML introduces unique challenges to content …

The fabrics of machine moderation: Studying the technical, normative, and organizational structure of Perspective API

B Rieder, Y Skop - Big Data & Society, 2021 - journals.sagepub.com
Over recent years, the stakes and complexity of online content moderation have been
steadily raised, swelling from concerns about personal conflict in smaller communities to …

Watching the Watchers: A Comparative Fairness Audit of Cloud-based Content Moderation Services

D Hartmann, A Oueslati, D Staufer - arXiv preprint arXiv:2406.14154, 2024 - arxiv.org
Online platforms face the challenge of moderating an ever-increasing volume of content,
including harmful hate speech. In the absence of clear legal definitions and a lack of …

[HTML][HTML] Commercial versus volunteer: Comparing user perceptions of toxicity and transparency in content moderation across social media platforms

CL Cook, A Patel, DY Wohn - Frontiers in Human Dynamics, 2021 - frontiersin.org
Content moderation is a critical service performed by a variety of people on social media,
protecting users from offensive or harmful content by reviewing and removing either the …

Explainability and Hate Speech: Structured Explanations Make Social Media Moderators Faster

A Calabrese, L Neves, N Shah, MW Bos… - arXiv preprint arXiv …, 2024 - arxiv.org
Content moderators play a key role in keeping the conversation on social media healthy.
While the high volume of content they need to judge represents a bottleneck to the …

Common sense or censorship: How algorithmic moderators and message type influence perceptions of online content deletion

J Gonçalves, I Weber, GM Masullo… - new media & …, 2023 - journals.sagepub.com
Hateful content online is a concern for social media platforms, policymakers, and the public.
This has led high-profile content platforms, such as Facebook, to adopt algorithmic content …