Toxicity Detection is NOT all you Need: Measuring the Gaps to Supporting Volunteer Content Moderators
YT Cao, LF Domingo, SA Gilbert, M Mazurek… - arXiv preprint arXiv …, 2023 - arxiv.org
Extensive efforts in automated approaches for content moderation have been focused on
developing models to identify toxic, offensive, and hateful content--with the aim of lightening …
developing models to identify toxic, offensive, and hateful content--with the aim of lightening …
A Critical Reflection on the Use of Toxicity Detection Algorithms in Proactive Content Moderation Systems
Toxicity detection algorithms, originally designed with reactive content moderation in mind,
are increasingly being deployed into proactive end-user interventions to moderate content …
are increasingly being deployed into proactive end-user interventions to moderate content …
Multilingual content moderation: A case study on Reddit
Content moderation is the process of flagging content based on pre-defined platform rules.
There has been a growing need for AI moderators to safeguard users as well as protect the …
There has been a growing need for AI moderators to safeguard users as well as protect the …
How We Define Harm Impacts Data Annotations: Explaining How Annotators Distinguish Hateful, Offensive, and Toxic Comments
Computational social science research has made advances in machine learning and natural
language processing that support content moderators in detecting harmful content. These …
language processing that support content moderators in detecting harmful content. These …
Algorithmic Arbitrariness in Content Moderation
Machine learning (ML) is widely used to moderate online content. Despite its scalability
relative to human moderation, the use of ML introduces unique challenges to content …
relative to human moderation, the use of ML introduces unique challenges to content …
The fabrics of machine moderation: Studying the technical, normative, and organizational structure of Perspective API
Over recent years, the stakes and complexity of online content moderation have been
steadily raised, swelling from concerns about personal conflict in smaller communities to …
steadily raised, swelling from concerns about personal conflict in smaller communities to …
Watching the Watchers: A Comparative Fairness Audit of Cloud-based Content Moderation Services
D Hartmann, A Oueslati, D Staufer - arXiv preprint arXiv:2406.14154, 2024 - arxiv.org
Online platforms face the challenge of moderating an ever-increasing volume of content,
including harmful hate speech. In the absence of clear legal definitions and a lack of …
including harmful hate speech. In the absence of clear legal definitions and a lack of …
[HTML][HTML] Commercial versus volunteer: Comparing user perceptions of toxicity and transparency in content moderation across social media platforms
Content moderation is a critical service performed by a variety of people on social media,
protecting users from offensive or harmful content by reviewing and removing either the …
protecting users from offensive or harmful content by reviewing and removing either the …
Explainability and Hate Speech: Structured Explanations Make Social Media Moderators Faster
Content moderators play a key role in keeping the conversation on social media healthy.
While the high volume of content they need to judge represents a bottleneck to the …
While the high volume of content they need to judge represents a bottleneck to the …
Common sense or censorship: How algorithmic moderators and message type influence perceptions of online content deletion
Hateful content online is a concern for social media platforms, policymakers, and the public.
This has led high-profile content platforms, such as Facebook, to adopt algorithmic content …
This has led high-profile content platforms, such as Facebook, to adopt algorithmic content …