Multilingual content moderation: A case study on Reddit

M Ye, K Sikka, K Atwell, S Hassan, A Divakaran… - arXiv preprint arXiv …, 2023 - arxiv.org
Content moderation is the process of flagging content based on pre-defined platform rules.
There has been a growing need for AI moderators to safeguard users as well as protect the …

Deeper attention to abusive user content moderation

J Pavlopoulos, P Malakasiotis… - Proceedings of the …, 2017 - aclanthology.org
Experimenting with a new dataset of 1.6 M user comments from a news portal and an
existing dataset of 115K Wikipedia talk page comments, we show that an RNN operating on …

The discourse of online content moderation: Investigating polarized user responses to changes in reddit's quarantine policy

Q Shen, C Rose - Proceedings of the third workshop on abusive …, 2019 - aclanthology.org
Recent concerns over abusive behavior on their platforms have pressured social media
companies to strengthen their content moderation policies. However, user opinions on these …

Beyond Trial-and-Error: Predicting User Abandonment After a Moderation Intervention

B Tessa, L Cima, A Trujillo, M Avvenuti… - arXiv preprint arXiv …, 2024 - arxiv.org
Current content moderation practices follow the\textit {trial-and-error} approach, meaning
that moderators apply sequences of interventions until they obtain the desired outcome …

One of many: Assessing user-level effects of moderation interventions on r/The_Donald

A Trujillo, S Cresci - Proceedings of the 15th ACM Web Science …, 2023 - dl.acm.org
Evaluating the effects of moderation interventions is a task of paramount importance, as it
allows assessing the success of content moderation processes. So far, intervention effects …

Watch your language: large language models and content moderation

D Kumar, Y AbuHashem, Z Durumeric - arXiv preprint arXiv:2309.14517, 2023 - arxiv.org
Large language models (LLMs) have exploded in popularity due to their ability to perform a
wide array of natural language tasks. Text-based content moderation is one LLM use case …

Toxicity Detection is NOT all you Need: Measuring the Gaps to Supporting Volunteer Content Moderators

YT Cao, LF Domingo, SA Gilbert, M Mazurek… - arXiv preprint arXiv …, 2023 - arxiv.org
Extensive efforts in automated approaches for content moderation have been focused on
developing models to identify toxic, offensive, and hateful content--with the aim of lightening …

Like trainer, like bot? Inheritance of bias in algorithmic content moderation

R Binns, M Veale, M Van Kleek, N Shadbolt - Social Informatics: 9th …, 2017 - Springer
The internet has become a central medium through which 'networked publics' express their
opinions and engage in debate. Offensive comments and personal attacks can inhibit …

To act or react: Investigating proactive strategies for online community moderation

H Habib, MB Musa, F Zaffar, R Nithyanand - arXiv preprint arXiv …, 2019 - arxiv.org
Reddit administrators have generally struggled to prevent or contain such discourse for
several reasons including:(1) the inability for a handful of human administrators to track and …

An interactive exploratory tool for the task of hate speech detection

A McMillan-Major, A Paullada… - Proceedings of the …, 2022 - aclanthology.org
With the growth of Automatic Content Moderation (ACM) on widely used social media
platforms, transparency into the design of moderation technology and policy is necessary for …