A literature review of textual hate speech detection methods and datasets

F Alkomah, X Ma - Information, 2022 - mdpi.com
Online toxic discourses could result in conflicts between groups or harm to online
communities. Hate speech is complex and multifaceted harmful or offensive content …

Detection and moderation of detrimental content on social media platforms: current status and future directions

VU Gongane, MV Munot, AD Anuse - Social Network Analysis and Mining, 2022 - Springer
Social Media has become a vital component of every individual's life in society opening a
preferred spectrum of virtual communication which provides an individual with a freedom to …

[HTML][HTML] ChatGPT: Jack of all trades, master of none

J Kocoń, I Cichecki, O Kaszyca, M Kochanek, D Szydło… - Information …, 2023 - Elsevier
OpenAI has released the Chat Generative Pre-trained Transformer (ChatGPT) and
revolutionized the approach in artificial intelligence to human-model interaction. The first …

Taxonomy of risks posed by language models

L Weidinger, J Uesato, M Rauh, C Griffin… - Proceedings of the …, 2022 - dl.acm.org
Responsible innovation on large-scale Language Models (LMs) requires foresight into and
in-depth understanding of the risks these models may pose. This paper develops a …

Toward a perspectivist turn in ground truthing for predictive computing

F Cabitza, A Campagner, V Basile - … of the AAAI Conference on Artificial …, 2023 - ojs.aaai.org
Abstract Most current Artificial Intelligence applications are based on supervised Machine
Learning (ML), which ultimately grounds on data annotated by small teams of experts or …

Is your toxicity my toxicity? exploring the impact of rater identity on toxicity annotation

N Goyal, ID Kivlichan, R Rosen… - Proceedings of the ACM …, 2022 - dl.acm.org
Machine learning models are commonly used to detect toxicity in online conversations.
These models are trained on datasets annotated by human raters. We explore how raters' …

Dices dataset: Diversity in conversational ai evaluation for safety

L Aroyo, A Taylor, M Diaz, C Homan… - Advances in …, 2024 - proceedings.neurips.cc
Abstract Machine learning approaches often require training and evaluation datasets with a
clear separation between positive and negative examples. This requirement overly …

Why don't you do it right? analysing annotators' disagreement in subjective tasks

M Sandri, E Leonardelli, S Tonelli… - Proceedings of the 17th …, 2023 - aclanthology.org
Annotators' disagreement in linguistic data has been recently the focus of multiple initiatives
aimed at raising awareness on issues related to 'majority voting'when aggregating diverging …

[HTML][HTML] Human-centered neural reasoning for subjective content processing: Hate speech, emotions, and humor

P Kazienko, J Bielaniewicz, M Gruza, K Kanclerz… - Information …, 2023 - Elsevier
Some tasks in content processing, eg, natural language processing (NLP), like hate or
offensive speech and emotional or funny text detection, are subjective by nature. Each …

[HTML][HTML] Detecting abusive comments at a fine-grained level in a low-resource language

BR Chakravarthi, R Priyadharshini, S Banerjee… - Natural Language …, 2023 - Elsevier
YouTube is a video-sharing and social media platform where users create profiles and
share videos for their followers to view, like, and comment on. Abusive comments on videos …