Combating misinformation in the age of llms: Opportunities and challenges

C Chen, K Shu - arXiv preprint arXiv:2311.05656, 2023 - arxiv.org
Misinformation such as fake news and rumors is a serious threat on information ecosystems
and public trust. The emergence of Large Language Models (LLMs) has great potential to …

Use of llms for illicit purposes: Threats, prevention measures, and vulnerabilities

M Mozes, X He, B Kleinberg, LD Griffin - arXiv preprint arXiv:2308.12833, 2023 - arxiv.org
Spurred by the recent rapid increase in the development and distribution of large language
models (LLMs) across industry and academia, much recent work has drawn attention to …

How does Twitter account moderation work? Dynamics of account creation and suspension on Twitter during major geopolitical events

F Pierri, L Luceri, E Chen, E Ferrara - EPJ Data Science, 2023 - epjds.epj.org
Social media moderation policies are often at the center of public debate, and their
implementation and enactment are sometimes surrounded by a veil of mystery …

Toxic language detection: A systematic review of Arabic datasets

I Bensalem, P Rosso, H Zitouni - Expert Systems, 2024 - Wiley Online Library
The detection of toxic language in the Arabic language has emerged as an active area of
research in recent years, and reviewing the existing datasets employed for training the …

Overview of the CLEF-2023 CheckThat! Lab Task 1 on Check-Worthiness of Multimodal and Multigenre Content

F Alam, A Barrón-Cedeño, GS Cheema, GK Shahi… - 2023 - dclibrary.mbzuai.ac.ae
We present an overview of CheckThat! Lab's 2023 Task 1, which is part of CLEF-2023. Task
1 asks to determine whether a text item, or a text coupled with an image, is check-worthy …

Overview of the CLEF-2024 CheckThat! lab task 1 on check-worthiness estimation of multigenre content

M Hasanain, R Suwaileh, S Weering, C Li… - 25th Working Notes of …, 2024 - research.rug.nl
We present an overview of the CheckThat! Lab 2024 Task 1, part of CLEF 2024. Task 1
involves determining whether a text item is check-worthy, with a special emphasis on COVID …

OffensEval 2023: Offensive language identification in the age of Large Language Models

M Zampieri, S Rosenthal, P Nakov… - Natural Language …, 2023 - cambridge.org
The OffensEval shared tasks organized as part of SemEval-2019–2020 were very popular,
attracting over 1300 participating teams. The two editions of the shared task helped advance …

Mossbench: Is your multimodal language model oversensitive to safe queries?

X Li, H Zhou, R Wang, T Zhou, M Cheng… - arXiv preprint arXiv …, 2024 - arxiv.org
Humans are prone to cognitive distortions--biased thinking patterns that lead to exaggerated
responses to specific stimuli, albeit in very different contexts. This paper demonstrates that …

Analyzing the use of large language models for content moderation with chatgpt examples

M Franco, O Gaggi, CE Palazzi - … of the 3rd International Workshop on …, 2023 - dl.acm.org
Content moderation systems are crucial in Online Social Networks (OSNs). Indeed, their role
is to keep platforms and their users safe from malicious activities. However, there is an …

Harm mitigation in recommender systems under user preference dynamics

J Chee, S Kalyanaraman, SK Ernala… - Proceedings of the 30th …, 2024 - dl.acm.org
We consider a recommender system that takes into account the interplay between
recommendations, the evolution of user interests, and harmful content. We model the impact …