Combating misinformation in the age of llms: Opportunities and challenges

C Chen, K Shu - AI Magazine, 2023 - Wiley Online Library
Misinformation such as fake news and rumors is a serious threat for information ecosystems
and public trust. The emergence of large language models (LLMs) has great potential to …

The llama 3 herd of models

A Dubey, A Jauhri, A Pandey, A Kadian… - arXiv preprint arXiv …, 2024 - arxiv.org
Modern artificial intelligence (AI) systems are powered by foundation models. This paper
presents a new set of foundation models, called Llama 3. It is a herd of language models …

Can llm-generated misinformation be detected?

C Chen, K Shu - arXiv preprint arXiv:2309.13788, 2023 - arxiv.org
The advent of Large Language Models (LLMs) has made a transformative impact. However,
the potential that LLMs such as ChatGPT can be exploited to generate misinformation has …

The AI Risk Repository: A Comprehensive Meta-Review, Database, and Taxonomy of Risks From Artificial Intelligence

P Slattery, AK Saeri, EAC Grundy, J Graham… - arXiv preprint arXiv …, 2024 - arxiv.org
The risks posed by Artificial Intelligence (AI) are of considerable concern to academics,
auditors, policymakers, AI companies, and the public. However, a lack of shared …

WalledEval: A Comprehensive Safety Evaluation Toolkit for Large Language Models

P Gupta, LQ Yau, HH Low, I Lee, HM Lim… - arXiv preprint arXiv …, 2024 - arxiv.org
WalledEval is a comprehensive AI safety testing toolkit designed to evaluate large language
models (LLMs). It accommodates a diverse range of models, including both open-weight …

Can Editing LLMs Inject Harm?

C Chen, B Huang, Z Li, Z Chen, S Lai, X Xu… - arXiv preprint arXiv …, 2024 - arxiv.org
Knowledge editing techniques have been increasingly adopted to efficiently correct the false
or outdated knowledge in Large Language Models (LLMs), due to the high cost of retraining …

Not My Voice! A Taxonomy of Ethical and Safety Harms of Speech Generators

W Hutiri, O Papakyriakopoulos, A Xiang - The 2024 ACM Conference on …, 2024 - dl.acm.org
The rapid and wide-scale adoption of AI to generate human speech poses a range of
significant ethical and safety risks to society that need to be addressed. For example, a …

Eureka: Evaluating and Understanding Large Foundation Models

V Balachandran, J Chen, N Joshi, B Nushi… - arXiv preprint arXiv …, 2024 - arxiv.org
Rigorous and reproducible evaluation is critical for assessing the state of the art and for
guiding scientific advances in Artificial Intelligence. Evaluation is challenging in practice due …

Authorship Attribution in the Era of LLMs: Problems, Methodologies, and Challenges

B Huang, C Chen, K Shu - arXiv preprint arXiv:2408.08946, 2024 - arxiv.org
Accurate attribution of authorship is crucial for maintaining the integrity of digital content,
improving forensic investigations, and mitigating the risks of misinformation and plagiarism …

Decoding Biases: Automated Methods and LLM Judges for Gender Bias Detection in Language Models

SH Kumar, S Sahay, S Mazumder, E Okur… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) have excelled at language understanding and generating
human-level text. However, even with supervised training and human alignment, these …