Towards trustworthy LLMs: a review on debiasing and dehallucinating in large language models

Z Lin, S Guan, W Zhang, H Zhang, Y Li… - Artificial Intelligence …, 2024 - Springer
Recently, large language models (LLMs) have attracted considerable attention due to their
remarkable capabilities. However, LLMs' generation of biased or hallucinatory content …

A Survey on Large Language Models with Multilingualism: Recent Advances and New Frontiers

K Huang, F Mo, H Li, Y Li, Y Zhang, W Yi, Y Mao… - arXiv preprint arXiv …, 2024 - arxiv.org
The rapid development of Large Language Models (LLMs) demonstrates remarkable
multilingual capabilities in natural language processing, attracting global attention in both …

Cogenesis: A framework collaborating large and small language models for secure context-aware instruction following

K Zhang, J Wang, E Hua, B Qi, N Ding… - arXiv preprint arXiv …, 2024 - arxiv.org
With the advancement of language models (LMs), their exposure to private data is
increasingly inevitable, and their deployment (especially for smaller ones) on personal …

Understanding and mitigating language confusion in llms

K Marchisio, WY Ko, A Bérard, T Dehaze… - arXiv preprint arXiv …, 2024 - arxiv.org
We investigate a surprising limitation of LLMs: their inability to consistently generate text in a
user's desired language. We create the Language Confusion Benchmark (LCB) to evaluate …

Mitigating hallucinations in large vision-language models with instruction contrastive decoding

X Wang, J Pan, L Ding, C Biemann - arXiv preprint arXiv:2403.18715, 2024 - arxiv.org
Large Vision-Language Models (LVLMs) are increasingly adept at generating contextually
detailed and coherent responses from visual inputs. However, their application in …

Language-specific neurons: The key to multilingual capabilities in large language models

T Tang, W Luo, H Huang, D Zhang, X Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Large language models (LLMs) demonstrate remarkable multilingual capabilities without
being pre-trained on specially curated multilingual parallel corpora. It remains a challenging …

Pseudo-RIS: Distinctive Pseudo-supervision Generation for Referring Image Segmentation

S Yu, PH Seo, J Son - arXiv preprint arXiv:2407.07412, 2024 - arxiv.org
We propose a new framework that automatically generates high-quality segmentation masks
with their referring expressions as pseudo supervisions for referring image segmentation …

Anti-LM Decoding for Zero-shot In-context Machine Translation

S Sia, A DeLucia, K Duh - arXiv preprint arXiv:2311.08324, 2023 - arxiv.org
Zero-shot In-context learning is the phenomenon where models can perform the task simply
given the instructions. However, pre-trained large language models are known to be poorly …

Enhanced Hallucination Detection in Neural Machine Translation through Simple Detector Aggregation

A Himmi, G Staerman, M Picot, P Colombo… - arXiv preprint arXiv …, 2024 - arxiv.org
Hallucinated translations pose significant threats and safety concerns when it comes to the
practical deployment of machine translation systems. Previous research works have …

Fast and Slow Generating: An Empirical Study on Large and Small Language Models Collaborative Decoding

K Zhang, J Wang, N Ding, B Qi, E Hua, X Lv… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) demonstrate impressive performance in diverse
applications, yet they face significant drawbacks, including high inference latency …