Atlas-Chat: Adapting Large Language Models for Low-Resource Moroccan Arabic Dialect

G Shang, H Abdine, Y Khoubrane, A Mohamed… - arXiv preprint arXiv …, 2024 - arxiv.org
We introduce Atlas-Chat, the first-ever collection of LLMs specifically developed for dialectal
Arabic. Focusing on Moroccan Arabic, also known as Darija, we construct our instruction …

Seallms 3: Open foundation and chat multilingual large language models for southeast asian languages

W Zhang, HP Chan, Y Zhao, M Aljunied, J Wang… - arXiv preprint arXiv …, 2024 - arxiv.org
Large Language Models (LLMs) have shown remarkable abilities across various tasks, yet
their development has predominantly centered on high-resource languages like English and …

Style-specific neurons for steering llms in text style transfer

W Lai, V Hangya, A Fraser - arXiv preprint arXiv:2410.00593, 2024 - arxiv.org
Text style transfer (TST) aims to modify the style of a text without altering its original
meaning. Large language models (LLMs) demonstrate superior performance across …

Towards understanding multi-task learning (generalization) of llms via detecting and exploring task-specific neurons

Y Leng, D Xiong - arXiv preprint arXiv:2407.06488, 2024 - arxiv.org
While large language models (LLMs) have demonstrated superior multi-task capabilities,
understanding the learning mechanisms behind this is still a challenging problem. In this …

Multilingual needle in a haystack: Investigating long-context behavior of multilingual large language models

A Hengle, P Bajpai, S Dan, T Chakraborty - arXiv preprint arXiv …, 2024 - arxiv.org
While recent large language models (LLMs) demonstrate remarkable abilities in responding
to queries in diverse languages, their ability to handle long multilingual contexts is …

Explain in Plain Language Questions with Indic Languages: Drawbacks, Affordances, and Opportunities

DH Smith IV, V Kumar, P Denny - arXiv preprint arXiv:2409.20297, 2024 - arxiv.org
Background: Introductory computer science courses use``Explain in Plain English''(EiPE)
activities to develop and assess students' code comprehension skills, but creating effective …

Getting More from Less: Large Language Models are Good Spontaneous Multilingual Learners

S Zhang, C Gao, W Zhu, J Chen, X Huang… - Proceedings of the …, 2024 - aclanthology.org
Abstract Recently, Large Language Models (LLMs) have shown impressive language
capabilities, while most of them have very unbalanced performance across different …

How Transliterations Improve Crosslingual Alignment

Y Liu, M Wang, AH Kargaran, A Imani, O Xhelili… - arXiv preprint arXiv …, 2024 - arxiv.org
Recent studies have shown that post-aligning multilingual pretrained language models
(mPLMs) using alignment objectives on both original and transliterated data can improve …

Exploring Large Language Models on Cross-Cultural Values in Connection with Training Methodology

M Kim, S Baek - arXiv preprint arXiv:2412.08846, 2024 - arxiv.org
Large language models (LLMs) closely interact with humans, and thus need an intimate
understanding of the cultural values of human society. In this paper, we explore how open …

Lens: Rethinking Multilingual Enhancement for Large Language Models

W Zhao, Y Hu, J Guo, X Sui, T Wu, Y Deng… - arXiv preprint arXiv …, 2024 - arxiv.org
Despite the growing global demand for large language models (LLMs) that serve users from
diverse linguistic backgrounds, most cutting-edge LLMs remain predominantly English …