BERT models for Arabic text classification: a systematic review

AS Alammary - Applied Sciences, 2022 - mdpi.com
Bidirectional Encoder Representations from Transformers (BERT) has gained increasing
attention from researchers and practitioners as it has proven to be an invaluable technique …

Adapting pre-trained language models to African languages via multilingual adaptive fine-tuning

JO Alabi, DI Adelani, M Mosbach, D Klakow - arXiv preprint arXiv …, 2022 - arxiv.org
Multilingual pre-trained language models (PLMs) have demonstrated impressive
performance on several downstream tasks for both high-resourced and low-resourced …

Fake news spreaders detection: Sometimes attention is not all you need

M Siino, E Di Nuovo, I Tinnirello, M La Cascia - Information, 2022 - mdpi.com
Guided by a corpus linguistics approach, in this article we present a comparative evaluation
of State-of-the-Art (SotA) models, with a special focus on Transformers, to address the task of …

Dziribert: a pre-trained language model for the algerian dialect

A Abdaoui, M Berrimi, M Oussalah… - arXiv preprint arXiv …, 2021 - arxiv.org
Pre-trained transformers are now the de facto models in Natural Language Processing given
their state-of-the-art results in many tasks and languages. However, most of the current …

Scandeval: A benchmark for Scandinavian natural language processing

DS Nielsen - arXiv preprint arXiv:2304.00906, 2023 - arxiv.org
This paper introduces a Scandinavian benchmarking platform, ScandEval, which can
benchmark any pretrained model on four different tasks in the Scandinavian languages. The …

Mini-model adaptation: Efficiently extending pretrained models to new languages via aligned shallow training

K Marchisio, P Lewis, Y Chen, M Artetxe - arXiv preprint arXiv:2212.10503, 2022 - arxiv.org
Prior work shows that it is possible to expand pretrained Masked Language Models (MLMs)
to new languages by learning a new set of embeddings, while keeping the transformer body …

You are what you write: Preserving privacy in the era of large language models

R Plant, V Giuffrida, D Gkatzia - arXiv preprint arXiv:2204.09391, 2022 - arxiv.org
Large scale adoption of large language models has introduced a new era of convenient
knowledge transfer for a slew of natural language processing tasks. However, these models …

Data-augmentation for bangla-english code-mixed sentiment analysis: Enhancing cross linguistic contextual understanding

M Tareq, MF Islam, S Deb, S Rahman… - IEEE Access, 2023 - ieeexplore.ieee.org
In today's digital world, automated sentiment analysis from online reviews can contribute to a
wide variety of decision-making processes. One example is examining typical perceptions of …

BLADE: combining vocabulary pruning and intermediate pretraining for scaleable neural CLIR

S Nair, E Yang, D Lawrie, J Mayfield… - Proceedings of the 46th …, 2023 - dl.acm.org
Learning sparse representations using pretrained language models enhances the
monolingual ranking effectiveness. Such representations are sparse vectors in the …

On the Usability of Transformers-based models for a French Question-Answering task

O Cattan, C Servan, S Rosset - arXiv preprint arXiv:2207.09150, 2022 - arxiv.org
For many tasks, state-of-the-art results have been achieved with Transformer-based
architectures, resulting in a paradigmatic shift in practices from the use of task-specific …