AlephBERT: Language model pre-training and evaluation from sub-word to sentence level

A Seker, E Bandel, D Bareket… - Proceedings of the …, 2022 - aclanthology.org
Abstract Large Pre-trained Language Models (PLMs) have become ubiquitous in the
development of language understanding technology and lie at the heart of many artificial …

Multitask learning for emotion and personality traits detection

Y Li, A Kazemeini, Y Mehta, E Cambria - Neurocomputing, 2022 - Elsevier
In recent years, deep learning-based automated personality traits detection has received a
lot of attention, especially now, due to the massive digital footprints of an individual …

On the frontlines: The impact of the COVID-19 pandemic on social workers' well-being

TM Schwartz Tayri - Social Work, 2023 - academic.oup.com
The crisis created by the spread of COVID-19 brought increasing needs and referrals to
social welfare services in many countries. However, at the same time, social services …

AlephBERT: A Hebrew large pre-trained language model to start-off your Hebrew NLP application with

A Seker, E Bandel, D Bareket, I Brusilovsky… - arXiv preprint arXiv …, 2021 - arxiv.org
Large Pre-trained Language Models (PLMs) have become ubiquitous in the development of
language understanding technology and lie at the heart of many artificial intelligence …

Quality at a glance: An audit of web-crawled multilingual datasets

J Kreutzer, I Caswell, L Wang, A Wahab… - Transactions of the …, 2022 - direct.mit.edu
With the success of large-scale pre-training and multilingual modeling in Natural Language
Processing (NLP), recent years have seen a proliferation of large, Web-mined text datasets …

Predicting Client Emotions and Therapist Interventions in Psychotherapy Dialogues

T Mayer, N Warikoo, A Eliassaf… - Proceedings of the …, 2024 - aclanthology.org
Abstract Natural Language Processing (NLP) can advance psychotherapy research by
scaling up therapy dialogue analysis as well as by allowing researchers to examine client …

Large pre-trained models with extra-large vocabularies: A contrastive analysis of hebrew bert models and a new one to outperform them all

E Gueta, A Shmidman, S Shmidman… - arXiv preprint arXiv …, 2022 - arxiv.org
We present a new pre-trained language model (PLM) for modern Hebrew, termed
AlephBERTGimmel, which employs a much larger vocabulary (128K items) than standard …

ParaShoot: A Hebrew question answering dataset

O Keren, O Levy - arXiv preprint arXiv:2109.11314, 2021 - arxiv.org
NLP research in Hebrew has largely focused on morphology and syntax, where rich
annotated datasets in the spirit of Universal Dependencies are available. Semantic datasets …

Offensive Hebrew corpus and detection using bert

N Hamad, M Jarrar, M Khalilia… - 2023 20th ACS/IEEE …, 2023 - ieeexplore.ieee.org
Offensive language detection has been well studied in many languages, but it is lagging
behind in low-resource languages, such as Hebrew. In this paper, we present a new …

Hero: Roberta and longformer hebrew language models

V Shalumov, H Haskey - arXiv preprint arXiv:2304.11077, 2023 - arxiv.org
In this paper, we fill in an existing gap in resources available to the Hebrew NLP community
by providing it with the largest so far pre-train dataset HeDC4, a state-of-the-art pre-trained …