BERTimbau: pretrained BERT models for Brazilian Portuguese

F Souza, R Nogueira, R Lotufo - … 2020, Rio Grande, Brazil, October 20–23 …, 2020 - Springer
Recent advances in language representation using neural networks have made it viable to
transfer the learned internal states of large pretrained language models (LMs) to …

mmarco: A multilingual version of the ms marco passage ranking dataset

L Bonifacio, V Jeronymo, HQ Abonizio… - arXiv preprint arXiv …, 2021 - arxiv.org
The MS MARCO ranking dataset has been widely used for training deep learning models for
IR tasks, achieving considerable effectiveness on diverse zero-shot scenarios. However, this …

BERT models for Brazilian Portuguese: Pretraining, evaluation and tokenization analysis

FC Souza, RF Nogueira, RA Lotufo - Applied Soft Computing, 2023 - Elsevier
Recent advances in language representation using neural networks have made it viable to
transfer the learned internal states of large pretrained language models (LMs) to …

Ptt5: Pretraining and validating the t5 model on brazilian portuguese data

D Carmo, M Piau, I Campiotti, R Nogueira… - arXiv preprint arXiv …, 2020 - arxiv.org
In natural language processing (NLP), there is a need for more resources in Portuguese,
since much of the data used in the state-of-the-art research is in other languages. In this …

A cost-benefit analysis of cross-lingual transfer methods

GM Rosa, LH Bonifacio, LR de Souza, R Lotufo… - arXiv preprint arXiv …, 2021 - arxiv.org
An effective method for cross-lingual transfer is to fine-tune a bilingual or multilingual model
on a supervised dataset in one language and evaluating it on another language in a zero …

Benchmarking natural language inference and semantic textual similarity for portuguese

P Fialho, L Coheur, P Quaresma - Information, 2020 - mdpi.com
Two sentences can be related in many different ways. Distinct tasks in natural language
processing aim to identify different semantic relations between sentences. We developed …

Answering fill-in-the-blank questions in Portuguese with transformer language models

H Gonçalo Oliveira - Progress in Artificial Intelligence: 20th EPIA …, 2021 - Springer
Despite different applications, transformer-based language models, like BERT and GPT,
learn about language by predicting missing parts of text. BERT is pretrained in Masked …

AIA-BDE: um corpo de perguntas, variações e outras anotações

HG Oliveira, AO Alves - Linguamática, 2021 - estudogeral.uc.pt
Apresentamos neste artigo o corpo AIA-BDE, que tem como principal objetivo a avalia¸ c˜ ao
de sistemas que procuram associar necessidades de informa¸ c˜ ao expressas em …

Semantic relations between sentences: from lexical to linguistically inspired semantic features and beyond

PMRP Fialho - 2023 - dspace.uevora.pt
This thesis is concerned with the identification of semantic equivalence between pairs of
natural language sentences, by studying and computing models to address Natural …

Transferência de estilo textual arbitrário em português

PB da Costa, I Paraboni - Linguamática, 2023 - linguamatica.com
Na Geração automática de língua natural, modelos de transferência de estilo textual
arbitrário objetivam a reescrita de um texto usando qualquer novo conjunto de …