This work identifies 18 foundational challenges in assuring the alignment and safety of large language models (LLMs). These challenges are organized into three different categories …
Petroni et al.(2019) demonstrated that it is possible to retrieve world facts from a pre-trained language model by expressing them as cloze-style prompts and interpret the model's …
Large language models have achieved success on a number of downstream tasks, particularly in a few and zero-shot manner. As a consequence, researchers have been …
Probing strategies have been shown to detect the presence of various linguistic features in large language models; in particular, semantic features intermediate to the" natural logic" …
Several studies have been carried out on revealing linguistic features captured by BERT. This is usually achieved by training a diagnostic classifier on the representations obtained …
Neural models achieve high performance on a variety of natural language processing (NLP) benchmark tasks. How models perform these tasks, though, is notoriously poorly …
Large NLP models have recently shown impressive performance in language understanding tasks, typically evaluated by their fine-tuned performance. Alternatively …
K Tatariya, V Araujo, T Bauwens… - arXiv preprint arXiv …, 2024 - arxiv.org
Pixel-based language models have emerged as a compelling alternative to subword-based language modelling, particularly because they can represent virtually any script. PIXEL, a …