Linguistic generalization and compositionality in modern artificial neural networks

M Baroni - … Transactions of the Royal Society B, 2020 - royalsocietypublishing.org
In the last decade, deep artificial neural networks have achieved astounding performance in
many natural language-processing tasks. Given the high productivity of language, these …

Compositionality decomposed: How do neural networks generalise?

D Hupkes, V Dankers, M Mul, E Bruni - Journal of Artificial Intelligence …, 2020 - jair.org
Despite a multitude of empirical studies, little consensus exists on whether neural networks
are able to generalise compositionally, a controversy that, in part, stems from a lack of …

State-of-the-art generalisation research in NLP: a taxonomy and review

D Hupkes, M Giulianelli, V Dankers, M Artetxe… - arXiv preprint arXiv …, 2022 - arxiv.org
The ability to generalise well is one of the primary desiderata of natural language
processing (NLP). Yet, what'good generalisation'entails and how it should be evaluated is …

A benchmark for systematic generalization in grounded language understanding

L Ruis, J Andreas, M Baroni… - Advances in neural …, 2020 - proceedings.neurips.cc
Humans easily interpret expressions that describe unfamiliar situations composed from
familiar parts (" greet the pink brontosaurus by the ferris wheel"). Modern neural networks, by …

Compositional generalization in semantic parsing: Pre-training vs. specialized architectures

D Furrer, M van Zee, N Scales, N Schärli - arXiv preprint arXiv:2007.08970, 2020 - arxiv.org
While mainstream machine learning methods are known to have limited ability to
compositionally generalize, new architectures and techniques continue to be proposed to …

Compositional generalization via neural-symbolic stack machines

X Chen, C Liang, AW Yu, D Song… - Advances in Neural …, 2020 - proceedings.neurips.cc
Despite achieving tremendous success, existing deep learning models have exposed
limitations in compositional generalization, the capability to learn compositional rules and …

Deep learning for laboratory earthquake prediction and autoregressive forecasting of fault zone stress

L Laurenti, E Tinti, F Galasso, L Franco… - Earth and Planetary …, 2022 - Elsevier
Earthquake forecasting and prediction have long and in some cases sordid histories but
recent work has rekindled interest based on advances in early warning, hazard assessment …

Permutation equivariant models for compositional generalization in language

J Gordon, D Lopez-Paz, M Baroni… - International …, 2019 - openreview.net
Humans understand novel sentences by composing meanings and roles of core language
components. In contrast, neural network models for natural language modeling fail when …

Sequence-to-sequence learning with latent neural grammars

Y Kim - Advances in Neural Information Processing …, 2021 - proceedings.neurips.cc
Sequence-to-sequence learning with neural networks has become the de facto standard for
sequence modeling. This approach typically models the local distribution over the next …

The neural data router: Adaptive control flow in transformers improves systematic generalization

R Csordás, K Irie, J Schmidhuber - arXiv preprint arXiv:2110.07732, 2021 - arxiv.org
Despite progress across a broad range of applications, Transformers have limited success
in systematic generalization. The situation is especially frustrating in the case of algorithmic …