Information theory as a bridge between language function and language form

R Futrell, M Hahn - Frontiers in Communication, 2022 - frontiersin.org
Formal and functional theories of language seem disparate, because formal theories answer
the question of what a language is, while functional theories answer the question of what …

[图书][B] Unbounded dependency constructions: Theoretical and experimental perspectives

RP Chaves, MT Putnam - 2020 - books.google.com
This book is about one of the most intriguing features of human communication systems: the
fact that words that go together in meaning can occur arbitrarily far away from each other. In …

Modeling word and morpheme order in natural language as an efficient trade-off of memory and surprisal.

M Hahn, J Degen, R Futrell - Psychological Review, 2021 - psycnet.apa.org
Memory limitations are known to constrain language comprehension and production, and
have been argued to account for crosslinguistic word order regularities. However, a …

Approximating CKY with Transformers

G Khalighinejad, O Liu, S Wiseman - arXiv preprint arXiv:2305.02386, 2023 - arxiv.org
We investigate the ability of transformer models to approximate the CKY algorithm, using
them to directly predict a sentence's parse and thus avoid the CKY algorithm's cubic …

A Procedure for Inferring a Minimalist Lexicon from an SMT Model of a Language Acquisition Device

S Indurkhya - International Conference on Grammatical …, 2023 - proceedings.mlr.press
We introduce a constraint-based procedure for inferring a Minimalist Grammar (MG) that falls
within the “Logic Grammar” framework. The procedure, implemented as a working computer …

Co-training an unsupervised constituency parser with weak supervision

N Maveli, SB Cohen - arXiv preprint arXiv:2110.02283, 2021 - arxiv.org
We introduce a method for unsupervised parsing that relies on bootstrapping classifiers to
identify if a node dominates a specific span in a sentence. There are two types of classifiers …

[PDF][PDF] Learning of Structurally Unambiguous Probabilistic Grammars

D Fisman, D Nitay… - Logical Methods in …, 2023 - lmcs.episciences.org
The problem of identifying a probabilistic context free grammar has two aspects: the first is
determining the grammar's topology (the rules of the grammar) and the second is estimating …

Estimating word co-occurrence probabilities from pretrained static embeddings using a log-bilinear model

R Futrell - Proceedings of the Workshop on Cognitive Modeling …, 2022 - aclanthology.org
We investigate how to use pretrained static word embeddings to deliver improved estimates
of bilexical co-occurrence probabilities: conditional probabilities of one word given a single …

Beyond Chomsky normal form: Extending strong learning algorithms for PCFGs

A Clark - International Conference on Grammatical Inference, 2021 - proceedings.mlr.press
We extend a recent consistent strong learning algorithm for a subclass of probabilistic
context-free grammars in Chomsky normal form,(Clark and Fijalkow, 2020) to a much larger …

[PDF][PDF] Strong learning of some Probabilistic Multiple Context-Free Grammars

A Clark - Proceedings of the 17th Meeting on the Mathematics of …, 2021 - aclanthology.org
This paper presents an algorithm for strong learning of probabilistic multiple context free
grammars from a positive sample of strings generated by the grammars. The algorithm is …