Information theory as a bridge between language function and language form

R Futrell, M Hahn - Frontiers in Communication, 2022 - frontiersin.org
Formal and functional theories of language seem disparate, because formal theories answer
the question of what a language is, while functional theories answer the question of what …

Topic models do not model topics: epistemological remarks and steps towards best practices

A Shadrova - Journal of Data Mining & Digital Humanities, 2021 - jdmdh.episciences.org
The social sciences and digital humanities have recently adopted the machine learning
technique of topic modeling to address research questions in their fields. This is problematic …

Information Theory Opens New Dimensions in Experimental Studies of Animal Behaviour and Communication

Z Reznikova - Animals, 2023 - mdpi.com
Simple Summary An information theory approach provides new methods for experimental
studying and analysis of animals' communication and behavioural sequences. One of the …

A challenge for contrastive L1/L2 corpus studies: Large inter-and intra-individual variation across morphological, but not global syntactic categories in task-based …

A Shadrova, P Linscheid, J Lukassek… - Frontiers in …, 2021 - frontiersin.org
In this paper, we present corpus data that questions the concept of native speaker
homogeneity as it is presumed in many studies using native speakers (L1) as a control …

RePair and all irreducible grammars are upper bounded by high-order empirical entropy

C Ochoa, G Navarro - IEEE Transactions on Information Theory, 2018 - ieeexplore.ieee.org
Irreducible grammars are a class of context-free grammars with well-known representatives,
such as Repair (with a few tweaks), Longest Match, Greedy, and Sequential. We show that a …

Complexity and entropy in legal language

R Friedrich - Frontiers in Physics, 2021 - frontiersin.org
We study the language of legal codes from different countries and legal traditions, using
concepts from physics, algorithmic complexity theory and information theory. We show that …

Estimating predictive rate–distortion curves via neural variational inference

M Hahn, R Futrell - Entropy, 2019 - mdpi.com
The Predictive Rate–Distortion curve quantifies the trade-off between compressing
information about the past of a stochastic process and predicting its future accurately …

A refutation of finite-state language models through Zipf's law for factual knowledge

Ł Dębowski - Entropy, 2021 - mdpi.com
We present a hypothetical argument against finite-state processes in statistical language
modeling that is based on semantics rather than syntax. In this theoretical model, we …

Measuring coselectional constraint in learner corpora: A graph-based approach

AV Shadrova - 2020 - edoc.hu-berlin.de
The thesis located in corpus linguistics analyzes the acquisition of coselectional constraint in
learners of German as a second language in a quasi-longitudinal design based on the …

Universal densities exist for every finite reference measure

Ł Dębowski - IEEE Transactions on Information Theory, 2023 - ieeexplore.ieee.org
As it is known, universal codes, which estimate the entropy rate consistently, exist for
stationary ergodic sources over finite alphabets but not over countably infinite ones. We …