Hyperbolic deep neural networks: A survey

W Peng, T Varanka, A Mostafa, H Shi… - IEEE Transactions on …, 2021 - ieeexplore.ieee.org
Recently, hyperbolic deep neural networks (HDNNs) have been gaining momentum as the
deep representations in the hyperbolic space provide high fidelity embeddings with few …

Poincaré embeddings for learning hierarchical representations

M Nickel, D Kiela - Advances in neural information …, 2017 - proceedings.neurips.cc
Abstract Representation learning has become an invaluable approach for learning from
symbolic data such as text and graphs. However, state-of-the-art embedding methods …

Meaning without reference in large language models

ST Piantadosi, F Hill - arXiv preprint arXiv:2208.02957, 2022 - arxiv.org
The widespread success of large language models (LLMs) has been met with skepticism
that they possess anything like human concepts or meanings. Contrary to claims that LLMs …

Poincar\'e glove: Hyperbolic word embeddings

A Tifrea, G Bécigneul, OE Ganea - arXiv preprint arXiv:1810.06546, 2018 - arxiv.org
Words are not created equal. In fact, they form an aristocratic graph with a latent hierarchical
structure that the next generation of unsupervised learned word embeddings should reveal …

Semantic specialization of distributional word vector spaces using monolingual and cross-lingual constraints

N Mrkšić, I Vulić, DÓ Séaghdha, I Leviant… - Transactions of the …, 2017 - direct.mit.edu
Abstract We present Attract-Repel, an algorithm for improving the semantic quality of word
vectors by injecting constraints extracted from lexical resources. Attract-Repel facilitates the …

Embedding text in hyperbolic spaces

B Dhingra, CJ Shallue, M Norouzi, AM Dai… - arXiv preprint arXiv …, 2018 - arxiv.org
Natural language text exhibits hierarchical structure in a variety of respects. Ideally, we could
incorporate our prior knowledge of this hierarchical structure into unsupervised learning …

Hearst patterns revisited: Automatic hypernym detection from large text corpora

S Roller, D Kiela, M Nickel - arXiv preprint arXiv:1806.03191, 2018 - arxiv.org
Methods for unsupervised hypernym detection may broadly be categorized according to two
paradigms: pattern-based and distributional methods. In this paper, we study the …

Dynamic meta-embeddings for improved sentence representations

D Kiela, C Wang, K Cho - arXiv preprint arXiv:1804.07983, 2018 - arxiv.org
While one of the first steps in many NLP systems is selecting what pre-trained word
embeddings to use, we argue that such a step is better left for neural networks to figure out …

Specialising word vectors for lexical entailment

I Vulić, N Mrkšić - arXiv preprint arXiv:1710.06371, 2017 - arxiv.org
We present LEAR (Lexical Entailment Attract-Repel), a novel post-processing method that
transforms any input word vector space to emphasise the asymmetric relation of lexical …

Hierarchical embeddings for hypernymy detection and directionality

KA Nguyen, M Köper, SS Walde, NT Vu - arXiv preprint arXiv:1707.07273, 2017 - arxiv.org
We present a novel neural model HyperVec to learn hierarchical embeddings for hypernymy
detection and directionality. While previous embeddings have shown limitations on …