Semantic Text Transmission via Prediction with Small Language Models: Cost-Similarity Trade-off

BA Madhabhavi, G Karevvanavar, RV Bhat… - arXiv preprint arXiv …, 2024 - arxiv.org
We consider the communication of natural language text from a source to a destination over
noiseless and character-erasure channels. We exploit language's inherent correlations and …

A fast and simple algorithm for training neural probabilistic language models

A Mnih, YW Teh - arXiv preprint arXiv:1206.6426, 2012 - arxiv.org
In spite of their superior performance, neural probabilistic language models (NPLMs) remain
far less widely used than n-gram models due to their notoriously long training times, which …

Positional encoding to control output sequence length

S Takase, N Okazaki - arXiv preprint arXiv:1904.07418, 2019 - arxiv.org
Neural encoder-decoder models have been successful in natural language generation
tasks. However, real applications of abstractive summarization must consider additional …

Semantics-native communication with contextual reasoning

H Seo, J Park, M Bennis, M Debbah - arXiv preprint arXiv:2108.05681, 2021 - arxiv.org
Spurred by a huge interest in the post-Shannon communication, it has recently been shown
that leveraging semantics can significantly improve the communication effectiveness across …

Lightweight Diffusion Models for Resource-Constrained Semantic Communication

G Pignata, E Grassucci, G Cicchetti… - arXiv preprint arXiv …, 2024 - arxiv.org
Recently, generative semantic communication models have proliferated as they are
revolutionizing semantic communication frameworks, improving their performance, and …

Enhancing neural data-to-text generation models with external background knowledge

S Chen, J Wang, X Feng, F Jiang, B Qin… - Proceedings of the …, 2019 - aclanthology.org
Recent neural models for data-to-text generation rely on massive parallel pairs of data and
text to learn the writing knowledge. They often assume that writing knowledge can be …

Lightweight Decoding Strategies for Increasing Specificity

KI Gero, C Kedzie, S Petridis, L Chilton - arXiv preprint arXiv:2110.11850, 2021 - arxiv.org
Language models are known to produce vague and generic outputs. We propose two
unsupervised decoding strategies based on either word-frequency or point-wise mutual …

A word prediction methodology for automatic sentence completion

C Spiccia, A Augello, G Pilato… - Proceedings of the 2015 …, 2015 - ieeexplore.ieee.org
Word prediction generally relies on n-grams occurrence statistics, which may have huge
data storage requirements and does not take into account the general meaning of the text …

: Increasing GPU Utilization during Generative Inference for Higher Throughput

Y Jin, CF Wu, D Brooks, GY Wei - Advances in Neural …, 2023 - proceedings.neurips.cc
Generating texts with a large language model (LLM) consumes massive amounts of
memory. Apart from the already-large model parameters, the key/value (KV) cache that …

Recyclegpt: An autoregressive language model with recyclable module

Y Jiang, Q He, X Zhuang, Z Wu, K Wang… - arXiv preprint arXiv …, 2023 - arxiv.org
Existing large language models have to run K times to generate a sequence of K tokens. In
this paper, we present RecycleGPT, a generative language model with fast decoding speed …