Language models are weak learners

H Manikandan, Y Jiang… - Advances in Neural …, 2023 - proceedings.neurips.cc
A central notion in practical and theoretical machine learning is that of a weak learner,
classifiers that achieve better-than-random performance (on any given distribution over …

Zero-Shot Continuous Prompt Transfer: Generalizing Task Semantics Across Language Models

Z Wu, Y Wu, L Mou - arXiv preprint arXiv:2310.01691, 2023 - arxiv.org
Prompt tuning in natural language processing (NLP) has become an increasingly popular
method for adapting large language models to specific tasks. However, the transferability of …

Unnatural language processing: How do language models handle machine-generated prompts?

C Kervadec, F Franzon, M Baroni - arXiv preprint arXiv:2310.15829, 2023 - arxiv.org
Language model prompt optimization research has shown that semantically and
grammatically well-formed manually crafted prompts are routinely outperformed by …

Latent Communication in Artificial Neural Networks

L Moschella - arXiv preprint arXiv:2406.11014, 2024 - arxiv.org
As NNs permeate various scientific and industrial domains, understanding the universality
and reusability of their representations becomes crucial. At their core, these networks create …

STARLING: Self-supervised Training of Text-based Reinforcement Learning Agent with Large Language Models

S Basavatia, K Murugesan, S Ratnakar - arXiv preprint arXiv:2406.05872, 2024 - arxiv.org
Interactive fiction games have emerged as an important application to improve the
generalization capabilities of language-based reinforcement learning (RL) agents. Existing …

[HTML][HTML] Out-of-distribution generalisation in machine learning

A Słowik - 2023 - repository.cam.ac.uk
Abstract Machine learning has proven extremely useful in many applications in recent years.
However, a lot of these success stories stem from evaluating the algorithms on data very …