Towards continual reinforcement learning: A review and perspectives

K Khetarpal, M Riemer, I Rish, D Precup - Journal of Artificial Intelligence …, 2022 - jair.org
In this article, we aim to provide a literature review of different formulations and approaches
to continual reinforcement learning (RL), also known as lifelong or non-stationary RL. We …

Meta-learning in natural and artificial intelligence

JX Wang - Current Opinion in Behavioral Sciences, 2021 - Elsevier
Highlights•Multiple scales of learning (and hence meta-learning) are ubiquitous in
nature.•Many existing lines of work in neuroscience and cognitive science touch upon …

The synthesizability of molecules proposed by generative models

W Gao, CW Coley - Journal of chemical information and modeling, 2020 - ACS Publications
The discovery of functional molecules is an expensive and time-consuming process,
exemplified by the rising costs of small molecule therapeutic discovery. One class of …

Mt-opt: Continuous multi-task robotic reinforcement learning at scale

D Kalashnikov, J Varley, Y Chebotar… - arXiv preprint arXiv …, 2021 - arxiv.org
General-purpose robotic systems must master a large repertoire of diverse skills to be useful
in a range of daily tasks. While reinforcement learning provides a powerful framework for …

Varibad: A very good method for bayes-adaptive deep rl via meta-learning

L Zintgraf, K Shiarlis, M Igl, S Schulze, Y Gal… - arXiv preprint arXiv …, 2019 - arxiv.org
Trading off exploration and exploitation in an unknown environment is key to maximising
expected return during learning. A Bayes-optimal policy, which does so optimally, conditions …

Human-timescale adaptation in an open-ended task space

AA Team, J Bauer, K Baumli, S Baveja… - arXiv preprint arXiv …, 2023 - arxiv.org
Foundation models have shown impressive adaptation and scalability in supervised and self-
supervised learning problems, but so far these successes have not fully translated to …

Transformers can do bayesian inference

S Müller, N Hollmann, SP Arango, J Grabocka… - arXiv preprint arXiv …, 2021 - arxiv.org
Currently, it is hard to reap the benefits of deep learning for Bayesian methods, which allow
the explicit specification of prior knowledge and accurately capture model uncertainty. We …

Human-timescale adaptation in an open-ended task space

J Bauer, K Baumli, F Behbahani… - International …, 2023 - proceedings.mlr.press
Foundation models have shown impressive adaptation and scalability in supervised and self-
supervised learning problems, but so far these successes have not fully translated to …

General-purpose in-context learning by meta-learning transformers

L Kirsch, J Harrison, J Sohl-Dickstein, L Metz - arXiv preprint arXiv …, 2022 - arxiv.org
Modern machine learning requires system designers to specify aspects of the learning
pipeline, such as losses, architectures, and optimizers. Meta-learning, or learning-to-learn …

[HTML][HTML] Replay and compositional computation

Z Kurth-Nelson, T Behrens, G Wayne, K Miller… - Neuron, 2023 - cell.com
Replay in the brain has been viewed as rehearsal or, more recently, as sampling from a
transition model. Here, we propose a new hypothesis: that replay is able to implement a form …