How to leverage unlabeled data in offline reinforcement learning

T Yu, A Kumar, Y Chebotar… - International …, 2022 - proceedings.mlr.press
Offline reinforcement learning (RL) can learn control policies from static datasets but, like
standard RL methods, it requires reward annotations for every transition. In many cases …

Conservative data sharing for multi-task offline reinforcement learning

T Yu, A Kumar, Y Chebotar… - Advances in …, 2021 - proceedings.neurips.cc
Offline reinforcement learning (RL) algorithms have shown promising results in domains
where abundant pre-collected data is available. However, prior methods focus on solving …

Pre-training for robots: Offline rl enables learning new tasks from a handful of trials

A Kumar, A Singh, F Ebert, M Nakamoto… - arXiv preprint arXiv …, 2022 - arxiv.org
Progress in deep learning highlights the tremendous potential of utilizing diverse robotic
datasets for attaining effective generalization and makes it enticing to consider leveraging …

Offline Meta Reinforcement Learning--Identifiability Challenges and Effective Data Collection Strategies

R Dorfman, I Shenfeld, A Tamar - Advances in Neural …, 2021 - proceedings.neurips.cc
Consider the following instance of the Offline Meta Reinforcement Learning (OMRL)
problem: given the complete training logs of $ N $ conventional RL agents, trained on $ N …

Metadiffuser: Diffusion model as conditional planner for offline meta-rl

F Ni, J Hao, Y Mu, Y Yuan, Y Zheng… - International …, 2023 - proceedings.mlr.press
Recently, diffusion model shines as a promising backbone for the sequence modeling
paradigm in offline reinforcement learning (RL). However, these works mostly lack the …

Don't start from scratch: Leveraging prior data to automate robotic reinforcement learning

HR Walke, JH Yang, A Yu, A Kumar… - … on Robot Learning, 2023 - proceedings.mlr.press
Reinforcement learning (RL) algorithms hold the promise of enabling autonomous skill
acquisition for robotic systems. However, in practice, real-world robotic RL typically requires …

[PDF][PDF] Discovering generalizable multi-agent coordination skills from multi-task offline data

F Zhang, C Jia, YC Li, L Yuan, Y Yu… - … Conference on Learning …, 2022 - drive.google.com
Cooperative multi-agent reinforcement learning (MARL) faces the challenge of adapting to
multiple tasks with varying agents and targets. Previous multi-task MARL approaches …

Model-based meta reinforcement learning using graph structured surrogate models and amortized policy search

Q Wang, H Van Hoof - International Conference on Machine …, 2022 - proceedings.mlr.press
Reinforcement learning is a promising paradigm for solving sequential decision-making
problems, but low data efficiency and weak generalization across tasks are bottlenecks in …

Generalizable Task Representation Learning for Offline Meta-Reinforcement Learning with Data Limitations

R Zhou, CX Gao, Z Zhang, Y Yu - … of the AAAI Conference on Artificial …, 2024 - ojs.aaai.org
Generalization and sample efficiency have been long-standing issues concerning
reinforcement learning, and thus the field of Offline Meta-Reinforcement Learning (OMRL) …

Context shift reduction for offline meta-reinforcement learning

Y Gao, R Zhang, J Guo, F Wu, Q Yi… - Advances in …, 2024 - proceedings.neurips.cc
Offline meta-reinforcement learning (OMRL) utilizes pre-collected offline datasets to
enhance the agent's generalization ability on unseen tasks. However, the context shift …