Active observing in continuous-time control

S Holt, A Hüyük… - Advances in Neural …, 2024 - proceedings.neurips.cc
The control of continuous-time environments while actively deciding when to take costly
observations in time is a crucial yet unexplored problem, particularly relevant to real-world …

A generalist dynamics model for control

I Schubert, J Zhang, J Bruce, S Bechtle… - arXiv preprint arXiv …, 2023 - arxiv.org
We investigate the use of transformer sequence models as dynamics models (TDMs) for
control. We find that TDMs exhibit strong generalization capabilities to unseen …

Inverse design for fluid-structure interactions using graph network simulators

K Allen, T Lopez-Guevara… - Advances in …, 2022 - proceedings.neurips.cc
Designing physical artifacts that serve a purpose---such as tools and other functional
structures---is central to engineering as well as everyday human behavior. Though …

Combining physics and deep learning to learn continuous-time dynamics models

M Lutter, J Peters - The International Journal of Robotics …, 2023 - journals.sagepub.com
Deep learning has been widely used within learning algorithms for robotics. One
disadvantage of deep networks is that these networks are black-box representations …

Value gradient weighted model-based reinforcement learning

C Voelcker, V Liao, A Garg, A Farahmand - arXiv preprint arXiv …, 2022 - arxiv.org
Model-based reinforcement learning (MBRL) is a sample efficient technique to obtain control
policies, yet unavoidable modeling errors often lead performance deterioration. The model …

Investigating the role of model-based learning in exploration and transfer

JC Walker, E Vértes, Y Li… - International …, 2023 - proceedings.mlr.press
State of the art reinforcement learning has enabled training agents on tasks of ever
increasing complexity. However, the current paradigm tends to favor training agents from …

Neural Laplace control for continuous-time delayed systems

S Holt, A Hüyük, Z Qian, H Sun… - International …, 2023 - proceedings.mlr.press
Many real-world offline reinforcement learning (RL) problems involve continuous-time
environments with delays. Such environments are characterized by two distinctive features …

Inferring smooth control: Monte carlo posterior policy iteration with gaussian processes

J Watson, J Peters - Conference on Robot Learning, 2023 - proceedings.mlr.press
Monte Carlo methods have become increasingly relevant for control of non-differentiable
systems, approximate dynamics models, and learning from data. These methods scale to …

Simplified temporal consistency reinforcement learning

Y Zhao, W Zhao, R Boney, J Kannala… - International …, 2023 - proceedings.mlr.press
Reinforcement learning (RL) is able to solve complex sequential decision-making tasks but
is currently limited by sample efficiency and required computation. To improve sample …

Get back here: Robust imitation by return-to-distribution planning

G Cideron, B Tabanpour, S Curi, S Girgin… - arXiv preprint arXiv …, 2023 - arxiv.org
We consider the Imitation Learning (IL) setup where expert data are not collected on the
actual deployment environment but on a different version. To address the resulting …