Maximum-likelihood inverse reinforcement learning with finite-time guarantees

S Zeng, C Li, A Garcia, M Hong - Advances in Neural …, 2022 - proceedings.neurips.cc
Inverse reinforcement learning (IRL) aims to recover the reward function and the associated
optimal policy that best fits observed sequences of states and actions implemented by an …

LS-IQ: Implicit reward regularization for inverse reinforcement learning

F Al-Hafez, D Tateo, O Arenz, G Zhao… - arXiv preprint arXiv …, 2023 - arxiv.org
Recent methods for imitation learning directly learn a $ Q $-function using an implicit reward
formulation rather than an explicit reward function. However, these methods generally …

Relative entropy inverse reinforcement learning

A Boularias, J Kober, J Peters - Proceedings of the fourteenth …, 2011 - proceedings.mlr.press
We consider the problem of imitation learning where the examples, demonstrated by an
expert, cover only a small part of a large state space. Inverse Reinforcement Learning (IRL) …

Identifiability and generalizability in constrained inverse reinforcement learning

A Schlaginhaufen… - … Conference on Machine …, 2023 - proceedings.mlr.press
Two main challenges in Reinforcement Learning (RL) are designing appropriate reward
functions and ensuring the safety of the learned policy. To address these challenges, we …

Clare: Conservative model-based reward learning for offline inverse reinforcement learning

S Yue, G Wang, W Shao, Z Zhang, S Lin, J Ren… - arXiv preprint arXiv …, 2023 - arxiv.org
This work aims to tackle a major challenge in offline Inverse Reinforcement Learning (IRL),
namely the reward extrapolation error, where the learned reward function may fail to explain …

Opirl: Sample efficient off-policy inverse reinforcement learning via distribution matching

H Hoshino, K Ota, A Kanezaki… - … Conference on Robotics …, 2022 - ieeexplore.ieee.org
Inverse Reinforcement Learning (IRL) is attractive in scenarios where reward engineering
can be tedious. However, prior IRL algorithms use on-policy transitions, which require …

Hybrid inverse reinforcement learning

J Ren, G Swamy, ZS Wu, JA Bagnell… - arXiv preprint arXiv …, 2024 - arxiv.org
The inverse reinforcement learning approach to imitation learning is a double-edged sword.
On the one hand, it can enable learning from a smaller number of expert demonstrations …

Compatible reward inverse reinforcement learning

AM Metelli, M Pirotta, M Restelli - Advances in neural …, 2017 - proceedings.neurips.cc
Abstract Inverse Reinforcement Learning (IRL) is an effective approach to recover a reward
function that explains the behavior of an expert by observing a set of demonstrations. This …

Maximum likelihood constraint inference for inverse reinforcement learning

DRR Scobee, SS Sastry - arXiv preprint arXiv:1909.05477, 2019 - arxiv.org
While most approaches to the problem of Inverse Reinforcement Learning (IRL) focus on
estimating a reward function that best explains an expert agent's policy or demonstrated …

Active exploration for inverse reinforcement learning

D Lindner, A Krause… - Advances in Neural …, 2022 - proceedings.neurips.cc
Abstract Inverse Reinforcement Learning (IRL) is a powerful paradigm for inferring a reward
function from expert demonstrations. Many IRL algorithms require a known transition model …