A Schlaginhaufen… - … Conference on Machine …, 2023 - proceedings.mlr.press
Two main challenges in Reinforcement Learning (RL) are designing appropriate reward functions and ensuring the safety of the learned policy. To address these challenges, we …
S Liu, M Zhu - Advances in Neural Information Processing …, 2022 - proceedings.neurips.cc
This paper considers the problem of recovering the policies of multiple interacting experts by estimating their reward functions and constraints where the demonstration data of the …
F Jarboui, V Perchet - arXiv preprint arXiv:2106.05068, 2021 - arxiv.org
The objective of offline RL is to learn optimal policies when a fixed exploratory demonstrations data-set is available and sampling additional observations is impossible …
S Liu, Y Qing, S Xu, H Wu, J Zhang, J Cong… - arXiv preprint arXiv …, 2023 - arxiv.org
Inverse Reinforcement Learning (IRL) aims to reconstruct the reward function from expert demonstrations to facilitate policy learning, and has demonstrated its remarkable success in …
DS Han, H Kim, H Lee, JH Ryu… - Advances in Neural …, 2022 - proceedings.neurips.cc
Recently, adversarial imitation learning has shown a scalable reward acquisition method for inverse reinforcement learning (IRL) problems. However, estimated reward signals often …
One of the challenges in applying reinforcement learning in a complex real-world environment lies in providing the agent with a sufficiently detailed reward function. Any …
Inverse reinforcement learning (IRL) aims to infer a reward from expert demonstrations, motivated by the idea that the reward, rather than the policy, is the most succinct and …
F Jarboui, V Perchet - arXiv preprint arXiv:2105.11812, 2021 - arxiv.org
The gloabal objective of inverse Reinforcement Learning (IRL) is to estimate the unknown cost function of some MDP base on observed trajectories generated by (approximate) …