Inverse reinforcement learning (IRL) aims to recover the reward function and the associated optimal policy that best fits observed sequences of states and actions implemented by an …
Imitation learning practitioners have often noted that conditioning policies on previous actions leads to a dramatic divergence between" held out" error and performance of the …
D Jarrett, A Hüyük… - … Conference on Machine …, 2021 - proceedings.mlr.press
Decision analysis deals with modeling and enhancing decision processes. A principal challenge in improving behavior is in obtaining a transparent* description* of existing …
Imitation learning methods seek to learn from an expert either through behavioral cloning (BC) for the policy or inverse reinforcement learning (IRL) for the reward. Such methods …
L Viano, A Kamoutsi, G Neu… - Advances in Neural …, 2022 - proceedings.neurips.cc
This work develops new algorithms with rigorous efficiency guarantees for infinite horizon imitation learning (IL) with linear function approximation without restrictive coherence …
C Cundy, S Ermon - arXiv preprint arXiv:2306.05426, 2023 - arxiv.org
In many domains, autoregressive models can attain high likelihood on the task of predicting the next observation. However, this maximum-likelihood (MLE) objective does not …
Training multiple agents to coordinate is an essential problem with applications in robotics, game theory, economics, and social sciences. However, most existing Multi-Agent …
Learning from observation (LfO) aims to imitate experts by learning from state-only demonstrations without requiring action labels. Existing adversarial imitation learning …
In a competitive game scenario, a set of agents have to learn decisions that maximize their goals and minimize their adversaries' goals at the same time. Besides dealing with the …